Sigmoid
Overall Progress
0%
ReLU, sigmoid, tanh, and softmax โ what they compute, when to use each, and why non-linearity is essential for deep networks.
Derivatives, chain rule, sigmoid and softmax โ with RL motivation and explained solutions.
Predict categories instead of numbers. Decision boundaries, sigmoid activation, and binary probability outputs.