Gradient Descent
Overall Progress
0%
Predict a continuous value with ลท = wx + b. Derive the MSE loss and compute one gradient descent step from scratch.
The optimization algorithm behind every trained ML model: iteratively follow the negative gradient to minimize a loss.
Binary classifier from scratch: sigmoid + cross-entropy loss + gradient update. The building block of softmax policies.