Optimization is the core of machine learning. Training a model essentially means minimizing or maximizing an objective (loss) function. For example:
Suppose we want to minimize:
f(x) = (x - 3)2
The derivative is f'(x) = 2(x - 3). Gradient Descent update rule:
xnew = xold - η * f'(x)
where η is the learning rate.
Minimize f(x) = (x - 3)² using Gradient Descent.