OPTIMIZATION _Lecture4_
OPTIMIZATION _Lecture4_
[AcSIR-01-ES-AD-009]
Presented by:
Dr. R. S. Bisht, Pr. Scientist,
[email protected]
CSIR-CBRI, Roorkee, India
Gradient Descent Algorithm
• Gradient Descent is a fundamental optimization algorithm used to
minimize functions by iteratively moving toward the function’s local
minimum. It is widely used in machine learning, especially for training
models by minimizing loss functions.
• The main idea is to update the variables or parameters of the function
in the direction of the steepest descent (i.e., the negative gradient),
which is the direction that reduces the function's value the fastest.
Gradient Descent Algorithm
• Gradient Calculation (Derivative)
• Update Rule:
• Learning Rate The learning rate α controls how fast we move towards
the minimum.
• If α is too small, convergence will be slow.
• If α is too large, we may overshoot the minimum and fail to converge.
• Convergence The algorithm stops when the updates to the
parameters become very small, meaning the cost function is
minimized, or after a fixed number of iterations.
gradient descent method
Mathematical Formulation:
Disadvantages:
• Sensitive to the choice of the learning rate α. If the learning rate is too
large, it may overshoot; if too small, convergence can be slow.