Deep Learning (MODULE-2)
Deep Learning (MODULE-2)
DEEP LEARNING
Module:2
Module:2
IMPROVING DEEP NEURAL NETWORKS
1. Mini-Batch gradient Descent
2. Exponential Weighted Averages
3. Gradient Descent with Momentum
4. RMSProp and Adam Optimization
5. Hyperparameter Tuning
6. Batch Normalization
7. Softmax Regression
8. Softmax Classifier
9. Deep Learning Frameworks
10. Data Augmentation
1.) Mini-Batch gradient Descent
Why Gradient Descent?
An algorithm to minimize a cost function by
optimizing its parameters.
We start with random guess and slowly move to
right or best answer.
Need – Parameter Optimization
Formula:
New value = old value - step size
Where,
step size = Learning rate x Slope.
1.) Mini-Batch gradient Descent
1.) Mini-Batch gradient Descent
Note:
If the learning rate is too high, you might jump
across the valley and end up on the other side,
possibly even higher up than you were before.
Substitute V98
2.) Exponentially Weighted Averages
Example:
2.) Exponentially Weighted Averages
2.) Exponentially Weighted Averages
Example:
3.) Gradient Descent with Momentum
Stochastic Gradient Descent / Batch Gradient Descent
In SGD /BGD:
3.) Gradient Descent with Momentum
In SGD /BGD:
3.) Gradient Descent with Momentum
In SGD /BGD:
3.) Gradient Descent with Momentum
Advantages:
(d) Robustness:
Adam optimization is relatively robust to
hyperparameter choices and works well across a
wide range of deep learning architectures.
5.) Hyperparameter Tuning
What is hyperparameter tuning?
5.) Hyperparameter Tuning
What is hyperparameter tuning?
Hyperparameter- Example:
5.) Hyperparameter Tuning
What is hyperparameter tuning?
5.) Hyperparameter Tuning
What is hyperparameter tuning?
When you’re training machine learning
models, each dataset and model needs a different
set of hyperparameters, which are a kind of
variable.
Total Model –Build and Tested: (10 Hyperparameter combination *10 Cross Validation = 100
5.) Hyperparameter Tuning (Regularization)
Let’s explore some more detailed explanations about the role of
Regularization:
1. Complexity Control: Regularization helps control model complexity by
preventing overfitting to training data, resulting in better generalization
to new data.