Deep Learning
Deep Learning
Answer- Deep neural networks are harder to train than shallow ones because they have
many layers, making it tough to adjust all the connections effectively. This complexity
leads to problems like vanishing or exploding gradients and increases the risk of over
fitting, where the network learns the training data too well.
2. What are some common issues encountered during training deep neural networks?
Answer- Common issues encountered during training deep neural networks include:
Answer- Greedy layer wise training in deep learning involves training one layer at a time,
gradually building up the network. Each layer is trained independently, focusing on
learning simpler representations before combining them in deeper layers. This approach
simplifies the training process by breaking down the problem into smaller, more
manageable parts, potentially aiding convergence and overcoming issues such as
vanishing or exploding gradients.
4. What are the advantages and disadvantages of greedy layer wise training?
Answer- Optimization in the context of training deep neural networks refers to the process of
finding the best set of parameters (weights and biases) for the network to minimize the
difference between the actual outputs and the predicted outputs. This involves adjusting the
parameters iteratively using optimization algorithms to improve the network's performance in
tasks such as classification, regression, or generation. The goal is to find the optimal
configuration that reduces errors and enhances the network's ability to generalize to unseen
data.
8. What are the computational challenges associated with using second order methods?
1. Preventing Over fitting: Deep neural networks have a large number of parameters,
making them prone to over fitting, especially when trained on limited data.
Regularization helps mitigate over fitting by penalizing complex models that fit the
training data too closely, thereby improving the model's ability to generalize to new,
unseen data.
2. Improving Generalization: By encouraging simpler model configurations,
regularization helps prevent the network from memorizing noise or irrelevant patterns
in the training data, leading to better generalization performance on unseen data.
3. Controlling Model Complexity: Regularization techniques introduce constraints on
the network's parameters, such as weight decay or dropout, which control the
complexity of the model. This helps prevent the network from becoming too complex
and ensures that it learns meaningful features that are relevant to the task at hand.
10. Explain the concept of batch normalization and its role in training deep neural
networks.
Answer- Batch normalization is a technique used in training deep neural networks to stabilize
and speed up the learning process. It works by normalizing the activations of each layer
within a mini-batch, transforming the inputs to have a mean of zero and a standard deviation
of one. This normalization is applied independently to each feature dimension.
Batch normalization helps in training deep neural networks by addressing several key
challenges:
1. Internal Covariate Shift: By normalizing the activations within each mini-batch, batch
normalization reduces the internal covariate shift, making the training process more stable
and allowing for higher learning rates.
2. Addressing Vanishing/Exploding Gradients: Batch normalization helps mitigate the
vanishing and exploding gradient problems by stabilizing the activations throughout the
network, enabling smoother and more consistent gradient flow during back propagation.
3. Regularization: Batch normalization acts as a form of regularization by adding noise to
the activations, similar to dropout, which helps prevent over fitting and improves the
generalization performance of the network.