Deep Learning
Deep Learning
and____________
Ans) The two initial waves of Deep Learning are Neural Networks (Connectionism) in the 1980s-
1990s and Deep Learning resurgence in the 2000s-2010s.
2) Why should one not view deep learning as an attempt to simulate the brain?
While deep learning (DL) is inspired by the structure and function of biological neural networks, it
should not be mistaken for an actual brain simulation. Here’s why:
o Biological neurons are complex electrochemical systems, while artificial neurons use
simple mathematical functions (e.g., weighted sums and activation functions).
o The brain operates with billions of interconnected neurons with dynamic synapses,
whereas deep learning models have static and predefined connections.
o The human brain learns from few examples (few-shot learning), while deep learning
often requires millions of labeled data points.
o The brain adapts in real-time, whereas deep learning models rely on stochastic
gradient descent (SGD) and batch processing.
o Deep learning models have fixed layers and weight structures once trained.
o A human brain operates on about 20 watts of power, while training deep networks
requires massive GPUs consuming kilowatts of energy.
o The brain processes information in massively parallel ways, while ANNs rely on
sequential matrix computations optimized for GPUs.
o They excel at specific tasks (e.g., image classification, NLP) but lack common sense
reasoning and cross-domain adaptability.
o A trained deep learning model cannot generalize beyond its dataset, while humans
can apply knowledge flexibly across different domains.
The brain uses spiking neurons, neuromodulation, and chemical signaling, which are absent
in artificial networks.
Concepts like emotion, consciousness, and intuition are not represented in deep learning
models.
3). Suppose that ŷ 0.9 and y = 1. What is the value of the logistic loss?
4)'a' is a (4,3) and 'b' is a (1,3) matrix. If we perform c = a * b, what will be the dimensions of
matrix 'c'?
is only valid if n=pn = p, and the resulting matrix will have dimensions:
(m×q)(m \times q)
Given:
Checking Compatibility:
For c=a∗bc = a * b, the number of columns in aa (which is 3) must match the number of
rows in bb (1).
Since 3 ≠ 1, this multiplication is not valid using standard matrix multiplication rules.
Alternative Interpretations:
. a) What will happen if we do not use any activation function(s) in a neural network?
If a neural network does not use any activation functions, it essentially reduces to a simple linear
model. Here's why:
1. Loss of Non-Linearity
Without activation functions, every layer in the network will perform only linear
transformations (i.e., matrix multiplication and addition).
Stacking multiple linear layers still results in a linear function, meaning the network cannot
learn complex patterns.
Example:
o This is still just a single linear transformation, meaning a deep network would be no
more powerful than a single-layer perceptron.
Most real-world problems (e.g., image recognition, language processing) require non-linear
decision boundaries.
Without activation functions, the model can only learn linear mappings, making it ineffective
for tasks like image classification or speech recognition.
3. No Universal Approximation
The Universal Approximation Theorem states that a neural network with at least one non-
linear activation function can approximate any function.
Without them, gradients remain constant across layers, making it difficult for deep networks
to learn efficiently.
Conclusion
Bottom Line: Activation functions introduce non-linearity, which is essential for deep learning.
Without them, a neural network is just a linear model.