Unit - V
Unit - V
Advanced Deep
Learning Topics
Dr. S Ruban
Dept of Software Technology
Introduction to
Deep Learning
• Working:
1. Input data is fed into the input layer.
2. Each neuron in a layer performs a weighted sum of its inputs and applies an
activation function (e.g., sigmoid, ReLU).
3. The output of one layer becomes the input to the next layer.
4. The network learns by adjusting the weights of connections between neurons
through backpropagation.
Representation Learning
Overview of Representation Learning
What it does Extracts meaningful patterns from raw data to create representations
that are easier for machines to understand and process
Why it's important Improves the performance of learning algorithms, especially when
dealing with high-dimensional data
How it works Replaces manual feature engineering, allowing machines to learn the
features and use them to perform a specific task
Approach:
3. Pre-train each layer using an unsupervised learning method (e.g., Restricted Boltzmann Machines (RBMs)).
Benefits:
How it works
• In transfer learning, a pre-trained model is fine-tuned for a new,
related task. The model's layers may be frozen or modified
depending on the target dataset. For example, if the target
dataset is small and distinct, most of the top layers may be
removed and new layers added.
Domain
Adaption
• Domain adaptation is a machine
learning technique that adjusts a
model trained on one domain to
work on a different domain. It's
used when there's a lack of
labelled data in the target
domain, but there's ample data in
the source domain.
• Adapting a model trained on a
source domain to perform well on
a target domain with different
data distributions.
• Example: Adapting a model
trained on English text to perform
well on Spanish text.
Domain Adaption Techniques
Applications:
Applications:
• Text summarization, Sentiment analysis,
Machine translation.
Distributed representations are a fundamental
concept in the field of machine learning and natural
Distribute language processing (NLP). They refer to a way of
representing data, typically words or phrases, as
d continuous vectors in a high-dimensional space.
representa
tion In distributed representations, also known as
embeddings, the idea is that the "meaning" or
"semantic content" of a data point is distributed
across multiple dimensions. For example, in NLP,
words with similar meanings are mapped to points
in the vector space that are close to each other.
Word Similarity: Measuring the semantic
similarity between words.
Application Text Classification: Categorizing
s of documents into predefined classes.
Distributed
Representa Machine Translation: Translating text
tions from one language to another.
Information Retrieval: Finding relevant
documents in response to a query.
Sentiment Analysis: Determining the
sentiment expressed in a piece of text.
Structured Probabilistic
Models for Deep Learning
What is a structured
probabilistic model
A structured probabilistic model is a way of describing a
probability distribution with graphs.
Describe Types:
Model • Bayesian Networks (Directed Acyclic Graphs).
• Markov Random Fields (Undirected Graphs).
Structure
Applications:
• Probabilistic inference, Speech recognition,
Gene network analysis.
Why Sampling?
• Allows inference when exact computations are
Sampling intractable.
from Techniques:
Graphical • Importance Sampling.
• Rejection Sampling.
Models • Gibbs Sampling (explained in later slides).
Challenges:
• Computational complexity, convergence
issues.
Monte Carlo Methods
Markov Chain Monte Carlo
Methods (MCMC)
A class of algorithms for sampling from probability
Definition: distributions based on constructing a Markov chain.
Directed Generative
Boltzmann Deep Belief
Generative Stochastic
Machines Networks
Nets Networks
Boltzmann machine
Definition: Advantages:
Generative models that directly define the High-quality samples.
likelihood function. Direct modeling of dependencies.
Example: PixelRNN, PixelCNN.
Directive Generative Nets (Conditional
Generative Adversarial Networks (cGANs))
• https://towardsdatascience.com/gibbs-sampling-8e4844560ae5
• https://www.sciencedirect.com/topics/computer-science/boltzmann-
machine
• https://cedar.buffalo.edu/~srihari/CSE676/20.10.1-DirectedGenNets.p
df
Thank You