Deep learning Module 1 Chapter 1
Deep learning Module 1 Chapter 1
MODULE 1
INTRODUCTION ON DEEP
LEARNING
OVERVIEW
• INTRODUCTION OF DEEP LEARNING
• DIFFERENTIATE BETWEEN AI VS ML VS DEEP LEARNING
• HISTORY OF DEEP LEARNING
• DEEP LEARNING MODEL
• HISTORICAL TRENDS IN DEEP LEARNING
• MACHINE LEARNING BASICS
• SUPERVISED LEARNING
• UNSUPERVISED LEARNING
• REINFORCEMENT LEARNING
• THE TASK T
• THE PERFORMANCE P
• THE EXPERIENCE E
Dr. Shiva Prasad KM Associate Professor CSE Department RYMEC Ballari M: 7899964163
9/25/2024
■ It's a powerful tool for solving complex problems in various domains, including image
recognition, natural language processing, and robotics.
■ Deep learning requires large amount of data to train and test the model which inturn
increases the time complexity and also requires high computational power.
AI VS ML VS DEEP LEARNING
■ Artificial Intelligence (AI): ■ Machine Learning (ML): ■ Deep Learning (DL):
– Definition: AI is the broad– Definition: ML is a subset of– Definition: DL is a subset of
field of creating intelligent AI that focuses on building ML that uses deep neural
machines that can mimic or algorithms that enable networks with multiple layers
simulate human intelligence. machines to learn from data to model complex data
– Example: Virtual assistants and improve their performance patterns.
like Siri and Alexa that can over time without explicit– Example: Self-driving cars
understand voice commands programming. that use deep learning models
and provide relevant responses– Example: A spam email filter to recognize road signs,
are AI systems. that learns from user pedestrians, and other vehicles
behaviour to identify spam in real time.
emails.
– 1943: The concept of artificial neurons was introduced by Warren McCulloch and Walter Pitts, who proposed a
mathematical model of how biological neurons might function, laying the groundwork for neural networks.
– 1958: Frank Rosenblatt developed the perceptron, an early model for binary classifiers that could learn simple
patterns. However, perceptron were limited and could not solve non-linear problems.
– Interest in neural networks declined during this period, partly due to limitations such as the inability to handle
non-linear problems. A famous book by Marvin Minsky and Seymour Papert in 1969 highlighted these
limitations.
– Back propagation (1986): A breakthrough came when Geoffrey Hinton, David Rumelhart, and Ronald
Williams rediscovered and popularized the back propagation algorithm, which allowed for training multi-
layer neural networks, helping to overcome previous challenges.
– With increased computing power and better algorithms, neural networks started to regain popularity.
Yann LeCun’s work on Convoluntional Neural Networks (CNNs), particularly for digit recognition,
showed the promise of deep networks in real-world tasks.
– 2006: Geoffrey Hinton introduced the concept of deep belief networks, which led to a resurgence of
interest in deep architectures. This marked the beginning of modern deep learning, where deeper
neural networks with more layers became feasible due to improved algorithms and hardware.
– 2012: Deep learning gained significant attention when Alex Krizhevsky, Ilya Sutskever, and
Geoffrey Hinton used CNNs to win the ImageNet competition by a large margin, revolutionizing
computer vision.
9/25/2024
– In 2014, Ian Good fellow introduced Generative Adversarial Networks (GANs), a groundbreaking
deep learning technique where two neural networks (a generator and a discriminator) compete
against each other to produce highly realistic data, such as images. GANs opened new possibilities in
areas like image synthesis, video generation, and even art.
– Deep learning is now at the core of advancements in various domains, including natural language
processing (e.g., Open AI's GPT series), autonomous driving, healthcare, and robotics. Transformers,
attention mechanisms, and transfer learning are key developments that continue to push the
boundaries of what deep learning can achieve.
– Hidden Layers: These are the intermediate layers that learn to extract features from the data. The
more hidden layers, the "deeper" the model. Hidden layers perform various transformations of the
input data using weights and activation functions.
– Output Layer: Produces the final result, like a classification label or regression value.
9/25/2024
■ There are three different types of Machine learning algorithms they are
■ Goal: The model learns to map input data to the correct output (label) and can generalize to predict outcomes on new, unseen
data.
■ Common Algorithms:
– Linear Regression: Used for predicting continuous values (e.g., predicting house prices).
– Logistic Regression: Used for binary classification problems (e.g., spam detection).
– Decision Trees: A tree-like model that makes decisions based on feature values.
– Support Vector Machines (SVM): Finds the optimal boundary (hyper plane) to separate data points into different classes.
– K-Nearest Neighbors (KNN): Classifies data points based on the majority label of the nearest neighbors.
■ Goal: The algorithm groups data or finds associations, without predefined labels.
■ Common Algorithms:
– Principal Component Analysis (PCA): Reduces the dimensionality of the data while preserving most
of the variance.
– Example: Grouping customers based on purchasing behaviour without prior knowledge of customer
categories.
9/25/2024
■ Goal: The agent learns to make a sequence of decisions to maximize cumulative rewards.
■ Common Algorithms:
– Q-Learning: A value-based approach where the agent learns the value of taking certain actions in
given states.
– Deep Q-Networks (DQN): Combines Q-learning with deep neural networks to handle complex
environments.
– Example: Teaching a robot to navigate a maze by rewarding it for reaching the goal and penalizing
it for hitting obstacles.
9/25/2024
THE TASK T
In the context of machine learning, Task T refers to the specific problem that a machine learning
model is designed to solve.
A task T can be defined as any problem where the goal is to map inputs to outputs in a way that
improves performance over time with experience.
■ Types of Tasks in Machine Learning
– Classification Tasks
– Regression Tasks
– Clustering Tasks
– Dimensionality Reduction Tasks
– Generative Tasks
– Sequence-to-Sequence Tasks
9/25/2024
THE PERFORMANCE P
■ Performance P in machine learning, is a measure of how well a model accomplishes its task T.
■ It is evaluated using various metrics, depending on the type of problem, and is influenced by factors
such as over fitting, under fitting, and generalization.
■ Achieving optimal performance requires balancing model complexity and ensuring that it generalizes
well to new data.
■ Key Aspects of Performance P
– Performance Metrics
– Training and Testing
– Over fitting and Under fitting
– Generalization
– Optimization and Learning
THE EXPERIENCE E
■ In machine learning, Experience (E) refers to the data and feedback a model uses to learn and
improve its performance on a task T.
■ Key Aspects of Experience E:
■ Training Data:
■ Learning from Data:
■ Feedback and Iteration
■ Generalization from Experience
■ Data Representation
■ Transfer Learning
9/25/2024
■ Linear Regression is a basic but essential algorithm for predicting continuous values. By learning
the best-fit line through the data, it captures the linear relationships between input features and the
target variable, making it useful for tasks like price prediction, demand forecasting, and more.
■ Example:
■ Problem: Predicting pizza price based on the size of the pizza.
– Input (X): The size of the pizza (e.g., diameter).
– Output (Y): The price of the pizza (continuous value).
■ Where:
– w is the weight (slope of the line), (How much Y Changes for the unit change in X)
– b is the bias (y-intercept of the line), (The value of Y when X=0)
– X is the input (size of the pizza)- Independent Variable ,
– Y is the predicted output (price of the Pizza )-Dependent Variable.
■ Training the Model
■ Objective: The goal of linear regression is to find the best values for w and b that minimize the
difference between the predicted prices and the actual prices in the training data.
9/25/2024
■ Gradient Descent: Linear regression uses optimization techniques like gradient descent to adjust
the weight w and bias b iteratively to minimize the loss function.
■ Probabilistic supervised learning in machine learning involves building models that predict the
probability distribution of target variables given input features.
■ This approach incorporates uncertainty into predictions, allowing the model to estimate not just a
single value but a range of possible outcomes with associated probabilities.
9/25/2024
■ Support Vector Machine (SVM) is a powerful supervised learning algorithm used for both
classification and regression tasks, although it is most commonly used for classification.
■ It works by finding the optimal boundary (or hyper plane) that separates different classes of data
points in a high-dimensional space.
■ Key Concepts of SVM
■ Objective:
– The goal of SVM is to find a hyperplane that maximally separates the classes of data points. In a
two-dimensional space, this hyperplane is a line, but in higher dimensions, it can be a plane or more
complex structure.
– SVM chooses the hyperplane with the maximum margin, which is the distance between the
hyperplane and the nearest data points from each class. These closest data points are called support
vectors.
9/25/2024
■ A natural approach to similarity based classification is K-Nearest Neighbor (K-NN) which is non-
parametric method used for both classification and regression problems.
■ KNN is one of the most basic yet essential classification algorithms in machine learning. It belongs
to the supervised learning domain and finds intense application in pattern recognition, data mining,
and intrusion detection.
■ It is simple and powerful non-parametric algorithm that predicts the category based on the test
instances according to the k training samples which are closer to the test instances and classifies it to
that category which has the largest probability.
■ There are various distance metrics used in K NN Algorithm they are Eucledian distance, Manhattan
Distance, Minkowski Distance
QUESTION BANK
MODULE 2
STAFF IN CHARGE
DR. SHIVA PRASAD KM
ASSOCIATE PROFESSOR
DEPARTMENT OF CSE
RYMEC BALLARI
Dr. Shiva Prasad KM Associate Professor CSE Department RYMEC Ballari M: 7899964163