Chapter 01 Notes
Chapter 01 Notes
Artificial Intelligence
‣ Artificial intelligence (AI) is the ability of a computer program or a machine to think and
learn. It is also a field of study that tries to make computers "smart." They work on their
own without being encoded with commands. John McCarthy came up with the name
"artificial intelligence" in 1955.
‣ These processes include learning (the acquisition of information and rules for using the
information), reasoning (using rules to reach approximate or definite conclusions), and
self-correction.
Types of AI
AI can be categorized as either weak or strong. Weak AI, also known as narrow AI, is an AI
system that is designed and trained for a particular task. Virtual personal assistants, such as
Apple's Siri, are a form of weak AI. Strong AI, also known as artificial general intelligence, is an
AI system with generalized human cognitive abilities. When presented with an unfamiliar task,
a strong AI system is able to find a solution without human intervention.
Machine learning is a field of artificial intelligence and the study of computer algorithms that
automatically improve through examples and experiences.
‣ “A field of research that develops algorithms that allow machines to learn from
data and execute actions that are not explicitly specified by code“ - Arthur
Samuel, 1959
‣ “A computer program is said to learn from experience E with respect to some
class of tasks T and performance measure P if its performance at tasks in T, as
measured by P, improved with experience E.“ - Tom Mitchell, 1977
Example: Spam filter is one of the machine learning programs that use samples of spam and
general mails to classify mails that are designated as spam by the user.
Basic terms
‣ Training set: samples used by the system for learning
‣ Training instance (or sample: each training data,
‣ Task T classifies if the new mail is spam
‣ Experience E is the training data
‣ Performance measure P should be directly defined by the user, and performance
measure is also called accuracy which is often used for classification tasks.
Example: The task T is to classify dogs and cats, and the performance P represents a measure of
classifying dogs and cats. E can be said to be “learning” if the performance of classifying dogs
and cats gradually improves through experience, or data (10,000 photos).
In standard machine learning methodology, a user prepares for training data set and extracts
relevant features such as edges and corners to train the model.
Then, the model references the features to classify new objects. This is an example of object
recognition and how computers distinguish cats and dogs with a significant level of accuracy.
Deep Learning
Workflow
On the other hand, in deep learning, a user does not need to extract features to feed images
into the model, and rather inputs images themselves directly into the deep learning model.
Deep learning model basically consist of ARTIFICIAL NEURAL NETWORK algorithm which learns
by itself to predict the object. Deep learning is a subtype of machine learning and usually more
complex than machine learning techniques. The key difference between machine learning and
deep learning is whether the model uses ARTIFICIAL NEURAL NETWORK or not.
In supervised learning, the machine is trained on a set of labeled data, which means that the
input data is paired with the desired output. The machine then learns to predict the output for
new input data. Supervised learning is often used for tasks such as classification, regression,
and object detection.
Key Points:
• Supervised learning allows collecting data and produces data output from previous
experiences.
• Helps to optimize performance criteria with the help of experience.
• Supervised machine learning helps to solve various types of real-world computation
problems.
• It performs classification and regression tasks.
• It allows estimating or mapping the result to a new sample.
• We have complete control over choosing the number of classes we want in the training
data.
• Disadvantages of Supervised learning
• Classifying big data can be challenging.
• Training for supervised learning needs a lot of computation time. So, it requires a lot of
time.
• Supervised learning cannot handle all complex tasks in Machine Learning.
• Computation time is vast for supervised learning.
• It requires a labelled data set.
• It requires a training process.
In unsupervised learning, the machine is trained on a set of unlabeled data, which means that
the input data is not paired with the desired output. The machine then learns to find patterns
and relationships in the data. Unsupervised learning is often used for tasks such as clustering,
dimensionality reduction, and anomaly detection.
Key Points
• Irrelevant Features
‣ The system can only learn when the features and those related to the training
data are sufficient.
‣ A vital element of a successful machine learning project is finding good features
to use for training, known as feature engineering.
‣ Feature selection: Choose the most useful feature for training
‣ Feature extraction: Combining features creates a more useful feature (dimension
reduction)
Underfitting and Overfitting
When we talk about the Machine Learning model, we actually talk about how well it performs
and its accuracy which is known as prediction errors. Let us consider that we are designing a
machine learning model. A model is said to be a good machine learning model if it generalizes
any new input data from the problem domain in a proper way. This helps us to make
predictions about future data, that the data model has never seen. Now, suppose we want to
check how well our machine learning model learns and generalizes to the new data. For that,
we have overfitting and underfitting, which are majorly responsible for the poor performances
of the machine learning algorithms.