0% found this document useful (0 votes)
45 views276 pages

Latest AI Book + Sample Paper 2025 PDF

The Class 10 Facilitator Handbook on Artificial Intelligence, curated with support from Intel, aims to prepare students for a future driven by AI by introducing foundational concepts, applications, and ethical considerations. It emphasizes the importance of skills-based education and project work to make students 'AI ready' and extends the AI curriculum from Class IX to Class X. The document outlines various topics, including decision-making, machine learning, data sciences, computer vision, and natural language processing, to enhance students' understanding of AI technologies.

Uploaded by

shivamlal13358
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
0% found this document useful (0 votes)
45 views276 pages

Latest AI Book + Sample Paper 2025 PDF

The Class 10 Facilitator Handbook on Artificial Intelligence, curated with support from Intel, aims to prepare students for a future driven by AI by introducing foundational concepts, applications, and ethical considerations. It emphasizes the importance of skills-based education and project work to make students 'AI ready' and extends the AI curriculum from Class IX to Class X. The document outlines various topics, including decision-making, machine learning, data sciences, computer vision, and natural language processing, to enhance students' understanding of AI technologies.

Uploaded by

shivamlal13358
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
You are on page 1/ 276

CLASS 10

FACILITATOR HANDBOOK
Artificial
Intelligence
Curriculum
Class 10
Curated with support from Intel®
Acknowledgements

Patrons:
• Sh. Ramesh Pokhriyal 'Nishank', Minister of Human Resource Development, Government
of India
• Sh. Dhotre Sanjay Shamrao, Minister of State for Human Resource Development,
Government of India

Human Resource Development, Government of India Advisory, Editorial and Creative Inputs:
• Ms. Anita Karwal, IAS, Chairperson, Central Board of Secondary Education
• Ms. Shweta Khurana, Director, Programs, Partnerships and Policy Group, Intel India

Guidance and Support:


• Sh. Anurag Tripathi, IRPS, Secretary, Central Board of Secondary Education
• Dr. Joseph Emmanuel, Director (Academics), Central Board of Secondary Education
• Dr. Biswajit Saha, Director (Skill Education & Training), Central Board of Secondary Education

Education Value adder, Curator and Coordinator:


• Sh. Ravinder Pal Singh, Joint Secretary, Department of Skill Education, Central Board of
Secondary Education

Content Curation Team:


• Ms. Sharon E. Kumar, Innovation and Education Consultant, Intel AI4Youth Program
• Ms. Ambika Saxena, Intel AI For Youth Coach
• Mr. Bhavik Khurana, Intel AI For Youth Coach
• Mr. Akshay Chawla, Intel AI For Youth Coach
• Mr. Shivam Agrawal, Intel AI For Youth Coach

Feedback By:
• Ms. Neelam Roy, ITL Public School, Delhi
• Ms. Mehreen Shamim, TGT, DPS Bangalore East, Bengaluru
• Ms. Saswati Sarangi, PGT Computer Science, RCIS Kalyan Nagar, Bengaluru
• Ms. Aayushi Agrawal, Salwan Girls School, Delhi
• Ms. Isha, HOD Computer Science, Salwan Public School, Delhi

Special Thanks To:

• Ms. Indu Khetrapal, Principal, Salwan Public School, Delhi


• Ms. Rekha Vinod, Principal, RCIS Kalyan Nagar, Bengaluru
• Ms. Manilla Carvalho, Principal, Delhi Public School – Bangalore East, Bengaluru
• Ms. Sudha Acharya, Principal, ITL Public School, Delhi
• Ms. Puneet Sardana, Vice-Principal, Salwan Girls School, Delhi
About the book
Artificial Intelligence (AI) is being widely recognised to be the power that will fuel the future global
digital economy. AI in the past few years has gained geo-strategic importance and a large number of
countries are striving hard to stay ahead with their policy initiatives to get their country ready.

India’s own AI strategy identifies AI as a n opportunity and solution provider for inclusive economic
growth and social development. The report also identifies the importance of skills-based education
(as opposed to knowledge intensive education), and the value of project related work in order to
“effectively harness the potential of AI in a sustainable manner” to make India’s next generation ‘AI
ready’.

As a beginning in this direction, CBSE introduced Artificial Intelligence as an optional subject at Class
IX from the Session 2019-2020 onwards. Also, to enhance the multidisciplinary approach in teaching-
learning so as to sensitize the new generation, it was decided that schools may start AI “Inspire
Module” of 12 hours at class VIII itself. CBSE has extended this subject to class X as well from the
Session 2020-2021.

CBSE is already offering various skill subjects at secondary and senior secondary level to upgrade the
skills and proficiency of the young generation and also to provide them awareness to explore various
career options. Ai secondary level, a skill subject may be offered as additional sixth subject along with
the existing five compulsory subjects.

CBSE acknowledges the initiative by Intel India in curating this Facilitator Handbook, the AI training
video and managing the subsequent trainings of trainers on the Artificial Intelligence Curriculum.

The aim is to strive together to make our students future ready and help them work on incorporating
Artificial Intelligence to improve their learning experience.
Table of Contents

Introduction to AI: Foundational Concepts ........................................................................................ 9


What is Intelligence? ....................................................................................................................... 9
Decision Making ............................................................................................................................ 12
How do you make decisions?.................................................................................................... 12
Make Your Choices! .................................................................................................................. 12
What is Artificial Intelligence? ...................................................................................................... 14
How do machines become Artificially Intelligent? ....................................................................... 14
Applications of Artificial Intelligence around us ........................................................................... 15
What is not AI? .............................................................................................................................. 16
Introduction to AI: Basics of AI ......................................................................................................... 18
AI, ML & DL.................................................................................................................................... 20
Introduction to AI Domains........................................................................................................... 21
Data Sciences ............................................................................................................................ 21
Computer Vision ....................................................................................................................... 21
Natural Language Processing .................................................................................................... 22
AI Ethics ......................................................................................................................................... 23
Moral Issues: Self-Driving Cars.................................................................................................. 23
Data Privacy .............................................................................................................................. 24
AI Bias ........................................................................................................................................ 26
AI Access ................................................................................................................................... 27
AI Project Cycle ................................................................................................................................. 29
Introduction .................................................................................................................................. 29
Problem Scoping ....................................................................................................................... 30
Data Acquisition ........................................................................................................................ 34
Data Exploration ....................................................................................................................... 35
Modelling .................................................................................................................................. 36
Learning Based Approach ......................................................................................................... 37
Evaluation ................................................................................................................................. 39
Neural Networks ........................................................................................................................... 40
Advance Python ................................................................................................................................ 42
Recap............................................................................................................................................. 42
Recap 1: Jupyter Notebook ....................................................................................................... 42
Introduction to Virtual Environments ....................................................................................... 43
Recap 2: Introduction to Python ............................................................................................... 47
Applications of Python .............................................................................................................. 48
Recap 3: Python Basics.............................................................................................................. 48
Python Packages ........................................................................................................................... 52
Data Sciences .................................................................................................................................... 54
Introduction .................................................................................................................................. 54
Applications of Data Sciences ....................................................................................................... 55
Getting Started.............................................................................................................................. 57
Revisiting AI Project Cycle ......................................................................................................... 57
Data Collection .......................................................................................................................... 62
Data Access ............................................................................................................................... 63
Basic Statistics with Python .......................................................................................................... 66
Data Visualisation ......................................................................................................................... 67
Data Sciences: Classification Model .............................................................................................. 71
Personality Prediction ............................................................................................................... 71
K-Nearest Neighbour: Explained ............................................................................................... 72
Computer Vision ............................................................................................................................... 75
Introduction .................................................................................................................................. 75
Applications of Computer Vision .................................................................................................. 76
Computer Vision: Getting Started................................................................................................. 78
Computer Vision Tasks .................................................................................................................. 78
Classification ............................................................................................................................. 78
Classification + Localisation ...................................................................................................... 78
Object Detection ....................................................................................................................... 78
Instance Segmentation ............................................................................................................. 78
Basics of Images ........................................................................................................................ 79
Basics of Pixels .......................................................................................................................... 79
Image Features.............................................................................................................................. 84
Introduction to OpenCV ................................................................................................................ 85
Convolution ................................................................................................................................... 86
Convolution : Explained ............................................................................................................ 88
Convolution Neural Networks (CNN) ............................................................................................ 91
Introduction .............................................................................................................................. 91
What is a Convolutional Neural Network ?............................................................................... 92
Convolution Layer ..................................................................................................................... 93
Rectified Linear Unit Function .................................................................................................. 94
Pooling Layer ............................................................................................................................. 95
Fully Connected Layer ............................................................................................................... 96
Natural Language Processing ............................................................................................................ 99
Introduction .................................................................................................................................. 99
Applications of Natural Language Processing ............................................................................. 100
Natural Language Processing: Getting Started ........................................................................... 101
Revisiting the AI Project Cycle................................................................................................. 101
Chatbots ...................................................................................................................................... 104
Human Language VS Computer Language .................................................................................. 105
Arrangement of the words and meaning................................................................................ 106
Multiple Meanings of a word .................................................................................................. 107
Perfect Syntax, no Meaning .................................................................................................... 107
Data Processing ........................................................................................................................... 108
Text Normalisation .................................................................................................................. 108
Bag of Words ........................................................................................................................... 112
TFIDF: Term Frequency & Inverse Document Frequency ....................................................... 114
Applications of TFIDF .................................................................................................................. 118
DIY – Do It Yourself! .................................................................................................................... 118
Evaluation ....................................................................................................................................... 119
Introduction ................................................................................................................................ 119
What is evaluation? .................................................................................................................... 119
Model Evaluation Terminologies ................................................................................................ 119
The Scenario............................................................................................................................ 119
Confusion matrix ......................................................................................................................... 122
Evaluation Methods .................................................................................................................... 123
Accuracy .................................................................................................................................. 123
Precision .................................................................................................................................. 124
Recall ....................................................................................................................................... 125
Which Metric is Important? .................................................................................................... 126
F1 Score ................................................................................................................................... 127
Introduction to AI: Foundational Concepts
What is Intelligence?
Humans have been developing machines which can make their lives easier. Machines are made with
an intent of accomplishing tasks which are either too tedious for humans or are time consuming.
Hence, machines help us by working for us, thereby sharing our load and making it easier for us to
fulfil such goals.

Life without machines today is unimaginable, and because of this, humans have been putting efforts
into making them even more sophisticated and smart. As a result, we are surrounded by smart devices
and gadgets like smartphones, smartwatches, smart TV, etc. But what makes them smart?

For example, how is a smartphone today different from the telephones we had in the last century?

* Images shown here are the property of individual organisations and are used here for reference purpose only.
Today’s phones can do much more than just call-up
people. They can help us in navigating, recommend
which songs we should listen to or which movies we
should watch according to our own likes and
dislikes. Our phones can help us connect with like-
minded people, make our selfies fun with face
filters, help us maintain a record of our health and
fitness and a lot more. These drastic technological
advancements lead us to recognize one key
concept: the concept of Artificial Intelligence.

What is Artificial Intelligence anyway? Well, the answer lies in the term itself. If we break up this term
up, we get the words “Artificial” and “Intelligence”. Artificial is something which is man-made, which
does not occur naturally. But what about Intelligence, how do we define that?

Humans are said to be an intelligent species, so what is it that makes us intelligent?

According to researchers, intelligence is the ‘ability to perceive or infer information, and to retain it as
knowledge to be applied towards adaptive behaviours within an environment or context.’

If we try to define intelligence with the help of its traits, these are the abilities that are involved in
intelligence:

* Images shown here are the property of individual organisations and are used here for reference purpose only.
Let us define each term mentioned above to get a proper understanding:

Mathematical •A person's ability to regulate, measure, and understand numerical


Logical Reasoning symbols, abstraction and logic.

Linguistic •Language processing skills both in terms of understanding or


Intelligence implementation in writing or verbally.

Spatial Visual •It is defined as the ability to perceive the visual world and the
Intelligence relationship of one object to another.

Kineasthetic •Ability that is related to how a person uses his limbs in a skilled
Intelligence manilr.

Musical •As the name suggests, this intelligence is about a person's ability to
Intelligence recognize and create sounds, rhythms, and sound patterns.

Intrapersonal •Describes how high the level of self-awareness someone has is.
Intelligence Starting from realizing weakness, strength, to his own feelings.

Existential •An additional category of intelligence relating to religious and


Intelligence spiritual awareness.

Naturalist •An additional category of intelligence relating to the


Intelligence ability to process information on the environment around us.

Interpersonal •Interpersonal intelligence is the ability to communicate with others


intelligence by understanding other people's feelings & influence of the person.

But even though one is more skilled in intelligence than the other, it should be noted that in fact all humans have all 9 of
these intelligences only at different levels. One might be an expert at painting, while the other might be an expert in
mathematical calculations. One is a musician, the other is an expert dancer.
In other words, we may define intelligence as:

• Ability to interact with the real world


o To perceive, understand and act
▪ Example: Speech Recognition – Understanding and synthesis
▪ Example: Image Recognition
▪ Example: Ability to take action: to have an effect
• Reasoning and planning
o Modelling the external world, given input
▪ Solving new problems, planning and making decisions
▪ Ability to deal with unexpected problems, uncertainties
• Learning and adaptation
o Continuous learning and adapting graph
▪ Our internal models are always being updated
▪ Example: Baby learning to categorize and recognise animals

For example, if someone starts talking to us, we know how to keep the conversation going. We can
understand what people mean and can reply in the same way. When we are hungry, we can come up
with various options on what to eat depending upon the food we have at our homes. When we read
something, we are able to understand its meaning and answer anything regarding it.

While understanding the term intelligence, it must be noticed that decision making comprises of a
crucial part of intelligence. Let us delve deeper into it.

Decision Making

You’re trapped. All the doors seem to have started shrinking and only one of them leads you out.
Which door would you pick?

How do you make decisions?


The basis of decision making depends upon the availability of information and how we experience and
understand it. For the purposes of this article, ‘information’ includes our past experience, intuition,
knowledge, and self-awareness.

We can’t make “good” decisions without information because then we have to deal with unknown
factors and face uncertainty, which leads us to make wild guesses, flipping coins, or rolling a dice.
Having knowledge, experience, or insights given a certain situation, helps us visualize what the
outcomes could be. and how we can achieve/avoid those outcomes.

Make Your Choices!

Scenario 1
You are locked inside a room with 3 doors to move out of the locked room and you need to find a safe
door to get your way out. Behind the 1st door is a lake with a deadly shark. The 2nd door has a mad
psychopath ready to kill with a weapon and the third one has a lion that has not eaten since the last 2
months.

* Images shown here are the property of individual organisations and are used here for reference purpose only.
Which door would you choose? and Why?

__________________________________________________________________________________
__________________________________________________________________________________
__________________________________________________________________________________
__________________________________________________________________________________
__________________________________________________________________________________

The answer is gate number 3. The reason being that since the lion has not eaten for 2 months, he
wouldn't have survived till now and would already be dead . This makes going out from gate 3 the
correct option.

Scenario 2
Aarti invited four of her friends to her House.. They hadn't seen each other in a long time, so they
chatted all night long and had a good time. In the morning, two of the friends Aarti had invited, died.
The police arrived at the house and found that both the friends were poisoned and that the poison
was in the strawberry pie. The three surviving friends told the police that they hadn't eaten the pie.
The police asked," Why didn’t you eat the pie ?". Shiv said, " I am allergic to strawberries.". Seema
said, " I am on a diet." And Aarti said, "I ate too many strawberries while cooking the pie, I just didn't
want anymore."

The policemen looked at the pictures of the party and immediately identified the murderer.

* Images shown here are the property of individual organisations and are used here for reference purpose only.
Look at the picture and identify who is the murderer? Also state why do you think this is the murderer?

__________________________________________________________________________________
__________________________________________________________________________________
__________________________________________________________________________________
__________________________________________________________________________________
__________________________________________________________________________________

The answer is Seema, can you guess how the police could tell? It’s because she said she is on a diet
and in the picture, she is eating a burger and fries which means she lied.

The above scenarios show that it’s the information which helps humans take good decisions.

What is Artificial Intelligence?


When a machine possesses the ability to mimic human
traits, i.e., make decisions, predict the future, learn and
improve on its own, it is said to have artificial
intelligence.
In other words, you can say that a machine is artificially
intelligent when it can accomplish tasks by itself -
collect data, understand it, analyse it, learn from it, and
improve it. You will get to know more about it in the
next unit.
But, what makes a machine intelligent?

How do machines become Artificially Intelligent?


Humans become more and more intelligent with time as they gain experiences during their lives.

* Images shown here are the property of individual organisations and are used here for reference purpose only.
For example, in elementary school, we learn about alphabets and eventually we move ahead to
making words with them. As we grow, we become more and more fluent in the language as we keep
learning new words and use them in our conversations.

Another example is how we learn walking.


Initially a baby struggles to walk. He takes
help from others while learning how to
walk and once he knows it, he keeps on
upgrading it by learning how to run, jump,
etc.

Similarly, machines also become


intelligent once they are trained with
some information which helps them
achieve their tasks. AI machines also keep
updating their knowledge to optimise
their output.

Applications of Artificial Intelligence around us


Whether we notice it or not, we are surrounded by machines that work on AI. They are becoming a
crucial part of our everyday life and provide us with an ease of having even some of the most
complicated and time-consuming tasks being done at the touch of a button or by the simple use of a
sensor.

Every now and then, we surf the internet for things on Google
without realizing how efficiently Google always responds to us
with accurate answers. Not only does it come up with results
to our search in a matter of seconds, it also suggests and auto-
corrects our typed sentences.

We nowadays have pocket assistants that can


do a lot of tasks at just one command. Alexa,
Google Assistant, Cortana, Siri are some very
common examples of the voice assistants which
are a major part of our digital devices.

To help us navigate to places, apps like UBER and Google Maps come in haman.
Thus, one no longer needs to stop repeatedly to ask for directions.

AI has completely enhanced the gaming experience for its users.


A lot of games nowadays are backed up with AI which helps in
enhancing the graphics, come up with new difficulty levels,
encourage gamers, etc.

* Images shown here are the property of individual organisations and are used here for reference purpose only.
AI has not only made our lives easier but has also been
taking care of our habits, likes, and dislikes. This is why
platforms like Netflix, Amazon, Spotify, YouTube etc.
show us recommendations on the basis of what we
like.

Well, the recommendations are not just limited to


our preferences, they even cater to our needs of
connecting with friends on social media platforms
with apps like Facebook and Instagram. They also
send us customized notifications about our online
shopping details, auto-create playlists according
to our requests and so on. Taking selfies was never
this fun as Snapchat filters make them look so
cool.

This isn’t all. AI is also being used to monitor


our health. A lot of chatbots and other health
apps are available, which continuously
monitor the physical and mental health of its
users.

These applications are not limited to smart devices but


also vary to humanoids like Sophia, the very first
humanoid robot sophisticated enough to get
citizenship, biometric security systems like the face
locks we have in our phones, real-time language
translators, weather forecasts, and whatnot! This list is
huge, and this module will go on forever if we keep
tabulating them. So, take some time, discuss with a
friend and identify more and more AI applications
around you!

What is not AI?


Since we have a lot of different technologies which exist around us in today’s time, it is very
common for us to misunderstand any other technology as AI. That is why, we need to have a
clear distinction between what is AI and what is not.
As we discussed earlier, any machine that has been trained with data and can make
decisions/predictions on its own can be termed as AI. Here, the term ‘training’ is important.

A fully automatic washing machine can work on its own, but it requires human
intervention to select the parameters of washing and to do the necessary preparation for
it to function correctly before each wash, which makes it an example of automation, not
AI.

* Images shown here are the property of individual organisations and are used here for reference purpose only.
An air conditioner can be turned on and off remotely with the help of internet but still
needs a human touch. This is an example of Internet of Things (IoT). Also, every now and
then we get to know about robots which might follow a path or maybe can avoid
obstacles but need to be primed accordingly each time.

We also get to see a lot of projects which can automate our surroundings with the
help of sensors. Here too, since the bot or the automation machine is not trained with
any data, it does not count as AI.

Also, it would be valid to say that not all the devices which are termed as "smart" are AI-enabled. For
example, a TV does not become AI-enabled if it is a smart one, it gets the power of AI when it is able
to think and process on its own.

Just as humans learn how to walk and then improve this skill with the help of their experiences, an AI
machine too gets trained first on the training data and then optimises itself according to its own
experiences which makes AI different from any other technological device/machine.

But well, surely these other technologies too can be integrated with AI to provide the users with a
much better and immersive experience!

Robotics and AI can definitely open the doors to humanoids and self-driving cars, AI when merged
with Internet of things can give rise to cloud computing of data and remote access of AI tools,
automation along with AI can help in achieving voice automated homes and so on. Such integrations
can help us get the best of both worlds!

* Images shown here are the property of individual organisations and are used here for reference purpose only.
Introduction to AI: Basics of AI
As discussed in the last chapter, Artificial Intelligence has always been a term which intrigues people
all over the world. Various organisations have coined their own versions of defining Artificial
Intelligence. Some of them are mentioned below:

NITI Aayog: National Strategy for Artificial Intelligence


AI refers to the ability of machines to perform cognitive tasks like thinking, perceiving, learning,
problem solving and decision making. Initially conceived as a technology that could mimic human
intelligence, AI has evolved in ways that far exceed its original conception. With incredible advances
made in data collection, processing and computation power, intelligent systems can now be deployed
to take over a variety of tasks, enable connectivity and enhance productivity.

World Economic Forum


Artificial intelligence (AI) is the software engine that drives the Fourth Industrial Revolution. Its impact
can already be seen in homes, businesses and political processes. In its embodied form of robots, it
will soon be driving cars, stocking warehouses and caring for the young and elderly. It holds the
promise of solving some of the most pressing issues facing society, but also presents challenges such
as inscrutable “black box” algorithms, unethical use of data and potential job displacement. As rapid
advances in machine learning (ML) increase the scope and scale of AI’s deployment across all aspects
of daily life, and as the technology itself can learn and change on its own, multi-stakeholder
collaboration is required to optimize accountability, transparency, privacy and impartiality to create
trust.

European Artificial Intelligence (AI) leadership, the path for an integrated vision
AI is not a well-defined technology and no universally agreed definition exists. It is rather a cover term
for techniques associated with data analysis and pattern recognition. AI is not a new technology,
having existed since the 1950s. While some markets, sectors and individual businesses are more
advanced than others, AI is still at a relatively early stage of development, so that the range of
potential applications, and the quality of most existing applications, have ample margins left for
further development and improvement.

Encyclopaedia Britannica
Artificial intelligence (AI), is the ability of a digital computer or computer-controlled robot to
perform tasks commonly associated with intelligent beings. The term is frequently applied to the
project of developing systems endowed with the intellectual processes characteristic of humans, such
as the ability to reason, discover meaning, generalize, or learn from past experience.
As you can see, Artificial Intelligence is a vast domain. Everyone looks at AI in a different way according
to their mindset. Now, according to your knowledge of AI, start filling the KWLH chart:

K • What I Know?

W • What I Want to know?


L • What have I learned?

H • How I learnt this?

What do you know about Artificial Intelligence (AI)?


__________________________________________________________________________________
__________________________________________________________________________________
__________________________________________________________________________________
__________________________________________________________________________________
__________________________________________________________________________________
__________________________________________________________________________________

What do you want to know about AI?

__________________________________________________________________________________
__________________________________________________________________________________
__________________________________________________________________________________
__________________________________________________________________________________
__________________________________________________________________________________
__________________________________________________________________________________

What have you learnt about AI?


__________________________________________________________________________________
__________________________________________________________________________________
__________________________________________________________________________________
__________________________________________________________________________________
__________________________________________________________________________________
__________________________________________________________________________________

How have you learnt this about AI?

__________________________________________________________________________________
__________________________________________________________________________________
__________________________________________________________________________________
__________________________________________________________________________________
__________________________________________________________________________________
__________________________________________________________________________________
In other words, AI can be defined as:

AI is a form of Intelligence; a type of technology and a field of study.

AI theory and development of computer systems (both machines and software) enables machines to
perform tasks that normally require human intelligence.

Artificial Intelligence covers a broad range of domains and applications and is expected to impact every
field in the future. Overall, its core idea is building machines and algorithms which are capable of
performing computational tasks that would otherwise require human like brain functions.

AI, ML & DL
As you have been progressing towards building AI readiness, you must have come across a very
common dilemma between Artificial Intelligence (AI) and Machine Learning (ML). Many times, these
terms are used interchangeably but are they the same? Is there no difference in Machine Learning
and Artificial Intelligence? Is Deep Learning (DL) Also Artificial Intelligence? What exactly is Deep
Learning? Let us see.

Artificial Intelligence (AI)


Refers to any technique that enables computers to mimic human intelligence. It gives the ability to
machines to recognize a human’s face; to move and manipulate objects; to understand the voice
commands by humans, and also do other tasks. The AI-enabled machines think algorithmically and
execute what they have been asked for intelligently.

Machine Learning (ML)


It is a subset of Artificial Intelligence which enables machines to improve at tasks with experience
(data). The intention of Machine Learning is to enable machines to learn by themselves using the
provided data and make accurate Predictions/ Decisions.

* Images shown here are the property of individual organisations and are used here for reference purpose only.
Deep Learning (DL)
It enables software to train itself to perform tasks with vast amounts of data. In Deep Learning, the
machine is trained with huge amounts of data which helps it in training itself around the data. Such
machines are intelligent enough to develop algorithms for themselves. Deep Learning is the most
advanced form of Artificial Intelligence out of these three. Then comes Machine Learning which is
intermediately intelligent and Artificial Intelligence covers all the concepts and algorithms which, in
some way or the other mimic human intelligence.

There are a lot of applications of AI out of which few are those which come under ML out of which
very few can be labelled as DL. Therefore, Machine Learning (ML) and Deep Learning (DL) are part of
Artificial Intelligence (AI), but not everything that is Machine learning will be Deep learning.

Introduction to AI Domains
Artificial Intelligence becomes intelligent according to the training which it gets. For training, the
machine is fed with datasets. According to the applications for which the AI algorithm is being
developed, the data which is fed into it changes. With respect to the type of data fed in the AI
model, AI models can be broadly categorised into three domains:

Data Sciences Computer Vision Natural Language Processing

Data Sciences
Data sciences is a domain of AI related to data systems and processes, in which the system collects
numerous data, maintains data sets and derives meaning/sense out of them.

The information extracted through data science can be used to make a decision about it.

Example of Data Science

Price Comparison Websites


These websites are being driven by lots and lots of data. If you have
ever used these websites, you would know, the convenience of
comparing the price of a product from multiple vendors at one
place. PriceGrabber, PriceRunner, Junglee, Shopzilla, DealTime are
some examples of price comparison websites. Now a days, price
comparison website can be found in almost every domain such as
technology, hospitality, automobiles, durables, apparels etc.

Computer Vision
Computer Vision, abbreviated as CV, is a domain of AI that depicts the capability of a machine to get
and analyse visual information and afterwards predict some decisions about it. The entire process
involves image acquiring, screening, analysing, identifying and extracting information. This extensive
processing helps computers to understand any visual content and act on it accordingly. In computer
vision, Input to machines can be photographs, videos and pictures from thermal or infrared sensors,
indicators and different sources.

* Images shown here are the property of individual organisations and are used here for reference purpose only.
Computer vision related projects translate digital visual data into descriptions. This data is then turned
into computer-readable language to aid the decision-making process. The main objective of this
domain of AI is to teach machines to collect information from pixels.

Examples of Computer Vision


Self-Driving cars/ Automatic Cars
CV systems scan live objects and analyse them, based on whether
the car decides to keep running or to stop.

Face Lock in Smartphones

Smartphones nowadays come with the feature of face locks in


which the smartphone’s owner can set up his/her face as an
unlocking mechanism for it. The front camera detects and captures
the face and saves its features during initiation. Next time onwards,
whenever the features match, the phone is unlocked.

Natural Language Processing


Natural Language Processing, abbreviated as NLP, is a branch of artificial intelligence that deals with
the interaction between computers and humans using the natural language. Natural language refers
to language that is spoken and written by people, and natural language processing (NLP) attempts to
extract information from the spoken and written word using algorithms.

The ultimate objective of NLP is to read, decipher, understand, and make sense of the human languages
in a manilr that is valuable.

Examples of Natural Language Processing

Email filters
Email filters are one of the most basic and
initial applications of NLP online. It started
out with spam filters, uncovering certain
words or phrases that signal a spam
message.

* Images shown here are the property of individual organisations and are used here for reference purpose only.
Smart assistants
Smart assistants like Apple’s Siri and Amazon’s Alexa recognize
patterns in speech, then infer meaning and provide a useful
response.

AI Ethics
Nowadays, we are moving from the Information era to Artificial Intelligence era. Now we do not use
data or information, but the intelligence collected from the data to build solutions. These solutions
can even recommend the next TV show or movies you should watch on Netflix.

We can proudly say that India is leading in the AI usage trends, so we need to keep aspects relating to
ethical practices in mind while developing solutions using AI. Let us understand some of the ethical
concerns in detail.

Moral Issues: Self-Driving Cars

Scenario 1:
Let us imagine that we are in year 2030. Self-Driving cars which are just a concept in today’s time are
now on roads. People like us are buying them for ease and are using it for our daily transits. Of-course
because of all the features which this car has, it is expensive. Now, let us assume, one day your father
is going to office in his self-driving car. He is sitting in the back seat as the car is driving itself. Suddenly,
a small boy comes in front of this car. The incident was so sudden that the car is only able to make
either of the two choices:

1. Go straight and hit the boy who has come in front of the car and injure him severely.

2. Take a sharp right turn to save the boy and smash the car into a metal pole thus damaging the car
as well as injuring the person sitting in it.

With the help of this scenario, we need to understand that the developer of the car goes through all
such dilemmas while developing the car’s algorithm. Thus, here the morality of the developer gets
transferred into the machine as what according to him/her is right would have a higher priority and
hence would be the selection made by the machine.

If you were in the place of this developer and if there was no other alternative to the situation, which
one of the two would you prioritise and why?

__________________________________________________________________________________
__________________________________________________________________________________
__________________________________________________________________________________
__________________________________________________________________________________
__________________________________________________________________________________
__________________________________________________________________________________
__________________________________________________________________________________

* Images shown here are the property of individual organisations and are used here for reference purpose only.
Scenario 2:
Let us now assume that the car has hit the boy who came in front of it. Considering this as an accident,
who should be held responsible for it? Why?

1. The person who bought this car


2. The Manufacturing Company
3. The developer who developed the car’s algorithm
4. The boy who came in front of the car and got severely injured

__________________________________________________________________________________
__________________________________________________________________________________
__________________________________________________________________________________
__________________________________________________________________________________
__________________________________________________________________________________
__________________________________________________________________________________
__________________________________________________________________________________
__________________________________________________________________________________

Here, the choices might differ from person to person and one must understand that nobody is wrong
in this case. Every person has a different perspective and hence he/she takes decisions according to
their moralities.

Data Privacy

The world of Artificial Intelligence revolves around Data. Every company whether small or big is mining
data from as many sources as possible. More than 70% of the data collected till now has been collected
in the last 3 years which shows how important data has become in recent times. It is not wrongly said
that Data is the new gold. This makes us think:

Where do we collect data from?

__________________________________________________________________________________
__________________________________________________________________________________
__________________________________________________________________________________
__________________________________________________________________________________
__________________________________________________________________________________
__________________________________________________________________________________
__________________________________________________________________________________
__________________________________________________________________________________

Why do we need to collect data?

__________________________________________________________________________________
__________________________________________________________________________________
__________________________________________________________________________________
__________________________________________________________________________________
__________________________________________________________________________________
__________________________________________________________________________________
__________________________________________________________________________________
__________________________________________________________________________________
One of the major sources of data for many major companies is the device which all of us have in our
hands all the time: Smartphones. Smartphones have nowadays become an integral part of our lives.
Most of us use smartphones more than we interact with people around us. Smartphones in today’s
era provide us with a lot of facilities and features which have made our lives easier. Feeling hungry?
Order food online. Want to shop but don’t have time to go out? Go shopping online. From booking
tickets to watching our favourite shows, everything is available in this one small box loaded with
technology.

Another feature of smartphones nowadays is that they provide us with customised recommendations
and notifications according to our choices. Let us understand this with the help of some examples:

1. When you are talking to your friend on a mobile network or on an app like WhatsApp. You tell
your friend that you wish to buy new shoes and are looking for suggestions from him/her. You
discuss about shoes and that is it. After some time, the online shopping websites start giving
you notifications to buy shoes! They start recommending some of their products and urge you
to you buy some.
2. If you search on Google for a trip to Kerala or any other destination, just after the search, all
the apps on your phone which support advertisements, will start sending messages about
packages that you can buy for the trip.
3. Even when you are not using your phone and talking to a person face-to-face about a book
you’ve read recently while the phone is kept in a locked mode nearby, the phone will end up
giving notifications about similar books or messages about the same book once you operate
it.

In all such examples, how does the smartphone get to know about the discussions and thoughts that
you have? Remember whenever you download an app and install it, it asks you for several permissions
to access your phone’s data in different ways. If you do not allow the app these permissions, you
normally cannot access it. And to access the app and make use of it, we sometimes don’t even give it
a thought and allow the app to get all the permissions that it wants. Hence every now and then, the
app has the permission to access various sensors which are there in your smartphone and gather data
about you and your surroundings. We forget that the smartphone which we use is a box full of sensors
which are powered all the time while the phone is switched on.

This leads us to a crucial question: Are we okay with sharing our data with the external world?

__________________________________________________________________________________
__________________________________________________________________________________
__________________________________________________________________________________
__________________________________________________________________________________

Why do these apps collect data?

__________________________________________________________________________________
__________________________________________________________________________________
__________________________________________________________________________________
__________________________________________________________________________________
__________________________________________________________________________________
__________________________________________________________________________________
__________________________________________________________________________________
__________________________________________________________________________________

We need to understand that the data which is collected by various applications is ethical as the
smartphone users agree to it (by clicking on allow when it asks for permission and by agreeing to all
the terms and conditions). But at the same time if one does not want to share his/her data with
anyone, he/she can opt for alternative applications which are of similar usage and keep your data
private. For example, an alternative to WhatsApp is the Telegram app which does not collect any data
from us. But since WhatsApp is more popular and used by the crowd, people go for it without thinking
twice.

AI Bias

Another aspect to AI Ethics is bias. Everyone has a bias of their own no matter how much one tries to
be unbiased, we in some way or the other have our own biases even towards smaller things. Biases
are not negative all the time. Sometimes, it is required to have a bias to control a situation and keep
things working.

When we talk about a machine, we know that it is artificial and cannot think on its own. It can have
intelligence, but we cannot expect a machine to have any biases of its own. Any bias can transfer from
the developer to the machine while the algorithm is being developed. Let us look at some of the
examples:

1. Majorly, all the virtual assistants have a female voice. It is only now that some companies have
understood this bias and have started giving options for male voices but since the virtual assistants
came into practice, female voices are always preferred for them over any other voice. Can you think
of some reasons for this?

__________________________________________________________________________________
__________________________________________________________________________________
__________________________________________________________________________________
__________________________________________________________________________________
__________________________________________________________________________________
__________________________________________________________________________________

2. If you search on Google for salons, the first few searches are mostly for female salons. This is based
on the assumption that if a person is searching fora salon, in all probability it would be a female. Do
you think this is a bias? If yes, then is it a Negative bias or Positive one?

__________________________________________________________________________________
__________________________________________________________________________________
__________________________________________________________________________________
__________________________________________________________________________________
__________________________________________________________________________________
__________________________________________________________________________________

Various other biases are also found in various systems which are not thought up by the machine but
have got transferred from the developer intentionally or unintentionally.
AI Access

Since Artificial Intelligence is still a budding technology, not everyone has the opportunity to access it.
The people who can afford AI enabled devices make the most of it while others who cannot are left
behind. Because of this, a gap has emerged between these two classes of people and it gets widened
with the rapid advancement of technology. Let us understand this with the help of some examples:

AI creates unemployment
AI is making people’s lives easier. Most of the things nowadays are done in just a few clicks. In no time
AI will manage to be able to do all the laborious tasks which we humans have been doing since long.
Maybe in the coming years, AI enabled machines will replace all the people who work as labourers.
This may start an era of mass unemployment where people having little or no skills may be left without
jobs and others who keep up with their skills according to what is required, will flourish.

This brings us to a crossroads. On one hand where AI is advancing and improving the lives of people
by working for them and doing some of their tasks, the other hand points towards the lives of people
who are dependent on laborious jobs and are not skilled to do anything else.

Should AI replace laborious jobs? Is there an alternative for major unemployment?

__________________________________________________________________________________
__________________________________________________________________________________
__________________________________________________________________________________
__________________________________________________________________________________
__________________________________________________________________________________
__________________________________________________________________________________

Should AI not replace laborious jobs? Will the lives of people improve if they keep on being unskilled?

__________________________________________________________________________________
__________________________________________________________________________________
__________________________________________________________________________________
__________________________________________________________________________________
__________________________________________________________________________________
__________________________________________________________________________________

Here, we need to understand that to overcome such an issue, one needs to be open to changes. As
technology is advancing with time, humans need to make sure that they are a step ahead and
understand this technology with its pros and cons.

AI for kids
As we all can see, kids nowadays are smart enough to understand technology from a very early age.
As their thinking capabilities increase, they start becoming techno-savvy and eventually they learn
everything more easily than an adult. But should technology be given to children so young?

Consider this: A young boy in class 3 has got some Maths homework to finish. He is sitting at a table
which has the Google chat bot - Alexa on it, and he is struggling with his homework. Soon, he starts
asking Alexa to answer all his questions. Alexa replies with answers and the boy simply writes them
down in his notebook.

While this scenario seems funny, it still has some concerns related to it. On one hand where it is good
that the boy knows how to use technology effectively, on the other hand he uses it to complete his
homework without really learning anything since he is not applying his brain to solve the Math
problems. So, while he is smart, he might not be getting educated properly.

Is it ethical to let the boy use technology to help in this manilr?

__________________________________________________________________________________
__________________________________________________________________________________
__________________________________________________________________________________
__________________________________________________________________________________
__________________________________________________________________________________
__________________________________________________________________________________

Conclusion
Despite AI’s promises to bring forth new opportunities, there are certain associated risks that need to
be mitigated appropriately and effectively. To give a better perspective, the ecosystem and the socio-
technical environment in which the AI systems are embedded needs to be more trustworthy.
AI Project Cycle
In this chapter, we will revisit the concept of AI Project Cycle.

Introduction
Let us assume that you have to make a greeting card for your mother as it is her birthday. You are very
excited about it and have thought of many ideas to execute the same. Let us look at some of the steps
which you might take to accomplish this task:

1. Look for some cool greeting card ideas from different sources. You might go online and
checkout some videos or you may ask someone who has knowledge about it.
2. After finalising the design, you would make a list of things that are required to make this card.
3. You will check if you have the material with you or not. If not, you could go and get all the
items required, ready for use.
4. Once you have everything with you, you would start making the card.
5. If you make a mistake in the card somewhere which cannot be rectified, you will discard it and
start remaking it.
6. Once the greeting card is made, you would gift it to your mother.
Are these steps relatable?

__________________________________________________________________________________
__________________________________________________________________________________

Do you think your steps might differ? If so, write them down!

__________________________________________________________________________________
__________________________________________________________________________________
__________________________________________________________________________________
__________________________________________________________________________________
__________________________________________________________________________________
__________________________________________________________________________________
__________________________________________________________________________________
__________________________________________________________________________________

These steps show how we plan to execute the tasks around us. Consciously or Subconsciously our
mind makes up plans for every task which we have to accomplish which is why things become clearer
in our mind. Similarly, if we have to develop an AI project, the AI Project Cycle provides us with an
appropriate framework which can lead us towards the goal. The AI Project Cycle mainly has 5 stages:

* Images shown here are the property of individual organisations and are used here for reference purpose only.
Starting with Problem Scoping, you set the goal for your AI project by stating the problem which you
wish to solve with it. Under problem scoping, we look at various parameters which affect the problem
we wish to solve so that the picture becomes clearer.

To proceed,

● You need to acquire data which will become the base of your project as it will help you in
understanding what the parameters that are related to problem scoping are.
● You go for data acquisition by collecting data from various reliable and authentic sources.
Since the data you collect would be in large quantities, you can try to give it a visual image of
different types of representations like graphs, databases, flow charts, maps, etc. This makes
it easier for you to interpret the patterns which your acquired data follows.
● After exploring the patterns, you can decide upon the type of model you would build to
achieve the goal. For this, you can research online and select various models which give a
suitable output.
● You can test the selected models and figure out which is the most efficient one.
● The most efficient model is now the base of your AI project and you can develop your
algorithm around it.
● Once the modelling is complete, you now need to test your model on some newly fetched
data. The results will help you in evaluating your model and improving it.
● Finally, after evaluation, the project cycle is now complete and what you get is your AI project.
Let us understand each stage of the AI Project Cycle in detail.

Problem Scoping
It is a fact that we are surrounded by problems. They could be small or big, sometimes ignored or
sometimes even critical. Many times, we become so used to a problem that it becomes a part of our
life. Identifying such a problem and having a vision to solve it, is what Problem Scoping is about. A lot
of times we are unable to observe any problem in our surroundings. In that case, we can take a look
at the Sustainable Development Goals. 17 goals have been announced by the United nations which
are termed as the Sustainable Development Goals. The aim is to achieve these goals by the end of
2030. A pledge to do so has been taken by all the member nations of the UN.

* Images shown here are the property of individual organisations and are used here for reference purpose only.
Here are the 17 SDGs. Let’s take a look:

As you can see, many goals correspond to the problems which we might observe around us too. One
should look for such problems and try to solve them as this would make many lives better and help
our country achieve these goals.

Scoping a problem is not that easy as we need to have a deeper understanding around it so that the
picture becomes clearer while we are working to solve it. Hence, we use the 4Ws Problem Canvas to
help us out.

4Ws Problem Canvas


The 4Ws Problem canvas helps in identifying the key elements related to the problem.

Who? What? Where? Why?

Let us go through each of the blocks one by one.

Who?
The “Who” block helps in analysing the people getting affected directly or indirectly due to it. Under
this, we find out who the ‘Stakeholders’ to this problem are and what we know about them.
Stakeholders are the people who face this problem and would be benefitted with the solution. Here is
the Who Canvas:

* Images shown here are the property of individual organisations and are used here for reference purpose only.
What?
Under the “What” block, you need to look into what you have on hand. At this stage, you need to
determine the nature of the problem. What is the problem and how do you know that it is a problem?
Under this block, you also gather evidence to prove that the problem you have selected actually exists.
Newspaper articles, Media, announcements, etc are some examples. Here is the What Canvas:

Where?
Now that you know who is associated with the problem and what the problem actually is; you need
to focus on the context/situation/location of the problem. This block will help you look into the
situation in which the problem arises, the context of it, and the locations where it is prominent. Here
is the Where Canvas:
Why?
You have finally listed down all the major elements that affect the problem directly. Now it is
convenient to understand who the people that would be benefitted by the solution are; what is to be
solved; and where will the solution be deployed. These three canvases now become the base of why
you want to solve this problem. Thus, in the “Why” canvas, think about the benefits which the
stakeholders would get from the solution and how it will benefit them as well as the society.

After filling the 4Ws Problem canvas, you now need to summarise all the cards into one template. The
Problem Statement Template helps us to summarise all the key points into one single Template so
that in future, whenever there is need to look back at the basis of the problem, we can take a look at
the Problem Statement Template and understand the key elements of it.
[stakeholder(s)] Who
Our

has /have a [issue, problem, need] What


problem that

when / while [context, situation] Where

An ideal [benefit of solution for them] Why


solution would

Data Acquisition
As we move ahead in the AI Project Cycle, we come across the second element which is : Data
Acquisition. As the term clearly mentions, this stage is about acquiring data for the project. Let us first
understand what is Data. Data can be a piece of information or facts and statistics collected together
for reference or analysis. Whenever we want an AI project to be able to predict an output, we need
to train it first using data.

For example, If you want to make an Artificially Intelligent system which can predict the salary of any
employee based on his previous salaries, you would feed the data of his previous salaries into the
machine. This is the data with which the machine can be trained. Now, once it is ready, it will predict
his next salary efficiently. The previous salary data here is known as Training Data while the next salary
prediction data set is known as the Testing Data.

For better efficiency of an AI project, the Training data needs to be relevant and authentic. In the
previous example, if the training data was not of the previous salaries but of his expenses, the machine
would not have predicted his next salary correctly since the whole training went wrong. Similarly, if
the previous salary data was not authentic, that is, it was not correct, then too the prediction could
have gone wrong. Hence….

For any AI project to be efficient, the training data should be authentic and relevant to the problem
statement scoped.
Data Features
Look at your problem statement once again and try to find the data features required to address this
issue. Data features refer to the type of data you want to collect. In our previous example, data
features would be salary amount, increment percentage, increment period, bonus, etc.

After mentioning the Data features, you get to know what sort of data is to be collected. Now, the
question arises- From where can we get this data? There can be various ways in which you can collect
data. Some of them are:

Surveys Web Scraping Sensors

API
Cameras Observations (Application Program
Interface)

Sometimes, you use the internet and try to acquire data for your project from some random websites.
Such data might not be authentic as its accuracy cannot be proved. Due to this, it becomes necessary
to find a reliable source of data from where some authentic information can be taken. At the same
time, we should keep in mind that the data which we collect is open-sourced and not someone’s
property. Extracting private data can be an offence. One of the most reliable and authentic sources of
information, are the open-sourced websites hosted by the government. These government portals
have general information collected in suitable format which can be downloaded and used wisely.

Some of the open-sourced Govt. portals are: data.gov.in, india.gov.in

Data Exploration
In the previous modules, you have set the goal of your project and have also found ways to acquire
data. While acquiring data, you must have noticed that the data is a complex entity – it is full of
numbers and if anyone wants to make some sense out of it, they have to work some patterns out of
it. For example, if you go to the library and pick up a random book, you first try to go through its
content quickly by turning pages and by reading the description before borrowing it for yourself,
because it helps you in understanding if the book is appropriate to your needs and interests or not.

Thus, to analyse the data, you need to visualise it in some user-friendly format so that you can:

● Quickly get a sense of the trends, relationships and patterns contained within the data.
● Define strategy for which model to use at a later stage.
● Communicate the same to others effectively. To visualise data, we can use various types of
visual representations.

Are you aware of visual representations of data? Fill them below:


Bar
Graphs

Visual
Representations

Modelling
In the previous module of Data exploration, we have seen various types of graphical representations
which can be used for representing different parameters of data. The graphical representation makes
the data understandable for humans as we can discover trends and patterns out of it. But when it
comes to machines accessing and analysing data, it needs the data in the most basic form of numbers
(which is binary – 0s and 1s) and when it comes to discovering patterns and trends in data, the machine
goes in for mathematical representations of the same. The ability to mathematically describe the
relationship between parameters is the heart of every AI model. Thus, whenever we talk about
developing AI models, it is the mathematical approach towards analysing data which we refer to.

Generally, AI models can be classified as follows:

Machine
Learning
Learning
Based Deep
AI Models
Learning
Rule Based

Rule Based Approach


Refers to the AI modelling where the rules are defined by the developer. The machine follows the
rules or instructions mentioned by the developer and performs its task accordingly. For example, we
have a dataset which tells us about the conditions on the basis of which we can decide if an elephant
may be spotted or not while on safari. The parameters are: Outlook, Temperature, Humidity and Wind.
Now, let’s take various possibilities of these parameters and see in which case the elephant may be
spotted and in which case it may not. After looking through all the cases, we feed this data in to the
machine along with the rules which tell the machine all the possibilities. The machine trains on this
data and now is ready to be tested. While testing the machine, we tell the machine that Outlook =
Overcast; Temperature = Normal; Humidity = Normal and Wind = Weak. On the basis of this testing
dataset, now the machine will be able to tell if the elephant has been spotted before or not and will
display the prediction to us. This is known as a rule-based approach because we fed the data along
with rules to the machine and the machine after getting trained on them is now able to predict
answers for the same. A drawback/feature for this approach is that the learning is static. The machine
once trained, does not take into consideration any changes made in the original training dataset. That
is, if you try testing the machine on a dataset which is different from the rules and data you fed it at
the training stage, the machine will fail and will not learn from its mistake. Once trained, the model
cannot improvise itself on the basis of feedbacks. Thus, machine learning gets introduced as an
extension to this as in that case, the machine adapts to change in data and rules and follows the
updated path only, while a rule-based model does what it has been taught once.

Learning Based Approach


Refers to the AI modelling where the machine learns by itself. Under the Learning Based approach,
the AI model gets trained on the data fed to it and then is able to design a model which is adaptive to
the change in data. That is, if the model is trained with X type of data and the machine designs the
algorithm around it, the model would modify itself according to the changes which occur in the data
so that all the exceptions are handled in this case. For example, suppose you have a dataset comprising
of 100 images of apples and bananas each. These images depict apples and bananas in various shapes
and sizes. These images are then labelled as either apple or banana so that all apple images are
labelled ‘apple’ and all the banana images have ‘banana’ as their label. Now, the AI model is trained
with this dataset and the model is programmed in such a way that it can distinguish between an apple
image and a banana image according to their features and can predict the label of any image which is
fed to it as an apple or a banana. After training, the machine is now fed with testing data. Now, the
testing data might not have similar images as the ones on which the model has been trained. So, the
model adapts to the features on which it has been trained and accordingly predicts if the image is of
an apple or banana. In this way, the machine learns by itself by adapting to the new data which is
flowing in. This is the machine learning approach which introduces the dynamicity in the model.

The learning-based approach can further be


Learning Based Approach

divided into three parts:

Supervised Learning
Supervised
In a supervised learning model, the dataset
Learning which is fed to the machine is labelled. In
other words, we can say that the dataset is
Unsupervised known to the person who is training the
Learning machine only then he/she is able to label the
data. A label is some information which can
Reinforcement be used as a tag for data. For example,
students get grades according to the marks
Learning they secure in examinations. These grades
are labels which categorise the students
according to their marks.
There are two types of Supervised Learning models:

Classification: Where the data is classified


according to the labels. For example, in the grading
system, students are classified on the basis of the
grades they obtain with respect to their marks in
the examination. This model works on discrete
dataset which means the data need not be
continuous.

Regression: Such models work on continuous data.


For example, if you wish to predict your next salary,
then you would put in the data of your previous
salary, any increments, etc., and would train the
model. Here, the data which has been fed to the
machine is continuous.

Unsupervised Learning
An unsupervised learning model works on unlabelled dataset. This means that the data which is fed
to the machine is random and there is a possibility that the person who is training the model does not
have any information regarding it. The unsupervised learning models are used to identify
relationships, patterns and trends out of the data which is fed into it. It helps the user in understanding
what the data is about and what are the major features identified by the machine in it.

For example, you have a random data of 1000 dog images and you wish to understand some pattern
out of it, you would feed this data into the unsupervised learning model and would train the machine
on it. After training, the machine would come up with patterns which it was able to identify out of it.
The Machine might come up with patterns which are already known to the user like colour or it might
even come up with something very unusual like the size of the dogs.

* Images shown here are the property of individual organisations and are used here for reference purpose only.
Unsupervised learning models can be further divided into two categories:

Clustering: Refers to the unsupervised learning


algorithm which can cluster the unknown data
according to the patterns or trends identified out of
it. The patterns observed might be the ones which
are known to the developer or it might even come
up with some unique patterns out of it.

Dimensionality Reduction: We humans are able to visualise upto 3-Dimensions only but according to
a lot of theories and algorithms, there are various entities which exist beyond 3-Dimensions. For
example, in Natural language Processing, the words are considered to be N-Dimensional entities.
Which means that we cannot visualise them as they exist beyond our visualisation ability. Hence, to
make sense out of it, we need to reduce their dimensions. Here, dimensionality reduction algorithm
is used.

As we reduce the dimension of an entity, the information which it contains starts getting distorted.
For example, if we have a ball in our hand, it is 3-Dimensions right now. But if we click its picture, the
data transforms to 2-D as an image is a 2-Dimensional entity. Now, as soon as we reduce one
dimension, at least 50% of the information is lost as now we will not know about the back of the ball.
Whether the ball was of same colour at the back or not? Or was it just a hemisphere? If we reduce the
dimensions further, more and more information will get lost.

Hence, to reduce the dimensions and still be able to make sense out of the data, we use Dimensionality
Reduction.

Evaluation
Once a model has been made and trained, it needs to go through proper testing so that one can
calculate the efficiency and performance of the model. Hence, the model is tested with the help of
Testing Data (which was separated out of the acquired dataset at Data Acquisition stage) and the
efficiency of the model is calculated on the basis of the parameters mentioned below:

Accuracy Precision Recall F1 Score

You will read more about this stage in Chapter 7.

* Images shown here are the property of individual organisations and are used here for reference purpose only.
Neural Networks
Neural networks are loosely modelled after how neurons in the human brain behave. The key
advantage of neural networks are that they are able to extract data features automatically without
needing the input of the programmer. A neural network is essentially a system of organizing machine
learning algorithms to perform certain tasks. It is a fast and efficient way to solve problems for which
the dataset is very large, such as in images.

As seen in the figure given, the larger Neural Networks tend to perform better with larger amounts of
data whereas the traditional machine learning algorithms stop improving after a certain saturation
point.

This is a representation of how neural networks work. A Neural Network is divided into multiple layers
and each layer is further divided into several blocks called nodes. Each node has its own task to
accomplish which is then passed to the next layer. The first layer of a Neural Network is known as the
input layer. The job of an input layer is to acquire data and feed it to the Neural Network. No
processing occurs at the input layer. Next to it, are the hidden layers. Hidden layers are the layers in
which the whole processing occurs. Their name essentially means that these layers are hidden and are
not visible to the user.

Each node of these hidden layers has its own machine learning algorithm which it executes on the
data received from the input layer. The processed output is then fed to the subsequent hidden layer

* Images shown here are the property of individual organisations and are used here for reference purpose only.
of the network. There can be multiple hidden layers in a neural network system and their number
depends upon the complexity of the function for which the network has been configured. Also, the
number of nodes in each layer can vary accordingly. The last hidden layer passes the final processed
data to the output layer which then gives it to the user as the final output. Similar to the input layer,
output layer too does not process the data which it acquires. It is meant for user-interface.

Some of the features of a Neural Network are listed below:


Advance Python
Recap
In this section, we will go through a quick refreshing session around Python concepts and Jupyter
notebook. Along with this we will talk about newer concepts like packages, virtual environments, etc.

Recap 1: Jupyter Notebook


The Jupyter Notebook is an incredibly powerful tool for
interactively developing and presenting AI related projects.
The Jupyter project is the successor to the earlier IPython
What is Jupyter Notebook, which was first published as a prototype in 2010.
Although it is possible to use many different programming
Notebook? languages within Jupyter Notebooks, Python remains the
most commonly used language for it. In other words, we can
say that the Jupyter Notebook is an open source web
application that can be used to create and share documents
that contain live code, equations, visualizations, and text.

The easiest way to install and start using Jupyter


Notebook is through Anaconda. Anaconda is the most
widely used Python distribution for data science and
comes pre-loaded with all the most popular libraries and
How to access
tools. With Anaconda, comes the Anaconda Navigator Jupyter Notebook?
through which we can scroll around all the applications
which come along with it. Jupyter notebook can easily be
accessed using the Anaconda Prompt with the help of a
local host.

To work with Jupyter Notebook, it is necessary to have a


kernel on which it operates. A kernel provides programming
Kernels in language support in Jupyter. IPython is the default kernel for
Jupyter Jupyter Notebook. Therefore, whenever we need to work
Notebook with Jupyter Notebook in a virtual environment, we first
need to install a kernel inside the environment in which the
Jupyter notebook will run.
Introduction to Virtual Environments

A virtual environment is a tool that helps to keep dependencies required by


different projects separated, by creating isolated Python virtual environments
What? for them. This is one of the most important tools that most of the Python
developers use.

Imagine a scenario where we are working on two Python-based projects and


one of them works on Python 2.7 and the other uses Python 3.7. In such
situations virtual environment can be really useful to maintain dependencies of
both the projects as the virtual environments will make sure that these
Why? dependencies are not conflicting with each other and no impact reaches the
base environment at any point in time. Thus, different projects developed in
the system might have another environment to keep their dependencies
isolated from each other.

Creating virtual environments is an easy task with Anaconda distribution. Steps


How? to create one are:

1. Open Anaconda Prompt.


2. As we open the Anaconda prompt, we can see that in the beginning of the prompt message, the
term (base) is written. This is the default environment in which the anaconda works. Now, we can
create our own virtual environment and use it so that the base does not get affected by anything
that is done in the virtual environment.

3. Let us now create a virtual environment named env. To create the environment, write
conda create -n env python=3.7

This code will create an environment named env and will install Python 3.7 and other basic packages
into it.
4. After some processing, the prompt will ask if we wish to proceed with installations or not. Type Y
on it and press Enter. Once we press Enter, the packages will start getting installed in the
environment.

5. Depending upon the internet speed, the downloading of packages might take varied time. The
processing screen will look like this:
6. Once all the packages are downloaded and installed, we will get a message like this:

7. This shows that our environment called env has been successfully created. Once an environment
has been successfully created, we can access it by writing the following:
conda activate env

This would activate the virtual environment and we can see the term written in brackets has changed
form (base) to (env). Now our virtual environment is ready to be used.

But, to open and work with Jupyter Notebooks in this environment, we need to install the packages
which help in working with Jupyter Notebook. These packages get installed by default in the base
environment when Anaconda gets installed.
To install Jupyter Notebook dependencies, we need to activate our virtual environment env and write:
conda install ipykernel nb_conda jupyter

It will again ask if we wish to proceed with the installations, type Y to begin the installations. Once the
installations are complete, we can start working with Jupyter notebooks in this environment.

Recap 2: Introduction to Python


In class 9, we were introduced to Python as the programming language which will be used for working
around AI. Let us recall the basics of Python.

Python is a programming language which was created by Guido Van Rossum in


Centrum Wiskunde & Informatica. The language was publicly released in 1991
and it got its name from a BBC comedy series from 1970s – ‘Monty Python’s Flying
What? Circus’. It can be used to follow both procedural approach and object-oriented
approach of programming. Python has a lot of functionalities which makes it so
popular to use.

Artificial intelligence is the trending technology of the future. We can see so many
applications around us. If we as individuals would also like to develop an AI
application, we will need to know a programming language. There are various
Why? programming languages like Lisp, Prolog, C++, Java and Python, which can be
used for developing applications of AI. Out of these, Python gains a maximum
popularity because of the following reasons:

Easy to learn, read and maintain

Python has few keywords, simple structure and a clearly defined syntax. Python allows anyone to learn
the language quickly. A program written in Python is fairly easy-to-maintain.

A Broad Standard library

Python has a huge bunch of libraries with plenty of built-in functions to solve a variety of problems.

Interactive Mode

Python has support for an interactive mode which allows interactive testing and debugging of snippets
of code.

Portability and Compatibility

Python can run on a wide variety of operating systems and hardware platforms, and has the same
interface on all platforms.

Extendable

We can add low-level modules to the Python interpreter. These modules enable programmers to add
to or customize their tools to be more efficient.
Databases and Scalable

Python provides interfaces to all major open source and commercial databases along with a better
structure and support for much larger programs than shell scripting.

Applications of Python
There exist a wide variety of applications when it comes to Python. Some of the applications are:

Recap 3: Python Basics


In class 9, as Python was introduced, we also discussed about some basic Python syntaxes which can
help us in writing codes in Python language. Let us brush up all the concepts once and see how we can
use them in coding.

1. Printing Statements

We can use Python to display outputs for any code we write. To print any statement, we use print()
function in Python.

2. Python Statements and Comments

Instructions written in the source code to execute are known as statements. These are the lines of
code which we write for the computer to work upon. For example, if we wish to print the addition of
two numbers, say 5 and 10, we would simply write:
print(5+10)

This is a Python statement as the computer would go through it and do the needful (which in this
case would be to calculate 5+10 and print it on the output screen)
On the other hand, there exist some statements which do not get executed by the computer. These
lines of code are skipped by the machine. They are known as comments. Comments are the
statements which are incorporated in the code to give a better understanding of code statements to
the user. To write a comment in Python, one can use # and then write anything after it. For example:
# This is a comment and will not be read by the machine.
print(5+10) # This is a statement and the machine will print the
summation.

Here, we can see that the first line is a comment as it starts with #. In the second line, we have an
executable statement followed by a comment which is written to explain the code. In this way, we can
add comments into our code so that anyone can understand the gist of it.

3. Keywords & Identifiers

In Python, there exist some words which are pre-defined and carry a specific meaning for the machine
by default. These words are known as keywords. Keywords cannot be changed at any point in time
and should not be used any other way except the default one, otherwise they create confusion and
might result in ambiguous outputs. Some of the Keywords are mentioned below:

Note that keywords are case-sensitive.

An identifier is any word which is variable. Identifiers can be declared by the user as per their
convenience of use and can vary according to the way the user wants. These words are not defined
and can be used in any way. Keywords cannot be used as identifiers. Some examples of keywords can
be: count, interest, x, ai_learning, Test, etc. Identifiers are also case-sensitive hence an identifier
named as Test would be different from an identifier named test.

4. Variables & Datatypes

A variable is a named location used to store data in the memory. It is helpful to think of variables as a
container that holds data which can be changed later throughout programming. Just like in
Mathematics, in Python too we can use variables to store values in it. The difference here is, that in
Python, the variables not only store numerical values, but can also contain different types of data.
For example:
X = 10 # X variable contains numerical data
Letters = ‘XYZ’ # Letters variable contains alphabetic data
number = 13.95 # number variable contains a decimal value
word = ‘k’ # word variable contains a character

All of these variables contain different types of data in them. The type of data is defined by the term
datatype in Python. There can be various types of data which are used in Python programming. Hence,
the machine identifies the type of variable according to the value which is stored inside it. Various
datatypes in Python can be:

5. Python inputs

In Python, not only can we display the output to the user, but we can also collect data from the user
and can pass it on to the Python script for further processing. To collect the data from the user at the
time of execution, input() function is used. While using the input function, the datatype of the
expected input is required to be mentioned so that the machine does not interpret the received data
in an incorrect manilr as the data taken as input from the user is considered to be a string (sequence
of characters) by default.

For example:
Str = input(<String>) # Python expects the input to be of string
datatype
Number = int(input(<string>)) # Input string gets converted to an
integer value before assignment
Value = float(input(<String>)) # Input string gets converted to a
decimal value before assignment

6. Python Operators

Operators are special symbols which represent computation. They are applied on operand(s), which
can be values or variables. Same operators can behave differently on different data types. Operators
when applied on operands form an expression. Operators are categorized as Arithmetic, Relational,
Logical and Assignment. Value and variables when used with operators are known as operands.
a. Arithmetic Operators

Operator Meaning Expression Result

+ Addition 10 + 20 30

- Subtraction 30 - 10 20

* Multiplication 30 * 100 300

/ Division 30 / 10 20.0

// Integer Division 25 // 10 2

% Remainder 25 % 10 5

** Raised to power 3 ** 2 9

b. Conditional Operators

Operator Meaning Expression Result


20 > 10 True
> Greater Than
15 > 25 False
20 < 45 True
< Less Than
20 < 10 False
5 == 5 True
== Equal To
5 == 6 False
67 != 45 True
!= Not Equal to
35 != 35 False
45 >= 45 True
>= Greater than or Equal to
23 >= 34 False
13 <= 24 True
<= Less than or equal to
13 <= 12 False

c. Logical Operators

Operator Meaning Expression Result


True and True True
and And operator
True and False False
True or False True
or Or operator
False or False False
not False True
not Not Operator
not True False

d. Assignment Operators

Operator Expression Equivalent to


= X=5 X=5
+= X +=5 X=X+5
-= X -= 5 X=X-5
*= X *= 5 X=X*5
/= X /= 5 X=X/5
7. Conditional Statements

While coding in Python, a lot of times we need to take decisions. For example, if a person needs to
create a calculator with the help of a Python code, he/she needs to take in 2 numbers from the user
and then ask the user about which function he/she wishes to operate. Now, according to the user’s
choice, the selection of function would change. In this case, we need the machine to understand what
should happen when. This is where conditional statements help. Conditional statements help the
machine in taking a decision according to the condition which gets fulfilled. There exist different types
of conditional statements in Python. Some of them are:

If statement If-else statement If-else ladder

According to the number of conditions and their dependency on each other, the relevant type of
conditional statement is used.

8. Looping

A lot of times, it happens that a task needs to be executed multiple number of times. For example, we
need to print hello 10 times on the output screen. One way of doing this is writing 10 print statements.
But this is time and space consuming. The other way, which is more efficient, is to use loop statements.
The loop statements help in iterating statements or a group of statements as many times as it is asked
for. In this case, we will simply write a loop which would start counting from 1 to 10. At every count,
it will print hello once on the screen and as soon as it reaches 10, the loop will stop executing. All this
can be done by just one loop statement.

Various types of looping mechanisms are available in Python. Some of them are:

While Loop
For Loop Do-While Loop

These were some of the basic concepts for writing a code in Python. We can explore these concepts
further by going through the experiential Jupyter notebook for this chapter. In that notebook, we will
get to explore Python basic concepts and we can also work around them to develop better
understanding around it.

Python Packages
A package is nothing but a space where we can find codes or functions or modules of similar type.
There are various packages readily available to use for free (perks of Python being an open-sourced
language) for various purposes.
To use any package in Python, we need to install it. Installing Python packages is easy. Steps for
package installation are:
1. Open Anaconda Navigator and activate your working environment.
2. Let us assume we wish to install the numpy package. To install this package, simply write:
conda install numpy

3. It will ask us to type Y if we wish to proceed with the installations. As soon as we type Y, the
installations will start and our package will be installed in our selected environment.
4. We can also install multiple packages all at once by mentioning all of them in one line. For
example, if we wish to install numpy, pandas and matplotlib package in our working
environment. For this, simply write:
conda install numpy pandas matplotlib

This code will install these three packages altogether in our environment.

Now, once the packages are installed, we can start using them by importing them in the file where
they are required. As soon as we open our Jupyter Notebook, include the package in the notebook by
writing the import command. Importing a package can be done in various ways:
import numpy

Meaning: Import numpy in the file to use its functionalities in the file to which it has been imported.

import numpy as np
Meaning: Import numpy and refer to it as np wherever it is used.

from numpy import array


Meaning: import only one functionality (array) from the whole numpy package. While this gives faster
processing, it limits the package’s usability.

from numpy import array as arr


Meaning: Import only one functionality (array) from the whole numpy package and refer to it as arr
wherever it is used. Some of the readily available packages are:

To develop a better understanding around these packages, let us go through the Jupyter Notebook
of package exploration and see how these packages can be used in Python.
Data Sciences
Introduction
As we have discussed earlier in class 9, Artificial Intelligence is a technology which completely depends
on data. It is the data which is fed into the machine which makes it intelligent. And depending upon
the type of data we have; AI can be classified into three broad domains:

• Data Sciences
Data • Working around numeric and alpha-numeric data.

• Computer Vision
CV • Working around image and visual data.

• Natural Language Processing


NLP • Working around textual and speech-based data.

Each domain has its own type of data which gets fed into the machine and hence has its own way of
working around it. Talking about Data Sciences, it is a concept to unify statistics, data analysis, machine
learning and their related methods in order to understand and analyse actual phenomena with data.
It employs techniques and theories drawn from many fields within the context of Mathematics,
Statistics, Computer Science, and Information Science.

Now before we get into the concepts of Data Sciences, let us experience this domain with the help of
the following game:

* Rock, Paper & Scissors: https://www.afiniti.com/corporate/rock-paper-


scissors

Go to this link and try to play the game of Rock, Paper Scissors against an AI model. The challenge here
is to win 20 games against AI before AI wins them against you.

Did you manage to win?

__________________________________________________________________________________
__________________________________________________________________________________
What was the strategy that you applied to win this game against the AI machine?

__________________________________________________________________________________
__________________________________________________________________________________
__________________________________________________________________________________
__________________________________________________________________________________

Was it different playing Rock, Paper & Scissors with an AI machine as compared to a human?

__________________________________________________________________________________
__________________________________________________________________________________
__________________________________________________________________________________
__________________________________________________________________________________

What approach was the machine following while playing against you?

__________________________________________________________________________________
__________________________________________________________________________________
__________________________________________________________________________________
__________________________________________________________________________________

Applications of Data Sciences


Data Science is not a new field. Data Sciences majorly work around analysing the data and when it
comes to AI, the analysis helps in making the machine intelligent enough to perform tasks by itself.
There exist various applications of Data Science in today’s world. Some of them are:

Fraud and Risk Detection*: The earliest applications of data


science were in Finance. Companies were fed up of bad debts and
losses every year. However, they had a lot of data which use to get
collected during the initial paperwork while sanctioning loans. They
decided to bring in data scientists in order to rescue them from
losses.
Over the years, banking companies learned to divide and conquer
data via customer profiling, past expenditures, and other essential
variables to analyse the probabilities of risk and default. Moreover,
it also helped them to push their banking products based on
customer’s purchasing power.

Genetics & Genomics*: Data Science applications also enable


an advanced level of treatment personalization through research
in genetics and genomics. The goal is to understand the impact
of the DNA on our health and find individual biological
connections between genetics, diseases, and drug response.
Data science techniques allow integration of different kinds of
data with genomic data in disease research, which provides a
deeper understanding of genetic issues in reactions to particular
drugs and diseases. As soon as we acquire reliable personal
genome data, we will achieve a deeper understanding of the
human DNA. The advanced genetic risk prediction will be a major step towards more individual care.

* Images shown here are the property of individual organisations and are used here for reference purpose only.
Internet Search*: When we talk about search engines, we think
‘Google’. Right? But there are many other search engines like
Yahoo, Bing, Ask, AOL, and so on. All these search engines
(including Google) make use of data science algorithms to deliver
the best result for our searched query in the fraction of a second.
Considering the fact that Google processes more than 20 petabytes
of data every day, had there been no data science, Google wouldn’t
have been the ‘Google’ we know today.

Targeted Advertising*: If you thought Search would have been


the biggest of all data science applications, here is a challenger –
the entire digital marketing spectrum. Starting from the display
banilrs on various websites to the digital billboards at the airports
– almost all of them are decided by using data science algorithms.
This is the reason why digital ads have been able to get a much
higher CTR (Call-Through Rate) than traditional advertisements.
They can be targeted based on a user’s past behaviour.

Website Recommendations:* Aren’t we all used to the


suggestions about similar products on Amazon? They not only
help us find relevant products from billions of products
available with them but also add a lot to the user experience.
A lot of companies have fervidly used this engine to promote
their products in accordance with the user’s interest and
relevance of information. Internet giants like Amazon, Twitter,
Google Play, Netflix, LinkedIn, IMDB and many more use this
system to improve the user experience. The recommendations
are made based on previous search results for a user.

Airline Route Planning*: The Airline


Industry across the world is known to
bear heavy losses. Except for a few airline
service providers, companies are
struggling to maintain their occupancy
ratio and operating profits. With high rise
in air-fuel prices and the need to offer
heavy discounts to customers, the
situation has got worse. It wasn’t long
before airline companies started using
Data Science to identify the strategic areas of improvements. Now, while using Data Science, the
airline companies can:

* Images shown here are the property of individual organisations and are used here for reference purpose only.
• Predict flight delay
• Decide which class of airplanes to buy
• Whether to directly land at the destination or take a halt in between (For example, A flight
can have a direct route from New Delhi to New York. Alternatively, it can also choose to halt
in any country.)
• Effectively drive customer loyalty programs

Getting Started
Data Sciences is a combination of Python and Mathematical concepts like Statistics, Data Analysis,
probability, etc. Concepts of Data Science can be used in developing applications around AI as it gives
a strong base for data analysis in Python.

Revisiting AI Project Cycle


But, before we get deeper into data analysis, let us recall how Data Sciences can be leveraged to solve
some of the pressing problems around us. For this, let us understand the AI project cycle framework
around Data Sciences with the help of an example.

Do you remember the AI Project Cycle?

Fill in all the stages of the cycle here:


The Scenario*

Humans are social animals. We tend to organise and/or participate in various kinds of social gatherings
all the time. We love eating out with friends and family because of which we can find restaurants
almost everywhere and out of these, many of the restaurants arrange for buffets to offer a variety of
food items to their customers. Be it small shops or big outlets, every restaurant prepares food in bulk
as they expect a good crowd to come and enjoy their food. But in most cases, after the day ends, a lot
of food is left which becomes unusable for the restaurant as they do not wish to serve stale food to
their customers the next day. So, every day, they prepare food in large quantities keeping in mind the
probable number of customers walking into their outlet. But if the expectations are not met, a good
amount of food gets wasted which eventually becomes a loss for the restaurant as they either have
to dump it or give it to hungry people for free. And if this daily loss is taken into account for a year, it
becomes quite a big amount.

Problem Scoping
Now that we have understood the scenario well, let us take a deeper look into the problem to find out
more about various factors around it. Let us fill up the 4Ws problem canvas to find out.

Who Canvas – Who is having the problem?

Who are the o Restaurants offering buffets


stakeholders? o Restaurant Chefs

o Restaurants cook food in bulk every day for their buffets to meet their
What do we
customer needs.
know about
o They estimate the number of customers that would walk into their
them?
restaurant every day.

* Images shown here are the property of individual organisations and are used here for reference purpose only.
What Canvas – What is the nature of their problem?

o Quite a large amount of food is leftover everyday unconsumed at the


What is the
restaurant which is either thrown away or given for free to needy people.
problem?
o Restaurants have to bear everyday losses for the unconsumed food.

How do you know o Restaurant Surveys have shown that restaurants face this problem of
it is a problem? food waste.

Where Canvas – Where does the problem arise?

What is the context/situation o Restaurants which serve buffet food


in which the stakeholders o At the end of the day, when no further food consumption is
experience this problem? possible

Why? – Why do you think it is a problem worth solving?

What would be of key o If the restaurant has a proper estimate of the quantity of food
value to the stakeholders? to be prepared every day, the food waste can be reduced.

How would it improve their o Less or no food would be left unconsumed.


situation? o Losses due to unconsumed food would reduce considerably.

Now that we have noted down all the factors around our problem, let us fill up the problem statement
template.

Our Restaurant Owners Who?


Have a problem of Losses due to food wastage What?
The food is left unconsumed due to improper
While Where?
estimation
Be to be able to predict the amount of food to be
An ideal solution would Why
prepared for every day consumption

The Problem statement template leads us towards the goal of our project which can now be stated
as:

“To be able to predict the quantity of food dishes to be


prepared for everyday consumption in restaurant buffets.”
Data Acquisition
After finalising the goal of our project, let us now move towards looking at various data features which
affect the problem in some way or the other. Since any AI-based project requires data for testing and
training, we need to understand what kind of data is to be collected to work towards the goal. In our
scenario, various factors that would affect the quantity of food to be prepared for the next day
consumption in buffets would be:

Quantity of
Total Number Dish
dish prepared
of Customers consumption
per day

Unconsumed Quantity of
dish quantity Price of dish dish for the
per day next day
Now let us understand how these factors are related to our problem statement. For this, we can use
the System Maps tool to figure out the relationship of elements with the project’s goal. Here is the
System map for our problem statement.
In this system map, you can see how the relationship of each element is defined with the goal of our
project. Recall that the positive arrows determine a direct relationship of elements while the negative
ones show an inverse relationship of elements.

After looking at the factors affecting our problem statement, now it’s time to take a look at the data
which is to be acquired for the goal. For this problem, a dataset covering all the elements mentioned
above is made for each dish prepared by the restaurant over a period of 30 days. This data is collected
offline in the form of a regular survey since this is a personalised dataset created just for one
restaurant’s needs.

Specifically, the data collected comes under the following categories: Name of the dish, Price of the
dish, Quantity of dish produced per day, Quantity of dish left unconsumed per day, Total number of
customers per day, Fixed customers per day, etc.

Data Exploration
After creating the database, we now need to look at the data collected and understand what is
required out of it. In this case, since the goal of our project is to be able to predict the quantity of food
to be prepared for the next day, we need to have the following data:

Quantity of
Quantity of that
unconsumed
Name of dish dish prepared per
portion of the dish
day
per day

Thus, we extract the required information from the curated dataset and clean it up in such a way that
there exist no errors or missing elements in it.

Modelling
Once the dataset is ready, we train our model on it. In this case, a regression model is chosen in which
the dataset is fed as a dataframe and is trained accordingly. Regression is a Supervised Learning model
which takes in continuous values of data over a period of time. Since in our case the data which we
have is a continuous data of 30 days, we can use the regression model so that it predicts the next
values to it in a similar manilr. In this case, the dataset of 30 days is divided in a ratio of 2:1 for training
and testing respectively. In this case, the model is first trained on the 20-day data and then gets
evaluated for the rest of the 10 days.

Evaluation
Once the model has been trained on the training dataset of 20 days, it is now time to see if the model
is working properly or not. Let us see how the model works and how is it tested.

Step 1: The trained model is fed data regards the name of the dish and the quantity produced for the
same.

Step 2: It is then fed data regards the quantity of food left unconsumed for the same dish on previous
occasions.

Step 3: The model then works upon the entries according to the training it got at the modelling stage.
Step 4: The Model predicts the quantity of food to be prepared for the next day.

Step 5: The prediction is compared to the testing dataset value. From the testing dataset, ideally, we
can say that the quantity of food to be produced for next day’s consumption should be the total
quantity minus the unconsumed quantity.

Step 6: The model is tested for 10 testing datasets kept aside while training.

Step 7: Prediction values of testing dataset is compared to the actual values.

Step 8: If the prediction value is same or almost similar to the actual values, the model is said to be
accurate. Otherwise, either the model selection is changed or the model is trained on more data for
better accuracy.

Once the model is able to achieve optimum efficiency, it is ready to be deployed in the restaurant for
real-time usage.

Data Collection
Data collection is nothing new which has come up in our lives. It has been in our society since ages.
Even when people did not have fair knowledge of calculations, records were still maintained in some
way or the other to keep an account of relevant things. Data collection is an exercise which does not
require even a tiny bit of technological knowledge. But when it comes to analysing the data, it
becomes a tedious process for humans as it is all about numbers and alpha-numerical data. That is
where Data Science comes into the picture. It not only gives us a clearer idea around the dataset, but
also adds value to it by providing deeper and clearer analyses around it. And as AI gets incorporated
in the process, predictions and suggestions by the machine become possible on the same.

Now that we have gone through an example of a Data Science based project, we have a bit of clarity
regarding the type of data that can be used to develop a Data Science related project. For the data
domain-based projects, majorly the type of data used is in numerical or alpha-numerical format and
such datasets are curated in the form of tables. Such databases are very commonly found in any
institution for record maintenance and other purposes. Some examples of datasets which you must
already be aware of are:

Banks Databases of loans issued, account holder, locker owners, employee


registrations, bank visitors, etc.

Usage details per day, cash denominations transaction details, visitor


ATM Machines
details, etc.

Movie details, tickets sold offline, tickets sold online, refreshment


Movie Theatres purchases, etc.

Now look around you and find out what are the different types of databases which are maintained in
the places mentioned below. Try surveying people who are responsible for the designated places to
get a better idea.

Your classroom Your school Your city


As you can see, all the type of data which has been mentioned above is in the form of tables. Tables
which contain numeric or alpha-numeric data. But this leads to a very critical dilemma: are these
datasets accessible to all? Should these databases be accessible to all? What are the various sources
of data from which we can gather such databases? Let’s find out!

Sources of Data
There exist various sources of data from where we can collect any type of data required and the data
collection process can be categorised in two ways: Offline and Online.

Offline Data Collection Online Data Collection


Sensors Open-sourced Government Portals
Surveys Reliable Websites (Kaggle)
Interviews World Organisations’ open-sourced statistical
Observations websites

While accessing data from any of the data sources, following points should be kept in mind:

1. Data which is available for public usage only should be taken up.
2. Personal datasets should only be used with the consent of the owner.
3. One should never breach someone’s privacy to collect data.
4. Data should only be taken form reliable sources as the data collected from random sources
can be wrong or unusable.
5. Reliable sources of data ensure the authenticity of data which helps in proper training of the
AI model.

Types of Data
For Data Science, usually the data is collected in the form of tables. These tabular datasets can be
stored in different formats. Some of the commonly used formats are:

1. CSV: CSV stands for comma separated values. It is a simple file format used to store tabular
data. Each line of this file is a data record and reach record consists of one or more fields which
are separated by commas. Since the values of records are separated by a comma, hence they
are known as CSV files.
2. Spreadsheet: A Spreadsheet is a piece of paper or a computer program which is used for
accounting and recording data using rows and columns into which information can be
entered. Microsoft excel is a program which helps in creating spreadsheets.
3. SQL: SQL is a programming language also known as Structured Query Language. It is a domain-
specific language used in programming and is designed for managing data held in different
kinds of DBMS (Database Management System) It is particularly useful in handling structured
data.
A lot of other formats of databases also exist, you can explore them online!

Data Access
After collecting the data, to be able to use it for programming purposes, we should know how to access
the same in a Python code. To make our lives easier, there exist various Python packages which help
us in accessing structured data (in tabular form) inside the code. Let us take a look at some of these
packages:
NumPy
NumPy, which stands for Numerical Python, is the fundamental package for Mathematical and logical
operations on arrays in Python. It is a commonly used package when it comes to working around
numbers. NumPy gives a wide range of arithmetic operations around numbers giving us an easier
approach in working with them. NumPy also works with arrays, which is nothing but a homogenous
collection of Data.

An array is nothing but a set of multiple values which are of same datatype. They can be numbers,
characters, booleans, etc. but only one datatype can be accessed through an array. In NumPy, the
arrays used are known as ND-arrays (N-Dimensional Arrays) as NumPy comes with a feature of
creating n-dimensional arrays in Python.

An array can easily be compared to a list. Let us take a look at how they are different:

NumPy Arrays Lists


1. Homogenous collection of Data. 1. Heterogenous collection of Data.
2. Can contain only one type of data, hence not 2. Can contain multiple types of data,
flexible with datatypes. hence flexible with datatypes.
3. Cannot be directly initialized. Can be operated 3. Can be directly initialized as it is a part
with Numpy package only. of Python syntax.
4. Direct numerical operations can be done. For 4. Direct numerical operations are not
example, dividing the whole array by 3 divides possible. For example, dividing the
every element by 3. whole list by 3 cannot divide every
5. Widely used for arithmetic operations. element by 3.
6. Arrays take less memory space. 5. Widely used for data management.
7. Functions like concatenation, appending, 6. Lists acquire more memory space.
reshaping, etc are not trivially possible with 7. Functions like concatenation,
arrays. appending, reshaping, etc are trivially
8. Example: To create a numpy array ‘A’: possible with lists.
8. Example: To create a list:
import numpy
A=numpy.array([1,2,3,4,5,6,7,8,9,0]) A = [1,2,3,4,5,6,7,8,9,0]

Pandas
Pandas is a software library written for the Python programming language for data manipulation and
analysis. In particular, it offers data structures and operations for manipulating numerical tables and
time series. The name is derived from the term "panel data", an econometrics term for data sets that
include observations over multiple time periods for the same individuals.

Pandas is well suited for many different kinds of data:

• Tabular data with heterogeneously-typed columns, as in an SQL table or Excel spreadsheet


• Ordered and unordered (not necessarily fixed-frequency) time series data.
• Arbitrary matrix data (homogeneously typed or heterogeneous) with row and column labels
• Any other form of observational / statistical data sets. The data actually need not be labelled
at all to be placed into a Pandas data structure
The two primary data structures of Pandas, Series (1-dimensional) and DataFrame (2-dimensional),
handle the vast majority of typical use cases in finance, statistics, social science, and many areas of
engineering. Pandas is built on top of NumPy and is intended to integrate well within a scientific
computing environment with many other 3rd party libraries.

Here are just a few of the things that pandas does well:

• Easy handling of missing data (represented as NaN) in floating point as well as non-floating
point data
• Size mutability: columns can be inserted and deleted from DataFrame and higher dimensional
objects
• Automatic and explicit data alignment: objects can be explicitly aligned to a set of labels, or
the user can simply ignore the labels and let Series, DataFrame, etc. automatically align the
data for you in computations
• Intelligent label-based slicing, fancy indexing, and subsetting of large data sets
• Intuitive merging and joining data sets
• Flexible reshaping and pivoting of data sets

Matplotlib*
Matplotlib is an amazing visualization library in Python for 2D plots of arrays. Matplotlib is a multi-
platform data visualization library built on NumPy arrays. One of the greatest benefits of visualization
is that it allows us visual access to huge amounts of data in easily digestible visuals. Matplotlib comes
with a wide variety of plots. Plots helps to understand trends, patterns, and to make correlations.
They’re typically instruments for reasoning about quantitative information. Some types of graphs that
we can make with this package are listed below:

Not just plotting, but you can also modify your plots the way you wish. You can stylise them and make
them more descriptive and communicable.

These packages help us in accessing the datasets we have and also in exploring them to develop a
better understanding of them.

* Images shown here are the property of individual organisations and are used here for reference purpose only.
Basic Statistics with Python
We have already understood that Data Sciences works around analysing data and performing tasks
around it. For analysing the numeric & alpha-numeric data used for this domain, mathematics comes
to our rescue. Basic statistical methods used in mathematics come quite hAmanin Python too for
analysing and working around such datasets. Statistical tools widely used in Python are:

Do you remember using these formulas in your class? Let us recall all of them here:

1. What is Mean? How is it calculated?

__________________________________________________________________________________
__________________________________________________________________________________
__________________________________________________________________________________
__________________________________________________________________________________

2. What is Median? How is it calculated?

__________________________________________________________________________________
__________________________________________________________________________________
__________________________________________________________________________________
__________________________________________________________________________________

3. What is Mode? How is it calculated?

__________________________________________________________________________________
__________________________________________________________________________________
__________________________________________________________________________________
__________________________________________________________________________________
4. What is Standard Deviation? How is it calculated?

__________________________________________________________________________________
__________________________________________________________________________________
__________________________________________________________________________________
__________________________________________________________________________________

5. What is Variance? How is it calculated?

__________________________________________________________________________________
__________________________________________________________________________________
__________________________________________________________________________________
__________________________________________________________________________________

Advantage of using Python packages is that we do not need to make our own formula or equation to
find out the results. There exist a lot of pre-defined functions with packages like NumPy which reduces
this trouble for us. All we need to do is write that function and pass on the data to it. It’s that simple!

Let us take a look at various Python syntaxes that can help us with the statistical work in data analysis.
Head to the Jupyter Notebook of Basic statistics with Python and start exploring! You may find the
Jupyter notebook here: http://bit.ly/data_notebook

Data Visualisation
While collecting data, it is possible that the data might come with some errors. Let us first take a look
at the types of issues we can face with data:

1. Erroneous Data: There are two ways in which the data can be erroneous:

• Incorrect values: The values in the dataset (at random places) are incorrect. For example, in
the column of phone number, there is a decimal value or in the marks column, there is a name
mentioned, etc. These are incorrect values that do not resemble the kind of data expected in
that position.
• Invalid or Null values: At some places, the values get corrupted and hence they become
invalid. Many times you will find NaN values in the dataset. These are null values which do not
hold any meaning and are not processible. That is why, these values (as and when
encountered) are removed from the database.
2. Missing Data: In some datasets, some cells remain empty. The values of these cells are missing and
hence the cells remain empty. Missing data cannot be interpreted as an error as the values here are
not erroneous or might not be missing because of any error.

3. Outliers: Data which does not fall in the range of a certain element are referred to as outliers. To
understand this better, let us take an example of marks of students in a class. Let us assume that a
student was absent for exams and hence has got 0 marks in it. If his marks are taken into account, the
whole class’s average would go down. To prevent this, the average is taken for the range of marks
from highest to lowest keeping this particular result separate. This makes sure that the average marks
of the class are true according to the data.

Analysing the data collected can be difficult as it is all about tables and numbers. While machines work
efficiently on numbers, humans need visual aid to understand and comprehend the information
passed. Hence, data visualisation is used to interpret the data collected and identify patterns and
trends out of it.
In Python, Matplotlib package helps in visualising the data and making some sense out of it. As we
have already discussed before, with the help of this package, we can plot various kinds of graphs. Let
us discuss some of them here:

Scatter plots are used to plot discontinuous data; that is, the data
which does not have any continuity in flow is termed as
Scatter Plot discontinuous. There exist gaps in data which introduce discontinuity.
A 2D scatter plot can display information maximum upto 4
parameters.

In this scatter plot, 2 axes (X and Y) are two different parameters. The colour of circles and the size
both represent 2 different parameters. Thus, just through one coordinate on the graph, one can
visualise 4 different parameters all at once.

* Images shown here are the property of individual organisations and are used here for reference purpose only.
It is one of the most commonly used graphical methods. From
students to scientists, everyone uses bar charts in some way or the
Bar Chart other. It is a very easy to draw yet informative graphical
representation. Various versions of bar chart exist like single bar
chart, double bar chart, etc.

This is an example of a double bar chart. The 2 axes depict two different parameters while bars of
different colours work with different entities ( in this case it is women and men). Bar chart also works
on discontinuous data and is made at uniform intervals.

Histograms are the accurate representation of a continuous data.

Histogram When it comes to plotting the variation in just one entity of a period
of time, histograms come into the picture. It represents the frequency
of the variable at different points of time with the help of the bins.

* Images shown here are the property of individual organisations and are used here for reference purpose only.
In the given example, the histogram is showing the variation in frequency of the entity plotted with
the help of XY plane. Here, at the left, the frequency of the element has been plotted and it is a
frequency map for the same. The colours show the transition from low to high and vice versa. Whereas
on the right, a continuous dataset has been plotted which might not be talking about the frequency
of occurrence of the element.

When the data is split according to its percentile throughout the

Box Plots range, box plots come in haman. Box plots also known as box and
whiskers plot conveniently display the distribution of data throughout
the range with the help of 4 quartiles.

* Images shown here are the property of individual organisations and are used here for reference purpose only.
Here as we can see, the plot contains a box and two lines at its left and right are termed as whiskers.
The plot has 5 different parts to it:

Quartile 1: From 0 percentile to 25th percentile – Here data lying between 0 and 25th percentile is
plotted. Now, if the data is close to each other, lets say 0 to 25th percentile data has been covered in
just 20-30 marks range, then the whisker would be smaller as the range is smaller. But if the range is
large that is 0-30 marks range, then the whisker would also get elongated as the range is longer.

Quartile 2: From 25th Percentile to 50th percentile – 50th percentile is termed as the mean of the whole
distribution and since the data falling in the range of 25th percentile to 75th percentile has minimum
deviation from the mean, it is plotted inside the box.

Quartile 3: From 50th percentile to 75th percentile – This range is again plotted in the box as its
deviation from the mean is less. Quartile 2 & 3 (from 25th percentile to 75th percentile) together
constitute the Inter Quartile Range (IQR). Also, depending upon the range of distribution, just like
whiskers, the length of box also varies if the data is less spread or more.

Quartile 4: From 75th percentile to 100th percentile – It is the whiskers plot for top 25 percentile data.

Outliers: The advantage of box plots is that they clearly show the outliers in a data distribution. Points
which do not lie in the range are plotted outside the graph as dots or circles and are termed as outliers
as they do not belong to the range of data. Since being out of range is not an error, that is why they
are still plotted on the graph for visualisation.

Let us now move ahead and experience data visualisation using Jupyter notebook. Matplotlib library
will help us in plotting all sorts of graphs while Numpy and Pandas will help us in analysing the data.

Data Sciences: Classification Model


In this section, we would be looking at one of the classification models used in Data Sciences. But
before we look into the technicalities of the code, let us play a game.

Personality Prediction
Step 1: Here is a map. Take a good look at it. In this map you can see the arrows determine a quality.
The qualities mentioned are:

1. Positive X-axis – People focussed: You focus more on people and try to deliver the best
experience to them.
2. Negative X-axis – Task focussed: You focus more on the task which is to be accomplished and
try to do your best to achieve that.
3. Positive Y-axis – Passive: You focus more on listening to people and understanding everything
that they say without interruption.
4. Negative Y-axis – Active: You actively participate in the discussions and make sure that you
make your point in-front of the crowd.
Think for a minute and understand which of these qualities you have in you. Now, take a chit and write
your name on it. Place this chit at a point in this map which best describes you. It can be placed
anywhere on the graph. Be honest about yourself and put it on the graph.

Step 2: Now that you have all put up your chits on the graph, it’s time to take a quick quiz. Go to this
link and finish the quiz on it individually: https://tinyurl.com/discanimal

On this link, you will find a personality prediction quiz. Take this quiz individually and try to answer all
the questions honestly. Do not take anyone’s help in it and do not discuss about it with anyone. Once
the quiz is finished, remember the animal which has been predicted for you. Write it somewhere and
do not show it to anyone. Keep it as your little secret.

Once everyone has gone through the quiz, go back to the board remove your chit, and draw the
symbol which corresponds to your animal in place of your chit. Here are the symbols:

Lion Otter Golden Retriever Beaver

 ⚫ ☺ 
Place these symbols at the locations where you had put up your names. Ask 4 students not to do so
and tell them to keep their animals a secret. Let their name chits be on the graph so that we can
predict their animals with the help of this map.

Now, we will try to use the nearest neighbour algorithm here and try to predict what can be the
possible animal(s) for these 4 unknowns. Now look that these 4 chits one by one. Which animal is
occurring the most in their vicinity? Do you think that if the m lion symbol is occurring the most near
their chit, then there is a good probability that their animal would also be a lion? Now let us try to
guess the animal for all 4 of them according to their nearest neighbours respectively. After guessing
the animals, ask these 4 students if the guess is right or not.

K-Nearest Neighbour: Explained


The k-nearest neighbours (KNN) algorithm is a simple, easy-to-implement supervised machine
learning algorithm that can be used to solve both classification and regression problems. The KNN
algorithm assumes that similar things exist in close proximity. In other words, similar things are near
to each other as the saying goes “Birds of a feather flock together”. Some features of KNN are:

• The KNN prediction model relies on the surrounding points or neighbours to determine its
class or group
• Utilises the properties of the majority of the nearest points to decide how to classify unknown
points
• Based on the concept that similar data points should be close to each other
The personality prediction activity was a brief introduction to KNN. As you recall, in that activity, we
tried to predict the animal for 4 students according to the animals which were the nearest to their
points. This is how in a lay-man’s language KNN works. Here, K is a variable which tells us about the
number of neighbours which are taken into account during prediction. It can be any integer value
starting from 1.

Let us look at another example to demystify this algorithm. Let us assume that we need to predict the
sweetness of a fruit according to the data which we have for the same type of fruit. So here we have
three maps to predict the same:
Here, X is the value which is to be predicted. The green dots depict sweet values and the blue ones
denote not sweet.

Let us try it out by ourselves first. Look at the map closely and decide whether X should be sweet or
not sweet?

Now, let us look at each graph one by one:

Here, we can see that K is taken as 1 which means that we are taking only 1 nearest

1 neighbour into consideration. The nearest value to X is a blue one hence 1-nearest
neighbour algorithm predicts that the fruit is not sweet.

In the 2nd graph, the value of K is 2. Taking 2 nearest nodes to X into consideration, we
see that one is sweet while the other one is not sweet. This makes it difficult for the
2 machine to make any predictions based on the nearest neighbour and hence the
machine is not able to give any prediction.

In the 3rd graph, the value of K becomes 3. Here, 3 nearest nodes to X are chosen out
3 of which 2 are green and 1 is blue. On the basis of this, the model is able to predict that
the fruit is sweet.

On the basis of this example, let us understand KNN better:

KNN tries to predict an unknown value on the basis of the known values. The model simply calculates
the distance between all the known points with the unknown point (by distance we mean to say the
different between two values) and takes up K number of points whose distance is minimum. And
according to it, the predictions are made.

Let us understand the significance of the number of neighbours:

1. As we decrease the value of K to 1, our predictions become less stable. Just think for a minute,
imagine K=1 and we have X surrounded by several greens and one blue, but the blue is the
single nearest neighbour. Reasonably, we would think X is most likely green, but because K=1,
KNN incorrectly predicts that it is blue.

* Images shown here are the property of individual organisations and are used here for reference purpose only.
2. Inversely, as we increase the value of K, our predictions become more stable due to majority
voting / averaging, and thus, more likely to make more accurate predictions (up to a certain
point). Eventually, we begin to witness an increasing number of errors. It is at this point we
know we have pushed the value of K too far.
3. In cases where we are taking a majority vote (e.g. picking the mode in a classification problem)
among labels, we usually make K an odd number to have a tiebreaker.
Computer Vision
Introduction
In the previous chapter, you studied the concepts of Artificial Intelligence for Data Sciences. It is a
concept to unify statistics, data analysis, machine learning and their related methods in order to
understand and analyse actual phenomena with data.

As we all know, artificial intelligence is a technique that enables computers to mimic human
intelligence. As humans we can see things, analyse it and then do the required action on the basis of
what we see.

But can machines do the same? Can machines have the eyes that humans have? If you answered Yes,
then you are absolutely right. The Computer Vision domain of Artificial Intelligence, enables machines
to see through images or visual data, process and analyse them on the basis of algorithms and
methods in order to analyse actual phenomena with images.

Now before we get into the concepts of Computer Vision, let us experience this domain with the help
of the following game:

* Emoji Scavenger Hunt :


https://emojiscavengerhunt.withgoogle.com/

Go to the link and try to play the game of Emoji Scavenger Hunt. The challenge here is to find 8 items
within the time limit to pass.

Did you manage to win?

__________________________________________________________________________________
__________________________________________________________________________________

What was the strategy that you applied to win this game?

__________________________________________________________________________________
__________________________________________________________________________________
__________________________________________________________________________________
__________________________________________________________________________________
__________________________________________________________________________________

Was the computer able to identify all the items you brought in front of it?

__________________________________________________________________________________
__________________________________________________________________________________
__________________________________________________________________________________
__________________________________________________________________________________
__________________________________________________________________________________

* Images shown here are the property of individual organisations and are used here for reference purpose only.
Did the lighting of the room affect the identifying of items by the machine?

__________________________________________________________________________________
__________________________________________________________________________________
__________________________________________________________________________________
__________________________________________________________________________________
__________________________________________________________________________________
__________________________________________________________________________________

Applications of Computer Vision


The concept of computer vision was first introduced in the 1970s. All these new applications of
computer vision excited everyone. Having said that, the computer vision technology advanced enough
to make these applications available to everyone at ease today. However, in recent years the world
witnessed a significant leap in technology that has put computer vision on the priority list of many
industries. Let us look at some of them:

Facial Recognition*: With the advent of smart cities and smart homes,
Computer Vision plays a vital role in making the home smarter. Security
being the most important application involves use of Computer Vision
for facial recognition. It can be either guest recognition or log
maintenance of the visitors.

It also finds its application in schools for an attendance system based on


facial recognition of students.

Face Filters*: The modern-day apps like Instagram and snapchat have
a lot of features based on the usage of computer vision. The
application of face filters is one among them. Through the camera the
machine or the algorithm is able to identify the facial dynamics of the
person and applies the facial filter selected.

Google’s Search by Image*: The maximum amount


of searching for data on Google’s search engine comes
from textual data, but at the same time it has an
interesting feature of getting search results through an
image. This uses Computer Vision as it compares
different features of the input image to the database
of images and give us the search result while at the
same time analysing various features of the image.

* Images shown here are the property of individual organisations and are used here for reference purpose only.
Computer Vision in Retail*: The retail field has been one of the
fastest growing field and at the same time is using Computer
Vision for making the user experience more fruitful. Retailers can
use Computer Vision techniques to track customers’ movements
through stores, analyse navigational routes and detect walking
patterns.
Inventory Management is another such application. Through
security camera image analysis, a Computer Vision algorithm can
generate a very accurate estimate of the items available in the
store. Also, it can analyse the use of shelf space to identify
suboptimal configurations and suggest better item placement.

Self-Driving Cars: Computer Vision is the fundamental


technology behind developing autonomous vehicles.
Most leading car manufacturers in the world are
reaping the benefits of investing in artificial intelligence
for developing on-road versions of hands-free
technology.

This involves the process of identifying the objects,


getting navigational routes and also at the same time
environment monitoring.

Medical Imaging*: For the last decades, computer-


supported medical imaging application has been a
trustworthy help for physicians. It doesn’t only
create and analyse images, but also becomes an
assistant and helps doctors with their interpretation.
The application is used to read and convert 2D scan
images into interactive 3D models that enable
medical professionals to gain a detailed
understanding of a patient’s health condition.

Google Translate App*: All you need to do to read signs in a


foreign language is to point your phone’s camera at the words and
let the Google Translate app tell you what it means in your preferred
language almost instantly. By using optical character recognition to
see the image and augmented reality to overlay an accurate
translation, this is a convenient tool that uses Computer Vision.

* Images shown here are the property of individual organisations and are used here for reference purpose only.
Computer Vision: Getting Started
Computer Vision is a domain of Artificial Intelligence, that deals with the images. It involves the
concepts of image processing and machine learning models to build a Computer Vision based
application.

Computer Vision Tasks


The various applications of Computer Vision are based on a certain number of tasks which are
performed to get certain information from the input image which can be directly used for prediction
or forms the base for further analysis. The tasks used in a computer vision application are :

For Single For Multiple


Objects Objects

Object
Classification
Detection

Classification + Instance
Localisation Segementation

Classification
Image Classification problem is the task of assigning an input image one label from a fixed set of
categories. This is one of the core problems in CV that, despite its simplicity, has a large variety of
practical applications.

Classification + Localisation
This is the task which involves both processes of identifying what object is present in the image and
at the same time identifying at what location that object is present in that image. It is used only for
single objects.

Object Detection
Object detection is the process of finding instances of real-world objects such as faces, bicycles, and
buildings in images or videos. Object detection algorithms typically use extracted features and
learning algorithms to recognize instances of an object category. It is commonly used in applications
such as image retrieval and automated vehicle parking systems.

Instance Segmentation
Instance Segmentation is the process of detecting instances of the objects, giving them a category and
then giving each pixel a label on the basis of that. A segmentation algorithm takes an image as input
and outputs a collection of regions (or segments).
Basics of Images
We all see a lot of images around us and use them daily either through our mobile phones or computer
system. But do we ask some basic questions to ourselves while we use them on such a regular basis.

Don’t know the answer yet? Don’t worry, in this section we will study about the basics of an image:

Basics of Pixels
The word “pixel” means a picture element. Every photograph, in digital form, is made up of pixels.
They are the smallest unit of information that make up a picture. Usually round or square, they are
typically arranged in a 2-dimensional grid.

In the image below, one portion has been magnified many times over so that you can see its individual
composition in pixels. As you can see, the pixels approximate the actual image. The more pixels you
have, the more closely the image resembles the original.

* Images shown here are the property of individual organisations and are used here for reference purpose only.
Resolution
The number of pixels in an image is sometimes called the resolution. When the term is used to describe
pixel count, one convention is to express resolution as the width by the height, for example a monitor
resolution of 1280×1024. This means there are 1280 pixels from one side to the other, and 1024 from
top to bottom.

Another convention is to express the number of pixels as a single number, like a 5 mega pixel camera
(a megapixel is a million pixels). This means the pixels along the width multiplied by the pixels along
the height of the image taken by the camera equals 5 million pixels. In the case of our 1280×1024
monitors, it could also be expressed as 1280 x 1024 = 1,310,720, or 1.31 megapixels.

Pixel value
Each of the pixels that represents an image stored inside a computer has a pixel value which describes
how bright that pixel is, and/or what colour it should be. The most common pixel format is the byte
image, where this number is stored as an 8-bit integer giving a range of possible values from 0 to 255.
Typically, zero is to be taken as no colour or black and 255 is taken to be full colour or white.

Why do we have a value of 255 ? In the computer systems, computer data is in the form of ones and
zeros, which we call the binary system. Each bit in a computer system can have either a zero or a one.

Since each pixel uses 1 byte of an image, which is equivalent to 8 bits of data. Since each bit can have
two possible values which tells us that the 8 bit can have 255 possibilities of values which starts from
0 and ends at 255.

* Images shown here are the property of individual organisations and are used here for reference purpose only.
Grayscale Images
Grayscale images are images which have a range of shades of gray without apparent colour. The
darkest possible shade is black, which is the total absence of colour or zero value of pixel. The lightest
possible shade is white, which is the total presence of colour or 255 value of a pixel . Intermediate
shades of gray are represented by equal brightness levels of the three primary colours.

A grayscale has each pixel of size 1 byte having a single plane of 2d array of pixels. The size of a
grayscale image is defined as the Height x Width of that image.

Let us look at an image to understand about grayscale images.

Here is an example of a grayscale image. as you check, the value of pixels are within the range of 0-
255.The computers store the images we see in the form of these numbers.

RGB Images
All the images that we see around are coloured images. These images are made up of three primary
colours Red, Green and Blue. All the colours that are present can be made by combining different
intensities of red, green and blue.

* Images shown here are the property of individual organisations and are used here for reference purpose only.
Let us experience!

Go to this online link https://www.w3schools.com/colors/colors_rgb.asp. On the basis of this online


tool, try and answer all the below mentioned questions.

1) What is the output colour when you put R=G=B=255 ?

___________________________________________________________________________

2) What is the output colour when you put R=G=B=0 ?

___________________________________________________________________________

3) How does the colour vary when you put either of the three as 0 and then keep on varying
the other two?

___________________________________________________________________________
___________________________________________________________________________
___________________________________________________________________________
___________________________________________________________________________

4) How does the output colour change when all the three colours are varied in same
proportion ?

___________________________________________________________________________
___________________________________________________________________________
___________________________________________________________________________
___________________________________________________________________________

5) What is the RGB value of your favourite colour from the colour palette?

___________________________________________________________________________

Were you able to answer all the questions? If yes, then you would have understood how every colour
we see around is made.

Now the question arises, how do computers store RGB images? Every RGB image is stored in the form
of three different channels called the R channel, G channel and the B channel.

Each plane separately has a number of pixels with each pixel value varying from 0 to 255. All the three
planes when combined together form a colour image. This means that in a RGB image, each pixel has
a set of three different values which together give colour to that particular pixel.
For Example,

As you can see, each colour image is stored in the form of three different channels, each having
different intensity. All three channels combine together to form a colour we see.

In the above given image, if we split the image into three different channels, namely Red (R), Green
(G) and Blue (B), the individual layers will have the following intensity of colours of the individual
pixels. These individual layers when stored in the memory looks like the image on the extreme right.
The images look in the grayscale image because each pixel has a value intensity of 0 to 255 and as
studied earlier, 0 is considered as black or no presence of colour and 255 means white or full presence
of colour. These three individual RGB values when combined together form the colour of each pixel.

Therefore, each pixel in the RGB image has three values to form the complete colour.

Task :
Go to the following link www.piskelapp.com and create your own pixel art. Try and make a GIF using
the online app for your own pixel art.

* Images shown here are the property of individual organisations and are used here for reference purpose only.
Image Features
In computer vision and image processing, a feature is a piece of information which is relevant for
solving the computational task related to a certain application. Features may be specific structures in
the image such as points, edges or objects.
For example:
Imagine that your security camera is capturing an image. At the top of the image we are given six small
patches of images. Our task is to find the exact location of those image patches in the image.
Take a pencil and mark the exact location of those patches in the image.

Were you able to find the exact location of all the patches?
__________________________________________________________________________________

Which one was the most difficult to find?

__________________________________________________________________________________
__________________________________________________________________________________

Which one was the easiest to find?

__________________________________________________________________________________
__________________________________________________________________________________

Let’s Reflect:
Let us take individual patches into account at once and then check the exact location of those patches.
For Patch A and B: The patch A and B are flat surfaces in the image and are spread over a lot of area.
They can be present at any location in a given area in the image.
For Patch C and D: The patches C and D are simpler as compared to A and B. They are edges of a
building and we can find an approximate location of these patches but finding the exact location is
still difficult. This is because the pattern is the same everywhere along the edge.
For Patch E and F: The patches E and F are the easiest to find in the image. The reason being that E
and F are some corners of the building. This is because at the corners, wherever we move this patch
it will look different.

* Images shown here are the property of individual organisations and are used here for reference purpose only.
Conclusion
In image processing, we can get a lot of features from the image. It can be either a blob, an edge or a
corner. These features help us to perform various tasks and then get the analysis done on the basis of
the application. Now the question that arises is which of the following are good features to be used?
As you saw in the previous activity, the features having the corners are easy to find as they can be
found only at a particular location in the image, whereas the edges which are spread over a line or an
edge look the same all along. This tells us that the corners are always good features to extract from
an image followed by the edges.
Let’s look at another example to understand this. Consider the images given below and apply the
concept of good features for the following.

In the above image how would we determine the exact location of each patch?
The blue patch is a flat area and difficult to find and track. Wherever you move the blue patch it looks
the same. The black patch has an edge. Moved along the edge (parallel to edge), it looks the same.
The red patch is a corner. Wherever you move the patch, it looks different, therefore it is unique.
Hence, corners are considered to be good features in an image.

Introduction to OpenCV
Now that we have learnt about image features and its importance in image processing, we will learn
about a tool we can use to extract these features from our image for further processing.
OpenCV or Open Source Computer Vision Library is that tool which helps a computer extract these
features from the images. It is used for all kinds of images and video processing and analysis. It is
capable of processing images and videos to identify objects, faces, or even handwriting.

In this chapter we will use OpenCV for basic image processing operations on
images such as resizing, cropping and many more.
To install OpenCV library, open anaconda prompt and then write the following
command:

pip install opencv-python

Now let us take a deep dive on the various functions of OpenCV to understand the various image
processing techniques. Head to Jupyter Notebook for introduction to OpenCV given on this link:
http://bit.ly/cv_notebook

* Images shown here are the property of individual organisations and are used here for reference purpose only.
Convolution
We have learnt that computers store images in numbers, and that pixels are arranged in a particular
manner to create the picture we can recognize. These pixels have value varying from 0 to 255 and the
value of the pixel determines the color of that pixel.

But what if we edit these numbers, will it bring a change to the image? The answer is yes. As we change
the values of these pixels, the image changes. This process of changing pixel values is the base of image
editing.

We all use a lot of image editing software like photoshop and at the same time use apps like Instagram
and snapchat, which apply filters to the image to enhance the quality of that image.

As you can see, different filters applied to an image change the pixel values evenly throughout the
image. How does this happen? This is done with the help of the process of convolution and the
convolution operator which is commonly used to create these effects.

Before we understand how the convolution operation works, let us try and create a theory for the
convolution operator by experiencing it using an online application.

Task
Go to the link http://matlabtricks.com/post-5/3x3-convolution-kernels-with-online-demo and at the
bottom of the page click on load “Click to Load Application”

* Images shown here are the property of individual organisations and are used here for reference purpose only.
Once the application is loaded try different filters and apply it on the image. Observe how the value
of the kernel is changing for different filters. Try these steps

1) Change all to positive values


2) Change all to negative values
3) Have a mixture of negative and positive values

Let us follow the following steps to understand how a convolution operator works. The steps to be
followed are:

Try experimenting with the following values to come up with a theory:

1) Make 4 numbers negative. Keep the rest as 0.


2) Now make one of them as positive.
3) Observe what happens.
4) Now make the second positive.

What theory do you propose for convolution on the basis of the observation?

__________________________________________________________________________________
__________________________________________________________________________________
__________________________________________________________________________________
__________________________________________________________________________________
__________________________________________________________________________________
__________________________________________________________________________________

It is time to test the theory. Change the location of the four numbers and follow the above mentioned
steps. Does your theory hold true?

__________________________________________________________________________________
__________________________________________________________________________________
__________________________________________________________________________________
__________________________________________________________________________________
__________________________________________________________________________________
__________________________________________________________________________________
If yes, change the picture and try whether the theory holds true or not. If it does not hold true, modify
your theory and keep trying until it satisfies all the conditions.

Let’s Discuss
What effect did you apply?

__________________________________________________________________________________
__________________________________________________________________________________
__________________________________________________________________________________
__________________________________________________________________________________
__________________________________________________________________________________

How did different kernels affect the image?

__________________________________________________________________________________
__________________________________________________________________________________
__________________________________________________________________________________
__________________________________________________________________________________
__________________________________________________________________________________

Why do you think we apply these effects?

__________________________________________________________________________________
__________________________________________________________________________________
__________________________________________________________________________________
__________________________________________________________________________________
__________________________________________________________________________________

How do you think the convolution operator works?

__________________________________________________________________________________
__________________________________________________________________________________
__________________________________________________________________________________
__________________________________________________________________________________
__________________________________________________________________________________

Convolution : Explained
Convolution is a simple Mathematical operation which is fundamental to many
common image processing operators. Convolution provides a way of `multiplying together' two arrays
of numbers, generally of different sizes, but of the same dimensionality, to produce a third array of
numbers of the same dimensionality.

An (image) convolution is simply an element-wise multiplication of image arrays and another array
called the kernel followed by sum.
As you can see here,

I = Image Array

K = Kernel Array

I * K = Resulting array after performing the convolution operator

Note: The Kernel is passed over the whole image to get the resulting array after convolution.

What is a Kernel?
A Kernel is a matrix, which is slid across the image and multiplied with the input such that the output
is enhanced in a certain desirable manner. Each kernel has a different value for different kind of effects
that we want to apply to an image.

In Image processing, we use the convolution operation to extract the features from the images which
can le later used for further processing especially in Convolution Neural Network (CNN), about which
we will study later in the chapter.

In this process, we overlap the centre of the image with the centre of the kernel to obtain the
convolution output. In the process of doing it, the output image becomes smaller as the overlapping
is done at the edge row and column of the image. What if we want the output image to be of exact
size of the input image, how can we achieve this?

To achieve this, we need to extend the edge values out by one in the original image while overlapping
the centres and performing the convolution. This will help us keep the input and output image of the
same size. While extending the edges, the pixel values are considered as zero.

* Images shown here are the property of individual organisations and are used here for reference purpose only.
Let’s try
In this section we will try performing the convolution operator on paper to understand how it works.
Fill the blank places of the output images by performing the convolution operation.

150 0 255 240 190 25 89 255

100 179 25 0 200 255 67 100

155 146 13 20 0 12 45 0

100 175 0 25 25 15 0 0
-1 0 -1
120 156 255 0 78 56 23 0 0 -1 0
-1 0 -1
115 113 25 90 0 80 56 155

135 190 115 116 178 0 145 165

123 255 255 0 255 255 255 0

* Images shown here are the property of individual organisations and are used here for reference purpose only.
Write Your Output Here :

Summary
1. Convolution is a common tool used for image editing.
2. It is an element wise multiplication of an image and a kernel to get the desired output.
3. In computer vision application, it is used in Convolutional Neural Network (CNN) to extract
image features.

Convolution Neural Networks (CNN)


Introduction
In class 9, you studied about the concepts of Neural Network. You played a neural network game to
understand how a neural network works.

Let’s recall

What is a Neural Network?

__________________________________________________________________________________
__________________________________________________________________________________
__________________________________________________________________________________
__________________________________________________________________________________
__________________________________________________________________________________
Fill in the names of different layers of Neural Network.

Did you get the answers right? In this section, we are going to study about one such neural network
which is Convolutional Neural Network (CNN). Many of the current computer vision applications use
a powerful neural network called the convolutional neural network.

What is a Convolutional Neural Network?

A Convolutional Neural Network (CNN) is a Deep Learning algorithm which can take in an input image,
assign importance (learnable weights and biases) to various aspects/objects in the image and be able
to differentiate one from the other.

The process of deploying a CNN is as follows:

* Images shown here are the property of individual organisations and are used here for reference purpose only.
In the above diagram, we give an input image, which is then processed through a CNN and then gives
prediction on the basis of the label given in the particular dataset.

The different layers of a Convolutional Neural Network (CNN) is as follows:

A convolutional neural network consists of the following layers:

1) Convolution Layer
2) Rectified linear Unit (ReLU)
3) Pooling Layer
4) Fully Connected Layer

Convolution Layer

It is the first layer of a CNN. The objective of the Convolution Operation is to extract the high-level
features such as edges, from the input image. CNN need not be limited to only one Convolutional
Layer. Conventionally, the first Convolution Layer is responsible for capturing the Low-Level features
such as edges, colour, gradient orientation, etc. With added layers, the architecture adapts to the
High-Level features as well, giving us a network which has the wholesome understanding of images in
the dataset.

It uses convolution operation on the images. In the convolution layer, there are several kernels that
are used to produce several features. The output of this layer is called the feature map. A feature map
is also called the activation map. We can use these terms interchangeably.
There’s several uses we derive from the feature map:
• We reduce the image size so that it can be processed more efficiently.
• We only focus on the features of the image that can help us in processing the image further.
For example, you might only need to recognize someone’s eyes, nose and mouth to recognize the
person. You might not need to see the whole face.

* Images shown here are the property of individual organisations and are used here for reference purpose only.
Rectified Linear Unit Function

The next layer in the Convolution Neural Network is the Rectified Linear Unit function or the ReLU
layer. After we get the feature map, it is then passed onto the ReLU layer. This layer simply gets rid of
all the negative numbers in the feature map and lets the positive number stay as it is.

The process of passing it to the ReLU layer introduces non – linearity in the feature map. Let us see it
through a graph.

If we see the two graphs side by side, the one on the left is a linear graph. This graph when passed
through the ReLU layer, gives the one on the right. The ReLU graph starts with a horizontal straight
line and then increases linearly as it reaches a positive number.

Now the question arises, why do we pass the feature map to the ReLU layer? it is to make the colour
change more obvious and more abrupt?

* Images shown here are the property of individual organisations and are used here for reference purpose only.
As shown in the above convolved image, there is a smooth grey gradient change from black to white.
After applying the ReLu function, we can see a more abrupt change in color which makes the edges
more obvious which acts as a better feature for the further layers in a CNN as it enhances the
activation layer.

Pooling Layer

Similar to the Convolutional Layer, the Pooling layer is responsible for reducing the spatial size of the
Convolved Feature while still retaining the important features.

There are two types of pooling which can be performed on an image.

1) Max Pooling : Max Pooling returns the maximum value from the portion of the image covered
by the Kernel.
2) Average Pooling: Max Pooling returns the maximum value from the portion of the image
covered by the Kernel.

* Images shown here are the property of individual organisations and are used here for reference purpose only.
The pooling layer is an important layer in the CNN as it performs a series of tasks which are as
follows :

1) Makes the image smaller and more manageable


2) Makes the image more resistant to small transformations, distortions and translations in the
input image.

A small difference in input image will create very similar pooled image.

Fully Connected Layer

The final layer in the CNN is the Fully Connected Layer (FCP). The objective of a fully connected layer
is to take the results of the convolution/pooling process and use them to classify the image into a label
(in a simple classification example).

The output of convolution/pooling is flattened into a single vector of values, each representing a
probability that a certain feature belongs to a label. For example, if the image is of a cat, features
representing things like whiskers or fur should have high probabilities for the label “cat”.

* Images shown here are the property of individual organisations and are used here for reference purpose only.
Let’s Summarize:

Write the whole process of how a CNN works on the basis of the above diagram.
__________________________________________________________________________________
__________________________________________________________________________________
__________________________________________________________________________________
__________________________________________________________________________________
__________________________________________________________________________________
__________________________________________________________________________________
__________________________________________________________________________________
__________________________________________________________________________________
__________________________________________________________________________________
__________________________________________________________________________________
__________________________________________________________________________________
__________________________________________________________________________________
__________________________________________________________________________________
__________________________________________________________________________________
__________________________________________________________________________________
__________________________________________________________________________________
__________________________________________________________________________________
__________________________________________________________________________________
__________________________________________________________________________________
__________________________________________________________________________________
__________________________________________________________________________________
__________________________________________________________________________________
__________________________________________________________________________________
__________________________________________________________________________________

* Images shown here are the property of individual organisations and are used here for reference purpose only.
Let’s Experience
Now let us see how this comes into practice. To see that, go to the link
http://scs.ryerson.ca/~aharley/vis/conv/flat.html

This is an online application of classifying different numbers. We need to analyse the different layers
in the application on the basis of the CNN that we have studied in the previous section.
Natural Language Processing
Introduction
Till now, we have explored two domains of AI: Data Science and Computer Vision. Both these domains
differ from each other in terms of the data on which they work. Data Science works around numbers
and tabular data while Computer Vision is all about visual data like images and videos. The third
domain, Natural Language Processing (commonly called NLP) takes in the data of Natural Languages
which humans use in their daily lives and operates on this.

Natural Language Processing, or NLP, is the sub-field of AI that is focused on enabling computers to
understand and process human languages. AI is a subfield of Linguistics, Computer Science,
Information Engineering, and Artificial Intelligence concerned with the interactions between
computers and human (natural) languages, in particular how to program computers to process and
analyse large amounts of natural language data.

But how do computers do that? How do they understand what we say in our language? This chapter
is all about demystifying the Natural Language Processing domain and understanding how it works.

Before we get deeper into NLP, let us experience it with the help of this AI Game:

Identify the mystery animal: http://bit.ly/iai4yma

Go to this link on Google Chrome, launch the experiment and try to identify the Mystery Animal by
asking the machine 20 Yes or No questions.

Were you able to guess the animal?

__________________________________________________________________________________
__________________________________________________________________________________

If yes, in how many questions were you able to guess it?

__________________________________________________________________________________
__________________________________________________________________________________

If no, how many times did you try playing this game?

__________________________________________________________________________________
__________________________________________________________________________________

What according to you was the task of the machine?

__________________________________________________________________________________
__________________________________________________________________________________
__________________________________________________________________________________
__________________________________________________________________________________

* Images shown here are the property of individual organisations and are used here for reference purpose only.
Were there any challenges that you faced while playing this game? If yes, list them down.

__________________________________________________________________________________
__________________________________________________________________________________
__________________________________________________________________________________
__________________________________________________________________________________
__________________________________________________________________________________
__________________________________________________________________________________
__________________________________________________________________________________
__________________________________________________________________________________

What approach must one follow to win this game?

__________________________________________________________________________________
__________________________________________________________________________________
__________________________________________________________________________________
__________________________________________________________________________________
__________________________________________________________________________________
__________________________________________________________________________________
__________________________________________________________________________________
__________________________________________________________________________________

Applications of Natural Language Processing


Since Artificial Intelligence nowadays is becoming an integral part of our lives, its applications are very
commonly used by the majority of people in their daily lives. Here are some of the applications of
Natural Language Processing which are used in the real-life scenario:

Automatic Summarization: Information overload is a real


problem when we need to access a specific, important piece
of information from a huge knowledge base. Automatic
summarization is relevant not only for summarizing the
meaning of documents and information, but also to
understand the emotional meanings within the information,
such as in collecting data from social media. Automatic
summarization is especially relevant when used to provide an
overview of a news item or blog post, while avoiding
redundancy from multiple sources and maximizing the
diversity of content obtained.

Sentiment Analysis: The goal of sentiment


analysis is to identify sentiment among several
posts or even in the same post where emotion is
not always explicitly expressed. Companies use
Natural Language Processing applications, such as
sentiment analysis, to identify opinions and
sentiment online to help them understand what
customers think about their products and services
(i.e., “I love the new iPhone” and, a few lines later
“But sometimes it doesn’t work well” where the
person is still talking about the iPhone) and overall

* Images shown here are the property of individual organisations and are used here for reference purpose only.
indicators of their reputation. Beyond determining simple polarity, sentiment analysis understands
sentiment in context to help better understand what’s behind an expressed opinion, which can be
extremely relevant in understanding and driving purchasing decisions.

Text classification: Text classification makes it possible to assign


predefined categories to a document and organize it to help you
find the information you need or simplify some activities. For
example, an application of text categorization is spam filtering in
email.

Virtual Assistants: Nowadays Google Assistant, Cortana,


Siri, Alexa, etc have become an integral part of our lives. Not
only can we talk to them but they also have the abilities to
make our lives easier. By accessing our data, they can help
us in keeping notes of our tasks, make calls for us, send
messages and a lot more. With the help of speech
recognition, these assistants can not only detect our speech
but can also make sense out of it. According to recent
researches, a lot more advancements are expected in this
field in the near future.

Natural Language Processing: Getting Started


Natural Language Processing is all about how machines try to understand and interpret human
language and operate accordingly. But how can Natural Language Processing be used to solve the
problems around us? Let us take a look.

Revisiting the AI Project Cycle


Let us try to understand how we can develop a project in Natural Language processing with the help
of an example.

The Scenario
The world is competitive nowadays. People face
competition in even the tiniest tasks and are expected to
give their best at every point in time. When people are
unable to meet these expectations, they get stressed and
could even go into depression. We get to hear a lot of cases
where people are depressed due to reasons like peer
pressure, studies, family issues, relationships, etc. and they
eventually get into something that is bad for them as well
as for others. So, to overcome this, cognitive behavioural
therapy (CBT) is considered to be one of the best methods
to address stress as it is easy to implement on people and
also gives good results. This therapy includes

* Images shown here are the property of individual organisations and are used here for reference purpose only.
understanding the behaviour and mindset of a person in their normal life. With the help of CBT,
therapists help people overcome their stress and live a happy life.
To understand more about the concept of this therapy, visit this link:
https://en.wikipedia.org/wiki/Cognitive_behavioral_therapy

Problem Scoping
CBT is a technique used by most therapists to cure patients out of stress and depression. But it has
been observed that people do not wish to seek the help of a psychiatrist willingly. They try to avoid
such interactions as much as possible. Thus, there is a need to bridge the gap between a person who
needs help and the psychiatrist. Let us look at various factors around this problem through the 4Ws
problem canvas.

Who Canvas – Who has the problem?

Who are the


o People who suffer from stress and are at the onset of depression.
stakeholders?

What do we know
o People who are going through stress are reluctant to consult a psychiatrist.
about them?

What Canvas – What is the nature of the problem?

What is the o People who need help are reluctant to consult a psychiatrist and hence live
problem? miserably.

How do you know o Studies around mental stress and depression available on various authentic
it is a problem? sources.

Where Canvas – Where does the problem arise?

What is the context/situation


o When they are going through a stressful period of time
in which the stakeholders
o Due to some unpleasant experiences
experience this problem?

Why Canvas – Why do you think it is a problem worth solving?

o People get a platform where they can talk and vent out their
What would be of key feelings anonymously
value to the stakeholders? o People get a medium that can interact with them and applies
primitive CBT on them and can suggest help whenever needed

How would it improve their o People would be able to vent out their stress
situation? o They would consider going to a psychiatrist whenever required
Now that we have gone through all the factors around the problem, the problem statement templates
go as follows:

Our People undergoing stress Who?


Have a problem of Not being able to share their feelings What?
While They need help in venting out their emotions Where?
Provide them a platform to share their thoughts
An ideal solution would Why
anonymously and suggest help whenever required

This leads us to the goal of our project which is:

“To create a chatbot which can interact with people, help them
to vent out their feelings and take them through primitive CBT.”

Data Acquisition
To understand the sentiments of people, we need to collect their conversational data so the machine
can interpret the words that they use and understand their meaning. Such data can be collected from
various means:

1. Surveys 2. Observing the therapist’s sessions


3. Databases available on the internet 4. Interviews, etc.
Data Exploration
Once the textual data has been collected, it needs to be processed and cleaned so that an easier
version can be sent to the machine. Thus, the text is normalised through various steps and is lowered
to minimum vocabulary since the machine does not require grammatically correct statements but the
essence of it.

Modelling
Once the text has been normalised, it is then fed to an NLP based AI model. Note that in NLP, modelling
requires data pre-processing only after which the data is fed to the machine. Depending upon the type
of chatbot we try to make, there are a lot of AI models available which help us build the foundation of
our project.

Evaluation
The model trained is then evaluated and the accuracy for the same is generated on the basis of the
relevance of the answers which the machine gives to the user’s responses. To understand the
efficiency of the model, the suggested answers by the chatbot are compared to the actual answers.
As you can see in the above diagram, the blue line talks about the model’s output while the green one
is the actual output along with the data samples.

The model’s output does not match the true function at all. Hence the model is said
Figure 1 to be underfitting and its accuracy is lower.

In the second one, the model’s performance matches well with the true function
Figure 2 which states that the model has optimum accuracy and the model is called a
perfect fit.

In the third case, model performance is trying to cover all the data samples even if
Figure 3 they are out of alignment to the true function. This model is said to be overfitting
and this too has a lower accuracy.

Once the model is evaluated thoroughly, it is then deployed in the form of an app which people can
use easily.

Chatbots
As we have seen earlier, one of the most common applications of Natural Language Processing is a
chatbot. There are a lot of chatbots available and many of them use the same approach as we used in
the scenario above.. Let us try some of the chatbots and see how they work.

• Mitsuku Bot*
https://www.pandorabots.com/mitsuku/

• CleverBot*
https://www.cleverbot.com/

• Jabberwacky*
http://www.jabberwacky.com/

• Haptik*
https://haptik.ai/contact-us

* Images shown here are the property of individual organisations and are used here for reference purpose only.
• Rose*
http://ec2-54-215-197-164.us-west-1.compute.amazonaws.com/speech.php

• Ochatbot*
https://www.ometrics.com/blog/list-of-fun-chatbots/

Let us discuss!
• Which chatbot did you try? Name any one.
• What is the purpose of this chatbot?
• How was the interaction with the chatbot?
• Did the chat feel like talking to a human or a robot? Why do you think so?
• Do you feel that the chatbot has a certain personality?
As you interact with more and more chatbots, you would realise that some of them are scripted or in
other words are traditional chatbots while others were AI-powered and had more knowledge. With
the help of this experience, we can understand that there are 2 types of chatbots around us: Script-
bot and Smart-bot. Let us understand what each of them mean in detail:

Script-bot Smart-bot
Script bots are easy to make Smart-bots are flexible and powerful
Script bots work around a script which is Smart bots work on bigger databases and other
programmed in them resources directly
Mostly they are free and are easy to integrate Smart bots learn with more data
to a messaging platform
No or little language processing skills Coding is required to take this up on board
Limited functionality Wide functionality

The story speaker activity which was done in class 9 can be considered as a script-bot as in that activity
we used to create a script around which the interactive story revolved. As soon as the machine got
triggered by the person, it used to follow the script and answer accordingly. Other examples of script
bot may include the bots which are deployed in the customer care section of various companies. Their
job is to answer some basic queries that they are coded for and connect them to human executives
once they are unable to handle the conversation.

On the other hand, all the assistants like Google Assistant, Alexa, Cortana, Siri, etc. can be taken as
smart bots as not only can they handle the conversations but can also manage to do other tasks which
makes them smarter.

Human Language VS Computer Language


Humans communicate through language which we process all the time. Our brain keeps on processing
the sounds that it hears around itself and tries to make sense out of them all the time. Even in the
classroom, as the teacher delivers the session, our brain is continuously processing everything and
storing it in some place. Also, while this is happening, when your friend whispers something, the focus
of your brain automatically shifts from the teacher’s speech to your friend’s conversation. So now, the
brain is processing both the sounds but is prioritising the one on which our interest lies.

* Images shown here are the property of individual organisations and are used here for reference purpose only.
The sound reaches the brain through a long channel. As a person speaks, the sound travels from his
mouth and goes to the listener’s eardrum. The sound striking the eardrum is converted into neuron
impulse, gets transported to the brain and then gets processed. After processing the signal, the brain
gains understanding around the meaning of it. If it is clear, the signal gets stored. Otherwise, the
listener asks for clarity to the speaker. This is how human languages are processed by humans.

On the other hand, the computer understands the language of numbers. Everything that is sent to the
machine has to be converted to numbers. And while typing, if a single mistake is made, the computer
throws an error and does not process that part. The communications made by the machines are very
basic and simple.

Now, if we want the machine to understand our language, how should this happen? What are the
possible difficulties a machine would face in processing natural language? Let us take a look at some
of them here:

Arrangement of the words and meaning


There are rules in human language. There are nouns, verbs, adverbs, adjectives. A word can be a noun
at one time and an adjective some other time. There are rules to provide structure to a language.

This is the issue related to the syntax of the language. Syntax refers to the grammatical structure of a
sentence. When the structure is present, we can start interpreting the message. Now we also want to
have the computer do this. One way to do this is to use the part-of-speech tagging. This allows the
computer to identify the different parts of a speech.

Besides the matter of arrangement, there’s also meaning behind the language we use. Human
communication is complex. There are multiple characteristics of the human language that might be
easy for a human to understand but extremely difficult for a computer to understand.

Analogy with programming language:

Different syntax, same semantics: 2+3 = 3+2

Here the way these statements are written is different, but their meanings are the same that is 5.

Different semantics, same syntax: 2/3 (Python 2.7) ≠ 2/3 (Python 3)

Here the statements written have the same syntax but their meanings are different. In Python 2.7,
this statement would result in 1 while in Python 3, it would give an output of 1.5.

Think of some other examples of different syntax and same semantics and vice-versa.

__________________________________________________________________________________
__________________________________________________________________________________
__________________________________________________________________________________
__________________________________________________________________________________
__________________________________________________________________________________
__________________________________________________________________________________
__________________________________________________________________________________
__________________________________________________________________________________
__________________________________________________________________________________
__________________________________________________________________________________
__________________________________________________________________________________
__________________________________________________________________________________
Multiple Meanings of a word
Let’s consider these three sentences:

His face turned red after he found out that he took the wrong bag
What does this mean? Is he feeling ashamed because he took another person’s bag instead of his? Is
he feeling angry because he did not manage to steal the bag that he has been targeting?

The red car zoomed past his nose


Probably talking about the color of the car

His face turns red after consuming the medicine


Is he having an allergic reaction? Or is he not able to bear the taste of that medicine?

Here we can see that context is important. We understand a sentence almost intuitively, depending
on our history of using the language, and the memories that have been built within. In all three
sentences, the word red has been used in three different ways which according to the context of the
statement changes its meaning completely. Thus, in natural language, it is important to understand
that a word can have multiple meanings and the meanings fit into the statement according to the
context of it.

Think of some other words which can have multiple meanings and use them in sentences.

__________________________________________________________________________________
__________________________________________________________________________________
__________________________________________________________________________________
__________________________________________________________________________________
__________________________________________________________________________________
__________________________________________________________________________________
__________________________________________________________________________________
__________________________________________________________________________________
__________________________________________________________________________________
__________________________________________________________________________________
__________________________________________________________________________________
__________________________________________________________________________________
__________________________________________________________________________________
__________________________________________________________________________________
__________________________________________________________________________________
__________________________________________________________________________________

Perfect Syntax, no Meaning


Sometimes, a statement can have a perfectly correct syntax but it does not mean anything. For
example, take a look at this statement:

Chickens feed extravagantly while the moon drinks tea.

This statement is correct grammatically but does this make any sense? In Human language, a perfect
balance of syntax and semantics is important for better understanding.
Think of some other sentences having correct syntax and incorrect semantics.

__________________________________________________________________________________
__________________________________________________________________________________
__________________________________________________________________________________
__________________________________________________________________________________
__________________________________________________________________________________
__________________________________________________________________________________
__________________________________________________________________________________
__________________________________________________________________________________
__________________________________________________________________________________
__________________________________________________________________________________
__________________________________________________________________________________
__________________________________________________________________________________
__________________________________________________________________________________
__________________________________________________________________________________
__________________________________________________________________________________
__________________________________________________________________________________

These are some of the challenges we might have to face if we try to teach computers how to
understand and interact in human language. So how does Natural Language Processing do this magic?

Data Processing
Humans interact with each other very easily. For us, the natural languages that we use are so
convenient that we speak them easily and understand them well too. But for computers, our
languages are very complex. As you have already gone through some of the complications in human
languages above, now it is time to see how Natural Language Processing makes it possible for the
machines to understand and speak in the Natural Languages just like humans.

Since we all know that the language of computers is Numerical, the very first step that comes to our
mind is to convert our language to numbers. This conversion takes a few steps to happen. The first
step to it is Text Normalisation. Since human languages are complex, we need to first of all simplify
them in order to make sure that the understanding becomes possible. Text Normalisation helps in
cleaning up the textual data in such a way that it comes down to a level where its complexity is lower
than the actual data. Let us go through Text Normalisation in detail.

Text Normalisation
In Text Normalisation, we undergo several steps to normalise the text to a lower level. Before we
begin, we need to understand that in this section, we will be working on a collection of written text.
That is, we will be working on text from multiple documents and the term used for the whole textual
data from all the documents altogether is known as corpus. Not only would we go through all the
steps of Text Normalisation, we would also work them out on a corpus. Let us take a look at the steps:

Sentence Segmentation
Under sentence segmentation, the whole corpus is divided into sentences. Each sentence is taken as
a different data so now the whole corpus gets reduced to sentences.
Tokenisation
After segmenting the sentences, each sentence is then further divided into tokens. Tokens is a term
used for any word or number or special character occurring in a sentence. Under tokenisation, every
word, number and special character is considered separately and each of them is now a separate
token.

Removing Stopwords, Special Characters and Numbers


In this step, the tokens which are not necessary are removed from the token list. What can be the
possible words which we might not require?

Stopwords are the words which occur very frequently in the corpus but do not add any value to it.
Humans use grammar to make their sentences meaningful for the other person to understand. But
grammatical words do not add any essence to the information which is to be transmitted through the
statement hence they come under stopwords. Some examples of stopwords are:

* Images shown here are the property of individual organisations and are used here for reference purpose only.
These words occur the most in any given corpus but talk very little or nothing about the context or the
meaning of it. Hence, to make it easier for the computer to focus on meaningful terms, these words
are removed.

Along with these words, a lot of times our corpus might have special characters and/or numbers. Now
it depends on the type of corpus that we are working on whether we should keep them in it or not.
For example, if you are working on a document containing email IDs, then you might not want to
remove the special characters and numbers whereas in some other textual data if these characters do
not make sense, then you can remove them along with the stopwords.

Converting text to a common case


After the stopwords removal, we convert the whole text into a similar case, preferably lower case.
This ensures that the case-sensitivity of the machine does not consider same words as different just
because of different cases.

Here in this example, the all the 6 forms of hello would be converted to lower case and hence would
be treated as the same word by the machine.

Stemming
In this step, the remaining words are reduced to their root words. In other words, stemming is the
process in which the affixes of words are removed and the words are converted to their base form.

* Images shown here are the property of individual organisations and are used here for reference purpose only.
Note that in stemming, the stemmed words (words which are we get after removing the affixes) might
not be meaningful. Here in this example as you can see: healed, healing and healer all were reduced
to heal but studies was reduced to studi after the affix removal which is not a meaningful word.
Stemming does not take into account if the stemmed word is meaningful or not. It just removes the
affixes hence it is faster.

Lemmatization
Stemming and lemmatization both are alternative processes to each other as the role of both the
processes is same – removal of affixes. But the difference between both of them is that in
lemmatization, the word we get after affix removal (also known as lemma) is a meaningful one.
Lemmatization makes sure that lemma is a word with meaning and hence it takes a longer time to
execute than stemming.

As you can see in the same example, the output for studies after affix removal has become study
instead of studi.

* Images shown here are the property of individual organisations and are used here for reference purpose only.
Difference between stemming and lemmatization can be summarized by this example:

With this we have normalised our text to tokens which are the simplest form of words present in the
corpus. Now it is time to convert the tokens into numbers. For this, we would use the Bag of Words
algorithm

Bag of Words
Bag of Words is a Natural Language Processing model which helps in extracting features out of the
text which can be helpful in machine learning algorithms. In bag of words, we get the occurrences of
each word and construct the vocabulary for the corpus.

This image gives us a brief overview about how bag of words works. Let us assume that the text on
the left in this image is the normalised corpus which we have got after going through all the steps of
text processing. Now, as we put this text into the bag of words algorithm, the algorithm returns to us
the unique words out of the corpus and their occurrences in it. As you can see at the right, it shows us
a list of words appearing in the corpus and the numbers corresponding to it shows how many times
the word has occurred in the text body. Thus, we can say that the bag of words gives us two things:

1. A vocabulary of words for the corpus

2. The frequency of these words (number of times it has occurred in the whole corpus).

Here calling this algorithm “bag” of words symbolises that the sequence of sentences or tokens does
not matter in this case as all we need are the unique words and their frequency in it.

* Images shown here are the property of individual organisations and are used here for reference purpose only.
Here is the step-by-step approach to implement bag of words algorithm:

1. Text Normalisation: Collect data and pre-process it


2. Create Dictionary: Make a list of all the unique words occurring in the corpus. (Vocabulary)
3. Create document vectors: For each document in the corpus, find out how many times the
word from the unique list of words has occurred.
4. Create document vectors for all the documents.
Let us go through all the steps with an example:

Step 1: Collecting data and pre-processing it.

Document 1: Aman and Anil are stressed

Document 2: Aman went to a therapist

Document 3: Anil went to download a health chatbot

Here are three documents having one sentence each. After text normalisation, the text becomes:

Document 1: [aman, and, anil, are, stressed]

Document 2: [aman, went, to, a, therapist]

Document 3: [anil, went, to, download, a, health, chatbot]

Note that no tokens have been removed in the stopwords removal step. It is because we have very
little data and since the frequency of all the words is almost the same, no word can be said to have
lesser value than the other.

Step 2: Create Dictionary

Go through all the steps and create a dictionary i.e., list down all the words which occur in all three
documents:

Dictionary:

aman and anil are stressed went

download health chatbot therapist a to

Note that even though some words are repeated in different documents, they are all written just once
as while creating the dictionary, we create the list of unique words.

Step 3: Create document vector

In this step, the vocabulary is written in the top row. Now, for each word in the document, if it matches
with the vocabulary, put a 1 under it. If the same word appears again, increment the previous value
by 1. And if the word does not occur in that document, put a 0 under it.

* Images shown here are the property of individual organisations and are used here for reference purpose only.
Since in the first document, we have words: aman, and, anil, are, stressed. So, all these words get a
value of 1 and rest of the words get a 0 value.

Step 4: Repeat for all documents

Same exercise has to be done for all the documents. Hence, the table becomes:

In this table, the header row contains the vocabulary of the corpus and three rows correspond to three
different documents. Take a look at this table and analyse the positioning of 0s and 1s in it.

Finally, this gives us the document vector table for our corpus. But the tokens have still not converted
to numbers. This leads us to the final steps of our algorithm: TFIDF.

TFIDF: Term Frequency & Inverse Document Frequency


Suppose you have a book. Which characters or words do you think would occur the most in it?

__________________________________________________________________________________
__________________________________________________________________________________

Bag of words algorithm gives us the frequency of words in each document we have in our corpus. It
gives us an idea that if the word is occurring more in a document, its value is more for that document.
For example, if I have a document on air pollution, air and pollution would be the words which occur
many times in it. And these words are valuable too as they give us some context around the document.
But let us suppose we have 10 documents and all of them talk about different issues. One is on women
empowerment, the other is on unemployment and so on. Do you think air and pollution would still be
one of the most occurring words in the whole corpus? If not, then which words do you think would
have the highest frequency in all of them?

And, this, is, the, etc. are the words which occur the most in almost all the documents. But these words
do not talk about the corpus at all. Though they are important for humans as they make the
statements understandable to us, for the machine they are a complete waste as they do not provide
us with any information regarding the corpus. Hence, these are termed as stopwords and are mostly
removed at the pre-processing stage only.

* Images shown here are the property of individual organisations and are used here for reference purpose only.
Take a look at this graph. It is a plot of occurrence of words versus their value. As you can see, if the
words have highest occurrence in all the documents of the corpus, they are said to have negligible
value hence they are termed as stop words. These words are mostly removed at the pre-processing
stage only. Now as we move ahead from the stopwords, the occurrence level drops drastically and the
words which have adequate occurrence in the corpus are said to have some amount of value and are
termed as frequent words. These words mostly talk about the document’s subject and their
occurrence is adequate in the corpus. Then as the occurrence of words drops further, the value of
such words rises. These words are termed as rare or valuable words. These words occur the least but
add the most value to the corpus. Hence, when we look at the text, we take frequent and rare words
into consideration.

Let us now demystify TFIDF. TFIDF stands for Term Frequency and Inverse Document Frequency. TFIDF
helps un in identifying the value for each word. Let us understand each term one by one.

Term Frequency
Term frequency is the frequency of a word in one document. Term frequency can easily be found from
the document vector table as in that table we mention the frequency of each word of the vocabulary
in each document.

* Images shown here are the property of individual organisations and are used here for reference purpose only.
Here, you can see that the frequency of each word for each document has been recorded in the table.
These numbers are nothing but the Term Frequencies!

Inverse Document Frequency


Now, let us look at the other half of TFIDF which is Inverse Document Frequency. For this, let us first
understand what does document frequency mean. Document Frequency is the number of documents
in which the word occurs irrespective of how many times it has occurred in those documents. The
document frequency for the exemplar vocabulary would be:

Here, you can see that the document frequency of ‘aman’, ‘anil’, ‘went’, ‘to’ and ‘a’ is 2 as they have
occurred in two documents. Rest of them occurred in just one document hence the document
frequency for them is one.

Talking about inverse document frequency, we need to put the document frequency in the
denominator while the total number of documents is the numerator. Here, the total number of
documents are 3, hence inverse document frequency becomes:

Finally, the formula of TFIDF for any word W becomes:

TFIDF(W) = TF(W) * log( IDF(W) )

Here, log is to the base of 10. Don’t worry! You don’t need to calculate the log values by yourself.
Simply use the log function in the calculator and find out!

Now, let’s multiply the IDF values to the TF values. Note that the TF values are for each document
while the IDF values are for the whole corpus. Hence, we need to multiply the IDF values to each row
of the document vector table.

* Images shown here are the property of individual organisations and are used here for reference purpose only.
Here, you can see that the IDF values for Aman in each row is the same and similar pattern is followed
for all the words of the vocabulary. After calculating all the values, we get:

Finally, the words have been converted to numbers. These numbers are the values of each for each
document. Here, you can see that since we have less amount of data, words like ‘are’ and ‘and’ also
have a high value. But as the IDF value increases, the value of that word decreases. That is, for
example:

Total Number of documents: 10

Number of documents in which ‘and’ occurs: 10

Therefore, IDF(and) = 10/10 = 1

Which means: log(1) = 0. Hence, the value of ‘and’ becomes 0.

On the other hand, number of documents in which ‘pollution’ occurs: 3

IDF(pollution) = 10/3 = 3.3333…

Which means: log(3.3333) = 0.522; which shows that the word ‘pollution’ has considerable value in
the corpus.

Summarising the concept, we can say that:

1. Words that occur in all the documents with high term frequencies have the least values and
are considered to be the stopwords.
2. For a word to have high TFIDF value, the word needs to have a high term frequency but less
document frequency which shows that the word is important for one document but is not a
common word for all documents.
3. These values help the computer understand which words are to be considered while
processing the natural language. The higher the value, the more important the word is for a
given corpus.

* Images shown here are the property of individual organisations and are used here for reference purpose only.
Applications of TFIDF
TFIDF is commonly used in the Natural Language Processing domain. Some of its applications are:

Document Information
Topic Modelling Stop word filtering
Classification Retrieval System

Helps in classifying the To extract the Helps in removing the


It helps in predicting
type and genre of a important information unnecessary words
the topic for a corpus.
document. out of a corpus. out of a text body.

DIY – Do It Yourself!
Here is a corpus for you to challenge yourself with the given tasks. Use the knowledge you have
gained in the above sections and try completing the whole exercise by yourself.

The Corpus
Document 1: We can use health chatbots for treating stress.

Document 2: We can use NLP to create chatbots and we will be making health chatbots now!

Document 3: Health Chatbots cannot replace human counsellors now. Yay >< !! @1nteLA!4Y

Accomplish the following challenges on the basis of the corpus given above. You can use the tools
available online for these challenges. Link for each tool is given below:

1. Sentence Segmentation: https://tinyurl.com/y36hd92n

2. Tokenisation: https://text-processing.com/demo/tokenize/

3. Stopwords removal: https://demos.datasciencedojo.com/demo/stopwords/

4. Lowercase conversion: https://caseconverter.com/

5. Stemming: http://textanalysisonline.com/nltk-porter-stemmer

6. Lemmatisation: http://textanalysisonline.com/spacy-word-lemmatize

7. Bag of Words: Create a document vector table for all documents.

8. Generate TFIDF values for all the words.

9. Find the words having highest value.

10. Find the words having the least value.


Evaluation
Introduction
Till now we have learnt about the 4 stages of AI project cycle, viz. Problem scoping, Data acquisition,
Data exploration and modelling. While in modelling we can make different types of models, how do
we check if one’s better than the other? That’s where Evaluation comes into play. In the Evaluation
stage, we will explore different methods of evaluating an AI model. Model Evaluation is an integral
part of the model development process. It helps to find the best model that represents our data and
how well the chosen model will work in the future

What is evaluation?
Evaluation is the process of understanding the reliability of any AI model, based on outputs by feeding
test dataset into the model and comparing with actual answers. There can be different Evaluation
techniques, depending of the type and purpose of the model. Remember that It’s not recommended
to use the data we used to build the model to evaluate it. This is because our model will simply
remember the whole training set, and will therefore always predict the correct label for any point in
the training set. This is known as overfitting.

Firstly, let us go through various terms which are very important to the evaluation process.

Model Evaluation Terminologies


There are various new terminologies which come into the picture when we work on evaluating our
model. Let’s explore them with an example of the Forest fire scenario.

The Scenario
Imagine that you have come up with an AI based prediction model which has been deployed in a forest
which is prone to forest fires. Now, the objective of the model is to predict whether a forest fire has
broken out in the forest or not. Now, to understand the efficiency of this model, we need to check if
the predictions which it makes are correct or not. Thus, there exist two conditions which we need to
ponder upon: Prediction and Reality. The prediction is the output which is given by the machine and
the reality is the real scenario in the forest when the prediction has been made. Now let us look at
various combinations that we can have with these two conditions.
Case 1: Is there a forest fire?

Here, we can see in the picture that a forest fire has broken out in the forest. The model predicts a Yes
which means there is a forest fire. The Prediction matches with the Reality. Hence, this condition is
termed as True Positive.

Case 2: Is there a forest fire?

Here there is no fire in the forest hence the reality is No. In this case, the machine too has predicted
it correctly as a No. Therefore, this condition is termed as True Negative.

* Images shown here are the property of individual organisations and are used here for reference purpose only.
Case 3: Is there a forest fire?

Here the reality is that there is no forest fire. But the machine has incorrectly predicted that there is
a forest fire. This case is termed as False Positive.

Case 4: Is there a forest fire?

* Images shown here are the property of individual organisations and are used here for reference purpose only.
Here, a forest fire has broken out in the forest because of which the Reality is Yes but the machine has
incorrectly predicted it as a No which means the machine predicts that there is no Forest Fire.
Therefore, this case becomes False Negative.

Confusion matrix
The result of comparison between the prediction and reality can be recorded in what we call the
confusion matrix. The confusion matrix allows us to understand the prediction results. Note that it is
not an evaluation metric but a record which can help in evaluation. Let us once again take a look at
the four conditions that we went through in the Forest Fire example:

Let us now take a look at the confusion matrix:

Prediction and Reality can be easily mapped together with the help of this confusion matrix.

* Images shown here are the property of individual organisations and are used here for reference purpose only.
Evaluation Methods
Now as we have gone through all the possible combinations of Prediction and Reality, let us see how
we can use these conditions to evaluate the model.

Accuracy
Accuracy is defined as the percentage of correct predictions out of all the observations. A prediction
can be said to be correct if it matches the reality. Here, we have two conditions in which the Prediction
matches with the Reality: True Positive and True Negative. Hence, the formula for Accuracy becomes:

Here, total observations cover all the possible cases of prediction that can be True Positive (TP), True
Negative (TN), False Positive (FP) and False Negative (FN).

As we can see, Accuracy talks about how true the predictions are by any model. Let us ponder:

Is high accuracy equivalent to good performance?

__________________________________________________________________________________
__________________________________________________________________________________

How much percentage of accuracy is reasonable to show good performance?

__________________________________________________________________________________
__________________________________________________________________________________

Let us go back to the Forest Fire example. Assume that the model always predicts that there is no fire.
But in reality, there is a 2% chance of forest fire breaking out. In this case, for 98 cases, the model will
be right but for those 2 cases in which there was a forest fire, then too the model predicted no fire.

Here,

True Positives = 0

True Negatives = 98

Total cases = 100

Therefore, accuracy becomes: (98 + 0) / 100 = 98%

* Images shown here are the property of individual organisations and are used here for reference purpose only.
This is a fairly high accuracy for an AI model. But this parameter is useless for us as the actual cases
where the fire broke out are not taken into account. Hence, there is a need to look at another
parameter which takes account of such cases as well.

Precision
Precision is defined as the percentage of true positive cases versus all the cases where the prediction
is true. That is, it takes into account the True Positives and False Positives.

Going back to the Forest Fire example, in this case, assume that the model always predicts that there
is a forest fire irrespective of the reality. In this case, all the Positive conditions would be taken into
account that is, True Positive (Prediction = Yes and Reality = Yes) and False Positive (Prediction = Yes
and Reality = No). In this case, the firefighters will check for the fire all the time to see if the alarm was
True or False.

You might recall the story of the boy who falsely cries out that there are wolves every time and so
when they actually arrive, no one comes to his rescue. Similarly, here if the Precision is low (which
means there are more False alarms than the actual ones) then the firefighters would get complacent
and might not go and check every time considering it could be a false alarm.

* Images shown here are the property of individual organisations and are used here for reference purpose only.
This makes Precision an important evaluation criteria. If Precision is high, this means the True Positive
cases are more, giving lesser False alarms.

But again, is good Precision equivalent to a good model performance? Why?

__________________________________________________________________________________
__________________________________________________________________________________
__________________________________________________________________________________
__________________________________________________________________________________
__________________________________________________________________________________
__________________________________________________________________________________

Let us consider that a model has 100% precision. Which means that whenever the machine says
there’s a fire, there is actually a fire (True Positive). In the same model, there can be a rare exceptional
case where there was actual fire but the system could not detect it. This is the case of a False Negative
condition. But the precision value would not be affected by it because it does not take FN into account.
Is precision then a good parameter for model performance?

Recall
Another parameter for evaluating the model’s performance is Recall. It can be defined as the fraction
of positive cases that are correctly identified. It majorly takes into account the true reality cases where
in Reality there was a fire but the machine either detected it correctly or it didn’t. That is, it considers
True Positives (There was a forest fire in reality and the model predicted a forest fire) and False
Negatives (There was a forest fire and the model didn’t predict it).

* Images shown here are the property of individual organisations and are used here for reference purpose only.
Now as we notice, we can see that the Numerator in both Precision and Recall is the same: True
Positives. But in the denominator, Precision counts the False Positives while Recall takes False
Negatives into consideration.

Let us ponder… Which one do you think is better? Precision or Recall? Why?

__________________________________________________________________________________
__________________________________________________________________________________
__________________________________________________________________________________
__________________________________________________________________________________
__________________________________________________________________________________
__________________________________________________________________________________

Which Metric is Important?


Choosing between Precision and Recall depends on the condition in which the model has been
deployed. In a case like Forest Fire, a False Negative can cost us a lot and is risky too. Imagine no alert
being given even when there is a Forest Fire. The whole forest might burn down.

Another case where a False Negative can be dangerous is Viral Outbreak. Imagine a deadly virus has
started spreading and the model which is supposed to predict a viral outbreak does not detect it. The
virus might spread widely and infect a lot of people.

On the other hand, there can be cases in which the False Positive condition costs us more than False
Negatives. One such case is Mining. Imagine a model telling you that there exists treasure at a point
and you keep on digging there but it turns out that it is a false alarm. Here, False Positive case
(predicting there is treasure but there is no treasure) can be very costly.

Similarly, let’s consider a model that predicts that a mail is spam or not. If the model always predicts
that the mail is spam, people would not look at it and eventually might lose important information.
Here also False Positive condition (Predicting the mail as spam while the mail is not spam) would have
a high cost.
Think of some more examples having:

• High False Negative cost


__________________________________________________________________________________
__________________________________________________________________________________
__________________________________________________________________________________
__________________________________________________________________________________

• High False Positive cost


__________________________________________________________________________________
__________________________________________________________________________________
__________________________________________________________________________________
__________________________________________________________________________________

To conclude the argument, we must say that if we want to know if our model’s performance is good,
we need these two measures: Recall and Precision. For some cases, you might have a High Precision
but Low Recall or Low Precision but High Recall. But since both the measures are important, there is
a need of a parameter which takes both Precision and Recall into account.

F1 Score
F1 score can be defined as the measure of balance between precision and recall.

Take a look at the formula and think of when can we get a perfect F1 score?

An ideal situation would be when we have a value of 1 (that is 100%) for both Precision and Recall. In
that case, the F1 score would also be an ideal 1 (100%). It is known as the perfect value for F1 Score.
As the values of both Precision and Recall ranges from 0 to 1, the F1 score also ranges from 0 to 1.
Let us explore the variations we can have in the F1 Score:

Precision Recall F1 Score

Low Low Low

Low High Low

High Low Low

High High High

In conclusion, we can say that a model has good performance if the F1 Score for that model is high.

Let’s practice!
Let us understand the evaluation parameters with the help of examples.

Challenge
Find out Accuracy, Precision, Recall and F1 Score for the given problems.

Scenario 1:
In schools, a lot of times it happens that there is no water to drink. At a few places, cases of water
shortage in schools are very common and prominent. Hence, an AI model is designed to predict if
there is going to be a water shortage in the school in the near future or not. The confusion matrix for
the same is:
Scenario 2:
Nowadays, the problem of floods has worsened in some parts of the country. Not only does it damage
the whole place but it also forces people to move out of their homes and relocate. To address this
issue, an AI model has been created which can predict if there is a chance of floods or not. The
confusion matrix for the same is:

Scenario 3:
A lot of times people face the problem of sudden downpour. People wash clothes and put them out
to dry but due to unexpected rain, their work gets wasted. Thus, an AI model has been created which
predicts if there will be rain or not. The confusion matrix for the same is:
Scenario 4:
Traffic Jams have become a common part of our lives nowadays. Living in an urban area means you
have to face traffic each and every time you get out on the road. Mostly, school students opt for buses
to go to school. Many times the bus gets late due to such jams and students are not able to reach their
school on time. Thus, an AI model is created to predict explicitly if there would be a traffic jam on their
way to school or not. The confusion matrix for the same is:
CBSE | DEPARTMENT OF SKILL EDUCATION

ARTIFICIAL INTELLIGENCE
QUESTION BANK – CLASS 10
CHAPTER 1: INTRODUCTION TO AI: FOUNDATIONAL CONCEPTS

One (01) Mark Questions


Fill in the Blanks:
1. The basis of decision making depends upon the availability of ____________and how we
experience and understand it. (information/data/conditions/ past experience/
knowledge/awareness.)
2. A machine can also become intelligent if it is trained with __________ which helps them
achieve their tasks (data)

True/False:
1. A machine is artificially intelligent when it can accomplish tasks by itself. (True)
2. Is a smart washing machine an example of an Artificially Intelligent devices? (False)
3. Platforms like Netflix, Amazon, Spotify, YouTube etc. show us recommendations on the
basis of what we like. (True)

Direct Question:
1. What do you understand by linguistic Intelligence?
Linguistic intelligence means intelligence to understand and interpret human natural
language and try to extract meaning out of it.
OR
Linguistic Intelligence refers to the ability of an individual to understand both written
and spoken language and the additional ability to write and speak the language too.
OR
This is the intelligence of language and communication. It includes the ability to speak,
articulate, and express, and convey one’s thoughts and feelings to the outside world in
one or more languages. This can be at an oral and written level. It also includes the
ability to listen to and to understand other people.

2. What do you understand by Interpersonal Intelligence?


Understanding human emotions, feelings and influenced by them is known as
interpersonal intelligence.
OR
Interpersonal intelligence is the ability to communicate with others by understanding
other people's feelings, being influenced by the person.
OR

CBSE Question Bank – AI – Class 10 – Chapter 1 Foundational Concepts 1


Interpersonal intelligence refers to the ability of a person to relate well with people
and manage relationships. It enables people to understand the needs and motivations
of those around them, which helps strengthen their overall influence.
OR
Interpersonal intelligence refers to the ability to understand social situations and the
behavior of other people.

3. Define Artificial Intelligence.


A machine is artificially intelligent when it can accomplish tasks by itself - collect data,
understand it, analyze it, learn from it, and improve it.
OR
When a machine possesses the ability to mimic human traits, i.e., make decisions,
predict the future, learn and improve on its own, it is said to have artificial intelligence.
OR
Artificial Intelligence is a way of making a computer, a computer-controlled robot, or a
software think intelligently, in a similar manner to how intelligent humans think.
OR
AI is a form of intelligence; a type of technology and a field of study. AI theory and
development of computer systems (both machines and software) are able to perform
tasks that normally require human intelligence.
OR
Artificial Intelligence works to implement human intelligence in machines: creating
systems that understand, think, learn, and behave like humans.
OR
Artificial Intelligence or AI for short, refers to any technique that enables computers to
mimic human intelligence. An artificially intelligent machine works on algorithms and
data fed to it and gives the desired output.

4. Mention two types of machines which have evolved with time.


Television/Mobile Phones/ Ceiling Fans/ Microwave ovens/ Headphones / Speakers/
Harvesters/ Refrigerators/Air Conditioners etc.
(1 mark for any two right answers)
5. What do you understand by mathematical and logical reasoning?
A person's ability to regulate, measure, and understand numerical symbols,
abstraction and logic.
OR
Reasoning is based on previous established facts. To establish a new fact or truth one
has to put it to the test of reasoning. If the new fact coincides with the previously
established facts, it is called logical or rational. It is the ability of a person to regulate,
measure and understand numerical symbols, abstraction and logic.

CBSE Question Bank – AI – Class 10 – Chapter 1 Foundational Concepts 2


Two (02) Mark Questions

1. Mention four examples of artificially intelligent applications in our


smartphones.
Phone Smart Lock / Snapchat filter / Shopping websites / Netflix / YouTube / Face
Detection / Google Maps / Emotions recognition / Google assistant / Natural language
recognition / image detection / beauty filters etc. (2 marks for any four right
examples)

2. How does a machine become Artificially Intelligent?


A machine becomes intelligent by training with data and algorithm. AI machines keep
updating their knowledge to optimize their output.
OR
Machines also become intelligent once they are trained with some information which
helps them achieve their tasks. AI machines also keep updating their knowledge to
optimize their output.

3. Mention four examples of machines that are not AI but confused with AI.
Or
Mention four examples of machines that are smart but not AI.
Automatic gates in shopping malls / remote control drones/ a fully automatic washing
machine/ Air Conditioner/ Refrigerator/ Robotic toy cars/ Television etc.

4. How does learning and adapting help an AI machine in improvising itself?


An artificially intelligent machine collects real time data and tries to figure out new
patterns in it. Machines learn in a similar way human being; by supervision or by
observation and respond according to past experiences in similar scenarios. A machine
learns from its mistakes. The more the machine gets trained on data, the more
accurate result it gives.
For example:
Any virtual assistant initially trained with few basic instructions, but with time, the
machine captures the data fed by the user, may be the wake-up time of the user,
sleeping time, dinner time and so on. Later in time, the machine gives reminders of
similar things on the basis of data and adapts these new commands.
OR
Just as humans learn how to walk and then improve this skill with the help of their
experiences, an AI machine too gets trained first on the training data and then
optimizes itself according to its own experiences which makes AI different from any
other technological device/machine.

5. Pick the odd one out and justify your answer:


a. Snap Chat Filter b. Face Lock in Phone
c. Chatbot d. Image search Option
Ans: Chatbot (1 marks), as it is NLP based, the other three are Computer vision based
(1marks for justification).

CBSE Question Bank – AI – Class 10 – Chapter 1 Foundational Concepts 3


6. Explain how AI works in the following areas (any two):
a. Google Search Engine b. Voice Assistants c. E-commerce websites

a. Google Search Engine:


With the help of AI, Google Search Engine has been turned into Intelligent search
which is a new network of systems that produces direct answers. It uses voice and
image searches and has incorporated deep learning to fasten the searches with more
accuracy.
b. Voice assistant:
AI is being used in voice assistants to recognize words spoken by the user. NLP has
capabilities like “Speech-to-Text” convert the natural language of the user into text for
further processing. As the digital assistant answers more and more queries, it “learns”
using ML algorithms. The more tasks it performs, its ML algorithms help it “learn”
from the tasks and the preferences of the user. As a result, the digital assistant
improves its performance over time.
c. E-commerce website:
With the use of big data, AI in E-Commerce is impacting customer choices by recording
the data of previous purchases, searched products, and online browsing habits.
Product recommendations provide multiple benefits for E-commerce retailers
including: Higher number of returning customers.

7. How has AI changed the gaming world?


AI has changed the world of gaming by making the game more intelligent by providing
them the ability to learn using machine learning algorithms. Games these days try to
understand human patterns and give responses on the basis of it and also give new
difficulty levels.
OR
AI has changed the gaming world in terms of feel and emotions. Some video games
react to player skill level. Depending on how well you do, adaptive AI ratchets the
game’s difficulty level up and down to give you a greater challenge when you need it or
to prevent you from rage-quitting in frustration. AI can also adapt to your playing style
by making the game more exciting.

8. Why training with information/Data is important in Artificial Intelligent


devices?
Similar to human beings, AI devices need experience to give better results and
improve in every next iteration. For giving better results, the machine should be
trained with some real data. The more the amount of accurate data, the better
predictions will be made by the machine. Hence, data is very important in AI devices.
OR
The AI devices need to be trained with information / Big data to produce the best
possible accurate results. All of AI’s learning happens only through this data. So, it
makes sense to have as big a dataset as is required to include variety, subtlety, and
nuance that makes the model viable for practical use. Before training, the model is just
a theorist.

CBSE Question Bank – AI – Class 10 – Chapter 1 Foundational Concepts 4


4 Mark Questions

1. What is Intelligence? Explain in brief any three types of intelligence that are
mainly perceived by human beings?
Intelligence is the ‘ability to perceive or infer information, and to retain it as knowledge
to be applied towards adaptive behavior within an environment or context.’
OR
Intelligence is the ability to interact with the world (speech, vision, motion,
manipulation), ability to model the world and to reason about it, ability to learn, ability
to make decisions and to adapt.
OR
Intelligence has been defined in many ways: It involves abstract reasoning, mental
representation, problem solving, and decision making, the ability to learn, emotional
knowledge, creativity, and adaptation to meet the demands of the environment
effectively.

As per major researches, there are mainly 9 types of Intelligence;


(i) Mathematical Logical Intelligence: A person's ability to regulate, measure,
and understand numerical symbols, abstraction and logic
(ii) Linguistic Intelligence: Language processing skills both in terms of
understanding orimplementation in writing or speech.
(iii) Spatial Visual Intelligence: It is defined as the ability to perceive the visual
world and the relationship of one object to another.
(iv) Kinesthetic Intelligence: Ability that is related to how a person uses his limbs
in a skilled manner.
(v) Musical Intelligence: As the name suggests, this intelligence is about a
person's ability to recognize and create sounds, rhythms, and sound patterns
(vi) Intrapersonal Intelligence: Describes the level of self-awareness someone has
starting from realizing weakness, strength, to recognizing his own feelings
(vii) Existential Intelligence: An additional category of intelligence relating to
religious andspiritual awareness.
(viii) Naturalist Intelligence: An additional category of intelligence relating to the
ability to process information on the environment around us.
(ix) Interpersonal Intelligence: Interpersonal intelligence is the ability to
communicate with others by understanding other people's feelings and the
influence of the person.

2. Differentiate between what is AI and what is not AI with the help of an example?
AI Machine Not AI machine
1. AI machines are trained with data 1. Smart machines which are not AI, do
and algorithm. not require training data, they work on
2. AI machines learn from mistakes and algorithms only.
experience. They try to improvise on 2. Smart machines work on fixed
their next iterations. algorithms and they always work with
3. AI machines can analyses the the same level of efficiency, which is
situation and can take decisions programmed into them.

CBSE Question Bank – AI – Class 10 – Chapter 1 Foundational Concepts 5


AI Machine Not AI machine
accordingly. 3. Machines which are not AI cannot take
4. AI based drones capture the real-time decisions on their own.
data during the flight, processes it in 4. An automatic door in a shopping mall,
real-time, and makes a human- seems to be AI-enabled, but it is built
independent decision based on the with only sensor technology.
processed data.
(Any other valid examples of AI and Non-AI machine can be considered.)

While we see a lot of AI applications around us, there still exist a lot of them which are
smart but not intelligent.
An AI enabled machine should not only recognize, but should also do something with
its gathered information. Artificial intelligence” must mean a human-made interface
with the power to reason and integrate knowledge. AI must demonstrate at least some
of the following behaviors associated with human intelligence: planning, learning,
reasoning, problem solving, knowledge representation, perception, motion,
manipulation and, to a lesser extent, social intelligence, and creativity.
Most IOT items are ordinary things outfitted with sensors and connected to the
Internet. For example, sensors in your office can recognize shadows or movements,
but that doesn’t make them an example of artificial intelligence. A fully automatic
washing machine can work on its own, but it requires human intervention to select the
parameters of washing and to do the necessary preparation for it to function correctly
before each wash, which makes it an example of automation, not AI.

3. How can AI be integrated with non-AI technologies? Explain with the help of an
example.
Today’s world is changing with the adoption of IOT (Internet of Things). IOT is helping
in prominently capturing a tremendous amount of data from multiple sources. The
convergence of AI (Artificial Intelligence) and IOT can redefine the way industries,
business, and economies function. AI enabled IoT creates intelligent machines that
simulate smart behavior and supports decision making with little or no human
interference.

While IOT provides data, artificial intelligence acquires the power to unlock responses,
offering both creativity and context to drive smart actions. Here are some examples:

Ex. 1: Self-driving Cars: Tesla’s self-driving cars are the best example of IoT and AI
working together. With the power of AI, self-driving cars predict the behavior
of pedestrians and cards in various circumstances. For example, they can
determine road conditions, optimal speed, weather and getting smarter with
each trip.
Ex. 2: Robots in Manufacturing: Manufacturing is one of the industries that already
embraced new technologies like IoT, artificial intelligence, facial recognition,
deep learning, Robots and many more. Robots employed in factories are
turning smarter with the support of implanted sensors, which facilitates data
transmission. Moreover, as the robots are provisioned with artificial
intelligence algorithms, they can learn from newer data. This approach not only
saves time and cost but also makes the manufacturing process better over time.

CBSE Question Bank – AI – Class 10 – Chapter 1 Foundational Concepts 6


Ex.3: Weather forecasting System: In a weather forecasting system, where IOT
temperature sensor and humidity sensors collect data from the physical world,
AI tries to figure out patterns from previous data collected and tries to interpret
and give accurate predictions of upcoming day weather.
Ex.4: Smart Drones: Initially drones were only able to capture photographs, these
were not AI drones. As the scientist used to analyze the data captured through
drones. Now the drones are incorporated with AI, which helps them to make
decisions also on the basis of the picture they capture.

*Any one example from above or any other matching example can be given
4. Read the given scenario and answer the questions that follow:
A farmer keeps rabbits in three large hutches that stand in a row in his backyard.
Each of the hutches is painted different colours – red, yellow and green. Until
recently, the number of rabbits in the green hutch was twice as large as the
number of rabbits in the yellow hutch. Then, one day, the farmer took five rabbits
out of the left-side hutch and gave them away to the local school’s pet corner. He
also took half of the rabbits that remained in the left-side hutch and moved them
to the red hutch.
a. What was the colour of the left-side hutch? Justify your answer with
explanation.
The answer is yellow.
Explanation: As we already know at the outset the number of rabbits in the green
hutch was twice as large as the number of rabbits in the yellow hutch. This means that
the number of rabbits in the green hutch was an even number. After the farmer
removed five rabbits from the left side hutch, then the number of rabbits that
remained there also became an even number. This is proven by the fact that it was
divisible by 2. Therefore, before those five were removed, the left side hutch contained
an uneven number of rabbits hence the left side hutch cannot be the green one, but
based on the given information, it cannot be the red one. Hence it is yellow.

5. A scenario is given to you below. Read it and answer the questions that follow:
Late one night, a car ran over a pedestrian in a narrow by street and drove away
without stopping. A policeman who saw the vehicle leave the scene of the accident
reported it moving at very high speed. The accident itself was witnessed by six
bystanders. They provided the following conflicting accounts of what had
happened:
- It was a blue car driven by a man;
- The car was moving at high speed and its headlights were turned off;
- The car did have license plates; it wasn’t going very fast;
- It was a Toyota and its headlights were turned off;
- The car didn’t have license plates; the driver was a woman;
- It was a grey Ford.
When the car and its driver were finally apprehended, it turned out that only one
of the six eyewitnesses gave a fully correct description. Each of the other five
provided one true and one false piece of information. Keeping that in mind, can
you determine the following:

CBSE Question Bank – AI – Class 10 – Chapter 1 Foundational Concepts 7


a. What was the car’s brand? Ans: FORD
b. What was the colour of the car? Ans: BLUE
c. Was the car going fast or slow? Ans: FAST
d. Did it have license plates? Ans: NO
e. Were its headlights turned on? Ans: NO
f. Was the driver a man or a woman? Ans: WOMAN

Explanation: Out of the statements of 6 bystanders, the third statement becomes false
as the policeman who saw the vehicle leave the scene of the accident reported it
moving at very high speed. Then eliminating all false statements of bystanders, the
above results can be extracted.

6. A firefighter has to get to a burning building as quickly as he can. There are three
paths that he can take. He can take his fire engine over a large hill (5 miles) at 8
miles per hour. He can take his fire engine through a windy road (7 miles) at 9
miles per hour. Or he can drive his fire engine along a dirt road which is 8 miles at
12 miles per hour. Which way should he choose? (speed=distance/time)
To reach the destination quickly, the fire fighter has to calculate the time required on
the basis of given data. Driving his fire engine 5 miles at 8 miles per hour takes 37.5
minutes. Driving his fire engine 7 miles at 9 miles per hour takes about 47 minutes.
Driving his fire engine 8 miles at 12 miles per hour takes 40 minutes So he should
choose to drive his fire engine over the hill.)

7. A thief has just found a pair of ancient treasure caves. One of the caves is filled
with unbelievable treasure and the other has a fire breathing monster that will
eat anyone who opens that cave. One cave has a black door decorated with
diamonds and the other cave has a brown door decorated with sapphires. Each of
the doors has an engraved description on top. The descriptions say:
a. Black Door: Monster is here.
b. Brown Door: Only One Door speaks the truth.
Which door should the thief open?
The treasure is in the Black door.
Explanation: Let us look at the description on the Brown door. It can be correct or
wrong.
Scenario 1: The description on the Brown door is true. Then the description on the
Black door has to be false. That means that the inscription on the Black door is false
and the cave with black door contains the treasure!
Scenario 2: The description on the Brown door is false. Then either both the
descriptions are false or both are true. Both cannot be true as that is impossible and
not consistent. That means that both descriptions are false.

CBSE Question Bank – AI – Class 10 – Chapter 1 Foundational Concepts 8


8. How intelligent robots are helping us in accomplishing dangerous jobs?
Robots let humans avoid some hurtful work:
(i) Lifting up heavy material at the construction site.
(ii) Stirring and mixing metals or liquids at a high temperature.
(iii) Collecting and packaging of radioactive waste.
(iv) Working in contaminated and dusty environments.

9. How AI helps in giving you personalized experience online?

AI based recommendations: AI uses advanced machine learning algorithms to analyze


browser history, page clicks, social interactions (likes, shares), past purchases, the
duration for which a page was viewed, location, etc. to gauge customer interests and
preferences. AI can help deliver product recommendations based on frequently bought
items, or related products. It can even help customize web pages and elements to suit a
customer’s needs. For instance, Netflix does intense behavior analysis based on
behavior and demographic data to determine the content that will resonate with their
customers.
Chatbots and Automated Messaging: AI-powered chatbots and messaging agents can
enhance the customer experience across channels. They can answer simple queries,
engage customers, efficiently handle multiple interactions,
Automated Service Interactions: AI-driven programs can send automated messages to
customers regarding a pending service, a part replacement, or a regular order.
Curating Select Products: Amazon has come up with the concept of the Amazon 4-star
retail store. Products that have received a multitude of 4-star ratings will be offered in
this physical store. Amazon will use its product recommendation engine to identify
trending products and customers’ favorites and bring them to a brick and mortar
setting.

CBSE Question Bank – AI – Class 10 – Chapter 1 Foundational Concepts 9


CBSE | DEPARTMENT OF SKILL EDUCATION
ARTIFICIAL INTELLIGENCE
QUESTION BANK – CLASS 10
CHAPTER 2: INTRODUCTION TO AI: BASICS OF AI
One (01) Mark Questions

Fill in the blanks:


1. One of the major sources of data for many major companies is the device which all of us
have in our hands all the time______________ (Smartphone/ Mobile Phones)
2. The world of Artificial Intelligence revolves around ___________ (Data)

True/False:
3. All the apps collect some kind of data. (True)

Direct Questions:
4. What do you understand by Machine Learning?
Machine Learning is a subset of Artificial Intelligence which enables machines to improve
at tasks with experience (data). The intention of Machine Learning is to enable machines
to learn by themselves using the provided data and make accurate Predictions/
Decisions.
OR
Machine learning focuses on the development of computer programs that can access data
and use it to learn for themselves.
OR
Machine learning is a data analytics technique that teaches computers to do what comes
naturally to humans and animals: learn from experience.

5. What do you understand by Deep Learning?


Deep Learning is the most advanced form of Artificial Intelligence. In Deep Learning, the
machine is trained with huge amounts of data which helps it in training itself around the
data. Such machines are intelligent enough to develop algorithms for themselves.
OR
Deep learning is an artificial intelligence (AI) function that imitates the workings of the
human brain in processing data and creating patterns for use in decision making.
OR
Deep learning is a subset of machine learning where artificial neural networks,
algorithms inspired by the human brain, learn from large amounts of data.

CBSE QUESTION BANK – AI – CLASS 10 – CHAPTER 2 INTRODUCTION TO AI: BASICS 1


OF AI
6. What are the three domains of AI?
● Data Science/ Big Data
● Computer Vision
● Natural Language Processing (NLP)

7. Name any two examples of Data science?


(Any two out of the following)

Price Comparison Websites/ Website Recommendations/ Fraud and Risk detection/


Internet search/ Personalized healthcare recommendations / Optimizing Traffic routes
in real-time / image tagging.

8. Name any two examples of Computer vision?


(Any two out of the following)

Self-Driving cars/ Autonomous vehicles Face Lock in Smartphones/ MedicalImaging/


Facial recognition /Security Systems / Waste Management / Satellite imaging.

9. Name any two examples of Natural Language Processing?


(Any two out of the following)

Email filters/Smart assistants/ Sentiment Analysis/Automatic Summarization/Search


results / Language translation / Digital phone calls.

10. Name any two examples of Machine Learning?


(Any two out of the following)
Virtual Personal Assistants, Recommendation systems like Netflix, Face Apps, Online
Fraud Detection

New Addition
MCQ (Correct answers are highlighted)

11. Snapchat filters use _____ and _____ to enhance your selfie with flowers, cat ears etc.
a) machine learning and deep learning
b) data and image processing
c) augmented reality and machine learning
d) NLP and computer vision

12. Based on the image below, choose the correct domain or domains of AI required for it:

CBSE QUESTION BANK – AI – CLASS 10 – CHAPTER 2 INTRODUCTION TO AI: BASICS 2


OF AI
a) Data
b) NLP
c) Computer Vision
d) Both (a) and (b)

13. Rock paper and scissors game is based on the following domain:
a) Data for AI
b) Natural Language Processing
c) Computer Vision
d) Image processing

14. Select a game which is based on Data Science domain of AI:


a) Rock Paper and Scissors b) Mystery Animal
c) Emoji Scavenger Hunt d) Pokémon

15. Identify the domain of AI in the following image:

a) Data Science
b) Natural Language Processing
c) Computer Vision
d) Rule Based

Two (02) Mark Questions


1. What is Data science? Give an example of it.
Data sciences is a domain of AI related to data systems and processes, in which the system
collects numerous data, maintains data sets and derives meaning/sense out of them. The
information extracted through data science can be used to make a decision about it.
OR
Data science is the field of study that combines domain expertise, programming skills, and
knowledge of mathematics and statistics to extract meaningful insights from data.
OR
Data Sciences, it is a concept to unify statistics, data analysis, machine learning and their
related methods in order to understand and analyses actual phenomena with data.

For example: a company that has petabytes of user data may use data science to develop
effective ways to store, manage, and analyze the data.
CBSE QUESTION BANK – AI – CLASS 10 – CHAPTER 2 INTRODUCTION TO AI: BASICS 3
OF AI
2. What is Computer Vision? Give an example of it.
Computer Vision, abbreviated as CV, is a domain of AI that depicts the capability of a
machine to get and analyze visual information and afterwards predict some decisions about
it. The entire process involves image acquiring, screening, analyzing, identifying and
extracting information.
OR
Computer vision is an interdisciplinary field that deals with how computers can be made to
gain high-level understanding from digital images or videos.
OR
The Computer Vision domain of Artificial Intelligence, enables machines to see through
images or visual data, process and analyze them on the basis of algorithms and methods in
order to analyze actual phenomena with images.

For Example: - Self-Driving cars/ Automatic Cars, Face Lock in Smartphones

3. What is Natural Language Processing? Give an example of it.


Natural Language Processing, abbreviated as NLP, is a branch of artificial intelligence that
deals with the interaction between machine/computers and humans using the natural
language. Natural language refers to language that is spoken and written by people, and
natural language processing (NLP) attempts to extract information from the spoken and
written word using algorithms.
OR
Natural Language Processing, or NLP, is the sub-field of AI that is focused on enabling
machine/computers to understand and process human languages. AI is a subfield of
Linguistics, Computer Science, Information Engineering, and Artificial Intelligence
concerned with the interactions between computers and human (natural) languages, in
particular how to program computers to process and analyze large amounts of natural
language data.
OR
In NLP, we teach machines how to understand and communicate in human
language. Natural language refers to speech analysis in both audible speeches, as well as text
of a language. NLP systems capture meaning from an input of words (sentences,
paragraphs, pages, etc.)

For Example: Email filters, Smart assistants: - Apple’s Siri and Amazon’s Alexa

4. Where do we collect data from?


Data can be collected from various sources like –
 Surveys
 Sensors
 Observations
 Web scrapping (Internet)
 Interviews
 Documents and records.
 Oral histories
CBSE QUESTION BANK – AI – CLASS 10 – CHAPTER 2 INTRODUCTION TO AI: BASICS 4
OF AI
5. Why do we need to collect data?
Data to a machine is similar to food for human being to function. The world of Artificial
Intelligence revolves around Data. Every company whether small or big is collecting data
from as many sources as possible. Data is called the New Gold today. It is through data
collection that a business or management has the quality information they need to make
informed decisions from further analysis, study, and research. Data collection allows them
to stay on top of trends, provide answers to problems, and analyze new insights to great
effect.

6. What is data mining? Explain with example.


Data mining is the process of analyzing large data sets and extracting the useful
information from it. Data mining is used by companies to turn raw data into useful
information. It is an interdisciplinary subfield of computer science and statistics with an
overall goal to extract information
OR
Data mining is an automatic or semi-automatic technical process that analyses large
amounts of scattered information to make sense of it and turn it into knowledge. It looks
for anomalies, patterns or correlations among millions of records to predict results, as
indicated by the SAS institute, a world leader in business analytics.

Example:
Price Comparison websites- They collect data about a product from different sites and then
analyze trends out of it and show up the most appropriate results.

Data mining is also known as Knowledge Discovery in Data (KDD)


To be moved to chapter no. 3

7. What do you understand by Data Privacy?


The world of Artificial Intelligence revolves around Data. Proper and ethical handling of
own data or user data is called data privacy. It is all about the rights of individuals with
respect to their personal information.

Data privacy or information privacy is a branch of data security concerned with the proper
handling of data – consent, notice, and regulatory obligations. More specifically,
practical data privacy concerns often revolve around: Whether or how data is shared with
third parties

8. Is data which is collected by various applications ethical in nature? Justify your


Yes, most of the times, the data collected by various applications is ethical in nature as the
users agree to it by clicking on allow when the application asks for various permissions.
They ask for our data for various facilities like - to show us personalized recommendations
and advertisements and to make their app more accurate and efficient.
OR

CBSE QUESTION BANK – AI – CLASS 10 – CHAPTER 2 INTRODUCTION TO AI: BASICS 5


OF AI
No, the data collected by various applications is not always ethical in nature. Sometimes, we
just share our data to non – trusted third party applications without reading what happens
to our data. This may lead to unethical use of our data. If one does not want to share his/her
data with anyone, he/she can opt for alternative applications which are of similar usage and
keep the data private. For example, an alternative to WhatsApp is the Telegram app which
does not collect any data from us.

Note: This is an open-ended question, so both the answers yes/no will be considered right
with correct justification.

9. Fill in the blanks for the image given below:

CBSE QUESTION BANK – AI – CLASS 10 – CHAPTER 2 INTRODUCTION TO AI: BASICS 6


OF AI
Three (03) Mark Questions
1. What do you understand by AI bias? Discuss in detail with some examples.
AI bias is the underlying prejudice in data that’s used to create AI algorithms, which can
ultimately result in discrimination and other social consequences.
AI Bias can creep into algorithms in several ways. AI systems learn to make decisions
based on training data, which can include biased human decisions or reflect historical or
social inequities, even if sensitive variables such as gender, race, or sexual orientation are
removed. Amazon stopped using a hiring algorithm after finding it favored applicants
based on words like “executed” or “captured” that were more commonly found on men’s
resumes, for example. Another source of bias is flawed data sampling, in which groups are
over- or underrepresented in the training data.
For Example
 Majorly, all the virtual assistants have a female voice. It is only now that some
companies have understood this bias and have started giving options for male voices
but since the virtual assistants came into practice, female voices are always preferred
for them over any other voice. Can you think of some reasons for this?
 If you search on Google for salons, the first few searches are mostly for female salons.
This is based on the assumption that if a person is searching for a salon, in all
probability it would be a female. Do you think this is a bias? If yes, then is it a Negative
bias or Positive one?

2. What do you understand by Data Privacy? Discuss in detail with some examples.
Data privacy, sometimes also referred to as information privacy, is an area of data
protection that concerns the proper handling of sensitive data including,
notably, personal data but also other confidential data, such as certain financial data and
intellectual property data, to meet regulatory requirements as well as protecting the
confidentiality and immutability of the data. It focuses on how to collect, process, share,
archive, and delete the data in accordance with the law.
Privacy, in the broadest sense, is the right of individuals, groups, or organizations
to control who can access, observe, or use something they own, such as their bodies,
property, ideas, data, or information.
Control is established through physical, social, or informational boundaries that help
prevent unwanted access, observation, or use. For example:
 A physical boundary, such as a locked front door, helps prevent others from entering a
building without explicit permission in the form of a key to unlock the door or a
person inside opening the door.
 A social boundary, such as a members-only club, only allows members to access and
use club resources.
 An informational boundary, such as a non-disclosure agreement, restricts what
information can be disclosed to others.
Privacy of information is extremely important in this digital age where everything is
interconnected and can be accessed and used easily. The possibilities of our private
information being extremely vulnerable are very real, which is why we require data
privacy.
CBSE QUESTION BANK – AI – CLASS 10 – CHAPTER 2 INTRODUCTION TO AI: BASICS 7
OF AI
3. What do you understand by AI, ML & DL? How are they different from each other?

a) Artificial Intelligence (AI)


AI is incorporating human intelligence to machines. Whenever a machine completes
tasks based on a set of rules that solve problems (algorithms), such an “intelligent”
behavior is what is called artificial intelligence.

b) Machine Learning (ML)


ML is a subset of AI that uses statistical learning algorithms to build smart systems. The
ML systems can automatically learn and improve without explicitly being programmed.

c) Deep Learning (DL)


In Deep Learning, the machine is trained with huge amounts of data which helps it in
training itself around the data. Such machines are intelligent enough to develop
algorithms for themselves.

How they differ?


 Deep Learning is the most advanced form of Artificial Intelligence out of these three. Then
comes Machine Learning which is intermediately intelligent and Artificial intelligence
covers all the concepts and algorithms which, in some way or the other mimic human
intelligence.
 Therefore, AI is the umbrella term which covers ML and DL.

4. Why do apps collect data in our phone?


One of the major sources of data for many major companies is the device which all of us have
in our hands all the time: Smartphones. Smartphones have nowadays become an integral
part of our lives. Most of us use smartphones more than we interact with people around us.
For the facilities that smartphones provide us, Apps need a lot of data which is collected
from the user like details about your face, browsing history, or your geographic location,
contact list etc. This data is collected with user’s consent which he/she gives at the time of
installing an app by clicking on “yes” or “allow” options which clearly means that we
ourselves are giving permissions to the Apps.
Permissions by themselves are harmless and even useful to provide users a good mobile
experience.
This data is collected to provide us with a lot of facilities and features which have made our
lives easier. Another reason to collect the data is to provide us with customized
recommendations and notifications according to our choices.
One more reason to collect the data is to make their app more accurate and efficient.

CBSE QUESTION BANK – AI – CLASS 10 – CHAPTER 2 INTRODUCTION TO AI: BASICS 8


OF AI
5. Should AI replace laborious jobs? Is there an alternative for major unemployment?

Yes, AI should replace laborious jobs.


● AI can replace laborious jobs like lifting of heavy items, working in mines etc.
● AI can indeed automate most repetitive and physical tasks.
● In future, AI would be a good option in the field of architecture and construction.
OR
No, AI should not replace laborious jobs completely as if it replaces laborious jobs
completely, then there will be no source of income for the daily wage workers due to
unemployment. So, industry owners can use some machines but more of man power. Hence
the production will not get affected as humans are smarter than machines since they were
the ones who invented AI.
Note: As this is an open-ended question so both the answers (yes/No) are correct but it
must be with correct justification.

Is there an alternative for major unemployment?


● AI taking over laborious jobs won’t create unemployment. It is just a groundless fear.
The standard view of technical change is that some jobs are displaced by the
substitution of machines for labour, but that the fear of total displacement is misplaced
because new jobs are created, largely due to the technology-fuelled increase in
productivity. Humans have always shifted away from work suitable for machines and to
other jobs.
● The basic fact is that technology eliminates jobs, not work. If this level of AI revolution
will happen, lots of job opportunities will be created. For example: 20-30 years ago,
being an accountant was a lucrative job, but AI took over this job but this created a lot
of opportunities, it raised the demand of a software engineer, data scientist, etc.
● It will open doors to skillful jobs rather than doing laborious tasks.
● Thus, we will be able to cope with the level of major unemployment, if AI took over
laborious jobs.

6. As Artificially Intelligent machines become more and more powerful, their ability to
accomplish tedious tasks is becoming better. Hence, it is now that AI machines have
started replacing humans in factories. While people see it in a negative way and say AI
has the power to bring mass unemployment and one day, machines would enslave
humans, on the other hand, other people say that machines are meant to ease our lives.
If machines over take monotonous and tedious tasks, humans should upgrade their
skills to remain their masters always.
What according to you is a better approach towards this ethical concern? Justify your
answer.
● AI taking over laborious jobs won’t create unemployment. It is just a groundless fear.
The standard view of technical change is that some jobs are displaced by the
substitution of machines for labour, but that the fear of total displacement is misplaced
because new jobs are created, largely due to the technology-fuelled increase in
productivity. Humans have always shifted away from work suitable for machines and to
other jobs.
CBSE QUESTION BANK – AI – CLASS 10 – CHAPTER 2 INTRODUCTION TO AI: BASICS 9
OF AI
● The basic fact is that technology eliminates jobs, not work. If this level of AI revolution
will happen, lots of job opportunities will be created. For example: 20-30 years ago,
being an accountant was a lucrative job, AI took over this job but this created a lot of
opportunities, it raised the demand of a software engineer, data scientist, etc.
● It will open doors to skillful jobs rather than doing laborious tasks.
● Thus, we will be able to cope with the level of major unemployment, if AI took over
laborious jobs.

7. List down various sensors that are present in a smartphone. Also list down the type
of data which gets collected through them.
 ACCELEROMETER [ helps running AR applications and track steps]
 GPS [ Location Data]
 Gyroscope [Orientation Data]
 Magnetometer [ Direction and Magnetic Field Data]
 Biometric Sensors [Fingerprint ,Iris, Face data]

New Additions

1. (Case Study) AI and robotics have raised some questions regarding liability. Take for
example the scenario of an ‘autonomous’ or AI-driven robot moving through a
factory. Another robot surprisingly crosses its way and our robot draws aside to
prevent collision. However, by this manoeuvre the robot injures a person.

a) Who can be held liable for damages caused by autonomous systems?


It is actually very difficult to blame anyone in such a scenario. Here is the situation
where AI Ethics come in to the picture. Here, the choices might differ from person to
person and one must understand that nobody is wrong in this case. Every person has a
different perspective and hence he/she takes decisions according to their moralities.
But still if someone is to be liable then it should be the programmer who has designed
the algorithm of the autonomous vehicle as he/she should have considered all the
exceptional conditions that could arise.

b) List two AI Ethics.


(Any two out of the following)
AI Bias, AI Access, Data privacy, AI for kids.

CBSE QUESTION BANK – AI – CLASS 10 – CHAPTER 2 INTRODUCTION TO AI: BASICS 10


OF AI
CBSE | DEPARTMENT OF SKILL EDUCATION

ARTIFICIAL INTELLIGENCE
QUESTION BANK – CLASS 10
CHAPTER 3: AI PROJECT CYCLE

One (01) Mark Questions

1. Name all the stages of an AI Project cycle.


Problem Scoping, Data Acquisition, Data Exploration, Modeling, Evaluation
2. What are sustainable development goals?
The Sustainable Development Goals (SDGs), also known as the Global Goals, were
adopted by all United Nations Member States in 2015 as a universal call to action to
end poverty, protect the planet and ensure that all people enjoy peace and prosperity.
OR
The Sustainable Development Goals (SDGs) or Global Goals are a collection of 17
interlinked goals designed to be a "blueprint to achieve a better and more sustainable
future for all" so that the future generations may live in peace and prosperity.
3. Name the 4Ws of problem canvases under the problem scoping stage of the AI
Project Cycle.
a. Who, b. what c. where d. why
4. What is Testing Dataset?
The dataset provided to the model ML. algorithm after training the algorithm
5. Mention the types of learning approaches for AI modeling.
Supervised, unsupervised and re-enforcement
6. What is the objective of evaluation stage?
It is to evaluate whether the ML algorithm is able to predict with high accuracy or not
before deployment.
7. Fill in the blank:
The analogy of an Artificial Neural Network can be made with _____________?
(Parallel Processing)
8. Which of the following is not an authentic source for data acquisition?
a. Sensors b. Surveys c. Web Scraping d. System Hacking
System Hacking
9. Which type of graphical representation suits best for continuous type of data
like monthly exam scores of a student?
Linear graph
10. Fill in the blank: Neural Network is a mesh of multiple _____________________.
Hidden Layers / Layers

CBSE Question Bank – AI – Class 10 – Chapter 3 AI Project Cycle 1


Two (02) Mark Questions

1. What are the two different approaches for AI modelling? Define them.
There are two approaches for AI Modelling; Rule Based and Learning Based.
The Rule based approach generates pre-defined outputs based on certain rules
programmed by humans. Whereas, machine learning approach has its own rules based
on the output and data used to train the models.
OR
Rule Based Approach Refers to the AI modelling where the relationship or patterns in
data are defined by the developer. The machine follows the rules or instructions
mentioned by the developer, and performs its task accordingly. Whereas in Learning
based approach, the relationship or patterns in data are not defined by the developer.
In this approach, random data is fed to the machine and it is left to the machine to
figure out patterns and trends out of it
2. What is a problem statement template and what is its significance?
The problem statement template gives a clear idea about the basic framework
required to achieve the goal. It is the 4Ws canvas which segregates; what is the
problem, where does it arise, who is affected, why is it a problem? It takes us straight
to the goal.
3. Explain any two SDGs in detail.
1. No Poverty: This is Goal 1 and strives to End poverty in all its forms everywhere
globally by 2030. The goal has a total of seven targets to be achieved.
2. Quality Education: This is Goal 4 which aspires to ensure inclusive and equitable
quality education and promote lifelong learning opportunities for all. It has 10 targets
to achieve.
* (Any two goals can be defined)
4. Mention the precautions to be taken while acquiring data for developing an AI
Project.
It should be from an authentic source, and accurate. Look for redundant and irrelevant
data parameters that does not take part in prediction.
5. What do you mean by Data Features?
The type of data to collect,It should be relevant data.
6. Write the names for missing stages in the given AI project cycle:

Problem scoping, Evaluation


7. Draw the icons of the following SDGs:
Gender Equality Clean Water and sanitation

CBSE Question Bank – AI – Class 10 – Chapter 3 AI Project Cycle 2


8. Draw the graphical representation of Classification AI model. Explain in brief.
Classification: The classification Model works on the labelled data. For example, we
have 3 coins of different denomination which are labelled according to their weight
then the model would look for the labelled features for predicting the output. This
model works on discrete dataset which means the data need not be continuous.
OR
In classification, data is categorized under different labels according to some
parameters given in input and then the labels are predicted for the data.

9. Draw the graphical representation of Regression AI model. Explain in brief.


Regression: These models work on continuous data to predict the output based on
patterns. For example, if you wish to predict your next salary, then you would put in
the data of your previous salary, any increments, etc., and would train the model.
Here, the data which has been fed to the machine is continuous.
OR
Regression is the process of finding a model for distinguishing the data into
continuous real values instead of using discrete values. It can also identify the
distribution movement depending on the historical data.

10. Draw the graphical representation of Clustering AI model. Explain in brief.


Clustering: It refers to the unsupervised learning algorithm which can cluster the
unknown data according to the patterns or trends identified out of it. The patterns
observed might be the ones which are known to the developer or it might even come
up with some unique patterns out of it.
OR
Clustering is the task of dividing the data points into a number of groups such that
data points in the same groups are more similar to other data points in the same

CBSE Question Bank – AI – Class 10 – Chapter 3 AI Project Cycle 3


group and dissimilar to the data points in other groups. It is basically a collection of
objects on the basis of similarity and dissimilarity between them.

11. Explain Data Exploration stage.


In this stage of project cycle, we try to interpret some useful information out of the
data we have acquired. For this purpose, we need to explore the data and try to put it
uniformly for a better understanding. This stage deals with validating or verification
of the collected data and to analyze that:
 The data is according to the specifications decided.
 The data is free from errors.
 The data is meeting our needs.

12. What are the features of an Artificial Neural Network?


Any Artificial Neural Network, irrespective of the style and logic of implementation,
has a few basic features as given below.
 The Artificial Neural Network systems are modelled on the human brain and nervous
system.
 They are able to automatically extract features without feeding the input by
programmer.
 Every node of layer in a Neural Network is compulsorily a machine learning algorithm.
 It is very useful to implement when solving problems for very huge datasets.
OR
 It can work with incomplete knowledge and may produce output even with
incomplete information.
 It has fault tolerance which means that corruption of one or more cells of ANN does
not prevent it from generating output.
 It has the ability to learn events and make decisions by commenting on similar events.
 It has Parallel processing capability i.e. ANN have numerical strength that can perform
more than one job at the same time.
OR
 Neural Networks have the ability to learn by themselves and produce the output that
is not limited to the input provided to them.
 The input is stored in its own networks instead of a database; hence the loss of data
does not affect its working.

CBSE Question Bank – AI – Class 10 – Chapter 3 AI Project Cycle 4


 These networks can learn from examples and apply them when a similar event arises,
making them able to work through real-time events.
 Even if a neuron is not responding or a piece of information is missing, the network
can detect the fault and still produce the output.
 They can perform multiple tasks in parallel without affecting the system performance

13. What is the purpose of getting AI Ready?


The world is changing with each day and we have huge data coming our way. The
purpose of getting AI ready means taking steps to collect data around relevant
systems, equipment, and procedures; and storing and curating that data in a way that
makes it easily accessible to others for use in future AI applications.
OR
The purpose of getting AI ready specifies the responsible and optimum use of huge
amount of data around us to create and implement into such systems and applications
which should make life of future generations more organized and sustainable. This
process may lead to better lives for mankind.

14. What are the different types of sources of data from where we can collect
reliable and authentic datasets? Explain in brief.
Data can be a piece of information or facts and statistics collected together for
reference or analysis. Whenever we want an AI project to be able to predict an output,
we need to train it first using data There could be many ways and sources from where
we can collect reliable and authentic datasets namely Surveys, Web scrapping,
Sensors, Cameras, Observations, Research, Investigation, API etc.
Sometimes Internet is also used to acquire data but the most important point to keep
in mind is that the data should be taken from reliable and authentic websites only.
Some reliable data sources are UN, Google scholar, Finance, CIA, Data.gov etc.

Four (04) Mark Questions

1. Explain the AI Project Cycle in detail.


The steps involved in AI project cycle are as given:
 The first step is Scope the Problem by which, you set the goal for your AI project by
stating the problem which you wish to solve with it. Under problem scoping, we look
at various parameters which affect the problem we wish to solve so that the picture
becomes clearer
 Next step is to acquire data which will become the base of your project as it will help
you in understanding what the parameters that are related to problem scoping.
 Next, you go for data acquisition by collecting data from various reliable and authentic
sources. Since the data you collect would be in large quantities, you can try to give it a
visual image of different types of representations like graphs, databases, flow charts,
maps, etc. This makes it easier for you to interpret the patterns in which your acquired
data follows.
 After exploring the patterns, you can decide upon the type of model you would build to
achieve the goal. For this, you can research online and select various models which
give a suitable output.
 You can test the selected models and figure out which is the most efficient one.
 The most efficient model is now the base of your AI project and you can develop your
algorithm around it.
 Once the modelling is complete, you now need to test your model on some newly
CBSE Question Bank – AI – Class 10 – Chapter 3 AI Project Cycle 5
fetched data. The results will help you in evaluating your model and hence improving
it.
Finally, after evaluation, the project cycle is now complete and what you get is your AI
project.

2. Explain the relation between data size and model performance of an Artificial
Neural Network.
The basis for any kind of AI development is BIG DATASET. The performance of any AI
based application depends on the data supplied
ANN models are also known as Learning models and are used for prediction purposes.
These are mostly developed without paying much cognizance to the size of datasets
that can produce models of high accuracy and better generalization. Although, the
general belief is that, large dataset is needed to construct a predictive learning model.
To describe a data set as large in size, perhaps, is circumstance dependent, thus, what
constitutes a dataset to be considered as being big or small is somehow vague.
In fact, the quantity of data partitioned for the purpose of training must be of good
representation of the entire sets and sufficient enough to span through the input
space. It must be authentic and relevant to give better model performance.

3. Draw the 4Ws problem canvas and explain each one of them briefly.
The 4Ws problem canvas is the basic template while scoping a problem and using this
canvas, the picture becomes clearer while we are working to solve it.
a) Who: The “Who” block helps you in analyzing the people getting affected directly or
indirectly due to it? Under this, you find out who the ‘stakeholders’ to this problem are
and what you know about them. Stakeholders are the people who face this problem
and would be benefitted with the solution.
b) What: Under the “What” block, you need to look into what you have on hand. At this
stage, you need to determine the nature of the problem. What is the problem and how
do you know that it is a problem?
c) Where: In this block, you need to focus on the context/situation/location of the
problem. It will help you look into the situation in which the problem arises, the
context of it, and the locations where it is prominent.
d) Why: in the “Why” canvas, think about the benefits which the stakeholders would get
from the solution and how would it benefit them as well as the society.

4. Differentiate between rule-based and learning-based AI modelling approaches.


Rule Based Approach: It refers to the AI modelling where the relationship or patterns
in data are defined by the developer. The machine follows the rules or instructions
mentioned by the developer, and performs its task accordingly.
For example, suppose you have a dataset comprising of 100 images of apples and 100
images of bananas. To train your machine, you feed this data into the machine and
label each image as either apple or banana. Now if you test the machine with the image
of an apple, it will compare the image with the trained data and according to the labels
of trained images, it will identify the test image as an apple. This is known as Rule
based approach. The rules given to the machine in this example are the labels given to
the machine for each image in the training dataset.
Learning Based Approach: In this approach, the machine learns by itself. It refers to
the AI modelling where the relationship or patterns in data are not defined by the
developer. In this approach, random data is fed to the machine and it is left on the
CBSE Question Bank – AI – Class 10 – Chapter 3 AI Project Cycle 6
machine to figure out patterns and trends out of it. Generally, this approach is followed
when the data is un labelled and too random for a human to make sense out of it.
For example, suppose you have a dataset of 1000 images of random stray dogs of your
area. You would put this into a learning approach-based AI machine and the machine
would come up with various patterns it has observed in the features of these 1000
images which you might not have even thought of!

5. What is an Artificial Neural Network? Explain the layers in an artificial neural


network.
Artificial Neural Network: Modeled in accordance with the human brain, a Neural
Network was built to mimic the functionality of a human brain. The human brain is a
neural network made up of multiple neurons, similarly, an Artificial Neural Network
(ANN) is made up of multiple perceptrons.
A neural network consists of three important layers:
Input Layer: As the name suggests, this layer accepts all the inputs provided by the
programmer.
Hidden Layer: Between the input and the output layer is a set of layers known as
Hidden layers. In this layer, computations are performed which result in the output.
There can be any number of hidden layers
Output Layer: The inputs go through a series of transformations via the hidden layer
which finally results in the output that is delivered via this layer.

6. What is the need of an AI Project Cycle? Explain.


Project cycle is the process of planning, organizing, coordinating, and finally
developing a project effectively throughout its phases, from planning through
execution then completion and review to achieve pre-defined objectives.
Our mind makes up plans for every task which we have to accomplish which is why
things become clearer in our mind. Similarly, if we have to develop an AI project, the AI
Project Cycle provides us with an appropriate framework which can lead us towards
the goal.
The major role of AI Project Cycle is to distribute the development of AI project in
various stages so that the development becomes easier, clearly understandable and
the steps / stages should become more specific to efficiently get the best possible
output. It mainly has 5 ordered stages which distribute the entire development in
specific and clear steps: These are Problem Scoping, Data Acquisition, Data
Exploration, Modelling and Evaluation.

7. Explain the following:


a. Supervised Learning b. Unsupervised Learning
 Supervised learning is an approach to creating artificial intelligence (AI), where the
program is given labelled input data and the expected output results.
OR
 Supervised learning is a learning in which we teach or train the machine using data
which is well labelled that means some data is already tagged with the correct answer.
After that, the machine is provided with a new set of examples (data) so that
supervised learning algorithm analyses the training data (set of training examples)
and produces a correct outcome from labelled data.

CBSE Question Bank – AI – Class 10 – Chapter 3 AI Project Cycle 7


OR
 In a supervised learning model, the dataset which is fed to the machine is labelled. It
means some data is already tagged with the correct answer. In other words, we can
say that the dataset is known to the person who is training the machine only then
he/she is able to label the data.
 Unsupervised Learning: An unsupervised learning model works on unlabeled
dataset. This means that the data which is fed to the machine is random and there is a
possibility that the person who is training the model does not have any information
regarding it. The unsupervised learning models are used to identify relationships,
patterns and trends out of the data which is fed into it. It helps the user in
understanding what the data is about and what are the major features identified by the
machine in it.
OR
 Unsupervised learning is the training of a machine using information that is neither
classified nor labelled and allowing the algorithm to act on that information without
guidance. Here the task of the machine is to group unsorted information according to
similarities, patterns and differences without any prior training of data.

8. Differentiate between classification and clustering algorithms with the help of


suitable examples.
Classification is a process of finding a function which helps in dividing the dataset into
classes based on different parameters. In Classification, a computer program is trained
on the training dataset and based on that training; it categorizes the data into different
classes. The task of the classification algorithm is to find the mapping function to map
the input(x) to the discrete output(y).
Example: The best example to understand the Classification problem is Email Spam
Detection. The model is trained on the basis of millions of emails on different
parameters, and whenever it receives a new email, it identifies whether the email is
spam or not. If the email is spam, then it is moved to the Spam folder.
Regression is a process of finding the correlations between dependent and
independent variables. It helps in predicting the continuous variables such as
prediction of Market Trends, prediction of House prices, etc. The task of the
Regression algorithm is to find the mapping function to map the input variable(x) to
the continuous output variable(y).
Example: Suppose we want to do weather forecasting, so for this, we will use the
Regression algorithm. In weather prediction, the model is trained on the past data, and
once the training is completed, it can easily predict the weather for future days.
OR
Classification is the process of finding or discovering a model (function) which helps in
separating the data into multiple categorical classes. In classification, the group
membership of the problem is identified, which means the data is categorized under
different labels according to some parameters and then the labels are predicted for the
data.
Regression is the process of finding a model or function for distinguishing the data into
continuous real values instead of using classes. Mathematically, with a regression
problem, one is trying to find the function approximation with the minimum error
deviation. In regression, the data numeric dependency is predicted to distinguish it.
The Regression analysis is the statistical model which is used to predict the numeric
data instead of labels. It can also identify the distribution movement depending on the
available data or historic data.
CBSE Question Bank – AI – Class 10 – Chapter 3 AI Project Cycle 8
OR
Key Differences between Classification and Regression
 The Classification process models a function through which the data is predicted in
discrete class labels. On the other hand, regression is the process of creating a model
which predicts continuous quantity.
 The classification algorithms involve decision tree, logistic regression, etc. In contrast,
regression tree (e.g. Random forest) and linear regression are the examples of
regression algorithms.
 Classification predicts unordered data while regression predicts ordered data.
 Regression can be evaluated using root mean square error. On the contrary,
classification is evaluated by measuring accuracy.

9. Five sustainable Development Goals are mentioned below. Write 2 problems under
each goal that you think should be addressed for achieving the goal.
a. Quality Education
b. Reduced Inequalities
c. Life on Land
d. No Poverty
e. Clean Water and Sanitation

a. Quality Education:
i. Providing education remotely, leveraging hi-tech, low-tech and no-tech approaches;
ii. Ensure coordinated responses and avoid overlapping efforts;
iii. Ensuring return of students to school when they reopen to avoid an upsurge in
dropout rates.

b. Reduced inequalities:
i. Reduction of relative economic inequalities inequality in some countries having
poorest and most vulnerable communities.
ii. Improving the situations in countries with weaker health systems.

c. Life on Land:
i. Prevention of Deforestation caused by humans and restoration of land
ii. Preventions and cure of diseases that are transmissible between animals and humans

d. No Poverty
i. Creation of Strong social protection systems to prevent people from falling into
poverty
ii. Reduction of social exclusion, and high vulnerability of certain populations to disasters
and diseases.
iii. Responsible distribution of resources.
CBSE Question Bank – AI – Class 10 – Chapter 3 AI Project Cycle 9
e. Clean Water and Sanitation
i. To increase access to clean drinking water and sanitation mostly in rural areas
ii. Managing our water sustainably to manage our production of food and energy.

10. Do ethics in AI hamper data acquisition stage? Justify your answer.


Data acquisition is the most important factor or stage as the entire project
development is based on the acquired data. There are several ethical issues which
must always be considered when planning any type of data collection.
We need to understand that the data which is collected is ethical only if the provider
agrees to provide. For example, in case of smartphone users, data is collected by
clicking on allow when it asks for permission and by agreeing to all the terms and
conditions. But at the same time if one does not want to share his/her data with
anyone then this ethical issue hampers the acquisition process and lowers the
accuracy or amount of data required for development.
Hence Regardless of the type of data collection, it is absolutely necessary to gain the
approval of the community from which the data will collected otherwise.

CBSE Question Bank – AI – Class 10 – Chapter 3 AI Project Cycle 10


CBSE | DEPARTMENT OF SKILL EDUCATION
ARTIFICIAL INTELLIGENCE
QUESTION BANK – CLASS 10

CHAPTER 7: NATURAL LANGUAGE PROCESSING

One (01) Mark Questions

1. What is a Chabot?
A chatbot is a computer program that's designed to simulate human conversation
through voice commands or text chats or both. Eg: Mitsuku Bot, Jabberwacky etc.
OR
A chatbot is a computer program that can learn over time how to best interact with
humans. It can answer questions and troubleshoot customer problems, evaluate and
qualify prospects, generate sales leads and increase sales on an ecommerce site.
OR
A chatbot is a computer program designed to simulate conversation with human users.
A chatbot is also known as an artificial conversational entity (ACE), chat robot, talk bot,
chatterbot or chatterbox.
OR
A chatbot is a software application used to conduct an on-line chat conversation via text
or text-to-speech, in lieu of providing direct contact with a live human agent.

2. What is the full form of NLP?


Natural Language Processing

3. While working with NLP what is the meaning of?


a. Syntax
b. Semantics
Syntax: Syntax refers to the grammatical structure of a sentence.
Semantics: It refers to the meaning of the sentence.

4. What is the difference between stemming and lemmatization?


Stemming is a technique used to extract the base form of the words by removing affixes
from them. It is just like cutting down the branches of a tree to its stems. For example,
the stem of the words eating, eats, eaten is eat.
Lemmatization is the grouping together of different forms of the same word. In search
queries, lemmatization allows end users to query any version of a base word and get
relevant results.
OR
Stemming is the process in which the affixes of words are removed and the words are
converted to their base form.

CBSE Question Bank – AI – Class 10 – Chapter- 7 Natural Language Processing 1


In lemmatization, the word we get after affix removal (also known as lemma) is a
meaningful one. Lemmatization makes sure that lemma is a word with meaning and
hence it takes a longer time to execute than stemming.
OR
Stemming algorithms work by cutting off the end or the beginning of the word, taking
into account a list of common prefixes and suffixes that can be found in an inflected
word.
Lemmatization on the other hand, takes into consideration the morphological analysis
of the words. To do so, it is necessary to have detailed dictionaries which the algorithm
can look through to link the form back to its lemma.

5. What is the full form of TFIDF?


Term Frequency and Inverse Document Frequency

6. What is meant by a dictionary in NLP?


Dictionary in NLP means a list of all the unique words occurring in the corpus. If some
words are repeated in different documents, they are all written just once as while
creating the dictionary.
7. What is term frequency?
Term frequency is the frequency of a word in one document. Term frequency can easily
be found from the document vector table as in that table we mention the frequency of
each word of the vocabulary in each document.

8. Which package is used for Natural Language Processing in Python programming?


Natural Language Toolkit (NLTK). NLTK is one of the leading platforms for building
Python programs that can work with human language data.

9. What is a document vector table?


Document Vector Table is used while implementing Bag of Words algorithm.
In a document vector table, the header row contains the vocabulary of the corpus and
other rows correspond to different documents.
If the document contains a particular word it is represented by 1 and absence of word is
represented by 0 value.
OR
Document Vector Table is a table containing the frequency of each word of the
vocabulary in each document.

10. What do you mean by corpus?


In Text Normalization, we undergo several steps to normalize the text to a lower level.
That is, we will be working on text from multiple documents and the term used for the
whole textual data from all the documents altogether is known as corpus.
OR
A corpus is a large and structured set of machine-readable texts that have been
produced in a natural communicative setting.
OR
A corpus can be defined as a collection of text documents. It can be thought of as just a
bunch of text files in a directory, often alongside many other directories of text files.

CBSE Question Bank – AI – Class 10 – Chapter- 7 Natural Language Processing 2


Two (02) Mark Questions

1. What are the types of data used for Natural Language Processing applications?
Natural Language Processing takes in the data of Natural Languages in the form of
written words and spoken words which humans use in their daily lives and operates on
this.

2. Differentiate between a script-bot and a smart-bot. (Any 2 differences)

Script-bot Smart-bot
 A scripted chatbot doesn’t carry  Smart bots are built on NLP and
even a glimpse of A.I ML.
 Script bots are easy to make  Smart –bots are comparatively
difficult to make.
 Script bot functioning is very  Smart-bots are flexible and
limited as they are less powerful. powerful.
 Script bots work around a script ● Smart bots work on bigger
which is programmed in them databases and other resources
directly
 No or little language processing ● NLP and Machine learning skills
skills are required.
 Limited functionality ● Wide functionality

3. Give an example of the following:


 Multiple meanings of a word
 Perfect syntax, no meaning
 Example of Multiple meanings of a word –
His face turns red after consuming the medicine
Meaning - Is he having an allergic reaction? Or is he not able to bear the taste of that
medicine?
 Example of Perfect syntax, no meaning-
Chickens feed extravagantly while the moon drinks tea.
This statement is correct grammatically but it does not make any sense. In Human
language, a perfect balance of syntax and semantics is important for better
understanding.

4. What is inverse document frequency?


To understand inverse document frequency, first we need to understand document
frequency.
Document Frequency is the number of documents in which the word occurs irrespective
of how many times it has occurred in those documents.
In case of inverse document frequency, we need to put the document frequency in the
denominator while the total number of documents is the numerator.
For example, if the document frequency of a word “AMAN” is 2 in a particular document
then its inverse document frequency will be 3/2. (Here no. of documents is 3)

CBSE Question Bank – AI – Class 10 – Chapter- 7 Natural Language Processing 3


5. Define the following:
● Stemming
● Lemmatization

Stemming: Stemming is a rudimentary rule-based process of stripping the suffixes


(“ing”, “ly”, “es”, “s” etc) from a word.
Stemming is a process of reducing words to their word stem, base or root form (for
example, books — book, looked — look).

Lemmatization: Lemmatization, on the other hand, is an organized & step by step


procedure of obtaining the root form of the word, it makes use of vocabulary (dictionary
importance of words) and morphological analysis (word structure and grammar
relations).

The aim of lemmatization, like stemming, is to reduce inflectional forms to a common


base form. As opposed to stemming, lemmatization does not simply chop off inflections.
Instead it uses lexical knowledge bases to get the correct base forms of words.
OR
Stemming is a technique used to extract the base form of the words by removing affixes
from them. It is just like cutting down the branches of a tree to its stems. For example,
the stem of the words eating, eats, eaten is eat.
Lemmatization is the grouping together of different forms of the same word. In search
queries, lemmatization allows end users to query any version of a base word and get
relevant results.
OR
Stemming is the process in which the affixes of words are removed and the words are
converted to their base form.
In lemmatization, the word we get after affix removal (also known as lemma) is a
meaningful one. Lemmatization makes sure that lemma is a word with meaning and
hence it takes a longer time to execute than stemming.
OR
Stemming algorithms work by cutting off the end or the beginning of the word, taking
into account a list of common prefixes and suffixes that can be found in an inflected
word.
Lemmatization on the other hand, takes into consideration the morphological analysis
of the words. To do so, it is necessary to have detailed dictionaries which the algorithm
can look through to link the form back to its lemma.

6. What do you mean by document vectors?


Document Vector contains the frequency of each word of the vocabulary in a particular
document.
In document vector vocabulary is written in the top row. Now, for each word in the
document, if it matches with the vocabulary, put a 1 under it. If the same word appears
again, increment the previous value by 1. And if the word does not occur in that
document, put a 0 under it.
7. What is TFIDF? Write its formula.
Term frequency–inverse document frequency, is a numerical statistic that is intended to
reflect how important a word is to a document in a collection or corpus.

CBSE Question Bank – AI – Class 10 – Chapter- 7 Natural Language Processing 4


The number of times a word appears in a document divided by the total number of
words in the document. Every document has its own term frequency.

8. Which words in a corpus have the highest values and which ones have the least?
Stop words like - and, this, is, the, etc. have highest values in a corpus. But these words
do not talk about the corpus at all. Hence, these are termed as stopwords and are mostly
removed at the pre-processing stage only.
Rare or valuable words occur the least but add the most importance to the corpus.
Hence, when we look at the text, we take frequent and rare words into consideration.

9. Does the vocabulary of a corpus remain the same before and after text
normalization? Why?
No, the vocabulary of a corpus does not remain the same before and after text
normalization. Reasons are –
● In normalization the text is normalized through various steps and is lowered to
minimum vocabulary since the machine does not require grammatically correct
statements but the essence of it.
● In normalization Stop words, Special Characters and Numbers are removed.
● In stemming the affixes of words are removed and the words are converted to their base
form.
So, after normalization, we get the reduced vocabulary.
10. What is the significance of converting the text into a common case?
In Text Normalization, we undergo several steps to normalize the text to a lower level.
After the removal of stop words, we convert the whole text into a similar case,
preferably lower case. This ensures that the case-sensitivity of the machine does not
consider same words as different just because of different cases.

11. Mention some applications of Natural Language Processing.


Natural Language Processing Applications-
● Sentiment Analysis.
● Chatbots & Virtual Assistants.
● Text Classification.
● Text Extraction.
● Machine Translation
● Text Summarization
● Market Intelligence
● Auto-Correct
CBSE Question Bank – AI – Class 10 – Chapter- 7 Natural Language Processing 5
12. What is the need of text normalization in NLP?
Since we all know that the language of computers is Numerical, the very first step that
comes to our mind is to convert our language to numbers.
This conversion takes a few steps to happen. The first step to it is Text Normalization.
Since human languages are complex, we need to first of all simplify them in order to
make sure that the understanding becomes possible. Text Normalization helps in
cleaning up the textual data in such a way that it comes down to a level where its
complexity is lower than the actual data.

13. Explain the concept of Bag of Words.


Bag of Words is a Natural Language Processing model which helps in extracting features
out of the text which can be helpful in machine learning algorithms. In bag of words, we
get the occurrences of each word and construct the vocabulary for the corpus.
Bag of Words just creates a set of vectors containing the count of word occurrences in
the document (reviews). Bag of Words vectors are easy to interpret.
14. Explain the relation between occurrence and value of a word.

plot of occurrence of words versus their value

As shown in the graph, occurrence and value of a word are inversely proportional. The
words which occur most (like stop words) have negligible value. As the occurrence of
words drops, the value of such words rises. These words are termed as rare or valuable
words. These words occur the least but add the most value to the corpus.

15. What are the applications of TFIDF?


TFIDF is commonly used in the Natural Language Processing domain. Some of its
applications are:
 Document Classification - Helps in classifying the type and genre of a document.
 Topic Modelling - It helps in predicting the topic for a corpus.
 Information Retrieval System - To extract the important information out of a corpus.
 Stop word filtering - Helps in removing the unnecessary words out of a text body.

16. What are stop words? Explain with the help of examples.
“Stop words” are the most common words in a language like “the”, “a”, “on”, “is”, “all”.
These words do not carry important meaning and are usually removed from texts. It is
possible to remove stop words using Natural Language Toolkit (NLTK), a suite of
libraries and programs for symbolic and statistical natural language processing.

CBSE Question Bank – AI – Class 10 – Chapter- 7 Natural Language Processing 6


17. Differentiate between Human Language and Computer Language.
Humans communicate through language which we process all the time. Our brain keeps
on processing the sounds that it hears around itself and tries to make sense out of them
all the time.
On the other hand, the computer understands the language of numbers. Everything that
is sent to the machine has to be converted to numbers. And while typing, if a single
mistake is made, the computer throws an error and does not process that part. The
communications made by the machines are very basic and simple.

Four 04 Mark Questions

1. Create a document vector table for the given corpus:


Document 1: We are going to Mumbai
Document 2: Mumbai is a famous place.
Document 3: We are going to a famous place.
Document 4: I am famous in Mumbai.

We Are going to Mumbai is a famous place I am in


1 1 1 1 1 0 0 0 0 0 0 0
0 0 0 0 1 1 1 1 1 0 0 0
1 1 1 1 0 0 1 1 1 0 0 0
0 0 0 0 1 0 0 1 0 1 1 1

2. Classify each of the images according to how well the model’s output matches the
data samples:

Here, the red dashed line is model’s output while the blue crosses are actual data
samples.

● The model’s output does not match the true function at all. Hence the model is said to be
under fitting and its accuracy is lower.
● In the second case, model performance is trying to cover all the data samples even if
they are out of alignment to the true function. This model is said to be over fitting and
this too has a lower accuracy
● In the third one, the model’s performance matches well with the true function which
states that the model has optimum accuracy and the model is called a perfect fit.

CBSE Question Bank – AI – Class 10 – Chapter- 7 Natural Language Processing 7


3. Explain how AI can play a role in sentiment analysis of human beings?
The goal of sentiment analysis is to identify sentiment among several posts or even in
the same post where emotion is not always explicitly expressed.
Companies use Natural Language Processing applications, such as sentiment analysis, to
identify opinions and sentiment online to help them understand what customers think
about their products and services (i.e., “I love the new iPhone” and, a few lines later “But
sometimes it doesn’t work well” where the person is still talking about the iPhone) and
overall *
Beyond determining simple polarity, sentiment analysis understands sentiment in
context to help better understand what’s behind an expressed opinion, which can be
extremely relevant in understanding and driving purchasing decisions.

4. Why are human languages complicated for a computer to understand? Explain.


The communications made by the machines are very basic and simple. Human
communication is complex. There are multiple characteristics of the human language
that might be easy for a human to understand but extremely difficult for a computer to
understand.
For machines it is difficult to understand our language. Let us take a look at some of
them here:
Arrangement of the words and meaning - There are rules in human language. There are
nouns, verbs, adverbs, adjectives. A word can be a noun at one time and an adjective
some other time. This can create difficulty while processing by computers.
Analogy with programming language- Different syntax, same semantics: 2+3 = 3+2 Here
the way these statements are written is different, but their meanings are the same that
is 5. Different semantics, same syntax: 2/3 (Python 2.7) ≠ 2/3 (Python 3) Here the
statements written have the same syntax but their meanings are different. In Python 2.7,
this statement would result in 1 while in Python 3, it would give an output of 1.5.
Multiple Meanings of a word - In natural language, it is important to understand that a
word can have multiple meanings and the meanings fit into the statement according to
the context of it.

CBSE Question Bank – AI – Class 10 – Chapter- 7 Natural Language Processing 8


Perfect Syntax, no Meaning - Sometimes, a statement can have a perfectly correct syntax
but it does not mean anything. In Human language, a perfect balance of syntax and
semantics is important for better understanding.
These are some of the challenges we might have to face if we try to teach computers
how to understand and interact in human language.

5. What are the steps of text Normalization? Explain them in brief.

Text Normalizationin Text Normalization, we undergo several steps to normalize the


text to a lower level.

Sentence Segmentation - Under sentence segmentation, the whole corpus is divided into
sentences. Each sentence is taken as a different data so now the whole corpus gets
reduced to sentences.

Tokenisation- After segmenting the sentences, each sentence is then further divided into
tokens. Tokens is a term used for any word or number or special character occurring in
a sentence. Under tokenisation, every word, number and special character is considered
separately and each of them is now a separate token.
Removing Stop words, Special Characters and Numbers - In this step, the tokens which
are not necessary are removed from the token list.
Converting text to a common case -After the stop words removal, we convert the whole
text into a similar case, preferably lower case. This ensures that the case-sensitivity of
the machine does not consider same words as different just because of different cases.
Stemming In this step, the remaining words are reduced to their root words. In other
words, stemming is the process in which the affixes of words are removed and the
words are converted to their base form.
Lemmatization -in lemmatization, the word we get after affix removal (also known as
lemma) is a meaningful one.

With this we have normalized our text to tokens which are the simplest form of words
present in the corpus. Now it is time to convert the tokens into numbers. For this, we
would use the Bag of Words algorithm

6. Through a step-by-step process, calculate TFIDF for the given corpus and mention
the word(s) having highest value.
Document 1: We are going to Mumbai
Document 2: Mumbai is a famous place.
Document 3: We are going to a famous place.
Document 4: I am famous in Mumbai.

Term Frequency
Term frequency is the frequency of a word in one document. Term frequency can easily
be found from the document vector table as in that table we mention the frequency of
each word of the vocabulary in each document.

CBSE Question Bank – AI – Class 10 – Chapter- 7 Natural Language Processing 9


We Are Going to Mumbai is a famous Place I am in
1 1 1 1 1 0 0 0 0 0 0 0
0 0 0 0 1 1 1 1 1 0 0 0
1 1 1 1 0 0 1 1 1 0 0 0
0 0 0 0 1 0 0 1 0 1 1 1
Inverse Document Frequency
The other half of TFIDF which is Inverse Document Frequency. For this, let us first
understand what does document frequency mean. Document Frequency is the number
of documents in which the word occurs irrespective of how many times it has occurred
in those documents. The document frequency for the exemplar vocabulary would be:

We Are going to Mumbai is a Famous place I am in


2 2 2 2 3 1 2 3 2 1 1 1

Talking about inverse document frequency, we need to put the document frequency in
the denominator while the total number of documents is the numerator. Here, the total
number of documents are 3, hence inverse document frequency becomes:

We Are going to Mumbai is a Famous Place I am in


4/2 4/2 4/2 4/2 4/3 4/1 4/2 4/3 4/2 4/1 4/1 4/1

The formula of TFIDF for any word W becomes:


TFIDF(W) = TF(W) * log (IDF(W))

The words having highest value are – Mumbai, Famous

7. Normalize the given text and comment on the vocabulary before and after the
normalization:
Raj and Vijay are best friends. They play together with other friends. Raj likes to
play football but Vijay prefers to play online games. Raj wants to be a footballer.
Vijay wants to become an online gamer.

Normalization of the given text:


Sentence Segmentation:
1. Raj and Vijay are best friends.
2. They play together with other friends.
3. Raj likes to play football but Vijay prefers to play online games.
4. Raj wants to be a footballer.
5. Vijay wants to become an online gamer.

CBSE Question Bank – AI – Class 10 – Chapter- 7 Natural Language Processing 10


Tokenization:

Raj and Vijay Raj and Vijay are best friends .


are best
friends.

They play They play Together with other friends .


together with
other friends

Same will be done for all sentences.


Removing Stop words, Special Characters and Numbers:
In this step, the tokens which are not necessary are removed from the token list.
So, the words and, are, to, an, (Punctuation) will be removed.

Converting text to a common case:


After the stop words removal, we convert the whole text into a similar case, preferably
lower case.
Here we don’t have words in different case so this step is not required for given text.
Stemming:
In this step, the remaining words are reduced to their root words. In other words,
stemming is the process in which the affixes of words are removed and the words are
converted to their base form.

Word Affixes Stem

Likes -s Like

Prefers -s Prefer

Wants -s want

In the given text Lemmatization is not required.


Given Text
Raj and Vijay are best friends. They play together with other friends. Raj likes to play
football but Vijay prefers to play online games. Raj wants to be a footballer. Vijay wants to
become an online gamer.
Normalized Text
Raj and Vijay best friends They play together with other friends Raj likes to play football
but Vijay prefers to play online games Raj wants to be a footballer Vijay wants to become
an online gamer

CBSE Question Bank – AI – Class 10 – Chapter- 7 Natural Language Processing 11


CBSE | DEPARTMENT OF SKILL EDUCATION
ARTIFICIAL INTELLIGENCE
QUESTION BANK – CLASS 10
CHAPTER 8: EVALUATION
One (01) Mark Questions

1. Define Evaluation.
Moving towards deploying the model in the real world, we test it in as many ways
aspossible. The stage of testing the models is known as EVALUATION.
OR
Evaluation is a process of understanding the reliability of any AI model, based on outputs
by feeding the test dataset into the model and comparing it with actual answers.
OR
Evaluation is a process that critically examines a program. It involves collecting and
analyzing information about a program’s activities, characteristics, and outcomes. Its
purpose is to make judgments about a program, to improve its effectiveness, and/or to
inform programming decisions.
2. Which two parameters are considered for Evaluation of a model?

Prediction and Reality are the two parameters considered for Evaluation of a model.
The “Prediction” is the output which is given by the machine and the “Reality”is the real
scenario, when the prediction has been made?

3. What is True Positive?


 The predicted value matches the actual value
 The actual value was positive and the model predicted a positive value

4. What is True Negative?


 The predicted value matches the actual value
 The actual value was negative and the model predicted a negative value

5. What is False Positive?


 The predicted value was falsely predicted
 The actual value was negative but the model predicted a positive value
 Also known as the Type 1 error

6. What is False Negative?


 The predicted value was falsely predicted
 The actual value was positive but the model predicted a negative value
 Also known as the Type 2 error

CBSE Question Bank – AI – Class 10 – Chapter- 8 Evaluation 1


Two (02) Mark Questions

1. What is meant by Overfitting of Data?


Overfitting is "the production of an analysis that corresponds too closely or exactly to a
particular set of data, and may therefore fail to fit additional data or predict future
observations reliably".
(OR)
An Overfitted Model is a statistical model that contains more parameters than can be
justified by the data. Here, to evaluate the AI model it is not necessary to use the data that
is used to build the model. Because AI Model remembers the whole training data set,
therefore it always predicts the correct label for any point in the training dataset. This is
known as Overfitting
(OR)
Models that use the training dataset during testing, will always results in correct output. This is
known as overfitting.

2. What is Accuracy? Mention its formula.


Accuracy is defined as the percentage of correct predictions out of all theobservations.
A prediction is said to be correct if it matches reality. Here we have twoconditions in which
the Prediction matches with the Reality, i.e., True Positive and True Negative. Therefore,
Formula for Accuracy is

Where TP = True Positives, TN = True Negatives, FP = False Positives, and FN = False


Negatives.

3. What is Precision? Mention its formula.

Precision is defined as the percentage of true positive cases versus all the cases where the
prediction is true.

That is, it takes into account the True Positives and False Positives.

CBSE Question Bank – AI – Class 10 – Chapter- 8 Evaluation 2


4. What is Recall? Mention its formula.
Recall is defined as the fraction of positive cases that are correctly Identified.

5. Why is evaluation important? Explain.


Importance of Evaluation
Evaluation is a process that critically examines a program. It involves collecting and
analyzing information about a program's activities, characteristics, and outcomes. Its
purpose is to make judgments about a program, to improve its effectiveness, and/or to
inform programming decisions.

 Evaluation is important to ensure that the model is operating correctly and optimally.
 Evaluation is an initiative to understand how well it achieves its goals.
 Evaluations help to determine what works well and what could be improved in a
program

6. How do you suggest which evaluation metric is more important for any case?
F 1 Evaluation metric is more important in any case. F1 score sort maintains a balance
between the precision and recall for the classifier. If the precision is low, the F1 is low
and if the recall is low again F1 score is low.
The F1 score is a number between 0 and 1 and is the harmonic mean of precision
and recall

When we have a value of 1 (that is 100%) for both Precision and Recall. The F1 score would
also be an ideal 1 (100%). It is known as the perfect value for F1 Score. As the values of
both Precision and Recall ranges from 0 to 1, the F1 score also ranges from 0 to 1.

7. Which evaluation metric would be crucial in the following cases? Justify your
answer.
a. Mail Spamming
b. Gold Mining
c. Viral Outbreak
Here, Mail Spamming and Gold Mining are related to FALSE POSITIVE cases which are
expensive at cost. But Viral Outbreak is a FALSE NEGATIVE case which infects a lot of
people on health and leads to expenditure of money too for checkups.
So, False Negative case (VIRAL OUTBREAK) are more crucial and dangerous when
compared to FALSE POSITIVE cases.
(OR)
CBSE Question Bank – AI – Class 10 – Chapter- 8 Evaluation 3
a. If the model always predicts that the mail is spam, people would not look at it and
eventually might lose important information. False Positive condition would have a
high cost. (predicting the mail as spam while the mail is not spam)
b. A model saying that there exists treasure at a point and you keep on digging there but
it turns out that it is a false alarm. False Positive case is very costly.
(predicting there is a treasure but there is no treasure)
c. A deadly virus has started spreading and the model which is supposed to predict a
viral outbreak does not detect it. The virus might spread widely and infect a lot of
people. Hence, False Negative can be dangerous

8. What are the possible reasons for an AI model not being efficient? Explain.
Reasons of an AI model not being efficient:

a. Lack of Training Data: If the data is not sufficient for developing an AI Model, or if the
data is missed while training the model, it will not be efficient.

b. Unauthenticated Data / Wrong Data: If the data is not authenticated and correct, then
the model will not give good results.
c. Inefficient coding / Wrong Algorithms: If the written algorithms are not correct
and relevant, Model will not give desired output. Not Tested: If the model is not
tested properly, then it will not be efficient.
d. Not Easy: If it is not easy to be implemented in production or scalable.
e. Less Accuracy: A model is not efficient if it gives less accuracy scores in production
or test data or if it is not able to generalize well on unseen data.
(Any three of the above can be selected)
9. Answer the following:

 Give an example where High Accuracy is not usable.


SCENARIO: An expensive robotic chicken crosses a very busy road a thousand times
per day. An ML model evaluates traffic patterns and predicts when this chicken can
safely cross the street with an accuracy of 99.99%.
Explanation: A 99.99% accuracy value on a very busy road strongly suggests that the
ML model is far better than chance. In some settings, however, the cost of making
even a small number of mistakes is still too high. 99.99% accuracy means that the
expensive chicken will need to be replaced, on average, every 10 days. (The chicken
might also cause extensive damage to cars that it hits.)
 Give an example where High Precision is not usable.

Example: “Predicting a mail as Spam or Not Spam”


False Positive: Mail is predicted as “spam” but it is “not spam”.
False Negative: Mail is predicted as “not spam” but it is “spam”.
Of course, too many False Negatives will make the spam filter ineffective but False
Positives may cause important mails to be missed and hence Precision is not usable.

CBSE Question Bank – AI – Class 10 – Chapter- 8 Evaluation 4


Four (04) Mark Questions

1. Deduce the formula of F1 Score? What is the need of its formulation?

The F1 Score, also called the F score or F measure, is a measure of a test’s accuracy.
It is calculated from the precision and recall of the test, where the precision is the number
of correctly identified positive results divided by the number of all positive results,
including those not identified correctly, and the recall is the number of correctly identified
positive results divided by the number of all samples that should have been identified as
positive.
The F1 score is defined as the weighted harmonic mean of the test’s precision and recall.
This score is calculated according to the formula.

Formula:

Necessary:
F-Measure provides a single score that balances both the concerns of precision and recall in
one number.
A good F1 score means that you have low false positives and low false negatives, so you’re
correctly identifying real threats, and you are not disturbed by false alarms.
An F1 score is considered perfect when it’s 1, while the model is a total failure when it’s 0.
F1 Score is a better metric to evaluate our model on real-life classification problems and
when imbalanced class distribution exists.

2. What is a confusion matrix? Explain in detail with the help of an example.

Confusion Matrix:
A Confusion Matrix is a table that is often used to describe the performance of a
classification model (or "classifier") on a set of test data for which the true values are
known.
(or)

A 2x2 matrix denoting the right and wrong predictions might help us analyse the rate of
success. This matrix is termed the Confusion Matrix.

Evaluation of the performance of a classification model is based on the counts of test


records correctly and incorrectly predicted by the model.

Therefore, Confusion Matrix provides a more insightful picture which is not only the
performance of a predictive model, but also which classes are being predicted correctly and
incorrectly, and what type of errors are being made.

The confusion matrix is useful for measuring Recall (also known as Sensitivity), Precision,
Accuracy and F1 Score.

CBSE Question Bank – AI – Class 10 – Chapter- 8 Evaluation 5


The following confusion matrix table illustrates how the 4-classification metrics are
calculated (TP, FP, FN, TN), and how our predicted value compared to the actual value in a
confusion matrix

Let’s decipher the matrix:

The target variable has two values: Positive or Negative

The columns represent the actual values of the target variable


The rows represent the predicted values of the target variable
True Positive, True Negative, False Positive and False Negative in a Confusion Matrix

True Positive (TP)

The predicted value matches the actual value


The actual value was positive and the model predicted a positive value
True Negative (TN)
The predicted value matches the actual value

The actual value was negative and the model predicted a negative value
False Positive (FP) – Type 1 error
The predicted value was falsely predicted
The actual value was negative but the model predicted a positive value ● Also known as the
Type 1 error
False Negative (FN) – Type 2 error

The predicted value was falsely predicted

The actual value was positive but the model predicted a negative value also known as the
Type 2 error

Example:

Case: Loan (Good loan & Bad loan)

CBSE Question Bank – AI – Class 10 – Chapter- 8 Evaluation 6


The result of TP will be that bad loans are correctly predicted as bad loans.

The result of TN will be that good loans are correctly predicted as good loans.

The result of FP will be that (actual) good loans are incorrectly predicted as bad loans.

The result of FN will be that (actual) bad loans are incorrectly predicted as good loans.
The banks would lose a bunch of money if the actual bad loans are predicted as good loans
due to loans not being repaid. On the other hand, banks won't be able to make more
revenue if the actual good loans are predicted as bad loans. Therefore, the cost of False
Negatives is much higher than the cost of False Positives.

3. Calculate Accuracy, Precision, Recall and F1 Score for the following Confusion Matrix
on Heart Attack Risk. Also suggest which metric would not be a good evaluation
parameter here and why?

The Confusion Reality: 1 Reality: 0


Matrix
Prediction: 1 50 20
Prediction: 0 10 20

The Confusion Reality: 1 Reality: 0


Matrix
Prediction: 1 50 20 70
Prediction: 0 10 20 30
60 40 100
Calculation:
Accuracy:
Accuracy is defined as the percentage of correct predictions out of all the observations

CBSE Question Bank – AI – Class 10 – Chapter- 8 Evaluation 7


Where True Positive (TP), True Negative (TN), False Positive (FP) and False Negative (FN).
Accuracy = (50+20) / (50+20+20+10)

= (70/100)
= 0.7

Precision:
Precision is defined as the percentage of true positive cases versus all the cases where the
prediction is true.

= (50 / (50 + 20))


= (50/70)
= 0.714
Recall: It is defined as the fraction of positive cases that are correctly identified.

= 50 / (50 + 60)
= 50 / 110
= 0.5
F1 Score:

F1 score is defined as the measure of balance between precision and recall.

= 2 * (0.714 *0.5) / (0.714 + 0.5)


= 2 * (0.357 / 1.214)

CBSE Question Bank – AI – Class 10 – Chapter- 8 Evaluation 8


= 2* (0.29406)
= 0.58

Therefore,
Accuracy= 0.7 Precision=0.714 Recall=0.5
F1 Score=0.588

Here within the test there is a tradeoff. But Recall is not a good Evaluation metric. Recall
metric needs to improve more.
Because,

False Positive (impacts Precision): A person is predicted as high risk but does not have
heart attack.

False Negative (impacts Recall): A person is predicted as low risk but has heart attack.
Therefore, False Negatives miss actual heart patients, hence recall metric need more
improvement.
False Negatives are more dangerous than False Positives.

4. Calculate Accuracy, Precision, Recall and F1 Score for the following Confusion Matrix on
Water Shortage in Schools: Also suggest which metric would not be a good evaluation
parameter here and why?

The Confusion Matrix (Water Shortage in Reality: 1 Reality: 0


School)
Prediction: 1 75 5
Prediction: 0 5 15

Reality: 1 Reality: 0
Prediction: 1 75 5 80
Prediction: 0 5 15 20
80 20 100

Calculation:

Accuracy
Accuracy is defined as the percentage of correct predictions out of all the observations

Where True Positive (TP), True Negative (TN), False Positive (FP) and False Negative
(FN).
CBSE Question Bank – AI – Class 10 – Chapter- 8 Evaluation 9
= (75+15) / (75+15+5+5)
= (90 / 100)
=0.9

Precision:
Precision is defined as the percentage of true positive cases versus all the cases where the
prediction is true.

= 75 / (75+5)
= 75 /80
= 0.9375

Recall:

It is defined as the fraction of positive cases that are correctly identified.

= 75 / (75+5)
= 75 /80
= 0.9375
F1 Score:
F1 score is defined as the measure of balance between precision and recall.
= 2 * ((0.9375 *0.9375) / (0.9375+0.9375)

Therefore,
= 2 * (0.8789 / 1.875)
= 2 * 0.46875
= 0.9375
Accuracy= 0.9% Precision=0.9375% Recall=0.9375%
F1 Score=0.

Here precision, recall, accuracy, f1 score all are same

CBSE Question Bank – AI – Class 10 – Chapter- 8 Evaluation 10


5. Calculate Accuracy, Precision, Recall and F1 Score for the following Confusion
Matrix on SPAM FILTERING: Also suggest which metric would not be a good
evaluation parameter here and why?
Confusion Matrix on SPAM Reality: 1 Reality: 0
FILTERING:
Prediction: 1 10 55
Prediction: 0 10 25

Confusion Matrix on SPAM Reality: 1 Reality: 0


FILTERING:
Prediction:1 10 55 65
Prediction: 0 10 25 35
20 80 100

Accuracy is defined as the percentage of correct predictions out of all the observations

Where True Positive (TP), True Negative (TN), False Positive (FP) and False Negative
(FN).
= (10 + 25) / (10+25+55+10)
= 35 / 100
= 0.35
Precision:
Precision is defined as the percentage of true positive cases versus all the cases where
the prediction is true.

= 10 / (10 +55)
= 10 /65
= 0.15
Recall:
It is defined as the fraction of positive cases that are correctly identified.

CBSE Question Bank – AI – Class 10 – Chapter- 8 Evaluation 11


F1
Score:
= 10 / (10 + 10)
= 10 / 20
= 0.5

F1 score is defined as the measure of balance between precision and recall.

Therefore,
= 2 * ((0.15 * 0.5) / (0.15 + 0.5))
= 2 * (0.075 / 0.65)
= 2 * 0.115
= 0.23
Accuracy= 0.35
Precision= 0.15
Recall= 0.5
F1 Score= 0.23

Here within the test there is a tradeoff. But Precision is not a good Evaluation metric.
Precision metric needs to improve more.
Because,

False Positive (impacts Precision): Mail is predicted as “spam” but it is not.

False Negative (impacts Recall): Mail is predicted as “not spam” but spam
Of course, too many False Negatives will make the Spam Filter ineffective. But False
Positives may cause important mails to be missed. Hence, Precision is more important
to improve

CBSE Question Bank – AI – Class 10 – Chapter- 8 Evaluation 12


CBSE | DEPARTMENT OF SKILL EDUCATION
ARTIFICIAL INTELLIGENCE (SUBJECT CODE - 417)
Blue-print for Sample Question Paper for Class X (Session 2022-2023)
Max. Time: 2 Hours Max. Marks: 50

PART A - EMPLOYABILITY SKILLS (10 MARKS):

OBJECTIVE TYPE SHORT ANSWER


UNIT QUESTIONS TYPE QUESTIONS TOTAL
NAME OF THE UNIT
NO. QUESTIONS
1 MARK EACH 2 MARKS EACH

1 Self-Management Skills - II 2 2 4

2 ICT Skills - II 2 1 3

3 Entrepreneurial Skills - II 2 2 4

TOTAL QUESTIONS 6 5 11

NO. OF QUESTIONS TO BE ANSWERED Any 4 Any 3 07

TOTAL MARKS 1x4=4 2x3=6 10 MARKS

PART B - SUBJECT SPECIFIC SKILLS (40 MARKS):

OBJECTIVE SHORT DESCRIPTIVE/


UNIT TYPE ANSWER TYPE LONG ANS. TYPE TOTAL
NAME OF THE UNIT QUESTIONS QUESTIONS QUESTIONS
NO. QUESTIONS
1 MARK EACH 2 MARKS EACH 4 MARKS EACH

1 Introduction to AI 6 2 2 10

2 AI Project Cycle 6 1 1 8

Natural Language
6 6 2 1 9
Processing

7 Evaluation 6 1 1 8

TOTAL QUESTIONS 24 6 5 35

NO. OF QUESTIONS
20 Any 4 Any 3 27
TO BE ANSWERED

TOTAL MARKS 1 x 20 = 20 2x4=8 4 x 3 = 12 40 MARKS

Page 1 of 9
CBSE | DEPARTMENT OF SKILL EDUCATION
ARTIFICIAL INTELLIGENCE (SUBJECT CODE - 417)

Sample Question Paper for Class X (Session 2022-2023)


Max. Time: 2 Hours Max. Marks: 50

General Instructions:
1. Please read the instructions carefully.
2. This Question Paper consists of 21 questions in two sections: Section A & Section B.
3. Section A has Objective type questions whereas Section B contains Subjective type questions.
4. Out of the given (5 + 16 =) 21 questions, a candidate has to answer (5 + 10 =) 15 questions in the
allotted (maximum) time of 2 hours.
5. All questions of a particular section must be attempted in the correct order.
6. SECTION A - OBJECTIVE TYPE QUESTIONS (24 MARKS):
i. This section has 05 questions.
ii. Marks allotted are mentioned against each question/part.
iii. There is no negative marking.
iv. Do as per the instructions given.
7. SECTION B – SUBJECTIVE TYPE QUESTIONS (26 MARKS):
i. This section has 16 questions.
ii. A candidate has to do 10 questions.
iii. Do as per the instructions given.
iv. Marks allotted are mentioned against each question/part.

Page 2 of 9
SECTION A: OBJECTIVE TYPE QUESTIONS

Q. 1 Answer any 4 out of the given 6 questions on Employability Skills (1 x 4 = 4 marks)

i. Pranjali gets up at 5 am and goes to her badminton classes. Then she comes home 1
and finishes her homework before going to school. She does this all by herself. No
one tells her to do it. This is an example of
(a) Self-motivation
(b) External motivation
(c) Both self and external motivation
(d) Not any specific type of motivation

ii. Which of the following can cause stress? 1


(a) Yoga and meditation
(b) Driving during rush hour
(c) Organized academic life
(d) Enjoying holidays with family

iii. What is the term used when you quickly click the left mouse button twice? 1
(a) Hover
(b) Drag and drop
(c) Double clicking
(d) Moving

iv. Ravi learnt that if a laptop gets overheated, the internal parts get damaged. What 1
happens if he leaves his device plugged in even after it is charged 100%?
(a) It can break
(b) It can stop functioning
(c) It can overheat
(d) Data can get corrupt

v. Srishti is a young woman who makes earrings. She buys jute from a farmer and 1
makes earrings from that. She sees that most women in her village do not work. So,
she hires two women to help her. As her orders increase, she hires three more
women to work for her. How was she helping her village grow?
(a) By selling earrings to women without a job
(b) By purchasing earrings from the local market
(c) By buying jute from the local farmer and by providing jobs to local women
(d) By attracting the women in her village with her creative earrings

vi. Mary has two people who work for her. Every day, she spends one hour with them 1
to learn about what they’ve done that day.
(a) Creates a new product
(b) Divides income
(c) Manages the business
(d) Takes risks

Page 3 of 9
Q. 2 Answer any 5 out of the given 6 questions (1 x 5 = 5 marks)

i. ____________ and __________ are AI based applications that help us in navigation. 1

ii. “This type of intelligence measure’s one’s awareness of the natural world 1
around them and their sensitivities to any changes that occur. It allows us to
identify the variation among two different species and understand how they are
related”.
Identify the type of intelligence described in the above sentence.

iii. Identify the incorrect statement(s) from the following: 1


(i) Deep learning is a subset of Machine Learning
(ii) Machine Learning is a subset of Deep Learning
(iii) Artificial Intelligence is a subset of Deep Learning
(iv) Deep Learning is the advanced form of AI and ML
(a) only (i)
(b) (ii) and (iii)
(c) (i) and (ii)
(d) Only (iii)

iv. Search engines not only predict what popular searches may apply to your query as 1
you start typing, but it looks at the whole picture and recognizes what you’re trying
to say rather than the exact search words. This is an example of
(a) Computer Vision
(b) Data Sciences
(c) Natural Language Processing
(d) Natural Language Understanding

v. When a user installs an app in the smartphone, it asks for access to gallery, contacts, 1
etc. After accepting this, it gives the user agreement which most users accept
without realizing the implications. What is the concern here?
(a) Data Privacy
(b) Unemployment
(c) AI bias
(d) No concern

vi. We can’t make “good” decisions without information. (True/False) 1

Q. 3 Answer any 5 out of the given 6 questions (1 x 5 = 5 marks)

i. __________ helps us to summarise all the key points into one single Template so 1
that in future, whenever there is a need to look back at the basis of the problem, we
can take a look at this and understand the key elements of it.

Page 4 of 9
ii. Divya was learning neural networks. She understood that there were three layers in 1
a neural network. Help her identify the layer that does processing in the neural
network.
(a) Output layer
(b) Hidden layer
(c) Input layer
(d) Data layer

iii. Smita is working on a project that involves over a lakh of records. Which of the 1
following should she use to make the best project?
(a) Traditional programming
(b) Manual processing
(c) IoT
(d) Neural networks

iv. For better efficiency of an AI project Training data should be _______ 1


i) Relevant
ii) Scattered
iii) Structured
iv) Authentic
Choose the correct option:
(a) Both i and ii
(b) Both i and iv
(c) Only i
(d) Only iv

v. The _______Sustainable Development Goals (SDGs) were launched at the United 1


Nations Sustainable Development Summit in New York in the year 2015, forming the
2030 Agenda for Sustainable Development.
(a) 17
(b) 15
(c) 13
(d) 19

vi. 1

Identify the algorithm based on the given graph


(a) Dimensionality reduction
(b) Classification
(c) Clustering
(d) Regression

Page 5 of 9
Q. 4 Answer any 5 out of the given 6 questions (1 x 5 = 5 marks)

i. _______________ helps in assigning a predefined category to a document, organize 1


it in such a way that helps customers to find information they want. For example
spam filtering in email, auto tagging on social media, categorization of news articles
etc.

ii. Which of the following is the type of data used by NLP applications? 1
(a) Images
(b) Numerical data
(c) Graphical data
(d) Text and Speech

iii. Ayushi was learning about NLP. She wanted to know the term used for the whole 1
textual data from all the documents altogether. Help her in identifying the term used
for it.

iv. What is the full form of TF-IDF? 1

v. A corpus contains 12 documents. How many document vectors will be there for 1
that corpus?
a. 12
b. 1
c. 24
d. 1/12

vi. Identify the type of chatbot with the information given below: 1
These bots work on pre-programmed instructions inside the application/machine
and are generally easy to develop. They are deployed in the customer care section
of various companies. Their job is to answer some basic queries that they are coded
for and connect them to human executives once they are unable to handle the
conversation.

Q. 5 Answer any 5 out of the given 6 questions (1 x 5 = 5 marks)

i. The output given by the AI machine is known as ________ (Prediction/ Reality) 1

ii. _____________ is used to record the result of comparison between the prediction 1
and reality. It is not an evaluation metric but a record which can help in evaluation.

iii. Raunak was learning the conditions that make up the confusion matrix. He came 1
across a scenario in which the machine that was supposed to predict an animal was
always predicting not an animal. What is this condition called?
(a) False Positive
(b) True Positive
(c) False Negative
(d) True Negative

Page 6 of 9
iv. Which two evaluation methods are used to calculate F1 Score? 1
(a) Precision and Accuracy
(b) Precision and Recall
(c) Accuracy and Recall
(d) Precision, F1 score

v. Which of the following statements is not true about overfitting models? 1


(a) This model learns the pattern and noise in the data to such extent that it
harms the performance of the model on the new dataset
(b) Training result is very good and the test result is poor
(c) It interprets noise as patterns in the data
(d) The training accuracy and test accuracy both are low

vi. Priya was confused with the terms used in the evaluation stage. Suggest her the term 1
used for the percentage of correct predictions out of all the observations.
(a) Accuracy
(b) Precision
(c) Recall
(d) F1 Score

SECTION B: SUBJECTIVE TYPE QUESTIONS

Answer any 3 out of the given 5 questions on Employability Skills (2 x 3 = 6 marks)


Answer each question in 20 – 30 words.
Q. 6 In SMART goals, what does ‘A’ stand for? Explain. 2

Q. 7 Sameera is always punctual at school. She has a regular schedule that she follows 2
every day. She plans for study and play time in advance. Enlist the four steps Sameera
must have followed for effective time management.

Q. 8 Enlist any two methods to protect our data on the computer. 2

Q. 9 What do entrepreneurs do when they run their business? Mention any two points. 2
Q. 10 Raj has a small convenience store in his locality. There are many other convenience 2
stores in the area. Yet, Raj’s store survives the competition and does well.
Which stage of an entrepreneur's career process can you relate this to? Explain.

Answer any 4 out of the given 6 questions in 20 – 30 words each (2 x 4 = 8 marks)


Q. 11 How do you understand whether a machine/application is AI based or not? Explain 2
with the help of an example.

Q. 12 If you do an image search for vacations on a popular search engine, the first few 2
searches mostly return the picture of beaches. What is the concern here? Explain.

Page 7 of 9
Q. 13 Suhana works for a company wherein she was assigned the task of developing a 2
project using AI project cycle. She knew that the first stage was scoping the problem.
Help her list the remaining stages that she must go through to develop the project.

Q. 14 What will be the results of conversion of the term, ‘happily’ in the process of 2
stemming and lemmatization? Which process takes longer time for execution?

Q. 15 What do we get from the “bag of words'' algorithm? 2

Q. 16 People of a village are totally dependent on the farmers for their daily food items. 2
Farmers grow new seeds by checking the weather conditions every year. An AI
model is being deployed in the village which predicts the chances of heavy rain to
alert farmers which helps them in doing the farming at the right time. Which
evaluation parameter out of precision, recall and F1 Score is best to evaluate the
performance of this AI model? Explain.

Answer any 3 out of the given 5 questions in 50– 80 words each (4 x 3 = 12 marks)

Q. 17 Ashwat is amazed to learn about his sister Ananya who is multi-talented and has 4
excelled in academics, music, dancing, sports and painting. He was quite curious
when Ananya told him that he too possessed all these intelligences like every human
being does, but only at different levels. He wondered which intelligence she was
talking about. Can you help Ashwat in learning about different types of intelligences
by naming and explaining any four types of intelligences?

Q. 18 Samarth attended a seminar on Artificial Intelligence and has now been asked to 4
write a report on his learnings from the seminar. Being a non-technical person, he
understood that the AI enabled machine uses data of different formats in many of
the daily based applications but failed to sync it with the right terminologies and
express the details. Help Samarth define Artificial Intelligence, list the three domains
of AI and the data that is used in these domains.

Q. 19 Neural networks are said to be modelled the way how neurons in the human brain 4
behave. A similar system is mimicked by the AI machine to perform certain tasks.
Explain how neural networks work in an AI model and mention any three features
of Neural Networks.

Q. 20 Samiksha, a student of class X was exploring the Natural Language Processing 4


domain. She got stuck while performing the text normalisation. Help her to
normalise the text on the segmented sentences given below:

Document 1: Akash and Ajay are best friends.


Document 2: Akash likes to play football but Ajay prefers to play online games.

Page 8 of 9
Q. 21 Automated trade industry has developed an AI model which predicts the selling 4
and purchasing of automobiles. During testing, the AI model came up with the
following predictions.

Reality
Confusion Matrix
Yes No

Yes 60 25
Predicted
No 05 10

(i) How many total tests have been performed in the above scenario?
(ii) Calculate precision, recall and F1 Score.

Page 9 of 9
CBSE | DEPARTMENT OF SKILL EDUCATION
ARTIFICIAL INTELLIGENCE (SUBJECT CODE - 417)
MARKING SCHEME FOR CLASS X (SESSION 2022-2023)
Max. Time: 2 Hours Max. Marks: 50
General Instructions:
1. Please read the instructions carefully.
2. This Question Paper consists of 21 questions in two sections – Section A & Section B.
3. Section A has Objective type questions whereas Section B contains Subjective type questions.
4. Out of the given (5 + 16 =) 21 questions, a candidate has to answer (5 + 10 =) 15 questions in the
allotted (maximum) time of 2 hours.
5. All questions of a particular section must be attempted in the correct order.
6. SECTION A - OBJECTIVE TYPE QUESTIONS (24 MARKS):
i. This section has 05 questions.
ii. There is no negative marking.
iii. Do as per the instructions given.
iv. Marks allotted are mentioned against each question/part.
7. SECTION B – SUBJECTIVE TYPE QUESTIONS (26 MARKS):
i. This section contains 16 questions.
ii. A candidate has to do 10 questions.
iii. Do as per the instructions given.
iv. Marks allotted are mentioned against each question/part.

SECTION A: OBJECTIVE TYPE QUESTIONS


Source Material
Page no.
Q. (NCERT/PSSCIVE/
QUESTION Unit/ Chap. No. of source Marks
No. CBSE Study
material
Material)
Q. 1 Answer any 4 out of the given 6 questions on Employability Skills (1 x 4 = 4 marks)
i. (a) Self-motivation Employability Unit 2 Self- 51 1
Skills Management
NCERT Skills

ii. (b) driving during rush hour Employability Unit 2 Self- 41 1


Skills Management
NCERT Skills

iii. (c) Double clicking Employability Unit 3 68 1


Skills Information
NCERT and
communication
Technology
Skills
iv. (c) It can overheat Employability Unit 3 78 1
Skills Information
NCERT and
communication

1
Technology
Skills
v. (c) By buying jute from the local farmer Employability Unit 4 87 1
and by providing jobs to local women Skills Entrepreneurial
NCERT Skills

vi. (c) Manages the business Employability Unit 4 94 1


Skills Entrepreneurial
NCERT Skills

Q. 2 Answer any 5 out of the given 6 questions (1 x 5 = 5 marks)


i. Google Maps, Ola, Uber Facilitator Unit 1 15 1
(any relevant application name with Handbook
similar functionality can be considered)
ii. Naturalist Intelligence Facilitator Unit 1 11 1
Handbook
iii. (b) (ii) and (iii) Facilitator Unit 1 20 1
Handbook
iv. (c) Natural Language Processing Facilitator Unit 1 22 1
Handbook
v. (a) Data Privacy Facilitator Unit 1 25 1
Handbook
vi. True Facilitator Unit 1 12 1
Handbook
Q. 3 Answer any 5 out of the given 6 questions (1 x 5 = 5 marks)
i. Problem Statement Template Facilitator Unit 2 33 1
Handbook
ii. (b) Hidden layer Facilitator Unit 2 40 1
Handbook
iii. (d) Neural networks Facilitator Unit 2 41 1
Handbook
iv. (b) Both i and iv Facilitator Unit 2 34 1
Handbook
v. (a) 17 Facilitator Unit 2 30 1
Handbook
vi. (c) Clustering Facilitator Unit 2 39 1
Handbook
Q. 4 Answer any 5 out of the given 6 questions (1 x 5 = 5 marks)
i. Text Classification Facilitator Unit 6 101 1
Handbook
ii. (d) Text and Speech Facilitator Unit 6 99 1
Handbook
iii. Corpus Facilitator Unit 6 108 1
Handbook
iv. Term Frequency Inverse Document Facilitator Unit 6 114 1
Frequency Handbook
v. (a) 12 Facilitator Unit 6 113 1
Handbook
vi. Script bot Facilitator Unit 6 105 1
Handbook
Q. 5 Answer any 5 out of the given 6 questions (1 x 5 = 5 marks)
i. Prediction Facilitator Unit 7 119 1
Handbook

2
ii. Confusion Matrix Facilitator Unit 7 122 1
Handbook
iii. (c) False Negative Facilitator Unit 7 121 1
Handbook
iv. (b) Precision and Recall Facilitator Unit 7 127 1
Handbook
v. (d) The training accuracy and test Facilitator Unit 7 119 1
accuracy both are low Handbook
vi. (a) Accuracy Facilitator Unit 7 123 1
Handbook

SECTION B: SUBJECTIVE TYPE QUESTIONS


Source Material
Page no.
Q. (NCERT/PSSCIVE Unit/ Chap.
QUESTION of source Marks
No. / CBSE Study No.
material
Material)
Answer any 3 out of the given 5 questions on Employability Skills in 20 – 30 words each (2 x 3 = 6 marks)
Q. 6 In SMART goals, A refers for Achievable, Employability Unit 2 Self- Page 55, 2
it means breaking down big goals into Skills Manageme 56
smaller parts will make the goal NCERT nt Skills
achievable.
For example Bigger Goal: “I want to
become a teacher in my school.”
Breaking it into smaller goals:
Complete higher secondary
Complete Graduation
Complete B.Ed.
Apply for jobs in the teaching field
(1 mark for identification; 1 mark for
explanation)

Q. 7 The four steps of effective time Employability Unit 2 Self- 60 2


management which Sameera must have Skills Manageme
followed are: NCERT nt Skills
(i) Organise
(ii) Prioritise
(iii) Control
(iv) Track
(½ mark for every step, ½ * 4 =2)

Q. 8 Two methods to protect our data on the Employability Unit 3 81, 2


computer: Skills Information 82
1. Use passwords to login to your NCERT and
computer. communica
tion
2. Install Anti-virus and Firewall
Technology
3. Encrypt Data Skills
4. Secure sites
(1 mark for each correct method, 1*2=2)

3
Q. 9 I. Fulfill Customer Needs Employability Unit 4 86, 2
II. Use Local Materials Skills Entrepre- 87
III. Help Society NCERT neurial
IV. Create Jobs Skills
V. Sharing of Wealth
VI. Lower Price of Products
(1 mark for each correct point, 1*2=2)

Q. 10 This stage is the Survive stage of an Employability Unit 4 100 2


entrepreneur's career process. In this Skills Entrepre-
stage, even if there are many NCERT neurial
entrepreneurs in the market, the new Skills
entrepreneur has to remain in a
competitive market.
(1 mark for mentioning the stage; 1
mark for correct explanation)

Answer any 4 out of the given 6 questions in 20 – 30 words each (2 x 4 = 8 marks)


Q. 11 Any machine that has been trained with Facilitator Unit 1 16, 2
data and can make decisions/predictions Handbook 17
on its own can be termed as AI.
Eg:The bot or the automation machine is
not trained with any data is not an AI
while a chatbot that understands and
processes human language is an AI.
(1 mark for correct explanation; ½ mark
for example of AI; ½ mark for example of
not AI)

Q. 12 In the given scenario, we are concerned Facilitator Unit 1 26 2


about the bias. When we talk about a Handbook
machine, we know that it is artificial and
cannot think on its own. It can have
intelligence, but we cannot expect a
machine to have any biases of its own.
Any bias can transfer from the developer
to the machine while the algorithm is
being developed.
(1 mark for mentioning the term bias; 1
mark for the correct explanation)

4
Q. 13 Steps of AI project life cycle: Facilitator Unit 2 29 2
1. Data Acquisition Handbook
2. Data Exploration
3. Modelling
4. Evaluation
(½ mark for mentioning each stage,
½ *4=2)
Q. 14 Stemming Lemmatization Facilitator Unit 6 110, 2
happily happi happy Handbook 111
Process that takes longer time for
execution is lemmatization
(½ marks each for identifying the correct
stem and lemma; 1 mark for identifying
the correct process)
Q. 15 Bag of words gives us two things: Facilitator Unit 6 112 2
1. A vocabulary of words for the corpus Handbook
2. The frequency of these words
(number of times it has occurred in the
whole corpus)
(1 mark for each point)
Q. 16 Let us take each of the factor into Facilitator Unit 7 126, 2
consideration at once, Handbook 127
If precision is considered, FN cases will not
be taken into account, so it will be of great
loss as if the machine will predict there will
be no heavy rain, but if the rain occurred, it
will be a big monetary loss due to damage to
crops.
If only recall is considered, then FP cases will
not be taken into account. This situation will
also cause a big amount of loss, as all people
of the village are dependent on farmers for
food, and if the model predicts there will be
heavy rain and the farmers may not grow
crops, it will affect the basic needs of the
people.
Hence F1 Score is the best suited parameter
to test this AI model, which is the balance
between Precision and Recall.
(1 mark for identifying the term F1 score; 1
mark for relevant explanation)
Answer any 3 out of the given 5 questions in 50– 80 words each (4 x 3 = 12 marks)
Q. 17 All humans possess 9 types of Facilitator Unit 1 11 4
intelligence but at different levels. They Handbook
are:
1. Mathematical Logical Reasoning:
ability to regulate, measure, and
understand numerical symbols,
abstraction and logic.
2. Linguistic Intelligence: Language
processing skills both in terms of

5
understanding or implementation in
writing or verbally.
3. Spatial Visual Intelligence : ability to
perceive the visual world and the
relationship of one object to another.
4. Kinesthetic Intelligence : ability that is
related to how a person uses his limbs in
a skilled manner.
5. Musical Intelligence : ability to
recognize and create sounds, rhythms,
and sound patterns.
6. Intrapersonal Intelligence : Describes
how high the level of self-awareness
someone has is. Starting from realizing
weakness, strength, to his own feelings.
7. Existential Intelligence : An additional
category of intelligence relating to
religious and spiritual awareness.
8. Naturalist Intelligence : An additional
category of intelligence relating to the
ability to process information on the
environment around us.
9. Interpersonal intelligence : ability to
communicate with others by
understanding other people's feelings &
influence of the person.
(½ mark for the naming the intelligence;
½ mark for the explanation of the same;
(½ + ½) * 4 = 4)
Q. 18 Artificial Intelligence (AI) refers to any Facilitator Unit 1 21 4
technique that enables computers to Handbook
mimic human intelligence i.e., make
decisions, predict the future, learn and
improve on its own.
With respect to the type of data fed in
the AI model, AI models can be broadly
categorised into three domains:
1. Data sciences
2. Computer vision
3. Natural Language Processing
Data Science takes input in the form of
numeric and alphanumeric data.
Computer Vision takes input in the form
of images and videos.
Natural Language Processing takes input
in the form of text and speech.
(1 mark for definition of AI; ½ mark each
for the names of the domains; ½ mark
each for the type of data input to
domains)

6
Q. 19 Neural networks are loosely modelled Facilitator Unit 2 40, 4
after how neurons in the human brain Handbook 41
behave.
The features of a neural network are :
1. They are able to extract data
features automatically without
needing the input of the
programmer.
2. A neural network is essentially a
system of organizing machine
learning algorithms to perform
certain tasks.
3. It is a fast and efficient way to
solve problems for which the
dataset is very large, such as in
images.
(1 mark for how neural networks are
modelled; 1 mark each for relevant
feature of neural network)
Q. 20 Facilitator Unit 6 108 - 4
1. Tokenisation Handbook 111
Akash, and, Ajay, are, best, friends
Akash, likes, to, play, football, but,
Ajay, prefers, to, play, online, games
2. Removal of stopwords
Akash, Ajay, best, friends
Akash, likes, play, football, Ajay,
prefers, play, online, games
3. converting text to a common case
akash, ajay, best, friends
akash, likes, play, football, ajay,
prefers, play, online, games
4. Stemming/Lemmatisation
akash, ajay, best, friend
akash, like, play, football, ajay, prefer,
play, online, game
(1 mark for each step; 1*4=4)
Q. 21 (i) TP=60, TN=10, FP=25, FN=5 Facilitator Unit 7 124- 4
60+25+5+10=100 total cases have been Handbook 127
performed

(ii) (Note: For calculating Precision,


Recall and F1 score, we need not
multiply the formula by 100 as all these
parameters need to range between 0 to
1)
Precision =TP/(TP+FP)
=60/(60+25)
=60/85
=0.7

7
Recall=TP/(TP+FN)
=60/(60+5)
=60/65
=0.92

F1 Score=2*Precision*Recall/
(Precision+Recall)
=2*0.7*0.92/(0.7+0.92)
=0.79

(1 mark for total number of cases; 1


mark each for the calculation of
precision, recall and F1 score)

8
CBSE | DEPARTMENT OF SKILL EDUCATION
ARTIFICIAL INTELLIGENCE (SUBJECT CODE - 417)
Blue-print for Sample Question Paper for Class X (Session 2023-2024)
Max. Time: 2 Hours Max. Marks: 50
PART A - EMPLOYABILITY SKILLS (10 MARKS):
OBJECTIVE TYPE SHORT ANSWER
UNIT QUESTIONS TYPE QUESTIONS TOTAL
NAME OF THE UNIT
NO. QUESTIONS
1 MARK EACH 2 MARKS EACH

1 Communication Skills -II 1 1 2

2 Self-Management Skills - II 2 1 3

3 ICT Skills - II 1 1 2

4 Entrepreneurial Skills - II 1 1 2

5 Green Skills - II 1 1 2

TOTAL QUESTIONS 6 5 11

NO. OF QUESTIONS TO BE ANSWERED Any 4 Any 3 07

TOTAL MARKS 1x4=4 2x3=6 10 MARKS

PART B - SUBJECT SPECIFIC SKILLS (40 MARKS):


SHORT ANSWER DESCRIPTIVE/
OBJECTIVE TYPE
UNIT TYPE LONG ANS. TYPE TOTAL
NAME OF THE UNIT QUESTIONS
NO. QUESTIONS QUESTIONS QUESTIONS
1 MARK EACH 2 MARKS EACH 4 MARKS EACH
1 Introduction to AI 5 1 1 7
2 AI Project Cycle 3 1 2 6
3 Advance Python - - - -
4 Data Sciences 3 1 - 4
5 Computer Vision 3 1 - 4

6 Natural Language 5 1 1 7
Processing
7 Evaluation 5 1 1 7

TOTAL QUESTIONS 24 6 5 35

NO. OF QUESTIONS 20 Any 4 Any 3 27


TO BE ANSWERED

TOTAL MARKS 1 x 20 = 20 2x4=8 4 x 3 = 12 40 MARKS

417 - ARTIFICIAL INTELLIGENCE – CLASS X PAGE 1 OF 8


CBSE | DEPARTMENT OF SKILL EDUCATION
ARTIFICIAL INTELLIGENCE (SUBJECT CODE - 417)

Sample Question Paper for Class X (Session 2023-2024)


Max. Time: 2 Hours Max. Marks: 50

General Instructions:
1. Please read the instructions carefully.
2. This Question Paper consists of 21 questions in two sections: Section A & Section B.
3. Section A has Objective type questions whereas Section B contains Subjective type questions.
4. Out of the given (5 + 16 =) 21 questions, a candidate has to answer (5 + 10 =) 15 questions in the
allotted (maximum) time of 2 hours.
5. All questions of a particular section must be attempted in the correct order.
6. SECTION A - OBJECTIVE TYPE QUESTIONS (24 MARKS):
i. This section has 05 questions.
ii. Marks allotted are mentioned against each question/part.
iii. There is no negative marking.
iv. Do as per the instructions given.
7. SECTION B – SUBJECTIVE TYPE QUESTIONS (26 MARKS):
i. This section has 16 questions.
ii. A candidate has to do 10 questions.
iii. Do as per the instructions given.
iv. Marks allotted are mentioned against each question/part.

417 - ARTIFICIAL INTELLIGENCE – CLASS X PAGE 2 OF 8


SECTION A: OBJECTIVE TYPE QUESTIONS

Q. 1 Answer any 4 out of the given 6 questions on Employability Skills (1 x 4 = 4 marks)

i. “M D Gulati started with a small shop with his focus, dedication and clear ideas, MDH 1
became one of the most popular brands in India besides having a good reputation all
over the world”.
Which self-management skill is clearly visible in the given statement?

ii. When you bring the mouse over a file in File Explorer, it will show the details of that file. 1
This is known as
(a) Drag and drop
(b) Double click
(c) Hover
(d) Single click

iii. Assertion(A): A doctor works for a renowned hospital. 1


Reason(R): The statement given above is an example of wage employment.
(a) Both A and R are correct and R is the correct explanation of A
(b) Both A and R are correct but R is NOT the correct explanation of A
(c) A is correct but R is not correct
(d) A is not correct but R is correct

iv. _______ the work is all about identifying and noting how we spent our time, and 1
analysing how to spend our time effectively.
(a) Organising
(b) Prioritising
(c) Controlling
(d) Tracking

v. Remya traveled to Sweden from India to pursue her higher education. But she doesn't 1
know how to speak Swedish (language of Sweden). Because of this, she was unable to
find a part time job. This is an example of _____________
(a) Interpersonal barrier
(b) Physical barrier
(c) Organisational barrier
(d) Linguistic barrier

vi. “Efforts are made to increase the solar power generation so that our electricity needs are met 1
and at the same time we do not pollute the environment or use up natural resources”.
Which SDG can you relate this statement to?
(a) Life on land
(b) Clean water and sanitation
(c) Affordable and clean energy
(d) Reduced inequalities

417 - ARTIFICIAL INTELLIGENCE – CLASS X PAGE 3 OF 8


Q. 2 Answer any 5 out of the given 6 questions (1 x 5 = 5 marks)

i. Assertion (A)– One can be a good singer while the other can be a great athlete. 1
Reason(R) – Humans possess different types of intelligences but at different levels.
(a) Both A and R are correct and R is the correct explanation of A
(b) Both A and R are correct but R is not the correct explanation of A
(c) A is correct but R is not correct
(d) A is not correct but R is correct.

ii. The Indian Government banned a few apps stating – “servers in the hostile nation are 1
receiving and using the acquired data improperly”.
Which terminology suits best for this action?
(a) AI Ethics
(b) Data Privacy
(c) AI Bias
(d) AI Access

iii. Statment1: There are four layers in a neural network. 1


Statement2:The first layer of the neural network is known as the output layer.
(a) Both Statement1 and Statement2 are correct
(b) Both Statement1 and Statement2 are incorrect
(c) Statement1 is correct but Statement2 is incorrect
(d) Statement2 is correct but Statement1 is incorrect

iv. Observe the given graph and fill in the blank: 1

__________ the neural network, better is the performance.

v. _______ is a simple file format that stores data separated by commas. 1


(a) jpg
(b) doc
(c) csv
(d) png

vi. A corpus contains 4 documents in which the word ‘diet’ was appearing once in 1
document1. Identify the term in which we can categorise the word ‘diet’.
(a) Stop word
(b) Rare word
(c) Frequent word
(d) Removable word

417 - ARTIFICIAL INTELLIGENCE – CLASS X PAGE 4 OF 8


Q. 3 Answer any 5 out of the given 6 questions (1 x 5 = 5 marks)

i. Read the examples given below- 1


i. Using Chat GPT to write an email
ii. Face unlock technology of mobile phones using camera
iii. Turning off lights with IoT device
iv. Hand sanitizer dispenser having sensor
Choose the options that are not AI
(a) i and ii
(b) iii and i
(c) iii and iv
(d) i, iii and iv

ii. _____ refer to the type of features that we want to collect. 1

iii. Which of the following is an application of data science? 1


(a) Text summarization
(b) Target Advertisements
(c) Face lock in smartphones
(d) Email filters

iv. ____ is the process of finding instances of real-world objects in images or videos. 1
(a) Instance segmentation
(b) Object detection
(c) Classification
(d) Image segmentation

v. Identify the given Chat bot type: 1


It learns from its environment and experience. It also builds on its capabilities based on
the knowledge. These can collaborate with humans, working along-side them and
learning from their behavior.

vi. F1 Score is the measure of the balance between 1


(a) Accuracy and Precision
(b) Precision and Recall
(c) Recall and Accuracy
(d) Recall and Reality

Q. 4 Answer any 5 out of the given 6 questions (1 x 5 = 5 marks)

i. Srishti learnt about AI terminologies but was not able to recollect the term that is used 1
to refer to machines that perform tasks with vast amounts of data using neural
networks. Help her with the correct term.

ii. Statment1: The output given by the AI model is known as reality. 1


Statement2:The real scenario is known as Prediction.
(a)Both Statement1 and Statement2 are correct
(b)Both Statement1 and Statement2 are incorrect
(c) Statement1 is correct but Statement2 is incorrect
(d) Statement2 is correct but Statement1 is incorrect

417 - ARTIFICIAL INTELLIGENCE – CLASS X PAGE 5 OF 8


iii. Rajat has made a model which predicts the performance of Indian Cricket players in 1
upcoming matches. He collected the data of players’ performance with respect to
stadium, bowlers, opponent team and health. His model works with good accuracy and
precision value. Which of the statement given below is incorrect?
(a) Data gathered with respect to stadium, bowlers, opponent team and health is
known as Testing Data.
(b) Data given to an AI model to check accuracy and precision is Testing Data.
(c) Training data and testing data are acquired in the Data Acquisition stage.
(d) Training data is always larger as compared to testing data.
iv. How many channels does a colour image have? 1

v. Which feature of NLP helps in understanding the emotions of the people mentioned 1
with the feedback?
(a) Virtual Assistants
(b) Sentiment Analysis
(c) Text classification
(d) Automatic Summarization

vi. Sarthak made a face mask detector system for which he had collected the dataset and 1
used all the dataset to train the model. Then, he used the same data to evaluate the
model which resulted in the correct answer all the time but was not able to perform
with unknown dataset.
Name the concept.

Q. 5 Answer any 5 out of the given 6 questions (1 x 5 = 5 marks)

i. Aditi, a student of class XII developed a chatbot that clarifies the doubts of Economics 1
students. She trained the software with lots of data sets catering to all difficulty levels.
If any student would type or ask questions related to Economics, the software would
give an instant reply. Identify the domain of AI in the given scenario.
(a) Computer Vision
(b) Data Science
(c) Natural Language Processing
(d) None of these

ii. Which evaluation parameter takes into consideration all the correct predictions? 1

iii. ______ means a picture element which is the smallest unit of information that makes 1
up a picture.
(a) Vision
(b) Pics
(c) Pixel
(d) Piskel

iv. What do you mean by syntax of a language? 1


(a) Meaning of a sentence
(b) Grammatical structure of a sentence
(c) Semantics of a sentence
(d) Synonym of a sentence

417 - ARTIFICIAL INTELLIGENCE – CLASS X PAGE 6 OF 8


v. Which algorithms result in two things, a vocabulary of words and frequency of the 1
words in the corpus?
(a) Sentence segmentation
(b) Tokenisation
(c) Bag of words
(d) Text normalisation

vi. Which one of the following scenario result in a high false positive cost? 1
(a) viral outbreak
(b) forest fire
(c) flood
(d) spam filter

SECTION B: SUBJECTIVE TYPE QUESTIONS


Answer any 3 out of the given 5 questions on Employability Skills (2 x 3 = 6 marks)
Answer each question in 20 – 30 words.
Q. 6 List two best practices for effective communication. 2

Q. 7 What is the importance of setting goals in life? 2

Q. 8 “The Trojan Horse was a wooden horse said to have been used by the Greeks during the 2
Trojan War to enter the city of Troy and win the war”.
What does Trojan horse mean in computer terminology?

Q. 9 How is society helping entrepreneurs in their business? 2

Q. 10 Mention any two human activities that lead to environmental degradation. 2

Answer any 4 out of the given 6 questions in 20 – 30 words each (2 x 4 = 8 marks)


Q. 11 All of us use smartphones. When we install a new app, it asks us for several permissions 2
to access our phone’s data in different ways. Why do apps collect such data?

Q. 12 Sirisha and Divisha want to make a model which will organize the unlabeled input data 2
into groups based on features. Which learning model should they use and why?

Q. 13 Ajay wants to access data from various sources. Suggest him any two points that he 2
needs to keep in mind while accessing data from any data source.

Q. 14 Explain the term resolution with an example. 2

Q. 15 Identify any two stop words which should not be removed from the given sentence and 2
why?
Get help and support whether you're shopping now or need help with a past purchase.
Contact us at [email protected] or on our website www.pwershel.com

Q. 16 Draw the confusion matrix for the following data 2


• the number of true positive = 100
417 - ARTIFICIAL INTELLIGENCE – CLASS X PAGE 7 OF 8
• the number of true negative 47
• the number of false positive = 62
• the number of false negative = 290

Answer any 3 out of the given 5 questions in 50– 80 words each (4 x 3 = 12 marks)
Q. 17 Your grandmother watches you use AI applications. She wants to understand more 4
about it. Help her understand the term artificial intelligence by giving the right definition
and explain to her with an example how machines become artificially intelligent.

Q. 18 Akhil wants to learn how to scope the problem for an AI Project. Explain him the 4
following:
(a) 4W Problem Canvas
(b) Problem Statement Template

Q. 19 Identify and explain the types of the learning-based approaches in the figures given 4
below.

Figure 1

Figure 2

Q. 20 We, human beings, can read, write and understand many languages. But computers can 4
understand only machine language. Do you think we might face any challenges if we try
to teach computers how to understand and interact in human languages? Explain.

Q. 21 An AI model made the following sales prediction for a new mobile phone which they 4
have recently launched:
Reality
Confusion Matrix Yes No

Yes 50 40
Prediction
No 12 10
(i) Identify the total number of wrong predictions made by the model.
(ii) Calculate precision, recall and F1 Score.

417 - ARTIFICIAL INTELLIGENCE – CLASS X PAGE 8 OF 8


CBSE | DEPARTMENT OF SKILL EDUCATION
ARTIFICIAL INTELLIGENCE (SUBJECT CODE - 417)
MARKING SCHEME FOR CLASS X (SESSION 2023-2024)
Max. Time: 2 Hours Max. Marks: 50
General Instructions:
1. Please read the instructions carefully.
2. This Question Paper consists of 21 questions in two sections – Section A & Section B.
3. Section A has Objective type questions whereas Section B contains Subjective type questions.
4. Out of the given (5 + 16 =) 21 questions, a candidate has to answer (5 + 10 =) 15 questions in the
allotted (maximum) time of 2 hours.
5. All questions of a particular section must be attempted in the correct order.
6. SECTION A – OBJECTIVE TYPE QUESTIONS (24 MARKS):
i. This section has 05 questions.
ii. There is no negative marking.
iii. Do as per the instructions given.
iv. Marks allotted are mentioned against each question/part.
7. SECTION B – SUBJECTIVE TYPE QUESTIONS (26 MARKS):
i. This section contains 16 questions.
ii. A candidate has to do 10 questions.
iii. Do as per the instructions given.
iv. Marks allotted are mentioned against each question/part.

SECTION A: OBJECTIVE TYPE QUESTIONS


Source Material
Page no.
Q. (NCERT/PSSCIVE/ Unit/
QUESTION of source Marks
No. CBSE Study Chap. No.
material
Material)
Q. 1 Answer any 4 out of the given 6 questions on Employability Skills (1 x 4 = 4 marks)
i. Self-Motivation NCERT 2 52 1

ii. (c) Hover NCERT 3 67 1

(a) Both A and R are correct and R is the


iii. NCERT 4 100 1
correct explanation of A
iv. (d) Tracking NCERT 2 60 1

v. (d) Linguistic barrier NCERT 1 21 1

vi. (c) Affordable and clean energy NCERT 5 111 1


Q. 2 Answer any 5 out of the given 6 questions (1 x 5 = 5 marks)
(a) Both A and R are correct and R is the Facilitator
i. 1 11 1
correct explanation of A Handbook
Facilitator
ii. (b) Data Privacy 1 24 1
Handbook

417 - ARTIFICIAL INTELLIGENCE – CLASS X PAGE 1 OF 8


(a) Both Statement1 and Statement2 are Facilitator
iii. 2 40 1
incorrect Handbook
Facilitator
iv. Larger 2 40 1
Handbook
Facilitator
v. (c) csv 4 63 1
Handbook
Facilitator
vi. (b) Rare word 6 115 1
Handbook
Q. 3 Answer any 5 out of the given 6 questions (1 x 5 = 5 marks)
Facilitator
i. (c) iii and iv 1 16 1
Handbook
Facilitator
ii. Data Features/Data 2 35 1
Handbook
Facilitator
iii. (b) Target Advertisements 4 56 1
Handbook
Facilitator
iv. (b) Object Detection 5 78 1
Handbook
Facilitator
v. Smart Bot 6 105 1
Handbook
Facilitator
vi. (b) Precision and Recall 7 127 1
Handbook
Q. 4 Answer any 5 out of the given 6 questions (1 x 5 = 5 marks)
Facilitator
i. Deep Learning 1 21 1
Handbook
(b) Both Statement1 and Statement2 are Facilitator
ii. 7 119 1
incorrect Handbook
(a) Data gathered with respect to
Facilitator
iii. stadium, bowlers, opponent team and 4 60 1
Handbook
health is known as Testing Data.
Facilitator
iv. Three Channels 5 83 1
Handbook
Facilitator
v. (b) Sentiment Analysis 6 100 1
Handbook
Facilitator
vi. Overfitting 7 119 1
Handbook
Q. 5 Answer any 5 out of the given 6 questions (1 x 5 = 5 marks)
Facilitator
i. (c) Natural Language Processing 1 22 1
Handbook
Facilitator
ii. Accuracy 7 123 1
Handbook
Facilitator
iii. (c) Pixel 5 79 1
Handbook
Facilitator
iv. (b) Grammatical structure of a sentence 6 106 1
Handbook
Facilitator
v. (c) Bag of words 6 112 1
Handbook
Facilitator
vi. (d) spam filter 7 126 1
Handbook

417 - ARTIFICIAL INTELLIGENCE – CLASS X PAGE 2 OF 8


SECTION B: SUBJECTIVE TYPE QUESTIONS
Source Unit/ Page no. Marks
Material Chap. No. of source
Q.
QUESTION (NCERT/PSSCIV material
No.
E/ CBSE Study
Material)
Answer any 3 out of the given 5 questions on Employability Skills in 20 – 30 words each (2 x 3 = 6 marks)
Q. 6 1. Use simple language NCERT 1 22 2
2. Be respectful of others’ opinions
3. Do not form assumptions on culture, religion
or geography
4. Try to communicate in person as much as
possible
5. Use visuals
6. Take help of a translator to overcome
differences in language
(2 marks for any two correct points from the
above)
Q. 7 1. The process of goal setting in your life helps NCERT 2 55 2
you decide on how to live your life, where
you want to be, and how you want to be in
the future.
2. It helps you to focus on the end result
instead of less important work.
3. This will make you successful in your career
and personal life.
(2 marks for any one valid justification given
above or any relevant answer)

Q. 8 A Trojan Horse is a type of malware which NCERT 3 81 2


disguises itself i.e., it appears to be a useful
software program but once it reaches a
computer it starts behaving like a virus and
destroys data.
(1 mark for acting like useful program and, 1
mark for the words
destroying/corrupting/deleting data)
or
(only ½ marks will be allotted if only
malware/harmful program/virus/ term is
mentioned)

Q. 9 Society is helping entrepreneurs by NCERT 4 85 2


1. Creating needs
2. Providing raw material
3. buying/selling of items
4. making money
(2 marks for any two valid points)

417 - ARTIFICIAL INTELLIGENCE – CLASS X PAGE 3 OF 8


Q. 10 1. Use of fertilisers, pesticides,etc for NCERT 5 103 2
increasing the production of crops.
2. Cutting down of forests for personal use
such as construction of buildings, factories
etc

(1 mark for each valid point, or any two relevant


answers other than the above)

Answer any 4 out of the given 6 questions in 20 – 30 words each (2 x 4 = 8 marks)


Q. 11 1. To provide customized notifications and Facilitator Unit 1 25, 26 2
recommendations. Handbook
2. To improve the efficiency and accuracy of the
app.

(2 marks for any one correct point with


explanation)

Q. 12 Clustering model/Unsupervised learning is used Facilitator Unit 2 39 2


to organize the unlabeled input data into groups Handbook
based on features.
Clustering is an unsupervised learning algorithm
which can cluster unknown data according to
the patterns or trends identified out of it. The
patterns observed might be the ones which are
known to the developer or it might even come
up with some unique patterns out of it.
(1 mark for identifying the name of the
algorithm and 1 mark for explanation)

Q. 13 While accessing data from any of the data Facilitator Unit 4 63 2


sources, following points should be kept in Handbook
mind:
1. Data which is available for public usage only
should be taken up.
2. Personal datasets should only be used with
the consent of the owner.
3. One should never breach someone’s privacy
to collect data.
4. Data should only be taken from reliable
sources as the data collected from random
sources can be wrong or unusable.
5. Reliable sources of data ensure the
authenticity of data which helps in the
proper training of the AI model.
6. Data should be relevant to the problem.

(any two; 1 mark for each valid point)

417 - ARTIFICIAL INTELLIGENCE – CLASS X PAGE 4 OF 8


Q. 14 Resolution of an image refers to the number of Facilitator Unit 5 80 2
pixels in an image, across the width and height. Handbook
For example a monitor resolution of 1280×1024.
This means there are 1280 pixels from one side
to the other, and 1024 from top to bottom.

(1 mark for explanation; 1 mark for example)


Q. 15 Stopwords in the given sentence which should Facilitator Unit 6 110 2
not be removed are: Handbook
@, . (fullstop) ,_(underscore) , 123(numbers)
These tokens are generally considered as
stopwords, but in the above sentence, these
tokens are part of email id. removing these
tokens may lead to invalid website address and
email ID. So these words should not be removed
from the above sentence.

(1 mark for identifying any two stop words from


the above, and 1 mark for the valid
justification.)
Q. 16 Facilitator Unit 7 122 2
Handbook
Confusion Reality
Matrix
Yes No

Prediction Yes 100 62

No 290 47
(½ marks each for mapping the values in the
correct section, ½ *4=2 marks)

Answer any 3 out of the given 5 questions in 50– 80 words each (4 x 3 = 12 marks)
Q. 17 When a machine possesses the ability to mimic Facilitator Unit 1 14, 15 4
human traits, i.e., make decisions, predict the Handbook
future, learn and improve on its own, it is said
to have artificial intelligence. In other words,
you can say that a machine is artificially
intelligent when it can accomplish tasks by
itself - collect data, understand it, analyse it,
learn from it, and improve it.

Machines become intelligent once they are


trained with some data which helps them
achieve their tasks. AI machines also keep
updating their knowledge to optimise their
output. For example, Netflix gives us
recommendations on the basis of what we like.
Whenever we start liking a new genre, it
updates and gives better suggestions.

417 - ARTIFICIAL INTELLIGENCE – CLASS X PAGE 5 OF 8


(2 marks for definition of Artificial intelligence
which includes any of the highlighted terms,
2 mark for an example explanation of how
machines become intelligent)
or
(only 1 mark for any AI machine example which
mimic human traits without explanation)

Q. 18 The 4Ws Problem canvas helps in identifying the Facilitator Unit 2 31-34 4
key elements related to the problem. The 4Ws Handbook
are Who, What, Where and Why
● The “Who” block helps in analysing the
people getting affected directly or
indirectly due to the problem.
● The “What” block helps us to determine
the nature of the problem.
● The “Where” block helps us to look into
the situation in which the problem
arises, the context of it, and the
locations where it is prominent.
● The “Why” block suggests to us the
benefits which the stakeholders would
get from the solution and how it will
benefit them as well as the society

Problem Statement Template


Our [stakeholders] Who

Have a problem [need] What


that

When/while [context/ Where


location/
situation]

An ideal solution [solution] Why


would be

(½ mark each for explanation of 4w s; 2 marks


for drawing the problem statement template
with correct words in it or explaining the
problem statement template)
or
(1 mark to be allotted if only 4Ws are written
without explanation)

Q. 19 The learning-based approaches shown in the Facilitator Unit 2 37, 38 4


given figures are Supervised learning and Handbook
Unsupervised learning.

417 - ARTIFICIAL INTELLIGENCE – CLASS X PAGE 6 OF 8


Figure 1: In a supervised learning model, the
dataset which is fed to the machine is labelled.
In other words, we can say that the dataset is
known to the person who is training the
machine only then he/she is able to label the
data. A label is some information which can be
used as a tag for data.
Here, labelled images of dog and cat are fed into
the model and trained. The model correctly
identifies the given input as dog.

Figure 2: An unsupervised learning model works


on unlabelled dataset. This means that the data
which is fed to the machine is random and there
is a possibility that the person who is training
the model does not have any information
regarding it. The unsupervised learning models
are used to identify relationships, patterns and
trends out of the data which is fed into it. It
helps the user in understanding what the data is
about and what are the major features
identified by the machine in it.
Here, images of a set of animals are fed into the
AI model and the model clusters them based on
similar features

(1 mark each for identifying each term


supervised learning and unsupervised learning;
1 mark per explanation of each term)

Q. 20 Yes, we might face any challenges if we try to Facilitator Unit 6 106, 107 4
teach computers how to understand and Handbook
interact in human languages.

The possible difficulties are:


1. Arrangement of the words and meaning -
the computer has to identify the different
parts of a speech. Also, it may be extremely
difficult for a computer to understand the
meaning behind the language we use.

2. Multiple Meanings of a word - same word


can be used in a number of different ways
which according to the context of the
statement changes its meaning completely.

3. Perfect Syntax, no Meaning - Sometimes, a


statement can have a perfectly correct
syntax but it does not mean anything. For
example, take a look at this statement:

417 - ARTIFICIAL INTELLIGENCE – CLASS X PAGE 7 OF 8


Chickens feed extravagantly while the moon
drinks tea.
This statement is correct grammatically but
does this make any sense? In Human
language, a perfect balance of syntax and
semantics is important for better
understanding.

(1 mark for Yes and 1 mark each for the points


on possible difficulties)

Q. 21 Ans: (i)the total number of wrong predictions Facilitator Unit 7 126,127 4


made by the model is the sum of false positive Handbook
and false negative. FP+FN=40+12= 52

(ii) Precision=TP/(TP+FP)

=50/(50+40)

50/90

=0.55

Recall=TP/(TP+FN)

=50/(50+12)

=50/62

=.81

F1 Score = 2*Precision*Recall/(Precision+Recall)

=2*0.55*.81/(.55+.81)

=.891/1.36

=0.65

(1 marks for part (i) and ½ mark for each


formula and ½ mark each for substitution of
values in part(ii))

Please note: the mathematical calculations can


be ignored

417 - ARTIFICIAL INTELLIGENCE – CLASS X PAGE 8 OF 8


CBSE | DEPARTMENT OF SKILL EDUCATION
ARTIFICIAL INTELLIGENCE (SUBJECT CODE: 417)
Blueprint for Sample Question Paper for Class X (Session 2020-2021)

Max. Time: 2 Hours Max. Marks: 50

PART A - EMPLOYABILITY SKILLS (10 MARKS):


OBJECTIVE SHORT ANSWER
UNIT TYPE TYPE TOTAL
NAME OF THE UNIT
NO. QUESTIONS QUESTIONS QUESTIONS
1 MARK EACH 2 MARKS EACH
1 Communication Skills-II 1 1 2
2 Self-Management Skills-II 2 1 3
Information and Communication
3 1 1 2
Technology Skills-II
4 Entrepreneurial Skills-II 1 1 2
5 Green Skills-II 1 1 2
TOTAL QUESTIONS 6 5 11
NO. OF QUESTIONS TO BE ANSWERED Any 4 Any 3 Any 7
TOTAL MARKS 1x4=4 2x3=6 10 MARKS

PART B - SUBJECT SPECIFIC SKILLS (40 MARKS):


DESCRIPTIVE/
OBJECTIVE SHORT
LONG ANS.
UNIT TYPE ANSWER TYPE TOTAL
NAME OF THE UNIT TYPE
NO. QUESTIONS QUESTIONS QUESTIONS
QUESTIONS
1 MARK EACH 2 MARKS EACH 4 MARKS EACH
Introduction to
I Artificial 6 1 1 8
Intelligence (AI)
II AI Project Cycle 6 2 1 9
Natural Language
VI 6 2 2 10
Processing
VII Evaluation 6 1 1 8
TOTAL QUESTIONS 24 6 5 35
NO. OF QUESTIONS
20 Any 4 Any 3 Any 27
TO BE ANSWERED
TOTAL MARKS 1 x 20 = 20 2x4=8 4 x 3 = 12 40 MARKS

Page No. 1 of 8
CBSE | DEPARTMENT OF SKILL EDUCATION
ARTIFICIAL INTELLIGENCE (SUBJECT CODE: 417)
Sample Question Paper for Class X (Session 2020-2021)

Max. Time: 2 Hours Max. Marks: 50

General Instructions:
1. Please read the instructions carefully.

2. This Question Paper consists of 21 questions in two sections: Section A & Section B.

3. Section A has Objective type questions whereas Section B contains Subjective type
questions.

4. Out of the given (5 + 16 =) 21 questions, a candidate has to answer (5 + 10 =) 15


questions in the allotted (maximum) time of 2 hours.

5. All questions of a particular section must be attempted in the correct order.

6. SECTION A - OBJECTIVE TYPE QUESTIONS (24 MARKS):

(i) This section has 05 questions.


(ii) Marks allotted are mentioned against each question/part.
(iii) There is no negative marking.
(iv) Do as per the instructions given.
7. SECTION B – SUBJECTIVE TYPE QUESTIONS (26 MARKS):

(i) This section has 16 questions.


(ii) A candidate has to do 10 questions.
(iii) Do as per the instructions given.
(iv) Marks allotted are mentioned against each question/part.

Page No. 2 of 8
SECTION A: OBJECTIVE TYPE QUESTIONS
Q. 1 Answer any 4 out of the given 6 questions on Employability Skills
(1 x 4 = 4 marks)
i ___________________ is the final component in the process of 1
communication as it defines the response given by the receiver to the sender.
a) Response
b) Request
c) Feedback
d) Notice

ii ___________________refers to focusing human efforts for maintaining a 1


healthy body and mind capable of better withstanding stressful situations
a) Mental Health
b) Emotional Health
c) Self-Management
d) Stress Management

iii Having conscious knowledge of your own self, capabilities, feelings and one’s 1
own character is called ____________.
a) Self-awareness
b) Self-motivation
c) Self-control
d) Independence

iv A ______________is a software program that attaches itself to other programs 1


and alters their behavior.
a) Operating system
b) Firewall
c) Antivirus
d) Computer Virus

v ______________refers to recruitment, employment, selection, training, 1


development and compensation of the employees with an organization.
a) Entrepreneurs
b) Management
c) Human Resource Management
d) Employer
vi ______________is caused when natural or a man-made disturbance disrupts 1
the natural balance of an ecosystem.
a) Pollution
b) Damage
c) Natural disaster
d) Ecological Imbalance

Q. 2 Answer any 5 out of the given 6 questions (1 x 5 = 5 marks)


i A _______________is divided into multiple layers and each layer is further 1
divided into several blocks called nodes.
a) Neural Networks
b) Convolutional Neural Network (CNN)
c) Machine learning algorithm
d) Hidden Layers
Page No. 3 of 8
ii The __________________canvas helps you in identifying the key elements 1
related to the problem.
a) Problem scoping
b) 4Ws Problem
c) Project cycle
d) Algorithm

iii _______is a domain of AI that depicts the capability of a machine to get and 1
analyse visual information and afterwards predict some decisions about it.
a) NLP
b) Data Sciences
c) Augmented Reality
d) Computer Vision

iv ____________is defined as the percentage of correct predictions out of all the 1


observations.
a) Predictions
b) Accuracy
c) Reality
d) F1 Score

v _________________is the sub-field of AI that is focused on enabling computers 1


to understand and process human languages.
a) Deep Learning
b) Machine Learning
c) NLP
d) Data Sciences

vi In___________________, the machine is trained with huge amounts of data 1


which helps it in training itself around the data.
a) Supervised Learning
b) Deep Learning
c) Classification
d) Unsupervised Learning

Q. 3 Answer any 5 out of the given 6 questions (1 x 5 = 5 marks)


i Expand CBT_______________ 1
a) Computer Behaved Training
b) Cognitive Behavioural Therapy
c) Consolidated Batch of trainers
d) Combined Basic Training
ii Name any 2 methods of collecting data. 1
a) Surveys and Interviews
b) Rumors and Myths
c) AI models and applications
d) Imagination and thoughts
iii What is the role of modelling in an NLP based AI model? 1
a) Modelling in NLP helps in processing of AI model
b) Modelling is required to make an AI model
c) In NLP, modelling requires data pre-processing only after which the
data is fed to the machine.
d) Modelling is used in simplification of data acquisition
Page No. 4 of 8
iv What will be the outcome, if the Prediction is “Yes” and it matches with 1
the Reality? What will be the outcome, if the Prediction is “Yes” and it
does not match the Reality?
a) True Positive, True Negative
b) True Negative, False Negative
c) True Negative, False Positive
d) True Positive, False Positive

v Recall-Evaluation method is 1
a) defined as the fraction of positive cases that are correctly
identified.
b) defined as the percentage of true positive cases versus all the cases
where the prediction is true.
c) defined as the percentage of correct predictions out of all the
observations.
d) comparison between the prediction and reality

vi Give 2 examples of Supervised Learning models. 1


a) Classification and Regression
b) Clustering and Dimensionality Reduction
c) Rule Based and Learning Based
d) Classification and Clustering

Q. 4 Answer any 5 out of the given 6 questions (1 x 5 = 5 marks)


i Define Machine Learning. 1
a) Machine learning is the study of computer algorithms that improve
automatically through experience.
b) Refers to any technique that enables computers to mimic human
intelligence.
c) Machine learning refers to computer systems (both machines and
software) enables machines to perform tasks for which it is
programmed.
d) Machine Learning refers to projects that allow the machine to work
on a particular logic.

ii Give one example of an application which uses augmented reality. 1

iii Differentiate between Prediction and Reality. 1


a) Prediction is the input given to the machine to receive the expected
result of the reality.
b) Prediction is the output given to match the reality.
c) The prediction is the output which is given by the machine and the
reality is the real scenario in which the prediction has been made.
d) Prediction and reality both can be used interchangeably.

iv The term Sentence Segmentation is 1


a) the whole corpus is divided into sentences
b) to undergo several steps to normalise the text to a lower level
c) in which each sentence is then further divided into tokens
d) the process in which the affixes of words are removed

Page No. 5 of 8
v Which of the following statements is true for the term Evaluation? 1
a) Helps in classifying the type and genre of a document.
b) It helps in predicting the topic for a corpus.
c) Helps in understanding the reliability of any AI model
d) Process to extract the important information out of a corpus.

vi Which of the following is not part of the AI Project Cycle? 1


a) Data Exploration
b) Modelling
c) Testing
d) Problem Scoping

Q. 5 Answer any 5 out of the given 6 questions (1 x 5 = 5 marks)


i ________________________ refers to the AI modelling where the machine 1
learns by itself.
a) Learning Based
b) Rule Based
c) Machine Learning
d) Data Sciences

ii Prediction and Reality can be easily mapped together with the help of : 1
a) Prediction
b) Reality
c) Accuracy
d) Confusion Matrix

iii ___________________ is an example of Applications of Natural Language 1


Processing.
a) Evaluation
b) Automatic Summarization
c) Deep Learning
d) Problem Scoping

iv _________________ is the last stage of the AI project Life cycle. 1


a) Problem Scoping
b) Evaluation
c) Modelling
d) Data Acquisition

v In __________________, the machine is trained with huge amounts of data 1


which helps it in training itself around the data.
a) Machine Learning
b) Artificial Intelligence
c) NLP
d) Deep Learning

vi In ___________, input to machines can be photographs, videos and pictures 1


from thermal or infrared sensors, indicators and different sources.
a) Computer Vision
b) Data Acquisition
c) Data Collection
d) Machine learning

Page No. 6 of 8
SECTION B: SUBJECTIVE TYPE QUESTIONS
Answer any 3 out of the given 5 questions on Employability Skills (2 x 3 = 6 marks)

Q. 6 Name the four main categories of Communication Styles. 2

Q. 7 List any 4 activities that help in stress management. 2

Q. 8 What are antivirus? Name any 2 antiviruses. 2

Q. 9 Name any 4 qualities of an entrepreneur. 2

Q. 10 Name any 4 man-made disruptions that cause ecological imbalance. 2

Answer any 4 out of the given 6 questions in 20 – 30 words each (2 x 4 = 8 marks)

Q. 11 Give 2 points of difference between a script-bot and a smart-bot 2

Q. 12 Define the term Machine Learning. Also give 2 applications of Machine 2


Learning in our daily lives.

Q. 13 Differentiate between Classification and Regression. 2

Q. 14 Explain the term Neural Networks. 2

Q. 15 Explain the term Text Normalisation in Data Processing. 2

Q. 16 What is F1 Score in Evaluation? 2

Answer any 3 out of the given 5 questions in 50– 80 words each (4 x 3 = 12 marks)

Q. 17 Categorize the following under Data Sciences, Machine Learning, Computer 4


Vision and NLP:

The latest technological advancements have made our lives convenient.


Google Home, Alexa and Siri have been a huge help to non-tech savvy
people. Features like Facial recognition and Facelock have added additional
security to our gadgets. These advancements have also contributed in
making our needs more approachable and convenient. Now you can even
check the prices with Price comparison websites and order groceries online
with chatbots. Did you know that you can even find how you are going to
look when you grow old? Faceapps and Snapchat filters have made this
possible!

Page No. 7 of 8
Q. 18 Create a 4W Project Canvas for the following: 4

As more and more new technologies get into play, risks will get more
concentrated into a common network. Cybersecurity becomes extremely
complicated in such scenarios and goes beyond the control of firewalls. It
will not be able to detect unusual activity and patterns including the
movement of data.

Think how AI algorithms can scrape through vast amounts of logs to identify
susceptible user behaviour. Use an AI project cycle to clearly identify the
scope, how you will collect data, model and evaluation parameters.

Q. 19 Differentiate between stemming and lemmatization. Explain with the help 4


of an example.

Q. 20 Write the applications of NLP (Natural Language Processing). (Any four) 4

Q. 21 Imagine that you have come up with an AI based prediction model which 4
has been deployed on the roads to check traffic jams. Now, the objective
of the model is to predict whether there will be a traffic jam or not. Now,
to understand the efficiency of this model, we need to check if the
predictions which it makes are correct or not. Thus, there exist two
conditions which we need to ponder upon: Prediction and Reality.

Traffic Jams have become a common part of our lives nowadays. Living in
an urban area means you have to face traffic each and every time you get
out on the road. Mostly, school students opt for buses to go to school. Many
times, the bus gets late due to such jams and the students are not able to
reach their school on time.

Considering all the possible situations make a Confusion Matrix for the
above situation.

Page No. 8 of 8
CBSE | DEPARTMENT OF SKILL EDUCATION
ARTIFICIAL INTELLIGENCE (SUBJECT CODE: 417)
Marking Scheme for the Sample Question Paper for Class X
(Session 2020-2021)

Max. Time: 2 Hours Max. Marks: 50

General Instructions:
1. Please read the instructions carefully.

2. This Question Paper consists of 21 questions in two sections: Section A & Section B.

3. Section A has Objective type questions whereas Section B contains Subjective type
questions.

4. Out of the given (5 + 16 =) 21 questions, a candidate has to answer (5 + 10 =) 15


questions in the allotted (maximum) time of 2 hours.

5. All questions of a particular section must be attempted in the correct order.

6. SECTION A - OBJECTIVE TYPE QUESTIONS (24 MARKS):

(i) This section has 05 questions.


(ii) Marks allotted are mentioned against each question/part.
(iii) There is no negative marking.
(iv) Do as per the instructions given.
7. SECTION B – SUBJECTIVE TYPE QUESTIONS (26 MARKS):

(i) This section has 16 questions.


(ii) A candidate has to do 10 questions.
(iii) Do as per the instructions given.
(iv) Marks allotted are mentioned against each question/part.

Page No. 1 of 11
SECTION A: OBJECTIVE TYPE QUESTIONS
Q. 1 Answer any 4 out of the given 6 questions on Employability Skills
(1 x 4 = 4 marks)
i ___________________ is the final component in the process of 1
communication as it defines the response given by the receiver to the sender.
a) Response
b) Request
c) Feedback
d) Notice

Ans: Feedback
ii ___________________refers to focusing human efforts for maintaining a 1
healthy body and mind capable of better withstanding stressful situations
a) Mental Health
b) Emotional Health
c) Self-Management
d) Stress Management

Ans: d) Stress Management


iii Having conscious knowledge of your own self, capabilities, feelings and one’s 1
own character is called ____________.
a) Self-awareness
b) Self-motivation
c) Self-control
d) Independence

Ans: a) Self-awareness
iv A ______________is a software program that attaches itself to other 1
programs and alters their behavior.
a) Operating system
b) Firewall
c) Antivirus
d) Computer Virus

Ans: d) Computer Virus


v ______________refers to recruitment, employment, selection, training, 1
development and compensation of the employees with an organization.
a) Entrepreneurs
b) Management
c) Human Resource Management
d) Employer
Ans: c) Human Resource Management
vi ______________is caused when natural or a man-made disturbance disrupts 1
the natural balance of an ecosystem.
a) Pollution
b) Damage
c) Natural disaster
d) Ecological Imbalance

Ans: d) Ecological Imbalance

Page No. 2 of 11
Q. 2 Answer any 5 out of the given 6 questions (1 x 5 = 5 marks)
i A _______________is divided into multiple layers and each layer is further 1
divided into several blocks called nodes.
a) Neural Networks
b) Convolutional Neural Network (CNN)
c) Machine learning algorithm
d) Hidden Layers

Ans: a) Neural Network


ii The __________________canvas helps you in identifying the key elements 1
related to the problem.
a) Problem scoping
b) 4Ws Problem
c) Project cycle
d) Algorithm

Ans: b) 4Ws Problem


iii _______is a domain of AI that depicts the capability of a machine to get and 1
analyse visual information and afterwards predict some decisions about it.
a) NLP
b) Data Sciences
c) Augmented Reality
d) Computer Vision

Ans: d) Computer Vision


iv ____________is defined as the percentage of correct predictions out of all 1
the observations.
a) Predictions
b) Accuracy
c) Reality
d) F1 Score
Ans: b) Accuracy
v _________________is the sub-field of AI that is focused on enabling 1
computers to understand and process human languages.
a) Deep Learning
b) Machine Learning
c) NLP
d) Data Sciences
Ans: c) NLP
vi In___________________, the machine is trained with huge amounts of data 1
which helps it in training itself around the data.
a) Supervised Learning
b) Deep Learning
c) Classification
d) Unsupervised Learning
Ans: b) Deep Learning

Page No. 3 of 11
Q. 3 Answer any 5 out of the given 6 questions (1 x 5 = 5 marks)
i Expand CBT_______________ 1
a) Computer Behaved Training
b) Cognitive Behavioural Therapy
c) Consolidated Batch of trainers
d) Combined Basic Training

Ans: b) Cognitive Behavioural Therapy


ii Name any 2 methods of collecting data. 1
a) Surveys and Interviews
b) Rumors and Myths
c) AI models and applications
d) Imagination and thoughts

Ans: a) Surveys and Interviews


(Any two of the following)
Surveys, Observing the therapist’s sessions, Databases available on the
internet, Interviews, etc.
iii What is the role of modelling in an NLP based AI model? 1
a) Modelling in NLP helps in processing of AI model
b) Modelling is required to make an AI model
c) In NLP, modelling requires data pre-processing only after which the
data is fed to the machine.
d) Modelling is used in simplification of data acquisition

Ans: c)In NLP, modelling requires data pre-processing only after which the
data is fed to the machine.
iv What will be the outcome, if the Prediction is “Yes” and it matches with the 1
Reality? What will be the outcome, if the Prediction is “Yes” and it does not
match the Reality?
a) True Positive, True Negative
b) True Negative, False Negative
c) True Negative, False Positive
d) True Positive, False Positive

Ans: d) True Positive, False Positive


v Recall-Evaluation method is 1
a) defined as the fraction of positive cases that are correctly identified.
b) defined as the percentage of true positive cases versus all the cases
where the prediction is true.
c) defined as the percentage of correct predictions out of all the
observations.
d) comparison between the prediction and reality

Ans: a) defined as the fraction of positive cases that are correctly identified.
vi Give 2 examples of Supervised Learning models. 1
a) Classification and Regression
b) Clustering and Dimensionality Reduction
c) Rule Based and Learning Based
d) Classification and Clustering

Ans: a) Classification and Regression

Page No. 4 of 11
Q. 4 Answer any 5 out of the given 6 questions (1 x 5 = 5 marks)
i Define Machine Learning. 1
a) Machine learning is the study of computer algorithms that improve
automatically through experience.
b) Refers to any technique that enables computers to mimic human
intelligence.
c) Machine learning refers to computer systems (both machines and
software) enables machines to perform tasks for which it is
programmed.
d) Machine Learning refers to projects that allow the machine to work on
a particular logic.

Ans: a) Machine learning is the study of computer algorithms that improve


automatically through experience.

ii Give one example of an application which uses augmented reality. 1

Ans: Self Driving Cars


iii Differentiate between Prediction and Reality. 1
a) Prediction is the input given to the machine to receive the expected
result of the reality.
b) Prediction is the output given to match the reality.
c) The prediction is the output which is given by the machine and the
reality is the real scenario in which the prediction has been made.
d) Prediction and reality both can be used interchangeably.

Ans: c) The prediction is the output which is given by the machine and the
reality is the real scenario in which the prediction has been made.
iv The term Sentence Segmentation is 1
a) the whole corpus is divided into sentences
b) to undergo several steps to normalise the text to a lower level
c) in which each sentence is then further divided into tokens
d) the process in which the affixes of words are removed

Ans: a) the whole corpus is divided into sentences.

v Which of the following statements is true for the term Evaluation? 1


a) Helps in classifying the type and genre of a document.
b) It helps in predicting the topic for a corpus.
c) Helps in understanding the reliability of any AI model
d) Process to extract the important information out of a corpus.

Ans: c) Helps in understanding the reliability of any AI model

vi Which of the following is not part of the AI Project Cycle? 1


a) Data Exploration
b) Modelling
c) Testing
d) Problem Scoping

Ans: (c) Testing

Page No. 5 of 11
Q. 5 Answer any 5 out of the given 6 questions (1 x 5 = 5 marks)
i ________________________ refers to the AI modelling where the machine 1
learns by itself.
a) Learning Based
b) Rule Based
c) Machine Learning
d) Data Sciences

Ans: (a) Learning Based

ii Prediction and Reality can be easily mapped together with the help of : 1
a) Prediction
b) Reality
c) Accuracy
d) Confusion Matrix

Ans: (d) Confusion Matrix

iii ___________________ is an example of Applications of Natural Language 1


Processing.
a) Evaluation
b) Automatic Summarization
c) Deep Learning
d) Problem Scoping

Ans: (b) Automatic Summarization

iv _________________ is the last stage of the AI project Life cycle. 1


a) Problem Scoping
b) Evaluation
c) Modelling
d) Data Acquisition

Ans: (b) Evaluation

v In __________________, the machine is trained with huge amounts of data 1


which helps it in training itself around the data.
a) Machine Learning
b) Artificial Intelligence
c) NLP
d) Deep Learning

Ans: (d) Deep Learning

vi In ___________, input to machines can be photographs, videos and pictures 1


from thermal or infrared sensors, indicators and different sources.
a) Computer Vision
b) Data Acquisition
c) Data Collection
d) Machine learning

Ans: (a) Computer Vision

Page No. 6 of 11
SECTION B: SUBJECTIVE TYPE QUESTIONS
Answer any 3 out of the given 5 questions on Employability Skills (2 x 3 = 6 marks)
Part A: Employability Skills
Q. 6 Name the four main categories of Communication Styles. 2

Ans:
Verbal, Non - Verbal, Written and Visual
Q. 7 List any 4 activities that help in stress management. 2

Ans:
(Any 4 out of the following or any other appropriate activity)
 Positive Thinking,
 Physical Exercise,
 Yoga,
 Meditation,
 Nature Walks,
 Vacations,
 Laughing aloud,
 Listening to good music
Q. 8 What are antivirus? Name any 2 antiviruses. 2

Ans:
 Antivirus software is a program designed to detect and remove
malicious programs from the computer.
 Examples: (Any 4 out of the following or any other correct name of
the antivirus):
Microsoft Security essentials, Microsoft Defender, McAfee Virus Scan,
Norton AntiVirus, Quick Heal.

Q. 9 Name any 4 qualities of an entrepreneur. 2

Ans:
(Any 4 out of the following)
 Hard working,
 Optimistic,
 Independent,
 Energetic,
 Self-confident,
 Perseverant
Q. 10 Name any 4 man-made disruptions that cause ecological imbalance. 2

Ans:
(Any 4 out of the following)
 Deforestation,
 Degradation of Land and Soil Erosion,
 Overexploitation of Resources,
 Industrial and Atmospheric Pollution,
 Faulty Mining Practices,
 E waste generation

Page No. 7 of 11
Answer any 4 out of the given 6 questions in 20 – 30 words each (2 x 4 = 8 marks)
Q. 11 Give 2 points of difference between a script-bot and a smart-bot 2

Ans:
Script-bot Smart-bot

Script bots are easy to make Smart-bots are flexible and powerful

Script bots work around a script Smart bots work on bigger databases
with instructions of program and other resources directly
stored inside them

Mostly are Free and are Easy to Smart bots learn on its own with more
Integrate data

No or very little language Coding is required to take this up on


processing skills board

Limited functionality Has wide functionality

Q. 12 Define the term Machine Learning. Also give 2 applications of Machine 2


Learning in our daily lives.

Ans:
Machine Learning: It is a subset of Artificial Intelligence which enables
machines to improve at tasks with experience (data). The intention of
Machine Learning is to enable machines to learn by themselves using the
provided data and make accurate Predictions/ Decisions.

Machine Learning is used in Snapchat Filters, NETFLIX recommendation


system.
Q. 13 Differentiate between Classification and Regression. 2

Ans:
Classification Regression

This model works on a discrete Such models work on continuous


dataset which means the data data.
need not be continuous.

For example, in the grading For example, if you wish to predict


system, students are classified on your next salary, then you would put
the basis of the grades they in the data of your previous salary,
obtain with respect to their marks any increments, etc and would train
in the examination. the model.

Page No. 8 of 11
Q. 14 Explain the term Text Normalisation in Data Processing. 2

Ans:
The first step in Data processing is Text Normalisation. Text Normalisation
helps in cleaning up the textual data in such a way that it comes down to a
level where its complexity is lower than the actual data. In this we undergo
several steps to normalise the text to a lower level. We work on text from
multiple documents and the term used for the whole textual data from all
the documents altogether is known as corpus.

Q. 15 Name any 2 applications of Natural Language Processing which are used in 2


the real-life scenario.

Ans: (Any 4 out of the following or any other appropriate activity)


 Automatic Summarization,
 Sentiment Analysis,
 Text classification,
 Virtual Assistants

Q. 16 What is F1 Score in Evaluation? 2

Ans: F1 score can be defined as the measure of balance between precision


and recall.

Precision * Recall
F1Score = 2 *
Precision + Recall

Answer any 3 out of the given 5 questions in 50– 80 words each (4 x 3 = 12 marks)
Q. 17 Categorize the following under Data Sciences, Machine Learning, Computer 4
Vision and NLP.

The latest technological advancements have made our lives convenient.


Google Home, Alexa and Siri have been a huge help to non-tech savvy people.
Features like Facial recognition and Facelock have added additional security
to our gadgets. These advancements have also contributed in making our
needs more approachable and convenient. Now you can even check the prices
with Price comparison websites and order groceries online with chatbots.
Did you know that you can even find how you are going to look when you grow
old? Faceapps and Snapchat filters have made this possible!

Ans:
 Alexa, Siri-NLP, Facial Recognition - Computer Vision
 Facelock - Computer Vision
 Price comparison websites - Data Sciences
 Chatbots - NLP
 Faceapps -NLP
 Snapchat Filters - Machine Learning

Page No. 9 of 11
Q. 18 Create a 4W Project Canvas for the following. 4

As more and more new technologies get into play, risks will get more
concentrated into a common network. Cybersecurity becomes extremely
complicated in such scenarios and goes beyond the control of firewalls. It
will not be able to detect unusual activity and patterns including the
movement of data.

Think how AI algorithms can scrape through vast amounts of logs to identify
susceptible user behaviour. Use an AI project cycle to clearly identify the
scope, how you will collect data, model and evaluation parameters.

Ans:

OUR [stakeholders] People who are using the new WHO


technology

HAS/ HAVE [issue, problem, need] Cyber security is the WHAT


PROBLEM need when so much of the flow of data is not
THAT monitored or escapes the antiviruses/ firewall
systems.

WHEN/ WHILE [context/situation] The problem is in the use WHERE


of the latest technology where vast amounts
of data is at risk.

AN IDEAL [benefit of solution to them] An effective AI WHY


SOLUTION system which is able to detect the flow of
WOULD data and also report unusual activity

Q. 19 Differentiate between stemming and lemmatization. Explain with the help 4


of an example.

Ans:
Stemming is the process in which the affixes of words are removed and the
words are converted to their base form.
In lemmatization, the word we get after affix removal (also known as lemma)
is a meaningful one. Lemmatization makes sure that lemma is a word with
meaning and hence it takes a longer time to execute than stemming.
The difference between the stemming and lemmatization can be depicted
by the following example:

CARING lemmatization CARE

CARING stemming CAR

Page No. 10 of 11
Q. 20 Write the applications of NLP (Natural Language Processing). (Any four) 4

Ans:
1. Automatic Summarization: Automatic summarization is relevant not only
for summarizing the meaning of documents and information, but also to
understand the emotional meanings within the information, such as in
collecting data from social media.
2. Sentiment Analysis: The goal of sentiment analysis is to identify
sentiment among several posts or even in the same post where emotion
is not always explicitly expressed.
3. Text classification : Text classification makes it possible to assign
predefined categories to a document and organize it to help you find the
information you need or simplify some activities.
4. Virtual Assistants: With the help of speech recognition, these assistants
can not only detect our speech but can also make sense out of it.
Q. 21 Imagine that you have come up with an AI based prediction model which has 4
been deployed on the roads to check traffic jams. Now, the objective of the
model is to predict whether there will be a traffic jam or not. Now, to
understand the efficiency of this model, we need to check if the predictions
which it makes are correct or not. Thus, there exist two conditions which
we need to ponder upon: Prediction and Reality.

Traffic Jams have become a common part of our lives nowadays. Living in
an urban area means you have to face traffic each and every time you get
out on the road. Mostly, school students opt for buses to go to school. Many
times, the bus gets late due to such jams and the students are not able to
reach their school on time.

Considering all the possible situations make a Confusion Matrix for the above
situation.
Ans:
Case 1: Is there a traffic Jam?
Prediction: Yes Reality: Yes
True Positive
Case 2: Is there a traffic Jam?
Prediction: No Reality: No
True Negative
Case 3: Is there a traffic Jam?
Prediction: Yes Reality: No
False Positive
Case 4: Is there a traffic Jam?
Prediction: No Reality: Yes
False Negative

Reality
Confusion Matrix
Yes No
Yes True Positive False Positive
Prediction
No False Negative True Negative

Page No. 11 of 11
CBSE | DEPARTMENT OF SKILL EDUCATION
ARTIFICIAL INTELLIGENCE (SUBJECT CODE 417)
CLASS X (SESSION 2021-2022)
BLUE-PRINT FOR SAMPLE QUESTION PAPER FOR TERM – II
Max. Time Allowed: 1 Hour (60 min) Max. Marks: 25
PART A - EMPLOYABILITY SKILLS (05 MARKS):

NO. OF NO. OF NO. OF


QUESTIONS - QUESTIONS - QUESTIONS TOTAL
UNIT
NAME OF THE UNIT VSA SA - LA NUMBER OF
NO.
(1 MARK (2 MARKS (4 MARKS QUESTIONS
EACH) EACH) EACH)

4 Entrepreneurial Skills-II 2 1 - 3

5 Green Skills-II 2 1 - 3

TOTAL QUESTIONS 4 2 - 06
NO. OF QUESTIONS TO BE
Any 3 Any 1 - 04
ANSWERED
TOTAL MARKS 3 x 1 = 3 Marks 1 x 2 = 2 Marks - 05 Marks

PART B - SUBJECT SPECIFIC SKILLS (20 MARKS):

NO. OF NO. OF NO. OF


QUESTIONS - QUESTIONS - QUESTIONS - TOTAL
UNIT
NAME OF THE UNIT VSA SA LA NUMBER OF
NO.
(1 MARK (2 MARKS (4 MARKS QUESTIONS
EACH) EACH) EACH)

6 Natural Language Processing 4 4 2 10

7 Evaluation
2 2 1 5

TOTAL QUESTIONS 06 06 03 15

NO. OF QUESTIONS TO BE
04 04 02 10
ANSWERED

4 x 1= 4x2= 2x4=
TOTAL MARKS 20 Marks
04 Marks 08 Marks 08 Marks

TOTAL MARKS 05 (Part A) + 20 (Part B) = 25 MARKS

417-X-SQP-Term II (2021-2022) 1
CBSE | DEPARTMENT OF SKILL EDUCATION
ARTIFICIAL INTELLIGENCE (SUBJECT CODE 417)
CLASS X (SESSION 2021-2022)
SAMPLE QUESTION PAPER FOR TERM - II
Max. Time Allowed: 1 Hour (60 min) Max. Marks: 25
General Instructions:
1. Please read the instructions carefully
2. This Question Paper is divided into 03 sections, viz., Section A, Section B and Section C.
3. Section A is of 05 marks and has 06 questions on Employability Skills.
a) Questions numbers 1 to 4 are one mark questions. Attempt any three questions.
b) Questions numbers 5 and 6 are two marks questions. Attempt any one question.
4. Section B is of 12 marks and has 12 questions on Subject Specific Skills.
a) Questions numbers 7 to 12 are one mark questions. Attempt any four questions.
b) Questions numbers 13 to 18 are two marks questions. Attempt any four questions.
5. Section C is of 08 marks and has 03 competency-based questions.
a) Questions numbers 19 to 21 are four marks questions. Attempt any two questions.
6. Do as per the instructions given in the respective sections.
7. Marks allotted are mentioned against each section/question.
SECTION A (3 + 2 = 5 marks)
Answer any 3 questions out of the given 4 questions. Each question is of the mark. 1x3=3
Q.1 Write any two qualities for being a good entrepreneur. 1
Q.2 What is sustainable development? 1
Q.3 Entrepreneurship has a positive impact on society. Write down any two. 1
Q.4 How many sustainable development goals were formulated by the United Nations? 1
Answer any 1 question out of the given 2 questions. Each question is of mark. 2x1=2
“Entrepreneurs are born, not made.” Do you agree with this statement? Justify your
Q.5 2
answer.
Q.6 Enlist any 2 SDGs which are formulated to address the problems related to water? 2

SECTION B (4 + 8 = 12 marks)

Answer any 04 questions out of the given 06 questions 1x4=4


What will be the output of the word “studies” if we do the following:
Q.7 a. Lemmatization 1
b. Stemming
How many tokens are there in the sentence given below?
Traffic Jams have become a common part of our lives nowadays. Living in an
Q.8 1
urban area means you have to face traffic each and every time you get out on the
road. Mostly, school students opt for buses to go to school.
Q.9 What is a corpus? 1
Identify any 2 stopwords in the given sentence:
Pollution is the introduction of contaminants into the natural environment
Q.10 1
that cause adverse change.The three types of pollution are air pollution, water
pollution and land pollution.
Q.11 Why should we avoid using the training data for evaluation? 1

Q.12 What should be the value of F1 score if the model needs to have 100% accuracy? 1

417-X-SQP-Term II (2021-2022) 2
Answer any 04 questions out of the given 06 questions 2x4=8
“Automatic summarization is used in NLP applications”. Is the given statement
Q.13 correct? Justify your answer with an example. 2

Give an example of a situation wherein false positive would have a high cost associated
Q.14 2
with it.
Q.15 Write any two applications of TFIDF 2
Q.16 Write down the steps to implement bag of words algorithm. 2
Q.17 What is a confusion matrix? What is it used for? 2
Explain from the given graph, how the value and occurrence of a word are related in a
corpus?

Q.18 2

SECTION C (2 x 4 = 8 marks)
(COMPETENCY-BASED QUESTIONS)
Answer any 02 questions out of the given 03 questions

Q.19 Through a step-by-step process, calculate TFIDF for the given corpus 4
Document 1: Johny Johny, Yes Papa,
Document 2: Eating sugar? No Papa
Document 3: Telling lies? No Papa
Document 4: Open your mouth, Ha! Ha! Ha!

Q.20 The world is competitive nowadays. People face competition in even the tiniest 4
tasks and are expected to give their best at every point in time. When people are
unable to meet these expectations, they get stressed and could even go into
depression. We get to hear a lot of cases where people are depressed due to
reasons like peer pressure, studies, family issues, relationships, etc. and they
eventually get into something that is bad for them as well as for others. So, to
overcome this, Cognitive Behavioural Therapy (CBT) is considered to be one of
the best methods to address stress as it is easy to implement on people and also
gives good results. This therapy includes understanding the behaviour and
mindset of a person in their normal life. With the help of CBT, therapists help
people overcome their stress and live a happy life.
For the situation given above,
1. Write the problem statement template
2. List any two sources from which data can be collected.
3. How do we explore the data?

417-X-SQP-Term II (2021-2022) 3
Q.21 Take a look at the confusion matrix: 4
The Confusion Reality

Matrix Yes No

Yes True False


Positive Positive
Prediction (TP) (FP)
No False True
Negative Negative
(FN) (TN)
How do you calculate F1 score?

417-X-SQP-Term II (2021-2022) 4
CBSE | DEPARTMENT OF SKILL EDUCATION
ARTIFICIAL INTELLIGENCE (SUBJECT CODE 417)
CLASS X (SESSION 2021-2022)
SAMPLE QUESTION PAPER FOR TERM - II
Max. Time Allowed: 1 Hour (60 min) Max. Marks: 25
General Instructions:
1. Please read the instructions carefully
2. This Question Paper is divided into 03 sections, viz., Section A, Section B and Section C.
3. Section A is of 05 marks and has 06 questions on Employability Skills.
a) Questions numbers 1 to 4 are one mark questions. Attempt any three questions.
b) Questions numbers 5 and 6 are two marks questions. Attempt any one question.
4. Section B is of 12 marks and has 12 questions on Subject Specific Skills.
a) Questions numbers 7 to 12 are one mark questions. Attempt any four questions.
b) Questions numbers 13 to 18 are two marks questions. Attempt any four questions.
5. Section C is of 08 marks and has 03 competency-based questions.
a) Questions numbers 19 to 21 are four marks questions. Attempt any two questions.
6. Do as per the instructions given in the respective sections.
7. Marks allotted are mentioned against each section/question.

SECTION A (3 + 2 = 5 marks)
Answer any 3 questions out of the given 4 questions. Each question is of the mark. 1x3=3
Write any two qualities for being a good entrepreneur.
Ans:
Any two of the following points-
• They are confident. They believe in themselves and their abilities.
• They keep trying new ideas in their business.
• They are patient.
Q.1 1
• They are creative and think differently about business ideas.
• They take responsibility for their actions.
• They make decisions after thinking about them.
• They work hard.
• They do not give up when they face a difficulty
(½ mark for each point; ½ x 2=1)
What is sustainable development?
Ans:
Sustainable development is the development that satisfies the needs of the present
Q.2 1
without compromising the capacity of future generations, guaranteeing the balance
between economic growth, care for the environment and social well-being.
(1 mark for correct answer/explanation)
Entrepreneurship has a positive impact on society. Write down any two.
Ans:
Q.3 1. Some of them work towards saving the environment. 1
2. Some of them give money to build schools and hospitals.
(½ mark for each point; ½ x 2=1)
How many sustainable development goals were formulated by the United
Nations?
Q.4 Ans: 1
There are 17 sustainable development goals formulated by the United Nations.
(1 mark for correct answer)

417-X-MS-Term II (2021-2022) 1
Answer any 1 question out of the given 2 questions. Each question is of mark. 2x1=2
“Entrepreneurs are born, not made.” Do you agree with this statement? Justify
your answer.
Ans:
No, this is a myth/misconception about entrepreneurship.
Q.5 2
Being an entrepreneur starts with a way of thinking. One must believe that anything
is possible and it shall be achieved. It starts with thinking of an idea that you want to
work on, making it different.
(1 mark for the option(NO);1 mark for correct explanation)
Enlist any 2 SDGs which are formulated to address the problems related to
water?
Ans:
Q.6 Clean water and sanitation 2
Life below water
Responsible consumption, and production
(any 2 SDG s related to water; 1 mark for each SDG)

SECTION B (4 + 8 = 12 marks)

Answer any 04 questions out of the given 06 questions 1x4=4

What will be the output of the word “studies” if we do the following:


a. Lemmatization
b. Stemming
Q.7 Ans: 1
The output of the word after lemmatization will be study.
The output of the word after stemming will be studi.
(½ mark for for lemmatization, ½ mark for stemming)
How many tokens are there in the sentence given below?
Traffic Jams have become a common part of our lives nowadays. Living in an
urban area means you have to face traffic each and every time you get out on the
Q.8 road. Mostly, school students opt for buses to go to school. 1
Ans:
46 tokens are there in the given sentence
(1 mark for correct answer)
What is a corpus?
Ans:
Q.9 The term used to describe the whole textual data from all the documents altogether is 1
known as corpus.
(1 mark for any correct explanation)
Identify any 2 stopwords in the given sentence:
Pollution is the introduction of contaminants into the natural environment
that cause adverse change.The three types of pollution are air pollution, water
Q.10 pollution and land pollution. 1
Ans:
Stopwords in the given sentence are: is, the, of, that, into, are, and
(any two correct answers; ½ mark each)
Why should we avoid using the training data for evaluation?
Ans:
Q.11 This is because our model will simply remember the whole training set, and will 1
therefore always predict the correct label for any point in the training set.
(1 mark for any correct explanation)
What should be the value of F1 score if the model needs to have 100% accuracy?.
Ans:
Q.12 1
The model will have an F1 score of 1 if it has to be 100% accurate.
(1 mark for correct answer)

417-X-MS-Term II (2021-2022) 2
Answer any 04 questions out of the given 06 questions 2x4=8

“Automatic summarization is used in NLP applications”. Is the given statement


correct? Justify your answer with an example.
Ans:
Yes, the given statement is correct. Automatic summarization is relevant not only
for summarizing the meaning of documents and information, but also to understand the
emotional meanings within the information, such as in collecting data from social media.
Q.13 2
Automatic summarization is especially relevant when used to provide an overview of a
news item or blog post, while avoiding redundancy from multiple sources and
maximizing the diversity of content obtained.

(1 mark for explanation, 1 mark for example)

Give an example of a situation wherein false positive would have a high cost
associated with it.
Ans:
Let us consider a model that predicts that a mail is spam or not. If the model always
predicts that the mail is spam, people would not look at it and eventually might lose
Q.14 important information. Here False Positive condition (Predicting the mail as spam while 2
the mail is not spam) would have a high cost.

(2 marks for any correct example with explanation; 1 marks can be given if only
explanation is written without example)

Write any two applications of TFIDF


Ans:
1. Document Classification
Helps in classifying the type and genre of a document.
2. Topic Modelling
It helps in predicting the topic for a corpus.
Q.15 3. Information Retrieval System 2
To extract the important information out of a corpus.
4. Stop word filtering
Helps in removing the unnecessary words out of a text body.

(1 mark for each application name/explanation)

Write down the steps to implement bag of words algorithm.


Ans:
The steps to implement bag of words algorithm are as follows:
1. Text Normalisation: Collect data and pre-process it
2. Create Dictionary: Make a list of all the unique words occurring in the corpus.
(Vocabulary)
Q.16 2
3. Create document vectors: For each document in the corpus, find out how many times
the word from the unique list of words has occurred.
4. Create document vectors for all the documents.

(½ mark for each step)

What is a confusion matrix? What is it used for?


Ans:
The confusion matrix is used to store the results of comparison between the prediction
and reality.From the confusion matrix, we can calculate parameters like recall, precision
Q.17 2
,F1 score which are used to evaluate the performance of an AI model.

(1 mark for definition, 1 mark for use)

417-X-MS-Term II (2021-2022) 3
Explain from the given graph, how the value and occurrence of a word are related
in a corpus?

Q.18 2

Ans:
As shown in the graph, occurrence and value of a word are inversely proportional. The
words which occur most (like stop words) have negligible value. As the occurrence of
words drops, the value of such words rises. These words are termed as rare or valuable
words. These words occur the least but add the most value to the corpus.
(complete explanation 2 marks)
SECTION C (2 x 4 = 8 marks)
(COMPETENCY-BASED QUESTIONS)
Answer any 02 questions out of the given 03 questions

Q.19 Through a step-by-step process, calculate TFIDF for the given corpus 4
Document 1: Johny Johny, Yes Papa,
Document 2: Eating sugar? No Papa
Document 3: Telling lies? No Papa
Document 4: Open your mouth, Ha! Ha! Ha!
Ans:
1. Create document vectors for the given documents (Term Frequency Table)
Johny Yes Papa Eating Sugar No Telling Lies Open your Mouth Ha

2 1 1 0 0 0 0 0 0 0 0 0

0 0 1 1 1 1 0 0 0 0 0 0

0 0 1 0 0 1 1 1 0 0 0 0

0 0 0 0 0 0 0 0 1 1 1 3

2. Record the occurrence of word in the document using term frequency table (Document
Frequency Table)
Johny Yes Papa Eating Sugar No Telling Lies Open your Mouth Ha

1 1 3 1 1 2 1 1 1 1 1 1

3. Draw the inverse document frequency table wherein, we need to put the document
frequency in the denominator while the total number of documents is the numerator.
Here, the total number of documents are 4, hence inverse document frequency becomes:

417-X-MS-Term II (2021-2022) 4
Johny Yes Papa Eating Sugar No Telling Lies Open your Mouth Ha

4/1 4/1 4/3 4/1 4/1 4/2 4/1 4/1 4/1 4/1 4/1 4/1

4. The formula of TFIDF for any word W becomes: TFIDF(W) = TF(W) * log (IDF(W))
Johny Yes Papa Eating Sugar No Telling Lies Open your Mouth Ha

2*log( 1*log( 1*log(4 0*log(4/1 0*log(4/ 0*lo 0*log(4/ 0*log( 0*log(4/ 0*log(4/ 0*log(4/ 0*log
4/1) 4/1) /3) ) 1 g(4/2 1 4/1) 1) 1) 1) (4/1)
)

0*log( 0*log( 1*log(4 1*log(4/1 1*log(4/ 1*lo 0*log(4/ 0*log( 0*log(4/ 0*log(4/ 0*log(4/ 0*log
4/1) 4/1) /3) ) 1) g(4/2 1) 4/1) 1) 1) 1) (4/1)
)

0*log( 0*log( 1*log(4 0*log(4/1 0*log(4/ 1*lo 1*log(4/ 1*log( 0*log(4/ 0*log(4/ 0*log(4/ 0*log
4/1) 4/1) /3) ) 1) g(4/2 1) 4/1) 1) 1) 1) (4/1)
)

0*log( 0*log( 0*log(4 0*log(4/1 0*log(4/ 0*lo 0*log(4/ 0*log( 1*log(4/ 1*log(4/ 1*log(4/ 3*log
4/1) 4/1) /3) ) 1) g(4/2 1) 4/1) 1) 1) 1) (4/1)
)

(1 mark for each correct table)


Q.20 The world is competitive nowadays. People face competition in even the tiniest 4
tasks and are expected to give their best at every point in time. When people are
unable to meet these expectations, they get stressed and could even go into
depression. We get to hear a lot of cases where people are depressed due to
reasons like peer pressure, studies, family issues, relationships, etc. and they
eventually get into something that is bad for them as well as for others. So, to
overcome this, Cognitive Behavioural Therapy (CBT) is considered to be one of
the best methods to address stress as it is easy to implement on people and also
gives good results. This therapy includes understanding the behaviour and
mindset of a person in their normal life. With the help of CBT, therapists help
people overcome their stress and live a happy life.
For the situation given above,
1. Write the problem statement template
2. List any two sources from which data can be collected.
3. How do we explore the data?
Ans
1. The problem statement template for the given scenario would be
Our people undergoing stress Who?

have a they are not being able to share their feelings What?
problem that

while they need help to vent out their emotions Where?

An ideal To provide a platform to share their thoughts Why?


solution would anonymously and suggest help whenever required.
be
2. Data can be collected from one of the following sources:
a. surveys
b. observing therapist’s sessions
c. databases available on the internet
d. interviews
3. Once the textual data has been collected, it needs to be processed and cleaned
so that an easier version can be sent to the machine. Thus, the text is
normalised through various steps and is lowered to minimum vocabulary since
the machine does not require grammatically correct statements but the essence
of it.
(2 marks for problem statement template; ½ marks for each data sources; 1 mark for
correct explanation of data exploration)

417-X-MS-Term II (2021-2022) 5
Q.21 Take a look at the confusion matrix: 4
The Confusion Reality

Matrix Yes No

Yes True False


Positive Positive
Prediction (TP) (FP)
No False True
Negative Negative
(FN) (TN)
How do you calculate F1 score?
Ans:
We begin the calculation by first using the formula to calculate Precision
Precision is defined as the percentage of true positive cases versus all the cases where
the prediction is true. That is, it takes into account the True Positives and False
Positives.
𝑇𝑟𝑢𝑒 𝑃𝑜𝑠𝑖𝑡𝑖𝑣𝑒
Precision = ×100%
𝐴𝑙𝑙 𝑃𝑟𝑒𝑑𝑖𝑐𝑡𝑒𝑑 𝑃𝑜𝑠𝑖𝑡𝑖𝑣𝑒𝑠

𝑇𝑃
Precision = ×100%
𝑇𝑃+𝐹𝑃
Next, we calculate recall as the fraction of positive cases that are correctly identified.
𝑇𝑟𝑢𝑒 𝑃𝑜𝑠𝑖𝑡𝑖𝑣𝑒
Recall =
𝑇𝑟𝑢𝑒 𝑃𝑜𝑠𝑖𝑡𝑖𝑣𝑒+𝐹𝑎𝑙𝑠𝑒 𝑁𝑒𝑔𝑎𝑡𝑖𝑣𝑒

𝑇𝑃
Recall =
𝑇𝑃+𝐹𝑁

Finally, we calculate the F1 score as the measure of balance between precision and
recall.
𝑃𝑟𝑒𝑐𝑖𝑠𝑖𝑜𝑛 ×𝑅𝑒𝑐𝑎𝑙𝑙
F1 score = 2 ×
𝑃𝑟𝑒𝑐𝑖𝑠𝑖𝑜𝑛 +𝑅𝑒𝑐𝑎𝑙𝑙

(1 mark for precision formula; 1 mark for recall formula; 1 mark for F1 score formula;1
mark for explanation)

417-X-MS-Term II (2021-2022) 6
CBSE | DEPARTMENT OF SKILL EDUCATION
ARTIFICIAL INTELLIGENCE (SUBJECT CODE - 417)
Blue-print for Sample Question Paper for Class X (Session 2023-2024)
Max. Time: 2 Hours Max. Marks: 50
PART A - EMPLOYABILITY SKILLS (10 MARKS):
OBJECTIVE TYPE SHORT ANSWER
UNIT QUESTIONS TYPE QUESTIONS TOTAL
NAME OF THE UNIT
NO. QUESTIONS
1 MARK EACH 2 MARKS EACH

1 Communication Skills -II 1 1 2

2 Self-Management Skills - II 2 1 3

3 ICT Skills - II 1 1 2

4 Entrepreneurial Skills - II 1 1 2

5 Green Skills - II 1 1 2

TOTAL QUESTIONS 6 5 11

NO. OF QUESTIONS TO BE ANSWERED Any 4 Any 3 07

TOTAL MARKS 1x4=4 2x3=6 10 MARKS

PART B - SUBJECT SPECIFIC SKILLS (40 MARKS):


SHORT ANSWER DESCRIPTIVE/
OBJECTIVE TYPE
UNIT TYPE LONG ANS. TYPE TOTAL
NAME OF THE UNIT QUESTIONS
NO. QUESTIONS QUESTIONS QUESTIONS
1 MARK EACH 2 MARKS EACH 4 MARKS EACH
1 Introduction to AI 5 1 1 7
2 AI Project Cycle 3 1 2 6
3 Advance Python - - - -
4 Data Sciences 3 1 - 4
5 Computer Vision 3 1 - 4

6 Natural Language 5 1 1 7
Processing
7 Evaluation 5 1 1 7

TOTAL QUESTIONS 24 6 5 35

NO. OF QUESTIONS 20 Any 4 Any 3 27


TO BE ANSWERED

TOTAL MARKS 1 x 20 = 20 2x4=8 4 x 3 = 12 40 MARKS

417 - ARTIFICIAL INTELLIGENCE – CLASS X PAGE 1 OF 8


CBSE | DEPARTMENT OF SKILL EDUCATION
ARTIFICIAL INTELLIGENCE (SUBJECT CODE - 417)

Sample Question Paper for Class X (Session 2023-2024)


Max. Time: 2 Hours Max. Marks: 50

General Instructions:
1. Please read the instructions carefully.
2. This Question Paper consists of 21 questions in two sections: Section A & Section B.
3. Section A has Objective type questions whereas Section B contains Subjective type questions.
4. Out of the given (5 + 16 =) 21 questions, a candidate has to answer (5 + 10 =) 15 questions in the
allotted (maximum) time of 2 hours.
5. All questions of a particular section must be attempted in the correct order.
6. SECTION A - OBJECTIVE TYPE QUESTIONS (24 MARKS):
i. This section has 05 questions.
ii. Marks allotted are mentioned against each question/part.
iii. There is no negative marking.
iv. Do as per the instructions given.
7. SECTION B – SUBJECTIVE TYPE QUESTIONS (26 MARKS):
i. This section has 16 questions.
ii. A candidate has to do 10 questions.
iii. Do as per the instructions given.
iv. Marks allotted are mentioned against each question/part.

417 - ARTIFICIAL INTELLIGENCE – CLASS X PAGE 2 OF 8


SECTION A: OBJECTIVE TYPE QUESTIONS

Q. 1 Answer any 4 out of the given 6 questions on Employability Skills (1 x 4 = 4 marks)

i. “M D Gulati started with a small shop with his focus, dedication and clear ideas, MDH 1
became one of the most popular brands in India besides having a good reputation all
over the world”.
Which self-management skill is clearly visible in the given statement?

ii. When you bring the mouse over a file in File Explorer, it will show the details of that file. 1
This is known as
(a) Drag and drop
(b) Double click
(c) Hover
(d) Single click

iii. Assertion(A): A doctor works for a renowned hospital. 1


Reason(R): The statement given above is an example of wage employment.
(a) Both A and R are correct and R is the correct explanation of A
(b) Both A and R are correct but R is NOT the correct explanation of A
(c) A is correct but R is not correct
(d) A is not correct but R is correct

iv. _______ the work is all about identifying and noting how we spent our time, and 1
analysing how to spend our time effectively.
(a) Organising
(b) Prioritising
(c) Controlling
(d) Tracking

v. Remya traveled to Sweden from India to pursue her higher education. But she doesn't 1
know how to speak Swedish (language of Sweden). Because of this, she was unable to
find a part time job. This is an example of _____________
(a) Interpersonal barrier
(b) Physical barrier
(c) Organisational barrier
(d) Linguistic barrier

vi. “Efforts are made to increase the solar power generation so that our electricity needs are met 1
and at the same time we do not pollute the environment or use up natural resources”.
Which SDG can you relate this statement to?
(a) Life on land
(b) Clean water and sanitation
(c) Affordable and clean energy
(d) Reduced inequalities

417 - ARTIFICIAL INTELLIGENCE – CLASS X PAGE 3 OF 8


Q. 2 Answer any 5 out of the given 6 questions (1 x 5 = 5 marks)

i. Assertion (A)– One can be a good singer while the other can be a great athlete. 1
Reason(R) – Humans possess different types of intelligences but at different levels.
(a) Both A and R are correct and R is the correct explanation of A
(b) Both A and R are correct but R is not the correct explanation of A
(c) A is correct but R is not correct
(d) A is not correct but R is correct.

ii. The Indian Government banned a few apps stating – “servers in the hostile nation are 1
receiving and using the acquired data improperly”.
Which terminology suits best for this action?
(a) AI Ethics
(b) Data Privacy
(c) AI Bias
(d) AI Access

iii. Statment1: There are four layers in a neural network. 1


Statement2:The first layer of the neural network is known as the output layer.
(a) Both Statement1 and Statement2 are correct
(b) Both Statement1 and Statement2 are incorrect
(c) Statement1 is correct but Statement2 is incorrect
(d) Statement2 is correct but Statement1 is incorrect

iv. Observe the given graph and fill in the blank: 1

__________ the neural network, better is the performance.

v. _______ is a simple file format that stores data separated by commas. 1


(a) jpg
(b) doc
(c) csv
(d) png

vi. A corpus contains 4 documents in which the word ‘diet’ was appearing once in 1
document1. Identify the term in which we can categorise the word ‘diet’.
(a) Stop word
(b) Rare word
(c) Frequent word
(d) Removable word

417 - ARTIFICIAL INTELLIGENCE – CLASS X PAGE 4 OF 8


Q. 3 Answer any 5 out of the given 6 questions (1 x 5 = 5 marks)

i. Read the examples given below- 1


i. Using Chat GPT to write an email
ii. Face unlock technology of mobile phones using camera
iii. Turning off lights with IoT device
iv. Hand sanitizer dispenser having sensor
Choose the options that are not AI
(a) i and ii
(b) iii and i
(c) iii and iv
(d) i, iii and iv

ii. _____ refer to the type of features that we want to collect. 1

iii. Which of the following is an application of data science? 1


(a) Text summarization
(b) Target Advertisements
(c) Face lock in smartphones
(d) Email filters

iv. ____ is the process of finding instances of real-world objects in images or videos. 1
(a) Instance segmentation
(b) Object detection
(c) Classification
(d) Image segmentation

v. Identify the given Chat bot type: 1


It learns from its environment and experience. It also builds on its capabilities based on
the knowledge. These can collaborate with humans, working along-side them and
learning from their behavior.

vi. F1 Score is the measure of the balance between 1


(a) Accuracy and Precision
(b) Precision and Recall
(c) Recall and Accuracy
(d) Recall and Reality

Q. 4 Answer any 5 out of the given 6 questions (1 x 5 = 5 marks)

i. Srishti learnt about AI terminologies but was not able to recollect the term that is used 1
to refer to machines that perform tasks with vast amounts of data using neural
networks. Help her with the correct term.

ii. Statment1: The output given by the AI model is known as reality. 1


Statement2:The real scenario is known as Prediction.
(a)Both Statement1 and Statement2 are correct
(b)Both Statement1 and Statement2 are incorrect
(c) Statement1 is correct but Statement2 is incorrect
(d) Statement2 is correct but Statement1 is incorrect

417 - ARTIFICIAL INTELLIGENCE – CLASS X PAGE 5 OF 8


iii. Rajat has made a model which predicts the performance of Indian Cricket players in 1
upcoming matches. He collected the data of players’ performance with respect to
stadium, bowlers, opponent team and health. His model works with good accuracy and
precision value. Which of the statement given below is incorrect?
(a) Data gathered with respect to stadium, bowlers, opponent team and health is
known as Testing Data.
(b) Data given to an AI model to check accuracy and precision is Testing Data.
(c) Training data and testing data are acquired in the Data Acquisition stage.
(d) Training data is always larger as compared to testing data.
iv. How many channels does a colour image have? 1

v. Which feature of NLP helps in understanding the emotions of the people mentioned 1
with the feedback?
(a) Virtual Assistants
(b) Sentiment Analysis
(c) Text classification
(d) Automatic Summarization

vi. Sarthak made a face mask detector system for which he had collected the dataset and 1
used all the dataset to train the model. Then, he used the same data to evaluate the
model which resulted in the correct answer all the time but was not able to perform
with unknown dataset.
Name the concept.

Q. 5 Answer any 5 out of the given 6 questions (1 x 5 = 5 marks)

i. Aditi, a student of class XII developed a chatbot that clarifies the doubts of Economics 1
students. She trained the software with lots of data sets catering to all difficulty levels.
If any student would type or ask questions related to Economics, the software would
give an instant reply. Identify the domain of AI in the given scenario.
(a) Computer Vision
(b) Data Science
(c) Natural Language Processing
(d) None of these

ii. Which evaluation parameter takes into consideration all the correct predictions? 1

iii. ______ means a picture element which is the smallest unit of information that makes 1
up a picture.
(a) Vision
(b) Pics
(c) Pixel
(d) Piskel

iv. What do you mean by syntax of a language? 1


(a) Meaning of a sentence
(b) Grammatical structure of a sentence
(c) Semantics of a sentence
(d) Synonym of a sentence

417 - ARTIFICIAL INTELLIGENCE – CLASS X PAGE 6 OF 8


v. Which algorithms result in two things, a vocabulary of words and frequency of the 1
words in the corpus?
(a) Sentence segmentation
(b) Tokenisation
(c) Bag of words
(d) Text normalisation

vi. Which one of the following scenario result in a high false positive cost? 1
(a) viral outbreak
(b) forest fire
(c) flood
(d) spam filter

SECTION B: SUBJECTIVE TYPE QUESTIONS


Answer any 3 out of the given 5 questions on Employability Skills (2 x 3 = 6 marks)
Answer each question in 20 – 30 words.
Q. 6 List two best practices for effective communication. 2

Q. 7 What is the importance of setting goals in life? 2

Q. 8 “The Trojan Horse was a wooden horse said to have been used by the Greeks during the 2
Trojan War to enter the city of Troy and win the war”.
What does Trojan horse mean in computer terminology?

Q. 9 How is society helping entrepreneurs in their business? 2

Q. 10 Mention any two human activities that lead to environmental degradation. 2

Answer any 4 out of the given 6 questions in 20 – 30 words each (2 x 4 = 8 marks)


Q. 11 All of us use smartphones. When we install a new app, it asks us for several permissions 2
to access our phone’s data in different ways. Why do apps collect such data?

Q. 12 Sirisha and Divisha want to make a model which will organize the unlabeled input data 2
into groups based on features. Which learning model should they use and why?

Q. 13 Ajay wants to access data from various sources. Suggest him any two points that he 2
needs to keep in mind while accessing data from any data source.

Q. 14 Explain the term resolution with an example. 2

Q. 15 Identify any two stop words which should not be removed from the given sentence and 2
why?
Get help and support whether you're shopping now or need help with a past purchase.
Contact us at [email protected] or on our website www.pwershel.com

Q. 16 Draw the confusion matrix for the following data 2


• the number of true positive = 100
417 - ARTIFICIAL INTELLIGENCE – CLASS X PAGE 7 OF 8
• the number of true negative 47
• the number of false positive = 62
• the number of false negative = 290

Answer any 3 out of the given 5 questions in 50– 80 words each (4 x 3 = 12 marks)
Q. 17 Your grandmother watches you use AI applications. She wants to understand more 4
about it. Help her understand the term artificial intelligence by giving the right definition
and explain to her with an example how machines become artificially intelligent.

Q. 18 Akhil wants to learn how to scope the problem for an AI Project. Explain him the 4
following:
(a) 4W Problem Canvas
(b) Problem Statement Template

Q. 19 Identify and explain the types of the learning-based approaches in the figures given 4
below.

Figure 1

Figure 2

Q. 20 We, human beings, can read, write and understand many languages. But computers can 4
understand only machine language. Do you think we might face any challenges if we try
to teach computers how to understand and interact in human languages? Explain.

Q. 21 An AI model made the following sales prediction for a new mobile phone which they 4
have recently launched:
Reality
Confusion Matrix Yes No

Yes 50 40
Prediction
No 12 10
(i) Identify the total number of wrong predictions made by the model.
(ii) Calculate precision, recall and F1 Score.

417 - ARTIFICIAL INTELLIGENCE – CLASS X PAGE 8 OF 8


CBSE | DEPARTMENT OF SKILL EDUCATION
ARTIFICIAL INTELLIGENCE (SUBJECT CODE - 417)
MARKING SCHEME FOR CLASS X (SESSION 2023-2024)
Max. Time: 2 Hours Max. Marks: 50
General Instructions:
1. Please read the instructions carefully.
2. This Question Paper consists of 21 questions in two sections – Section A & Section B.
3. Section A has Objective type questions whereas Section B contains Subjective type questions.
4. Out of the given (5 + 16 =) 21 questions, a candidate has to answer (5 + 10 =) 15 questions in the
allotted (maximum) time of 2 hours.
5. All questions of a particular section must be attempted in the correct order.
6. SECTION A – OBJECTIVE TYPE QUESTIONS (24 MARKS):
i. This section has 05 questions.
ii. There is no negative marking.
iii. Do as per the instructions given.
iv. Marks allotted are mentioned against each question/part.
7. SECTION B – SUBJECTIVE TYPE QUESTIONS (26 MARKS):
i. This section contains 16 questions.
ii. A candidate has to do 10 questions.
iii. Do as per the instructions given.
iv. Marks allotted are mentioned against each question/part.

SECTION A: OBJECTIVE TYPE QUESTIONS


Source Material
Page no.
Q. (NCERT/PSSCIVE/ Unit/
QUESTION of source Marks
No. CBSE Study Chap. No.
material
Material)
Q. 1 Answer any 4 out of the given 6 questions on Employability Skills (1 x 4 = 4 marks)
i. Self-Motivation NCERT 2 52 1

ii. (c) Hover NCERT 3 67 1

(a) Both A and R are correct and R is the


iii. NCERT 4 100 1
correct explanation of A
iv. (d) Tracking NCERT 2 60 1

v. (d) Linguistic barrier NCERT 1 21 1

vi. (c) Affordable and clean energy NCERT 5 111 1


Q. 2 Answer any 5 out of the given 6 questions (1 x 5 = 5 marks)
(a) Both A and R are correct and R is the Facilitator
i. 1 11 1
correct explanation of A Handbook
Facilitator
ii. (b) Data Privacy 1 24 1
Handbook

417 - ARTIFICIAL INTELLIGENCE – CLASS X PAGE 1 OF 8


(a) Both Statement1 and Statement2 are Facilitator
iii. 2 40 1
incorrect Handbook
Facilitator
iv. Larger 2 40 1
Handbook
Facilitator
v. (c) csv 4 63 1
Handbook
Facilitator
vi. (b) Rare word 6 115 1
Handbook
Q. 3 Answer any 5 out of the given 6 questions (1 x 5 = 5 marks)
Facilitator
i. (c) iii and iv 1 16 1
Handbook
Facilitator
ii. Data Features/Data 2 35 1
Handbook
Facilitator
iii. (b) Target Advertisements 4 56 1
Handbook
Facilitator
iv. (b) Object Detection 5 78 1
Handbook
Facilitator
v. Smart Bot 6 105 1
Handbook
Facilitator
vi. (b) Precision and Recall 7 127 1
Handbook
Q. 4 Answer any 5 out of the given 6 questions (1 x 5 = 5 marks)
Facilitator
i. Deep Learning 1 21 1
Handbook
(b) Both Statement1 and Statement2 are Facilitator
ii. 7 119 1
incorrect Handbook
(a) Data gathered with respect to
Facilitator
iii. stadium, bowlers, opponent team and 4 60 1
Handbook
health is known as Testing Data.
Facilitator
iv. Three Channels 5 83 1
Handbook
Facilitator
v. (b) Sentiment Analysis 6 100 1
Handbook
Facilitator
vi. Overfitting 7 119 1
Handbook
Q. 5 Answer any 5 out of the given 6 questions (1 x 5 = 5 marks)
Facilitator
i. (c) Natural Language Processing 1 22 1
Handbook
Facilitator
ii. Accuracy 7 123 1
Handbook
Facilitator
iii. (c) Pixel 5 79 1
Handbook
Facilitator
iv. (b) Grammatical structure of a sentence 6 106 1
Handbook
Facilitator
v. (c) Bag of words 6 112 1
Handbook
Facilitator
vi. (d) spam filter 7 126 1
Handbook

417 - ARTIFICIAL INTELLIGENCE – CLASS X PAGE 2 OF 8


SECTION B: SUBJECTIVE TYPE QUESTIONS
Source Unit/ Page no. Marks
Material Chap. No. of source
Q.
QUESTION (NCERT/PSSCIV material
No.
E/ CBSE Study
Material)
Answer any 3 out of the given 5 questions on Employability Skills in 20 – 30 words each (2 x 3 = 6 marks)
Q. 6 1. Use simple language NCERT 1 22 2
2. Be respectful of others’ opinions
3. Do not form assumptions on culture, religion
or geography
4. Try to communicate in person as much as
possible
5. Use visuals
6. Take help of a translator to overcome
differences in language
(2 marks for any two correct points from the
above)
Q. 7 1. The process of goal setting in your life helps NCERT 2 55 2
you decide on how to live your life, where
you want to be, and how you want to be in
the future.
2. It helps you to focus on the end result
instead of less important work.
3. This will make you successful in your career
and personal life.
(2 marks for any one valid justification given
above or any relevant answer)

Q. 8 A Trojan Horse is a type of malware which NCERT 3 81 2


disguises itself i.e., it appears to be a useful
software program but once it reaches a
computer it starts behaving like a virus and
destroys data.
(1 mark for acting like useful program and, 1
mark for the words
destroying/corrupting/deleting data)
or
(only ½ marks will be allotted if only
malware/harmful program/virus/ term is
mentioned)

Q. 9 Society is helping entrepreneurs by NCERT 4 85 2


1. Creating needs
2. Providing raw material
3. buying/selling of items
4. making money
(2 marks for any two valid points)

417 - ARTIFICIAL INTELLIGENCE – CLASS X PAGE 3 OF 8


Q. 10 1. Use of fertilisers, pesticides,etc for NCERT 5 103 2
increasing the production of crops.
2. Cutting down of forests for personal use
such as construction of buildings, factories
etc

(1 mark for each valid point, or any two relevant


answers other than the above)

Answer any 4 out of the given 6 questions in 20 – 30 words each (2 x 4 = 8 marks)


Q. 11 1. To provide customized notifications and Facilitator Unit 1 25, 26 2
recommendations. Handbook
2. To improve the efficiency and accuracy of the
app.

(2 marks for any one correct point with


explanation)

Q. 12 Clustering model/Unsupervised learning is used Facilitator Unit 2 39 2


to organize the unlabeled input data into groups Handbook
based on features.
Clustering is an unsupervised learning algorithm
which can cluster unknown data according to
the patterns or trends identified out of it. The
patterns observed might be the ones which are
known to the developer or it might even come
up with some unique patterns out of it.
(1 mark for identifying the name of the
algorithm and 1 mark for explanation)

Q. 13 While accessing data from any of the data Facilitator Unit 4 63 2


sources, following points should be kept in Handbook
mind:
1. Data which is available for public usage only
should be taken up.
2. Personal datasets should only be used with
the consent of the owner.
3. One should never breach someone’s privacy
to collect data.
4. Data should only be taken from reliable
sources as the data collected from random
sources can be wrong or unusable.
5. Reliable sources of data ensure the
authenticity of data which helps in the
proper training of the AI model.
6. Data should be relevant to the problem.

(any two; 1 mark for each valid point)

417 - ARTIFICIAL INTELLIGENCE – CLASS X PAGE 4 OF 8


Q. 14 Resolution of an image refers to the number of Facilitator Unit 5 80 2
pixels in an image, across the width and height. Handbook
For example a monitor resolution of 1280×1024.
This means there are 1280 pixels from one side
to the other, and 1024 from top to bottom.

(1 mark for explanation; 1 mark for example)


Q. 15 Stopwords in the given sentence which should Facilitator Unit 6 110 2
not be removed are: Handbook
@, . (fullstop) ,_(underscore) , 123(numbers)
These tokens are generally considered as
stopwords, but in the above sentence, these
tokens are part of email id. removing these
tokens may lead to invalid website address and
email ID. So these words should not be removed
from the above sentence.

(1 mark for identifying any two stop words from


the above, and 1 mark for the valid
justification.)
Q. 16 Facilitator Unit 7 122 2
Handbook
Confusion Reality
Matrix
Yes No

Prediction Yes 100 62

No 290 47
(½ marks each for mapping the values in the
correct section, ½ *4=2 marks)

Answer any 3 out of the given 5 questions in 50– 80 words each (4 x 3 = 12 marks)
Q. 17 When a machine possesses the ability to mimic Facilitator Unit 1 14, 15 4
human traits, i.e., make decisions, predict the Handbook
future, learn and improve on its own, it is said
to have artificial intelligence. In other words,
you can say that a machine is artificially
intelligent when it can accomplish tasks by
itself - collect data, understand it, analyse it,
learn from it, and improve it.

Machines become intelligent once they are


trained with some data which helps them
achieve their tasks. AI machines also keep
updating their knowledge to optimise their
output. For example, Netflix gives us
recommendations on the basis of what we like.
Whenever we start liking a new genre, it
updates and gives better suggestions.

417 - ARTIFICIAL INTELLIGENCE – CLASS X PAGE 5 OF 8


(2 marks for definition of Artificial intelligence
which includes any of the highlighted terms,
2 mark for an example explanation of how
machines become intelligent)
or
(only 1 mark for any AI machine example which
mimic human traits without explanation)

Q. 18 The 4Ws Problem canvas helps in identifying the Facilitator Unit 2 31-34 4
key elements related to the problem. The 4Ws Handbook
are Who, What, Where and Why
● The “Who” block helps in analysing the
people getting affected directly or
indirectly due to the problem.
● The “What” block helps us to determine
the nature of the problem.
● The “Where” block helps us to look into
the situation in which the problem
arises, the context of it, and the
locations where it is prominent.
● The “Why” block suggests to us the
benefits which the stakeholders would
get from the solution and how it will
benefit them as well as the society

Problem Statement Template


Our [stakeholders] Who

Have a problem [need] What


that

When/while [context/ Where


location/
situation]

An ideal solution [solution] Why


would be

(½ mark each for explanation of 4w s; 2 marks


for drawing the problem statement template
with correct words in it or explaining the
problem statement template)
or
(1 mark to be allotted if only 4Ws are written
without explanation)

Q. 19 The learning-based approaches shown in the Facilitator Unit 2 37, 38 4


given figures are Supervised learning and Handbook
Unsupervised learning.

417 - ARTIFICIAL INTELLIGENCE – CLASS X PAGE 6 OF 8


Figure 1: In a supervised learning model, the
dataset which is fed to the machine is labelled.
In other words, we can say that the dataset is
known to the person who is training the
machine only then he/she is able to label the
data. A label is some information which can be
used as a tag for data.
Here, labelled images of dog and cat are fed into
the model and trained. The model correctly
identifies the given input as dog.

Figure 2: An unsupervised learning model works


on unlabelled dataset. This means that the data
which is fed to the machine is random and there
is a possibility that the person who is training
the model does not have any information
regarding it. The unsupervised learning models
are used to identify relationships, patterns and
trends out of the data which is fed into it. It
helps the user in understanding what the data is
about and what are the major features
identified by the machine in it.
Here, images of a set of animals are fed into the
AI model and the model clusters them based on
similar features

(1 mark each for identifying each term


supervised learning and unsupervised learning;
1 mark per explanation of each term)

Q. 20 Yes, we might face any challenges if we try to Facilitator Unit 6 106, 107 4
teach computers how to understand and Handbook
interact in human languages.

The possible difficulties are:


1. Arrangement of the words and meaning -
the computer has to identify the different
parts of a speech. Also, it may be extremely
difficult for a computer to understand the
meaning behind the language we use.

2. Multiple Meanings of a word - same word


can be used in a number of different ways
which according to the context of the
statement changes its meaning completely.

3. Perfect Syntax, no Meaning - Sometimes, a


statement can have a perfectly correct
syntax but it does not mean anything. For
example, take a look at this statement:

417 - ARTIFICIAL INTELLIGENCE – CLASS X PAGE 7 OF 8


Chickens feed extravagantly while the moon
drinks tea.
This statement is correct grammatically but
does this make any sense? In Human
language, a perfect balance of syntax and
semantics is important for better
understanding.

(1 mark for Yes and 1 mark each for the points


on possible difficulties)

Q. 21 Ans: (i)the total number of wrong predictions Facilitator Unit 7 126,127 4


made by the model is the sum of false positive Handbook
and false negative. FP+FN=40+12= 52

(ii) Precision=TP/(TP+FP)

=50/(50+40)

50/90

=0.55

Recall=TP/(TP+FN)

=50/(50+12)

=50/62

=.81

F1 Score = 2*Precision*Recall/(Precision+Recall)

=2*0.55*.81/(.55+.81)

=.891/1.36

=0.65

(1 marks for part (i) and ½ mark for each


formula and ½ mark each for substitution of
values in part(ii))

Please note: the mathematical calculations can


be ignored

417 - ARTIFICIAL INTELLIGENCE – CLASS X PAGE 8 OF 8


CBSE | DEPARTMENT OF SKILL EDUCATION
ARTIFICIAL INTELLIGENCE (SUBJECT CODE - 417)
BLUE PRINT FOR SAMPLE QUESTION PAPER - 1 (PRACTICE PAPER-1)
CLASS X (SESSION 2024-2025)
Max. Time: 2 Hours Max. Marks: 50
PART A - EMPLOYABILITY SKILLS (10 MARKS):
OBJECTIVE TYPE SHORT ANSWER
UNIT QUESTIONS TYPE QUESTIONS TOTAL
NAME OF THE UNIT
NO. QUESTIONS
1 MARK EACH 2 MARKS EACH

1 Communication Skills -II 1 1 2

2 Self-Management Skills - II 2 1 3

3 ICT Skills - II 1 1 2

4 Entrepreneurial Skills - II 1 1 2

5 Green Skills - II 1 1 2

TOTAL QUESTIONS 6 5 11

NO. OF QUESTIONS TO BE ANSWERED Any 4 Any 3 07

TOTAL MARKS 1x4=4 2x3=6 10 MARKS

PART B - SUBJECT SPECIFIC SKILLS (40 MARKS):


SHORT ANSWER DESCRIPTIVE/
OBJECTIVE TYPE
UNIT TYPE LONG ANS. TYPE TOTAL
NAME OF THE UNIT QUESTIONS
NO. QUESTIONS QUESTIONS QUESTIONS
1 MARK EACH 2 MARKS EACH 4 MARKS EACH
1 Introduction to AI 5 1 1 7
2 AI Project Cycle 3 1 2 6
3 Advance Python - - - -
4 Data Sciences 3 1 - 4
5 Computer Vision 3 1 - 4
6 Natural Language 5 1 1 7
Processing
7 Evaluation 5 1 1 7
TOTAL QUESTIONS 24 6 5 35
NO. OF QUESTIONS 20 Any 4 Any 3 27
TO BE ANSWERED
TOTAL MARKS 1 x 20 = 20 2x4=8 4 x 3 = 12 40 MARKS

417 - ARTIFICIAL INTELLIGENCE – CLASS X PAGE 1 OF 8


CBSE | DEPARTMENT OF SKILL EDUCATION
ARTIFICIAL INTELLIGENCE (SUBJECT CODE - 417)
SAMPLE QUESTION PAPER - 1 (PRACTICE PAPER-1)
CLASS X (SESSION 2024-2025)
Max. Time: 2 Hours Max. Marks: 50

General Instructions:
1. Please read the instructions carefully.
2. This Question Paper consists of 21 questions in two sections: Section A & Section B.
3. Section A has Objective type questions whereas Section B contains Subjective type questions.
4. Out of the given (5 + 16 =) 21 questions, a candidate has to answer (5 + 10 =) 15 questions in the
allotted (maximum) time of 2 hours.
5. All questions of a particular section must be attempted in the correct order.
6. SECTION A - OBJECTIVE TYPE QUESTIONS (24 MARKS):
i. This section has 05 questions.
ii. Marks allotted are mentioned against each question/part.
iii. There is no negative marking.
iv. Do as per the instructions given.
7. SECTION B – SUBJECTIVE TYPE QUESTIONS (26 MARKS):
i. This section has 16 questions.
ii. A candidate has to do 10 questions.
iii. Do as per the instructions given.
iv. Marks allotted are mentioned against each question/part.

417 - ARTIFICIAL INTELLIGENCE – CLASS X PAGE 2 OF 8


SECTION A: OBJECTIVE TYPE QUESTIONS

Q. 1 Answer any 4 out of the given 6 questions on Employability Skills (1 x 4 = 4 marks)

i. “M D Gulati started with a small shop with his focus, dedication and clear ideas, MDH 1
became one of the most popular brands in India besides having a good reputation all
over the world”.
Which self-management skill is clearly visible in the given statement?
ii. When you bring the mouse over a file in File Explorer, it will show the details of that file. 1
This is known as
(a) Drag and drop
(b) Double click
(c) Hover
(d) Single click

iii. Assertion(A): A doctor works for a renowned hospital. 1


Reason(R): The statement given above is an example of wage employment.
(a) Both A and R are correct and R is the correct explanation of A
(b) Both A and R are correct but R is NOT the correct explanation of A
(c) A is correct but R is not correct
(d) A is not correct but R is correct

iv. the work is all about identifying and noting how we spent our time, and 1
analysing how to spend our time effectively.
(a) Organising
(b) Prioritising
(c) Controlling
(d) Tracking

v. Remya traveled to Sweden from India to pursue her higher education. But she doesn't 1
know how to speak Swedish (language of Sweden). Because of this, she was unable to
find a part time job. This is an example of
(a) Interpersonal barrier
(b) Physical barrier
(c) Organisational barrier
(d) Linguistic barrier

vi. “Efforts are made to increase the solar power generation so that our electricity needs are met 1
and at the same time we do not pollute the environment or use up natural resources”.
Which SDG can you relate this statement to?
(a) Life on land
(b) Clean water and sanitation
(c) Affordable and clean energy
(d) Reduced inequalities

417 - ARTIFICIAL INTELLIGENCE – CLASS X PAGE 3 OF 8


Q. 2 Answer any 5 out of the given 6 questions (1 x 5 = 5 marks)

i. Assertion (A)– One can be a good singer while the other can be a great athlete. 1
Reason(R) – Humans possess different types of intelligences but at different levels.
(a) Both A and R are correct and R is the correct explanation of A
(b) Both A and R are correct but R is not the correct explanation of A
(c) A is correct but R is not correct
(d) A is not correct but R is correct.

ii. The Indian Government banned a few apps stating – “servers in the hostile nation are 1
receiving and using the acquired data improperly”.
Which terminology suits best for this action?
(a) AI Ethics
(b) Data Privacy
(c) AI Bias
(d) AI Access

iii. Statment1: There are four layers in a neural network. 1


Statement2:The first layer of the neural network is known as the output layer.
(a) Both Statement1 and Statement2 are correct
(b) Both Statement1 and Statement2 are incorrect
(c) Statement1 is correct but Statement2 is incorrect
(d) Statement2 is correct but Statement1 is incorrect

iv. Observe the given graph and fill in the blank: 1

the neural network, better is the performance.

v. is a simple file format that stores data separated by commas. 1


(a) jpg
(b) doc
(c) csv
(d) png

vi. A corpus contains 4 documents in which the word ‘diet’ was appearing once in 1
document1. Identify the term in which we can categorise the word ‘diet’.
(a) Stop word
(b) Rare word
(c) Frequent word
(d) Removable word

417 - ARTIFICIAL INTELLIGENCE – CLASS X PAGE 4 OF 8


Q. 3 Answer any 5 out of the given 6 questions (1 x 5 = 5 marks)

i. Read the examples given below- 1


i. Using Chat GPT to write an email
ii. Face unlock technology of mobile phones using camera
iii. Turning off lights with IoT device
iv. Hand sanitizer dispenser having sensor
Choose the options that are not AI
(a) i and ii
(b) iii and i
(c) iii and iv
(d) i, iii and iv

ii. refer to the type of features that we want to collect. 1

iii. Which of the following is an application of data science? 1


(a) Text summarization
(b) Target Advertisements
(c) Face lock in smartphones
(d) Email filters

iv. is the process of finding instances of real-world objects in images or videos. 1


(a) Instance segmentation
(b) Object detection
(c) Classification
(d) Image segmentation

v. Identify the given Chat bot type: 1


It learns from its environment and experience. It also builds on its capabilities based on
the knowledge. These can collaborate with humans, working along-side them and
learning from their behavior.

vi. F1 Score is the measure of the balance between 1


(a) Accuracy and Precision
(b) Precision and Recall
(c) Recall and Accuracy
(d) Recall and Reality

Q. 4 Answer any 5 out of the given 6 questions (1 x 5 = 5 marks)

i. Srishti learnt about AI terminologies but was not able to recollect the term that is used 1
to refer to machines that perform tasks with vast amounts of data using neural
networks. Help her with the correct term.

ii. Statment1: The output given by the AI model is known as reality. 1


Statement2:The real scenario is known as Prediction.
(a) Both Statement1 and Statement2 are correct
(b) Both Statement1 and Statement2 are incorrect
(c) Statement1 is correct but Statement2 is incorrect
(d) Statement2 is correct but Statement1 is incorrect

417 - ARTIFICIAL INTELLIGENCE – CLASS X PAGE 5 OF 8


iii. Rajat has made a model which predicts the performance of Indian Cricket players in 1
upcoming matches. He collected the data of players’ performance with respect to
stadium, bowlers, opponent team and health. His model works with good accuracy and
precision value. Which of the statement given below is incorrect?
(a) Data gathered with respect to stadium, bowlers, opponent team and health is
known as Testing Data.
(b) Data given to an AI model to check accuracy and precision is Testing Data.
(c) Training data and testing data are acquired in the Data Acquisition stage.
(d) Training data is always larger as compared to testing data.
iv. How many channels does a colour image have? 1

v. Which feature of NLP helps in understanding the emotions of the people mentioned 1
with the feedback?
(a) Virtual Assistants
(b) Sentiment Analysis
(c) Text classification
(d) Automatic Summarization

vi. Sarthak made a face mask detector system for which he had collected the dataset and 1
used all the dataset to train the model. Then, he used the same data to evaluate the
model which resulted in the correct answer all the time but was not able to perform
with unknown dataset.
Name the concept.

Q. 5 Answer any 5 out of the given 6 questions (1 x 5 = 5 marks)

i. Aditi, a student of class XII developed a chatbot that clarifies the doubts of Economics 1
students. She trained the software with lots of data sets catering to all difficulty levels.
If any student would type or ask questions related to Economics, the software would
give an instant reply. Identify the domain of AI in the given scenario.
(a) Computer Vision
(b) Data Science
(c) Natural Language Processing
(d) None of these

ii. Which evaluation parameter takes into consideration all the correct predictions? 1

iii. means a picture element which is the smallest unit of information that makes 1
up a picture.
(a) Vision
(b) Pics
(c) Pixel
(d) Piskel

iv. What do you mean by syntax of a language? 1


(a) Meaning of a sentence
(b) Grammatical structure of a sentence
(c) Semantics of a sentence
(d) Synonym of a sentence

417 - ARTIFICIAL INTELLIGENCE – CLASS X PAGE 6 OF 8


v. Which algorithms result in two things, a vocabulary of words and frequency of the 1
words in the corpus?
(a) Sentence segmentation
(b) Tokenisation
(c) Bag of words
(d) Text normalisation

vi. Which one of the following scenario result in a high false positive cost? 1
(a) viral outbreak
(b) forest fire
(c) flood
(d) spam filter

SECTION B: SUBJECTIVE TYPE QUESTIONS


Answer any 3 out of the given 5 questions on Employability Skills (2 x 3 = 6 marks)
Answer each question in 20 – 30 words.
Q. 6 List two best practices for effective communication. 2

Q. 7 What is the importance of setting goals in life? 2

Q. 8 “The Trojan Horse was a wooden horse said to have been used by the Greeks during the 2
Trojan War to enter the city of Troy and win the war”.
What does Trojan horse mean in computer terminology?
Q. 9 How is society helping entrepreneurs in their business? 2

Q. 10 Mention any two human activities that lead to environmental degradation. 2

Answer any 4 out of the given 6 questions in 20 – 30 words each (2 x 4 = 8 marks)


Q. 11 All of us use smartphones. When we install a new app, it asks us for several permissions 2
to access our phone’s data in different ways. Why do apps collect such data?

Q. 12 Sirisha and Divisha want to make a model which will organize the unlabeled input data 2
into groups based on features. Which learning model should they use and why?

Q. 13 Ajay wants to access data from various sources. Suggest him any two points that he 2
needs to keep in mind while accessing data from any data source.

Q. 14 Explain the term resolution with an example. 2

Q. 15 Identify any two stop words which should not be removed from the given sentence and 2
why?
Get help and support whether you're shopping now or need help with a past purchase.
Contact us at [email protected] or on our website www.pwershel.com

Q. 16 Draw the confusion matrix for the following data 2


• the number of true positive = 100
417 - ARTIFICIAL INTELLIGENCE – CLASS X PAGE 7 OF 8
• the number of true negative 47
• the number of false positive = 62
• the number of false negative = 290

Answer any 3 out of the given 5 questions in 50– 80 words each (4 x 3 = 12 marks)
Q. 17 Your grandmother watches you use AI applications. She wants to understand more 4
about it. Help her understand the term artificial intelligence by giving the right definition
and explain to her with an example how machines become artificially intelligent.

Q. 18 Akhil wants to learn how to scope the problem for an AI Project. Explain him the 4
following:
(a) 4W Problem Canvas
(b) Problem Statement Template

Q. 19 Identify and explain the types of the learning-based approaches in the figures given 4
below.

Figure 1

Figure 2

Q. 20 We, human beings, can read, write and understand many languages. But computers can 4
understand only machine language. Do you think we might face any challenges if we try
to teach computers how to understand and interact in human languages? Explain.

Q. 21 An AI model made the following sales prediction for a new mobile phone which they 4
have recently launched:
Reality
Confusion Matrix
Yes No

Yes 50 40
Prediction
No 12 10
(i) Identify the total number of wrong predictions made by the model.
(ii) Calculate precision, recall and F1 Score.

417 - ARTIFICIAL INTELLIGENCE – CLASS X PAGE 8 OF 8


CBSE | DEPARTMENT OF SKILL EDUCATION
ARTIFICIAL INTELLIGENCE (SUBJECT CODE - 417)
MARKING SCHEME FOR CLASS X (SESSION 2024-2025)
Max. Time: 2 Hours Max. Marks: 50
General Instructions:
1. Please read the instructions carefully.
2. This Question Paper consists of 21 questions in two sections – Section A & Section B.
3. Section A has Objective type questions whereas Section B contains Subjective type questions.
4. Out of the given (5 + 16 =) 21 questions, a candidate has to answer (5 + 10 =) 15 questions in the
allotted (maximum) time of 2 hours.
5. All questions of a particular section must be attempted in the correct order.
6. SECTION A – OBJECTIVE TYPE QUESTIONS (24 MARKS):
i. This section has 05 questions.
ii. There is no negative marking.
iii. Do as per the instructions given.
iv. Marks allotted are mentioned against each question/part.
7. SECTION B – SUBJECTIVE TYPE QUESTIONS (26 MARKS):
i. This section contains 16 questions.
ii. A candidate has to do 10 questions.
iii. Do as per the instructions given.
iv. Marks allotted are mentioned against each question/part.

SECTION A: OBJECTIVE TYPE QUESTIONS


Source Material
Page no.
Q. (NCERT/PSSCIVE/ Unit/
QUESTION of source Marks
No. CBSE Study Chap. No.
material
Material)
Q. 1 Answer any 4 out of the given 6 questions on Employability Skills (1 x 4 = 4 marks)

i. Self-Motivation NCERT 2 52 1

ii. (c) Hover NCERT 3 67 1

(a) Both A and R are correct and R is the


iii. NCERT 4 100 1
correct explanation of A
iv. (d) Tracking NCERT 2 60 1

v. (d) Linguistic barrier NCERT 1 21 1

vi. (c) Affordable and clean energy NCERT 5 111 1


Q. 2 Answer any 5 out of the given 6 questions (1 x 5 = 5 marks)
(a) Both A and R are correct and R is the Facilitator
i. 1 11 1
correct explanation of A Handbook
Facilitator
ii. (b) Data Privacy 1 24 1
Handbook

417 - ARTIFICIAL INTELLIGENCE – X PAGE 1 OF 8


(a) Both Statement1 and Statement2 are Facilitator
iii. 2 40 1
incorrect Handbook
Facilitator
iv. Larger 2 40 1
Handbook
Facilitator
v. (c) csv 4 63 1
Handbook
Facilitator
vi. (b) Rare word 6 115 1
Handbook
Q. 3 Answer any 5 out of the given 6 questions (1 x 5 = 5 marks)
Facilitator
i. (c) iii and iv 1 16 1
Handbook
Facilitator
ii. Data Features/Data 2 35 1
Handbook
Facilitator
iii. (b) Target Advertisements 4 56 1
Handbook
Facilitator
iv. (b) Object Detection 5 78 1
Handbook
Facilitator
v. Smart Bot 6 105 1
Handbook
Facilitator
vi. (b) Precision and Recall 7 127 1
Handbook
Q. 4 Answer any 5 out of the given 6 questions (1 x 5 = 5 marks)
Facilitator
i. Deep Learning 1 21 1
Handbook
(b) Both Statement1 and Statement2 are Facilitator
ii. 7 119 1
incorrect Handbook
(a) Data gathered with respect to
Facilitator
iii. stadium, bowlers, opponent team and 4 60 1
Handbook
health is known as Testing Data.
Facilitator
iv. Three Channels 5 83 1
Handbook
Facilitator
v. (b) Sentiment Analysis 6 100 1
Handbook
Facilitator
vi. Overfitting 7 119 1
Handbook
Q. 5 Answer any 5 out of the given 6 questions (1 x 5 = 5 marks)
Facilitator
i. (c) Natural Language Processing 1 22 1
Handbook
Facilitator
ii. Accuracy 7 123 1
Handbook
Facilitator
iii. (c) Pixel 5 79 1
Handbook
Facilitator
iv. (b) Grammatical structure of a sentence 6 106 1
Handbook
Facilitator
v. (c) Bag of words 6 112 1
Handbook
(d) spam filter Facilitator
vi. 7 126 1
Handbook

417 - ARTIFICIAL INTELLIGENCE – X PAGE 2 OF 8


SECTION B: SUBJECTIVE TYPE QUESTIONS
Source Unit/ Page no. Marks
Material Chap. No. of source
Q.
QUESTION (NCERT/PSSCIV material
No.
E/ CBSE Study
Material)
Answer any 3 out of the given 5 questions on Employability Skills in 20 – 30 words each (2 x 3 = 6 marks)
Q. 6 1. Use simple language NCERT 1 22 2
2. Be respectful of others’ opinions
3. Do not form assumptions on culture, religion
or geography
4. Try to communicate in person as much as
possible
5. Use visuals
6. Take help of a translator to overcome
differences in language
(2 marks for any two correct points from the
above)
Q. 7 1. The process of goal setting in your life helps NCERT 2 55 2
you decide on how to live your life, where
you want to be, and how you want to be in
the future.
2. It helps you to focus on the end result
instead of less important work.
3. This will make you successful in your career
and personal life.
(2 marks for any one valid justification given
above or any relevant answer)

Q. 8 A Trojan Horse is a type of malware which NCERT 3 81 2


disguises itself i.e., it appears to be a useful
software program but once it reaches a
computer it starts behaving like a virus and
destroys data.
(1 mark for acting like useful program and, 1
mark for the words
destroying/corrupting/deleting data)
or
(only ½ marks will be allotted if only
malware/harmful program/virus/ term is
mentioned)

Q. 9 Society is helping entrepreneurs by NCERT 4 85 2


1. Creating needs
2. Providing raw material
3. buying/selling of items
4. making money
(2 marks for any two valid points)

417 - ARTIFICIAL INTELLIGENCE – X PAGE 3 OF 8


Q. 10 1. Use of fertilisers, pesticides,etc for NCERT 5 103 2
increasing the production of crops.
2. Cutting down of forests for personal use
such as construction of buildings, factories
etc

(1 mark for each valid point, or any two relevant


answers other than the above)

Answer any 4 out of the given 6 questions in 20 – 30 words each (2 x 4 = 8 marks)


Q. 11 1. To provide customized notifications and Facilitator Unit 1 25, 26 2
recommendations. Handbook
2. To improve the efficiency and accuracy of the
app.

(2 marks for any one correct point with


explanation)

Q. 12 Clustering model/Unsupervised learning is used Facilitator Unit 2 39 2


to organize the unlabeled input data into groups Handbook
based on features.
Clustering is an unsupervised learning algorithm
which can cluster unknown data according to
the patterns or trends identified out of it. The
patterns observed might be the ones which are
known to the developer or it might even come
up with some unique patterns out of it.
(1 mark for identifying the name of the
algorithm and 1 mark for explanation)

Q. 13 While accessing data from any of the data Facilitator Unit 4 63 2


sources, following points should be kept in Handbook
mind:
1. Data which is available for public usage only
should be taken up.
2. Personal datasets should only be used with
the consent of the owner.
3. One should never breach someone’s privacy
to collect data.
4. Data should only be taken from reliable
sources as the data collected from random
sources can be wrong or unusable.
5. Reliable sources of data ensure the
authenticity of data which helps in the
proper training of the AI model.
6. Data should be relevant to the problem.

(any two; 1 mark for each valid point)

417 - ARTIFICIAL INTELLIGENCE – X PAGE 4 OF 8


Q. 14 Resolution of an image refers to the number of Facilitator Unit 5 80 2
pixels in an image, across the width and height. Handbook
For example a monitor resolution of 1280×1024.
This means there are 1280 pixels from one side
to the other, and 1024 from top to bottom.

(1 mark for explanation; 1 mark for example)


Q. 15 Stopwords in the given sentence which should Facilitator Unit 6 110 2
not be removed are: Handbook
@, . (fullstop) ,_(underscore) , 123(numbers)
These tokens are generally considered as
stopwords, but in the above sentence, these
tokens are part of email id. removing these
tokens may lead to invalid website address and
email ID. So these words should not be removed
from the above sentence.

(1 mark for identifying any two stop words from


the above, and 1 mark for the valid
justification.)
Q. 16 Facilitator Unit 7 122 2
Handbook
Confusion Reality
Matrix
Yes No

Prediction Yes 100 62

No 290 47
(½ marks each for mapping the values in the
correct section, ½ *4=2 marks)

Answer any 3 out of the given 5 questions in 50– 80 words each (4 x 3 = 12 marks)
Q. 17 When a machine possesses the ability to mimic Facilitator Unit 1 14, 15 4
human traits, i.e., make decisions, predict the Handbook
future, learn and improve on its own, it is said
to have artificial intelligence. In other words,
you can say that a machine is artificially
intelligent when it can accomplish tasks by
itself - collect data, understand it, analyse it,
learn from it, and improve it.

Machines become intelligent once they are


trained with some data which helps them
achieve their tasks. AI machines also keep
updating their knowledge to optimise their
output. For example, Netflix gives us
recommendations on the basis of what we like.
Whenever we start liking a new genre, it
updates and gives better suggestions.

417 - ARTIFICIAL INTELLIGENCE – X PAGE 5 OF 8


(2 marks for definition of Artificial intelligence
which includes any of the highlighted terms,
2 mark for an example explanation of how
machines become intelligent)
or
(only 1 mark for any AI machine example which
mimic human traits without explanation)

Q. 18 The 4Ws Problem canvas helps in identifying the Facilitator Unit 2 31-34 4
key elements related to the problem. The 4Ws Handbook
are Who, What, Where and Why
● The “Who” block helps in analysing the
people getting affected directly or
indirectly due to the problem.
● The “What” block helps us to determine
the nature of the problem.
● The “Where” block helps us to look into
the situation in which the problem
arises, the context of it, and the
locations where it is prominent.
● The “Why” block suggests to us the
benefits which the stakeholders would
get from the solution and how it will
benefit them as well as the society

Problem Statement Template


Our [stakeholders] Who

Have a problem [need] What


that

When/while [context/ Where


location/
situation]

An ideal solution [solution] Why


would be

(½ mark each for explanation of 4w s; 2 marks


for drawing the problem statement template
with correct words in it or explaining the
problem statement template)
or
(1 mark to be allotted if only 4Ws are written
without explanation)

Q. 19 The learning-based approaches shown in the Facilitator Unit 2 37, 38 4


given figures are Supervised learning and Handbook
Unsupervised learning.

417 - ARTIFICIAL INTELLIGENCE – X PAGE 6 OF 8


Figure 1: In a supervised learning model, the
dataset which is fed to the machine is labelled.
In other words, we can say that the dataset is
known to the person who is training the
machine only then he/she is able to label the
data. A label is some information which can be
used as a tag for data.
Here, labelled images of dog and cat are fed into
the model and trained. The model correctly
identifies the given input as dog.

Figure 2: An unsupervised learning model works


on unlabelled dataset. This means that the data
which is fed to the machine is random and there
is a possibility that the person who is training
the model does not have any information
regarding it. The unsupervised learning models
are used to identify relationships, patterns and
trends out of the data which is fed into it. It
helps the user in understanding what the data is
about and what are the major features
identified by the machine in it.
Here, images of a set of animals are fed into the
AI model and the model clusters them based on
similar features

(1 mark each for identifying each term


supervised learning and unsupervised learning;
1 mark per explanation of each term)

Q. 20 Yes, we might face any challenges if we try to Facilitator Unit 6 106, 107 4
teach computers how to understand and Handbook
interact in human languages.

The possible difficulties are:


1. Arrangement of the words and meaning -
the computer has to identify the different
parts of a speech. Also, it may be extremely
difficult for a computer to understand the
meaning behind the language we use.

2. Multiple Meanings of a word - same word


can be used in a number of different ways
which according to the context of the
statement changes its meaning completely.

3. Perfect Syntax, no Meaning - Sometimes, a


statement can have a perfectly correct
syntax but it does not mean anything. For
example, take a look at this statement:

417 - ARTIFICIAL INTELLIGENCE – X PAGE 7 OF 8


Chickens feed extravagantly while the moon
drinks tea.
This statement is correct grammatically but
does this make any sense? In Human
language, a perfect balance of syntax and
semantics is important for better
understanding.

(1 mark for Yes and 1 mark each for the points


on possible difficulties)

Q. 21 Ans: (i)the total number of wrong predictions Facilitator Unit 7 126,127 4


made by the model is the sum of false positive Handbook
and false negative. FP+FN=40+12= 52

(ii) Precision=TP/(TP+FP)

=50/(50+40)

50/90

=0.55

Recall=TP/(TP+FN)

=50/(50+12)

=50/62

=.81

F1 Score = 2*Precision*Recall/(Precision+Recall)

=2*0.55*.81/(.55+.81)

=.891/1.36

=0.65

(1 marks for part (i) and ½ mark for each


formula and ½ mark each for substitution of
values in part(ii))

Please note: the mathematical calculations can


be ignored

417 - ARTIFICIAL INTELLIGENCE – X PAGE 8 OF 8

You might also like