AWS Major Project.docx (1)
AWS Major Project.docx (1)
ABSTRACT
EXPECTED OUTCOMES vi
ABSTRACT ix
1.3.1 Chatbot 2
1.11 SUMMARY 11
2.1 PREAMBLE 12
1.5.1 Processor 28
1.5.2 Ram 29
1.8 DATASET 31
1.9 SUMMARY 32
4.10 PREAMBLE 33
4.2.1 Dataset 34
4.4 SUMMARY 35
5.1 PREAMBLE 36
5.3.2 Classification 38
5.4.1c Response 41
5.4.3 Tokenization 41
5.5 SUMMARY 46
CHAPTER 6: EXPERIMENTAL RESULTS 46-25
6.1 PREAMBLE 47
6.2 RESULTS 47
6.4 APPLICATIONS 54
6.5 SUMMARY 55
6.1 PREAMBLE 47
6.2 RESULTS 47
6.4 APPLICATIONS 54
6.5 SUMMARY 54
7.1 CONCLUSION 55
REFERENCES 56-58
B1-B4
5.1 Dataset 66
6.11 s
.
i
t
m C
a
6o
p
rt
of
all
cl
as
se
s
LIST OF TABLES
PARAMETERS
LIST OF ABBRIVATIONS
NB - Naïve Bayes
RL - Reinforcement Learning
CHAPTER 1
INTRODUCTION
1.1 PREAMBLE
Customer query solving is became a challenging task when comes to support team.
Especially when the main challenge comes to customer support team is to solve the
customer queries in a fast and efficient manner. Even though there is a high man power to
solve the queries a person can solve only a single query when comes to queuing technique
when two or more users post the query at the same time there is chance of time latency and
sometimes the query cannot be solved. Here to address this problem there is a need of a
technology medium which can solve user queries in a fast and efficient manner.
Conversation interface with the help of machine learning models and a dataset should be
considered in which a set of question and answers should be entered such that when a user
enter a query it involves a set of phases which includes elimination of unnecessary
symbols, special characters which is called text preprocessing.
Query resolving became a challenging task when comes to support team. Fast query
solving in time-efficient is the most challenging task. Even though technologies have been
evolved there is huge latency when comes to customer service. Because a person can only
solve single query at a time but when two or more users enter a query at the same time
there is not a chance of solving the two or more queries at a time.Rasa interface is used to
built a chatbot service but this is time-consuming and cost- effective.
1.3.1 Chatbot
The Chatbots, or informal interfaces as they're additionally notable, gift a brand
new approach for people to act with pc systems. historically, to urge a matter answered
by a software system program concerned employing a computer programme, or filling
out a type. A chatbot permits a user to easily raise queries within the same manner that
they'd address a person.
Most industrial chatbots square measure keen about platforms created by the
technology giants for his or her tongue process. These embody Amazon Lex, Microsoft
psychological feature Services, Google Cloud tongue API, Facebook DeepText, and
IBM Watson. Platforms wherever chatbots square measure deployed embody Facebook
traveler, Skype, and Slack, among several others.
`
There are various types of chatbots in which it varies for various industries and fields.
1.3.2a Menu/button-based chatbots
If you'll be able to predict the kinds of queries your customers might raise, a linguistic
chatbot could be the answer for you. Linguistic or rules-based chatbots produce
informal automation flows exploitation if/then logic. First, you have got to outline the
language conditions of your chatbots. Conditions may be created to assess the words,
the order of the words, synonyms, and more. If the incoming question matches the
conditions outlined by your chatbot, your customers will receive the acceptable
facilitate in no time. However, it’s your job to make sure that every permutation and
combination of every question is outlined, otherwise, the chatbot won't perceive your
customer’s input. this is often why a linguistic model, whereas implausibly common,
may be slow to
develop. These chatbots demand rigidity and specificity.
Unlike menu-based chatbots, keyword recognition-based chatbots will hear what users
kind and respond fitly. These chatbots utilize customizable keywords Associate in
Nursingd Associate in Nursing AI application - linguistic communication process
(NLP) to see a way to serve an acceptable response to the user.
These styles of chatbots disappoint once they have to be compelled to answer heaps of
comparable queries. The IP chatbots can begin to slide once there are keyword
redundancies between many connected queries.
It is quite standard to ascertain chatbot examples that ar a hybrid of keyword
recognition- based and menu/button-based. These chatbots offer users with the
selection to undertake to raise their queries directly or use the chatbot’s menu buttons if
the keyword recognition practicality is yielding poor results or the user needs some
steerage to search out their answers.
Fig 1.4: Illustration of a Keyword-based chatbot[4]
For example, a discourse chatbot that permits users to order food; the chatbot can store
the information from every oral communication and learn what the user likes to order.
The result's that eventually once a user chats with this chatbot, it'll bear in mind their
commonest order, their delivery address, and their payment data and simply raise if
they’d wish to repeat this order. rather than having to reply to many queries the user
simply needs to answer with ‘Yes’ and also the food is ready!
While this food ordering example is elementary, it's straightforward to check simply
however powerful oral communication context may be once controlled with AI and
cubic centimeter. the final word goal of ANy chatbot ought to be to supply an
improved user expertise over the choice of the establishment. investment oral
communication context is one amongst the simplest ways that to shorten processes like
these via these chatbots.
Fig 1.5 : A picture of a machine learning chatbot[5]
To make informal interfaces even a lot of vernacular, businesses area unit currently
setting out to use voice-based chatbots or voice bots. Voice bots are on the increase
for the last few years, with virtual assistants like Apple’s Siri, to Amazon’s Alexa,
and why? due to the convenience they carry. It’s abundant easier for a client to talk
instead of sort. A voice-activated chatbot brings resistance experiences on to the top
client.
Can hold multiple conversations at once: Chatbots will converse at the same time
with thousands of consumers. This will increase business productivity and eliminates
wait
times.
Cost-effective: A chatbot may be a quicker and cheaper one-time investment than
making a zealous, cross-platform app or hiring extra staff. additionally, chatbots will
cut back pricey issues caused by human error. User acquisition prices conjointly
decrease with a chatbot's ability to retort inside seconds.
Saves time: Chatbots will alter tasks performed oft and at specific times. this offers
staff time to specialise in additional necessary tasks and prevents customers from
waiting to receive responses.
Proactive customer interaction: Within the past, organizations relied on passive
client interaction and waited for consumers to succeed in out 1st. With chatbots,
organizations will move proactively, as bots will initiate conversations and monitor
however customers use the websites and landing pages. Organizations will then use the
knowledge gathered from watching to supply specific incentives to consumers,
facilitate users navigate the location and answer future queries.
Monitors and analyzes consumer data: Chatbots collect feedback from every
interaction to assist businesses improve their services and product or optimize their
websites. Bots may record user information to trace behaviors and buying patterns.
This data offers organizations insight into a way to higher market their product and
services, yet as common obstacles that customers face throughout the shopping for
method.
Improves customer engagement: Most firms already engage their customers through
social media. Chatbots will build this engagement additional interactive. consumers
seldom see the individuals inside businesses, thus chatbots open a channel wherever
customers will have interaction while not the strain of interacting with another person.
Eases quantifiability to international markets: Chatbots will solve
client considerations and queries in multiple languages. Their 24/7 access permits
customers to use them notwithstanding time or geographical zone.
Expands the customer base: Chatbots will improve lead generation, qualification and
nurturing. Chatbots will raise queries throughout the buyer's journey and supply data
which will persuade the user and build a lead. Chatbots will then give potential client
data to the sales team, United Nations agency will have interaction with the leads. The
bots will improve conversion rates and make sure the lead's journey flows within the
right direction -- toward an acquisition.
Measures lead qualifications: Chatbots will facilitate sales groups confirm a lead's
qualifications mistreatment known key performance indicators, like budget, timeline
and
resources. this will forestall firms from holdup on unqualified leads and long customers.
1.3.4 Origin of chatbots
ELIZA: It's thought-about to be the primary chatbot within the history of computing
that was developed by Joseph Weizenbaum at Massachusetts Institute of Technology
(MIT). it had been in 1994 that the term ‘Chatterbot” was coined. ELIZA operates by
recognizing key words or phrases from the input to breed a response victimization
those keywords from pre-programmed responses. for example, if a person's says that
‘My mother cooks sensible food’. ELIZA would obtain the word ‘mother’, associated
respond by asking an open- over question ‘Tell Maine additional regarding your
family’. This created associate illusion of understanding associated having an
interaction with a true soul although the method was a mechanized one.
ALICE: It had been developed in 1995 by Richard Wallace. in contrast to Eliza, the
ALICE chatbot was able to use linguistic communication process, that allowed for
additional refined speech communication. it had been revolutionary, though, for being
ASCII text file. Developers may use AIML (artificial intelligence markup language) to
make their own chatbots battery-powered by ALICE.
Research in AI has given rise to varied colloquial interfaces in machine learning and
linguistic communication process. Over the previous couple of years, there as been
associate exponential growth of tools to style, mock, build, deploy, manage, and
legalize chatbots.
One of the major challenge in the field of customer support especially in query
resolving is one of the important challenge. To solve this challenge a service chatbot
has been implemented to solve user queries in fast and efficient manner.But a chatbot
only solves a query at a time and every chatbot differs in the operation and
performance. To improve this our goal is to implement a universal chatbot which can
solve queries in fast and efficient manner such that the service can solve many queries
for different industry.
⮚ To implement a universal chatbot which can solve user queries of various domains.
⮚ To evaluate the performance of the proposed model by response time of the
given user query.
⮚ To compare the performance of the proposed model with existing methods in the
current context.
1.7 PROPOSED METHODOLOGY
A chat bot is an artificial intelligence (AI) software that can simulate a natural
language conversation (or chat) with a user via messaging applications, websites and
mobile apps, or by telephone . A chat bot is a computer application that makes use of
auditory or textual techniques to conduct a communication in an herbal language,
apprehend the user’s purpose, and provide a response based totally on the
organization’s business policies and knowledge. The main character that lies in back
of chat bots is NLP and Machine Learning. When a question is presented to a chat
bot, a chain of complicated algorithms method the acquired input, recognize what the
person is asking, and based totally on that determines the answer suitable to the
query. Chat bot have to rely upon the capacity of the algorithms to come across the
complexity of each textual content and spoken words. Some chat bots perform
thoroughly to the human. However, handling complicated conversation is a big
challenge; where there may be a utilization of diversity figures of speech, it may be
difficult for machines to understand From all the existing methods every chatbot is
designed for only a particular type of service. But any chatbot main aim is to deliver
the response delivery to the user.But a chatbot can provide only particular set of
services when the implementation differs from application to application. To solve
this challenge and to deliver the response to the customers regardless with the
application a universal chatbot can be implemented to ensure any query can be
resolved in time and efficient manner.
⮚ Chatbot helps to resolve the queries faster and efficient in which a chatbot can
help more efficiently.
● This project deals only with voice based resolving is only applicable.
● This project is mainly deals with devices in which some situations which
is cost effective.
1.11 SUMMARY
LITERATURE SURVEY
2.1 PREAMBLE
The following section summarizes the history of those works which are done
previously, highlighting the strengths and weaknesses of each method. Various chatbot
implementation methods were proposed and reviewed here.
● Human supervised learning
● Artificial neural networks
● Cognitive behavior therepy
● Sentiment analysis
● Natural language processing
● Deep learning
● Reinforcement learning(RL)
● Cognitive informatics
● Intent classification
● Deep learning
● Neural machine translation
● Natural language generation with attention mechanism
In this chapter, several methods have been proposed for chatbot implementation.
Among the most recently published works are those presented as follows:
Supervised, unattended, and reinforcement learning area unit the 3 main classes
of machine learning. associate degree example of supervised learning is human-guided
machine learning, that creates a model employing a assortment of coaching knowledge
that has been labeled by humans. In supervised learning, the algorithmic rule picks up
a group of inputs and also the correct outputs that keep company with them. though the
believability of the coaching knowledge wont to generate a machine learning model is
probable to be oppose, the model should still be valid for accuracy before it are often
deployed. Active learning and semi-supervised learning area unit different subsets of
supervised learning wherever the machine learning model is improved with every new
correction or piece of knowledge gathered. Humans will play a job during this.
Subject matter specialists speed up the training method by instructing the
technology in period through a method referred to as human-guided machine learning.
as an example, an individual are often asked to weigh in and supply comments if the
machine learning model encounters any knowledge that it's unsure regarding. The
model then incorporates what it's learned from this input to supply a future prediction
that's a lot of correct. Humans area unit solely concerned once it's necessary for
coaching or validation. Human-guided machine learning builds data from the bottom
up by 1st using algorithms to try to to the labour-intensive method of discovering
relationships inside the info. this suggests that, inevitably, as machine learning
accuracy rises, thus can the quantity of your time somebody's has to pay corporal
punishment a given work.
Giovanni Almeida Santos et.al [2] proposed the Chatbot Management method,
technique for content management on chatbot systems. The planned methodology
relies on the experiences nonheritable with the event of Evatalk, the chatbot for the
Brazilian Virtual faculty of presidency. the main focus of this technique is to evolve the
chatbot content through the analysis of user interactions, permitting a cyclic and
human-supervised method. we tend to divided the planned methodology into 3 distinct
phases, namely, manage, build, and analyze. Moreover, the planned methodology
presents a transparent de_nition of the roles of the chatbot team. we tend to validate the
planned methodology beside the creation of the Evatalk chatbot, whose quantity of
interactions was of twenty- two,771 for the one,698,957 listed attendees within the
Brazillian Virtual faculty of presidency in 2020. the appliance of the methodology on
Evatalk's chatbot brought positive results: we tend to reduced the chatbot's human
hand-off rate from forty four.43% to 30.16%, the chatbot's content examples
accumulated by a hundred and sixtieth while maintaining a high proportion of
con_dence in its responses and keeping the user satisfaction collected in conversations
stable.
forms the basis of cognitive informatics.A set of foundational theories that serve as
the basis for information and knowledge-based science and engineering disciplines
can be developed from this by cognitive informatics.
A modular cognitive agent architecture for instructional question-answering that
includes social dialogue (small talk) and is enhanced for a particular knowledge topic
was proposed by D. Carlander-Reuterfelt et al [14]. As a personal assistant, this
system has been set up to help students understand data science and machine learning
skills. A human-like interface is used to train machine learning models and natural
language understanding algorithms as part of its implementation.
In the field of AI, ML, and chatbots, intent classification refers to the process
of categorising the customer's intent using language analysis. For example, a
consumer who asks, "How can I locate my order status?" in a chat window is
probably seeking, well, order status. The machine recognises this purpose and directs
the user to a human or automated system that can answer their question. Classifying
consumer intent correctly up front results in a lot speedier and more frictionless
experience for the customer, hence intent classification is important (as well as any
agents involved). Poorly formulated customer inquiries do not waste the time of the
agent, and more significantly, they do not waste the time of the consumer. Effective
intent classification goes beyond merely determining the client's requirements.
In conversational systems, the initial stage is typically intent classification. It is the
technique of assigning queries to a known class and is designed to make user query
comprehension easier. Different extraction procedures may be raised by query
classes. The approach taken to find the solution to a question like "Who is Alan
Turin?" is probably different from the approach taken to find the solution to a
question like "Who invented the Turing machine?" The first question asks for a
description, more specifically a definition of "Alan Turing," whereas the second
anticipates the name of a specific person, "Alan Turing."
In order to create an interactive user interface, Addi Ait-Mlouk et al. [10] first design
and develop an architecture. Second, we provide a machine learning strategy based
on intent classification and NLU to comprehend user intents and produce SPARQL
queries. We specifically process a new social network dataset (i.e., myPersonality)
and include it into the current knowledge bases to increase the chatbot's capacity for
processing analytical inquiries. The system is flexible, allows for the intuitive
formulation and execution of various activities for a wide range of topics, has several
knowledge bases, and can be expanded with a new domain if needed. Additionally,
evaluation and application cases for the chatbot are offered to demonstrate how it
enables interactive semantic data for a variety of real-world application scenarios and
highlights the suggested method for a knowledge.
2.10 DEEP LEARNING
Text translation from one language to another is the main goal of the
computational linguistics discipline known as machine translation (MT). Neural
Machine Translation (NMT) has emerged as the most potent algorithm to carry out
this task because to the strength of deep learning. Tech businesses all across the
world are
investing heavily in NMT, even though Google Translate is the main industry
example. This cutting-edge approach uses deep learning to develop a model that can
translate between any two languages by using enormous datasets of translated
sentences. Numerous NMT variants are currently being researched and used in the
industry as a result of the extensive research conducted in recent years. one of the
more traditional and established forms of NMT.
Numerous studies have sought to train neural networks on symbolic processing in
order to use it for inference since the advent of deep learning. Additionally, research
comparable to the present study have been performed to learn propositional logic,
first- order predicate logic, and to do unification. However, because they couldn't be
put into practise, these research were only able to offer method suggestions. Studies
to train symbolic processing with graph networks and feedforward networks have
been carried out and reported with the advent of deep learning. When using graph
networks for symbolic learning, it was necessary to assume that forms of formulae
were already present in the data and to give the networks preliminary versions of
atomic formulas or formulas. With regard to symbolic learning.
A system similar to Prolog is employed for symbolic processing. When a Prolog
processing system receives a query, it makes use of knowledge base facts and rules to
deduce an answer. A question has one or more objectives. A Prolog processing
system uses backward reasoning to infer goals.The actions that a Prolog processing
system takes to deduce responses are briefly described.
04 Addiait- A Knowledge Graph Based Machine learning Chatbot for law There should be
mlouk ChatBot for Natural algorithms and firm helps you a scope to add
et.al [10] Language Understanding SPARQL capture accurate more text-based
Over Linked Data of case details. data sources
Analyze answers with privacy
to improve client preservation and
service with law integrating with
helpdesk bot. social based
apps.
05 Mary A chat messaging service Conversational The use of study does not
Bates [11] that helps doctors and other interfaces with chatbots in the attempt to
health professionals obtain AI&ML algorithms field of medical evaluate an
information about the safety with NLP. analysis helps to exhaustive list of
of drugs. diagnose existing health
according to chatbots
different users.
06 Vorada A design to act as a Natural language Improve Chatbot
Socatiyan precedent for other works at processing with engagement with should be able
urak et.al home and abroad to addition to LAW chatbot for law to solve many
[12] perpetuate awareness of codes related to act. firms queries of the
sexual violence. users by more
interaction.
07 Daniel Advances in cognitive Training of machine Advantages of Need an
Carlander computing enable a new learning models and cognitive improvement
- natural language assistants in of human-
way of interaction that interactions in
Reuterfelt processing involves education and the
accelerates insight from chatbots.
[14] for building human- corresponding
existing information interaction based challenges in
sources, thereby chatbots. implementation.
contributing to the process
of learning.
08 Hiroshi To learn symbolic Symbolic processing Able to learn Symbolic
Honda processing with deep performed by the symbolic processing to
[16] learning and to build prolog processing processing with analyse large-
question answering systems system with neural deep learning and scale data
by means of learned models. machine translation. to build question present on the
answering Web and
systems by inductive
means of learned inference
models. with deep
learning-base
d symbolic
processing
09 Kapil Chatbot must provide Windows Azure Widespread of Time efficient
Tajane services to the customers in Compute , Windows use of chatbots chatbots should
et.al [24] many aspects for delivering Azure Storage can increase the be made to
the data. ,Fabric Controller , interactions. decrease time
Microsoft SQL complexity.
Azure, Windows
Azure Platform
AppFabric , Content
Delivery Network
,Connect
10 Achmad The chatbot application Selenium feature The chatbot will Raspberry Pi
Ramaditiy system is crucial, especially with python be able to read all that is unable to
a et.al for quickly and effectively programming. the messages at access
[26] conveying information to a once and reply to WhatsApp
large number of consumers. the messages. should have its
This the WhatsApp operating system
application for the purposes replaced with
of this Chatbot technology. either Windows
10 or Linux.
2. 16 SUMMARY
SYSTEM REQUIREMENTS
3.1 PREAMBLE
The previous chapter describes the literature review related to different types
of methods and algorithms which helps to build the chatbot which have been proposed
in the earlier paper, comparision of the models in existing papers and feature
comparision to give an brief insight in chapter 2. This chapter describes the different
kinds of tools used and different requirements to develop the proposed system.
Software and hardware requirements vary based on the development model. This
works provides details about tools used like front-end tools, back-end tools, hardware
tools, and different system requirements like hardware and software requirements of
the system.
⮚ Python is Interactive - You can take a seat down at a Python activate and
have interaction with the interpreter immediately to jot down your programs.
3.5.1 Processor
⮚ Step1: Click on Google Chrome and on the address block type Google Colab.
⮚ Step 2: Then click on new notebook on file menu and give the
file name to the notebook.
● The FAQ and a normal dataset contains more queries and responses in which
it should be able to respond to the particular queries.
● All the responses and questions are related to many domains and it is not
permitted to one service of a domain so as to make a universal solver.
3.9 SUMMARY
SYSTEM DESIGN
4.1 PREAMBLE
The previous chapter describes the different requirements and tools used for
developing the proposed system. It also consists of the hardware requirements like
processor, memory, and software requirements like python and frontend and backend
tools. This chapter deals with the design of the proposed system. It includes thesystem
architecture, block diagram, and various diagrams to represent the design of the
system.
The system architecture is the conceptual model that defines the structure,
behavior, and views of a system. An architecture description is a formal description
and representation of a system, organized in a way that supports reasoning about the
structures and behaviors of the system.
The below diagram represents the system architecture for the phases of flow of
the query response to the user. The proposed system consists of two basic steps:
training and validation of the model. The procedure is as follows: In the first phase
training of the knowledge base is involved. After the data has been stored in the
knowledge base the second step in the training phase involves data analysis in which
it checks which kind of data is present in the KB and checks if the data requires any
necessary cleaning or organizations. After the data analysis phase the third step
involves data preprocessing technique in which natural language processing pipeline
plays a keyrole in which it cleans the data by removing unnecessary
symbols,punctions etc.,. After cleaning of the dataset it moves to the sentence
transformer which is a machine learning technique in which it converts the data into
vector or machine understandable form.When the data is converted in embedded form
the embedded data will be stored in the embedded data storage.It can be explained in
the figure 4.1 given below.
When the first phase has been completed the old dataset will be converted into
a new dataset which is fully cleaned and organised.
Figure 4.1: Block Diagram of the Model
A system architecture can consist of a system and the sub-systems developed,
that will work together to implement the overall system.
Above figure explains the training phase.The second phase of the architecture
is testing of the system. The data which is stored in the embedded storage should be
fetched. When the data is fetched it is forwaded to the similarity engine in which in
this phase cosine similarity technique is considered in which it compared the given
user query and stored embedded query and stored into the lookup table.These steps
are performed before the user have entered the query.When the user enters the query it
first enters into the data analysis and next to preprocessing phase and next to sentence
transformer in which after this phase it will be moved to similarity engine phase and
remaining phases will be performed. Compared with the query given by the user the
related response will be given to the user by lookup table as in fig 4.2.
4.2.1 Dataset
A new dataset should be made with frequently asked questions in which the
dataset queries which is related to different domains.The datasets should contains
more number of queries inorder to solve more number of queries.
4.2.2 Data Pre-processing
The FAQ dataset which contains the data should not contain unnecessary
punctuations, unnecessary symbols, numerical digits in which it forms more fuzzy in
organizing the data. For data pre-processing essential python libraries like linguistic
libraries should be imported to remove the unnecessary punctuations and symbols.
The design of the system is explained in chapter 4. This chapter provides information about the
system design. Here, the system architecture diagram of the proposed system has been
discussed. This chapter deals with the design of the proposed system.It includes the system
architecture, to represent the design of the system. This chapter contains main architecture,
sub-systems of the architecture which have been discussed above in the proposed system.
CHAPTER 5
IMPLEMENTATION
5.1 PREAMBLE
In the previous chapter, the proposed system design for developing a chatbot
model has been discussed. Once when the model design and requirements are known,
the next step is implementation. The implementation is the first step in the phase of
development. This chapter briefly discusses the step-by-step process of
implementation for the development of the proposed system.
To solve this problem and to improve the model accuracy LSTM algorithm in deep
learning can be introduced which helps to learn long term dependencies and helps in
pattern recognition. After training the model and testing of the model, the model
computes model accuracy, overall accuracy and all the confusion matrix metrics, f1_score
as evaluation parameters. The dataset consists of queries along with necessary responses
in the form of Json file.
The Proposed system for pattern recognition helps to resolve the queries faster accurately
and the evaluated results have been found.
The proposed method has 7 main modules which are dataset collection, data
visualization, data pre-processing, tokenization, LSTM model preparation, model
validation and computing evaluation parameters. First the dataset is collected in the
form of raw data which are queries along with corresponding responses. The second
module is all about data visualization and all the modules are as follows. The modules
are given below:
1. Dataset collection
2. Data Visualization
3. Data Pre-processing
4. Tokenization
5. Implementation of LSTM model
6. Working of LSTM model
7. Computing evaluation parameters
The dataset considered here is in the form of Json file and it consists of
different types of queries and responses. The Json file consists of Non-Preprocessed
data. There is only a single Json file and the data should be pre-processed.
In this study, the training and testing have been done by the model on the
dataset which includes and corresponding responses. The testing of the model
includes training of the model with the Json file.
The following are explained below with brief summary starting from the
dataset pre-processing, to the machine learning algorithm
The dataset consists of queries with responses in which the input data is in the
form of raw data or unprocessed data. In the dataset, the data consists of unnecessary
words, symbols and punctuations. These punctuations and unnecessary symbols
creates unnecessary space in the memory. The increase in memory space can create
huge time complexity. To reduce this and to make the data clear and precise data
preprocessing should be done on the dataset.
The feature extraction techniques are mathematical models based on various text
processing properties. The different types of features include process of converting
raw data into numerical features that can be processed while preserving the
information in the original dataset. This type of feature extraction can be used by
applying different machine learning techniques on raw data.
The proposed work has four main modules which are dataset collection,
dataset pre-processing, Feature Extraction, and applying Machine learning algorithms
for classification. First, the dataset will be loaded. The dataset collected will then be
pre- processed and the pre-processed data will be followed to training of the model
with the dataset. Later, Machine learning algorithms are used for the classification of
input data. Lastly, the results from the various machine learning algorithms are
evaluated. The modules are given below:
Data visualization is the process of displaying the input dataset. There are 3
different classes for a single data. They are:
6. Input Query
7. Tag name
8.Response
Input Query is nothing but a question form which helps in pattern recognition.
Depending on the input query the response is generated. The input query given here in
the json file will be the intents for training of the model.
Tag name refers to the value which identifies which query or input data refers
to which tag value. The tag name is said to be the key to identify the output values. A
tag is simply an indicator. If users are tagged as customers they are filter customers
otherwise they are not related to customers. Tag are single name entities that can be
attached to the conversation.
Response is defined as the value in which it depends on the input query. Depending
on the input the output response is given. If the model initiates a conversation it is
said to be proactive message.
5.4.3 Tokenization
A common LSTM unit is composed of a cell, an input gate, an output gate and
a forget gate. The cell is essential to remember values over an arbitrary intervals of
time and the three gates are used to regulate the information into and out of the cell.
Forgets gates decide what information to discard from a previous state by assigning a
previous state, compared to a current input in which it values are between 0 and 1.
A rounded value of 1 means to keep the information in the cell and the value 0
means to discard it. Inputs gates are essential to decide which set of information to
store in the current state, using the same system as forget states. Output gates are used
to control which type of information in the current state to output by assigning a value
from 0 to 1 to the information, considering the previous and current states. Selectively
outputting the relevant information from the current state allows the LSTM network
to maintain useful information and long-term dependencies which help to make
predictions both in current and future time-steps.
The below flow chart shows the working process of a LSTM algorithm with
five types of layers which are continuously repeated for the input.
The main idea is to implement a unique ideal network which can store long-
term dependencies. The main difference between RNN and LSTM is which can store
information in the memory for long periods of time. In this LSTM has an advantage
over RNN such that LSTM can handle the information in memory for longer periods
of time as compare to RNN. RNNs consists of feedback loops in the recurrent layer.
This type of network helps to maintain information in the memory for over time. But
it is very difficult to train standard RNNs to solve problems which require learning if
long- term dependencies. This is because the gradient of the loss function decays
exponentially with time known as varnishing gradient problem.
Sigmoid
Hidden layer
Hidden layer is simply defined as the layer between input layers and output
layers, where artificial neurons take in a set of weighted inputs and produce an
output through an activation function. The size of the hidden layer would be near to
512 units for each LSTM cell.
Activation Layer
5.5 SUMMARY
7.1 CONCLUSION
55
CHAPTER 6
EXPERIMENTAL RESULTS
6.1 PREAMBLE
The previous chapter discussed the implementation of the project and the
different modules while implementing the project. In this chapter, we discuss the
experimental results of the developed model. And the experimental results of a few
existing systems are being compared to know the efficiency of the proposed system
6.2 RESULTS
To generate and give the responses to the Queries using Assistive Technology.
The model A query Resolving Chat Bot Using Assistive Technology Approach using
LSTM algorithm was proposed this model makes use of FAQ dataset in the form of
Json file. The Json file contains intents and the intents contains three files. a) tag ,b)
input , c)responses and every tag represents the key value There is only one folder i.e.,
training json file. The validation of the model will be performed by the validate given
by the user.
Query Resolving using the LSTM model. the performance of the model of the
LSTM Design was optimized based on the performance of previously produced
architecture. This study resulted in a new model .By using LSTM architecture, the
performance of the model is increasing and produces the results efficiently.
The model summary is shown below, which has the description of the layersof
the model and their output shapes, parameters, and also which layer it is going to
connect. The layers of the model are namely the Input layer, LSTM layer, Flatten
layer, Dense layer, and these layers are fully explained in the previous chapters. The
model summary is shown as:
The model summary explains how each layer of the model is functioning.
Depending on the layers in the model the output shape depends. For every layer the
activation function helps to perform the operations within the range of the activation
function.
Figure 6.1: Model Summary
Train model took 400 iterations or epochs and using this module will plot the
accuracy, loss graph of the model at each epoch/iteration
After the training has been completed, for testing the model it should be
validated and the validation will be done by the validate (i.e., input) given
by the user. During validation the user will give a query as input and the
trained model will generate the response which is exactly related to the
query given by the user. Below figure shows how the responses have been
given by the model to the validate.
Figure 6.4: Output of the validate phase
And the results of the output dataset is shown below, which displays the result
of the output dataset consists of actual and predicted values along with probabilities.
Below is the graphs of the Accuracy, and Loss of training of data of around
400 Epochs.
In the below graph x-axis represents EPOCH and the y-axis represents
accuracy, loss values in the below graph blue line represent training accuracy, and loss
and the orange line represents training set loss. In the below graph at each iteration the
training accuracy is increasing whereas the training loss graph is decreasing.
Figure 6.6: Model Accuracy Graph
To determine the overall accuracy of the model a confusion matrix will be generated from the
output dataset. The output dataset as mentioned in figure 6.5 is used to generate the confusion matrix as shown
in below.
To visualize the confusion matrix clearly a heat map is generated for the
confusion matrix which helps to visualize the data as shown in below figure.
Figure 6.8: Heatmap
From the above heatmap the overall accuracy of the model will be calculated
as mentioned in below figure.
When comparing the evaluation parameter of the proposed system with the
existing systems the proposed system gives the efficient values as it is shown in the
below table.
⮚ Specificity:
⮚ Accuracy:
Accuracy= (TP+TN)/(TP+FP+FN+TN)
⮚ Recall:
The ability of the model to find all the relevant cases within a dataset. The recall is
the percentage that can define the number of true positives divided by number of
true positives plus number of false negatives, which can be calculated using the
formula.
Recall=TP/(TP+FN)
⮚ F1 Score:
The harmonic mean of precision and recall which is better measure than accuracy.
F1 Score becomes high only when both precision and recall are high.
⮚ False Positives :
The Positive outcomes that the model predicted incorrectly is known as
False positives.
⮚ False Negatives:
The false negative is an outcome where the model incorrectly predicts
the negative class.
FN= (Colwise FN)- TP
TABLE 6.1: CLASSIFICATION OF EVALUATION PARAMETERS
6.4 APPLICATIONS
⮚ Sentiment Analysis
⮚ Speech Recognition
⮚ Language Modelling
This chapter discusses the experimental results of the proposed work and
also explains the graphs and comparison results of the proposed system.
APPENDIX A
SOURCE CODE
19CSEA016.ipynb
import tensorflow as tf
import numpy as np
import pandas as pd
import json
import nltk
from tensorflow.keras.preprocessing.text import Tokenizer
from tensorflow.keras.layers import Input, Embedding, LSTM, Dense, GlobalMaxPooling1
D, Flatten
from tensorflow.keras.models import Model
import matplotlib.pyplot as plt
#print the
data data
#pre-processing
#removing
punctuations import
string
data['inputs']=data['inputs'].apply(lambda wrd:[ltrs.lower() for ltrs
i n wrd if ltrs not in string.punctuation])
data['inputs']=data['inputs'].apply(lambda wrd: ''.join(wrd))
data
input_shape=x_train.shape[1
] input_shape
#define vocabulary
vocabulary=len(tokenizer.word_index)
print("number of unique words : ",
vocabulary)
output_length=le.classes_.shape[0]
print("output length : ",output_length)
print('unique words : ',tokenizer.word_index)
i=Input(shape=(input_shape,))
x=Embedding(vocabulary+1,10)(i)
x=LSTM(10,return_sequences=True)(x)
x=Flatten()(x)
x=Dense(output_length,activation="softmax")(x
) model=Model(i,x)
model.summary()
#commpiling the
model
model.compile(loss="sparse_categorical_crossentropy",optimizer='adam',
m etrics=['accuracy'])
train=model.fit(x_train,y_train,epochs=400
#Generate Heatmap
import seaborn as
sn
from sklearn.metrics import
classification_report import pandas as pd
#make predictions
model_pred=pd.read_excel("/content/output.xlsx",index_col=[0])
model_pred1=pd.read_excel("/content/Model2_Output.xlsx",index_col=[0])
model_pred.head()
#confusion matrix
confusion_model=pd.crosstab(model_pred.Predicted,model_pred.Actual)
confusion_model1=pd.crosstab(model_pred1.Predicted_class,model_pred1.A
c tual_class)
#heatmap
fig=plt.figure(figsize=(17,9)
) ax1=plt.subplot(121)
sn.heatmap(confusion_model,annot=True,cmap='Blues')
ax1.set_title('Output Model 1')
ax2=plt.subplot(122)
sn.heatmap(confusion_model1,annot=True,cmap='Blues'
) ax2.set_title('Output Model 2')
for i in
range(confusion_model.shape[0]):
TP=confusion_model.iloc[i,i]
FP=confusion_model.iloc[i,:].sum()-TP
FN=confusion_model.iloc[:,i].sum()-TP
TN=confusion_model.sum().sum()-TP-FP-F
N
accuracy=(TP+TN)/confusion_model.sum().sum(
) precision=TP/(TP+FP)
recall=TP/(TP+FN)
F1_score=(2*precision*recall)/(precision+recall
) print('class : ',confusion_model.index[i])
print('True positive for class '+confusion_model.index[i]+ ': ',TP)
print('False positive for class '+confusion_model.index[i]+ ':
',FP) print('False negative for class '+confusion_model.index[i]+
': ',FN) print('True negative for class '+confusion_model.index[i]+
': ',TN) print('Accuracy : ',accuracy)
print('Precision :
',precision) print('recall :
',recall) print('F1 Score :
',F1_score)
pd.DataFrame(classification_report(model_pred.Actual,model_pred.Predic
t ed,output_dict=True)).T
#pd.DataFrame(classification_report(model_pred1.Actual_class,model_pre
d 1.Predicted_class,output_dict=True)).T
plt.xlabel("model")
plt.ylabel("Accuracy of the
model")
plt.title("Model Comparision of Simple RNN and
LSTM") plt.show()
#model testing
print('welcome to spark AI chatbot(type exit to quit')
#testing the
model import
random while True:
texts_p=[]
prediction_input=input('you :
') if prediction_input=='exit':
print('Thank
you') break
#removing punctuations and converting into lower case
prediction_input=[letters.lower() for letters in prediction_input
if
letters not in string.punctuation]
prediction_input=''.join(prediction_input
) texts_p.append(prediction_input)
prediction_input=tokenizer.texts_to_sequences(texts_p)
prediction_input=np.array(prediction_input).reshape(-1)
prediction_input=pad_sequences([prediction_input],input_shape
)
response_tag=le.inverse_transform([output])[0]
print("spark :
",random.choice(responses[response_tag]))
APPENDIX B
SCREENSHOTS
In the below screen at the final epoch LSTM accuracy is 0.9831%, and now
see the main output in the below screen.
In the above graph x-axis represents model and the y-axis represents accuracy
of the model.
APPENDIX D
%%writefile content.json
{"intents"
: [
{
"tag":"greetings",
"input":["hello","hi there","nice to meet you","hi","hey there"
,"hi there"],
"responses":[
"hi, welcome",
"this is spark how can i help
you", "hello how may i help you",
"hello my friend, how can i help you"
]
},
{
"tag":"whoareyou",
"input":["who are you?","what is your name?","are you a chatbot
?","what can i call
you?"],
"responses":[
"I am a virtual assistant",
"my name is spark, i am a virtual
assistant", "you can call me spark"
]
},
{
"tag":"ai",
"input":["what is artificial intelligence?","what is ai?","arti
ficial
intelligence?","ai"],
"responses":[
"The theory and development of computer systems able to
perfo rm tasks normally requiring human intelligence, such as visual
percep tion, speech recognition, decision-
making, and translation between languages.",
"Artificial intelligence is the simulation of human
intellige nce processes by machines, especially computer systems.",
"AI is the ability of a machine to display human-
like capabilities such as reasoning, learning, planning and creativit
y. AI enables technical systems to perceive their environment, deal
w ith what they perceive, solve problems and act to achieve a
specific goal."
]
},
{
"tag":"ml",
"input":["what is machine learning?","what is ml?","tell me abo
ut machine
learning?","ml"],
"responses":[
"Machine learning is a branch of artificial intelligence
(AI) and computer science which focuses on the use of data and
algorithms to imitate the way that humans learn, gradually
improving its accura
cy.",
"Machine learning is a subfield of artificial intelligence,
w hich is broadly defined as the capability of a machine to imitate
int elligent human behavior. Artificial intelligence systems are
used to perform complex tasks in a way that is similar to how humans
solve pr oblems.",
"Machine learning (ML) is a type of artificial intelligence
( AI) that allows software applications to become more accurate at
pred icting outcomes without being explicitly programmed to do so.
Machine learning algorithms use historical data as input to predict
new outp
ut values."
]
},
{
"tag":"computer vision",
"input":["what is computer vision?","tell me about computer vis
ion?","computer
vision"],
"responses":[
"A field of artificial intelligence (AI) that enables
compute rs and systems to derive meaningful information from digital
images, videos and other visual inputs.",
"Computer vision is a field of artificial intelligence that
t rains computers to interpret and understand the visual world.
Using d igital images from cameras and videos and deep learning
models, machi nes can accurately identify and classify objects — and
then react to what they see",
"Computer vision is a field of AI that trains computers to
ca pture and interpret information from image and video data. By
applyin g machine learning (ML) models to images, computers can
classify obje cts and respond—
like unlocking your smartphone when it recognizes your face.",
"Computer vision is used to detect and classify objects (e.g.
, road signs or traffic lights), create 3D maps or motion estimation,
and played a key role in making autonomous vehicles a reality. Self-
driving cars collect data on their surroundings from sensors and came
ras, interpret it, and respond accordingly."
]
},
{
"tag":"nlp",
"input":["what is nlp?","what is natural language processing?",
"tell me about natural language processing?","natural language
proces sing?","nlp"],
"responses":[
"The application of computational techniques to the
analysis and synthesis of natural language and speech.",
"Natural language processing (NLP) refers to the branch of
co mputer science—
and more specifically, the branch of artificial intelligence or AI—
concerned with giving computers the ability to understand text and
sp oken words in much the same way human beings can.",
"NLP drives computer programs that translate text from one
la nguage to another, respond to spoken commands, and summarize
large vo lumes of text rapidly—even in real time."
]
},
{ "tag":"computer",
"input":["what is computer?","tell me
about computer?","compute
r"],
"responses":[
"A computer is a device that accepts
information (in the form
of digitized data) and manipulates it for some result based on a
pro gram, software, or sequence of instructions on how the data is
to be processed.",
"A computer is an electronic machine that processes raw
data to give information as output."
]
},
{
"tag":"cpu",
"input":["what is cpu?","tell me
"], about cpu","cpu","types of cpu
"responses":[
"CPU is considered as the brain of
the computer. · CPU perfor
ms all types of data processing operations.",
"Central Processing Unit (CPU) The CPU is the brain of a
comp uter, containing all the circuitry needed to process input,
store dat a, and output results.",
"The major types of CPU are classified as single-core,
dual- core, Quad-core, Hexa core, Octa-core, and Deca core
processor."
]
},
{
"tag":"alu",
"input":["what is alu?","tell me about alu?","natural language
processing?","alu"],
"responses":[
"An arithmetic-
logic unit is the part of a central processing unit that carries out
arithmetic and logic operations on the operands in computer
instructi on words.",
"An arithmetic logic unit (ALU) is a digital circuit used to
perform arithmetic and logic operations. It represents the
fundamenta l building block of the central processing unit (CPU) of
a computer."
,
"The ALU performs simple addition, subtraction,
multiplicatio n, division, and logic operations, such as OR and AND.
The memory sto res the program's instructions and data."
]
},
{ "tag":"python",
"input":["what is python?","modules in
python","python","module
s"],
"responses":[
"Python is an interpreted,
object-oriented, high-
level programming language with dynamic semantics.",
"python modules are os,sys,data time,math,random,JSON"
]
},
{
"tag":"html",
"input":["html","what is html?","tell me about
html"], "responses":[
"HTML is an acronym which stands for Hyper Text Markup Langua
ge which is used for creating web pages and web applications",
"HTML is the standard markup language for creating Web pages.
HTML describes the structure of a Web page. HTML consists of a
serie s of elements."
]
},
{
"tag":"sql",
"input":["what is
sql?","sql"], "responses":[
"SQL stands for Structured Query Language. It is the
standard language used to maintain the relational database and
perform many d
ifferent data manipulation operations on the data.",
"it is the standard language for relational database
manageme nt systems. SQL statements are used to perform tasks such
as update d ata on a database, or retrieve data from a database."
]
},
{
"tag":"types in sql",
"input":["what are the types in sql","types in
sql"], "responses":[
"Data Definition Language, Data Query Language, Data Manipula
tion Language, Data Control Language,Transaction Control Language."
]
},
{
"tag":"dbms",
"input":["what is dbms?","tell me about
dbms?","dbms"], "responses":[
"DBMS is a collection of programs that facilitates users to c
reate and maintain a database.",
"A database management system (or DBMS) is essentially
nothin g more than a computerized data-keeping system"
]
},
{
"tag":"types in dbms",
"input":["types in dbms","dbms
types"], "responses":[
"Relational database,Object oriented database,Hierarchical da
tabase,Network database."
]
},
{
"tag":"oops",
"input":["what is
oops","oops"], "responses":[
"Object-
oriented programming is based on the concept of objects In object-
oriented programming data structures, or objects are defined, each
wi th its own properties or attributes"
]
},
{
"tag":"pillars of oops",
"input":["four pillars of oops","pillars of
oops"], "responses":[
"Abstraction ,Inheritance ,Polymorphism ,Encapsulation"
]
},
{
"tag":"cancer",
"input":["what is a
cancer","cancer"], "responses":[
"Cancer is a disease caused when cells divide uncontrollably
and spread into surrounding tissues"
]
},
{
"tag":"causes of cancer",
"input":["main causes of cancer","cause of
cancer"], "responses":[
"Leading risk factors for preventable cancers are smoking, ge
tting too much ultraviolet (UV) radiation from the sun or tanning
bed s, being overweight or having obesity, and drinking too much
alcohol. "
]
},
{
"tag":"cardiology",
"input":["what is cardiology","cardiology","What is the role of
cardiology"],
"responses":[
"Cardiology is a medical specialty and a branch of internal
m edicine concerned with disorders of the heart.",
"Cardiology is the study of the heart. In medicine,
cardiolog y is a branch of medicine that deals with disorders of the
heart and the cardiovascular system.",
"Cardiologists diagnose, assess and treat patients with
defec ts and diseases of the heart and the blood vessels, which are
known a s the cardiovascular system."
]
},
{
"tag":"cardiovascular disease",
"input":["what is cardiovascular disease","cardiovascular disea
se","tell me about cardiovascular
disease","cvd"], "responses": [
"Heart and blood vessel disease (also called heart disease)
i ncludes numerous problems, many of which are related to a process
cal led atherosclerosis",
"Cardiovascular disease (CVD) is a general term that
describe s a disease of the heart or blood vessels. Blood flow to
the heart, b rain or body can be reduced because of a: blood clot
build-
up of fatty deposits inside an artery, leading to the artery
hardenin g and narrowing"]
},
{
"tag":"symptoms",
"input": ["symptoms of cardiovascular disease","main cause of
c ardiovascular disease"],
"responses":[
"Chest pain, chest tightness, chest pressure and chest
discom fort (angina)",
"Shortness of breath.Pain in the neck, jaw, throat, upper
bel ly area or back,Pain, numbness, weakness or coldness in the legs
or a rms if the blood vessels in those body areas are narrowed.",
"High blood pressure, high blood cholesterol, and smoking
are key risk factors for heart disease"]
},
{
"tag":"effects",
"input":["effects of cardiovascular disease","effects of
cvd"], "responses":[
"The most important behavioural risk factors of heart
disease and stroke are unhealthy diet, physical inactivity, tobacco
use and
harmful use of alcohol.",
"The effects of behavioural risk factors may show up in
indiv iduals as raised blood pressure, raised blood glucose, raised
blood l ipids, and overweight and obesity",
"Patients with history of CVD may experience various
physical and emotional symptoms such as fatigue, edema, and sleeping
difficul ties that limit their physical and social activities which
will in tu
rn result in poor quality of life."]
},
{
"tag":"treatments for cardiovascular disease",
"input":["treatments for cardiovascular disease"],
"responses":["Eat a healthy diet. Maintaining a heart healthy
d
iet is key for preventing heart disease.Exercise regularly ,Control
y our blood pressure ,Keep cholesterol under control ,Maintain a
health y weight,Limit alcohol intake ,Don't smoke ,Manage stress."]
},
{
"tag":"ophthalmology",
"input":["what is
ophthalmology","ophthalmology"], "responses":[
"An ophthalmologist is a medical or osteopathic doctor who
sp ecializes in eye and vision care",
"Ophthalmology is a surgical subspecialty within medicine
tha t deals with the diagnosis and treatment of eye disorders."]
},
{
"tag":"types of ophthalmology diseases",
"input":["types of ophthalmology
ses"], diseases","ophthalmology disea
"responses":[
"Refractive Errors ,Age-
Related Macular Degeneration ,Cataract ,Diabetic Retinopathy
,Glaucom a ,Amblyopia ,Strabismus.",
"These refractive errors are the most common eye problems
are Refractive errors include nearsightedness (myopia),
farsightedness (
hyperopia) and distorted vision at all distances (astigmatism)."]
},
{
"tag":"treatments for ophthalmology",
"input":["treatments for
ophthalmology","treatment"], "responses":[
"Prescription of glasses or contact lenses to improve vision
,Medical treatment with pharmaceutical agents ,Surgical procedures
su ch as complex microsurgery.",
"Ophthalmologists prescribe medications to treat eye diseases
, implement laser therapy, and perform surgery when needed.
Ophthalmo logists provide both primary and specialty eye care -
medical and sur gical."]
},
{
"tag":"dermatology",
"input":["what is
dermatology","dermatology"], "responses":[
"A dermatology is a medical doctor who specializes in
treatin g the skin, hair, and nails. Dermatologists care for people
of all ag es",
"Dermatology is the branch of medicine dealing with the
skin. It is a speciality with both medical and surgical aspects. A
dermato logist is a specialist medical doctor who manages diseases
related to
skin, hair, nails, and some cosmetic problems."]
},
{
"tag":"paraben",
"input":["what is
paraben","paraben"], "responses":[
"Parabens are a class of widely used preservatives in
cosmeti c and pharmaceutical products. Chemically, they are a series
of parah ydroxybenzoates or esters",
"Parabens are a group of chemicals widely used as
artificial preservatives in cosmetic and body care products."]
},
{
"tag":"aws",
"input":["what is
aws","aws"], "responses":[
"AWS (Amazon Web Services) is a comprehensive, evolving
cloud computing platform provided by Amazon that includes a mixture
of inf
rastructure-as-a-service (IaaS), platform-as-a-
service (PaaS) and packaged-software-as-a-service (SaaS)
offerings.", "Amazon Web Services (AWS) is the world's most
comprehensive
and broadly adopted cloud platform, offering over 200 fully
featured services from data centers globally."]
},
{
"tag":"cloud computing",
"input":["what is cloud computing","cloud
computing"], "responses":[
"Cloud computing is the on-
demand delivery of IT resources over the Internet with
pay-as-you- go pricing",
"Cloud computing is the on-
demand availability of computer system resources, especially data
sto rage (cloud storage) and computing power, without direct active
manag ement by the user."]
},
{
"tag":"electrical engineering",
"input":["what is electrical
engineering","eee"], "responses":[
"Electrical engineering is an engineering discipline
concerne d with the study, design, and application of equipment,
devices, and systems which use electricity ,electronics, and
electromagnetism",
"Electrical engineers design, develop, test and manage the
ma nufacturing of electrical equipment, from electric motors and
navigat ion systems to power generation equipment and the electrical
componen ts of vehicles and personal devices."]
},
{
"tag":"ece",
"input":["what is
ece","ece"], "responses":[
"ECE is Electronics and Communication Engineering. ECE is an
engineering branch that includes the production, design, research
and
testing of electronic devices and technology issues related to
telec ommunications systems, electronics, computers and related
industrial sectors.",
"Electronics and Communications Engineering (ECE) involves
re searching, designing, developing, and testing electronic
equipment us ed in various systems. Electronics and Communications
engineers also conceptualize and oversee the manufacturing of
communications and bro adcast systems."]
},
{
"tag":"cse",
"input": ["what is
"], cse","cse","computer science and
engineering
"responses":[
"The full form of CSE is Computer
Science Engineering. It is
an engineering course that involves several computer science
aspects essential for the creation of the computer system.",
"Computer Science engineering is a course that deals with
var ious aspects of computer science and its applications such as
designi ng, developing, and implementing softwares, websites,
networking, alg orithms, and more."]
},
{
"tag":"physics",
"input":["what is
physics","physics"], "responses":[
"Physics is the branch of science that deals with the
structu re of matter and how the fundamental constituents of the
universe int eract.",
"Physics is the natural science that studies matter, its
fund amental constituents, its motion and behavior through space and
time,
and the related entities."]
},
{
"tag":"chemistry",
"input": ["what is
chemistry","chemistry"], "responses":[
"Chemistry is a branch of natural science that deals
principa lly with the properties of substances, the changes they
undergo, and the natural laws that describe these changes.",
"Chemistry is a science that deals with the composition,
stru cture, and properties of substances and with the changes that
they go
through. : chemical composition, properties, or processes."]
},
{
"tag":"web development",
"input":["what is web development","web
development"], "responses":[
"Web development refers to the creating, building, and
mainta ining of websites. It includes aspects such as web design,
web publis hing, web programming, and database management.",
"It is the creation of an application that works over the
int ernet i.e. websites.",
"Web development is the work involved in developing a
website for the internet or an interanet (a private network). Web
developmen t can range from developing a simple single static page
of plain next to complex web application, electronic businesses, and
social networ
k services."]
},
{
"tag":"paralysis",
"input":["what is
paralysis","paralysis"], "responses":[
"Paralysis occurs when you're unable to make voluntary
muscle movements. A nervous system problem causes paralysis.",
"Paralysis is the loss of the ability to move some or all of
your body. It can have lots of different causes, some of which can
be
serious."]
},
{
"tag":"brain tumor",
"input":["what is brain tumor","brain
tumor"], "responses":[
"A brain tumor is a collection, or mass, of abnormal cells
in your brain. Your skull, which encloses your brain, is very
rigid.",
"A brain tumor is a mass or growth of abnormal cells in your
brain. Many different types of brain tumors exist. Some brain tumors
are noncancerous (benign), and some brain tumors are cancerous
(malig nant)."]
},
{
"tag":"lung cancer",
"input": ["what is lung cancer","lung
cancer"], "responses":[
"Cancer is a disease in which cells in the body grow out of
c ontrol. When cancer starts in the lungs, it is called lung
cancer.",
"Lung cancer is a type of cancer that begins in the lungs.
Yo ur lungs are two spongy organs in your chest that take in oxygen
when
you inhale and release carbon dioxide when you exhale."]
},
{
"tag":"blood cancer",
"input":["what is blood cancer","types of blood
cancer"], "responses":[
"Blood cancer is a type of cancer that affects your blood
cel ls. Leukaemia, lymphoma and myeloma are some of the most common
types
of blood cancer.",
"Lymphoma ,Myelodysplastic Syndromes (MDS)
,Myeloproliferativ e Disorder (MPD) ,Multiple Myeloma."]
},
{
"tag":"skin cancer",
"input":["what is skin
cancer","cancer"], "responses":[
"Skin cancer is the out-of-
control growth of abnormal cells in the epidermis, the outermost
skin layer, caused by unrepaired DNA damage that triggers.",
"Skin cancer that forms in melanocytes (skin cells that make
pigment) is called melanoma. Skin cancer that forms in the lower
part
of the epidermis (the outer layer of the skin) is called basal
cell carcinoma."]
},
{
"tag":"malaria",
"input":["what is
malaria","malaria"], "responses":[
"Malaria is a disease caused by a parasite. The parasite is
s pread to humans through the bites of infected mosquitoes."]
},
{
"tag":"virus",
"input":["what is
virus","virus"], "responses":[
"A virus is an infectious microbe consisting of a segment
of nucleic acid (either DNA or RNA) surrounded by a protein coat",
"A virus is a submicroscopic infectious agent that
replicates only inside the living cells of an organism."]
},
{
"tag":"population",
"input":["population of india","india
population"], "responses":["140.76 crores"]
},
{
"tag":"largest country",
"input":["which is the largest country in the world","largest c
ountry in the world"],
"responses":["russia"
]
}
]
}
APPENDIX D
STUDENT CONTRIBUTION
S.No ACTIVITY 19K61A0549 19K61A0530 19K61A0505 20K65A0506
1 Title Conformation ✔ ✔ ✔ ✔
2 Literature Survey ✔ ✔ ✔ ✔
3 Problem ✔ ✔ ✔ ✔
Formulation
4 Requirements ✔ ✔ ✔ ✔
Gathering
5 Designing ✔ ✔
6 Implementation ✔ ✔
7 Documantation ✔ ✔ ✔ ✔
APPENDIX E
OUTCOME
DESCRIPTION
NO
Develop problem formation and design skills for engineering and real-
CO1
world problems.
Collect and Generate ideas through literature survey on current
CO2 research areas which help to analyse and present to impart knowledge
in different fields.
Import knowledge on software & hardware to meet industry
CO3
perspective needs and standards.
Create interest to carry out research on innovative ideas as a lifelong
CO4
learning.
Ability to work with team, and enrich presentation and communication
CO5
skills.
CO6 Create a platform that makes students employable.
COs/POs PO1 PO2 PO3 PO4 PO5 PO6 PO7 PO8 PO9 PO10 PO11 PO12 PSO1 PSO2
CO1 3 0 0 1 0 1 1 1 3 3 0 1 0 0
CO2 3 3 0 0 0 2 0 0 3 2 0 0 0 0
CO3 2 0 1 1 3 0 0 0 3 2 0 0 0 0
CO4 3 0 0 3 3 0 3 1 3 3 1 1 0 0
CO5 2 0 0 0 2 0 0 0 3 3 0 3 0 0
CO6 2 1 0 0 3 1 0 3 3 2 2 2 0 0
Overall
3 1 1 1 2 1 1 0 3 2 1 1 0 0
Course
PROGRAM OUTCOMES (POs)
Programme Educational
RELEVANCE
PEOs Objectives
To get the project executed, all
Graduates will be able to analyze, the team have done analysis and
PEO 1 design, and develop advanced research-oriented surveys to
computer applications to provide a frame the solution and identify
solution to real-world problems. the limitations.
CO PO PI Relevance
PO9 9.5.2 This work is carried out by all the team members.
CO1 - -
CO2 - -
CO3 - -
CO4 - -
CO5 - -
CO6 - -
D9