By DR Narayana Swamy Ramaiah Professor, Dept of CSE SCSE, FET, JAIN Deemed To Be University
By DR Narayana Swamy Ramaiah Professor, Dept of CSE SCSE, FET, JAIN Deemed To Be University
Intelligent Agent
By
• Intelligent Agent
An intelligent agent is a GOAL-DIRECTED AGENT. It perceives its environment through its sensors using the
observations and built-in knowledge, acts upon the environment through its actuators.
• Rational Agent
A rational agent is an agent which takes the right action for every perception. By
doing so, it maximizes the performance measure, which makes an agent be the most
successful.
• It is expected from an intelligent agent to act in a way that maximizes its performance
measure. Therefore, the rationality of an agent depends on four things:
The performance measure which defines the criterion of success.
The agent’s built-in knowledge about the environment.
The actions that the agent can perform.
The agent’s percept sequence until now.
For example: score in exams depends on the question paper as well as our knowledge.
• Note: Rationality maximizes the expected performance, while perfection maximizes the
actual performance which leads to omniscience.
Autonomy
Information
Exploration Learning Autonomy
Gathering
Our definition requires a rational agent not only to explore and gather information but also to learn as much as
possible from what it perceives. The agent’s initial configuration could reflect some prior knowledge of the
environment, but as the agent gains experience this may be modified and augmented. to the extent of autonomy.
• Omniscient Agent
An omniscient agent is an agent which knows the actual outcome of its action in advance. However, such agents are
impossible in the real world.
• Software Agents
It is a software program which works in a dynamic environment. These agents are also known as Softbots because
all body parts of software agents are software only.
• Therefore, an agent is the combination of the architecture and the program i.e.
• Note: The difference between the agent program and agent function is that an agent program takes the
current percept as input, whereas an agent function takes the entire percept history.
Types of Agent Programs
• Varying in the level of intelligence and complexity of the task, the following
four types of agents are there:
Simple reflex agents: It is the simplest agent which acts according to the current percept only,
pays no attention to the rest of the percept history. The agent function of this type relies on the
condition-action rule – “If condition, then action.” It makes correct decisions only if the
environment is fully observable. These agents cannot ignore infinite loop when the
environment is partially observable but can escape from infinite loops if the agents randomize
its actions.
Example: iDraw, a drawing robot which converts the typed characters into writing without storing the past data.
• Note: Simple reflex agents do not maintain the internal state and do not depend
on the percept theory.
• Model-based agent: These type of agents can handle partially observable environments by maintaining
some internal states. The internal state depends on the percept history, which reflects at least some of the
unobserved aspects of the current state. Therefore, as time passes, the internal state needs to be updated
which requires two types of knowledge or information to be encoded in an agent program i.e., the evolution
of the world on its own and the effects of the agent’s actions.
Example: When a person walks in a lane, he maps the pathway in his mind.
Goal-based agents: It is not sufficient to have the current state information unless the goal is not
decided. Therefore, a goal-based agent selects a way among multiple possibilities that helps it to reach
its goal.
Note: With the help of searching and planning (subfields of AI), it becomes easy for the Goal-based agent to reach its destination.
• Utility-based agents: These types of agents are concerned about the performance measure. The agent
selects those actions which maximize the performance measure and devote towards the goal.
Example: The main goal of chess playing is to ‘check-and-mate’ the king, but the player completes several small goals previously.
Note: Utility-based agents keep track of its environment, and before reaching its main goal, it completes several tiny goals that may
come in between the path.
Learning agents: The main task of these agents is to teach the agent machines to operate in an unknown
environment and gain as much knowledge as they can.
• Note: The objective of a Learning agent is to improve the overall performance of the agent.
Working of an agent program’s components
• The function of agent components is to answer some basic questions like “What is the
world like now?”, “what do my actions do?” etc.
• We can represent the environment inherited by the agent in various ways by distinguishing
on an axis of increasing expressive power and complexity as discussed below:
Atomic Representation: Here, we cannot divide each state of the world. So, it does not have
any internal structure. Search, and game-playing, Hidden Markov Models, and Markov
decision process all work with the atomic representation.
Factored Representation: Here, each state is split into a fixed set of attributes or variables
having a value. It allows us to represent uncertainty. Constraint satisfaction, propositional
logic, Bayesian networks, and machine learning algorithms work with the Factored
representation.
Note: Two different factored states can share some variables like current GPS location, but two different
atomic states cannot do so.
• Structured Representation: Here, we can explicitly describe various and
varying relationships between different objects which exist in the world.
Relational databases and first-order logic, first-order probability models, natural language
understanding underlie structured representation.