Professional Documents
Culture Documents
L04-06 Intelligent Agents 23AIE231M Intro To AI&DS 20231004
L04-06 Intelligent Agents 23AIE231M Intro To AI&DS 20231004
Intelligent Agents
Unit 02: L04-L06
In an episodic task environment, the agent receives a percept and then performs a single
action. The action is independent of actions in previous episodes. E.g. a defective part on an
assembly line once detected, is removed, regardless of previous actions of the checking robot.
In a sequential task environment, current decisions affect all future decisions. E.g. driverless
car or chess. Such task environments are much more difficult than episodic task environments.
If the environment can change while an agent is deliberating, then we say the environment is
dynamic for that agent; otherwise, it is static. to do next. Chess is static.
This key challenge for AI is to find out how to write programs that, to the extent possible,
produce rational behavior from a smallish program rather than from a vast table.
• Utility-based agents
1. SIMPLE REFLEX AGENTS: They choose actions on the basis of the current percept. The table
shows pseudocode for agent function tabulated in slide 5.
The program is quite small as compared to the agent function. This reduction is a natural
outcome of ignoring percept history which reduces possibility space from 3𝑇 to just 3.
20 Puja Dutta, CIE, Amit Agarwal, EEE | TIFAC-CORE in Cyber security
6. Agent Types [2/8]
A further small reduction comes from the fact that actions for whether to suck or not do
nothing depend on which cell the agent is in. Thus, the instantaneous percept reduces from
𝐴𝑐𝑙𝑒𝑎𝑛 , 𝐴𝑑𝑖𝑟𝑡𝑦 , 𝐵𝑐𝑙𝑒𝑎𝑛 , 𝐵𝑑𝑖𝑟𝑡𝑦 to 𝐴, 𝐵, 𝐷𝑖𝑟𝑡𝑦? .
The INTERPRET-INPUT function generates an abstracted description of the current state from
the percept, and the RULE-MATCH function returns the first rule in the set of rules that
matches the given state description.
21 Puja Dutta, CIE, Amit Agarwal, EEE | TIFAC-CORE in Cyber security
6. Agent Types [3/8]
In the figure:
Rectangles denote the current internal state of the agent’s decision process.
Ovals denote the background information used in the process.
Paris Riots
2018
➢ Partial observation may lead to infinite looping (vacuum example). Randomization can help
escape infinite looping.
Thus, can we say that some degree of irrationality, in some contexts, is more intelligent?
In the case of the video on Paris 2018 Riots, one could a) keep some frames in the memory.
However, this is necessary but not sufficient.
One needs to how some knowledge of the world, i.e., b) a model of the world, that will lead to
questions, such as for example:
➢ if the smoke or fire is so massive then what is it that is burning?
➢ given that Paris is densely populated, do I see reports of people dead or injured given the rather
strong impression of unrest the image conveys?
➢ do I see any reports of loss of something treasured, important or expensive?
➢…
Finally, we need b) a model of how the world changes due to the action of the agent.
Q2: Define in 3 lines each term: agent, agent function, agent program, rationality, autonomy,
reflex agent, model-based agent, goal-based agent, utility-based agent, learning agent.
Q3: Write the objective function you seek to maximize over the next 5 years after graduation
with respect to your personal life (only).