The document describes different dimensions along which task environments can vary, including whether they are fully or partially observable, deterministic or stochastic, episodic or sequential, static or dynamic, discrete or continuous, and involve single or multiple agents. It then provides examples of different task environments and how they vary along these dimensions. The document goes on to describe different types of agents from reactive to learning agents. It states the goal is to design agents for different task environments by constructing a function that takes the current state, percepts, and knowledge base to output an action. Finally, it discusses deductive, inductive, and hybrid approaches to developing agents.
The document describes different dimensions along which task environments can vary, including whether they are fully or partially observable, deterministic or stochastic, episodic or sequential, static or dynamic, discrete or continuous, and involve single or multiple agents. It then provides examples of different task environments and how they vary along these dimensions. The document goes on to describe different types of agents from reactive to learning agents. It states the goal is to design agents for different task environments by constructing a function that takes the current state, percepts, and knowledge base to output an action. Finally, it discusses deductive, inductive, and hybrid approaches to developing agents.
The document describes different dimensions along which task environments can vary, including whether they are fully or partially observable, deterministic or stochastic, episodic or sequential, static or dynamic, discrete or continuous, and involve single or multiple agents. It then provides examples of different task environments and how they vary along these dimensions. The document goes on to describe different types of agents from reactive to learning agents. It states the goal is to design agents for different task environments by constructing a function that takes the current state, percepts, and knowledge base to output an action. Finally, it discusses deductive, inductive, and hybrid approaches to developing agents.
The task environment can be characterized by the following
attributes: I Fully observable vs partially observable. I Deterministic vs non-deterministic. If s, s1 ∈ S are states and a ∈ A is an action then if (s, a) → s1 , s1 completely determined by s and a - except for other agents - then it is deterministic. Called stochastic - if it is non-deterministic and some aspects of the environment can be modelled via probability distributions (e.g. games of chance). I Episodic (indp. sense-act chunks) vs sequential (curr. action affects future) I Dynamic (environment can change - even during sense-act cycle) vs static. Called semi-dynamic if environment does not change with time but agents performance does. I Discrete vs continuous - nature of time applied to state, environment, actions. I Single agent vs multiple agents. Cooperative vs competitive within a multi-agent environment. Examples: Dimensions of variation
Task Env Obs(F/P) Det.(D/S) Epis(E/S) Static(S/D) Disc.(D/C) No. of Ags(1/M)
Crossword F D S S D 1 Tournament chess F Str. S Semi D M Self driving car P S S D C M Med. diagnosis P S S D C 1 Part picking P S E D C 1 Image analysis F D E Semi D 1 Plant controller P S S D C 1 Tutor P S S D D M Chem. struc. P S S S D 1 Agent types Agents can be classified into some broad agent types. The following agent types are common (from the simple to complex). I Reactive or reflex agents. Modelled by table driven representations. I Rule driven agents. Modelled by situation-action rules. I Model driven agents. Constructs models of the task environment, effect of actions on the environment and changes in the environment due to actions. I Goal driven agents. Have models and also goals. Typically, one goal. I Utility based agents. Generalization of goal driven agents. Optimize utility in the presence of multiple conflicting goals using trade-offs. I Learning agents. f can keep changing based on changing dependencies. Even when dependencies are assumed constant they are not known a priori so they may have to be learnt from data. Focus of the course: AI as agent design
Course goals: Design agents in different task environments.
Formally, construct function f : S × P × K → A, where S is set of states, P is set of percepts, K is the knowledge base (history), A set of actions. f is the findAction function which takes as argument: current state, percepts (from sensing), knowledge (past history, other information) and outputs an action. To successfully build agents we need appropriate representations for S, P, K and, A and suitable algorithms that implement f . Two main approaches
I Deductive or top-down. We have knowledge about the task
domain and it has to be implemented as an agent. I Inductive or bottom-up. We do not have knowledge about the domain and an agent has to learn how to act. I A third hybrid situation is one where we have partial knowledge about the task domain but learning is also needed to be able to do the task.