Foundations of artificial intelligence - Chapter 2: Intelligent agents includes Agents and environments, Rationality; PEAS (Performance measure, Environment, Actuators, Sensors); Environment types, Agent types.
Trang 1FOUNDATIONS OF ARTIFICIAL
INTELLIGENCE
Chapter 2 Intelligent Agents
Trang 3An agent is anything that can be viewed as
perceiving its environment through sensors and
acting upon that environment through actuators
Example:
Human agent:
eyes, ears, and other organs for sensors;
hands, legs, mouth, and other body parts for actuators
Robotic agent:
cameras and infrared range finders for sensors;
various motors for actuators
Trang 4Agents and environments
Agents include human, robots, softbots, thermostats, etc.
The agent function maps from percept histories to
actions: [f: P* A]
The agent program runs on the physical architecture
Trang 5Vacuum-cleaner world
Environment: square A and B
Percepts: location and contents, e.g., [A,Dirty]
Actions: Left, Right, Suck, NoOp
Trang 6The vacuum-cleaner world
What is the right function? Can it be implemented in
function REFLEX-VACUUM-AGENT ([location, status]) return an action
if status == Dirty then return Suck else if location == A then return Right else if location == B then return Left
Percept sequence Action
Trang 7Rational agents
A rational agent is one that does the right thing based
on what it can perceive and the actions it can
perform
What is the right thing?
Approximation: The right action is the one that will cause the
agent to be most successful
Measure of success?
Performance measure: An objective criterion for
success of an agent's behavior
E.g., performance measure of a vacuum-cleaner agent could be amount of dirt cleaned up, amount of time taken, amount of
electricity consumed, amount of noise generated, etc
Trang 8Rational agents
What is rational at a given time depends on four
things:
Performance measure,
Percept sequence to date (sensors)
Prior environment knowledge,
Actions,
Rational Agent: For each possible percept sequence,
a rational agent should select an action that is
expected to maximize its performance measure,
given the evidence provided by the percept sequence and whatever built-in knowledge the agent has.
Trang 9information gathering, exploration
An agent is autonomous if its behavior is determined
by its own experience (with ability to learn and
adapt)
Rational exploration, learning, autonomy
Trang 10To design a rational agent we must specify its task environment.
PEAS description of the environment:
Performance measure: Goals/desires the agent should try to
achieve
Environment: in which the agent exists
Actuators: Actions which may act the environment
Sensors: Percepts/observations of the environment
Trang 11Environment: Roads, other traffic, pedestrians, customers
Actuators: Steering wheel, accelerator, brake, signal, horn
Sensors: Cameras, sonar, speedometer, odometer, engine
sensors, keyboard
Internet shopping
Performance measure: price, quality, appropriateness, efficiencyEnvironment: current and future WWW sites, vendors, shippersActuators: display to user, follow URL, ll inform
Sensors: HTML pages (text, graphics, scripts)
Trang 12Agent: Medical diagnosis system
Performance measure: Healthy patient, minimize costs, lawsuitsEnvironment: Patient, hospital, staff
Actuators: Screen display (questions, tests, diagnoses,
treatments, referrals)
Sensors: Keyboard (entry of symptoms, findings, patient's
answers)
Agent: Part-picking robot
Performance measure: Percentage of parts in correct bins
Environment: Conveyor belt with parts, bins
Actuators: Jointed arm and hand
Sensors: Camera, joint angle sensors
Trang 13Agent: Interactive English tutor
Performance measure: Maximize student's score on test
Environment: Set of students
Actuators: Screen display (exercises, suggestions, corrections)Sensors: Keyboard (typed words)
Trang 14Environment types
Fully observable (vs partially observable):
An agent's sensors give it access to the complete state of the environment at each point in time
Deterministic (vs stochastic):
The next state of the environment is completely determined by the current state and the action executed by the agent (If the environment is deterministic except for the actions of other
agents, then the environment is strategic)
Episodic (vs sequential):
The agent's experience is divided into atomic "episodes" (each episode consists of the agent perceiving and then performing a single action), and the choice of action in each episode depends only on the episode itself
Trang 15Environment types
Static (vs dynamic):
The environment is unchanged while an agent is deliberating (The environment is semidynamic if the environment itself does not change with the passage of time but the agent's
performance score does)
Discrete (vs continuous):
A limited number of distinct, clearly defined percepts and
actions
Single agent (vs multiagent):
An agent operating by itself in an environment
Trang 16Environment types
Solitaire Image-Analysis
system
Intenet shopping
Trang 17Environment types
Solitaire Image-Analysis
system
Internet shopping
Fully vs partially observable: an environment is full
observable when the sensors can detect all aspects
that are relevant to the choice of action
Trang 18Environment types
Solitaire Image-Analysis
system
Intenet shopping
Taxi
Episodic??
Static??
Deterministic vs stochastic: if the next environment
state is completely determined by the current state the executed action then the environment is deterministic
Trang 19Environment types
Solitaire Image-Analysis
system
Intenet shopping
Taxi
Static??
Discrete??
Episodic vs sequential : In an episodic environment the
agent’s experience can be divided into atomic steps where the agents perceives and then performs a single action The choice of action depends only on the episode itself
Trang 20Environment types
Solitaire Image-Analysis
system
Intenet shopping
Taxi
Static vs dynamic : If the environment can change while the agent is choosing an action, the environment is dynamic Semi-dynamic if the agent’s performance changes even
when the environment remains the same
Trang 21Environment types
Solitaire Image-Analysis
system
Intenet shopping
Taxi
Discrete vs continuous : This distinction can be applied to the state of the environment, the way time is handled and to the percepts/actions of the agent.
Trang 22Environment types
Solitaire Image-Analysis
system
Internet shopping
Taxi
Single vs multi-agent : Does the environment contain other agents who are also maximizing some performance
measure that depends on the current agent’s actions?
Trang 23Environment types
The environment type largely determines the agent design
The simplest environment is
Fully observable, deterministic, episodic, static, discrete and
single-agent
Most real situations are:
Partially observable, stochastic, sequential, dynamic, continuous and multi-agent
Trang 24Agent functions and programs
How does the inside of the agent work?
Agent = architecture + program
An agent is completely specified by the agent
function mapping percept sequences to actions
One agent function (or a small equivalence class) is rational
Aim: find a way to implement the rational agent function
concisely
All agents have the same skeleton:
Input = current percepts
Output = action
Program= manipulates input to produce output
Trang 25Even with learning, need a long time to learn the table entries
Function TABLE-DRIVEN_AGENT(percept) returns an action
static: percepts, a sequence initially empty
table, a table of actions, indexed by percept sequence
append percept to the end of percepts
action LOOKUP(percepts, table)
return action
Trang 26Agent types
Four basic types in order of increasing generality:
Simple reflex agents
Model-based reflex agents
Goal-based agents
Utility-based agents
All these can be turned into learning agents.
Trang 27Agent types; simple reflex
Select action on the
basis of only the current percept.
E.g the agent
vacuum-Large reduction in possible
percept/action situations(next page) Implemented through
condition-action rules
If dirty then suck
Trang 28The vacuum-cleaner world
function REFLEX-VACUUM-AGENT ([location, status]) return an action
if status == Dirty then return Suck
else if location == A then return Right
else if location == B then return Left
Trang 29Agent types; simple reflex
Will only work if the environment is fully observable
otherwise infinite loops may occur.
function SIMPLE-REFLEX-AGENT(percept) returns an action
static: rules, a set of condition-action rules
state INTERPRET-INPUT(percept) rule RULE-MATCH(state, rule)
action RULE-ACTION[rule]
return action
Trang 30Agent types; reflex and state
To tackle partially observable
environments.
Maintain internal state
Over time update state using world knowledge
How does the world change
How do actions affect world.
Model of World
Trang 31Agent types; reflex and state
function REFLEX-AGENT-WITH-STATE(percept) returns an action
static: rules, a set of condition-action rules
state, a description of the current world state action, the most recent action.
state UPDATE-STATE(state, action, percept)
rule RULE-MATCH(state, rule)
action RULE-ACTION[rule]
return action
Trang 32Agent types; goal-based
The agent needs a goal to know which situations are
desirable.
Things become difficult when long sequences of actions are required to find the goal.
Typically investigated in search and planning
research.
Major difference: future
is taken into account
Is more flexible since
Trang 33Agent types; utility-based
Certain goals can be reached in different ways.
Some are better, have a higher utility.
Utility function maps a (sequence of) state(s) onto a real number.
Improves on goals:
Selecting between conflicting goals Select appropriately between several goals based on likelihood of success.
Trang 34Agent types; learning
All previous programs describe methods for selecting
agent-actions.
Yet it does not explain the origin of these programs
Learning mechanisms can be used to perform this task.
Teach them instead of instructing them.
Advantage is the robustness of the program toward initially unknown environments.
Trang 35Agent types; learning
Trang 36Agents interact with environments through actuators and sensorsThe agent function describes what the agent does in all
circumstances
The performance measure evaluates the environment sequence
A perfectly rational agent maximizes expected performance
Agent programs implement (some) agent functions
PEAS descriptions dene task environments
Environments are categorized along several dimensions:
observable? deterministic? episodic? static? discrete? agent?
single-Several basic agent architectures exist: