1. Trang chủ
  2. » Công Nghệ Thông Tin

Foundations of artificial intelligence - Chapter 2: Intelligent agents

36 58 0

Đang tải... (xem toàn văn)

Tài liệu hạn chế xem trước, để xem đầy đủ mời bạn chọn Tải xuống

THÔNG TIN TÀI LIỆU

Thông tin cơ bản

Định dạng
Số trang 36
Dung lượng 1,65 MB

Các công cụ chuyển đổi và chỉnh sửa cho tài liệu này

Nội dung

Foundations of artificial intelligence - Chapter 2: Intelligent agents includes Agents and environments, Rationality; PEAS (Performance measure, Environment, Actuators, Sensors); Environment types, Agent types.

Trang 1

FOUNDATIONS OF ARTIFICIAL

INTELLIGENCE

Chapter 2 Intelligent Agents

Trang 3

An agent is anything that can be viewed as

perceiving its environment through sensors and

acting upon that environment through actuators

Example:

Human agent:

eyes, ears, and other organs for sensors;

hands, legs, mouth, and other body parts for actuators

Robotic agent:

cameras and infrared range finders for sensors;

various motors for actuators

Trang 4

Agents and environments

Agents include human, robots, softbots, thermostats, etc.

The agent function maps from percept histories to

actions: [f: P*  A]

The agent program runs on the physical architecture

Trang 5

Vacuum-cleaner world

Environment: square A and B

Percepts: location and contents, e.g., [A,Dirty]

Actions: Left, Right, Suck, NoOp

Trang 6

The vacuum-cleaner world

What is the right function? Can it be implemented in

function REFLEX-VACUUM-AGENT ([location, status]) return an action

if status == Dirty then return Suck else if location == A then return Right else if location == B then return Left

Percept sequence Action

Trang 7

Rational agents

A rational agent is one that does the right thing based

on what it can perceive and the actions it can

perform

What is the right thing?

Approximation: The right action is the one that will cause the

agent to be most successful

Measure of success?

Performance measure: An objective criterion for

success of an agent's behavior

E.g., performance measure of a vacuum-cleaner agent could be amount of dirt cleaned up, amount of time taken, amount of

electricity consumed, amount of noise generated, etc

Trang 8

Rational agents

What is rational at a given time depends on four

things:

Performance measure,

Percept sequence to date (sensors)

Prior environment knowledge,

Actions,

Rational Agent: For each possible percept sequence,

a rational agent should select an action that is

expected to maximize its performance measure,

given the evidence provided by the percept sequence and whatever built-in knowledge the agent has.

Trang 9

information gathering, exploration

An agent is autonomous if its behavior is determined

by its own experience (with ability to learn and

adapt)

Rational  exploration, learning, autonomy

Trang 10

To design a rational agent we must specify its task environment.

PEAS description of the environment:

Performance measure: Goals/desires the agent should try to

achieve

Environment: in which the agent exists

Actuators: Actions which may act the environment

Sensors: Percepts/observations of the environment

Trang 11

Environment: Roads, other traffic, pedestrians, customers

Actuators: Steering wheel, accelerator, brake, signal, horn

Sensors: Cameras, sonar, speedometer, odometer, engine

sensors, keyboard

Internet shopping

Performance measure: price, quality, appropriateness, efficiencyEnvironment: current and future WWW sites, vendors, shippersActuators: display to user, follow URL, ll inform

Sensors: HTML pages (text, graphics, scripts)

Trang 12

Agent: Medical diagnosis system

Performance measure: Healthy patient, minimize costs, lawsuitsEnvironment: Patient, hospital, staff

Actuators: Screen display (questions, tests, diagnoses,

treatments, referrals)

Sensors: Keyboard (entry of symptoms, findings, patient's

answers)

Agent: Part-picking robot

Performance measure: Percentage of parts in correct bins

Environment: Conveyor belt with parts, bins

Actuators: Jointed arm and hand

Sensors: Camera, joint angle sensors

Trang 13

Agent: Interactive English tutor

Performance measure: Maximize student's score on test

Environment: Set of students

Actuators: Screen display (exercises, suggestions, corrections)Sensors: Keyboard (typed words)

Trang 14

Environment types

Fully observable (vs partially observable):

An agent's sensors give it access to the complete state of the environment at each point in time

Deterministic (vs stochastic):

The next state of the environment is completely determined by the current state and the action executed by the agent (If the environment is deterministic except for the actions of other

agents, then the environment is strategic)

Episodic (vs sequential):

The agent's experience is divided into atomic "episodes" (each episode consists of the agent perceiving and then performing a single action), and the choice of action in each episode depends only on the episode itself

Trang 15

Environment types

Static (vs dynamic):

The environment is unchanged while an agent is deliberating (The environment is semidynamic if the environment itself does not change with the passage of time but the agent's

performance score does)

Discrete (vs continuous):

A limited number of distinct, clearly defined percepts and

actions

Single agent (vs multiagent):

An agent operating by itself in an environment

Trang 16

Environment types

Solitaire Image-Analysis

system

Intenet shopping

Trang 17

Environment types

Solitaire Image-Analysis

system

Internet shopping

Fully vs partially observable: an environment is full

observable when the sensors can detect all aspects

that are relevant to the choice of action

Trang 18

Environment types

Solitaire Image-Analysis

system

Intenet shopping

Taxi

Episodic??

Static??

Deterministic vs stochastic: if the next environment

state is completely determined by the current state the executed action then the environment is deterministic

Trang 19

Environment types

Solitaire Image-Analysis

system

Intenet shopping

Taxi

Static??

Discrete??

Episodic vs sequential : In an episodic environment the

agent’s experience can be divided into atomic steps where the agents perceives and then performs a single action The choice of action depends only on the episode itself

Trang 20

Environment types

Solitaire Image-Analysis

system

Intenet shopping

Taxi

Static vs dynamic : If the environment can change while the agent is choosing an action, the environment is dynamic Semi-dynamic if the agent’s performance changes even

when the environment remains the same

Trang 21

Environment types

Solitaire Image-Analysis

system

Intenet shopping

Taxi

Discrete vs continuous : This distinction can be applied to the state of the environment, the way time is handled and to the percepts/actions of the agent.

Trang 22

Environment types

Solitaire Image-Analysis

system

Internet shopping

Taxi

Single vs multi-agent : Does the environment contain other agents who are also maximizing some performance

measure that depends on the current agent’s actions?

Trang 23

Environment types

The environment type largely determines the agent design

The simplest environment is

Fully observable, deterministic, episodic, static, discrete and

single-agent

Most real situations are:

Partially observable, stochastic, sequential, dynamic, continuous and multi-agent

Trang 24

Agent functions and programs

How does the inside of the agent work?

Agent = architecture + program

An agent is completely specified by the agent

function mapping percept sequences to actions

One agent function (or a small equivalence class) is rational

Aim: find a way to implement the rational agent function

concisely

All agents have the same skeleton:

Input = current percepts

Output = action

Program= manipulates input to produce output

Trang 25

Even with learning, need a long time to learn the table entries

Function TABLE-DRIVEN_AGENT(percept) returns an action

static: percepts, a sequence initially empty

table, a table of actions, indexed by percept sequence

append percept to the end of percepts

action  LOOKUP(percepts, table)

return action

Trang 26

Agent types

Four basic types in order of increasing generality:

Simple reflex agents

Model-based reflex agents

Goal-based agents

Utility-based agents

All these can be turned into learning agents.

Trang 27

Agent types; simple reflex

Select action on the

basis of only the current percept.

E.g the agent

vacuum-Large reduction in possible

percept/action situations(next page) Implemented through

condition-action rules

If dirty then suck

Trang 28

The vacuum-cleaner world

function REFLEX-VACUUM-AGENT ([location, status]) return an action

if status == Dirty then return Suck

else if location == A then return Right

else if location == B then return Left

Trang 29

Agent types; simple reflex

Will only work if the environment is fully observable

otherwise infinite loops may occur.

function SIMPLE-REFLEX-AGENT(percept) returns an action

static: rules, a set of condition-action rules

state  INTERPRET-INPUT(percept) rule  RULE-MATCH(state, rule)

action  RULE-ACTION[rule]

return action

Trang 30

Agent types; reflex and state

To tackle partially observable

environments.

Maintain internal state

Over time update state using world knowledge

How does the world change

How do actions affect world.

 Model of World

Trang 31

Agent types; reflex and state

function REFLEX-AGENT-WITH-STATE(percept) returns an action

static: rules, a set of condition-action rules

state, a description of the current world state action, the most recent action.

state  UPDATE-STATE(state, action, percept)

rule  RULE-MATCH(state, rule)

action  RULE-ACTION[rule]

return action

Trang 32

Agent types; goal-based

The agent needs a goal to know which situations are

desirable.

Things become difficult when long sequences of actions are required to find the goal.

Typically investigated in search and planning

research.

Major difference: future

is taken into account

Is more flexible since

Trang 33

Agent types; utility-based

Certain goals can be reached in different ways.

Some are better, have a higher utility.

Utility function maps a (sequence of) state(s) onto a real number.

Improves on goals:

Selecting between conflicting goals Select appropriately between several goals based on likelihood of success.

Trang 34

Agent types; learning

All previous programs describe methods for selecting

agent-actions.

Yet it does not explain the origin of these programs

Learning mechanisms can be used to perform this task.

Teach them instead of instructing them.

Advantage is the robustness of the program toward initially unknown environments.

Trang 35

Agent types; learning

Trang 36

Agents interact with environments through actuators and sensorsThe agent function describes what the agent does in all

circumstances

The performance measure evaluates the environment sequence

A perfectly rational agent maximizes expected performance

Agent programs implement (some) agent functions

PEAS descriptions dene task environments

Environments are categorized along several dimensions:

observable? deterministic? episodic? static? discrete? agent?

single-Several basic agent architectures exist:

Ngày đăng: 30/01/2020, 04:07

TỪ KHÓA LIÊN QUAN