Download PDF handout of power point slides

Survey
yes no Was this document useful for you?
   Thank you for your participation!

* Your assessment is very important for improving the workof artificial intelligence, which forms the content of this project

Document related concepts

Technological singularity wikipedia , lookup

Artificial intelligence in video games wikipedia , lookup

Philosophy of artificial intelligence wikipedia , lookup

Agent (The Matrix) wikipedia , lookup

Ethics of artificial intelligence wikipedia , lookup

Intelligence explosion wikipedia , lookup

History of artificial intelligence wikipedia , lookup

Existential risk from artificial general intelligence wikipedia , lookup

Cognitive model wikipedia , lookup

Embodied cognitive science wikipedia , lookup

Transcript
Agents
sensors
percepts
Intelligent Agents
?
environment
agent
actions
R
Russell
ll and
dN
Norvig:
i 2
actuators
• Agent – perceives the environment through sensors and
acts on it through actuators
• Percept – agent’s perceptual input (the basis for its
actions)
• Percept Sequence – complete history of what has been
perceived.
CISC4/681 Introduction to Artificial Intelligence
1
Agent Function
CISC4/681 Introduction to Artificial Intelligence
2
Vacuum Cleaner World
sensors
A
percepts
B
?
environment
actions
agent
actuators
• Agent Function – maps a give percept sequence
into an action; describes what the agent does.
• Externally – Table of actions
• Internally – Agent Program
CISC4/681 Introduction to Artificial Intelligence
3
Agent Characterization
CISC4/681 Introduction to Artificial Intelligence
4
Vacuum Cleaner World
• Meant to be a tool for analyzing systems –
not characterizing them as agent versus
non-agent
• Lots of things can be characterized as
agents (artifacts) that act on the world
• AI operates where
– Artifacts have significant computational
resources
– Task Environments require nontrivial decision
making
CISC4/681 Introduction to Artificial Intelligence
• Percepts: which square (A or B); dirt?
• Actions: move right, move left, suck, do nothing
• Agent function: maps percept sequence into
actions
• Agent program: function’s implementation
• How should the program act?
5
A
B
• Percepts: which square (A or B); dirt?
• Actions: move right, move left, suck, do nothing
• Agent function: maps percept sequence into
actions
• Agent program: function’s implementation
• How should the program act?
CISC4/681 Introduction to Artificial Intelligence
6
1
Rational Agent – does the right thing
Rationality Depends on:
What does that mean? One that behaves
as well as possible given the Environment
in which it acts. How should success be
measured? On consequences.
• Performance measure
1. Performance measure that defines criterion for
success.
2. Agent’s prior knowledge of the environment.
3. Actions the agent can perform.
4 Agent
4.
Agent’s
s percept sequence to date
date.
– Embodies criterion for success
• Amount of dirt cleaned?
• Cleaned floors?
– Generally defined in terms of desired effect on
environment (not on actions of agent)
– Defining measure not always easy!
CISC4/681 Introduction to Artificial Intelligence
7
Rationality
For each possible percept sequence, a rational
agent should select an action that is expected to
maximize its performance measure, given the
evidence provided by the percept sequence and
whatever built-in knowledge the agent has.
CISC4/681 Introduction to Artificial Intelligence
Rationality
For each possible percept sequence, a
rational agent should select an action that
is expected to maximize its performance
measure, given the evidence provided by
percept
p sequence
q
and whatever builtthe p
in knowledge the agent has.
• Notice the rationality is dependent on
EXPECTED maximization.
• Agent might need to learn how the
environment changes, what action
sequences to put together, etc…
• Notice that an agent may be rational
because the designer thought of
everything, or it may have learned it itself
(more autonomous)
CISC4/681 Introduction to Artificial Intelligence
CISC4/681 Introduction to Artificial Intelligence
9
For each possible percept sequence, a
rational agent should select an action that
is expected to maximize its performance
measure, given the evidence provided by
percept
p sequence
q
and whatever builtthe p
in knowledge the agent has.
10
Properties of Task Environments
(affect appropriate agent design)
Task Environment
• The “problems” for which rational agents
are the “solutions”
PEAS Description of Task Environment
• Performance
P f
Measure
M
• Environment
• Actuators (actions)
• Sensors (what can be perceived)
CISC4/681 Introduction to Artificial Intelligence
8
• Fully observable vs partially observable
– Fully observable gives access to complete
state of the environment
– Complete state means aspects relevant to
action choice
– global vs local dirt sensor
11
CISC4/681 Introduction to Artificial Intelligence
12
2
Properties of Task Environments
(affect appropriate agent design)
• Deterministic vs Stochastic
• Single Agent vs Multi-agent
– Deterministic – next state completely
determined by current state and action
– Uncertainty may arise because of defective
actions or partially observable state (i.e.,
agent might not see everything that affects
the outcome of an action).
– Single Agent – crossword puzzle
– Multi-agent – chess, taxi driving? (are other
drivers best described as maximizing a
performance element?)
– Multi-agent means other agents may be
competitive or cooperative and may require
communication
– Multi-agent may need communication
CISC4/681 Introduction to Artificial Intelligence
13
Properties of Task Environments
(affect appropriate agent design)
14
• Static vs Dynamic
– Episodic the agent’s experience divided into
atomic episodes
– Next
N t episode
i d nott d
dependent
d t on actions
ti
ttaken
k
in previous episode. E.g., assembly line
– Sequential – current action may affect future
actions. E.g., playing chess, taxi
– short-term actions have long-term effects
– must think ahead in choosing an action
15
Properties of Task Environments
(affect appropriate agent design)
– does environment change while agent is
deliberating?
– Static
St ti – crossword
d puzzle
l
– Dynamic – taxi driver
CISC4/681 Introduction to Artificial Intelligence
16
Properties of Task Environments
(affect appropriate agent design)
• Known vs Unknown
• Discrete vs Continuous.
Can refer to
– the state of the environment ((chess has finite
number of discrete states)
– the way time is handled (taxi driving
continuous – speed and location of taxi sweep
through range of continuous values)
– percepts and actions (taxi driving continuous
– steering angles)
CISC4/681 Introduction to Artificial Intelligence
CISC4/681 Introduction to Artificial Intelligence
Properties of Task Environments
(affect appropriate agent design)
• Episodic vs Sequential
CISC4/681 Introduction to Artificial Intelligence
Properties of Task Environments
(affect appropriate agent design)
17
- This does not refer to the environment itself,
but rather the agent’s knowledge of it and how
it changes
changes.
- If unknown, the agent may need to learn
CISC4/681 Introduction to Artificial Intelligence
18
3
Properties of Task Environments
(affect appropriate agent design)
Fully observable
Deterministic
Episodic
Static
Discrete
Single agent
• Easy: Fully observable, Deterministic,
Episodic, Static, Discrete, Single agent.
• Hard: Partially observable, Sochastic,
Sequential, Dynamic, Continuous, MultiAgent
CISC4/681 Introduction to Artificial Intelligence
Environment types
Chess with
a clock
Yes
Strategic
No
Semi
Yes
No
Chess without
a clock
Yes
Strategic
No
Yes
Yes
No
Taxi driving
No
No
No
No
No
No
• The environment type largely determines the agent
design
19
Agent Programs
• The real world is (of course) partially observable,
stochastic, sequential, dynamic, continuous, multi-agent
Possible Agent Program
• Need to develop agents – programs that
take the current percept as input from the
sensors and return an action to the
actuators
actuators.
CISC4/681 Introduction to Artificial Intelligence
21
Agent Programs
22
Simple Reflective Agent
23
Agent
Sensors
What the world
is like now
Condition−action rules
What action I
should do now
Envi
vironment
• Need to develop agents – programs that
take the current percept as input from the
sensors and return an action to the
actuators
actuators.
• The key challenge for AI is to find out how
to write programs that, to the extent
possible, produce rational behavior from a
small amount of code.
CISC4/681 Introduction to Artificial Intelligence
CISC4/681 Introduction to Artificial Intelligence
Actuators
CISC4/681 Introduction to Artificial Intelligence
24
4
Simple Reflexive Agent
Simply Reflexive Vacuum Agent
•
•
•
•
Handles simplest kind of world
Agent embodies a set of condition-action rules
If percept then action
Agent simply takes in a percept
percept, determines
which action could be applied, and does that
action.
• NOTE:
• Action dependent on current percept only
• Only works in fully observable environment
CISC4/681 Introduction to Artificial Intelligence
• Implements the agent function (described
in earlier table)
25
Model-Based Reflex Agent
• Upon getting a percept
Sensors
What my actions do
Condition−action rules
Agent
What action I
should do now
– Update the state (given the current state, the
action you just did, and the observations)
– Choose a rule to apply (whose conditions
match the state)
– Schedule the action associated with the
chosen rule
Envi
vironment
What the world
is like now
26
Model-Based Reflex Agent
State
How the world evolves
CISC4/681 Introduction to Artificial Intelligence
Actuators
CISC4/681 Introduction to Artificial Intelligence
27
CISC4/681 Introduction to Artificial Intelligence
Goal Based Agent
Utility-Based Agent
Sensors
Sensors
State
State
What it will be like
if I do action A
Agent
CISC4/681 Introduction to Artificial Intelligence
What action I
should do now
How the world evolves
What the world
is like now
What my actions do
What it will be like
if I do action A
Utility
How happy I will be
in such a state
What action I
should do now
Agent
Actuators
29
CISC4/681 Introduction to Artificial Intelligence
Envi
vironment
What my actions do
Envi
vironment
How the world evolves
What the world
is like now
Goals
28
Actuators
30
5
Learning Agent Components
Learning agents
1. Learning Element – responsible for making
improvements (on what ever aspect is being
learned…)
2. Performance Element – responsible for
selecting external actions. In previous parts,
this was the entire agent!
3. Critic – gives feedback on how agent is going
and determines how performance element
should be modified to do better in the future
4. Problem Generator – suggests actions for new
and informative experiences.
CISC4/681 Introduction to Artificial Intelligence
Summary Chapter 2
Summary (cont)
• Agents interact with environments through
actuators and sensors
• The agent function describes what the
agent does in all circumstances.
• The
Th performance
f
measure evaluates
l
the
h
environment sequence.
• A perfectly rational agent maximizes
expected performance.
• Agent programs implement (some) agent
functions.
CISC4/681 Introduction to Artificial Intelligence
32
33
• PEAS descriptions define task
environments.
• Environments are categorized along
several dimensions:
– Observable? Deterministic? Episodic? Static?
Discrete? Single-agent?
• Several basic agent architectures exist:
– Reflex, reflex with state, goal-based, utilitybased
CISC4/681 Introduction to Artificial Intelligence
34
6