SlideShare une entreprise Scribd logo
1  sur  39
Artificial
Intelligence I :
Intelligent Agents
Lecturer: Tom Lenaerts
Institut de Recherches Interdisciplinaires et de
Développements en Intelligence Artificielle
(IRIDIA)
Université Libre de Bruxelles
TLo (IRIDIA) 2October 13, 2015
Outline
 Agents and environments.
 The vacuum-cleaner world
 The concept of rational behavior.
 Environments.
 Agent structure.
TLo (IRIDIA) 3October 13, 2015
Agents and environments
 Agents include human, robots,
softbots, thermostats, etc.
 The agent function maps
percept sequence to actions
 An agent can perceive its own
actions, but not always it
effects.
f : P* → A
TLo (IRIDIA) 4October 13, 2015
Agents and environments
 The agent function will
internally be represented by
the agent program.
 The agent program runs on the
physical architecture to
produce f.
TLo (IRIDIA) 5October 13, 2015
The vacuum-cleaner world
 Environment: square A and B
 Percepts: [location and content] e.g. [A, Dirty]
 Actions: left, right, suck, and no-op
TLo (IRIDIA) 6October 13, 2015
The vacuum-cleaner world
Percept sequence Action
[A,Clean] Right
[A, Dirty] Suck
[B, Clean] Left
[B, Dirty] Suck
[A, Clean],[A, Clean] Right
[A, Clean],[A, Dirty] Suck
… …
TLo (IRIDIA) 7October 13, 2015
The vacuum-cleaner world
function REFLEX-VACUUM-AGENT ([location, status]) return an action
if status == Dirty then return Suck
else if location == A then return Right
else if location == B then return Left
What is the right function? Can it be implemented in a small agent
program?
TLo (IRIDIA) 8October 13, 2015
The concept of rationality
 A rational agent is one that does the right thing.
 Every entry in the table is filled out correctly.
 What is the right thing?
 Approximation: the most succesfull agent.
 Measure of success?
 Performance measure should be objective
 E.g. the amount of dirt cleaned within a certain time.
 E.g. how clean the floor is.
 …
 Performance measure according to what is wanted in the
environment instead of how the agents should behave.
TLo (IRIDIA) 9October 13, 2015
Rationality
 What is rational at a given time depends on four things:
 Performance measure,
 Prior environment knowledge,
 Actions,
 Percept sequence to date (sensors).
 DEF: A rational agent chooses whichever action
maximizes the expected value of the performance measure
given the percept sequence to date and prior environment
knowledge.
TLo (IRIDIA) 10October 13, 2015
Rationality
 Rationality ≠ omniscience
 An omniscient agent knows the actual outcome of
its actions.
 Rationality ≠ perfection
 Rationality maximizes expected performance,
while perfection maximizes actual performance.
TLo (IRIDIA) 11October 13, 2015
Rationality
 The proposed definition requires:
 Information gathering/exploration
 To maximize future rewards
 Learn from percepts
 Extending prior knowledge
 Agent autonomy
 Compensate for incorrect prior knowledge
TLo (IRIDIA) 12October 13, 2015
Environments
 To design a rational agent we must specify its task
environment.
 PEAS description of the environment:
 Performance
 Environment
 Actuators
 Sensors
TLo (IRIDIA) 13October 13, 2015
Environments
 E.g. Fully automated taxi:
 PEAS description of the environment:
 Performance
 Safety, destination, profits, legality, comfort
 Environment
 Streets/freeways, other traffic, pedestrians, weather,, …
 Actuators
 Steering, accelerating, brake, horn, speaker/display,…
 Sensors
 Video, sonar, speedometer, engine sensors, keyboard, GPS, …
TLo (IRIDIA) 14October 13, 2015
Environment types
Solitaire Backgammom Intenet shopping Taxi
Observable??
Deterministic??
Episodic??
Static??
Discrete??
Single-agent??
TLo (IRIDIA) 15October 13, 2015
Environment types
Solitaire Backgammom Intenet shopping Taxi
Observable??
Deterministic??
Episodic??
Static??
Discrete??
Single-agent??
Fully vs. partially observable: an environment is full observable when the
sensors can detect all aspects that are relevant to the choice of action.
TLo (IRIDIA) 16October 13, 2015
Environment types
Solitaire Backgammom Intenet shopping Taxi
Observable?? FULL FULL PARTIAL PARTIAL
Deterministic??
Episodic??
Static??
Discrete??
Single-agent??
Fully vs. partially observable: an environment is full observable when the
sensors can detect all aspects that are relevant to the choice of action.
TLo (IRIDIA) 17October 13, 2015
Environment types
Solitaire Backgammom Intenet shopping Taxi
Observable?? FULL FULL PARTIAL PARTIAL
Deterministic??
Episodic??
Static??
Discrete??
Single-agent??
Deterministic vs. stochastic: if the next environment state is completely
determined by the current state the executed action then the environment is
deterministic.
TLo (IRIDIA) 18October 13, 2015
Environment types
Solitaire Backgammom Intenet shopping Taxi
Observable?? FULL FULL PARTIAL PARTIAL
Deterministic?? YES NO YES NO
Episodic??
Static??
Discrete??
Single-agent??
Deterministic vs. stochastic: if the next environment state is completely
determined by the current state the executed action then the environment is
deterministic.
TLo (IRIDIA) 19October 13, 2015
Environment types
Solitaire Backgammom Intenet shopping Taxi
Observable?? FULL FULL PARTIAL PARTIAL
Deterministic?? YES NO YES NO
Episodic??
Static??
Discrete??
Single-agent??
Episodic vs. sequential: In an episodic environment the agent’s experience
can be divided into atomic steps where the agents perceives and then performs
A single action. The choice of action depends only on the episode itself
TLo (IRIDIA) 20October 13, 2015
Environment types
Solitaire Backgammom Intenet shopping Taxi
Observable?? FULL FULL PARTIAL PARTIAL
Deterministic?? YES NO YES NO
Episodic?? NO NO NO NO
Static??
Discrete??
Single-agent??
Episodic vs. sequential: In an episodic environment the agent’s experience
can be divided into atomic steps where the agents perceives and then performs
A single action. The choice of action depends only on the episode itself
TLo (IRIDIA) 21October 13, 2015
Environment types
Solitaire Backgammom Intenet shopping Taxi
Observable?? FULL FULL PARTIAL PARTIAL
Deterministic?? YES NO YES NO
Episodic?? NO NO NO NO
Static??
Discrete??
Single-agent??
Static vs. dynamic: If the environment can change while the agent is choosing
an action, the environment is dynamic. Semi-dynamic if the agent’s performance
changes even when the environment remains the same.
TLo (IRIDIA) 22October 13, 2015
Environment types
Solitaire Backgammom Intenet shopping Taxi
Observable?? FULL FULL PARTIAL PARTIAL
Deterministic?? YES NO YES NO
Episodic?? NO NO NO NO
Static?? YES YES SEMI NO
Discrete??
Single-agent??
Static vs. dynamic: If the environment can change while the agent is choosing
an action, the environment is dynamic. Semi-dynamic if the agent’s performance
changes even when the environment remains the same.
TLo (IRIDIA) 23October 13, 2015
Environment types
Solitaire Backgammom Intenet shopping Taxi
Observable?? FULL FULL PARTIAL PARTIAL
Deterministic?? YES NO YES NO
Episodic?? NO NO NO NO
Static?? YES YES SEMI NO
Discrete??
Single-agent??
Discrete vs. continuous: This distinction can be applied to the state of the
environment, the way time is handled and to the percepts/actions of the agent.
TLo (IRIDIA) 24October 13, 2015
Environment types
Solitaire Backgammom Intenet shopping Taxi
Observable?? FULL FULL PARTIAL PARTIAL
Deterministic?? YES NO YES NO
Episodic?? NO NO NO NO
Static?? YES YES SEMI NO
Discrete?? YES YES YES NO
Single-agent??
Discrete vs. continuous: This distinction can be applied to the state of the
environment, the way time is handled and to the percepts/actions of the agent.
TLo (IRIDIA) 25October 13, 2015
Environment types
Solitaire Backgammom Intenet shopping Taxi
Observable?? FULL FULL PARTIAL PARTIAL
Deterministic?? YES NO YES NO
Episodic?? NO NO NO NO
Static?? YES YES SEMI NO
Discrete?? YES YES YES NO
Single-agent??
Single vs. multi-agent: Does the environment contain other agents who
are also maximizing some performance measure that depends on the
current agent’s actions?
TLo (IRIDIA) 26October 13, 2015
Environment types
Solitaire Backgammom Intenet shopping Taxi
Observable?? FULL FULL PARTIAL PARTIAL
Deterministic?? YES NO YES NO
Episodic?? NO NO NO NO
Static?? YES YES SEMI NO
Discrete?? YES YES YES NO
Single-agent?? YES NO NO NO
Single vs. multi-agent: Does the environment contain other agents who
are also maximizing some performance measure that depends on the
current agent’s actions?
TLo (IRIDIA) 27October 13, 2015
Environment types
 The simplest environment is
 Fully observable, deterministic, episodic, static,
discrete and single-agent.
 Most real situations are:
 Partially observable, stochastic, sequential,
dynamic, continuous and multi-agent.
TLo (IRIDIA) 28October 13, 2015
Agent types
 How does the inside of the agent work?
 Agent = architecture + program
 All agents have the same skeleton:
 Input = current percepts
 Output = action
 Program= manipulates input to produce output
 Note difference with agent function.
TLo (IRIDIA) 29October 13, 2015
Agent types
Function TABLE-DRIVEN_AGENT(percept) returns an action
static: percepts, a sequence initially empty
table, a table of actions, indexed by percept sequence
append percept to the end of percepts
action ← LOOKUP(percepts, table)
return action
This approach is doomed to failure
TLo (IRIDIA) 30October 13, 2015
Agent types
 Four basic kind of agent programs will be
discussed:
 Simple reflex agents
 Model-based reflex agents
 Goal-based agents
 Utility-based agents
 All these can be turned into learning agents.
TLo (IRIDIA) 31October 13, 2015
Agent types; simple reflex
 Select action on the basis
of only the current
percept.
 E.g. the vacuum-agent
 Large reduction in
possible percept/action
situations(next page).
 Implemented through
condition-action rules
 If dirty then suck
TLo (IRIDIA) 32October 13, 2015
The vacuum-cleaner world
function REFLEX-VACUUM-AGENT ([location, status]) return an action
if status == Dirty then return Suck
else if location == A then return Right
else if location == B then return Left
Reduction from 4T
to 4 entries
TLo (IRIDIA) 33October 13, 2015
Agent types; simple reflex
function SIMPLE-REFLEX-AGENT(percept) returns an action
static: rules, a set of condition-action rules
state ← INTERPRET-INPUT(percept)
rule ← RULE-MATCH(state, rule)
action ← RULE-ACTION[rule]
return action
Will only work if the environment is fully observable otherwise
infinite loops may occur.
TLo (IRIDIA) 34October 13, 2015
Agent types; reflex and state
 To tackle partially
observable environments.
 Maintain internal state
 Over time update state
using world knowledge
 How does the world
change.
 How do actions affect
world.
⇒ Model of World
TLo (IRIDIA) 35October 13, 2015
Agent types; reflex and state
function REFLEX-AGENT-WITH-STATE(percept) returns an action
static: rules, a set of condition-action rules
state, a description of the current world state
action, the most recent action.
state ← UPDATE-STATE(state, action, percept)
rule ← RULE-MATCH(state, rule)
action ← RULE-ACTION[rule]
return action
TLo (IRIDIA) 36October 13, 2015
Agent types; goal-based
 The agent needs a goal to know
which situations are desirable.
 Things become difficult when
long sequences of actions are
required to find the goal.
 Typically investigated in search
and planning research.
 Major difference: future is taken
into account
 Is more flexible since knowledge
is represented explicitly and can
be manipulated.
TLo (IRIDIA) 37October 13, 2015
Agent types; utility-based
 Certain goals can be reached in
different ways.
 Some are better, have a
higher utility.
 Utility function maps a
(sequence of) state(s) onto a
real number.
 Improves on goals:
 Selecting between
conflicting goals
 Select appropriately
between several goals
based on likelihood of
success.
TLo (IRIDIA) 38October 13, 2015
Agent types; learning
 All previous agent-programs
describe methods for selecting
actions.
 Yet it does not explain the
origin of these programs.
 Learning mechanisms can
be used to perform this
task.
 Teach them instead of
instructing them.
 Advantage is the
robustness of the program
toward initially unknown
environments.
TLo (IRIDIA) 39October 13, 2015
Agent types; learning
 Learning element: introduce
improvements in performance
element.
 Critic provides feedback on
agents performance based on
fixed performance standard.
 Performance element: selecting
actions based on percepts.
 Corresponds to the previous
agent programs
 Problem generator: suggests
actions that will lead to new and
informative experiences.
 Exploration vs. exploitation

Contenu connexe

Tendances

Tendances (20)

Intelligent agent
Intelligent agentIntelligent agent
Intelligent agent
 
State Space Search in ai
State Space Search in aiState Space Search in ai
State Space Search in ai
 
Artificial Intelligence Searching Techniques
Artificial Intelligence Searching TechniquesArtificial Intelligence Searching Techniques
Artificial Intelligence Searching Techniques
 
AI
AIAI
AI
 
Agents in Artificial intelligence
Agents in Artificial intelligence Agents in Artificial intelligence
Agents in Artificial intelligence
 
AI Lecture 3 (solving problems by searching)
AI Lecture 3 (solving problems by searching)AI Lecture 3 (solving problems by searching)
AI Lecture 3 (solving problems by searching)
 
The structure of agents
The structure of agentsThe structure of agents
The structure of agents
 
Informed and Uninformed search Strategies
Informed and Uninformed search StrategiesInformed and Uninformed search Strategies
Informed and Uninformed search Strategies
 
Problem Solving
Problem Solving Problem Solving
Problem Solving
 
Planning
PlanningPlanning
Planning
 
Unit3:Informed and Uninformed search
Unit3:Informed and Uninformed searchUnit3:Informed and Uninformed search
Unit3:Informed and Uninformed search
 
Planning
Planning Planning
Planning
 
Structure of agents
Structure of agentsStructure of agents
Structure of agents
 
AI Lecture 4 (informed search and exploration)
AI Lecture 4 (informed search and exploration)AI Lecture 4 (informed search and exploration)
AI Lecture 4 (informed search and exploration)
 
Artificial Intelligence -- Search Algorithms
Artificial Intelligence-- Search Algorithms Artificial Intelligence-- Search Algorithms
Artificial Intelligence -- Search Algorithms
 
State space search
State space searchState space search
State space search
 
Informed search
Informed searchInformed search
Informed search
 
Intelligent Agents
Intelligent Agents Intelligent Agents
Intelligent Agents
 
Problem solving agents
Problem solving agentsProblem solving agents
Problem solving agents
 
AI search techniques
AI search techniquesAI search techniques
AI search techniques
 

En vedette (12)

3 problem-solving-
3 problem-solving-3 problem-solving-
3 problem-solving-
 
AI: Introduction to artificial intelligence
AI: Introduction to artificial intelligenceAI: Introduction to artificial intelligence
AI: Introduction to artificial intelligence
 
Lecture 4- Agent types
Lecture 4- Agent typesLecture 4- Agent types
Lecture 4- Agent types
 
Nonparametric statistics
Nonparametric statisticsNonparametric statistics
Nonparametric statistics
 
non parametric statistics
non parametric statisticsnon parametric statistics
non parametric statistics
 
AI: Planning and AI
AI: Planning and AIAI: Planning and AI
AI: Planning and AI
 
AI: Learning in AI
AI: Learning in AI AI: Learning in AI
AI: Learning in AI
 
Non-Parametric Tests
Non-Parametric TestsNon-Parametric Tests
Non-Parametric Tests
 
Non parametric tests
Non parametric testsNon parametric tests
Non parametric tests
 
Artificial Intelligence Presentation
Artificial Intelligence PresentationArtificial Intelligence Presentation
Artificial Intelligence Presentation
 
Artificial Intelligence
Artificial IntelligenceArtificial Intelligence
Artificial Intelligence
 
Artificial Intelligence Presentation
Artificial Intelligence PresentationArtificial Intelligence Presentation
Artificial Intelligence Presentation
 

Similaire à 2-Agents- Artificial Intelligence

Introduction To Artificial Intelligence
Introduction To Artificial IntelligenceIntroduction To Artificial Intelligence
Introduction To Artificial Intelligence
NeHal VeRma
 
ai-slides-1233566181695672-2 (1).pdf
ai-slides-1233566181695672-2 (1).pdfai-slides-1233566181695672-2 (1).pdf
ai-slides-1233566181695672-2 (1).pdf
ShivareddyGangam
 
topic-2intelligentagents-190426041219.pdf
topic-2intelligentagents-190426041219.pdftopic-2intelligentagents-190426041219.pdf
topic-2intelligentagents-190426041219.pdf
ShivareddyGangam
 
introduction to inteligent IntelligentAgent.ppt
introduction to inteligent IntelligentAgent.pptintroduction to inteligent IntelligentAgent.ppt
introduction to inteligent IntelligentAgent.ppt
dejene3
 
introduction to machine learning
introduction to machine learningintroduction to machine learning
introduction to machine learning
colleges
 
Jarrar.lecture notes.aai.2011s.ch2.intelligentagents
Jarrar.lecture notes.aai.2011s.ch2.intelligentagentsJarrar.lecture notes.aai.2011s.ch2.intelligentagents
Jarrar.lecture notes.aai.2011s.ch2.intelligentagents
PalGov
 
Jarrar.lecture notes.aai.2011s.ch2.intelligentagents
Jarrar.lecture notes.aai.2011s.ch2.intelligentagentsJarrar.lecture notes.aai.2011s.ch2.intelligentagents
Jarrar.lecture notes.aai.2011s.ch2.intelligentagents
PalGov
 

Similaire à 2-Agents- Artificial Intelligence (20)

agents in ai ppt
agents in ai pptagents in ai ppt
agents in ai ppt
 
Introduction To Artificial Intelligence
Introduction To Artificial IntelligenceIntroduction To Artificial Intelligence
Introduction To Artificial Intelligence
 
Introduction To Artificial Intelligence
Introduction To Artificial IntelligenceIntroduction To Artificial Intelligence
Introduction To Artificial Intelligence
 
AI Lecture 2 (intelligent agents)
AI Lecture 2 (intelligent agents)AI Lecture 2 (intelligent agents)
AI Lecture 2 (intelligent agents)
 
Artificial Intelligent Agents
Artificial Intelligent AgentsArtificial Intelligent Agents
Artificial Intelligent Agents
 
CoThink RATIO Approach and Tripod Beta
CoThink RATIO Approach and Tripod BetaCoThink RATIO Approach and Tripod Beta
CoThink RATIO Approach and Tripod Beta
 
chapter2.pptx
chapter2.pptxchapter2.pptx
chapter2.pptx
 
Unit 1.ppt
Unit 1.pptUnit 1.ppt
Unit 1.ppt
 
AI_Ch2.pptx
AI_Ch2.pptxAI_Ch2.pptx
AI_Ch2.pptx
 
ai-slides-1233566181695672-2 (1).pdf
ai-slides-1233566181695672-2 (1).pdfai-slides-1233566181695672-2 (1).pdf
ai-slides-1233566181695672-2 (1).pdf
 
Lecture 4 (1).pptx
Lecture 4 (1).pptxLecture 4 (1).pptx
Lecture 4 (1).pptx
 
topic-2intelligentagents-190426041219.pdf
topic-2intelligentagents-190426041219.pdftopic-2intelligentagents-190426041219.pdf
topic-2intelligentagents-190426041219.pdf
 
introduction to inteligent IntelligentAgent.ppt
introduction to inteligent IntelligentAgent.pptintroduction to inteligent IntelligentAgent.ppt
introduction to inteligent IntelligentAgent.ppt
 
Types of environment
Types of environmentTypes of environment
Types of environment
 
introduction to machine learning
introduction to machine learningintroduction to machine learning
introduction to machine learning
 
Artificial intelligence introduction
Artificial intelligence introductionArtificial intelligence introduction
Artificial intelligence introduction
 
chapter2-(Intelligent Agents).ppt
chapter2-(Intelligent Agents).pptchapter2-(Intelligent Agents).ppt
chapter2-(Intelligent Agents).ppt
 
AI Agents, Agents in Artificial Intelligence
AI Agents, Agents in Artificial IntelligenceAI Agents, Agents in Artificial Intelligence
AI Agents, Agents in Artificial Intelligence
 
Jarrar.lecture notes.aai.2011s.ch2.intelligentagents
Jarrar.lecture notes.aai.2011s.ch2.intelligentagentsJarrar.lecture notes.aai.2011s.ch2.intelligentagents
Jarrar.lecture notes.aai.2011s.ch2.intelligentagents
 
Jarrar.lecture notes.aai.2011s.ch2.intelligentagents
Jarrar.lecture notes.aai.2011s.ch2.intelligentagentsJarrar.lecture notes.aai.2011s.ch2.intelligentagents
Jarrar.lecture notes.aai.2011s.ch2.intelligentagents
 

Plus de Mhd Sb (10)

AI ch6
AI ch6AI ch6
AI ch6
 
AI ch5
AI ch5AI ch5
AI ch5
 
AI ch4
AI ch4AI ch4
AI ch4
 
AI ch3
AI ch3AI ch3
AI ch3
 
Ai ch2
Ai ch2Ai ch2
Ai ch2
 
Ai ch1
Ai ch1Ai ch1
Ai ch1
 
6 games
6 games6 games
6 games
 
5 csp
5 csp5 csp
5 csp
 
4 informed-search
4 informed-search4 informed-search
4 informed-search
 
Artificial Intelligence
Artificial IntelligenceArtificial Intelligence
Artificial Intelligence
 

Dernier

AKTU Computer Networks notes --- Unit 3.pdf
AKTU Computer Networks notes ---  Unit 3.pdfAKTU Computer Networks notes ---  Unit 3.pdf
AKTU Computer Networks notes --- Unit 3.pdf
ankushspencer015
 
VIP Call Girls Ankleshwar 7001035870 Whatsapp Number, 24/07 Booking
VIP Call Girls Ankleshwar 7001035870 Whatsapp Number, 24/07 BookingVIP Call Girls Ankleshwar 7001035870 Whatsapp Number, 24/07 Booking
VIP Call Girls Ankleshwar 7001035870 Whatsapp Number, 24/07 Booking
dharasingh5698
 
UNIT-V FMM.HYDRAULIC TURBINE - Construction and working
UNIT-V FMM.HYDRAULIC TURBINE - Construction and workingUNIT-V FMM.HYDRAULIC TURBINE - Construction and working
UNIT-V FMM.HYDRAULIC TURBINE - Construction and working
rknatarajan
 

Dernier (20)

AKTU Computer Networks notes --- Unit 3.pdf
AKTU Computer Networks notes ---  Unit 3.pdfAKTU Computer Networks notes ---  Unit 3.pdf
AKTU Computer Networks notes --- Unit 3.pdf
 
data_management_and _data_science_cheat_sheet.pdf
data_management_and _data_science_cheat_sheet.pdfdata_management_and _data_science_cheat_sheet.pdf
data_management_and _data_science_cheat_sheet.pdf
 
BSides Seattle 2024 - Stopping Ethan Hunt From Taking Your Data.pptx
BSides Seattle 2024 - Stopping Ethan Hunt From Taking Your Data.pptxBSides Seattle 2024 - Stopping Ethan Hunt From Taking Your Data.pptx
BSides Seattle 2024 - Stopping Ethan Hunt From Taking Your Data.pptx
 
The Most Attractive Pune Call Girls Manchar 8250192130 Will You Miss This Cha...
The Most Attractive Pune Call Girls Manchar 8250192130 Will You Miss This Cha...The Most Attractive Pune Call Girls Manchar 8250192130 Will You Miss This Cha...
The Most Attractive Pune Call Girls Manchar 8250192130 Will You Miss This Cha...
 
CCS335 _ Neural Networks and Deep Learning Laboratory_Lab Complete Record
CCS335 _ Neural Networks and Deep Learning Laboratory_Lab Complete RecordCCS335 _ Neural Networks and Deep Learning Laboratory_Lab Complete Record
CCS335 _ Neural Networks and Deep Learning Laboratory_Lab Complete Record
 
UNIT-III FMM. DIMENSIONAL ANALYSIS
UNIT-III FMM.        DIMENSIONAL ANALYSISUNIT-III FMM.        DIMENSIONAL ANALYSIS
UNIT-III FMM. DIMENSIONAL ANALYSIS
 
Booking open Available Pune Call Girls Pargaon 6297143586 Call Hot Indian Gi...
Booking open Available Pune Call Girls Pargaon  6297143586 Call Hot Indian Gi...Booking open Available Pune Call Girls Pargaon  6297143586 Call Hot Indian Gi...
Booking open Available Pune Call Girls Pargaon 6297143586 Call Hot Indian Gi...
 
Coefficient of Thermal Expansion and their Importance.pptx
Coefficient of Thermal Expansion and their Importance.pptxCoefficient of Thermal Expansion and their Importance.pptx
Coefficient of Thermal Expansion and their Importance.pptx
 
VIP Call Girls Ankleshwar 7001035870 Whatsapp Number, 24/07 Booking
VIP Call Girls Ankleshwar 7001035870 Whatsapp Number, 24/07 BookingVIP Call Girls Ankleshwar 7001035870 Whatsapp Number, 24/07 Booking
VIP Call Girls Ankleshwar 7001035870 Whatsapp Number, 24/07 Booking
 
Thermal Engineering Unit - I & II . ppt
Thermal Engineering  Unit - I & II . pptThermal Engineering  Unit - I & II . ppt
Thermal Engineering Unit - I & II . ppt
 
chapter 5.pptx: drainage and irrigation engineering
chapter 5.pptx: drainage and irrigation engineeringchapter 5.pptx: drainage and irrigation engineering
chapter 5.pptx: drainage and irrigation engineering
 
Call Girls Walvekar Nagar Call Me 7737669865 Budget Friendly No Advance Booking
Call Girls Walvekar Nagar Call Me 7737669865 Budget Friendly No Advance BookingCall Girls Walvekar Nagar Call Me 7737669865 Budget Friendly No Advance Booking
Call Girls Walvekar Nagar Call Me 7737669865 Budget Friendly No Advance Booking
 
Java Programming :Event Handling(Types of Events)
Java Programming :Event Handling(Types of Events)Java Programming :Event Handling(Types of Events)
Java Programming :Event Handling(Types of Events)
 
Call for Papers - International Journal of Intelligent Systems and Applicatio...
Call for Papers - International Journal of Intelligent Systems and Applicatio...Call for Papers - International Journal of Intelligent Systems and Applicatio...
Call for Papers - International Journal of Intelligent Systems and Applicatio...
 
Intze Overhead Water Tank Design by Working Stress - IS Method.pdf
Intze Overhead Water Tank  Design by Working Stress - IS Method.pdfIntze Overhead Water Tank  Design by Working Stress - IS Method.pdf
Intze Overhead Water Tank Design by Working Stress - IS Method.pdf
 
NFPA 5000 2024 standard .
NFPA 5000 2024 standard                                  .NFPA 5000 2024 standard                                  .
NFPA 5000 2024 standard .
 
UNIT-V FMM.HYDRAULIC TURBINE - Construction and working
UNIT-V FMM.HYDRAULIC TURBINE - Construction and workingUNIT-V FMM.HYDRAULIC TURBINE - Construction and working
UNIT-V FMM.HYDRAULIC TURBINE - Construction and working
 
Call Girls Pimpri Chinchwad Call Me 7737669865 Budget Friendly No Advance Boo...
Call Girls Pimpri Chinchwad Call Me 7737669865 Budget Friendly No Advance Boo...Call Girls Pimpri Chinchwad Call Me 7737669865 Budget Friendly No Advance Boo...
Call Girls Pimpri Chinchwad Call Me 7737669865 Budget Friendly No Advance Boo...
 
Online banking management system project.pdf
Online banking management system project.pdfOnline banking management system project.pdf
Online banking management system project.pdf
 
Booking open Available Pune Call Girls Koregaon Park 6297143586 Call Hot Ind...
Booking open Available Pune Call Girls Koregaon Park  6297143586 Call Hot Ind...Booking open Available Pune Call Girls Koregaon Park  6297143586 Call Hot Ind...
Booking open Available Pune Call Girls Koregaon Park 6297143586 Call Hot Ind...
 

2-Agents- Artificial Intelligence

  • 1. Artificial Intelligence I : Intelligent Agents Lecturer: Tom Lenaerts Institut de Recherches Interdisciplinaires et de Développements en Intelligence Artificielle (IRIDIA) Université Libre de Bruxelles
  • 2. TLo (IRIDIA) 2October 13, 2015 Outline  Agents and environments.  The vacuum-cleaner world  The concept of rational behavior.  Environments.  Agent structure.
  • 3. TLo (IRIDIA) 3October 13, 2015 Agents and environments  Agents include human, robots, softbots, thermostats, etc.  The agent function maps percept sequence to actions  An agent can perceive its own actions, but not always it effects. f : P* → A
  • 4. TLo (IRIDIA) 4October 13, 2015 Agents and environments  The agent function will internally be represented by the agent program.  The agent program runs on the physical architecture to produce f.
  • 5. TLo (IRIDIA) 5October 13, 2015 The vacuum-cleaner world  Environment: square A and B  Percepts: [location and content] e.g. [A, Dirty]  Actions: left, right, suck, and no-op
  • 6. TLo (IRIDIA) 6October 13, 2015 The vacuum-cleaner world Percept sequence Action [A,Clean] Right [A, Dirty] Suck [B, Clean] Left [B, Dirty] Suck [A, Clean],[A, Clean] Right [A, Clean],[A, Dirty] Suck … …
  • 7. TLo (IRIDIA) 7October 13, 2015 The vacuum-cleaner world function REFLEX-VACUUM-AGENT ([location, status]) return an action if status == Dirty then return Suck else if location == A then return Right else if location == B then return Left What is the right function? Can it be implemented in a small agent program?
  • 8. TLo (IRIDIA) 8October 13, 2015 The concept of rationality  A rational agent is one that does the right thing.  Every entry in the table is filled out correctly.  What is the right thing?  Approximation: the most succesfull agent.  Measure of success?  Performance measure should be objective  E.g. the amount of dirt cleaned within a certain time.  E.g. how clean the floor is.  …  Performance measure according to what is wanted in the environment instead of how the agents should behave.
  • 9. TLo (IRIDIA) 9October 13, 2015 Rationality  What is rational at a given time depends on four things:  Performance measure,  Prior environment knowledge,  Actions,  Percept sequence to date (sensors).  DEF: A rational agent chooses whichever action maximizes the expected value of the performance measure given the percept sequence to date and prior environment knowledge.
  • 10. TLo (IRIDIA) 10October 13, 2015 Rationality  Rationality ≠ omniscience  An omniscient agent knows the actual outcome of its actions.  Rationality ≠ perfection  Rationality maximizes expected performance, while perfection maximizes actual performance.
  • 11. TLo (IRIDIA) 11October 13, 2015 Rationality  The proposed definition requires:  Information gathering/exploration  To maximize future rewards  Learn from percepts  Extending prior knowledge  Agent autonomy  Compensate for incorrect prior knowledge
  • 12. TLo (IRIDIA) 12October 13, 2015 Environments  To design a rational agent we must specify its task environment.  PEAS description of the environment:  Performance  Environment  Actuators  Sensors
  • 13. TLo (IRIDIA) 13October 13, 2015 Environments  E.g. Fully automated taxi:  PEAS description of the environment:  Performance  Safety, destination, profits, legality, comfort  Environment  Streets/freeways, other traffic, pedestrians, weather,, …  Actuators  Steering, accelerating, brake, horn, speaker/display,…  Sensors  Video, sonar, speedometer, engine sensors, keyboard, GPS, …
  • 14. TLo (IRIDIA) 14October 13, 2015 Environment types Solitaire Backgammom Intenet shopping Taxi Observable?? Deterministic?? Episodic?? Static?? Discrete?? Single-agent??
  • 15. TLo (IRIDIA) 15October 13, 2015 Environment types Solitaire Backgammom Intenet shopping Taxi Observable?? Deterministic?? Episodic?? Static?? Discrete?? Single-agent?? Fully vs. partially observable: an environment is full observable when the sensors can detect all aspects that are relevant to the choice of action.
  • 16. TLo (IRIDIA) 16October 13, 2015 Environment types Solitaire Backgammom Intenet shopping Taxi Observable?? FULL FULL PARTIAL PARTIAL Deterministic?? Episodic?? Static?? Discrete?? Single-agent?? Fully vs. partially observable: an environment is full observable when the sensors can detect all aspects that are relevant to the choice of action.
  • 17. TLo (IRIDIA) 17October 13, 2015 Environment types Solitaire Backgammom Intenet shopping Taxi Observable?? FULL FULL PARTIAL PARTIAL Deterministic?? Episodic?? Static?? Discrete?? Single-agent?? Deterministic vs. stochastic: if the next environment state is completely determined by the current state the executed action then the environment is deterministic.
  • 18. TLo (IRIDIA) 18October 13, 2015 Environment types Solitaire Backgammom Intenet shopping Taxi Observable?? FULL FULL PARTIAL PARTIAL Deterministic?? YES NO YES NO Episodic?? Static?? Discrete?? Single-agent?? Deterministic vs. stochastic: if the next environment state is completely determined by the current state the executed action then the environment is deterministic.
  • 19. TLo (IRIDIA) 19October 13, 2015 Environment types Solitaire Backgammom Intenet shopping Taxi Observable?? FULL FULL PARTIAL PARTIAL Deterministic?? YES NO YES NO Episodic?? Static?? Discrete?? Single-agent?? Episodic vs. sequential: In an episodic environment the agent’s experience can be divided into atomic steps where the agents perceives and then performs A single action. The choice of action depends only on the episode itself
  • 20. TLo (IRIDIA) 20October 13, 2015 Environment types Solitaire Backgammom Intenet shopping Taxi Observable?? FULL FULL PARTIAL PARTIAL Deterministic?? YES NO YES NO Episodic?? NO NO NO NO Static?? Discrete?? Single-agent?? Episodic vs. sequential: In an episodic environment the agent’s experience can be divided into atomic steps where the agents perceives and then performs A single action. The choice of action depends only on the episode itself
  • 21. TLo (IRIDIA) 21October 13, 2015 Environment types Solitaire Backgammom Intenet shopping Taxi Observable?? FULL FULL PARTIAL PARTIAL Deterministic?? YES NO YES NO Episodic?? NO NO NO NO Static?? Discrete?? Single-agent?? Static vs. dynamic: If the environment can change while the agent is choosing an action, the environment is dynamic. Semi-dynamic if the agent’s performance changes even when the environment remains the same.
  • 22. TLo (IRIDIA) 22October 13, 2015 Environment types Solitaire Backgammom Intenet shopping Taxi Observable?? FULL FULL PARTIAL PARTIAL Deterministic?? YES NO YES NO Episodic?? NO NO NO NO Static?? YES YES SEMI NO Discrete?? Single-agent?? Static vs. dynamic: If the environment can change while the agent is choosing an action, the environment is dynamic. Semi-dynamic if the agent’s performance changes even when the environment remains the same.
  • 23. TLo (IRIDIA) 23October 13, 2015 Environment types Solitaire Backgammom Intenet shopping Taxi Observable?? FULL FULL PARTIAL PARTIAL Deterministic?? YES NO YES NO Episodic?? NO NO NO NO Static?? YES YES SEMI NO Discrete?? Single-agent?? Discrete vs. continuous: This distinction can be applied to the state of the environment, the way time is handled and to the percepts/actions of the agent.
  • 24. TLo (IRIDIA) 24October 13, 2015 Environment types Solitaire Backgammom Intenet shopping Taxi Observable?? FULL FULL PARTIAL PARTIAL Deterministic?? YES NO YES NO Episodic?? NO NO NO NO Static?? YES YES SEMI NO Discrete?? YES YES YES NO Single-agent?? Discrete vs. continuous: This distinction can be applied to the state of the environment, the way time is handled and to the percepts/actions of the agent.
  • 25. TLo (IRIDIA) 25October 13, 2015 Environment types Solitaire Backgammom Intenet shopping Taxi Observable?? FULL FULL PARTIAL PARTIAL Deterministic?? YES NO YES NO Episodic?? NO NO NO NO Static?? YES YES SEMI NO Discrete?? YES YES YES NO Single-agent?? Single vs. multi-agent: Does the environment contain other agents who are also maximizing some performance measure that depends on the current agent’s actions?
  • 26. TLo (IRIDIA) 26October 13, 2015 Environment types Solitaire Backgammom Intenet shopping Taxi Observable?? FULL FULL PARTIAL PARTIAL Deterministic?? YES NO YES NO Episodic?? NO NO NO NO Static?? YES YES SEMI NO Discrete?? YES YES YES NO Single-agent?? YES NO NO NO Single vs. multi-agent: Does the environment contain other agents who are also maximizing some performance measure that depends on the current agent’s actions?
  • 27. TLo (IRIDIA) 27October 13, 2015 Environment types  The simplest environment is  Fully observable, deterministic, episodic, static, discrete and single-agent.  Most real situations are:  Partially observable, stochastic, sequential, dynamic, continuous and multi-agent.
  • 28. TLo (IRIDIA) 28October 13, 2015 Agent types  How does the inside of the agent work?  Agent = architecture + program  All agents have the same skeleton:  Input = current percepts  Output = action  Program= manipulates input to produce output  Note difference with agent function.
  • 29. TLo (IRIDIA) 29October 13, 2015 Agent types Function TABLE-DRIVEN_AGENT(percept) returns an action static: percepts, a sequence initially empty table, a table of actions, indexed by percept sequence append percept to the end of percepts action ← LOOKUP(percepts, table) return action This approach is doomed to failure
  • 30. TLo (IRIDIA) 30October 13, 2015 Agent types  Four basic kind of agent programs will be discussed:  Simple reflex agents  Model-based reflex agents  Goal-based agents  Utility-based agents  All these can be turned into learning agents.
  • 31. TLo (IRIDIA) 31October 13, 2015 Agent types; simple reflex  Select action on the basis of only the current percept.  E.g. the vacuum-agent  Large reduction in possible percept/action situations(next page).  Implemented through condition-action rules  If dirty then suck
  • 32. TLo (IRIDIA) 32October 13, 2015 The vacuum-cleaner world function REFLEX-VACUUM-AGENT ([location, status]) return an action if status == Dirty then return Suck else if location == A then return Right else if location == B then return Left Reduction from 4T to 4 entries
  • 33. TLo (IRIDIA) 33October 13, 2015 Agent types; simple reflex function SIMPLE-REFLEX-AGENT(percept) returns an action static: rules, a set of condition-action rules state ← INTERPRET-INPUT(percept) rule ← RULE-MATCH(state, rule) action ← RULE-ACTION[rule] return action Will only work if the environment is fully observable otherwise infinite loops may occur.
  • 34. TLo (IRIDIA) 34October 13, 2015 Agent types; reflex and state  To tackle partially observable environments.  Maintain internal state  Over time update state using world knowledge  How does the world change.  How do actions affect world. ⇒ Model of World
  • 35. TLo (IRIDIA) 35October 13, 2015 Agent types; reflex and state function REFLEX-AGENT-WITH-STATE(percept) returns an action static: rules, a set of condition-action rules state, a description of the current world state action, the most recent action. state ← UPDATE-STATE(state, action, percept) rule ← RULE-MATCH(state, rule) action ← RULE-ACTION[rule] return action
  • 36. TLo (IRIDIA) 36October 13, 2015 Agent types; goal-based  The agent needs a goal to know which situations are desirable.  Things become difficult when long sequences of actions are required to find the goal.  Typically investigated in search and planning research.  Major difference: future is taken into account  Is more flexible since knowledge is represented explicitly and can be manipulated.
  • 37. TLo (IRIDIA) 37October 13, 2015 Agent types; utility-based  Certain goals can be reached in different ways.  Some are better, have a higher utility.  Utility function maps a (sequence of) state(s) onto a real number.  Improves on goals:  Selecting between conflicting goals  Select appropriately between several goals based on likelihood of success.
  • 38. TLo (IRIDIA) 38October 13, 2015 Agent types; learning  All previous agent-programs describe methods for selecting actions.  Yet it does not explain the origin of these programs.  Learning mechanisms can be used to perform this task.  Teach them instead of instructing them.  Advantage is the robustness of the program toward initially unknown environments.
  • 39. TLo (IRIDIA) 39October 13, 2015 Agent types; learning  Learning element: introduce improvements in performance element.  Critic provides feedback on agents performance based on fixed performance standard.  Performance element: selecting actions based on percepts.  Corresponds to the previous agent programs  Problem generator: suggests actions that will lead to new and informative experiences.  Exploration vs. exploitation

Notes de l'éditeur

  1. Volledige observeerbare omgevingen zijn handig omdat de agent geen Interne staat moet bijhouden over de omgeving Partieel observeerbare omgevingen door noise of niet goed afgestelde sensoren of ontbrekende informatie
  2. Als de omgeving gedeeltelijk observeerbaar is kan deze stochastisch lijken. Dus het is beter om te kijken of de omgebing determinitisch of stochastisch is vanuit het standpunt van de agent. Als de omgeving deterministisch is behalve de acties van de andere agenten dan is de omgeving strategisch.
  3. If the agent performance score changes ahen time passes, the environment is semi-dynamic.
  4. If the agent performance score changes ahen time passes, the environment is semi-dynamic.
  5. If the agent performance score changes ahen time passes, the environment is semi-dynamic.
  6. If the agent performance score changes ahen time passes, the environment is semi-dynamic.
  7. If the agent performance score changes ahen time passes, the environment is semi-dynamic.
  8. If the agent performance score changes ahen time passes, the environment is semi-dynamic.