Knowledge-based agents

Transcription

Knowledge-based agents
Games & Agents 2012-2013
Games & Agents
Lecture 2
Agents and Intelligence
Jan Broersen
J.M.Broersen@uu.nl
Tuesday, May 28, 13
Games & Agents 2012-2013
The subject of today
I will give a very broad overview of different ideas from AI about
endowing agents with intelligence.
First: a movie about how the game-industry looks at intelligence
J.M.Broersen@uu.nl
Tuesday, May 28, 13
Games & Agents 2012-2013
The AI in Killzone 2
J.M.Broersen@uu.nl
Tuesday, May 28, 13
Intelligence
Games & Agents 2012-2013
•
In the trailer:
– (Agent) intelligence: (1) enemies pin you down and rush up to
you, (2) you do not see the same thing twice, (3) the
environment is destructible (4) the AI responds to changes
in the environment, (5) lean and peak system
– Multi-agent intelligence: buddies in the game that tell you to
take cover, saving your skin
•
In AI:
– Intelligence = intelligent behavior (we endorse Turing’s test)
– Intelligent behavior = selecting the ‘intelligent choices’ from
all the ones that are feasible and performing them
– Agent theory: intelligence is a function from percepts to
actions; the function can either be programmed or learned
J.M.Broersen@uu.nl
Tuesday, May 28, 13
Games & Agents 2012-2013
Simple reflex agents
J.M.Broersen@uu.nl
Tuesday, May 28, 13
Games & Agents 2012-2013
Simple Reflex Agent
J.M.Broersen@uu.nl
Tuesday, May 28, 13
Decision tree
Games & Agents 2012-2013
Conditionalization of actions is
of the simplest kind
In the tree: decisions at the
squares depend only on
observations at these squares.
Cycle: sense, act
J.M.Broersen@uu.nl
Tuesday, May 28, 13
Are these used in computer games?
Games & Agents 2012-2013
Yes: these are the scripted agents
Where do agent actions in scripts conditionalize on?
Very simple information like the player being nearby
However, intelligent choices should
1.
2.
3.
4.
be conditional on the state of the environment
be conditional on the state of other agents
be conditional on internal states
adapt after earlier experiences
J.M.Broersen@uu.nl
Tuesday, May 28, 13
Shortcomings simple reflex agents
Games & Agents 2012-2013
(1) No conditionalization on appropriate circumstances
(2) No situation awareness
(3) No looking ahead for decision making
=> agents need a model of the world, that is, we have to give them
knowledge
=> agents need a goal in life, that is, we have to give them
motivations (desires, intentions, obligations, etc.)
J.M.Broersen@uu.nl
Tuesday, May 28, 13
Games & Agents 2012-2013
Knowledge-based agents
J.M.Broersen@uu.nl
Tuesday, May 28, 13
Games & Agents 2012-2013
Reflex + State Agent
J.M.Broersen@uu.nl
Tuesday, May 28, 13
Games & Agents 2012-2013
Decision tree
Decisions conditionalize not
directly on the percepts, but
on the agent’s model of the
world (recalls what happened
before and what it has done)
How the world is, may now be
derived instead of observed
Allows for ‘reasoning’ before
acting
Cycle: sense, reason, act
J.M.Broersen@uu.nl
Tuesday, May 28, 13
Games & Agents 2012-2013
Knowledge-based agents
Perceived as a major break-through in the late 50ies, just after the
birth of artificial intelligence as a discipline (maybe the same
brake-through still has to be realized in game-AI..)
Model not only knowledge, but also lack thereof (uncertainty)
Logic based approach 1: epistemic logic
Logic based approach 2: non-monotonic reasoning
Bayesian approach: beliefs as probabilities
J.M.Broersen@uu.nl
Tuesday, May 28, 13
Are these used in computer games?
Games & Agents 2012-2013
Hardly (do you know of examples?)
However, if we agree the cognitive SOAR architecture gives us
knowledge based agents, then we do have an interesting
example.
Paper: [It knows what you’re going to do: adding anticipation to a
Quakebot, John Laird]
J.M.Broersen@uu.nl
Tuesday, May 28, 13
Games & Agents 2012-2013
A cognitive architecture: SOAR
Close to ACT-R
SOAR is claimed to be:
a candidate unified theory of cognition (Newell, 1990),
and
an architecture for intelligent behaviour
Decisions are never precompiled into uninterruptible
sequences => continuous (and concurrent) interaction
with the environment and with multiple internal
processes.
J.M.Broersen@uu.nl
Tuesday, May 28, 13
The SOAR Architecture
Games & Agents 2012-2013
SOAR is an ‘AI engine’ for making and executing decisions
Basic objects of decision: operators (we would call them actions)
Operators are either environmental actions, internal actions,
complex actions, or abstract goals.
SOAR is completely rule-based
So, far, very much like 3APL/2APL
J.M.Broersen@uu.nl
Tuesday, May 28, 13
The SOAR Architecture
Long-Term Memories
Games & Agents 2012-2013
Procedural/Skill
Semantic
Learning
Rule
Learning
Short-Term Memory
Control
Procedure
Action
Perception
J.M.Broersen@uu.nl
Tuesday, May 28, 13
Semantic/Concept
Body
Episodic
Episodic
Learning
Games & Agents 2012-2013
But, unlike in 3APL/2APL:
There is no deliberation or control language! Also there is no
predefined order in which to apply rules: the system is
completely data-driven.
The current state (‘working memory’) consists of perceptions,
elaborations on it, and everything else derived by earlier rule
applications.
SOAR continually and concurrently proposes, selects and applies
operator rules to the current state.
Abstract (goal) operators that cannot be executed directly generate
sub-states for which additional operators are proposed, selected
and applied (remember HTN-planning)
J.M.Broersen@uu.nl
Tuesday, May 28, 13
Adding anticipation to a Quakebot
Games & Agents 2012-2013
•
•
•
Central idea: enable a Quakebot to project itself in its
enemies to predict its behavior and to take advantage of
that.
Similar to a 1 ply deep minimax or alfa-beta algorithm
(Chess, game theory) where the idea is also to use the
same evaluation function for opponents.
Add three new sets of rules to SOAR bots:
1. For deciding when to start predicting
2. For generating expected opponent behavior
3. For proposing when to stop predicting and for how to
take advantage of the prediction (e.g. ambush)
J.M.Broersen@uu.nl
Tuesday, May 28, 13
When to start predicting? (task 1)
Games & Agents 2012-2013
Not always..
Decided by a special set of rules
Which are up to the Quakebot designer..
J.M.Broersen@uu.nl
Tuesday, May 28, 13
Games & Agents 2012-2013
Generating expected behavior (task 2)
First the quake bot generates an expected internal state of its
opponent, by applying special rules on his perceptions of the
other (his position, health, armor level, current weapon).
Then he applies his own behavioral rules on this state, thereby
projecting himself in the other!
The behavioral projection works on the generated internal state and
projected environment, which is less computational intensive
than the ‘real thing’.
J.M.Broersen@uu.nl
Tuesday, May 28, 13
Games & Agents 2012-2013
Predictions, and how they are used (task 3)
Predictions should not go on forever: special rules decide when to
stop. For instance if an ambush is not possible due to
environmental conditions or the timing.
Predictions also stop when they encounter too much uncertainty
(impasses in the prediction!).
In the paper’s set-up the technique is only used to hunt, ambush and
deny power-ups.
J.M.Broersen@uu.nl
Tuesday, May 28, 13
Games & Agents 2012-2013
Limitations
Anticipations is restricted to situations deemed worthwhile by the
programmer.
So, tactics are not ‘discovered’ but preprogrammed by the Quake
bot designer.
Alternative: always do predictions, and add general evaluation
functions and planning ⇒ too expensive, computationally.
J.M.Broersen@uu.nl
Tuesday, May 28, 13
Conclusions SOAR application to Quake
Games & Agents 2012-2013
SOAR quake bots use about 750 rules.
Adding anticipation requires only 4 rules for deciding when to start
predicting, 16 to generate an approximate opponent internal
state, 5 to use the prediction.
So adding anticipation turns out to be quite easy, from a
programming point of view.
But, computational much more intensive.
J.M.Broersen@uu.nl
Tuesday, May 28, 13
Shortcomings of knowledge-based agents
Games & Agents 2012-2013
Goals are implicit in the set of rules
Agents decide without explicitly looking ahead in the tree
J.M.Broersen@uu.nl
Tuesday, May 28, 13
Games & Agents 2012-2013
Goal-based Agent
J.M.Broersen@uu.nl
Tuesday, May 28, 13
Deliberation cycles for the different agent types
Games & Agents 2012-2013
Simple reflex agents: sense, act
Reflex agents with state: sense, reason, act
Goal-based agents: sense, update beliefs, select goals, select
intentions, etc..., act
J.M.Broersen@uu.nl
Tuesday, May 28, 13
Games & Agents 2012-2013
The tree view: planning
J.M.Broersen@uu.nl
Tuesday, May 28, 13
Games & Agents 2012-2013
Are these agents used in computer games?
Yes: path-planning using A*
Knowledge: the network of waypoints
Goal: finding a shortest route
However, this is a very specific type of planning, because actions
(movements) and goals (wanting to be somewhere else)
themselves are very specific
Knowledge about waypoints is simply assumed (and not learned
through observation)
Because the problem is so specific, we might even consider to
calculate most of it offline for a given fixed environment..
J.M.Broersen@uu.nl
Tuesday, May 28, 13
Planning
Games & Agents 2012-2013
See course “intelligent systems”
State space planning versus plan space planning (HTN planning)
Planning under uncertainty (conformant planning, contingency
planning, etc.)
Main focus of research: controlling complexity, heuristics
J.M.Broersen@uu.nl
Tuesday, May 28, 13
BDI-theory
Games & Agents 2012-2013
Prominent theory: Belief-Desire-Intention
Issue: how do all three depend on each other?
Example: commitment strategies [Cohen and Levesque]: you keep
an intention as long as you belief the associated desire to be
attainable
Example: realism: you do not desire things you believe to be true
already; you do not believe things only because they match your
desire (wishful thinking)
J.M.Broersen@uu.nl
Tuesday, May 28, 13
3APL/2APL
Games & Agents 2012-2013
Inspired by BDI
Goal-base, Belief-base, and Plan-base (intentions)
A language to specify the deliberation cycle
Intuitions from Prolog
J.M.Broersen@uu.nl
Tuesday, May 28, 13
Shortcomings
Games & Agents 2012-2013
Knowledge is not ‘fine grained’
Goals are not ‘fine grained’
J.M.Broersen@uu.nl
Tuesday, May 28, 13
Games & Agents 2012-2013
Utility-based agents
J.M.Broersen@uu.nl
Tuesday, May 28, 13
Games & Agents 2012-2013
Utility-based agent
J.M.Broersen@uu.nl
Tuesday, May 28, 13
Games & Agents 2012-2013
Maximal expected utility
• Cycle: sense, find optimal
tactic given beliefs,
utilities, act
• Reasoning is now in terms
of utilities, chances and
look aheads: MEU, MDPs,
PoMDPs
J.M.Broersen@uu.nl
Tuesday, May 28, 13
Games & Agents 2012-2013
MDPs and PoMDPs
Again see the theory explained in the course “intelligent
systems” (Policies, Bellman equations, value iteration, policy
iteration, etc.)
I do not know of any example from computer games where this
agent theory is applied (do you?)
J.M.Broersen@uu.nl
Tuesday, May 28, 13
Bayesian update
Games & Agents 2012-2013
Bayes' theorem adjusts probabilities given new evidence in the following way:
Where H represents a specific hypothesis, which may or may not be some null
hypothesis.
P(H) is called the prior probability of H that was inferred before new evidence,
E, became available.
P(E | H) is called the conditional probability of seeing the evidence E if the
hypothesis H happens to be true. It is also called a likelihood function when it
is considered as a function of H for fixed E.
P(E) is called the marginal probability of E: the a priori probability of
witnessing the new evidence E under all possible hypotheses.
P(H | E) is called the posterior probability of H given E.
J.M.Broersen@uu.nl
Tuesday, May 28, 13
Games & Agents 2012-2013
The three prisoners
Three prisoners A, B and C are told they have a 2/3 chance of being
executed.
At a certain point the decision is made by the director of the prison.
Then the director visits prisoner A. A asks: will B or C be
executed? The director answers: B
What is A’s chance of survival?
Answer: unclear!
J.M.Broersen@uu.nl
Tuesday, May 28, 13
Games & Agents 2012-2013
The three doors
The Monty Hall dilemma
Conditionalizing on extra information
The chances for getting the car when not changing are 1/3. When
changing, the chances are 2/3! (1/3 for initially correct and
changing to the wrong door and 2/3 for initially wrong after
which changing guarantees the good outcome).
J.M.Broersen@uu.nl
Tuesday, May 28, 13
Shortcomings
Games & Agents 2012-2013
• However, No learning (like all of the previous)
• Or, is there?
• Any update of the model of the world can be seen as a
learning action
J.M.Broersen@uu.nl
Tuesday, May 28, 13
Games & Agents 2012-2013
Learning Agent
Can be combined with any of the previous
J.M.Broersen@uu.nl
Tuesday, May 28, 13
Games & Agents 2012-2013
Trends in Neural Networks
There has always been the “black box” criticism towards neural
networks (I call it “the tweaking problem”)
Current trend: add structure to neural networks
How should such structure look like?
Inspiration from biology?
NEAT: the structure of the neural network evolves determined by a
genetic algorithm
Liquid state machines
Neuro-symbolic computing
J.M.Broersen@uu.nl
Tuesday, May 28, 13
Games & Agents 2012-2013
Rationality versus emotion
J.M.Broersen@uu.nl
Tuesday, May 28, 13
Games & Agents 2012-2013
Some functions of emotions
Frijda: emotions are (search) heuristics.
Emotions may thus be said to guide planning.
E.g. happiness may keep an agent on the right track, because it
makes him selecting similar choices to reinforce his happiness.
E.g. anger or distress may guide an agent into selecting choices he
would otherwise not consider. Those choices maybe the solution
to his problem.
A thought: A* is an optimistic algorithm. Because of that it is
complete (always finds a solution if it exists) and optimal (it will
always find the ‘best’ solution). Optimism seems to be related to
positive emotions...
J.M.Broersen@uu.nl
Tuesday, May 28, 13
Games & Agents 2012-2013
Emotions and rationality
What is the relation between rationality and emotion? Both are
involved in selecting (intelligent) decisions.
For instance, rationally you estimate it is better to kill your enemy,
or he will probably kill you. But, emotionally you feel extremely
bad about killing.
Or, rationally you estimate it is good to punish your child,
emotionally you detest it.
There is an example case of a man without emotion due to brain
damage. He kept an average IQ, but could no longer decide on
simple things like what to wear.
J.M.Broersen@uu.nl
Tuesday, May 28, 13
Games & Agents 2012-2013
The Multi-Agent paradigm
J.M.Broersen@uu.nl
Tuesday, May 28, 13
Computing perspective
Games & Agents 2012-2013
Multi-agent systems constitute a new computational paradigm
based on:
Distribution of computational power
Interconnection and communication
Autonomy and Intelligence
Delegation and distribution of control
Anthropomorphic concepts
J.M.Broersen@uu.nl
Tuesday, May 28, 13
Software engineering perspective
Games & Agents 2012-2013
Multi-agent systems are a new software engineering paradigm.
Analysis ⇒ Design ⇒ Implementation ⇒ Test
The aim is to provide high-level abstraction, to model and develop complex
systems
70’s: Structural analysis methodology
80’s / 90’s: Object-oriented methodology
90’s / 00’s: Agent-oriented methodology
J.M.Broersen@uu.nl
Tuesday, May 28, 13
Artificial intelligence perspective
Games & Agents 2012-2013
Multi-agent systems and Artificial Intelligence do not coincide.
Artificial Intelligence
Planning
Learning
Vision
Language understanding
Multi-agent Systems
Understand and model social intelligence and emergent behavior
Interaction and Communication
Social concepts: obligation, norms, responsibilities, etc.
Optimal solutions can be obtained by co-ordination and co-operation
Simulation can verify social and economic theory
J.M.Broersen@uu.nl
Tuesday, May 28, 13
Games & Agents 2012-2013
Characteristics of multi-agent systems
Multi-agent systems consist of a number of interacting autonomous agents.
Multi-agent systems are designed to achieve some global goal.
Agents need special abilities to cooperate, coordinate, and negotiate to achieve
their objectives.
Multi-agent systems are specified in terms of high-level abstract concepts such as
role, permission, responsibility, and interaction.
Applications for multi-agent systems are, for example, power management
systems, transportation systems, and auctions.
J.M.Broersen@uu.nl
Tuesday, May 28, 13
Autonomous agents
Games & Agents 2012-2013
Autonomous agents are intelligent agents from a multi-agent perspective.
Autonomous agents are active, social, and adaptable computer systems situated in
some dynamic environment that autonomously perform actions to achieve
their objectives.
1.
2.
3.
4.
Reactivity: responding to changes in the environment
Pro-activity (deliberativeness): goal-directed behavior
Social awareness: interaction and communication
Rationality: maximization of utilities
•
•
Agents decide which action to perform based on their internal state.
The internal state of agents can be specified in terms of high-level abstract
concepts such as belief, desire, goal, intention, plan, and action.
J.M.Broersen@uu.nl
Tuesday, May 28, 13
Games & Agents 2012-2013
The End
J.M.Broersen@uu.nl
Tuesday, May 28, 13