CSCE 420 Lecture 24
Jump to navigation
Jump to search
« previous | Tuesday, April 16, 2013 | next »
Agent Architectures
Decision-making Implementation ()
- Sense
- Decide
- Act
- repeat...
Similar to other fields'
- Observe
- Orient
- Decide
- Act
- repeat...
Simple Reflex
- sensor input
- what does world look like now? → state
- what action should I do now? → action
- perform action
- Table-lookup: determine actions in a table based on sensor inputs (doesn't scale well)
- Rule-based Agent: Implement with set of "condition → action" rules (requires conjunctive antecedents)
- Rules should be ordered by priority
- Stop after first match
Genghis Robot (Rob Brooks, MIT) has simple, reactive controllers for legs:
- push backward
- lift up
- move forward
- if stuck, retract and lift higher
- set down
- repeat...
Model-Based
Maintain an internal representation of the world state
- sensor input
- update internal state
- what does the world look like now? → state
- what action should I do now? → action
- perform actions
Goal-Based
Describe in KB what goals need to be accomplished
- Try to infer for all
- use Situation Calculus to encode preconditions and effects of actions in
OR use planners to derive sequence of actions (plan) by state-space search (e.g. A*)
In either case, output of actuators follows a plan
Utility-Based
- Utility Function
- Transition function
- defines outcomes of actions
- could be probablistic (distribution over successor states)
- Reward/Cost function
- payoff of action
- Goal: maximize reward over time: , where is a geometric weighting scalar/constant that helps series to converge.
- smaller would look for immediate payoff
- larger would be longer term
Plans are encoded in policies (mapping states to actions)
Computing optimal policy that maximizes long-term discounted reward
- evaluation via Bellman equations:
- Value function gives utility of each state (which depends on its neigboring states)
- Therefore values are all Coupled