Learning Agents Overview Learning important aspects Learning in - - PDF document

learning agents
SMART_READER_LITE
LIVE PREVIEW

Learning Agents Overview Learning important aspects Learning in - - PDF document

CPE/CSC 580-S06 Artificial Intelligence Intelligent Agents Learning Agents Overview Learning important aspects Learning in Agents goal, types; individual agents, multi-agent systems Learning Agent Model components, representation,


slide-1
SLIDE 1

CPE/CSC 580-S06 Artificial Intelligence – Intelligent Agents

Learning Agents

Overview

Learning important aspects Learning in Agents goal, types; individual agents, multi-agent systems Learning Agent Model components, representation, feedback, prior knowledge Learning Methods inductive learning, neural networks, reinforcement learning, genetic algorithms Knowledge and Learning explanation-based learning, relevance information

Franz J. Kurfess, Cal Poly SLO 152

slide-2
SLIDE 2

CPE/CSC 580-S06 Artificial Intelligence – Intelligent Agents

Learning

acquisition of new knowledge and skills

  • n the agent’s own initiative

incorporation of new knowledge into the existing knowledge performed by the system itself not only injected by the developer performance improvement simply accumulating knowledge isn’t sufficient

Franz J. Kurfess, Cal Poly SLO 153

slide-3
SLIDE 3

CPE/CSC 580-S06 Artificial Intelligence – Intelligent Agents

Learning in Agents

improved performance through learning

learning modify the internal knowledge goal improvement of future performance types of learning memorization, self-observation, generalization, exploration, creation of new theories, meta-learning levels of learning value-action pairs representation of a function general first-order logic theories

Franz J. Kurfess, Cal Poly SLO 154

slide-4
SLIDE 4

CPE/CSC 580-S06 Artificial Intelligence – Intelligent Agents

Learning Agent Model

conceptual components

learning element responsible for making improvements performance element selection of external actions: takes in percepts and decides on actions critic evaluation of the performance according to a fixed standard problem generator suggests exploratory actions new experiences with potential benefits

Franz J. Kurfess, Cal Poly SLO 155

slide-5
SLIDE 5

CPE/CSC 580-S06 Artificial Intelligence – Intelligent Agents

Diagram [?] p. 526

Franz J. Kurfess, Cal Poly SLO 155

slide-6
SLIDE 6

CPE/CSC 580-S06 Artificial Intelligence – Intelligent Agents

Learning Element

how to improve performance

performance element affected components internal representation used for components to improved feedback from the environment from a teacher prior knowledge about the environment / domain

Franz J. Kurfess, Cal Poly SLO 156

slide-7
SLIDE 7

CPE/CSC 580-S06 Artificial Intelligence – Intelligent Agents

Performance Element Components

relevant for learning

mapping function from percepts and internal state to actions inference mechanism infer relevant properties of the world from percepts changes in the world information about the way the world evolves effects of actions results of possible actions the agent can take utility information desirability of world / internal states action-value information desirability of actions in particular states goals classes of desirable states

Franz J. Kurfess, Cal Poly SLO 157

slide-8
SLIDE 8

CPE/CSC 580-S06 Artificial Intelligence – Intelligent Agents

utility maximization

Franz J. Kurfess, Cal Poly SLO 158

slide-9
SLIDE 9

CPE/CSC 580-S06 Artificial Intelligence – Intelligent Agents

Representation

used in a component

deterministic linear weighted polynomials logic propositional, first order probabilistic belief networks, decision theory learning algorithms need to be adapted to the particular representation

Franz J. Kurfess, Cal Poly SLO 159

slide-10
SLIDE 10

CPE/CSC 580-S06 Artificial Intelligence – Intelligent Agents

Feedback

about the desired outcome

supervised learning inputs and outputs of percepts can be perceived immediately reinforcement learning an evaluation of the action (hint) becomes available not necessarily immediately no direct information about the correct action unsupervised learning no hint about correct outputs

Franz J. Kurfess, Cal Poly SLO 160

slide-11
SLIDE 11

CPE/CSC 580-S06 Artificial Intelligence – Intelligent Agents

Inductive Learning

learning from examples

reflex agent direct mapping from percepts to actions inductive inference given a collection of examples for a function f, return a function h (hypothesis) that approximates f bias preference for one hypothesis over another usually large number of possible consistent hypotheses incremental learning new examples are integrated as they arrive

Franz J. Kurfess, Cal Poly SLO 161

slide-12
SLIDE 12

CPE/CSC 580-S06 Artificial Intelligence – Intelligent Agents

Decision Trees

deriving decisions from examples

goal take a situation described by a set of properties,] and produce a yes/no decision goal predicate Boolean function defining the goal expressiveness propositional logic efficiency more compact than truth tables in many cases exponential in some cases (parity, majority)

Franz J. Kurfess, Cal Poly SLO 162

slide-13
SLIDE 13

CPE/CSC 580-S06 Artificial Intelligence – Intelligent Agents

Induction

for decision trees

example described by the values of the attributes and the value of the goal predicate (classification) training set set of examples used for training test set set of examples used for evaluation different from the training set algorithm classify into positive and negative sets select the most important attribute split the tree, and apply the algorithm recursively to the subtrees

Franz J. Kurfess, Cal Poly SLO 163

slide-14
SLIDE 14

CPE/CSC 580-S06 Artificial Intelligence – Intelligent Agents

Performance Evaluation

for inductive learning algorithms

goals reproduce classification of the training set predict classification of unseen examples example set size must be reasonably large average prediction quality for different sizes of training sets and randomly selected training sets learning curve (”happy curve”) plots average prediction quality as a function of the size of the training set training and test data should be kept separate, and each run of the algorithm should be independent of the others

Franz J. Kurfess, Cal Poly SLO 164

slide-15
SLIDE 15

CPE/CSC 580-S06 Artificial Intelligence – Intelligent Agents

Examples

decision tree learning

Gasoil design of oil platform equipment expert system with 2500 rules generated from existing designs using a flight simulator program generated from examples of skilled human pilots somewhat better performance that the teachers (for regular tasks) not so good for rare, complex tasks

Franz J. Kurfess, Cal Poly SLO 165

slide-16
SLIDE 16

CPE/CSC 580-S06 Artificial Intelligence – Intelligent Agents

Neural Networks: see separate slides

Franz J. Kurfess, Cal Poly SLO 165

slide-17
SLIDE 17

CPE/CSC 580-S06 Artificial Intelligence – Intelligent Agents

Reinforcement Learning

learning from success and failure

reinforcement or punishment feedback about the outcome of actions no direct feedback about the correctness of an action possibly delayed rewards as percepts must be recognized as special percepts, not just another sensory input can be components of the utility, or hints

Franz J. Kurfess, Cal Poly SLO 166

slide-18
SLIDE 18

CPE/CSC 580-S06 Artificial Intelligence – Intelligent Agents

Variations

in the learning task

environment accessible or not prior knowledge internal model of the environment knowledge about effects of actions utility information passive learner watches the environment without actions active learner act based upon learned information problem generation for exploring the environment exploration trade-off between immediate and future benefits

Franz J. Kurfess, Cal Poly SLO 167

slide-19
SLIDE 19

CPE/CSC 580-S06 Artificial Intelligence – Intelligent Agents

Generalization

in reinforcement learning

implicit representation more compact form than a table for input-output values input generalization apply learned information to unknown states trade-off between the size of the hypothesis space and the time to learn a function

Franz J. Kurfess, Cal Poly SLO 168

slide-20
SLIDE 20

CPE/CSC 580-S06 Artificial Intelligence – Intelligent Agents

Examples

  • f reinforcement learning

game-playing TD-gammon: neural network with 80 hidden units, 300,000 training games and precomputed features added to the input representation plays

  • n par with the top three human players

worldwide robot control cart-pole balancing (inverted pendulum)

Franz J. Kurfess, Cal Poly SLO 169

slide-21
SLIDE 21

CPE/CSC 580-S06 Artificial Intelligence – Intelligent Agents

Genetic Algorithms

as a variation of reinforcement learning

basic idea selection and reproduction operators are applied to sets of individuals reward successful reproduction agent is a species, not an individual fitness function takes an individual, returns a real number algorithm parallel search in the space of individuals for one that maximizes the fitness function selection strategy random, probability of selection is proportional to fitness reproduction selected individuals are randomly paired

Franz J. Kurfess, Cal Poly SLO 170

slide-22
SLIDE 22

CPE/CSC 580-S06 Artificial Intelligence – Intelligent Agents

cross-over: gene sequences are split at the same point and crossed mutation: each gene can be altered with small probability

Franz J. Kurfess, Cal Poly SLO 171

slide-23
SLIDE 23

CPE/CSC 580-S06 Artificial Intelligence – Intelligent Agents

Knowledge and Learning

learning with prior knowledge

learning methods take advantage of prior knowledge about the environment learning level general first-order logic theories as opposed to function learning description conjunction of all example specifications classification conjunction of all example evaluations hypothesis newly generated theory entailment constraint together with descriptions, the hypothesis must entail classifications

Franz J. Kurfess, Cal Poly SLO 172

slide-24
SLIDE 24

CPE/CSC 580-S06 Artificial Intelligence – Intelligent Agents

essential step toward truly autonomous intelligent agents

Franz J. Kurfess, Cal Poly SLO 173

slide-25
SLIDE 25

CPE/CSC 580-S06 Artificial Intelligence – Intelligent Agents

Explanation-Based Learning

potentially known information is made explicit

usage known theories are converted into directly applicable knowledge (”aha-effect”) entailment constraint background entails hypothesis, which together with the example descriptions entails classificaitns

Franz J. Kurfess, Cal Poly SLO 174

slide-26
SLIDE 26

CPE/CSC 580-S06 Artificial Intelligence – Intelligent Agents

Example: Gary Larson sketch where Thad the caveman shows his mates how to use a stick to grill a lizard

Franz J. Kurfess, Cal Poly SLO 174

slide-27
SLIDE 27

CPE/CSC 580-S06 Artificial Intelligence – Intelligent Agents

Relevance-Based Learning

points out relevant features

functional dependencies generalization that derives the value of one predicate from another one learning method deductive does not create any new knowledge main effect limitation of the hypothesis space allows deductive generalizations from single examples

Franz J. Kurfess, Cal Poly SLO 175

slide-28
SLIDE 28

CPE/CSC 580-S06 Artificial Intelligence – Intelligent Agents

Examples: inference from nationality to spoken language

Franz J. Kurfess, Cal Poly SLO 175

slide-29
SLIDE 29

CPE/CSC 580-S06 Artificial Intelligence – Intelligent Agents

Inductive Logic Programming

learning based on predicate logic

learning methods discovery of new predicates and new knowledge extensions of decision trees to predicate logic main effects reduction of the hypothesis space to include

  • nly theories that are consistent with prior

knowledge smaller hypotheses by using prior knowledge to formulate new rules

Franz J. Kurfess, Cal Poly SLO 176

slide-30
SLIDE 30

CPE/CSC 580-S06 Artificial Intelligence – Intelligent Agents

Example: Learning of family relationships generating the difinition of Grandparent becomes much easier if Parent is available too complex for decision-tree learning

Franz J. Kurfess, Cal Poly SLO 176

slide-31
SLIDE 31

CPE/CSC 580-S06 Artificial Intelligence – Intelligent Agents

Learning in Multi-Agent Systems

as opposed to learning in individual agents

principal categories differentiating features learning coordination learning with other agents

Franz J. Kurfess, Cal Poly SLO 177

slide-32
SLIDE 32

CPE/CSC 580-S06 Artificial Intelligence – Intelligent Agents

Principal Categories

  • f learning in multi-agent systems

centralized learning

  • ne agent performs all relevant activities

no interaction required may include multiple agents with the same learning goals decentralized learning several agents are engaged in the same learning process requires interaction and collaboration between agents

Franz J. Kurfess, Cal Poly SLO 178

slide-33
SLIDE 33

CPE/CSC 580-S06 Artificial Intelligence – Intelligent Agents

Differentiating Features

for multi-agent learning approaches

degree of decentralization distributed, parallelized interaction-specific features level, persistence, frequency, patterns, variability

  • f interaction

involvement-specific features relevance of involvement roles played by agents goal-specific features type of improvement intended by learning compatibility of learning goals across different agents learning method rote learning, learning by instruction, example, practice, analogy, discovery learning feedback

Franz J. Kurfess, Cal Poly SLO 179

slide-34
SLIDE 34

CPE/CSC 580-S06 Artificial Intelligence – Intelligent Agents

supervised learning reinforcement learning unsupervised learning

Franz J. Kurfess, Cal Poly SLO 180

slide-35
SLIDE 35

CPE/CSC 580-S06 Artificial Intelligence – Intelligent Agents

Credit Assignment

Whose fault was it?

inter-agent credit assignment several agents are involved in a learning activity that results in a performance change who is responsible for the change? intra-agent credit assignment which of the internal components of an agent involved in learning is responsible for a performance change? general question what action carried out by which agent contributed to what extent to the performance change? related question what knowledge, what inferences, and what decisions led to an action

Franz J. Kurfess, Cal Poly SLO 181

slide-36
SLIDE 36

CPE/CSC 580-S06 Artificial Intelligence – Intelligent Agents

Learning and Activity Coordination

improving the performance of the overall system

reinforcement learning agents try to maximize the amount of reinforcement they receive from the environment or from an instructor isolated reinforcement learning individual agents use reinforcement learning to achieve their own goals no communication or collaboration about the learning processes or results collaborative reinforcement learning agents communicate to decide on individual and group actions agents have some insight into each others learning processes, and share some results

Franz J. Kurfess, Cal Poly SLO 182

slide-37
SLIDE 37

CPE/CSC 580-S06 Artificial Intelligence – Intelligent Agents

Learning with Other Agents

  • rganizational roles

assignments of roles in teams in order to learn more effectively environmental conditions adaptation to changes in the environment mutual exchange of pertinent information especially for buyer and seller agents in electronic marketplaces team competitions improvements in playing against competitors learning from more experienced agents

Franz J. Kurfess, Cal Poly SLO 183

slide-38
SLIDE 38

CPE/CSC 580-S06 Artificial Intelligence – Intelligent Agents

Explanation-Based Learning

to improve cooperative problem solving

inefficiencies in coordinated behavior identification of underlying causes rectification of decisions or actions learning analysis problem-solving traces are collected and analyzed explanations are generated for relevant decisions agent models the explanations should be generated with the model of the agent in mind

Franz J. Kurfess, Cal Poly SLO 184

slide-39
SLIDE 39

CPE/CSC 580-S06 Artificial Intelligence – Intelligent Agents

Learning and Communication

learn more and less by exchanging information

reduced communication learning may lead to less conversations among agents instead of asking, agents learn themselves reduced learning communcation can reduce the need for learning instead of learning, agents ask other agents balancing learning and communication may depend on bottlenecks in computational power or bandwidth

Franz J. Kurfess, Cal Poly SLO 185

slide-40
SLIDE 40

CPE/CSC 580-S06 Artificial Intelligence – Intelligent Agents

Summary - Learning Agents

Learning in agents acquisition and generation of new knowledge performance improvement as goal Learning Agent Model components, representation, feedback, prior knowledge Learning Methods inductive learning, neural networks, reinforcement learning, genetic algorithms Knowledge and Learning explanation-based learning, relevance information inductive logic programming, maching learning Learning in Multi-Agent Systems de-centralized, interaction, integration of results

Franz J. Kurfess, Cal Poly SLO 186