learning agent learning agents
play

Learning Agent Learning Agents An Agent that observes its - PDF document

Learning Agent Learning Agents An Agent that observes its performance and adapts its decision-making to improve its performance in the future. MSE 2400 EaLiCaRA Dr. Tom Way MSE 2400 Evolution & Learning 2 Agent Simple Reflex Agent


  1. Learning Agent Learning Agents • An Agent that observes its performance and adapts its decision-making to improve its performance in the future. MSE 2400 EaLiCaRA Dr. Tom Way MSE 2400 Evolution & Learning 2 Agent Simple Reflex Agent • Something that does something • Computational Agent – a computer that does something MSE 2400 Evolution & Learning 3 MSE 2400 Evolution & Learning 4 Simple Reflex Agent Model-based Reflex Agent • The action to be selected only depends on the most recent percept, not a percept sequence • As a result, these agents are stateless devices which do not have memory of past world states MSE 2400 Evolution & Learning 5 MSE 2400 Evolution & Learning 6 1

  2. Model-based Reflex Agent Goal-based Agent • Have internal state which is used to keep track of past states of the world (i.e., percept sequences may determine action) • Can assist an agent deal with at least some of the observed aspects of the current state MSE 2400 Evolution & Learning 7 MSE 2400 Evolution & Learning 8 Goal-based Agent Utility-based Agent • Agent can act differently depending on what the final state should look like • Example: automated taxi driver will act differently depending on where the passenger wants to go MSE 2400 Evolution & Learning 9 MSE 2400 Evolution & Learning 10 Utility-based Agent Learning Agent (in general) • An agent's utility function is an internalization of the performance measure (which is external) • Performance and utility may differ if the environment is not completely observable or deterministic MSE 2400 Evolution & Learning 11 MSE 2400 Evolution & Learning 12 2

  3. Learning Agent Parts (1) Learning Agent Parts (2) • Environment – world around the agent • Problem generator – test what is known • Sensors – data input, senses • Performance element – considers all that is known so far, refines what is known • Critic – evaluates the input from sensors • Changes – new information • Feedback – refined input, extracted info • Knowledge – improved ideas & concepts • Learning element – stores knowledge • Actuators – probes environment, triggers • Learning goals – tells what to learn gathering of input in new ways MSE 2400 Evolution & Learning 13 MSE 2400 Evolution & Learning 14 Intelligent Agents should… Classes of Intelligent Agents (1) • accommodate new problem solving rules incrementally • Decision Agents – for decision making • adapt online and in real time • Input Agents - that process and make • be able to analyze itself in terms of behavior, error and sense of sensor inputs (neural networks) success. • Processing Agents - solve a problem like • learn and improve through interaction with the environment (embodiment) speech recognition • learn quickly from large amounts of data • Spatial Agents - relate to physical world • have memory-based exemplar storage and retrieval capacities • have parameters to represent short and long term memory, age, forgetting, etc. MSE 2400 Evolution & Learning 15 MSE 2400 Evolution & Learning 16 Classes of Intelligent Agents (2) Classes of Intelligent Agents (3) • World Agents - incorporate a combination • Physical Agents - entity which percepts of all the other classes of agents to allow through sensors and acts through autonomous behaviors actuators. • Believable agents - exhibits a personality • Temporal Agents - uses time based stored via the use of an artificial character for the information to offer instructions to a interaction computer program or human being and uses feedback to adjust its next behaviors. MSE 2400 Evolution & Learning 17 MSE 2400 Evolution & Learning 18 3

  4. How Learning Agents Acquire Knowledge How Learning Agents Acquire Concepts (1) • Supervised Learning • Incremental Learning: update hypothesis – Agent told by teacher what is best action for a model only when new examples are given situation, then generalizes concept F(x) encountered • Inductive Learning • Feedback Learning: agent gets feedback – Given some outputs of F(x), agent builds h(x) on quality of actions it chooses given the that approximates F on all examples seen so h(x) it learned so far. far is SUPPOSED to be a good approximation for as yet unseen examples MSE 2400 Evolution & Learning 19 MSE 2400 Evolution & Learning 20 Examples How Learning Agents Acquire Concepts (2) • Reinforcement Learning: rewards / • Eliza - http://www.simonebaldassarri.com/eliza/eliza.html • Mike - http://www.rong-chang.com/tutor_mike.htm punishments prod agent into learning • iEinstein - • Credit Assignment Problem: agent doesn’t http://www.pandorabots.com/pandora/talk?botid=ea77c0 always know what the best (as opposed to 200e365cfb • More Cleverbots - https://www.existor.com/en/ just good) actions are, nor which rewards • Chatbots - http://www.chatbots.org/ are due to which actions. MSE 2400 Evolution & Learning 21 MSE 2400 Evolution & Learning 22 4

Download Presentation
Download Policy: The content available on the website is offered to you 'AS IS' for your personal information and use only. It cannot be commercialized, licensed, or distributed on other websites without prior consent from the author. To download a presentation, simply click this link. If you encounter any difficulties during the download process, it's possible that the publisher has removed the file from their server.

Recommend


More recommend