introduction to artificial intelligence planning under
play

Introduction to Artificial Intelligence Planning under Uncertainty - PowerPoint PPT Presentation

Introduction to Artificial Intelligence Planning under Uncertainty Janyl Jumadinova November 2, 2016 Goals and Preferences 2/17 Preferences Actions result in outcomes . Agents have preferences over outcomes. 3/17 Preferences


  1. Introduction to Artificial Intelligence Planning under Uncertainty Janyl Jumadinova November 2, 2016

  2. Goals and Preferences 2/17

  3. Preferences ◮ Actions result in outcomes . ◮ Agents have preferences over outcomes. 3/17

  4. Preferences ◮ Actions result in outcomes . ◮ Agents have preferences over outcomes. A rational agent will do the action that has the best outcome for them 3/17

  5. Preferences ◮ Actions result in outcomes . ◮ Agents have preferences over outcomes. A rational agent will do the action that has the best outcome for them ◮ Sometimes agents don’t know the outcomes of the actions, but they still need to compare actions ◮ Agents have to act. (Doing nothing is (often) an action) 3/17

  6. Preferences over Outcomes 4/17

  7. Lotteries ◮ An agent may not know the outcomes of their actions, but only have a probability distribution of the outcomes. ◮ A lottery is a probability distribution over outcomes. [ p 1 : o 1 , p 2 : o 2 , ..., p k : o k ], where o i are outcomes and p i ≥ 0 s.t. � i p i = 1 ◮ The lottery specifies that outcome o i occurs with probability p i . 5/17

  8. Measure of Preference ◮ We would like a measure of preference that can be combined with probabilities: value ([ p : o 1 , 1 − p : o 2 ]) = p × value ( o 1 ) + (1 − p ) × value ( o 2 ) 6/17

  9. Measure of Preference ◮ We would like a measure of preference that can be combined with probabilities: value ([ p : o 1 , 1 − p : o 2 ]) = p × value ( o 1 ) + (1 − p ) × value ( o 2 ) ◮ Money does not act this way: $1 , 000 , 000 or [0 . 5 : $1 , 0 . 5 : 2 , 000 , 000]? 6/17

  10. Theorem ◮ If preferences follow the preceding properties, then preferences can be measured by a function: utility : outcomes → [0 , 1] such that 7/17

  11. Utility as a function of money 8/17

  12. Additive Utility ◮ Suppose the outcomes can be described in terms of features X 1 , ..., X n . ◮ An additive utility is one that can be decomposed into set of factors: u ( X 1 , ..., X n ) = f 1 ( X 1 ) + ... + f n ( X n ). 9/17

  13. Additive Utility ◮ Suppose the outcomes can be described in terms of features X 1 , ..., X n . ◮ An additive utility is one that can be decomposed into set of factors: u ( X 1 , ..., X n ) = f 1 ( X 1 ) + ... + f n ( X n ). ◮ This assumes additive independence . ◮ Strong assumption : contribution of each feature doesnt depend on other features. 9/17

  14. Additive Utility ◮ An additive utility has a canonical representation: u ( X 1 , ..., X n ) = w 1 × u 1 ( X 1 ) + ... + w n u n ( X n ). 10/17

  15. Additive Utility ◮ An additive utility has a canonical representation: u ( X 1 , ..., X n ) = w 1 × u 1 ( X 1 ) + ... + w n u n ( X n ). ◮ If best i is the best value of X i , u i ( X i = best i ) = 1. ◮ If worst i is the worst value of X i , u i ( X i = worst i ) = 0. 10/17

  16. Additive Utility ◮ An additive utility has a canonical representation: u ( X 1 , ..., X n ) = w 1 × u 1 ( X 1 ) + ... + w n u n ( X n ). ◮ If best i is the best value of X i , u i ( X i = best i ) = 1. ◮ If worst i is the worst value of X i , u i ( X i = worst i ) = 0. ◮ w i are weights, � i w i = 1. ◮ The weights reflect the relative importance of features. We can determine weights by comparing outcomes. 10/17

  17. Utility and Time ◮ Would you prefer $1000 today or $1000 next year? 11/17

  18. Utility and Time ◮ Would you prefer $1000 today or $1000 next year? ◮ What price would you pay now to have an eternity of happiness? 11/17

  19. Utility and Time ◮ Would you prefer $1000 today or $1000 next year? ◮ What price would you pay now to have an eternity of happiness? ◮ How can you trade off pleasures today with pleasures in the future? 11/17

  20. Utility and Time 12/17

  21. Rewards and Values 13/17

  22. Rewards and Values 14/17

  23. Framing Effects 15/17

  24. Framing Effects 16/17

  25. Framing Effects 17/17

Download Presentation
Download Policy: The content available on the website is offered to you 'AS IS' for your personal information and use only. It cannot be commercialized, licensed, or distributed on other websites without prior consent from the author. To download a presentation, simply click this link. If you encounter any difficulties during the download process, it's possible that the publisher has removed the file from their server.

Recommend


More recommend