CS 486/686 Introduction to Artifjcial Intelligence Alice Gao - - PowerPoint PPT Presentation

cs 486 686 introduction to artifjcial intelligence
SMART_READER_LITE
LIVE PREVIEW

CS 486/686 Introduction to Artifjcial Intelligence Alice Gao - - PowerPoint PPT Presentation

1/22 CS 486/686 Introduction to Artifjcial Intelligence Alice Gao Lecture 2 Readings: R & N 2.1, 2.2, 2.3 (esp 2.3.2) Based on work by K. Leyton-Brown, K. Larson, and P. van Beek 2/22 Outline Learning goals Rational Agents Properties


slide-1
SLIDE 1

1/22

CS 486/686 Introduction to Artifjcial Intelligence

Alice Gao

Lecture 2 Readings: R & N 2.1, 2.2, 2.3 (esp 2.3.2) Based on work by K. Leyton-Brown, K. Larson, and P. van Beek

slide-2
SLIDE 2

2/22

Outline

Learning goals Rational Agents Properties of Task Environments Revisiting the learning goals

slide-3
SLIDE 3

3/22

Learning goals - CS 486/686 Lecture 2

By the end of the lecture, you should be able to

▶ Given examples of sensors and actuators. ▶ Defjne rational agents. ▶ Given a task environment, describe its properties. ▶ Given a property, give examples of task environments that

have this property.

slide-4
SLIDE 4

4/22

Agents

▶ Interact with the environment. ▶ Perceive the environment using sensors. ▶ Act on the environment using actuators.

As a human, what sensors and actuators do we have? Consider a software agent. What sensors and actuators does it have?

slide-5
SLIDE 5

5/22

Defjnition of a rational agent

For each possible percept sequence, a rational agent should select an action that is expected to maximize its performance measure, given the evidence provided by the percept sequence and whatever prior knowledge the agent has.

slide-6
SLIDE 6

6/22

Properties of Task Environments

The problems: the task environments The solutions: the rational agents Properties of the task environment:

▶ Fully observable v.s. partially observable ▶ Deterministic v.s. stochastic ▶ Static v.s. dynamic ▶ Episodic v.s. sequential ▶ Known v.s. unknown ▶ Single agent v.s. multi-agent

slide-7
SLIDE 7

7/22

Uncertainty

Given the observations, can the agent determine the state?

▶ Fully observable: The agent knows the state of the world from

the observations.

▶ Partially observable: Many states are possible given an

  • bservation.
slide-8
SLIDE 8

8/22

CQ: Fully versus Partial Observability

CQ: Which pair of environments has difgerent observability? (A) Poker and autonomous cars (B) Chess and medical diagnosis (C) Crossword puzzle and Go

slide-9
SLIDE 9

9/22

Examples of Uncertainty

Come up with some additional examples yourself. Fully observable: Partially observable:

slide-10
SLIDE 10

10/22

Uncertain dynamics

Given the current state and an action, can the agent predict the next state?

▶ Deterministic: The next state is completely determined given

the current state and the action.

▶ Stochastic: The current state and an action can lead to

multiple possible next states.

slide-11
SLIDE 11

11/22

CQ: Deterministic versus Stochastic

CQ: Consider Chess and Poker. Which of the following is correct? (A) Both are deterministic. (B) Both are stochastic. (C) Chess is deterministic. Poker is stochastic. (D) Chess is stochastic. Poker is deterministic.

slide-12
SLIDE 12

12/22

Examples of uncertain dynamics

Come up with some additional examples yourself. Deterministic: Stochastic:

slide-13
SLIDE 13

13/22

An uncertain environment

An environment is uncertain if

▶ It is not fully observable, or ▶ It is not deterministic.

slide-14
SLIDE 14

14/22

Can the environment change?

Can the environment change while the agent interacts with it?

▶ Static: The environment does not change. ▶ Dynamic: The environment changes while the agent interacts

with it.

slide-15
SLIDE 15

15/22

CQ: Static versus dynamic

CQ: Consider autonomous cars and medical diagnosis. Which of the following statement is correct? (A) Both are static. (B) Both are dynamic. (C) Autonomous cars is static. Medical diagnosis is dynamic. (D) Autonomous cars is dynamic. Medical diagnosis is static.

slide-16
SLIDE 16

16/22

Examples of changing environments

Come up with some additional examples yourself. Static: Dynamic

slide-17
SLIDE 17

17/22

Long-term consequence of actions

Can the agent’s current action afgect future actions?

▶ Episodic: The current action does not afgect future actions. ▶ Sequential: The current action could afgect all future actions.

slide-18
SLIDE 18

18/22

CQ: Episodic v.s. Sequential

CQ: Consider crossword puzzle and image classifjcation. Which of the following statement is correct? (A) Both are episodic. (B) Both are sequential. (C) Crossword puzzle is episodic. Image classifjcation is sequential. (D) Crossword puzzle is sequential. Image classifjcation is episodic.

slide-19
SLIDE 19

19/22

Learning the rules of the environment

Does the agent know the rules of the environment?

▶ Known: The agent knows all the rules of the environment. ▶ Unknown: The agent does not know all the rules of the

environment.

slide-20
SLIDE 20

20/22

Number of agents

Does the agent consider all other agents to be part of the environment?

▶ Single agent: The agent assumes that any other agents are

part of the environment.

▶ Multi-agent: The agent explicitly models other agents and

reasons strategically about the other agents.

slide-21
SLIDE 21

21/22

CQ: Single or multi agent

CQ: Is autonomous cars single agent or multi-agent? (A) Defjnitely single agent. (B) Defjnitely multi-agent. (C) It depends.

slide-22
SLIDE 22

22/22

Revisiting the learning goals

By the end of the lecture, you should be able to

▶ Given examples of sensors and actuators. ▶ Defjne rational agents. ▶ Given a task environment, describe its properties. ▶ Given a property, give examples of task environments that

have this property.