Markov Decision Processes CS60077: Reinforcement Learning Abir Das - - PowerPoint PPT Presentation

markov decision processes
SMART_READER_LITE
LIVE PREVIEW

Markov Decision Processes CS60077: Reinforcement Learning Abir Das - - PowerPoint PPT Presentation

Markov Decision Processes CS60077: Reinforcement Learning Abir Das IIT Kharagpur July 26, Aug 01, 02, 08, 2019 Agenda Terminology Markov Decision Process Agenda Understand definitions and notation to be used in the course. Understand


slide-1
SLIDE 1

Markov Decision Processes

CS60077: Reinforcement Learning Abir Das

IIT Kharagpur

July 26, Aug 01, 02, 08, 2019

slide-2
SLIDE 2

Agenda Terminology Markov Decision Process

Agenda

§ Understand definitions and notation to be used in the course. § Understand definition and setup of sequential decision problems.

Abir Das (IIT Kharagpur) CS60077 July 26, Aug 01, 02, 08, 2019 2 / 43

slide-3
SLIDE 3

Agenda Terminology Markov Decision Process

Resources

§ Reinforcement Learning by David Silver [Link] § Deep Reinforcement Learning by Sergey Levine [Link] § SB: Chapter 3

Abir Das (IIT Kharagpur) CS60077 July 26, Aug 01, 02, 08, 2019 3 / 43

slide-4
SLIDE 4

Agenda Terminology Markov Decision Process

Terminology and Notation

Figure credit: S. Levine - CS 294-112 Course, UC Berkeley Abir Das (IIT Kharagpur) CS60077 July 26, Aug 01, 02, 08, 2019 4 / 43

slide-5
SLIDE 5

Agenda Terminology Markov Decision Process

Terminology and Notation

Figure credit: S. Levine - CS 294-112 Course, UC Berkeley Abir Das (IIT Kharagpur) CS60077 July 26, Aug 01, 02, 08, 2019 4 / 43

slide-6
SLIDE 6

Agenda Terminology Markov Decision Process

Terminology and Notation

Figure credit: S. Levine - CS 294-112 Course, UC Berkeley Abir Das (IIT Kharagpur) CS60077 July 26, Aug 01, 02, 08, 2019 4 / 43

slide-7
SLIDE 7

Agenda Terminology Markov Decision Process

Terminology and Notation

  • 1. run away
  • 3. pet
  • 2. ignore

Figure credit: S. Levine - CS 294-112 Course, UC Berkeley Abir Das (IIT Kharagpur) CS60077 July 26, Aug 01, 02, 08, 2019 4 / 43

slide-8
SLIDE 8

Agenda Terminology Markov Decision Process

Terminology and Notation

Figure credit: S. Levine - CS 294-112 Course, UC Berkeley Abir Das (IIT Kharagpur) CS60077 July 26, Aug 01, 02, 08, 2019 4 / 43

slide-9
SLIDE 9

Agenda Terminology Markov Decision Process

Terminology and Notation

Figure credit: S. Levine - CS 294-112 Course, UC Berkeley Abir Das (IIT Kharagpur) CS60077 July 26, Aug 01, 02, 08, 2019 4 / 43

slide-10
SLIDE 10

Agenda Terminology Markov Decision Process

Terminology and Notation

Figure credit: S. Levine - CS 294-112 Course, UC Berkeley Abir Das (IIT Kharagpur) CS60077 July 26, Aug 01, 02, 08, 2019 4 / 43

slide-11
SLIDE 11

Agenda Terminology Markov Decision Process

Terminology and Notation

Figure credit: S. Levine - CS 294-112 Course, UC Berkeley Abir Das (IIT Kharagpur) CS60077 July 26, Aug 01, 02, 08, 2019 4 / 43

slide-12
SLIDE 12

Agenda Terminology Markov Decision Process

Terminology and Notation

Figure credit: S. Levine - CS 294-112 Course, UC Berkeley Abir Das (IIT Kharagpur) CS60077 July 26, Aug 01, 02, 08, 2019 4 / 43

slide-13
SLIDE 13

Agenda Terminology Markov Decision Process

Terminology and Notation

Figure credit: S. Levine - CS 294-112 Course, UC Berkeley Abir Das (IIT Kharagpur) CS60077 July 26, Aug 01, 02, 08, 2019 4 / 43

slide-14
SLIDE 14

Agenda Terminology Markov Decision Process

Markov Property

The future is independent of the past given the present. Definition A state St is Markov if and only if P(St+1|St) = P(St+1|St, St−1, St−2, · · · , S1)

Andrey Markov

§ Once the present state is known, the history may be thrown away § The current state is a sufficient statistic of the future

Abir Das (IIT Kharagpur) CS60077 July 26, Aug 01, 02, 08, 2019 5 / 43

slide-15
SLIDE 15

Agenda Terminology Markov Decision Process

Markov Chain

A Markov Chain or Markov Process is temporal process i.e., a sequence of random states S1, S2, · · · where the states obey the Markov property. Definition A Markov Process is a tuple S, P, where § S is the state space (can be continuous or discrete) § P is the state transition probability matrix. P also called an operator

P =      P11 P12 · · · P1n P21 P22 · · · P2n . . . . . . ... . . . Pn1 Pn2 · · · Pnn     

where Pss′ = P(St+1 = s′|St = s)

Abir Das (IIT Kharagpur) CS60077 July 26, Aug 01, 02, 08, 2019 6 / 43

slide-16
SLIDE 16

Agenda Terminology Markov Decision Process

Markov Chain

P =      P11 P12 · · · P1n P21 P22 · · · P2n . . . . . . ... . . . Pn1 Pn2 · · · Pnn     

Let µt,i = P(St = si) and µt =

  • µt,1, µt,2, · · · , µt,n

T , i.e., µt is a vector

  • f probabilities, then µt+1 = PT µt

     µt+1,1 µt+1,2 . . . µt+1,n      =      P11 P12 · · · P1n P21 P22 · · · P2n . . . . . . ... . . . Pn1 Pn2 · · · Pnn     

T 

    µt,1 µt,2 . . . µt,n     

Abir Das (IIT Kharagpur) CS60077 July 26, Aug 01, 02, 08, 2019 7 / 43

slide-17
SLIDE 17

Agenda Terminology Markov Decision Process

Markov Chain

P =      P11 P12 · · · P1n P21 P22 · · · P2n . . . . . . ... . . . Pn1 Pn2 · · · Pnn     

Let µt,i = P(St = si) and µt =

  • µt,1, µt,2, · · · , µt,n

T , i.e., µt is a vector

  • f probabilities, then µt+1 = PT µt

     µt+1,1 µt+1,2 . . . µt+1,n      =      P11 P12 · · · P1n P21 P22 · · · P2n . . . . . . ... . . . Pn1 Pn2 · · · Pnn     

T 

    µt,1 µt,2 . . . µt,n     

𝑞 𝑡# 𝑡#$%)

Abir Das (IIT Kharagpur) CS60077 July 26, Aug 01, 02, 08, 2019 7 / 43

slide-18
SLIDE 18

Agenda Terminology Markov Decision Process

Student Markov Process

Class 1 Class 2 Class 3 Pass Pub Facebook Sleep

0.9 0.1 0.5 0.5 0.8 0.6 0.2 0.4 0.4 0.4 1.0 0.2

Figure credit: David Silver, DeepMind Abir Das (IIT Kharagpur) CS60077 July 26, Aug 01, 02, 08, 2019 8 / 43

slide-19
SLIDE 19

Agenda Terminology Markov Decision Process

Student Markov Process - Episodes

Class 1 Class 2 Class 3 Pass Pub Facebook Sleep 0.9 0.1 0.5 0.5 0.8 0.6 0.2 0.4 0.4 0.4 1.0 0.2

Figure credit: David Silver, DeepMind

Sample episodes for Student Markov process starting from S1 = C1 § C1 C2 C3 Pass Sleep § C1 FB FB C1 C2 Sleep § C1 C2 C3 Pub C2 C3 Pass Sleep § C1 FB FB C1 C2 C3 Pub C1 FB FB FB C1 C2 C3 Pub C2 Sleep

Abir Das (IIT Kharagpur) CS60077 July 26, Aug 01, 02, 08, 2019 9 / 43

slide-20
SLIDE 20

Agenda Terminology Markov Decision Process

Student Markov Process - Transition Matrix

Class 1 Class 2 Class 3 Pass Pub Facebook Sleep 0.9 0.1 0.5 0.5 0.8 0.6 0.2 0.4 0.4 0.4 1.0 0.2

Figure credit: David Silver, DeepMind

         

C1 C2 C3 P ass P ub F B Sleep C1

0.5 0.5

C2

0.8 0.2

C3

0.6 0.4

P ass

1.0

P ub

0.2 0.4 0.4

F B

0.1 0.9

Sleep

1.0          

Abir Das (IIT Kharagpur) CS60077 July 26, Aug 01, 02, 08, 2019 10 / 43

slide-21
SLIDE 21

Agenda Terminology Markov Decision Process

Markov Reward Process

A Markov reward process is a Markov process with rewards. Definition A Markov Reward Process is a tuple S, P, R, γ, where § S is the state space (can be continuous or discrete) § P is the state transition probability matrix. P also called an operator. Pss′ = P(St+1 = s′|St = s) § R is a reward function, R = E

  • Rt+1|St = s
  • = R(s)

§ γ is a discount factor, γ ∈

  • 0, 1
  • Abir Das (IIT Kharagpur)

CS60077 July 26, Aug 01, 02, 08, 2019 11 / 43

slide-22
SLIDE 22

Agenda Terminology Markov Decision Process

Markov Reward Process

A Markov reward process is a Markov process with rewards. Definition A Markov Reward Process is a tuple S, P, R, γ, where § S is the state space (can be continuous or discrete) § P is the state transition probability matrix. P also called an operator. Pss′ = P(St+1 = s′|St = s) § R is a reward function, R = E

  • Rt+1|St = s
  • = R(s)

§ γ is a discount factor, γ ∈

  • 0, 1
  • Abir Das (IIT Kharagpur)

CS60077 July 26, Aug 01, 02, 08, 2019 11 / 43

slide-23
SLIDE 23

Agenda Terminology Markov Decision Process

Student Markov Reward Process

Class 1 Class 2 Class 3 Pass Pub Facebook Sleep

0.9 0.1 0.5 0.5 0.8 0.6 0.2 0.4 0.4 1.0

R=0 R=-1 R=-2 R=-2 R=-2 R=+10 R=+1

0.2 0.4

Figure credit: David Silver, DeepMind Abir Das (IIT Kharagpur) CS60077 July 26, Aug 01, 02, 08, 2019 12 / 43

slide-24
SLIDE 24

Agenda Terminology Markov Decision Process

Return

Definition The return Gt is the total discounted reward from timestep t. Gt = Rt+1 + γRt+2 + · · · =

  • k=0

γkRt+k+1 (1) § γ ∈

  • 0, 1
  • is the discounted present value of the future rewards.

§ Immediate rewards are valued above delayed rewards.

◮ γ close to 0 leads to “myopic” evaluation. ◮ γ close to 1 leads to “far-sighted” evaluation.

Abir Das (IIT Kharagpur) CS60077 July 26, Aug 01, 02, 08, 2019 13 / 43

slide-25
SLIDE 25

Agenda Terminology Markov Decision Process

Why Discount?

Most Markov reward and decision processes are discounted. Why?

Abir Das (IIT Kharagpur) CS60077 July 26, Aug 01, 02, 08, 2019 14 / 43

slide-26
SLIDE 26

Agenda Terminology Markov Decision Process

Why Discount?

Most Markov reward and decision processes are discounted. Why? § Uncertainty about the future may not be fully represented

Abir Das (IIT Kharagpur) CS60077 July 26, Aug 01, 02, 08, 2019 14 / 43

slide-27
SLIDE 27

Agenda Terminology Markov Decision Process

Why Discount?

Most Markov reward and decision processes are discounted. Why? § Uncertainty about the future may not be fully represented § Immediate rewards are valued above delayed rewards.

Abir Das (IIT Kharagpur) CS60077 July 26, Aug 01, 02, 08, 2019 14 / 43

slide-28
SLIDE 28

Agenda Terminology Markov Decision Process

Why Discount?

Most Markov reward and decision processes are discounted. Why? § Uncertainty about the future may not be fully represented § Immediate rewards are valued above delayed rewards. § Avoids infinite returns in cyclic Markov processes or infinite horizon problems.

Abir Das (IIT Kharagpur) CS60077 July 26, Aug 01, 02, 08, 2019 14 / 43

slide-29
SLIDE 29

Agenda Terminology Markov Decision Process

Why Discount?

Most Markov reward and decision processes are discounted. Why? § Uncertainty about the future may not be fully represented § Immediate rewards are valued above delayed rewards. § Avoids infinite returns in cyclic Markov processes or infinite horizon problems. § Mathematically convenient. We can use stationarity property to better effect.

Abir Das (IIT Kharagpur) CS60077 July 26, Aug 01, 02, 08, 2019 14 / 43

slide-30
SLIDE 30

Agenda Terminology Markov Decision Process

Why Discount?

Most Markov reward and decision processes are discounted. Why? § Uncertainty about the future may not be fully represented § Immediate rewards are valued above delayed rewards. § Avoids infinite returns in cyclic Markov processes or infinite horizon problems. § Mathematically convenient. We can use stationarity property to better effect. It is sometimes possible to use average rewards also to bound the return to finite values. More of it to follow when we discuss Markov Decision Process

Abir Das (IIT Kharagpur) CS60077 July 26, Aug 01, 02, 08, 2019 14 / 43

slide-31
SLIDE 31

Agenda Terminology Markov Decision Process

Value Function

The value function v(s) gives the long-term value of state s Definition The state value function v(s) of an MRP is the expected return starting from state s v(s) = E

  • Gt|St = s
  • (2)

Abir Das (IIT Kharagpur) CS60077 July 26, Aug 01, 02, 08, 2019 15 / 43

slide-32
SLIDE 32

Agenda Terminology Markov Decision Process

Example Student MRP Returns

Sample returns for Student MRP: Starting from S1 = C1 with γ = 1

2

G1 = R2 + γR3 + · · · + γT−1RT+1

§ C1 C2 C3 Pass Sleep § C1 FB FB C1 C2 Sleep § C1 C2 C3 Pub C2 C3 Pass Sleep § C1 FB FB C1 C2 C3 Pub C1 FB FB FB C1 C2 C3 Pub C2 Sleep § −2− 1

2 ∗2− 1 4 ∗2+ 1 8 ∗10 = −2.25

§ −2− 1

2 ∗1− 1 4 ∗1− 1 8 ∗2− 1 16 ∗2 =

−3.125 § −2 − 1

2 ∗ 2 − 1 4 ∗ 2 + 1 8 ∗ 1 − − 1 16 ∗

2 − 1

32 ∗ 2 + 1 64 ∗ 10 = −3.41

§ −2 − 1

2 ∗ 1 − 1 4 ∗ 1 − 1 8 ∗ 2 − 1 16 ∗

2 + · · · = −3.20

Abir Das (IIT Kharagpur) CS60077 July 26, Aug 01, 02, 08, 2019 16 / 43

slide-33
SLIDE 33

Agenda Terminology Markov Decision Process

State-Value Function for Student MRP (1)

Class 1 Class 2 Class 3 Pass Pub Facebook Sleep

0.9 0.1 0.5 0.5 0.8 0.6 0.2 0.4 0.4 1.0

R=0 R=-1 R=-2 R=-2 R=-2 R=+10 R=+1

  • 1
  • 2
  • 2
  • 2

10 1 V(s) for 𝛿 = 0

0.2 0.4

Figure credit: David Silver, DeepMind Abir Das (IIT Kharagpur) CS60077 July 26, Aug 01, 02, 08, 2019 17 / 43

slide-34
SLIDE 34

Agenda Terminology Markov Decision Process

State-Value Function for Student MRP (2)

Class 1 Class 2 Class 3 Pass Pub Facebook Sleep

0.9 0.1 0.5 0.5 0.8 0.6 0.2 0.4 0.4 1.0

R=0 R=-1 R=-2 R=-2 R=-2 R=+10 R=+1

  • 7.6
  • 5.0

0.9 4.1 10 1.9 V(s) for 𝛿 = 0.9

0.2 0.4

Figure credit: David Silver, DeepMind Abir Das (IIT Kharagpur) CS60077 July 26, Aug 01, 02, 08, 2019 18 / 43

slide-35
SLIDE 35

Agenda Terminology Markov Decision Process

State-Value Function for Student MRP (3)

Class 1 Class 2 Class 3 Pass Pub Facebook Sleep

0.9 0.1 0.5 0.5 0.8 0.6 0.2 0.4 0.4 1.0

R=0 R=-1 R=-2 R=-2 R=-2 R=+10 R=+1

  • 23
  • 13

1.5 4.3 10 +0.8 V(s) for 𝛿 = 1

0.2 0.4

Figure credit: David Silver, DeepMind Abir Das (IIT Kharagpur) CS60077 July 26, Aug 01, 02, 08, 2019 19 / 43

slide-36
SLIDE 36

Agenda Terminology Markov Decision Process

Bellman Equation for MRPs

The value function can be decomposed into two parts: § immediate reward R(s) § discounted value of successor state γv(s′) v(s) = R(s) + γEs′∈S

  • v(s′)
  • = R(s) + γ
  • s′∈S

Pss′v(s′) (3)

Abir Das (IIT Kharagpur) CS60077 July 26, Aug 01, 02, 08, 2019 20 / 43

slide-37
SLIDE 37

Agenda Terminology Markov Decision Process

Bellman Equation for MRPs

The value function can be decomposed into two parts: § immediate reward R(s) § discounted value of successor state γv(s′) v(s) = R(s) + γEs′∈S

  • v(s′)
  • = R(s) + γ
  • s′∈S

Pss′v(s′) (3)

s’ s’’ V(s’) V(s) V(s’’) r s

Abir Das (IIT Kharagpur) CS60077 July 26, Aug 01, 02, 08, 2019 20 / 43

slide-38
SLIDE 38

Agenda Terminology Markov Decision Process

Bellman Equation for MRPs - Proof

v(s)=E

  • Gt|St = s
  • =E
  • Rt+1 + γRt+2 + γ2Rt+3 + γ3Rt+4 + · · · |St = s
  • Abir Das (IIT Kharagpur)

CS60077 July 26, Aug 01, 02, 08, 2019 21 / 43

slide-39
SLIDE 39

Agenda Terminology Markov Decision Process

Bellman Equation for MRPs - Proof

v(s)=E

  • Gt|St = s
  • =E
  • Rt+1 + γRt+2 + γ2Rt+3 + γ3Rt+4 + · · · |St = s
  • =E
  • Rt+1(St) + γRt+2(St+1) + γ2Rt+3(St+2) + γ3Rt+4(St+3) + · · · |St = s
  • Abir Das (IIT Kharagpur)

CS60077 July 26, Aug 01, 02, 08, 2019 21 / 43

slide-40
SLIDE 40

Agenda Terminology Markov Decision Process

Bellman Equation for MRPs - Proof

v(s)=E

  • Gt|St = s
  • =E
  • Rt+1 + γRt+2 + γ2Rt+3 + γ3Rt+4 + · · · |St = s
  • =E
  • Rt+1(St) + γRt+2(St+1) + γ2Rt+3(St+2) + γ3Rt+4(St+3) + · · · |St = s
  • =
  • St+1,St+2,···
  • P(St+1, St+2, · · · |St = s)
  • Rt+1(St) + γRt+2(St+1)+

γ2Rt+3(St+2) + γ3Rt+4(St+3) + · · ·

  • Abir Das (IIT Kharagpur)

CS60077 July 26, Aug 01, 02, 08, 2019 21 / 43

slide-41
SLIDE 41

Agenda Terminology Markov Decision Process

Bellman Equation for MRPs - Proof

v(s)=E

  • Gt|St = s
  • =E
  • Rt+1 + γRt+2 + γ2Rt+3 + γ3Rt+4 + · · · |St = s
  • =E
  • Rt+1(St) + γRt+2(St+1) + γ2Rt+3(St+2) + γ3Rt+4(St+3) + · · · |St = s
  • =
  • St+1,St+2,···
  • P(St+1, St+2, · · · |St = s)
  • Rt+1(St) + γRt+2(St+1)+

γ2Rt+3(St+2) + γ3Rt+4(St+3) + · · ·

  • =
  • St+1,St+2,···

P(St+1, St+2, · · · |St = s)Rt+1(St)+ γ

  • St+1,St+2,···
  • P(

St+1, St+2, · · · |St =s )

  • Rt+2(St+1) + γRt+3(St+2)+

γ2Rt+4(St+3) + · · ·

  • Abir Das (IIT Kharagpur)

CS60077 July 26, Aug 01, 02, 08, 2019 21 / 43

slide-42
SLIDE 42

Agenda Terminology Markov Decision Process

Bellman Equation for MRPs - Proof

= Rt+1(St)

  • St+1,St+2,···

P(St+1, St+2, · · · |St = s)+ γ

  • St+1,St+2,···
  • P(

St+1, St+2, · · · |St =s )

  • Rt+2(St+1) + γRt+3(St+2)+

γ2Rt+4(St+3) + · · ·

  • Abir Das (IIT Kharagpur)

CS60077 July 26, Aug 01, 02, 08, 2019 22 / 43

slide-43
SLIDE 43

Agenda Terminology Markov Decision Process

Bellman Equation for MRPs - Proof

= Rt+1(St) ✘✘✘✘✘✘✘✘✘✘✘✘✘✘ ✘ ✿1

  • St+1,St+2,···

P(St+1, St+2, · · · |St = s)+ γ

  • St+1,St+2,···
  • P(

St+1, St+2, · · · |St =s )

  • Rt+2(St+1) + γRt+3(St+2)+

γ2Rt+4(St+3) + · · ·

  • = Rt+1(St) + γ
  • St+1,St+2,···
  • P(

St+1, St+2, · · · |St =s )

  • Rt+2(St+1) + γRt+3(St+2)+

γ2Rt+4(St+3) + · · ·

  • Abir Das (IIT Kharagpur)

CS60077 July 26, Aug 01, 02, 08, 2019 22 / 43

slide-44
SLIDE 44

Agenda Terminology Markov Decision Process

Bellman Equation for MRPs - Proof

= Rt+1(St) ✘✘✘✘✘✘✘✘✘✘✘✘✘✘ ✘ ✿1

  • St+1,St+2,···

P(St+1, St+2, · · · |St = s)+ γ

  • St+1,St+2,···
  • P(

St+1, St+2, · · · |St =s )

  • Rt+2(St+1) + γRt+3(St+2)+

γ2Rt+4(St+3) + · · ·

  • = Rt+1(St) + γ
  • St+1,St+2,···
  • P(

St+1, St+2, · · · |St =s )

  • Rt+2(St+1) + γRt+3(St+2)+

γ2Rt+4(St+3) + · · ·

  • = Rt+1(St) + γ
  • St+1,St+2,···
  • P(

St+2, · · · |St+1, St =s )P(St+1|St =s)

  • Rt+2(

St+1)+ γRt+3( St+2) + γ2Rt+4(St+3) + · · ·

  • Abir Das (IIT Kharagpur)

CS60077 July 26, Aug 01, 02, 08, 2019 22 / 43

slide-45
SLIDE 45

Agenda Terminology Markov Decision Process

Bellman Equation for MRPs - Proof

= Rt+1(St) + γ

  • St+1,St+2,···
  • P(

St+2, · · · |St+1)P(St+1|St =s)

  • Rt+2(

St+1) +γRt+3( St+2)+ γ2Rt+4(St+3) + · · ·

  • [Conditional independence (Ref eq. (7))]

Abir Das (IIT Kharagpur) CS60077 July 26, Aug 01, 02, 08, 2019 23 / 43

slide-46
SLIDE 46

Agenda Terminology Markov Decision Process

Bellman Equation for MRPs - Proof

= Rt+1(St) + γ

  • St+1,St+2,···
  • P(

St+2, · · · |St+1)P(St+1|St =s)

  • Rt+2(

St+1) +γRt+3( St+2)+ γ2Rt+4(St+3) + · · ·

  • [Conditional independence (Ref eq. (7))]

= Rt+1(St) + γ

  • St+1
  • St+2,St+3,···
  • P(

St+2, · · · |St+1)P(St+1|St =s)

  • Rt+2(

St+1)+ γRt+3( St+2) + γ2Rt+4(St+3) + · · ·

  • Abir Das (IIT Kharagpur)

CS60077 July 26, Aug 01, 02, 08, 2019 23 / 43

slide-47
SLIDE 47

Agenda Terminology Markov Decision Process

Bellman Equation for MRPs - Proof

= Rt+1(St) + γ

  • St+1,St+2,···
  • P(

St+2, · · · |St+1)P(St+1|St =s)

  • Rt+2(

St+1) +γRt+3( St+2)+ γ2Rt+4(St+3) + · · ·

  • [Conditional independence (Ref eq. (7))]

= Rt+1(St) + γ

  • St+1
  • St+2,St+3,···
  • P(

St+2, · · · |St+1)P(St+1|St =s)

  • Rt+2(

St+1)+ γRt+3( St+2) + γ2Rt+4(St+3) + · · ·

  • = Rt+1(St) + γ
  • St+1

P(St+1|St =s)

  • St+2,St+3,···
  • P(

St+2, · · · |St+1)

  • Rt+2(

St+1)+ γRt+3( St+2) + γ2Rt+4(St+3) + · · ·

  • Abir Das (IIT Kharagpur)

CS60077 July 26, Aug 01, 02, 08, 2019 23 / 43

slide-48
SLIDE 48

Agenda Terminology Markov Decision Process

Bellman Equation for MRPs - Proof

= Rt+1(St) + γ

  • St+1,St+2,···
  • P(

St+2, · · · |St+1)P(St+1|St =s)

  • Rt+2(

St+1) +γRt+3( St+2)+ γ2Rt+4(St+3) + · · ·

  • [Conditional independence (Ref eq. (7))]

= Rt+1(St) + γ

  • St+1
  • St+2,St+3,···
  • P(

St+2, · · · |St+1)P(St+1|St =s)

  • Rt+2(

St+1)+ γRt+3( St+2) + γ2Rt+4(St+3) + · · ·

  • = Rt+1(St) + γ
  • St+1

P(St+1|St =s)

  • St+2,St+3,···
  • P(

St+2, · · · |St+1)

  • Rt+2(

St+1)+ γRt+3( St+2) + γ2Rt+4(St+3) + · · ·

  • = Rt+1(St) + γ
  • St+1

P(St+1|St =s)v(St+1)

Abir Das (IIT Kharagpur) CS60077 July 26, Aug 01, 02, 08, 2019 23 / 43

slide-49
SLIDE 49

Agenda Terminology Markov Decision Process

Bellman Equation for MRPs - Proof

= Rt+1(St) + γ

  • St+1,St+2,···
  • P(

St+2, · · · |St+1)P(St+1|St =s)

  • Rt+2(

St+1) +γRt+3( St+2)+ γ2Rt+4(St+3) + · · ·

  • [Conditional independence (Ref eq. (7))]

= Rt+1(St) + γ

  • St+1
  • St+2,St+3,···
  • P(

St+2, · · · |St+1)P(St+1|St =s)

  • Rt+2(

St+1)+ γRt+3( St+2) + γ2Rt+4(St+3) + · · ·

  • = Rt+1(St) + γ
  • St+1

P(St+1|St =s)

  • St+2,St+3,···
  • P(

St+2, · · · |St+1)

  • Rt+2(

St+1)+ γRt+3( St+2) + γ2Rt+4(St+3) + · · ·

  • = Rt+1(St) + γ
  • St+1

P(St+1|St =s)v(St+1) = Rt+1(St =s) + γ

  • s′∈S

P(St+1 =s′|St =s)v(St+1 =s′)

Abir Das (IIT Kharagpur) CS60077 July 26, Aug 01, 02, 08, 2019 23 / 43

slide-50
SLIDE 50

Agenda Terminology Markov Decision Process

Bellman Equation for MRPs - Proof

= Rt+1(St) + γ

  • St+1,St+2,···
  • P(

St+2, · · · |St+1)P(St+1|St =s)

  • Rt+2(

St+1) +γRt+3( St+2)+ γ2Rt+4(St+3) + · · ·

  • [Conditional independence (Ref eq. (7))]

= Rt+1(St) + γ

  • St+1
  • St+2,St+3,···
  • P(

St+2, · · · |St+1)P(St+1|St =s)

  • Rt+2(

St+1)+ γRt+3( St+2) + γ2Rt+4(St+3) + · · ·

  • = Rt+1(St) + γ
  • St+1

P(St+1|St =s)

  • St+2,St+3,···
  • P(

St+2, · · · |St+1)

  • Rt+2(

St+1)+ γRt+3( St+2) + γ2Rt+4(St+3) + · · ·

  • = Rt+1(St) + γ
  • St+1

P(St+1|St =s)v(St+1) = Rt+1(St =s) + γ

  • s′∈S

P(St+1 =s′|St =s)v(St+1 =s′)= R(s) + γ

  • s′∈S

Pss′v(s′)

Abir Das (IIT Kharagpur) CS60077 July 26, Aug 01, 02, 08, 2019 23 / 43

slide-51
SLIDE 51

Agenda Terminology Markov Decision Process

Bellman Equation in Matrix Form

So, we have seen, v(s) = R(s) + γ

  • s′∈S

Pss′v(s′) Where are the time subscripts? Hint: Think about (1). Definition of value function, (2). Expectation operation.

Abir Das (IIT Kharagpur) CS60077 July 26, Aug 01, 02, 08, 2019 24 / 43

slide-52
SLIDE 52

Agenda Terminology Markov Decision Process

Bellman Equation in Matrix Form

So, we have seen, v(s) = R(s) + γ

  • s′∈S

Pss′v(s′) Where are the time subscripts? Hint: Think about (1). Definition of value function, (2). Expectation operation. The Bellman equation can be expressed concisely using matrices. v = R + γPv where v and R are column vectors with one entry per state.

     v(s1) v(s2) . . . v(sn)      =      R(s1) R(s2) . . . R(sn)      + γ      P11 P12 · · · P1n P21 P22 · · · P2n . . . . . . ... . . . Pn1 Pn2 · · · Pnn           v(s1) v(s2) . . . v(sn)     

Abir Das (IIT Kharagpur) CS60077 July 26, Aug 01, 02, 08, 2019 24 / 43

slide-53
SLIDE 53

Agenda Terminology Markov Decision Process

Solving Bellman Equation

§ The Bellman equation being a linear equation, it can be solved directly. v = R + γPv

  • I − γP
  • v = R

v =

  • I − γP

−1R § As computational complexity is O(n3) for n states, direct solution is

  • nly feasible for small MRPs.

§ There are many iterative methods for large MRPs, e.g., Dynamic programing, Monte-Carlo, Temporal difference learning

Abir Das (IIT Kharagpur) CS60077 July 26, Aug 01, 02, 08, 2019 25 / 43

slide-54
SLIDE 54

Agenda Terminology Markov Decision Process

Existence of Solution to Bellman Equation

§ We need to show that

  • I − γP
  • is invertible and for that we will use

the following result from linear algebra - The inverse of a matrix exists if and only if all its eigenvalues are non-zero.

✶ ✶ ✶

Abir Das (IIT Kharagpur) CS60077 July 26, Aug 01, 02, 08, 2019 26 / 43

slide-55
SLIDE 55

Agenda Terminology Markov Decision Process

Existence of Solution to Bellman Equation

§ We need to show that

  • I − γP
  • is invertible and for that we will use

the following result from linear algebra - The inverse of a matrix exists if and only if all its eigenvalues are non-zero. § For a stochastic matrix (row sum equal to 1 and all entries are ≥ 0), the largest eigenvalue is 1.

✶ ✶ ✶

Abir Das (IIT Kharagpur) CS60077 July 26, Aug 01, 02, 08, 2019 26 / 43

slide-56
SLIDE 56

Agenda Terminology Markov Decision Process

Existence of Solution to Bellman Equation

§ We need to show that

  • I − γP
  • is invertible and for that we will use

the following result from linear algebra - The inverse of a matrix exists if and only if all its eigenvalues are non-zero. § For a stochastic matrix (row sum equal to 1 and all entries are ≥ 0), the largest eigenvalue is 1.

Proof As P is a stchoastic matrix, P✶ = ✶ where ✶ = [1, 1, · · · 1]T . This means 1 is an eigenvalue of P. Now, lets suppose ∃ λ > 1 and non-zero x such that Px = λx. Since the rows of P are non-negative and sum to 1, each element of vector Px is a convex combination of the components of the vector x. A convex combination can’t be greater than xmax, the largest component of x. However, as λ>1, at least one element (λxmax) in the R.H.S. (i.e., in λx) is greater than xmax. This is a contradiction and so λ>1 is not possible.

Abir Das (IIT Kharagpur) CS60077 July 26, Aug 01, 02, 08, 2019 26 / 43

slide-57
SLIDE 57

Agenda Terminology Markov Decision Process

Existence of Solution to Bellman Equation

§ So the largest eigenvalue of P is 1.

Abir Das (IIT Kharagpur) CS60077 July 26, Aug 01, 02, 08, 2019 27 / 43

slide-58
SLIDE 58

Agenda Terminology Markov Decision Process

Existence of Solution to Bellman Equation

§ So the largest eigenvalue of P is 1.

Theorem and its proof For all eigenvalues λi of a square matrix A and corresponding eigenvectors vi such that Avi = λivi, eig(I + γA) = 1 + γλi [γ is any scalar] Proof: Avi = λivi γAvi = γλivi vi + γAvi = vi + γλivi (I + γA)vi = (1 + γλi)vi

Abir Das (IIT Kharagpur) CS60077 July 26, Aug 01, 02, 08, 2019 27 / 43

slide-59
SLIDE 59

Agenda Terminology Markov Decision Process

Existence of Solution to Bellman Equation

§ So the largest eigenvalue of P is 1.

Theorem and its proof For all eigenvalues λi of a square matrix A and corresponding eigenvectors vi such that Avi = λivi, eig(I + γA) = 1 + γλi [γ is any scalar] Proof: Avi = λivi γAvi = γλivi vi + γAvi = vi + γλivi (I + γA)vi = (1 + γλi)vi

§ So the smallest eigenvalue of

  • I − γP
  • is 1 − γ. For γ < 1 which is

> 0. And hence,

  • I − γP
  • is invertible.

Abir Das (IIT Kharagpur) CS60077 July 26, Aug 01, 02, 08, 2019 27 / 43

slide-60
SLIDE 60

Agenda Terminology Markov Decision Process

Markov Decision Process

A Markov decision process is a Markov reward process with actions. Definition A Markov Decision Process is a tuple S, A, P, R, γ, where § S is the state space (can be continuous or discrete) § A is the action space (can be continuous or discrete) § P is the state transition probability matrix. Pa

ss′ = P(St+1 = s′|St = s, At = a) = p(s′/s, a)

§ R is a reward function, R = E

  • Rt+1|St = s, At = a
  • = R(s, a)

§ γ is a discount factor, γ ∈

  • 0, 1
  • Abir Das (IIT Kharagpur)

CS60077 July 26, Aug 01, 02, 08, 2019 28 / 43

slide-61
SLIDE 61

Agenda Terminology Markov Decision Process

Example: Student MDP

0.2 0.4 R = -1

Facebook

0.4 R = -1

Quit

R = -1

Facebook

R = 0

Sleep

R = -2

Study

R = -2

Study

R = +1

Pub

R = +10

Study

Figure credit: David Silver, DeepMind Abir Das (IIT Kharagpur) CS60077 July 26, Aug 01, 02, 08, 2019 29 / 43

slide-62
SLIDE 62

Agenda Terminology Markov Decision Process

Policy

Definition A policy π is a distribution over actions given states, π(a/s) = P

  • At = a|St = s
  • § The Markov property means the policy depends on the current state

(not the history) § The policy can be either deterministic or stochastic § The policy can be either stationary or non-stationary

Abir Das (IIT Kharagpur) CS60077 July 26, Aug 01, 02, 08, 2019 30 / 43

slide-63
SLIDE 63

Agenda Terminology Markov Decision Process

Policy

§ For a deterministic environment p(s′/s, a) = 1, else for stochastic environment 0 ≤ p(s′/s, a) ≤ 1 § In a stochastic environment, there is always some chance to end up in s′ starting from state s and taking any action.

S → s’

a

§ So, probability of ending up in state s′ from s irrespective of the action (i.e., taking any action according to the policy), = probability

  • f taking action 1 from state s× probability of ending up in state s′

taking action 1 + probability of taking action 2 from state s× probability of ending up in state s′ taking action 2 + · · · § This means pπ(s′|s) =

a

π(a|s)p(s′|s, a) § Similarly, the one-step expected reward for following policy π is given by rπ(s) =

a

π(a|s)r(s, a) § Side note: The above is given by rπ(s) =

a

π(a|s)

s′ p(s′|s, a)r(s, a, s′)

when reward is a function of the transiting state s′ also.

Abir Das (IIT Kharagpur) CS60077 July 26, Aug 01, 02, 08, 2019 31 / 43

slide-64
SLIDE 64

Agenda Terminology Markov Decision Process

Value Functions

Definition The state-value function vπ(s) of an MDP is the expected return starting from state s, and then following policy π vπ(s) = Eπ

  • Gt|St = s
  • (4)

Definition The action-value function qπ(s, a) of an MDP is the expected return starting from state s, taking action a, and then following policy π qπ(s, a) = Eπ

  • Gt|St = s, At = a
  • (5)

Abir Das (IIT Kharagpur) CS60077 July 26, Aug 01, 02, 08, 2019 32 / 43

slide-65
SLIDE 65

Agenda Terminology Markov Decision Process

Example: State-Value function for Student MDP

0.2 0.4 R = -1

Facebook

0.4 R = -1

Quit

R = -1

Facebook

R = 0

Sleep

R = -2

Study

R = -2

Study

R = +1

Pub

R = +10

Study

  • 2.3
  • 1.3

2.7 7.4 𝑤" 𝑡 for 𝜌 𝑏 𝑡 = 0.5, 𝛿 = 1

Figure credit: David Silver, DeepMind Abir Das (IIT Kharagpur) CS60077 July 26, Aug 01, 02, 08, 2019 33 / 43

slide-66
SLIDE 66

Agenda Terminology Markov Decision Process

Relation between vπ and qπ

𝑤"(𝑡) 𝑡 𝑟"(𝑡, 𝑏) a a′𝑟"(𝑡, 𝑏′)

vπ(s) =

  • a∈A

π(a|s)qπ(s, a)

Abir Das (IIT Kharagpur) CS60077 July 26, Aug 01, 02, 08, 2019 34 / 43

slide-67
SLIDE 67

Agenda Terminology Markov Decision Process

Relation between vπ and qπ

𝑤"(𝑡) 𝑡 𝑟"(𝑡, 𝑏) a a′𝑟"(𝑡, 𝑏′)

vπ(s) =

  • a∈A

π(a|s)qπ(s, a)

𝑟"(𝑡, 𝑏) 𝑡 𝑤"(𝑡′) 𝑡′ 𝑡′′ 𝑤"(𝑡′′)

𝑠

qπ(s, a) = r(s, a)+γ

  • s′∈S

p(s′|s, a)vπ(s′)

Abir Das (IIT Kharagpur) CS60077 July 26, Aug 01, 02, 08, 2019 34 / 43

slide-68
SLIDE 68

Agenda Terminology Markov Decision Process

Relation between vπ and qπ

𝑤"(𝑡) 𝑡 𝑟"(𝑡, 𝑏) a a′𝑟"(𝑡, 𝑏′)

vπ(s) =

  • a∈A

π(a|s)qπ(s, a)

𝑟"(𝑡, 𝑏) 𝑡 𝑤"(𝑡′) 𝑡′ 𝑡′′ 𝑤"(𝑡′′)

𝑠

qπ(s, a) = r(s, a)+γ

  • s′∈S

p(s′|s, a)vπ(s′)

𝑤"(𝑡) 𝑡 𝑏 𝑤"(𝑡′) 𝑡′

𝑠

vπ(s) =

  • a∈A

π(a|s)

  • r(s, a) +

γ

  • s′∈S

p(s′|s, a)vπ(s′)

  • Abir Das (IIT Kharagpur)

CS60077 July 26, Aug 01, 02, 08, 2019 34 / 43

slide-69
SLIDE 69

Agenda Terminology Markov Decision Process

Relation between vπ and qπ

𝑤"(𝑡) 𝑡 𝑟"(𝑡, 𝑏) a a′𝑟"(𝑡, 𝑏′)

vπ(s) =

  • a∈A

π(a|s)qπ(s, a)

𝑟"(𝑡, 𝑏) 𝑡 𝑤"(𝑡′) 𝑡′ 𝑡′′ 𝑤"(𝑡′′)

𝑠

qπ(s, a) = r(s, a)+γ

  • s′∈S

p(s′|s, a)vπ(s′)

𝑤"(𝑡) 𝑡 𝑏 𝑤"(𝑡′) 𝑡′

𝑠

vπ(s) =

  • a∈A

π(a|s)

  • r(s, a) +

γ

  • s′∈S

p(s′|s, a)vπ(s′)

  • 𝑟"(𝑡, 𝑏)

𝑡 𝑡′

𝑠

𝑟"(𝑡′, 𝑏′)

qπ(s, a) = r(s, a) + γ

  • s′∈S

p(s′|s, a)

a′∈A

π(a′|s′)qπ(s′, a′)

  • Abir Das (IIT Kharagpur)

CS60077 July 26, Aug 01, 02, 08, 2019 34 / 43

slide-70
SLIDE 70

Agenda Terminology Markov Decision Process

Bellman Expectation Equations

Like MRPs, the value function can be decomposed into two parts - immediate reward r(s) and the discounted value of successor state γv(s′). But, as action is involved in MDP, the form is a little different. vπ(s) =

  • a∈A

π(a|s)

  • s′∈S

p(s′|s, a)

  • r(s, a, s′) + γvπ(s′)
  • [when r is a function of s, a, s′]

=

  • a∈A

π(a|s)

  • r(s, a) + γ
  • s′∈S

p(s′|s, a)vπ(s′)

  • [when r is a function of s, a]

= r(s) + γ

  • a∈A

π(a|s)

  • s′∈S

p(s′|s, a)vπ(s′) [when r is a function of s] (6)

Abir Das (IIT Kharagpur) CS60077 July 26, Aug 01, 02, 08, 2019 35 / 43

slide-71
SLIDE 71

Agenda Terminology Markov Decision Process

Bellman Expectation Equations

qπ(s, a)=Eπ

  • Gt|St = s, at = a
  • [eqn. 3.13 in SB]

=Eπ

  • rt+1 + γrt+2 + γ2rt+3...|St = s, at = a
  • =Eπ
  • rt+1 + γ(rt+2 + γrt+3...)|St = s, at = a
  • =Eπ
  • rt+1 + γGt+1|St = s, at = a
  • [By definition, eqn. 3.11 in SB]

=Eπ

  • rt+1|St = s, at = a
  • + γEπ
  • Gt+1|St = s, at = a
  • Abir Das (IIT Kharagpur)

CS60077 July 26, Aug 01, 02, 08, 2019 36 / 43

slide-72
SLIDE 72

Agenda Terminology Markov Decision Process

Bellman Expectation Equations

qπ(s, a)=Eπ

  • Gt|St = s, at = a
  • [eqn. 3.13 in SB]

=Eπ

  • rt+1 + γrt+2 + γ2rt+3...|St = s, at = a
  • =Eπ
  • rt+1 + γ(rt+2 + γrt+3...)|St = s, at = a
  • =Eπ
  • rt+1 + γGt+1|St = s, at = a
  • [By definition, eqn. 3.11 in SB]

=Eπ

  • rt+1|St = s, at = a
  • + γEπ
  • Gt+1|St = s, at = a
  • =Eπ
  • rt+1|St = s, at = a
  • +

γEπ

  • Gt+1|St = s, at = a, St+1 = s′, at+1 = a′

|St = s, at = a

  • (Above applies the formula E
  • Y |X
  • = E
  • E
  • Y |X, Z
  • |X
  • )

[Get the intuition behind the formula in this youtube link]

Abir Das (IIT Kharagpur) CS60077 July 26, Aug 01, 02, 08, 2019 36 / 43

slide-73
SLIDE 73

Agenda Terminology Markov Decision Process

Bellman Expectation Equations

qπ(s, a)=Eπ

  • Gt|St = s, at = a
  • [eqn. 3.13 in SB]

=Eπ

  • rt+1 + γrt+2 + γ2rt+3...|St = s, at = a
  • =Eπ
  • rt+1 + γ(rt+2 + γrt+3...)|St = s, at = a
  • =Eπ
  • rt+1 + γGt+1|St = s, at = a
  • [By definition, eqn. 3.11 in SB]

=Eπ

  • rt+1|St = s, at = a
  • + γEπ
  • Gt+1|St = s, at = a
  • =Eπ
  • rt+1|St = s, at = a
  • +

γEπ

  • Gt+1|St = s, at = a, St+1 = s′, at+1 = a′

|St = s, at = a

  • (Above applies the formula E
  • Y |X
  • = E
  • E
  • Y |X, Z
  • |X
  • )

[Get the intuition behind the formula in this youtube link] =Eπ

  • rt+1|St = s, at = a
  • +

γEπ

  • Gt+1|St+1 = s′, at+1 = a′

|St = s, at = a

  • [Gt+1depends only on st+1 and at+1]

Abir Das (IIT Kharagpur) CS60077 July 26, Aug 01, 02, 08, 2019 36 / 43

slide-74
SLIDE 74

Agenda Terminology Markov Decision Process

Bellman Expectation Equations

qπ(s, a)=Eπ

  • Gt|St = s, at = a
  • [eqn. 3.13 in SB]

=Eπ

  • rt+1 + γrt+2 + γ2rt+3...|St = s, at = a
  • =Eπ
  • rt+1 + γ(rt+2 + γrt+3...)|St = s, at = a
  • =Eπ
  • rt+1 + γGt+1|St = s, at = a
  • [By definition, eqn. 3.11 in SB]

=Eπ

  • rt+1|St = s, at = a
  • + γEπ
  • Gt+1|St = s, at = a
  • =Eπ
  • rt+1|St = s, at = a
  • +

γEπ

  • Gt+1|St = s, at = a, St+1 = s′, at+1 = a′

|St = s, at = a

  • (Above applies the formula E
  • Y |X
  • = E
  • E
  • Y |X, Z
  • |X
  • )

[Get the intuition behind the formula in this youtube link] =Eπ

  • rt+1|St = s, at = a
  • +

γEπ

  • Gt+1|St+1 = s′, at+1 = a′

|St = s, at = a

  • [Gt+1depends only on st+1 and at+1]

= Eπ

  • rt+1|St =s, at =a
  • + γEπ
  • qπ(s′, a′)|St =s, at =a
  • [Using definition of qπ]

Abir Das (IIT Kharagpur) CS60077 July 26, Aug 01, 02, 08, 2019 36 / 43

slide-75
SLIDE 75

Agenda Terminology Markov Decision Process

Bellman Expectation Equations

qπ(s, a)=Eπ

  • Gt|St = s, at = a
  • [eqn. 3.13 in SB]

=Eπ

  • rt+1 + γrt+2 + γ2rt+3...|St = s, at = a
  • =Eπ
  • rt+1 + γ(rt+2 + γrt+3...)|St = s, at = a
  • =Eπ
  • rt+1 + γGt+1|St = s, at = a
  • [By definition, eqn. 3.11 in SB]

=Eπ

  • rt+1|St = s, at = a
  • + γEπ
  • Gt+1|St = s, at = a
  • =Eπ
  • rt+1|St = s, at = a
  • +

γEπ

  • Gt+1|St = s, at = a, St+1 = s′, at+1 = a′

|St = s, at = a

  • (Above applies the formula E
  • Y |X
  • = E
  • E
  • Y |X, Z
  • |X
  • )

[Get the intuition behind the formula in this youtube link] =Eπ

  • rt+1|St = s, at = a
  • +

γEπ

  • Gt+1|St+1 = s′, at+1 = a′

|St = s, at = a

  • [Gt+1depends only on st+1 and at+1]

= Eπ

  • rt+1|St =s, at =a
  • + γEπ
  • qπ(s′, a′)|St =s, at =a
  • [Using definition of qπ]

Abir Das (IIT Kharagpur) CS60077 July 26, Aug 01, 02, 08, 2019 36 / 43

slide-76
SLIDE 76

Agenda Terminology Markov Decision Process

Bellman Expectation Equations

qπ(s, a)=Eπ

  • Gt|St = s, at = a
  • [eqn. 3.13 in SB]

=Eπ

  • rt+1 + γrt+2 + γ2rt+3...|St = s, at = a
  • =Eπ
  • rt+1 + γ(rt+2 + γrt+3...)|St = s, at = a
  • =Eπ
  • rt+1 + γGt+1|St = s, at = a
  • [By definition, eqn. 3.11 in SB]

=Eπ

  • rt+1|St = s, at = a
  • + γEπ
  • Gt+1|St = s, at = a
  • =Eπ
  • rt+1|St = s, at = a
  • +

γEπ

  • Gt+1|St = s, at = a, St+1 = s′, at+1 = a′

|St = s, at = a

  • (Above applies the formula E
  • Y |X
  • = E
  • E
  • Y |X, Z
  • |X
  • )

[Get the intuition behind the formula in this youtube link] =Eπ

  • rt+1|St = s, at = a
  • +

γEπ

  • Gt+1|St+1 = s′, at+1 = a′

|St = s, at = a

  • [Gt+1depends only on st+1 and at+1]

= Eπ

  • rt+1|St =s, at =a
  • + γEπ
  • qπ(s′, a′)|St =s, at =a
  • [Using definition of qπ]

Abir Das (IIT Kharagpur) CS60077 July 26, Aug 01, 02, 08, 2019 36 / 43

slide-77
SLIDE 77

Agenda Terminology Markov Decision Process

Bellman Expectation Equations

= r(s, a) +

  • s′∈S
  • a′∈A

qπ(s′, a′)p(a′, s′|s, a) = r(s, a) +

  • s′∈S
  • a′∈A

qπ(s′, a′)p(a′|s′, s, a)p(s′|s, a) = r(s, a) +

  • s′∈S
  • a′∈A

qπ(s′, a′)p(a′|s′)p(s′|s, a) [Markov property]

Abir Das (IIT Kharagpur) CS60077 July 26, Aug 01, 02, 08, 2019 37 / 43

slide-78
SLIDE 78

Agenda Terminology Markov Decision Process

Bellman Expectation Equations

= r(s, a) +

  • s′∈S
  • a′∈A

qπ(s′, a′)p(a′, s′|s, a) = r(s, a) +

  • s′∈S
  • a′∈A

qπ(s′, a′)p(a′|s′, s, a)p(s′|s, a) = r(s, a) +

  • s′∈S
  • a′∈A

qπ(s′, a′)p(a′|s′)p(s′|s, a) [Markov property] = r(s, a) +

  • s′∈S

p(s′|s, a)

  • a′∈A

qπ(s′, a′)p(a′|s′)

Abir Das (IIT Kharagpur) CS60077 July 26, Aug 01, 02, 08, 2019 37 / 43

slide-79
SLIDE 79

Agenda Terminology Markov Decision Process

Bellman Expectation Equations

= r(s, a) +

  • s′∈S
  • a′∈A

qπ(s′, a′)p(a′, s′|s, a) = r(s, a) +

  • s′∈S
  • a′∈A

qπ(s′, a′)p(a′|s′, s, a)p(s′|s, a) = r(s, a) +

  • s′∈S
  • a′∈A

qπ(s′, a′)p(a′|s′)p(s′|s, a) [Markov property] = r(s, a) +

  • s′∈S

p(s′|s, a)

  • a′∈A

qπ(s′, a′)p(a′|s′)

Abir Das (IIT Kharagpur) CS60077 July 26, Aug 01, 02, 08, 2019 37 / 43

slide-80
SLIDE 80

Agenda Terminology Markov Decision Process

Bellman Expectation Equations

= r(s, a) +

  • s′∈S
  • a′∈A

qπ(s′, a′)p(a′, s′|s, a) = r(s, a) +

  • s′∈S
  • a′∈A

qπ(s′, a′)p(a′|s′, s, a)p(s′|s, a) = r(s, a) +

  • s′∈S
  • a′∈A

qπ(s′, a′)p(a′|s′)p(s′|s, a) [Markov property] = r(s, a) +

  • s′∈S

p(s′|s, a)

  • a′∈A

qπ(s′, a′)p(a′|s′)

Abir Das (IIT Kharagpur) CS60077 July 26, Aug 01, 02, 08, 2019 37 / 43

slide-81
SLIDE 81

Agenda Terminology Markov Decision Process

Bellman Expectation Equation for Student MDP

0.2 0.4 R = -1

Facebook

0.4 R = -1

Quit

R = -1

Facebook

R = 0

Sleep

R = -2

Study

R = -2

Study

R = +1

Pub

R = +10

Study

  • 2.3
  • 1.3

2.7 7.4 𝑤" 𝑡 for 𝜌 𝑏 𝑡 = 0.5, 𝛿 = 1

7.4 = 0.5*{10+0} + 0.5*{1+1*(-0.2*1.3+0.4*2.7+0.4*7.4)}

Figure credit: David Silver, DeepMind Abir Das (IIT Kharagpur) CS60077 July 26, Aug 01, 02, 08, 2019 38 / 43

slide-82
SLIDE 82

Agenda Terminology Markov Decision Process

Optimal Policies and Optimal Value Functions

§ Solving a reinforcement learning task means, roughly, finding a policy that achieves a lot of reward (maximum) over the long run. § The notion of maximality leads to optimality in MDPs.

Abir Das (IIT Kharagpur) CS60077 July 26, Aug 01, 02, 08, 2019 39 / 43

slide-83
SLIDE 83

Agenda Terminology Markov Decision Process

Optimal Policies and Optimal Value Functions

§ Solving a reinforcement learning task means, roughly, finding a policy that achieves a lot of reward (maximum) over the long run. § The notion of maximality leads to optimality in MDPs. § What is meant by a policy is better than some other policy?

Abir Das (IIT Kharagpur) CS60077 July 26, Aug 01, 02, 08, 2019 39 / 43

slide-84
SLIDE 84

Agenda Terminology Markov Decision Process

Optimal Policies and Optimal Value Functions

§ Solving a reinforcement learning task means, roughly, finding a policy that achieves a lot of reward (maximum) over the long run. § The notion of maximality leads to optimality in MDPs. § What is meant by a policy is better than some other policy? § A policy π is defined to be better than or equal to a policy π′ if its expected return is greater than or equal to that of π′ for all states. Definition π ≥ π′ iff vπ(s) ≥ vπ′(s), ∀s ∈ S

Abir Das (IIT Kharagpur) CS60077 July 26, Aug 01, 02, 08, 2019 39 / 43

slide-85
SLIDE 85

Agenda Terminology Markov Decision Process

Optimal Policies and Optimal Value Functions

Definition The optimal state-value function v∗(s) is the maximum state-value function over all policies v∗(s) = max

π

vπ(s), ∀s ∈ S The optimal action-value function q∗(s, a) is the maximum action-value function over all policies q∗(s, a) = max

π

qπ(s, a), ∀s ∈ S and , ∀a ∈ A § An MDP is “solved” when we know the optimal value function

Abir Das (IIT Kharagpur) CS60077 July 26, Aug 01, 02, 08, 2019 40 / 43

slide-86
SLIDE 86

Agenda Terminology Markov Decision Process

Optimal Action-Value Function for Student MDP

0.2 0.4 R = -1

Facebook

0.4 R = -1

Quit

R = -1

Facebook

R = 0

Sleep

R = -2

Study

R = -2

Study

R = +1

Pub

R = +10

Study

6 6 8 10 𝜌∗ 𝑏 𝑡 for 𝛿 = 1

𝑟∗ = 5

𝑟∗ = 6

𝑟∗ = 5 𝑟∗ = 0 𝑟∗ = 6 𝑟∗ = 8 𝑟∗ = 10 𝑟∗ = 9.4 Figure credit: David Silver, DeepMind Abir Das (IIT Kharagpur) CS60077 July 26, Aug 01, 02, 08, 2019 41 / 43

slide-87
SLIDE 87

Agenda Terminology Markov Decision Process

Optimal Policy

Theorem For any Markov Decision Process § There exists an optimal policy π∗ that is better than or equal to all

  • ther policies, π∗ ≥ π, ∀π

§ All optimal policies achieve the optimal value function vπ∗(s) = v∗(s) § All optimal policies achieve the optimal action-value function qπ∗(s, a) = q∗(s, a)

Abir Das (IIT Kharagpur) CS60077 July 26, Aug 01, 02, 08, 2019 42 / 43

slide-88
SLIDE 88

Agenda Terminology Markov Decision Process

Optimal Policy

Theorem For any Markov Decision Process § There exists an optimal policy π∗ that is better than or equal to all

  • ther policies, π∗ ≥ π, ∀π

§ All optimal policies achieve the optimal value function vπ∗(s) = v∗(s) § All optimal policies achieve the optimal action-value function qπ∗(s, a) = q∗(s, a) An optimal policy can be found by maximising over q∗(s, a). π∗(a|s) =    1 if a = arg max

a∈A

q∗(s, a)

  • therwise

§ There is always a deterministic optimal policy for any MDP. § If we know q∗(s, a), we immediately have the optimal policy.

Abir Das (IIT Kharagpur) CS60077 July 26, Aug 01, 02, 08, 2019 42 / 43

slide-89
SLIDE 89

Agenda Terminology Markov Decision Process

Relation between v∗ and q∗

𝑤∗(𝑡) 𝑡 𝑟∗(𝑡, 𝑏) a a′𝑟∗(𝑡, 𝑏′)

v∗(s) = max

a∈A q∗(s, a)

Abir Das (IIT Kharagpur) CS60077 July 26, Aug 01, 02, 08, 2019 43 / 43

slide-90
SLIDE 90

Agenda Terminology Markov Decision Process

Relation between v∗ and q∗

𝑤∗(𝑡) 𝑡 𝑟∗(𝑡, 𝑏) a a′𝑟∗(𝑡, 𝑏′)

v∗(s) = max

a∈A q∗(s, a)

𝑟∗(𝑡, 𝑏) 𝑡 𝑤∗(𝑡′) 𝑡′ 𝑡′′ 𝑤∗(𝑡′′)

𝑠

q∗(s, a) = r(s, a)+γ

  • s′∈S

p(s′|s, a)v∗(s′)

Abir Das (IIT Kharagpur) CS60077 July 26, Aug 01, 02, 08, 2019 43 / 43

slide-91
SLIDE 91

Agenda Terminology Markov Decision Process

Relation between v∗ and q∗

𝑤∗(𝑡) 𝑡 𝑟∗(𝑡, 𝑏) a a′𝑟∗(𝑡, 𝑏′)

v∗(s) = max

a∈A q∗(s, a)

𝑟∗(𝑡, 𝑏) 𝑡 𝑤∗(𝑡′) 𝑡′ 𝑡′′ 𝑤∗(𝑡′′)

𝑠

q∗(s, a) = r(s, a)+γ

  • s′∈S

p(s′|s, a)v∗(s′)

𝑤∗(𝑡) 𝑡 𝑏 𝑤∗(𝑡′) 𝑡′

𝑠

v∗(s) = max

a∈A

  • r(s, a) +

γ

  • s′∈S

p(s′|s, a)v∗(s′)

  • Abir Das (IIT Kharagpur)

CS60077 July 26, Aug 01, 02, 08, 2019 43 / 43

slide-92
SLIDE 92

Agenda Terminology Markov Decision Process

Relation between v∗ and q∗

𝑤∗(𝑡) 𝑡 𝑟∗(𝑡, 𝑏) a a′𝑟∗(𝑡, 𝑏′)

v∗(s) = max

a∈A q∗(s, a)

𝑟∗(𝑡, 𝑏) 𝑡 𝑤∗(𝑡′) 𝑡′ 𝑡′′ 𝑤∗(𝑡′′)

𝑠

q∗(s, a) = r(s, a)+γ

  • s′∈S

p(s′|s, a)v∗(s′)

𝑤∗(𝑡) 𝑡 𝑏 𝑤∗(𝑡′) 𝑡′

𝑠

v∗(s) = max

a∈A

  • r(s, a) +

γ

  • s′∈S

p(s′|s, a)v∗(s′)

  • 𝑟∗(𝑡, 𝑏)

𝑡 𝑡′

𝑠

𝑟∗(𝑡′, 𝑏′) 𝑏′

q∗(s, a) = r(s, a) + γ

  • s′∈S

p(s′|s, a) max

a′∈A q∗(s′, a′)

Abir Das (IIT Kharagpur) CS60077 July 26, Aug 01, 02, 08, 2019 43 / 43

slide-93
SLIDE 93

Appendices Conditional Independence Eigenvalues

Appendices

Appendices

Abir Das (IIT Kharagpur) CS60077 July 26, Aug 01, 02, 08, 2019 1 / 3

slide-94
SLIDE 94

Appendices Conditional Independence Eigenvalues

  • 1. Independence

Independence A⊥ ⊥B = ⇒ P(A|B) = P(A) Conditional Independence A⊥ ⊥B|C = ⇒ P(A|B, C) = P(A|C) Proof: P(A|B, C) = P(A, B, C) P(B, C) = P(A, B|C)✟✟

P(C) P(B|C)✟✟

P(C) (7) = P(A|C)P(B|C) P(B|C) [ From definition of conditional independence] = P(A|C)

Abir Das (IIT Kharagpur) CS60077 July 26, Aug 01, 02, 08, 2019 2 / 3

slide-95
SLIDE 95

Appendices Conditional Independence Eigenvalues

  • 2. Independence

Theorem Eigenvalues of the transpose AT are the same as the eigenvalues of A Proof Eigenvalues of a matrix are roots of its characteristic polynomial. Hence if the matrices A and AT have the same characteristic polynomial, then they have the same eigenvalues. det(AT − λI) = det(AT − λIT ) (8) = det(A − λI)T = det(A − λI) [Since det(A) = det(AT )]

Abir Das (IIT Kharagpur) CS60077 July 26, Aug 01, 02, 08, 2019 3 / 3