Calibrated Model-Based Deep Reinforcement Learning IC ML 2019 Ali - - PowerPoint PPT Presentation

calibrated model based deep reinforcement learning
SMART_READER_LITE
LIVE PREVIEW

Calibrated Model-Based Deep Reinforcement Learning IC ML 2019 Ali - - PowerPoint PPT Presentation

Calibrated Model-Based Deep Reinforcement Learning IC ML 2019 Ali Malik*, Volodymyr Kuleshov*, Jiaming Song, Danny Nemer, Harlan Seymour, Stefano Ermon June 13, 2019 *equal contribution Overview Importance of predictive uncertainty


slide-1
SLIDE 1

Calibrated Model-Based Deep Reinforcement Learning

Ali Malik*, Volodymyr Kuleshov*, Jiaming Song, Danny Nemer, Harlan Seymour, Stefano Ermon June 13, 2019

IC ML 2019

*equal contribution

slide-2
SLIDE 2

Overview

  • Importance of predictive uncertainty
  • Which uncertainties matter for MBRL?
  • Calibration in MBRL
  • Recalibrating MBRL
  • Results
slide-3
SLIDE 3

Importance of Predictive Uncertainty

Assessing uncertainty is crucial in modern decision-making systems

RL + Control

Kahn et al. (2018) Chua et al. (2018)

Safety

Berkenkamp et al. (2017)

Medicine

Saria (2018) Heckerman et al. (1989) Safe exploration Diagnosis, risk prediction, treatment recommendation. Obstacle avoidance, reward planning

slide-4
SLIDE 4

Importance of Predictive Uncertainty

Assessing uncertainty is crucial in modern decision-making systems

Autonomous Driving

Smith & Cheeseman (1986) McAllister et al. (2017) Segmentation, object detection, depth estimation.

Upper Confidence Bounds

Auer et al. (2002) Li et al. (2010) Balancing exploration and exploitation

slide-5
SLIDE 5

Importance of Predictive Uncertainty

Modelling uncertainty accurately is crucial

Key question: Which uncertainties are important in Model-Based Reinforcement Learning?

slide-6
SLIDE 6

What constitutes good probabilistic forecasts?

Literature on proper scoring rules suggest two important factors

Sharpness Calibration

100

%

100

%

100

%

100

%

Sharp

Predictive distributions should be focused i.e have low variance Uncertainty should be empirically accurate i.e. true value should fall in a p% confidence interval p% of the time

75

%

75

%

75

%

75

%

Calibrated

slide-7
SLIDE 7

Calibration

66

%

Forecaster

66

%

66

%

I should be correct 66% of the time For things I’m 66% sure about Calibration measures reliability of probabilistic claims.

slide-8
SLIDE 8

Calibration

For regression: 66

%

Forecaster

66

%

66

%

I should be correct 66% of the time For things I’m 66% sure about Calibration measures reliability of probabilistic claims. Predicted probability for credible interval True probability of Y falling in the interval 90

%

=

slide-9
SLIDE 9

Calibration vs Sharpness

There is an inherent trade-off between calibration and sharpness What should we prioritise?

Claim: In model-based reinforcement learning, uncertainties should be calibrated

slide-10
SLIDE 10

Importance of Calibration

Calibration is really important in model-based reinforcement learning.

Planning

Calibrated uncertainties lead to better estimates of expectation. Theorem: The value of policy ! for an MDP under the true dynamics " is equal to the value of the policy under some other dynamics # " that are calibrated with respect to the MDP.

slide-11
SLIDE 11

Importance of Calibration

Calibration is really important in model-based reinforcement learning.

Exploration

Many exploration/exploitation algorithms use Upper Confidence Bounds (UCBs) to guide choices: True reward

  • f arm

Calibrated reward

Uncalib. reward

Calibration naturally improves UCBs, resulting in better exploration.

slide-12
SLIDE 12

Calibrating Model-Based RL

Uncertainties derived from modern neural networks are often uncalibrated. We can recalibrate any forecaster using work by Kuleshov et al (2018):

New Forecast

R(Ft(y))

Input

Predictor

H : X → (Y → [0, 1])

Can be any model (seen as black box)

Forecast

Ft(y)

F : Y → [0, 1]

Uncalibrated CDF

Recalibrator R : [0, 1] → [0, 1]

Transforms probabilities coming out of F

R Recalibration

slide-13
SLIDE 13

Deriving the Ideal Recalibrator

We learn a mapping between predicted and true (empirical) probabilities. 60% quantile 40% quantile 70% quantile 45% quantile 80% quantile 55% quantile

what model predicts what data says

P(FX(Y ) ≤ p)

p

… …

Fact: Ideal recalibrator is R(p) = P(Y ≤ FX-1(p)).

p = P(Y ≤ F −1

X (p))

Calibration

slide-14
SLIDE 14

Calibrating Model-Based RL

This gives the following algorithm for MBRL:

Calibrated MBRL

Train calibrated transition model !

" from observations by repeatedly: 1.Explore: Collect observations using current transition

model.

2.LearnModel: Retrain transition model using new

  • bservations.

3.LearnCalib: Learn recalibrator # on held-out subset of

  • bservations.

4.Recalibrate: Set ! " = # % ! "

slide-15
SLIDE 15

Results: Contextual Bandits

We can apply this scheme to the LinUCB algorithm for contextual bandits:

Recalibration consistently improves the exploration/exploitation balance in contextual bandits tasks.

slide-16
SLIDE 16

Results: MuJoCo Continuous Control

We calibrate the probabilistic ensemble model from Chua et al. 2018 and show noticeable improvement in sample complexity across different tasks:

Recalibration improves the sample complexity in continuous control tasks.

slide-17
SLIDE 17

Results: Inventory Planning

We also calibrate a Bayesian DenseNet tasked with controlling the inventory of perishable goods in a store Inventory Position (state) Sales, Spoilage (state transitions) Inventory Position (state) Inventory Position (state) Sales, Spoilage (state transitions) Shipment (store decision) Shipment (store decision)

… …

Reward: Sales revenue, minus shipment costs.

slide-18
SLIDE 18

Thank you!

Stop by poster #36 for more details