time Markov Models Mohmmadsadegh Mohagheghi And Behrang Chaboki - - PowerPoint PPT Presentation

โ–ถ
time markov models
SMART_READER_LITE
LIVE PREVIEW

time Markov Models Mohmmadsadegh Mohagheghi And Behrang Chaboki - - PowerPoint PPT Presentation

Dirac-based Reduction Techniques for Quantitative Analysis of Discrete- time Markov Models Mohmmadsadegh Mohagheghi And Behrang Chaboki Vali-e-Asr University of Rafsanjan Probabilistic Model Checking Probabilistic Model Checking A Markov


slide-1
SLIDE 1

Dirac-based Reduction Techniques for Quantitative Analysis of Discrete- time Markov Models

Mohmmadsadegh Mohagheghi And Behrang Chaboki Vali-e-Asr University of Rafsanjan

slide-2
SLIDE 2

Probabilistic Model Checking

slide-3
SLIDE 3

Probabilistic Model Checking

  • A Markov Decision Process (MDP) is a tuple

M = (S, s0, Act, P, R) where:

  • S is a set of states,
  • s0 is the initial state,
  • Act is a finite set of actions,
  • P is a probabilistic transition function,
  • R is a reward function.
slide-4
SLIDE 4

Probabilistic Model Checking

  • A policy is used to resolve non-deterministic

choices of an MDP.

  • A policy ๐œŒ: ๐‘‡ โ†’ ๐ต๐‘‘๐‘ข selects one action for

each state s.

  • For an MDP M, every possible policy ๐œŒ induces

a quotient DTMC.

slide-5
SLIDE 5

Numeric Computations

  • Reachability Probabilities: The (maximal or

minimal) probability of finally reaching one of the goal states [For MDPs]

  • Expected Rewards: The (maximal or minimal)

expectation of accumulated rewards until reaching a goal state

slide-6
SLIDE 6

Numeric Computations

  • Extremal Reachablity Probabilities

& Expected Rewards

Solving a Linear Program (Exact Solutions) Using Iterative Methods (In practice)

slide-7
SLIDE 7

Jacobi Iterative Method (DTMCs)

  • Starting from an initial vector

๐‘Š of value-states, in each iteration and every state ๐‘ก โˆˆ ๐‘‡ update ๐‘Š๐‘ก as:

๏ƒฅ ๏ƒŽ

๏€ฝ

ฮฑ) Post(s, s' s' s

).V s' ฮฑ, P(s, V

๏ƒฅ ๏ƒŽ

๏€ซ ๏€ฝ

ฮฑ) Post(s, s' s' s

).V s' ฮฑ, P(s, R(s) V

slide-8
SLIDE 8

Value Iteration (MDPs)

  • Starting from an initial vector

๐‘Šof value-states, in each iteration and every state ๐‘ก โˆˆ ๐‘‡ update ๐‘Š๐‘ก as:

) ).V s' ฮฑ, P(s, ( Max V

ฮฑ) Post(s, s' s' Act(s) ฮฑ s

๏ƒฅ ๏ƒŽ

๏ƒŽ

๏€ฝ ) ).V s' ฮฑ, P(s, (R(s) Max V

ฮฑ) Post(s, s' s' Act(s) ฮฑ s

๏ƒฅ ๏ƒŽ

๏ƒŽ

๏€ซ ๏€ฝ

slide-9
SLIDE 9

Value Iteration (MDPs)

  • Starting from an initial vector

๐‘Š of value-states, in each iteration and every state s of S update ๐‘Š๐‘ก as:

  • Terminate criterion:

for some tiny ฮต (10^-6)

) ).V s' ฮฑ, P(s, (R(s) Max V

ฮฑ) Post(s, s' s' Act(s) ฮฑ s

๏ƒฅ ๏ƒŽ

๏ƒŽ

๏€ซ ๏€ฝ

๏ฅ ๏‚ฃ

๏ƒŽ

) V

  • (V

Max

  • ld

s s S s

slide-10
SLIDE 10

Policy Iteration

Select a policy ๐œŒ repeat Compute the values of the induced DTMC Update ๐œŒ Until no change in policies

slide-11
SLIDE 11

Dirac-based Reduction Technique

  • Idea: If ๐‘„(๐‘ก, ๐‘กโ€ฒ) = 1 in a DTMC, reachability

probability of s and ๐‘กโ€ฒ are equal.

slide-12
SLIDE 12

Dirac-based Reduction Technique

  • Dirac transitions are used to classify ๐‘‡.
  • The states of each class are connected with

Dirac transitions and have the same reachability probabilities.

  • Apply iterative commutations on the reduced

DTMC

slide-13
SLIDE 13

Dirac-based Reduction Technique

  • DTMC reduction can be used for policy

iteration

  • Time complexity: Linear in the size of DTMCs
slide-14
SLIDE 14

Dirac-based Reduction Technique

  • Expected rewards: If ๐‘„(๐‘ก, ๐‘กโ€ฒ) = 1 then ๐‘Š

๐‘กโ€ฒ =

๐‘Š

๐‘ก + ๐‘†(๐‘ก)

  • State-rewards should be modified for reduced

DTMCs

slide-15
SLIDE 15

Experimental Results

  • We implemented Dirac-based methods in

PRISM.

  • Available in:

https://github.com/sadeghrk/prism/tree/Dira cBased-Improving

slide-16
SLIDE 16

Experimental Results

slide-17
SLIDE 17

Questions?