Weighted Automata Extraction from Recurrent Neural Networks via - - PowerPoint PPT Presentation

β–Ά
weighted automata extraction
SMART_READER_LITE
LIVE PREVIEW

Weighted Automata Extraction from Recurrent Neural Networks via - - PowerPoint PPT Presentation

Weighted Automata Extraction from Recurrent Neural Networks via Regression on State Spaces Takamasa Okudono, Masaki Waga, Taro Sekiyama, Ichiro Hasuo SOKENDAI, the Graduate University for Advanced Studies, Japan /National Institute of


slide-1
SLIDE 1

Weighted Automata Extraction from Recurrent Neural Networks via Regression on State Spaces

Takamasa Okudono, Masaki Waga, Taro Sekiyama, Ichiro Hasuo SOKENDAI, the Graduate University for Advanced Studies, Japan /National Institute of Informatics, Japan LearnAut19, Vancouver, Canada 23 June 2019

1

slide-2
SLIDE 2

RNN

RNN is a neural network equipped with a internal state

Drawing by FranΓ§ois Deloche (CC BY-SA 4.0)

2

slide-3
SLIDE 3

Goal

Input: RNN 𝑆 whose output is in ℝ (defines 𝑔

𝑆: Ξ£βˆ— β†’ ℝ)

Output: WFA 𝐡(𝑆) (defines 𝑔

𝐡(𝑆): Ξ£βˆ— β†’ ℝ) s.t. 𝑔 𝐡(𝑆) ≃ 𝑔 𝑆

RNN WFA Initial state Final func. Transition func. Transition matrix Initial vector Final vector

3

slide-4
SLIDE 4

Motivation

  • Getting lighter (faster to infer) model of an RNN
  • Because the inference of RNNs are sometimes heavy
  • Investigate the behavior of RNN 𝑆 via the extracted WFA 𝐡 𝑆
  • WFA equips many operations and leads to model checking?
  • In research line of RNN⇔DFA conversion as an acceptor
  • Ours is a quantitative extension

4

slide-5
SLIDE 5

Contribution

  • Proposed a method to apply Balle and Mohri’s algorithm for

the extraction

  • The key is checking if 𝑆 ≃ 𝐡 by using regression
  • Our method extracts +7% more accurate models than the

baseline

  • The extracted WFAs are about 1,000 times faster to infer

than the target RNNs

5

slide-6
SLIDE 6
  • Def. of RNN (Mathematically, in this work)

RNN 𝑆 (of alphabet Ξ£ and dimension 𝑒) consists of

  • 𝛽 ∈ ℝ𝑒: Initial state
  • 𝛾: ℝ𝑒 β†’ ℝ: Final function
  • 𝑕𝑆: ℝ𝑒 Γ— Ξ£ β†’ ℝ𝑒: Transition function
  • 𝑕𝑆: ℝ𝑒 Γ— Ξ£βˆ— β†’ ℝ𝑒 is induced recursivelyβ– 

𝑔

𝑆: Ξ£βˆ— β†’ ℝ is induced by 𝑔 𝑆 π‘₯1 … π‘₯𝑂 = 𝛾 ∘ 𝑕𝑆(𝛽, π‘₯1 … π‘₯𝑂)

The configuration for π‘₯1 … π‘₯𝑂 is defined by πœ€π‘† π‘₯1 … π‘₯𝑂 = 𝑕𝑆 𝛽, π‘₯1 … π‘₯𝑂

Need not to be linear β€œinternal state”

6

slide-7
SLIDE 7
  • Def. of Weighted Finite Automaton (WFA)

WFA 𝐡 (of size π‘œ and alphabet Ξ£) consists of

  • 𝛽 ∈ β„π‘œ: Initial vector
  • 𝛾 ∈ β„π‘œ: Final vector
  • 𝐡𝜏 ∈ β„π‘œΓ—π‘œ: Transition matrix (𝜏 ∈ Ξ£) β– 

WFA 𝐡 is a formalism to define 𝑔

𝐡: Ξ£βˆ— β†’ ℝ (c.f.) A DFA is a formalism to define 𝑔: Ξ£βˆ— β†’ 2 WFA is an extension of DFA via the matrix representation.

7

slide-8
SLIDE 8
  • Def. of WFA
  • WFA 𝐡 induces the function 𝑔

𝐡: Ξ£βˆ— β†’ ℝ as

𝑔

𝐡 π‘₯1 … π‘₯𝑂 = 𝛽𝐡π‘₯1 … 𝐡π‘₯𝑂𝛾

  • The configuration (β€œinternal state”) of WFA 𝐡 is

πœ€π΅ π‘₯1 … π‘₯𝑂 = 𝛽𝐡π‘₯1 … 𝐡π‘₯𝑂 ∈ β„π‘œ For example:

  • Ξ£ = 0, 1 , 𝛽 = 0.8

0.2 , 𝛾 = 0.9 0.7 , 𝐡0 = 0 1 1 0 , 𝐡1 = 0.9 0.1 0.5 0.5

  • 𝑔

𝐡 10 = 0.8

0.2 0.9 0.1 0.5 0.5 1 1 0.9 0.7 = 0.736

  • πœ€π΅ 10 = 0.8

0.2 0.9 0.1 0.5 0.5 1 1 0 = 0.18 0.82

8

slide-9
SLIDE 9

RNN and WFA

RNN 𝑆 (of alphabet Ξ£ and dimension 𝑒) consists of

  • 𝛽 ∈ ℝ𝑒: Initial state
  • 𝛾: ℝ𝑒 β†’ ℝ: Final function
  • 𝑕𝑆: ℝ𝑒 Γ— Ξ£ β†’ ℝ𝑒: Transition functionβ– 

WFA 𝐡 (of alphabet Ξ£ and size π‘œ) consists of

  • 𝛽 ∈ β„π‘œ: Initial vector
  • 𝛾 ∈ β„π‘œ: Final vector
  • 𝐡𝜏 ∈ β„π‘œΓ—π‘œ: Transition matrix (𝜏 ∈ Ξ£) β– 

Similar formalism! Can we approximate RNN by WFA?

9

slide-10
SLIDE 10

Goal Input: RNN 𝑆 whose output is in ℝ (defines 𝑔

𝑆: Ξ£βˆ— β†’ ℝ)

Output: WFA 𝐡(𝑆) (defines 𝑔

𝐡(𝑆): Ξ£βˆ— β†’ ℝ) s.t. 𝑔 𝐡(𝑆) ≃ 𝑔 𝑆

Approach: Use Balle and Mohri’s algorithm

  • The challenge is to give the procedure to check if 𝑔

𝐡 ≃ 𝑔 𝑆 for

a candidate WFA 𝐡

Goal and Our Approach

10

slide-11
SLIDE 11

Balle and Mohri’s Algorithm

An extension of Angluin’s L* Algorithm for WFA

  • Input:
  • Membership query procedure m: Ξ£βˆ— β†’ ℝ
  • Equivalence query procedure e: WFAs β†’ Equivalent βŠ” Ξ£βˆ—
  • Output:
  • Minimal WFA 𝐡′
  • Property: Given WFA 𝐡, if 𝑛 = 𝑔

𝐡 and

𝑓 ሚ 𝐡 = α‰ŠEquivalent ; 𝑔

𝐡 = 𝑔 ΰ·¨ 𝐡

π‘₯ ; 𝑔

𝐡 π‘₯ β‰  𝑔 ΰ·¨ 𝐡(π‘₯)

then, it terminates by calling 𝑛, 𝑓 polynomial times and 𝑔

𝐡 = 𝑔 𝐡′

Called β€œCounterexample”

11

slide-12
SLIDE 12

Idea of Overall Architecture (Detailed)

Implement

  • Membership query 𝑛 to be the RNN’s induced function 𝑔

𝑆

  • Equivalence query 𝑓 to be

𝑓 ሚ 𝐡 = α‰ŠEquivalent ; 𝑔

𝑆 ≃ 𝑔 ΰ·¨ 𝐡

π‘₯ ; 𝑔

𝑆 π‘₯ β‰  𝑔 ΰ·¨ 𝐡(π‘₯)

Then we would be able to get a WFA ሚ 𝐡 s.t. 𝑔

𝑆 ≃ 𝑔 ΰ·¨ 𝐡 !

Generally it cannot be β€œ=β€œ

But how can we implement such an equivalence query 𝑓?

12

slide-13
SLIDE 13

How do we know 𝑔

𝑆 ≃ 𝑔 𝐡?

𝑔

𝑆 π‘₯ ≃ 𝑔 𝐡(π‘₯)

⇔ 𝛾𝑆 ∘ πœ€π‘† 𝛽𝑆, π‘₯1 … π‘₯π‘œ ≃ πœ€π΅(π‘₯1 … π‘₯π‘œ)𝛾𝐡 Both calculate their configurations (β€œinternal states”) If there is a β€œgood” relation between πœ€π‘† and πœ€π΅, 𝐡 and 𝑆 would behave similarly

13

slide-14
SLIDE 14

β€œGood” relation between πœ€π‘† and πœ€π΅

  • This work views π‘ž: ℝ𝑒 β†’ β„π‘œ satisfying the following property

as a good relation:

βˆ€π‘₯ ∈ Ξ£βˆ—. p πœ€π‘† w ≃ πœ€π΅(π‘₯)

14

slide-15
SLIDE 15

Equivalence Query by approximating π‘ž

Let’s approximate configuration translator π‘ž: ℝ𝑒 β†’ β„π‘œ such that βˆ€π‘₯ ∈ Ξ£βˆ—. p πœ€π‘† w ≃ πœ€π΅(π‘₯) by applying regression on sampled data. The data is sampled by observing Ξ£βˆ— in Breadth-First Search.

15

slide-16
SLIDE 16

Relation π‘ž between 𝑆 and 𝐡

  • config. space of 𝑆 (ℝ𝑒)
  • config. space of 𝐡 (β„π‘œ)

・𝛽𝑆 ・𝛽𝐡

16

slide-17
SLIDE 17

Relation π‘ž between 𝑆 and 𝐡

  • config. space of 𝑆 (ℝ𝑒)
  • config. space of 𝐡 (β„π‘œ)

・𝛽𝑆 ・𝛽𝐡

17

slide-18
SLIDE 18

Relation π‘ž between 𝑆 and 𝐡

  • config. space of 𝑆 (ℝ𝑒)
  • config. space of 𝐡 (β„π‘œ)

・𝛽𝑆 ・𝛽𝐡 γƒ»πœ€π‘†(0) γƒ»πœ€π΅(0)

18

slide-19
SLIDE 19

Relation π‘ž between 𝑆 and 𝐡

  • config. space of 𝑆 (ℝ𝑒)
  • config. space of 𝐡 (β„π‘œ)

・𝛽𝑆 ・𝛽𝐡 γƒ»πœ€π‘†(0) γƒ»πœ€π΅(0)

19

slide-20
SLIDE 20

Relation π‘ž between 𝑆 and 𝐡

  • config. space of 𝑆 (ℝ𝑒)
  • config. space of 𝐡 (β„π‘œ)

・𝛽𝑆 ・𝛽𝐡 γƒ»πœ€π‘†(0) γƒ»πœ€π΅(0) γƒ»πœ€π‘†(1) γƒ»πœ€π΅(1)

20

slide-21
SLIDE 21

Relation π‘ž between 𝑆 and 𝐡

  • config. space of 𝑆 (ℝ𝑒)
  • config. space of 𝐡 (β„π‘œ)

・𝛽𝑆 ・𝛽𝐡 γƒ»πœ€π‘†(0) γƒ»πœ€π΅(0) γƒ»πœ€π‘†(1) γƒ»πœ€π΅(1)

21

slide-22
SLIDE 22

Relation π‘ž between 𝑆 and 𝐡

  • config. space of 𝑆 (ℝ𝑒)
  • config. space of 𝐡 (β„π‘œ)

・𝛽𝑆 ・𝛽𝐡 γƒ»πœ€π‘†(0) γƒ»πœ€π΅(0) γƒ»πœ€π‘†(1) γƒ»πœ€π΅(1) γƒ»πœ€π‘†(00) γƒ»πœ€π΅(00)

22

slide-23
SLIDE 23

Relation π‘ž between 𝑆 and 𝐡

  • config. space of 𝑆 (ℝ𝑒)
  • config. space of 𝐡 (β„π‘œ)

・𝛽𝑆 ・𝛽𝐡 γƒ»πœ€π‘†(0) γƒ»πœ€π΅(0) γƒ»πœ€π‘†(1) γƒ»πœ€π΅(1) γƒ»πœ€π‘†(00) γƒ»πœ€π΅(00)

23

slide-24
SLIDE 24

Relation π‘ž between 𝑆 and 𝐡

  • config. space of 𝑆 (ℝ𝑒)
  • config. space of 𝐡 (β„π‘œ)

・𝛽𝑆 ・𝛽𝐡 γƒ»πœ€π‘†(0) γƒ»πœ€π΅ 0 ≃ πœ€π΅(01) γƒ»πœ€π‘†(1) γƒ»πœ€π΅(1) γƒ»πœ€π‘†(00) γƒ»πœ€π΅(00) γƒ»πœ€π‘†(01)

24

slide-25
SLIDE 25

Relation π‘ž between 𝑆 and 𝐡

  • config. space of 𝑆 (ℝ𝑒)
  • config. space of 𝐡 (β„π‘œ)

・𝛽𝑆 ・𝛽𝐡 γƒ»πœ€π‘†(0) γƒ»πœ€π΅ 0 ≃ πœ€π΅(01) γƒ»πœ€π‘†(1) γƒ»πœ€π΅(1) γƒ»πœ€π‘†(00) γƒ»πœ€π΅(00) γƒ»πœ€π‘†(01)

25

slide-26
SLIDE 26

BFS-based Equivalence Query

Pop w from queue Add w’s next words to queue

Equivalence query proceeds based on Breadth-First Search

26

slide-27
SLIDE 27

Maintaining π‘ž

Pop w from queue Add w’s next words to queue Check if π‘ž should be refined Refine π‘ž NO YES We want it to satisfy βˆ€π‘₯ ∈ 𝑋. p πœ€π‘† w ≃ πœ€π΅(π‘₯)

27

slide-28
SLIDE 28

Check if π‘ž should be refined

  • config. space of 𝑆 (ℝ𝑒)
  • config. space of 𝐡 (β„π‘œ)

γƒ»πœ€π‘† π‘₯β€² γƒ»πœ€π΅ π‘₯β€² = π‘ž(πœ€π΅ π‘₯β€² ) π‘ž π‘₯β€²: a word already visited in the BFS loop π‘₯: a word just popped

28

slide-29
SLIDE 29

Check if π‘ž should be refined

  • config. space of 𝑆 (ℝ𝑒)
  • config. space of 𝐡 (β„π‘œ)

γƒ»πœ€π‘† π‘₯β€² γƒ»πœ€π΅ π‘₯β€² = π‘ž πœ€π΅ π‘₯β€² = πœ€π΅(π‘₯) γƒ»πœ€π‘†(π‘₯) π‘ž π‘₯β€²: a word already visited in the BFS loop π‘₯: a word just popped

29

slide-30
SLIDE 30

Check if π‘ž should be refined

  • config. space of 𝑆 (ℝ𝑒)
  • config. space of 𝐡 (β„π‘œ)

γƒ»πœ€π‘† π‘₯β€² γƒ»πœ€π΅ π‘₯β€² = π‘ž πœ€π΅ π‘₯β€² = πœ€π΅ π‘₯ = π‘ž(πœ€π‘† π‘₯ ) γƒ»πœ€π‘†(π‘₯)

  • config. space of 𝑆 (ℝ𝑒)
  • config. space of 𝐡 (β„π‘œ)

γƒ»πœ€π‘† π‘₯β€² γƒ»πœ€π΅ π‘₯β€² = π‘ž πœ€π΅ π‘₯β€² = πœ€π΅(π‘₯) γƒ»πœ€π‘†(π‘₯) π‘ž π‘ž π‘ž γƒ»π‘ž(πœ€π‘† π‘₯ ) π‘ž π‘₯β€²: a word already visited in the BFS loop π‘₯: a word just popped

30

slide-31
SLIDE 31

Check if π‘ž should be refined

  • config. space of 𝑆 (ℝ𝑒)
  • config. space of 𝐡 (β„π‘œ)

γƒ»πœ€π‘† π‘₯β€² γƒ»πœ€π΅ π‘₯β€² = π‘ž πœ€π΅ π‘₯β€² = πœ€π΅ π‘₯ = π‘ž(πœ€π‘† π‘₯ ) γƒ»πœ€π‘†(π‘₯)

  • config. space of 𝑆 (ℝ𝑒)
  • config. space of 𝐡 (β„π‘œ)

γƒ»πœ€π‘† π‘₯β€² γƒ»πœ€π΅ π‘₯β€² = π‘ž πœ€π΅ π‘₯β€² = πœ€π΅(π‘₯) γƒ»πœ€π‘†(π‘₯) π‘ž π‘ž π‘ž γƒ»π‘ž(πœ€π‘† π‘₯ ) π‘ž

↓This Violates π‘ž πœ€π‘† π‘₯ = πœ€π΅(π‘₯)

π‘₯β€²: a word already visited in the BFS loop π‘₯: a word just popped

31

slide-32
SLIDE 32

Maintaining π‘ž

Pop w from queue Add w’s next words to queue Check if π‘ž should be refined Refine π‘ž NO YES We want it to satisfy βˆ€π‘₯ ∈ 𝑋. p πœ€π‘† w ≃ πœ€π΅(π‘₯)

32

slide-33
SLIDE 33

Check if π‘ž should be refined Refine π‘ž

Finding Counterexample

Pop w from queue Add w’s next words to queue NO YES Check if 𝑔

𝑆 π‘₯

= 𝑔

𝐡(π‘₯)

If 𝑔

𝑆 π‘₯ β‰  𝑔 𝐡(π‘₯), returns π‘₯ as a counterexample of the

equivalence query. 𝑓 𝐡 = α‰Š Equivalent ; 𝑔

𝑆 ≃ 𝑔 𝐡

π‘₯β€²β€² ; 𝑔

𝑆 π‘₯β€²β€² β‰  𝑔 𝐡(π‘₯β€²β€²)

33

slide-34
SLIDE 34

Check if π‘ž should be refined Refine π‘ž

Returning β€œEquivalent”

Pop w from queue Add w’s next words to queue NO YES Check if 𝑔

𝑆 π‘₯

= 𝑔

𝐡(π‘₯)

If there are many (𝑁 = 5 times) visited words π‘₯β€² ∈ {Visited words} π‘ž ∘ πœ€π‘† π‘₯ = π‘ž ∘ πœ€π‘† π‘₯β€² }, the next words of π‘₯ is not added (Pruning the subtree under π‘₯ in BFS)

34

slide-35
SLIDE 35

Check if π‘ž should be refined Refine π‘ž Add w’s next words to queue NO YES

Returning β€œEquivalent”

Pop w from queue Check if 𝑔

𝑆 π‘₯

= 𝑔

𝐡(π‘₯)

When the queue is empty, all the trees are pruned and it returns β€œEquivalent”. 𝑓 𝐡 = α‰Š Equivalent ; 𝑔

𝑆 ≃ 𝑔 𝐡

π‘₯β€²β€² ; 𝑔

𝑆 π‘₯β€²β€² β‰  𝑔 𝐡(π‘₯β€²β€²)

35

slide-36
SLIDE 36

Experiments (Target RNNs)

90 target RNNs to evaluate our algorithm are made by

  • 1. Generate a random WFA 𝐡 of size π‘œ ∈ {10, 20, 30} and alphabet

Σ of size a ∈ {10,15,20,30,40,50}

  • 2. Learn RNN 𝑆(𝐡) from 𝐡
  • 3. Repeat Step 1-2 for each (π‘œ, 𝑑) 5 times.

RNNs consist of two-stacked LSTM with 50 cells.

36

slide-37
SLIDE 37

Experiments (Settings)

Methods

  • Our algorithm with 𝑁 = 5
  • Baseline algorithm (comes later)

Evaluation

  • Time to extract (timeout=10,000 sec.)
  • Accuracy
  • If 𝑔

𝑆 π‘₯ βˆ’ 𝑔 𝐡 𝑆 (π‘₯) < 0.05 then it is β€œcorrect”

  • Calculated by randomly generating 1000 words
  • Time to infer the words in 𝑆 𝐡 , 𝐡(𝑆 𝐡 )

37

slide-38
SLIDE 38

Experiments (Baseline algorithm)

Pop w from queue Add w’s next words to queue Check if 𝑔

𝑆 π‘₯

= 𝑔

𝐡(π‘₯)

If 𝑔

𝑆 π‘₯ β‰  𝑔 𝐡(π‘₯), returns π‘₯ as a counterexample of the

equivalence query. 𝑓 𝐡 = α‰Š Equivalent ; 𝑔

𝑆 ≃ 𝑔 𝐡

π‘₯β€²β€² ; 𝑔

𝑆 π‘₯β€²β€² β‰  𝑔 𝐡(π‘₯β€²β€²)

38

slide-39
SLIDE 39

Experiments (Baseline algorithm)

Pop w from queue Add w’s next words to queue Check if 𝑔

𝑆 π‘₯

= 𝑔

𝐡(π‘₯)

If 𝑔

𝑆 π‘₯ = 𝑔 𝐡(π‘₯) in a row (1000 times), returns β„Ž as a counterexample of the

equivalence query. 𝑓 𝐡 = α‰Š Equivalent ; 𝑔

𝑆 ≃ 𝑔 𝐡

π‘₯β€²β€² ; 𝑔

𝑆 π‘₯β€²β€² β‰  𝑔 𝐡(π‘₯β€²β€²)

(If this happens, queue is preserved for the next invoke of eq-query)

39

slide-40
SLIDE 40

Result (Overall)

Difference of accuracy and extracting time between ours and baseline

40

slide-41
SLIDE 41

Result (Overall)

Average (and Std) Ours(M=5) Baseline Accuracy[%] 81.9% (std=18.8%) 74.1% (std=22.9%) Time [s] 8805 (std=2220) 6277 (std=2966)

  • The accuracy of β€œOurs (M=5)” exceeded those of β€œBaseline”

in 59 tasks.

  • The extracting time of β€œOurs (M=5)” longer than those of

β€œBaseline” in 80 tasks.

  • (90 tasks in total)

41

slide-42
SLIDE 42

Result (WFA size π‘œ = 10)

Difference of accuracy and extracting time between ours and baseline

42

slide-43
SLIDE 43

Result (alphabet size 𝑏 = 10)

Difference of accuracy and extracting time between ours and baseline

43

slide-44
SLIDE 44

Time to Infer a Value from a Word

  • To test our motivation β€œGetting lighter (faster to infer) model
  • f an RNN” is feasible.
  • We compared the time to compute 𝑔

𝑆(π‘₯) and 𝑔 𝐡 𝑆 (π‘₯) for

1,000 words whose lengths are ≀ 20. Average Time on RNN 𝑆 [s] 32.0 (std=2.0) Time on WFA 𝐡(𝑆) [s] 0.028 (std=0.007)

44

slide-45
SLIDE 45

Conclusion

  • Proposed a method to extract the WFA 𝐡(𝑆) from a given

RNN 𝑆 so that 𝑔

𝐡(𝑆) ≃ 𝑔 𝑆.

  • Compared our method to the baseline algorithm in the

accuracy and time

  • Our algorithm achieved higher accuracy and took more time than the

baseline.

  • The extracted WFA 𝐡 𝑆 took less time to infer values than

the original RNN 𝑆

45

slide-46
SLIDE 46

Future Work

  • Adding experiment
  • To reveal the overall tendency clearly
  • To reveal what is happening when the accuracy is quite low
  • Adding the idea of bisimulation to π‘ž
  • Think of questionable parts in the loop?
  • Refining π‘ž at the different timing could be better?
  • Modifying Balle and Mohri’s algorithm to generate

probabilistic WFA

  • Finding good hyper parameter 𝑁 experimentally or

theoretically

46

slide-47
SLIDE 47

β€œChecking if π‘ž is OK” could be like this?

  • config. space of 𝑆 (ℝ𝑒)
  • config. space of 𝐡 (β„π‘œ)

γƒ»πœ€π‘† β„Žβ€² γƒ»πœ€π΅ β„Žβ€² = π‘ž(πœ€π΅ β„Žβ€² ) γƒ»πœ€π‘†(β„Ž) γƒ»πœ€π΅ β„Ž = πœ€π‘† β„Ž

  • config. space of 𝑆 (ℝ𝑒)
  • config. space of 𝐡 (β„π‘œ)

γƒ»πœ€π‘† β„Žβ€² γƒ»πœ€π΅ β„Žβ€² = π‘ž(πœ€π΅ β„Žβ€² ) γƒ»πœ€π‘†(β„Ž) γƒ»πœ€π΅(β„Ž) π‘ž π‘ž π‘ž γƒ»π‘ž(πœ€π‘†(β„Ž))

↓This Violates π‘ž πœ€π‘† β„Ž = πœ€π΅(β„Ž)

π‘ž

47

slide-48
SLIDE 48
  • Def. of WFA
  • WFA 𝐡 is probabilistic if
  • 𝛽 β‹… 𝟐 = 1
  • For all 𝜏 ∈ Ξ£, the sums of rows are 1
  • 0 ≀ 𝛾 ≀ 1 β– 

For example:

  • Ξ£ = 0, 1 , 𝛽 = 0.8

0.2 , 𝛾 = 0.9 0.7 , 𝐡0 = 0 1 1 0 , 𝐡1 = 0.9 0.1 0.5 0.5

48