Exploring interpretable LSTM neural networks over multi-variable - - PowerPoint PPT Presentation

exploring interpretable lstm neural networks over multi
SMART_READER_LITE
LIVE PREVIEW

Exploring interpretable LSTM neural networks over multi-variable - - PowerPoint PPT Presentation

Exploring interpretable LSTM neural networks over multi-variable data Sebastian U. Stich (MLO, EPFL) on behalf of the authors Tian Guo, COSS, ETH Zurich Tao Lin, MLO, EPFL Nino Antulov-Fantulin, COSS, ETH Zurich | | 13 June 2019 1


slide-1
SLIDE 1

| |

Sebastian U. Stich (MLO, EPFL)

  • n behalf of the authors

Tian Guo, COSS, ETH Zurich Tao Lin, MLO, EPFL Nino Antulov-Fantulin, COSS, ETH Zurich

13 June 2019 1

Exploring interpretable LSTM neural networks

  • ver multi-variable data
slide-2
SLIDE 2

| | 13 June 2019 2

Problem formulation

  • Multi-variable time series
  • Target and exogenous variables
  • Predictive model
slide-3
SLIDE 3

| | 13 June 2019 3

Problem formulation

  • Weak interpretability of RNNs on multi-variable data
  • Multi-variable input to hidden states

i.e. vectors

  • No correspondence between hidden states

and input variables

  • Different dynamics of variables are mingled

in hidden states

slide-4
SLIDE 4

| |

  • Interpretable prediction model on multi-variable time series
  • Accurate
  • Capture different dynamics of input variables
  • Interpretable
  • Variable importance w.r.t. predictive power

i.e. which variable is more important for RNNs to perform prediction

  • Temporal importance of each variable

i.e. short or long-term correlation to the target

13 June 2019 4

Problem formulation

slide-5
SLIDE 5

| | 13 June 2019 5

Interpretable multi-variable LSTM

  • IMV-LSTM
  • Key ideas:
slide-6
SLIDE 6

| | 13 June 2019 6

IMV-LSTM

  • IMV-LSTM with variable-wise hidden states
  • Conventional LSTM with hidden vectors
slide-7
SLIDE 7

| | 13 June 2019 7

Results

  • Variable importance
  • Learned during the training
  • The higher the value, the more important
  • Variable-wise temporal importance
  • The lighter the color, the more important
slide-8
SLIDE 8

| |

  • Explored the internal structures of LSTMs to enable variable-wise hidden

states.

  • Developed mixture attention and associated learning procedure to quantify

variable importance and variable-wise temporal importance w.r.t. the target.

  • Extensive experiments provide insights into achieving superior prediction

performance and importance interpretation for LSTM.

13 June 2019 8

Conclusion

slide-9
SLIDE 9

| | 13 June 2019 9

Backup

Network architecture: Mixture attention to model generative process of the target: