exploring interpretable lstm neural networks over multi
play

Exploring interpretable LSTM neural networks over multi-variable - PowerPoint PPT Presentation

Exploring interpretable LSTM neural networks over multi-variable data Sebastian U. Stich (MLO, EPFL) on behalf of the authors Tian Guo, COSS, ETH Zurich Tao Lin, MLO, EPFL Nino Antulov-Fantulin, COSS, ETH Zurich | | 13 June 2019 1


  1. Exploring interpretable LSTM neural networks over multi-variable data Sebastian U. Stich (MLO, EPFL) on behalf of the authors Tian Guo, COSS, ETH Zurich Tao Lin, MLO, EPFL Nino Antulov-Fantulin, COSS, ETH Zurich | | 13 June 2019 1

  2. Problem formulation  Multi-variable time series  Target and exogenous variables  Predictive model | 13 June 2019 | 2

  3. Problem formulation  Weak interpretability of RNNs on multi-variable data  Multi-variable input to hidden states i.e. vectors  No correspondence between hidden states and input variables  Different dynamics of variables are mingled in hidden states | 13 June 2019 | 3

  4. Problem formulation  Interpretable prediction model on multi-variable time series  Accurate  Capture different dynamics of input variables  Interpretable  Variable importance w.r.t. predictive power i.e. which variable is more important for RNNs to perform prediction  Temporal importance of each variable i.e. short or long-term correlation to the target | | 13 June 2019 4

  5. Interpretable multi-variable LSTM  IMV-LSTM  Key ideas: | 13 June 2019 | 5

  6. IMV-LSTM  IMV-LSTM with variable-wise hidden states  Conventional LSTM with hidden vectors | 13 June 2019 | 6

  7. Results  Variable importance  Learned during the training  The higher the value, the more important  Variable-wise temporal importance  The lighter the color, the more important | 13 June 2019 | 7

  8. Conclusion  Explored the internal structures of LSTMs to enable variable-wise hidden states.  Developed mixture attention and associated learning procedure to quantify variable importance and variable-wise temporal importance w.r.t. the target.  Extensive experiments provide insights into achieving superior prediction performance and importance interpretation for LSTM. | | 13 June 2019 8

  9. Backup Network architecture: Mixture attention to model generative process of the target: | 13 June 2019 | 9

Download Presentation
Download Policy: The content available on the website is offered to you 'AS IS' for your personal information and use only. It cannot be commercialized, licensed, or distributed on other websites without prior consent from the author. To download a presentation, simply click this link. If you encounter any difficulties during the download process, it's possible that the publisher has removed the file from their server.

Recommend


More recommend