DL Approaches to Tim ime Series Data Miro Enev, DL Solution - - PowerPoint PPT Presentation

dl approaches to tim ime series data
SMART_READER_LITE
LIVE PREVIEW

DL Approaches to Tim ime Series Data Miro Enev, DL Solution - - PowerPoint PPT Presentation

DL Approaches to Tim ime Series Data Miro Enev, DL Solution Architect Jeff Weiss, Director West SAs Agenda Define Time Series [ Examples & Brief Summary of Considerations ] Semi-supervised Anomaly Detection [ with Deep Autoencoders


slide-1
SLIDE 1

DL Approaches to Tim ime Series Data

Miro Enev, DL Solution Architect Jeff Weiss, Director West SAs

slide-2
SLIDE 2

Agenda

  • Define Time Series

[ Examples & Brief Summary of Considerations ]

  • Semi-supervised Anomaly Detection

[ with Deep Autoencoders ]

  • Ensemble event detection and classification

[ with MLPs & CNN-MLPs ]

  • Prediction

[ with Dual Attentional RNNs (DA-RNNs) ]

slide-3
SLIDE 3

Time Series [ [ One Defi finition ]

  • A time series is a series of values recorded over equally spaced time intervals.
  • The amount of time between observations is the sampling interval or sampling rate.
  • The time series represent some underlying, partially-random (“stochastic”)

process which generated the data.

  • We want to use the data to make guesses (“inferences”) about the process, and

want to make reliable guesses while being clear about the uncertainty involved.

  • The complication is that each observation is dependent on all the other observations, and in

fact its this dependence that we want to learn and draw inferences about.

slide-4
SLIDE 4

Popularity of f French Names

slide-5
SLIDE 5

NVIDIA Stock

https://www.google.com/finance?chdnp=0&chdd=0&chds=1& chdv=1&chvs=Linear&chdeh=0&chfdeh=0&chdet=1494360000 000&chddm=98532&chls=IntervalBasedLine&q=NASDAQ:NVD A&ntsp=0&ei=N_kRWemBAoqxjAHRgZDwBA

slide-6
SLIDE 6

Categories of f Time Series Analysis

  • Prediction/Forecasting
  • Weather, Sales, Inventory, Financial Markets
  • Anomaly Detection
  • Predictive Maintenance
  • Reconstruction
  • Virtual Sensors
  • Classification
  • Event/Activity Recognition
slide-7
SLIDE 7

Time Series Best Practices

  • Think hard about the input(s)
  • Time and frequency domain representations
  • Lags and windowing (multi-time scale)
  • Study correlational structure
  • Use semi/unsupervised approaches [ free ]
  • Feature Building, Data Buckets, Anomaly Detection
  • Gather and/or generate ground truth
  • Class imbalance & sample weighting
  • Try DL fabrics with increasing complexity & Ensemble
  • Use a non DL baseline [ ARIMA, SVMs, random forests ]
  • Use forward chained cross-validation for evaluation
  • Train attentional mechanism*
slide-8
SLIDE 8

DL for Anomaly Detection

slide-9
SLIDE 9

Deep Autoencoder Anomaly Detection

13 Sensors, 100Hz, NASA Dataset, .5 seconds window, 650 dimensions per sample [ 256, 196, 136, 76, 14 ]

Anomaly Detection and Fault Disambiguation in Large Flight Data: A Multi-modal Deep Auto-encoder Approach, K. Reddy et al, United Technologies Research Center (PHM16)

slide-10
SLIDE 10

Reconstruction & Feature Learning

Exceptionally low normalized RMS reconstruction error (0.04 – 0.09)

Anomaly Detection and Fault Disambiguation in Large Flight Data: A Multi-modal Deep Auto-encoder Approach, K. Reddy et al, United Technologies Research Center (PHM16)

Sample Reconstructions

slide-11
SLIDE 11

Anomaly Detection and Fault Disambiguation in Large Flight Data: A Multi-modal Deep Auto-encoder Approach, K. Reddy et al, United Technologies Research Center (PHM16)

11-layer 14-dimensional bottleneck DAE yields 97.8% true positive detection rate with 0.0% false alarm Artifcially created anomalies [ Spall Fault ; Ballscrew Jam ]

Anomaly Analysis

slide-12
SLIDE 12

Anomaly Detection and Fault Disambiguation in Large Flight Data: A Multi-modal Deep Auto-encoder Approach, K. Reddy et al, United Technologies Research Center (PHM16)

11-layer 14-dimensional bottleneck DAE yields 97.8% true positive detection rate with 0.0% false alarm Artificially created anomalies [ Spall Fault ; Ballscrew Jam ]

Anomaly Analysis/Interpretation

slide-13
SLIDE 13

DL for Time Series Classification

slide-14
SLIDE 14

Io IoT Case Study – Water Disagg ggregation

slide-15
SLIDE 15
slide-16
SLIDE 16
slide-17
SLIDE 17

Frequency Domain Utilization

slide-18
SLIDE 18

Automated Ground Truth Building

Open Edges Close Edges

slide-19
SLIDE 19
slide-20
SLIDE 20

Event/Edge Detection

slide-21
SLIDE 21

Open Edge Clo Close ed edge

Edge Characterization, , Example 1

slide-22
SLIDE 22

Edge Characterization, , Example 2

Clo Close ed edge Open Edge

slide-23
SLIDE 23
slide-24
SLIDE 24

DL for Time Series Prediction

slide-25
SLIDE 25
slide-26
SLIDE 26

Dual Attention RNNs

slide-27
SLIDE 27

Performance @ NASDAQ 100

  • In NASDAQ 100 Stock dataset, we collect the stock prices of 81 major corporations under NASDAQ 100, which are

used as the driving time series. The index value of NASDAQ 100 is used as the target series. The frequency of the data collection is one-minute. This data covers the period from July 26, 2016 to December 26, 2016, in total 104

  • days. Each day contains 390 data points from the opening to closing of the market. In our experiments, we use the

first 90 days as the training set and the following seven days as the validation.

slide-28
SLIDE 28

Performance @ SML 2010

SML 2010 is a public dataset used for indoor temperature forecasting. This dataset is collected from a monitor system mounted in a domestic house. We use room temperature as the target series and select 16 relevant driving series which contains approximately 40 days of monitoring data. The data was sampled every minute and was smoothed with 15 minute means. In our experiment, we use the first 3200 data points as the training set, the following 400 data points as the validation set, and the last 537 data points as the test set.

slide-29
SLIDE 29

Attention Mechanism at Work

Train Set [ NASDAQ ] Test Set [ NASDAQ ]

slide-30
SLIDE 30

Thanks!