time series forecasting with a learning algorithm an
play

Time Series Forecasting With a Learning Algorithm: An Approximate - PowerPoint PPT Presentation

Time Series Forecasting With a Learning Algorithm: An Approximate Dynamic Programming Approach Ricardo Collado 1 an Creamer 1 Germ 1 School of Business Stevens Institute of Technology Hoboken, New Jersey R. Collado (Stevens) Learning Time


  1. Time Series Forecasting With a Learning Algorithm: An Approximate Dynamic Programming Approach Ricardo Collado 1 an Creamer 1 Germ´ 1 School of Business Stevens Institute of Technology Hoboken, New Jersey R. Collado (Stevens) Learning Time Series and Dynamic Programming September 10, 2019 1 / 34

  2. Introduction Time Series Drivers: • Historical data: • Incorporated in classical time series forecast methods • Works best when the underlying model is fix • Exogenous processes: • Not included in historical observations • Difficult to incorporate via classical methods • Could indicate changes in the underlying model R. Collado (Stevens) Learning Time Series and Dynamic Programming September 10, 2019 2 / 34

  3. Introduction Techniques to handle changes due to external forces: • Jump Diffusion Models • Regime Switching Methods • System of Weighted Experts • Others . . . These methods do not directly integrate alternative data sources available to us R. Collado (Stevens) Learning Time Series and Dynamic Programming September 10, 2019 3 / 34

  4. Introduction Alternative data sources: • Text & News Analysis • Social Networks Data • Sentiment Analysis R. Collado (Stevens) Learning Time Series and Dynamic Programming September 10, 2019 4 / 34

  5. Introduction Alternative data sources: • Text & News Analysis • Social Networks Data • Sentiment Analysis R. Collado (Stevens) Learning Time Series and Dynamic Programming September 10, 2019 4 / 34

  6. Introduction Alternative data sources: • Text & News Analysis • Social Networks Data • Sentiment Analysis R. Collado (Stevens) Learning Time Series and Dynamic Programming September 10, 2019 4 / 34

  7. Introduction We study time series forecast methods that are: • Dynamic • Context-Based • Capable of Integrating Social, Text, and Sentiment Data In this presentation we develop: • Stochastic dynamic programming model for time series forecast • Rely on an “external forecast” for future values • External forecast allows to incorporate alternative data sources R. Collado (Stevens) Learning Time Series and Dynamic Programming September 10, 2019 5 / 34

  8. Traditional Approach Time Series Fitting Process R. Collado (Stevens) Learning Time Series and Dynamic Programming September 10, 2019 6 / 34

  9. Traditional Approach s 0 = x 0 a 0 = φ ∗ 1 X 1 = φ ∗ 1 x 0 + � 1 A = { φ ∈ R } s 0 R. Collado (Stevens) Learning Time Series and Dynamic Programming September 10, 2019 7 / 34

  10. Traditional Approach s 0 = x 0 a 0 = φ ∗ 1 X 1 = φ ∗ 1 x 0 + � 1 s 0 R. Collado (Stevens) Learning Time Series and Dynamic Programming September 10, 2019 7 / 34

  11. Traditional Approach s 1 = x 1 a 1 = φ ∗ 2 X 2 = φ ∗ 2 x 1 + � 2 s 1 s 0 R. Collado (Stevens) Learning Time Series and Dynamic Programming September 10, 2019 7 / 34

  12. Traditional Approach s 2 = x 2 a 2 = φ ∗ 3 X 3 = φ ∗ 3 x 2 + � 3 s 2 s 1 s 0 R. Collado (Stevens) Learning Time Series and Dynamic Programming September 10, 2019 7 / 34

  13. Traditional Approach s 3 = x 3 a 3 = φ ∗ 4 X 4 = φ ∗ 4 x 3 + � 4 s 3 s 2 s 1 s 0 R. Collado (Stevens) Learning Time Series and Dynamic Programming September 10, 2019 7 / 34

  14. Traditional Approach s 4 = x 4 a 4 = φ ∗ 5 X 5 = φ ∗ 5 x 4 + � 5 s 3 s 2 s 1 s 0 R. Collado (Stevens) Learning Time Series and Dynamic Programming September 10, 2019 7 / 34

  15. Traditional Time Series Fitting Main Problem: � T � � min c t ( s t , a t ) , π ∈ Π E t =1 where a t = π t ( x 1 , . . . , x t ) is an admissible fitting policy. • The time series model is parametrized by Θ ⊆ R d • A ( s ) = Θ for all states s • c t ( s , a ) is the result of a goodness of fit test for the observations s = ( x 0 , . . . , x t ) and model selection a = θ • Solution via Bellman’s optimality equations R. Collado (Stevens) Learning Time Series and Dynamic Programming September 10, 2019 8 / 34

  16. Traditional Time Series Fitting Conditions to guarantee optimality: • The set of actions A ( s ) is compact • The cost functions c t ( s , · ) are lower semicontinuous • For every measurable selection a t ( · ) ∈ A t ( · ), the functions s �→ c t ( s , a t ( s )) and c T ( · ) are elements of L 1 ( S , B S , P 0 ) • The DP stochastic kernel function Q t ( s , · ) is continuous R. Collado (Stevens) Learning Time Series and Dynamic Programming September 10, 2019 9 / 34

  17. Traditional Time Series Fitting Value Functions: Value functions v t : S → R , t = 1 , . . . , T , given recursively by: v T ( s ) = c T ( s ) v t ( s ) = min a ∈ A t ( s ) { c t ( s , a ) + E [ v t +1 | s , a ] } , for all s ∈ S and t = T − 1 , . . . , 0. Bellman’s Optimality Equations: Then an optimal Markov policy π ∗ = { π ∗ 0 , . . . , π ∗ T − 1 } exists and satisfies the equations: π ∗ t ( s ) ∈ arg min { c t ( s , a ) + E [ v t +1 | s , a ] } , s ∈ S , t = T − 1 , . . . , 0 . a ∈ A t ( s ) Conversely, any measurable solution of these is an optimal Markov policy π ∗ . R. Collado (Stevens) Learning Time Series and Dynamic Programming September 10, 2019 10 / 34

  18. Traditional Time Series Fitting Notice that: • Our choice of model does not affect future observations and cost. • So, E [ v | s , a ] = E [ v | s , a ′ ], for any ( s , a ) , ( s , a ′ ) ∈ graph( A ). • Therefore we can rewrite the optimal policy as: π ∗ t ( s ) ∈ arg min { c t ( s , a ) } , s ∈ S , t = T − 1 , . . . , 0 . a ∈ A t ( s ) R. Collado (Stevens) Learning Time Series and Dynamic Programming September 10, 2019 11 / 34

  19. Traditional Time Series Fitting Notice that: • Our choice of model does not affect future observations and cost. • So, E [ v | s , a ] = E [ v | s , a ′ ], for any ( s , a ) , ( s , a ′ ) ∈ graph( A ). • Therefore we can rewrite the optimal policy as: π ∗ t ( s ) ∈ arg min { c t ( s , a ) } , s ∈ S , t = T − 1 , . . . , 0 . a ∈ A t ( s ) • The optimal policy π ∗ is purely myopic R. Collado (Stevens) Learning Time Series and Dynamic Programming September 10, 2019 11 / 34

  20. Traditional Time Series Fitting Q: How to break with the myopic policy? R. Collado (Stevens) Learning Time Series and Dynamic Programming September 10, 2019 12 / 34

  21. Traditional Time Series Fitting A: Introduce a new Markov model R. Collado (Stevens) Learning Time Series and Dynamic Programming September 10, 2019 13 / 34

  22. New Markov Model New Markov Model: • Given stochastic process { X t | t = 0 , . . . , T } , s.t. X 0 = { φ 0 } • Time series model parameterized by Θ ⊆ R d • State space:  �  �  x t observation from X t ,   �  � S t = ( x t , h t − 1 , θ t − 1 ) h t = x 0 , . . . , x t − 1 sample sequence , �  �    � θ t − 1 = ( φ 1 , . . . , φ p ) ∈ Θ • Action space: A ( s ) = Θ for all states s ∈ S t R. Collado (Stevens) Learning Time Series and Dynamic Programming September 10, 2019 14 / 34

  23. New Markov Model Cost function: c t ( s , θ t ) = γ ( s t , θ t ) + r δ ( s t , θ t − 1 , θ t ) • γ : Goodness of fit test • δ : Penalty on changes from previous model selection • r ≥ 0: Scaling factor used to balance fit and penalty R. Collado (Stevens) Learning Time Series and Dynamic Programming September 10, 2019 15 / 34

  24. New Markov Model Example: � � �� � � � ��� χ 2 ( θ t | h t − 1 , x t ) + r � E [ P θ t | h t − 1 , x t ] − E � h t − 1 , x t 1 − exp − λ P θ t − 1 , where r , λ ≥ 0. • γ ( s t , θ t ) := χ 2 ( θ t | h t − 1 , x t ) � � � � �� �� � E [ P θ t | h t − 1 , x t ] − E � h t − 1 , x t • δ ( s t , θ t − 1 , θ t ) := 1 − exp − λ P θ t − 1 R. Collado (Stevens) Learning Time Series and Dynamic Programming September 10, 2019 16 / 34

  25. New Markov Model Dynamic Optimization Model: s 0 = x 0 a 0 = φ ∗ 1 X 1 = φ ∗ 1 x 0 + � 1 A = { φ ∈ R } s 0 R. Collado (Stevens) Learning Time Series and Dynamic Programming September 10, 2019 17 / 34

  26. New Markov Model Dynamic Optimization Model: s 0 = x 0 a 0 = φ ∗ 1 γ enforces good fit X 1 = φ ∗ 1 x 0 + � 1 s 0 downstream δ adds cost for changing model R. Collado (Stevens) Learning Time Series and Dynamic Programming September 10, 2019 17 / 34

  27. New Markov Model Dynamic Optimization Model: s 0 = x 0 a 0 = φ ∗ 1 γ enforces good fit X 1 = φ ∗ 1 x 0 + � 1 s 0 downstream δ adds cost for changing model R. Collado (Stevens) Learning Time Series and Dynamic Programming September 10, 2019 17 / 34

  28. New Markov Model Dynamic Optimization Model: s 1 = x 1 a 1 = φ ∗ 2 X 2 = φ ∗ 2 x 1 + � 2 s 1 s 0 R. Collado (Stevens) Learning Time Series and Dynamic Programming September 10, 2019 17 / 34

  29. New Markov Model Dynamic Optimization Model: s 2 = x 2 a 2 = φ ∗ 3 X 3 = φ ∗ 3 x 2 + � 3 s 2 s 1 s 0 R. Collado (Stevens) Learning Time Series and Dynamic Programming September 10, 2019 17 / 34

  30. New Markov Model Dynamic Optimization Model: s 3 = x 3 a 3 = φ ∗ 4 X 4 = φ ∗ 4 x 3 + � 4 s 3 s 2 s 1 s 0 R. Collado (Stevens) Learning Time Series and Dynamic Programming September 10, 2019 17 / 34

  31. Lookahead Methods Introducing Exogenous Information Via Lookahead Methods R. Collado (Stevens) Learning Time Series and Dynamic Programming September 10, 2019 18 / 34

Download Presentation
Download Policy: The content available on the website is offered to you 'AS IS' for your personal information and use only. It cannot be commercialized, licensed, or distributed on other websites without prior consent from the author. To download a presentation, simply click this link. If you encounter any difficulties during the download process, it's possible that the publisher has removed the file from their server.

Recommend


More recommend