TRADING USING DEEP LEARNING MAN VS MACHINE Orders By Algorithms - - PowerPoint PPT Presentation

โ–ถ
trading using deep learning man vs machine
SMART_READER_LITE
LIVE PREVIEW

TRADING USING DEEP LEARNING MAN VS MACHINE Orders By Algorithms - - PowerPoint PPT Presentation

TRADING USING DEEP LEARNING MAN VS MACHINE Orders By Algorithms 84% Orders By Human 16% TRADING USING DEEP LEARNING Artificial Neural Networks Neural networks are a family of models inspired by biological brain structure and are used to


slide-1
SLIDE 1

TRADING USING DEEP LEARNING

slide-2
SLIDE 2

MAN VS MACHINE

TRADING USING DEEP LEARNING

84%

Orders By Algorithms

16%

Orders By Human

slide-3
SLIDE 3

Artificial Neural Networks

Neural networks are a family of models inspired by biological brain structure and are used to estimate or approximate functions that can depend on a large number

  • f inputs and are generally unknown.

Recent breakthroughs in artificial neural networks led to a modern renascence in AI.

TRADING USING DEEP LEARNING

slide-4
SLIDE 4

DEEP LEARNING SUPERIORITY

96.92%

Deep Learning

94.9%

Human

DEEP META LEARNING

ref: http://www.image-net.org/challenges/LSVRC/

slide-5
SLIDE 5

GRADIENT DESCENT

๐น = Error of the network

๐‘ฅ๐‘ข = ๐‘ฅ๐‘ขโˆ’1 โˆ’ ๐›ฟ ๐œ–๐น ๐œ–๐‘ฅ

๐‘‹ = Weight matrix representing the filters

DEEP META LEARNING

slide-6
SLIDE 6

GRADIENT BASED MODELS

Legend

๐‘ฆ0

๐‘”

0(๐‘ฆ0, ๐‘ฅ0)

๐‘”

1(๐‘ฆ1, ๐‘ฅ1)

๐‘”

2(๐‘ฆ2, ๐‘ฅ2)

๐‘”

๐‘œ ๐‘ฆ๐‘œ, ๐‘ฅ๐‘œ = เทœ

๐‘ง

๐‘”

๐‘œโˆ’1(๐‘ฆ๐‘œโˆ’1, ๐‘ฅ๐‘œโˆ’1)

๐‘”

๐‘œโˆ’2(๐‘ฆ๐‘œโˆ’2, ๐‘ฅ๐‘œโˆ’2)

๐‘ฅ0 ๐‘ฅ1 ๐‘ฅ๐‘œ ๐‘ฅ๐‘œโˆ’1

๐น = ๐‘š เทœ ๐‘ง, ๐‘ง ๐‘ง

๐‘š เทœ ๐‘ง, ๐‘ง

  • Loss Function

๐‘ฆ0 - Features Vector ๐‘ฆ๐‘— - Output of ๐‘— layer ๐‘ฅ๐‘— - Weights of ๐‘— layer ๐‘ง โ€“ Ground Truth เทœ ๐‘ง โ€“ Model Output ๐น โ€“ Loss Surface

๐œ–๐น ๐œ–๐‘ฆ๐‘œ = ๐œ–๐‘š เทœ ๐‘ง, ๐‘ง ๐œ–๐‘ฆ๐‘œ ๐œ–๐น ๐œ–๐‘ฅ๐‘œ = ๐œ–๐น ๐œ–๐‘ฆ๐‘œ ๐œ–๐‘”

๐‘œ ๐‘ฆ๐‘œโˆ’1, ๐‘ฅ๐‘œ

๐œ–๐‘ฅ๐‘œ ๐œ–๐น ๐œ–๐‘ฆ๐‘œโˆ’1 = ๐œ–๐น ๐œ–๐‘ฆ๐‘œ ๐œ–๐‘”

๐‘œ ๐‘ฆ๐‘œโˆ’1, ๐‘ฅ๐‘œ

๐‘ฆ๐‘œโˆ’1

๐‘”โ€“ Activation Function

๐œ–๐น ๐œ–๐‘ฆ๐‘œโˆ’2 = ๐œ–๐น ๐œ–๐‘ฆ๐‘œโˆ’1 ๐œ–๐‘”

๐‘œโˆ’1 ๐‘ฆ๐‘œโˆ’2, ๐‘ฅ๐‘œโˆ’1

๐‘ฆ๐‘œโˆ’2 ๐œ–๐น ๐œ–๐‘ฅ๐‘œโˆ’1 = ๐œ–๐น ๐œ–๐‘ฆ๐‘œโˆ’1 ๐œ–๐‘”

๐‘œ ๐‘ฆ๐‘œโˆ’2, ๐‘ฅ๐‘œโˆ’1

๐œ–๐‘ฅ๐‘œโˆ’1

โ€ฆ โ€ฆ ๐บ๐‘๐‘ ๐‘ฅ๐‘๐‘ ๐‘’ ๐‘„๐‘ ๐‘๐‘ž๐‘๐‘•๐‘๐‘ข๐‘—๐‘๐‘œ

๐ถ๐‘๐‘‘๐‘™ ๐‘„๐‘ ๐‘๐‘ž๐‘๐‘•๐‘๐‘ข๐‘—๐‘๐‘œ

1: Forward Propagation 2: Loss Calculation 3: Optimization

DEEP META LEARNING

slide-7
SLIDE 7

Supervised Learning in a nutshell.

CAT Learning From Examples.

slide-8
SLIDE 8

FINNANCIAL PREDICTION PITFALLS

Much Data

Possible relevant data from many markets is incredibly large.

No Theory

Complex non-linear interactions in the data are not well specified by financial theory.

Noisy Data

Noise In financial data Is very common and sometimes distinguishing noise from behavior is hard.

Importance

Data Importance is questionable and determination of meaningful data is hard.

Overfitting

Overfitted easily, most models have poor predictive capabilities On financial data.

Behavior

Behavior of financial markets change all the time and can be really unpredictable. TRADING USING DEEP LEARNING

slide-9
SLIDE 9

WHY DEEP LEARNING?

Much Data

Possible relevant data from many markets is incredibly large.

No Theory

Complex non-linear interactions in the data are not well specified by financial theory.

Noisy Data

Noise In financial data Is very common and sometimes distinguishing noise from behavior is hard.

Importance

Data Importance is questionable and determination of meaningful data is hard.

Overfitting

Overfitted easily, most models have poor predictive capabilities On financial data.

Behavior

Behavior of financial markets change all the time and can be really unpredictable. TRADING USING DEEP LEARNING

slide-10
SLIDE 10

160.5 161 161.5 162 162.5 163 163.5 164 35:09.9 36:57.2 38:04.2 40:03.3 41:06.0 42:41.7 44:12.7 45:35.9 47:39.1 49:05.9 50:17.6 51:31.1 53:04.0 54:36.3 56:31.8 58:07.0 59:29.3 00:57.2 02:31.0 04:05.5 08:16.0 10:11.6 12:09.7 17:17.8 20:37.3 22:09.9 28:36.1 30:52.1 33:07.0 36:46.1 39:29.7 42:05.1 46:26.1 49:55.8 53:08.2 54:59.0 57:12.2 00:35.6 04:10.1 06:56.1 09:30.2 12:35.7 15:44.1 19:03.8 21:05.0 23:13.6 28:12.2 32:33.9 38:22.9 43:34.5 46:47.4 52:54.0 00:33.2 09:54.6 22:43.6 33:06.1 44:02.0 59:27.2 03:38.7 14:35.6 .5 1 .5 2 .5 3 .5 4 35:09.9 37:08.7 38:59.2 40:38.7 42:17.6 43:49.0 45:35.9 47:55.2 49:32.2 51:02.2 52:25.6 54:24.2 56:31.8 58:27.3 59:58.8 01:43.3 03:22.2 08:03.3 10:11.6 14:04.0 17:37.5 21:19.2 26:03.8 30:46.0 33:07.0 36:54.2 40:00.4 43:27.5 48:22.7 52:26.2 54:59.0 57:33.4 01:00.6 04:41.4 08:23.5 11:20.4 15:44.1 19:15.9 21:24.3 25:00.7 30:25.5 37:26.4 43:34.5 47:56.6 53:40.0 01:46.4 21:12.6 31:31.7 44:02.0 59:33.3 12:50.0

BACK TO FINANCE

?

TRADING USING DEEP LEARNING

slide-11
SLIDE 11

Strategy Universe

Strategy Configurations Configuration

DEEP REINFORCEMENT LEARNING

Trading Decision Utility

1 - buy 0 - hold

  • 1 - Sell

P&L / Drawdown

slide-12
SLIDE 12

DEEP LEARNING IN FINANCE

Technical analysis Might of might not work, One thing for sure: Very hard to generalize.

Technical Analysis

talib.SMA(โ€ฆ talib.MOM(โ€ฆ

TA-Lib : Technical Analysis Library

slide-13
SLIDE 13

Successful Technical Trading Agents Using Genetic Programming Farnsworth , 2004

DEEP LEARNING IN FINANCE

Surprisingly, Genetic programing can be very successful when It comes to financial strategies gp = SymbolicRegressor(... gp.fit(X_train, y_train)

slide-14
SLIDE 14

DEEP LEARNING IN FINANCE

slide-15
SLIDE 15

APPLYING DEEP LEARNING TO ENHANCE MOMENTUM TRADING STRATEGIES IN STOCKS

DEEP LEARNING IN FINANCE

slide-16
SLIDE 16

APPLYING DEEP LEARNING TO ENHANCE MOMENTUM TRADING STRATEGIES IN STOCKS L Takeuchi, 2013

๐‘Œ๐‘—: ฯƒ๐‘ขโˆ’12

๐‘ข+๐‘—

๐‘๐‘—๐‘’๐‘ขโˆ’๐‘๐‘ก๐‘™๐‘ขโˆ’12 ๐‘๐‘ก๐‘™๐‘ขโˆ’12

โˆ’๐œˆ ๐œ

| ๐‘— โˆˆ (1,11), โˆช 1 ๐‘—๐‘” ๐‘ข ๐‘—๐‘œ ๐‘˜๐‘๐‘œ๐‘ฃ๐‘๐‘ ๐‘ง ๐‘“๐‘š๐‘ก๐‘“ 0

๐บ๐‘“๐‘๐‘ข๐‘ฃ๐‘ ๐‘“ ๐‘Š๐‘“๐‘‘๐‘ข๐‘๐‘ 

FEATURE ENGINEERI NG MODEL RESULT S

DEEP LEARNING IN FINANCE

๐‘Œ๐‘—: ฯƒ๐‘ขโˆ’12

๐‘ข+๐‘—

๐‘๐‘—๐‘’๐‘ขโˆ’๐‘๐‘ก๐‘™๐‘ขโˆ’12 ๐‘๐‘ก๐‘™๐‘ขโˆ’12

โˆ’๐œˆ ๐œ

| ๐‘— โˆˆ (1,11), โˆช 1 ๐‘—๐‘” ๐‘ข ๐‘—๐‘œ ๐‘˜๐‘๐‘œ๐‘ฃ๐‘๐‘ ๐‘ง ๐‘“๐‘š๐‘ก๐‘“ 0

๐ป๐‘ ๐‘๐‘ฃ๐‘œ๐‘’ ๐‘ˆ๐‘ ๐‘ฃ๐‘ขโ„Ž

40 33 4 50 2 40 33 ๐‘Œ:

๐‘๐‘—๐‘’๐‘ข+1 โˆ’ ๐‘๐‘ก๐‘™๐‘ขโˆ’12 ๐‘๐‘ก๐‘™๐‘ขโˆ’12 > เท

๐‘ขโˆ’12 ๐‘ข+๐‘— ๐‘๐‘—๐‘’๐‘ข โˆ’ ๐‘๐‘ก๐‘™๐‘ขโˆ’12

๐‘๐‘ก๐‘™๐‘ขโˆ’12

๐ป๐‘ ๐‘๐‘ฃ๐‘œ๐‘’ ๐‘ˆ๐‘ ๐‘ฃ๐‘ขโ„Ž

12 Monthly Returns For every month: Daily cumulative returns Z-score Against other cumulative returns of other stocks Flag if January Layers Structure Hyper Parameters

K - Fold Stacked RBMs Layers size found by grid search Not Written

Confusion Matrix

True False

Predicted True Predicted False 22.38% 19.19% 30.97% 27.45%

Precision: 61.224% Recall: 53.659% Accuracy: 53.061%

slide-17
SLIDE 17

DEEP MODELING COMPLEX COUPLINGS WITHIN FINANCIAL MARKETS

DEEP LEARNING IN FINANCE

slide-18
SLIDE 18

DEEP MODELING COMPLEX COUPLINGS WITHIN FINANCIAL MARKETS

๐‘Œ๐‘—: ๐‘‡๐‘— โˆช ๐บ๐‘—

๐บ๐‘“๐‘๐‘ข๐‘ฃ๐‘ ๐‘“ ๐‘Š๐‘“๐‘‘๐‘ข๐‘๐‘ 

FEATURE ENGINEERI NG MODEL RESULT S

DEEP LEARNING IN FINANCE

R B M

S T O C K

Used Deep Belief Network to find hidden couplings between markets Used Past Prices of stocks and forex as features Unsupervised Learning Model Layers Structure Hyper Parameters

DBN of Stacked RBMs Optimizer: SGD Loss: Negative Log Likelihood Note: No Cross Validation in paper

F O R E X

R B M R B M R B M

slide-19
SLIDE 19

DEEP LEARNING FOR MULTIVARIATE FINANCIAL TIME SERIES

DEEP LEARNING IN FINANCE

slide-20
SLIDE 20

DEEP LEARNING FOR MULTIVARIATE FINANCIAL TIME SERIES

๐‘Œ๐‘—:

log

๐‘๐‘—๐‘’๐‘ข ๐‘๐‘ก๐‘™๐‘ขโˆ’1

โˆ’๐œˆ ๐œ

| ๐‘— โˆˆ (โˆ’33, โˆ’2), โˆช 1 ๐‘—๐‘” ๐‘ข ๐‘—๐‘œ ๐‘˜๐‘๐‘œ๐‘ฃ๐‘๐‘ ๐‘ง ๐‘“๐‘š๐‘ก๐‘“ 0

๐บ๐‘“๐‘๐‘ข๐‘ฃ๐‘ ๐‘“ ๐‘Š๐‘“๐‘‘๐‘ข๐‘๐‘ 

FEATURE ENGINEERI NG MODEL RESULT S

DEEP LEARNING IN FINANCE

๐‘: 1 ๐‘—๐‘” ๐‘„๐‘ ๐‘—๐‘‘๐‘“ ๐‘—๐‘ก ๐‘๐‘ค๐‘“๐‘  ๐‘›๐‘“๐‘’๐‘—๐‘๐‘œ ๐‘๐‘ข ๐‘ข + 1 ๐ป๐‘ ๐‘๐‘ฃ๐‘œ๐‘’ ๐‘ˆ๐‘ ๐‘ฃ๐‘ขโ„Ž

D E N S e

1

๐ป๐‘ ๐‘๐‘ฃ๐‘œ๐‘’ ๐‘ˆ๐‘ ๐‘ฃ๐‘ขโ„Ž

Matrix of log returns over all the stocks Z-score Against other stocks log returns Flag if January Layers Structure Hyper Parameters

DBN Connected to MLP Optimizer: ADAGrad CrosVal: Tarining: 70%, Valid: 15%, Test:

15%

Loss: Negative Log Likelihood

R B M R B M R B M

โ€œDeepโ€ Belief Net Fully Connected

Activation: ๐‘ข๐‘๐‘œโ„Ž Regularization: ๐‘€1

slide-21
SLIDE 21

IMPLEMENTING DEEP NEURAL NETWORKS FOR FINANCIAL MARKET PREDICTION

DEEP LEARNING IN FINANCE

slide-22
SLIDE 22

IMPLEMENTING DEEP NEURAL NETWORKS FOR FINANCIAL MARKET PREDICTION

๐‘Œ: แˆซ

๐‘— ๐‘—โˆ’100 ๐‘ž๐‘ข โˆ’ ๐‘ž๐‘ขโˆ’1

๐‘ž๐‘ขโˆ’1 โˆช แˆซ

๐‘”=5 100

๐‘๐ต( ๐‘Œ, ๐‘”) โˆช แˆซ

๐‘˜=1 ๐‘

๐œ(๐‘Œ, ๐‘Œ

๐‘˜)

๐บ๐‘“๐‘๐‘ข๐‘ฃ๐‘ ๐‘“ ๐‘Š๐‘“๐‘‘๐‘ข๐‘๐‘ 

FEATURE ENGINEERI NG MODEL RESULT S

DEEP LEARNING IN FINANCE

๐‘ โˆˆ 1,0, โˆ’1 ๐‘”๐‘๐‘  ๐‘๐‘ฃ๐‘ง, ๐‘ก๐‘“๐‘š๐‘š, โ„Ž๐‘๐‘š๐‘’ ๐ป๐‘ ๐‘๐‘ฃ๐‘œ๐‘’ ๐‘ˆ๐‘ ๐‘ฃ๐‘ขโ„Ž

1

๐ป๐‘ ๐‘๐‘ฃ๐‘œ๐‘’ ๐‘ˆ๐‘ ๐‘ฃ๐‘ขโ„Ž

All moving averages from 5 to 100 List of 100 lagged prices Pearson correlation between the returns (all 100) of the stock and all the other stocks(45) Layers Structure Hyper Parameters

Simple Fully Connected Optimizer: SGD CrosVal: Tarining: 80%, Test:

20%

Loss: Categorical Cross Entropy

1000

100 135

Fully Connected

Activation: ReLU, ๐‘ก๐‘๐‘”๐‘ข๐‘›๐‘๐‘ฆ

Training Algorithm: Walk forward

Accuracy: 73% F1 Score: 0.4

9895
slide-23
SLIDE 23

DEEP LEARNING IN FINANCE

DEEP LEARNING IN FINANCE

slide-24
SLIDE 24

DEEP LEARNING IN FINANCE Heaton et al, 2016

๐‘ž๐‘—๐‘˜: | ๐‘— โˆˆ (0,365), j โˆˆ (1,500)

๐บ๐‘“๐‘๐‘ข๐‘ฃ๐‘ ๐‘“ ๐‘Š๐‘“๐‘‘๐‘ข๐‘๐‘ 

FEATURE ENGINEERI NG MODEL RESULT S

DEEP LEARNING IN FINANCE

๐‘ž๐‘—๐‘˜: | ๐‘— โˆˆ (0,365), j โˆˆ (1,500)

๐ป๐‘ ๐‘๐‘ฃ๐‘œ๐‘’ ๐‘ˆ๐‘ ๐‘ฃ๐‘ขโ„Ž

50

4 4 2

๐‘ž๐‘—๐‘˜

๐‘ก&๐‘ž500: | j โˆˆ (1,500) ๐ป๐‘ ๐‘๐‘ฃ๐‘œ๐‘’ ๐‘ˆ๐‘ ๐‘ฃ๐‘ขโ„Ž

Trained an auto encoder Used it to find stock close to the market encoded Used those with deep architecture to find s&p500

Layers Structure Hyper Parameters

AutoEncod er DFP Policy Sparsity indicator: 0.1

50

slide-25
SLIDE 25

Stationarity

Definition

Let ๐‘ฆ๐‘ข be a stochastic process and let ๐บ

๐‘ฆ ๐‘ฆ๐‘ข1+๐œ, . . . . , ๐‘ฆ๐‘ข๐‘™+๐œ represent the cumulative distribution

function of the joint distribution of ๐‘ฆ๐‘ข at times ๐‘ข1 + ๐œ, . . . . . . . , ๐‘ข๐‘™ + ๐œ. ๐’š๐’– is said to be strictly stationary if, for all k, for all ๐œ and for all ๐‘ข1. . . . . . ๐‘ข๐‘™

๐‘ฎ๐’š ๐’š๐’–๐Ÿ+๐Š, . . . . , ๐’š๐’–๐’+๐Š = ๐‘ฎ๐’€ ๐’š๐’–๐Ÿ, . . . . , ๐’š๐’–๐’

In other words

Shifting the time origin by an amount ๐œ has no effect on the joint distribution which depends only on the intervals between ๐‘ข1 . . . . . . . ๐‘ข๐‘™

slide-26
SLIDE 26

Stationarizing A Time Series

V A L U E F O R E C A S T I N G

slide-27
SLIDE 27

"Stationarizing" A Time Series

V A L U E F O R E C A S T I N G

๐‘ง๐‘— = ๐‘ง๐‘๐‘ ๐‘—๐‘•๐‘— ๐‘ง๐‘— = [ln(๐‘ง๐‘๐‘ ๐‘—๐‘•๐‘—)]โ€ฒ

slide-28
SLIDE 28

Algorithm Structure

Learning Model Logarithm & Differentiation

Integration & Exponentiation

V A L U E F O R E C A S T I N G

Segmentation Concatenation

๐‘ง๐‘— = [ln(๐‘ง๐‘๐‘ ๐‘—๐‘•๐‘—)]โ€ฒ ๐‘ง๐‘ž๐‘ ๐‘“๐‘’๐‘— = ๐‘“ืฌ เทœ

๐‘ง๐‘—

Features Prediction

Nonstationary Stochastic Series Stationary Stochastic Series Segments Regression Prediction Stationary Stochastic Prediction Nonstationary Final Stochastic Prediction

slide-29
SLIDE 29

4D Financial Graph Data

Assets

DEEP REINFORCEMENT LEARNING

Time Features

slide-30
SLIDE 30

DEEP LEARNING IN PRODUCTION

slide-31
SLIDE 31

ALGOTRADING PITFALLS

Slippage is the difference between where the computer signaled the entry and exit for a trade and where actual clients, with actual money, entered and exited.

Slippage

Commission is a service charge assessed by a broker

  • r investment advisor in return

for providing investment advice and/or handling the purchase or sale of a security.

Commission

DEEP LEARNING IN FINANCE

slide-32
SLIDE 32

Theoretical Motivations

Smoothness Assumption

Points that are close to each other are more likely to share a label.

  • Allow to easily interpolate between examples.
  • The root for the curse of dimensionality

Depth Assumption

Depth is a double edge sword.

  • Depth can add exponentially (comparing to width) more predictive power to the network but only

if done right.

Distributed Representations

Localist models are very inefficient whenever the data has componential structure

  • Distributed representations are useful for efficient combining of features to learn the underlying

mechanism that the labels are derived from.

Kitchen sink approach

Possible relevant features from many market can be incredibly large.

DEEP LEARNING IN FINANCE

slide-33
SLIDE 33

UNDERFITTING

Finance models often have poor results

Solve this by smart risk management

โ€œIt works if you remove this situationโ€

Save yourself some money, it just doesnโ€™t work.

Feature Engineering VS Raw Data

A good question. DEEP LEARNING IN FINANCE

Data: Unlimited

Image Finance

Overtraining : Hard Data: Limited Overtraining: Easy

$

slide-34
SLIDE 34

โ€œYou are basically risking millions on something no

  • ne in the world fully

understands?โ€

"Why Should I Trust You?": Explaining the Predictions of Any Classifier โ€“ Ribeiro, et al 2015

slide-35
SLIDE 35

DEEP LEARNING IN FINANCE

MARKOV PROPERTY

๐‘„ ๐‘Œ๐‘œ = ๐‘ฆ๐‘œ ๐‘Œ๐‘œโˆ’1 = ๐‘ฆ๐‘œโˆ’1, ๐‘Œ๐‘œโˆ’2 = ๐‘ฆ๐‘œโˆ’2, . . . . , ๐‘Œ0 = ๐‘ฆ0) = ๐‘„(๐‘Œ๐‘œ = ๐‘ฆ๐‘œ|๐‘Œ๐‘œโˆ’๐‘ข = ๐‘ฆ๐‘œโˆ’๐‘ข)

Markov Property

slide-36
SLIDE 36

DEEP LEARNING IN FINANCE

RNN PITFALLS

Example: Jane walked into the room John walked in too. It was late at night. Jane said hi to _____ The difficulty in training recurrent neural nets โ€“ Bengio, et al 2013 Tricks:

  • Init the weight matrices to the identity Matrix
  • Set the activation function to RELU
  • Norm clipping the exploding gradient

เทœ ๐‘• โ†

๐œ–๐œ ๐œ–๐‘ฅ

๐’‹๐’ˆ ๐‘• โ‰ฅ ๐‘ขโ„Ž๐‘ ๐‘“๐‘กโ„Ž๐‘๐‘š๐‘’ ๐’–๐’Š๐’‡๐’ เทœ ๐‘• โ†

๐‘ขโ„Ž๐‘ ๐‘“๐‘กโ„Ž๐‘๐‘š๐‘’ || เทœ ๐‘•||

เทœ ๐‘• A Simple Way to Initialize Recurrent Networks of Rectified Linear Unitsโ€“ Hinton, et al 2015

slide-37
SLIDE 37
slide-38
SLIDE 38

DEEP LEARNING IN FINANCE

DEBUGGING DEEP NETS

Understanding the difficulty in training deep feedforward neural netsโ€“ Bengio, et al 2013

Mean and standard deviation of the activation (output of the sigmoid) during learning, for 4 hidden layers . The top hidden layer quickly saturates at 0 (slowing down all learning), but then slowly desaturates ~ epoch 100.

slide-39
SLIDE 39

Other Uses of Deep Learning

Extracting price impact of news

Stress signal on Individual Asset from text (Word2Vec) and fundamental data

Non-Linear Portfolio Replication

Using Fewer Instruments

Reinforcement Learning

For Continues Algo optimization

Space Embedding

For Correlation between Assets

DEEP LEARNING IN FINANCE

slide-40
SLIDE 40

GOOG AAPL MMM

... ...

Deep Learning Orders

Trading / Training On Site

Research In House

Data Storage GPU CLUSTER Testing Data

Market Market

slide-41
SLIDE 41

Under fitting

Underfitting occurs when a statistical model or machine learning algorithm cannot capture the underlying trend of the data. Intuitively, underfitting

  • ccurs when the model or the algorithm does not fit

the data well enough.

Underftting is either caused by trying to fit a too simple model or by not training the model enough. Calculate 1+1 What is 1? I've

  • nly seen

equations with 4

TRADING USING DEEP LEARNING

slide-42
SLIDE 42

Represented As Tensor Multiplications Computed On NVidia GPUs Model Training Algorithm

USING THE GPU

TRADING USING DEEP LEARNING

X

slide-43
SLIDE 43

THE IMPORTANCE OF POWER

TRADING USING DEEP LEARNING

Performance

Computational Speed allows us to further train our model to yield better accuracy in the finite amount of training time we have.

Memory Capacity

High Memory Capacity is crucial to our systems, allowing us to construct wider and deeper models and integrate more data into the models.

Interface Speed

The Speed of the card interface defines how many model โ€œshiftsโ€ can we perform in one trading cycle.

slide-44
SLIDE 44

CPU vs GPU

Compiled with .nvcc GDDR5 โ€“ 7.0 Gbps Compiled with .gcc DDR4 - 2400Mhz

TRADING USING DEEP LEARNING

Matrix multiplication complexityโ€“ ๐‘ƒ(๐‘œ3) Epoch takes: 7500 sec (avg) Matrix multiplication complexityโ€“ ๐‘ƒ(๐‘œ)

[1]

Epoch takes: 500 sec (avg)

[1] - Understanding the Efficiency of GPU Algorithms for Matrix-Matrix Multiplication - Fatahalian, et al, 2004

slide-45
SLIDE 45

POWER TRAINING ACCURACY RETURNS = = =

slide-46
SLIDE 46
slide-47
SLIDE 47
slide-48
SLIDE 48

Deep Learning Algorithmic Trading Paper Review GPU Based Trading Recap

slide-49
SLIDE 49

Questions?