Latent and Network Models with Applications to Finance Jingchen Liu - - PowerPoint PPT Presentation

latent and network models with applications to finance
SMART_READER_LITE
LIVE PREVIEW

Latent and Network Models with Applications to Finance Jingchen Liu - - PowerPoint PPT Presentation

Latent and Network Models with Applications to Finance Jingchen Liu Department of Statistics Columbia University Joint work with Yunxiao Chen, Xiaoou Li, and Zhiliang Ying At ISFA-Columbia Workshop, June 28, 2016 November 16, 2015 1 / 44


slide-1
SLIDE 1

Latent and Network Models with Applications to Finance

Jingchen Liu

Department of Statistics Columbia University Joint work with Yunxiao Chen, Xiaoou Li, and Zhiliang Ying At ISFA-Columbia Workshop, June 28, 2016

November 16, 2015

1 / 44

slide-2
SLIDE 2

Modeling multivariate distribution

◮ Multivariate random vector: (R1, ..., RJ) ◮ Continuous vectors: multivariate Gaussian, multivariate t-distribution... ◮ Categorical vectors: loglinear model... ◮ Copula ◮ Regression

2 / 44

slide-3
SLIDE 3

Modeling multivariate distribution

◮ Multivariate random vector: (R1, ..., RJ) ◮ Continuous vectors: multivariate Gaussian, multivariate t-distribution... ◮ Categorical vectors: loglinear model... ◮ Copula ◮ Regression

2 / 44

slide-4
SLIDE 4

Modeling multivariate distribution

◮ Multivariate random vector: (R1, ..., RJ) ◮ Continuous vectors: multivariate Gaussian, multivariate t-distribution... ◮ Categorical vectors: loglinear model... ◮ Copula ◮ Regression

2 / 44

slide-5
SLIDE 5

Modeling multivariate distribution

◮ Multivariate random vector: (R1, ..., RJ) ◮ Continuous vectors: multivariate Gaussian, multivariate t-distribution... ◮ Categorical vectors: loglinear model... ◮ Copula ◮ Regression

2 / 44

slide-6
SLIDE 6

Modeling multivariate distribution

◮ Multivariate random vector: (R1, ..., RJ) ◮ Continuous vectors: multivariate Gaussian, multivariate t-distribution... ◮ Categorical vectors: loglinear model... ◮ Copula ◮ Regression

2 / 44

slide-7
SLIDE 7

Latent variable modeling

◮ There exists α such that f (R1, ..., RJ|α) is simple. ◮ What is considered as simple? ◮ Independence, small variance,...

3 / 44

slide-8
SLIDE 8

Latent variable modeling

◮ There exists α such that f (R1, ..., RJ|α) is simple. ◮ What is considered as simple? ◮ Independence, small variance,...

3 / 44

slide-9
SLIDE 9

Latent variable modeling

◮ There exists α such that f (R1, ..., RJ|α) is simple. ◮ What is considered as simple? ◮ Independence, small variance,...

3 / 44

slide-10
SLIDE 10

Graphical representation

4 / 44

slide-11
SLIDE 11

Local independence

f (R1, ..., RJ|α) =

  • j

f (Rj|α)

5 / 44

slide-12
SLIDE 12

Applications

◮ Finance, political sciences ◮ Education ◮ Psychiatry/psychology ◮ Marketing and e-commerce

6 / 44

slide-13
SLIDE 13

Applications

◮ Finance, political sciences ◮ Education ◮ Psychiatry/psychology ◮ Marketing and e-commerce

6 / 44

slide-14
SLIDE 14

Applications

◮ Finance, political sciences ◮ Education ◮ Psychiatry/psychology ◮ Marketing and e-commerce

6 / 44

slide-15
SLIDE 15

Applications

◮ Finance, political sciences ◮ Education ◮ Psychiatry/psychology ◮ Marketing and e-commerce

6 / 44

slide-16
SLIDE 16

Linear factor models

◮ (R1, ..., RJ) is continous. ◮ Linear factor models: α = (α1, ..., αK)

Rj = a⊤

j α + εj

◮ Principle component analysis

7 / 44

slide-17
SLIDE 17

Linear factor models

◮ (R1, ..., RJ) is continous. ◮ Linear factor models: α = (α1, ..., αK)

Rj = a⊤

j α + εj

◮ Principle component analysis

7 / 44

slide-18
SLIDE 18

Linear factor models

◮ (R1, ..., RJ) is continous. ◮ Linear factor models: α = (α1, ..., αK)

Rj = a⊤

j α + εj

◮ Principle component analysis

7 / 44

slide-19
SLIDE 19

Categorical variable and item response theory model

◮ Binary Ri ∈ {0, 1}. ◮ P(Rj = 1|α) =

ea⊤

j α−bj

1+e

a⊤ j α−bj ,

α ∈ RK

  • 4
  • 2

2 4 0.0 0.2 0.4 0.6 0.8 1.0 x y

8 / 44

slide-20
SLIDE 20

Categorical variable and item response theory model

◮ Binary Ri ∈ {0, 1}. ◮ P(Rj = 1|α) =

ea⊤

j α−bj

1+e

a⊤ j α−bj ,

α ∈ RK

  • 4
  • 2

2 4 0.0 0.2 0.4 0.6 0.8 1.0 x y

8 / 44

slide-21
SLIDE 21

Stock Price Structure

◮ Data1: 97 stocks selected from S&P100 in 1013 trading days

from 2009 to 2014.

◮ Data2: 117 stocks selected from SSE180 (Shanghai Stock

Exchange) in 1159 trading days from 2009 to 2014.

9 / 44

slide-22
SLIDE 22

Stock Price Structure

◮ Data1: 97 stocks selected from S&P100 in 1013 trading days

from 2009 to 2014.

◮ Data2: 117 stocks selected from SSE180 (Shanghai Stock

Exchange) in 1159 trading days from 2009 to 2014.

9 / 44

slide-23
SLIDE 23

Exploratory Analysis

The heatmap of stock-stock cor- relation (Data 1; based on daily log return); stocks have been re-

  • rdered

e.g.

◮ The block circled by blue contains

mostly the energy companies:

APA (Apache Corp), APC (Anadarko Petroleum), BHI (Baker Hughes), COP (Conoco Phillips), CVX (Chevron), DVN (Devon), ... ◮ The block circled by black contains

the financial companies:

C (citi), BAC (BOA), MS (Morgan Stanley), BK(Bank of New York Mellon), JPM (JP Morgan), ...

10 / 44

slide-24
SLIDE 24

Linear factor model

◮ Linear factor models

Rj = a⊤

j α + εj ◮ Fama-French model:

R = Rf + β(K − Rf ) + bsSMB + bvHML + α

11 / 44

slide-25
SLIDE 25

Linear factor model

◮ Linear factor models

Rj = a⊤

j α + εj ◮ Fama-French model:

R = Rf + β(K − Rf ) + bsSMB + bvHML + α

11 / 44

slide-26
SLIDE 26

Linear factor model

◮ (R1, ..., RJ) is not multivariate Gaussian

in many ways if J is large!

◮ Marginal tail, joint tail, asymmetric correlation... ◮ Too many factors!

12 / 44

slide-27
SLIDE 27

Linear factor model

◮ (R1, ..., RJ) is not multivariate Gaussian

in many ways if J is large!

◮ Marginal tail, joint tail, asymmetric correlation... ◮ Too many factors!

12 / 44

slide-28
SLIDE 28

Linear factor model

◮ (R1, ..., RJ) is not multivariate Gaussian

in many ways if J is large!

◮ Marginal tail, joint tail, asymmetric correlation... ◮ Too many factors!

12 / 44

slide-29
SLIDE 29

Nonlinear factor model

◮ Dichotomize Rji = 1 if Sclose i

> Sopen

i

for stock j on day i

◮ P(Rj = 1|α) = ea⊤

j α−bj

1+e

a⊤ j α−bj ,

α ∈ RK

13 / 44

slide-30
SLIDE 30

Nonlinear factor model

◮ Dichotomize Rji = 1 if Sclose i

> Sopen

i

for stock j on day i

◮ P(Rj = 1|α) = ea⊤

j α−bj

1+e

a⊤ j α−bj ,

α ∈ RK

13 / 44

slide-31
SLIDE 31

Latent graphical model

14 / 44

slide-32
SLIDE 32

Latent graphical model

15 / 44

slide-33
SLIDE 33

Latent graphical model

16 / 44

slide-34
SLIDE 34

Issues to concern

◮ Parametric/nonparametric models: latent variable and graph ◮ Inference: identifiability

17 / 44

slide-35
SLIDE 35

Issues to concern

◮ Parametric/nonparametric models: latent variable and graph ◮ Inference: identifiability

17 / 44

slide-36
SLIDE 36

The latent variable component – IRT model

◮ Alternative formulation:

P(Rj = 1|α) = ea⊤

j α−bj

1 + ea⊤

j α−bj

⇔ P(Rj|α) ∝ eRj (a⊤

j α−bj )

◮ Local independence

P(R1, ..., RJ|α) =

J

  • j=1

P(Rj|α) ∝ e

  • j Rj (a⊤

j α−bj )

18 / 44

slide-37
SLIDE 37

The latent variable component – IRT model

◮ Alternative formulation:

P(Rj = 1|α) = ea⊤

j α−bj

1 + ea⊤

j α−bj

⇔ P(Rj|α) ∝ eRj (a⊤

j α−bj )

◮ Local independence

P(R1, ..., RJ|α) =

J

  • j=1

P(Rj|α) ∝ e

  • j Rj (a⊤

j α−bj )

18 / 44

slide-38
SLIDE 38

Graphical component component – Ising model

P(R1, ..., RJ|S) ∝ e

1 2

  • i,j sij Ri Rj

◮ Physics ◮ Graphical representation

19 / 44

slide-39
SLIDE 39

Graphical component component – Ising model

P(R1, ..., RJ|S) ∝ e

1 2

  • i,j sij Ri Rj

◮ Physics ◮ Graphical representation

19 / 44

slide-40
SLIDE 40

Latent graphical model: IRT model + Ising model

◮ Nonlocal independence

P(R1, ..., RJ|α) ∝ e

  • j Rj (a⊤

j α−bj )+ 1 2

  • i,j sij Ri Rj

◮ Simplification: R2

j = Rj

P(R1, ..., RJ|α) ∝ e

  • j Rj a⊤

j α+ 1 2

  • i,j sij Ri Rj

20 / 44

slide-41
SLIDE 41

Latent graphical model: IRT model + Ising model

◮ Nonlocal independence

P(R1, ..., RJ|α) ∝ e

  • j Rj (a⊤

j α−bj )+ 1 2

  • i,j sij Ri Rj

◮ Simplification: R2

j = Rj

P(R1, ..., RJ|α) ∝ e

  • j Rj a⊤

j α+ 1 2

  • i,j sij Ri Rj

20 / 44

slide-42
SLIDE 42

Latent variable and network modeling

◮ The item response function

fA,S(R|α) ∝ exp{α⊤AR + 1 2R⊤SR} where AK×J = (a1, ..., aJ) and SJ×J = (sij)

◮ Population (prior) distribution such that

fA,S(R, α) ∝ exp{−|α|2/2 + α⊤AR + 1 2R⊤SR}

21 / 44

slide-43
SLIDE 43

Latent variable and network modeling

◮ The item response function

fA,S(R|α) ∝ exp{α⊤AR + 1 2R⊤SR} where AK×J = (a1, ..., aJ) and SJ×J = (sij)

◮ Population (prior) distribution such that

fA,S(R, α) ∝ exp{−|α|2/2 + α⊤AR + 1 2R⊤SR}

21 / 44

slide-44
SLIDE 44

Latent variable and network modeling

◮ Marginalized likelihood

L(A, S) =

  • f (R, α)dα ∝ exp{1

2R⊤(A⊤A + S)R}

◮ Let LJ×J = A⊤A

L(L, S) = f (R|L, S) ∝ exp{1 2R⊤(L + S)R}

22 / 44

slide-45
SLIDE 45

Latent variable and network modeling

◮ Marginalized likelihood

L(A, S) =

  • f (R, α)dα ∝ exp{1

2R⊤(A⊤A + S)R}

◮ Let LJ×J = A⊤A

L(L, S) = f (R|L, S) ∝ exp{1 2R⊤(L + S)R}

22 / 44

slide-46
SLIDE 46

Identifiability

◮ Identifiability of L and S ◮ Low dimension latent factor :

LJ×J = A⊤A is positive semi-definite of rank K ≪ J

◮ Small remaining dependence

S is sparse

23 / 44

slide-47
SLIDE 47

Identifiability

◮ Identifiability of L and S ◮ Low dimension latent factor :

LJ×J = A⊤A is positive semi-definite of rank K ≪ J

◮ Small remaining dependence

S is sparse

23 / 44

slide-48
SLIDE 48

Identifiability

◮ Identifiability of L and S ◮ Low dimension latent factor :

LJ×J = A⊤A is positive semi-definite of rank K ≪ J

◮ Small remaining dependence

S is sparse

23 / 44

slide-49
SLIDE 49

Regularized estimation

◮ Sparse network:

O(S)1 =

  • i=j

|sij|

◮ Rank of L = T ⊤ΛT = number of nonzero eigenvalues

L∗

J

  • i=1

|λi| =

J

  • i=1

λi = Trace(L)

24 / 44

slide-50
SLIDE 50

Regularized estimation

◮ Sparse network:

O(S)1 =

  • i=j

|sij|

◮ Rank of L = T ⊤ΛT = number of nonzero eigenvalues

L∗

J

  • i=1

|λi| =

J

  • i=1

λi = Trace(L)

24 / 44

slide-51
SLIDE 51

Regularized pseudo-likelihood estimation

max

L,S {log L(L, S) − γO(S)1 − δL∗}

25 / 44

slide-52
SLIDE 52

S&P 100

◮ ˆ

K = 1: a unidimensional market factor is preferred (general market movement).

◮ Cor(ˆ

θ, rs&p) = 0.77.

26 / 44

slide-53
SLIDE 53

S&P 100

◮ ˆ

K = 1: a unidimensional market factor is preferred (general market movement).

◮ Cor(ˆ

θ, rs&p) = 0.77.

26 / 44

slide-54
SLIDE 54

S&P 100

27 / 44

slide-55
SLIDE 55

S&P 100

Figure: The estimated graph has 462 edges.

28 / 44

slide-56
SLIDE 56

Structure Learning of S&P Stock Movement

Most positive stock-pairs:

Stock 1 Stock 2 ˆ sij cor MA (Mastercard) V (Visa) 1.90 0.53 NSC (Norfolk Southern) UNP (Union Pacific) 1.90 0.60 HD (Home Depot) LOW (Lowe’s Cos) 1.81 0.51 AEP (American Electric Power) SO (Southern) 1.54 0.48 FDX (FedEx) UPS (United Parcel Service) 1.47 0.52 CVX (Chevron) XOM (Exxon Mobil) 1.44 0.59 BAC (Bank of America) C (Citigroup) 1.40 0.56 T (AT&T) VZ (Verizon) 1.40 0.45 INTC (Intel) TXN (Texas Instruments) 1.36 0.48 LMT (Lockheed Martin) RTN (Raytheon) 1.32 0.49 BHI (Baker Hughes) HAL (Halliburton) 1.31 0.58 HAL (Halliburton) SLB (Schlumberger) 1.12 0.57 JPM (JP Morgan) WFC (Wells Fargo) 1.09 0.57 KO (Coca-Cola) PEP (Pepsi) 1.03 0.39 USB (US Bancorp) WFC (Wells Fargo) 1.02 0.54 AXP (American Express) COF (Capital One) 0.96 0.44

29 / 44

slide-57
SLIDE 57

Structure Learning of S&P Stock Movement

Most negative stock-pairs:

Stock 1 Stock 2 ˆ sij cor DOW (Dow Chemical) LLY (Eli Lilly)

  • 0.73

0.13 AMZN (Amazon) BAC (Bank of America)

  • 0.69

0.17 GILD (Gilead Sciences) PEP (Pepsi)

  • 0.61

0.10 DVN (Devon Energy) SBUX (Starbucks)

  • 0.60

0.20 GILD (Gilead Sciences) SPG (Simon Property Group)

  • 0.55

0.10 CAT (Caterpillar) PEP (Pepsi)

  • 0.53

0.13 F (Ford Motor) PFE (Pfizer)

  • 0.53

0.17 LLY (Eli Lilly) WFC (Wells Fargo)

  • 0.53

0.13

30 / 44

slide-58
SLIDE 58

Structure Learning of S&P Stock Movement

Possible explanations: different reactions to the business cycle.

◮ Eli Lilly VS Dow Chemical: Health care stocks are viewed as more

defensive investments compared with more cyclically sensitive stocks (such as Dow Chemical), since consumers are less likely to cut back

  • n healthcare expenses during times of economic stress.

◮ BOA VS Amazon: Financial stocks tend to do well at the beginning

  • f the business cycle, and may weaken as the economy heads into
  • recession. However, Amazon benefited from the general shift to
  • nline commerce and the careful shopping behavior that consumers

were exhibiting during the downturn (Bloomberg, 2009).

31 / 44

slide-59
SLIDE 59

Structure Learning of S&P Stock Movement

Maximal graph cliques (stock clusters): 10 6-stock cliques, 13 5-stock cliques, 64 4-stock cliques, and 94 3-stock cliques. e.g.

Stocks 1 HAL (Halliburton), BHI (Baker Hughes), Energy NOV (National Oilwell Varco), SLB (Schlumberger), DVN (Devon), APC (Anadarko Petroleum) 2 BAC (Bank of America), JPM (JP Morgan), Finance WFC (Wells Fargo), MS (Morgan Stanley), C (Citigroup) 3 BA (Boeing), GD (General Dynamics), Industrials LMT (Lockheed Martin), HON (Honeywell), RTN (Raytheon) 4 EBAY (eBay), EMC (EMC), TXN (Texas Instruments), IT (although AMZN is listed QCOM (QUALCOMM), AMZN (Amazon)

in Consumer Discretionary)

5 CAT (Caterpillar), BHI (Baker Hughes), SLB (Schlumberger), Industrial NOV (National Oilwell Varco) & energy

32 / 44

slide-60
SLIDE 60

Structure Learning of Chinese Stock Movement

Results:

◮ ˆ

K = 1 and the estimated graph has 208 edges.

33 / 44

slide-61
SLIDE 61

Structure Learning of Chinese Stock Movement

Most positive pairs:

34 / 44

slide-62
SLIDE 62

Structure Learning of Chinese Stock Movement

◮ There are stock pairs with moderate correlation but high

residual association according to the graphical model (marked by yellow).

◮ Interesting patterns may be discovered: e.g., there seems to

be nothing in common between “Zhongxin Zhengquan” and “Liaoning Chengda” (marked by green). However, “Zhongxin Zhengquan” kept to be one of the top ten shareholders of “Liaoning Chengda” during the period being considered.

35 / 44

slide-63
SLIDE 63

Structure Learning of Chinese Stock Movement

Maximal graph cliques (stock clusters): 5 6-stock cliques, 6 5-stock cliques, 18 4-stock cliques, and 23-stock cliques. e.g.

36 / 44

slide-64
SLIDE 64

Senate voting

◮ The US senate voting record from 108th congress (2003)

consists of 100 senators voting to 459 bills.

◮ Major factor: party membership ◮ Graph: personal relationship

37 / 44

slide-65
SLIDE 65

Main factor

Figure: 0: Democratic, 1: Republican

38 / 44

slide-66
SLIDE 66

The graph

39 / 44

slide-67
SLIDE 67

Senate voting

◮ Positive links: same state or close relationship ◮ Example: Edwards.D.NC. & Kerry.D.MA.

John Edwards became the 2004 Democratic candidate for vice president, the running mate of presidential nominee Senator John Kerry of Massachusetts.

◮ Example2: Baucus.D.MT. & Breaux.D.LA.

Both of them supported the Bush’s administration’s effort to enact a Medicare drug benefit.

40 / 44

slide-68
SLIDE 68

Senate voting

◮ Positive links: same state or close relationship ◮ Example: Edwards.D.NC. & Kerry.D.MA.

John Edwards became the 2004 Democratic candidate for vice president, the running mate of presidential nominee Senator John Kerry of Massachusetts.

◮ Example2: Baucus.D.MT. & Breaux.D.LA.

Both of them supported the Bush’s administration’s effort to enact a Medicare drug benefit.

40 / 44

slide-69
SLIDE 69

Senate voting

◮ Positive links: same state or close relationship ◮ Example: Edwards.D.NC. & Kerry.D.MA.

John Edwards became the 2004 Democratic candidate for vice president, the running mate of presidential nominee Senator John Kerry of Massachusetts.

◮ Example2: Baucus.D.MT. & Breaux.D.LA.

Both of them supported the Bush’s administration’s effort to enact a Medicare drug benefit.

40 / 44

slide-70
SLIDE 70

Senate voting

◮ Positive link while negative correlation. ◮ Example: Lieberman.D.CT. & Stevens.R.AK.

Here is what Lieberman said about Stevens’s death (2010). “... I have lost a dear friend. I am deeply saddened by Ted’s death. I knew him for many years as a valued friend, a neighbor and a colleague. We shared many great experiences and I am grateful for all of the wisdom he offered me personally ?”

41 / 44

slide-71
SLIDE 71

Summary

◮ Latent variable models ◮ Conditional graphical model

42 / 44