financial econometrics econ 40357 regression review time
play

Financial Econometrics Econ 40357 Regression review, Time-series - PowerPoint PPT Presentation

Financial Econometrics Econ 40357 Regression review, Time-series regression Some Necessary Matrix Algebra (sorry, cant avoid this) N.C. Mark University of Notre Dame and NBER 2020 1 / 32 Regression review A time series is a sequence of


  1. Financial Econometrics Econ 40357 Regression review, Time-series regression Some Necessary Matrix Algebra (sorry, can’t avoid this) N.C. Mark University of Notre Dame and NBER 2020 1 / 32

  2. Regression review A time series is a sequence of observations over time. Let T be sample size. We write the sequence as { y t } T t = 1 We use notation such as this = E ( y t ) µ y σ 2 = Var ( y t ) y 2 / 32

  3. Regression in population We have in mind a joint distribution between two time series, y t and x t . This is our model. In finance, we are less concerned about exogeneity, instrumental variables, and establishing cause and effect. We are more concerned about understanding reduced form correlations. Understanding the statistical dependence across time series, and dependence of observations across time. The cross-moments of the joint distribution. 3 / 32

  4. Regression in population Write the population regression as E ( y t | x t ) � �� � y t = α + β x t + ǫ t (1) The systematic part of regression is also called the projection . ǫ t is projection error . Assume error is iid but not necessarily normal (what does this mean?). � � 0 , σ 2 ǫ t is i.i.d. ǫ Think of fitted part of regression as conditional expectation . Conditional expectation is the best predictor. Prediction means the same thing as forecast We use regression for things like computing betas , which measure exposure of an asset to risk factors . 4 / 32

  5. Regression in population Take expectation of y t = α + β x t + ǫ t 1 E ( y t ) = α + β E ( x t ) + E ( ǫ t ) Using the short-hand notation, µ y = α + βµ x rearrange to get, α = µ y − βµ x y t ≡ y t − µ y and ˜ x t ≡ x t − µ x . The ‘tilde’ represents the Define ˜ 2 variable expressed as a deviation from its mean. Substitute this expression for α back into the regression (eq. 1). Doing so gives the regression in deviations from mean form. y t = β ˜ ˜ x t + ǫ t (2) 5 / 32

  6. Regression in population Multiply both sides of eq.(2) by ˜ x t , then take expectations on both sides, y t ˜ ˜ x t = β ˜ x t ˜ x t + ǫ t ˜ x t E ( ˜ x t ) = β E ( ˜ x t ) + E ( ǫ t ˜ x t ) y t ˜ x t ˜ Solve for β β = E ( ˜ x t ) 2 = Cov ( y t , x t ) x t ) y t ˜ σ y , x = σ y , x σ y σ y = = (3) ρ y , x E ( ˜ Var ( x t ) σ 2 σ y σ x σ x σ x x � �� � � �� � ���� � �� � Interpretation Interpretation Notation Algebra 6 / 32

  7. Regression in population Take conditional expectation on both sides of original regression conditional on x t . E ( y t | x t ) = α + β x t There’s a theorem that says the conditional expectation is the best linear predictor of y t conditional on x t . The best! Since regression is conditional expectation, this motivates using it as the forecast function. 7 / 32

  8. Estimation of α and β by least squares Eviews will do this work for you. Least squares estimates are the sample counterparts to the population parameters. What does this mean? In deviations from the mean form, solution to least squares problem is β = ∑ T T ∑ T 1 x t ˜ y t t = 1 ˜ x t ˜ y t t = 1 ˜ ˆ = ∑ T T ∑ T x 2 1 x 2 t = 1 ˜ t = 1 ˜ t t y t = ˆ ˜ β ˜ x t + ˆ ǫ t ǫ t is residual , not error. ˆ β is a random variable. To see this, make ˆ substitutions, β = ∑ ˜ x t ( β ˜ x t + ǫ t ) = β + ∑ ˜ x t ǫ t ˆ x 2 x 2 ∑ ˜ ∑ ˜ t t β is a linear combination of ǫ ′ ˆ t s , which are random variables. Therefore ˆ β is a random variable, with a distribution. 8 / 32

  9. Inference in Time Series Regression is a bit Different What is statistical inference? We want the sample standard deviation of ˆ β . It is called the standard error of ˆ β , because ˆ β is a statistic. What is a statistic? ˆ β divided by standard error is the t-ratio . We are mainly interested in t-ratios. Not interested in F-statistics. Nobody’s opinion ever was changed by having a significant F-statistic and insignifcant t-ratios. What is statistical significance? In your previous econometrics class, you assumed x t is exogenous . This alllows you to treat them as constants, and therefore, randomness in ˆ β is induced only by ǫ t . In time-series, we can’t do that (If y t is Amazon returns and x t is the market return, how can we say the market is exogenous? or if x t = y t − 1 , how can we treat x t as constant? No way!). 9 / 32

  10. Inference in Time Series Regression Let’s pretend the ˜ x t are exogenous. Even more inappropriately, let’s pretend they are non-stochastic constants. � � ∑ T T � � t = 1 ˜ x t ǫ t 1 x 2 1 σ 2 x 2 2 σ 2 x 2 T σ 2 ∑ Var = ǫ + ˜ ǫ + · · · ˜ ˜ ǫ � � 2 ∑ T x 2 t = 1 ˜ ∑ T x 2 t = 1 ˜ t = 1 t t x 2 σ 2 ∑ ˜ t � 2 σ 2 ǫ = ǫ = � x 2 ∑ ˜ ∑ T x 2 t = 1 ˜ t t The standard deviation of the term is � � ∑ T t = 1 ˜ x t ǫ t σ ǫ = sd � ∑ T x 2 t = 1 ˜ x 2 ∑ ˜ t t The standard error of the term, and hence of ˆ β is � = � ˆ σ ǫ ˆ se β � x 2 ∑ ˜ t 10 / 32

  11. where we estimate ˆ σ ǫ with the sample standard deviation of the regression residuals ˆ ǫ t . This particular formula is true only when the errors are iid , and for large (infinite) sample sizes. Why do we use? Because we can find the answer for large samples, and we hope that it is a good approximation to the exact true (but unknown) distribution 11 / 32

  12. Inference in Time Series Regression So time-series econometricians do a thing called asymptotic theory. They ask how the numerator and denominator, T numerator: 1 � � 0 , σ 2 ∑ x t ǫ t → N ˜ ǫ Q T t = 1 T denominator: 1 x 2 ∑ ˜ t → Q T t = 1 behave as T → ∞ . It is very complicated business involves a lot of high-level math. Fortunately, at the end of the day, what comes out of all this is the same thing we learned in your first econometrics class! 12 / 32

  13. Least squares estimation of α and β That is, we pretend we have an infinite sample size ( T = ∞ , ) in which case, t = ( ˆ β − β ) � ∼ N ( 0 , 1 ) � ˆ s . e . β � = � ˆ σ ǫ ˆ se β (4) � x 2 ∑ ˜ t ǫ = 1 σ 2 ǫ 2 T ∑ ˆ (5) ˆ t The difference is we don’t consult the t-table or worry about degrees of freedom. We consult the standard normal table. The strategy: The exact t-distribution is unknown (why?), so we use the asymptotic distribution (why?) and hope it is a good approximation to the unknown distribution. Finally, we are also interested in R 2 , the measure of goodness of fit. y 2 ǫ 2 SST = ∑ ˜ R 2 = SSR = 1 − ∑ ˆ = 1 − SSE t t x 2 x 2 SST ∑ ˜ ∑ ˜ t t 13 / 32

  14. Story behind the t-test In the 1890s, William Gosset was studying chemical properties of barley with small samples, for the Guinness company (yes, that Guinness). He showed his results to the great statistician Karl Pearson at University College London, who mentored him. Gossett published his work in the journal Biometrika , using the pseudonym Student, because he would have gotten in trouble at Guinness if he used his real name. 14 / 32

  15. t-test review: two sided test 15 / 32

  16. t-test review: one sided test 16 / 32

  17. Some matrix algebra Scalar : a single number   a 11 a 12  is a ( 3 × 2 ) Matrix: a two-dimensional array. A = a 21 a 22  a 31 a 23 matrix–that is, 3 rows and 2 columns. We say the number of rows then columns. a 11 is the (1,1) element of A , and is a scalar. The subscripts of the elements tell us which row and column they are from. Vector : a one-dimensional array. If we take the first column of A   a 11 and call it A 1 =  , it is a ( 3 × 1 ) column vector . If we a 21  a 31 � � take the second row of A and call it A 2 = a 21 a 22 , it is a ( 1 × 2 ) row vector . 17 / 32

  18. Square matrix: An m × n matrix is square if m = n .   a 11 a 12 a 13  is a square matrix. The diagonal of the A = a 21 a 22 a 23  a 31 a 32 a 33 matrix A are the elements a 11 , a 22 , a 33 . It only makes sense to talk about the diagonal of square matrices. Symmetric matrix . For a square matrix, if the elements a ij = a ji , for i � = j , then the matrix is symmetric. (notice the correspondence of the bold entries).   2 3 4 A = 3 10 6   4 6 11 18 / 32

  19. Transpose of a matrix. The i − th row becomes the i − th column. The transpose of an ( m × n ) matrix is ( n × × m ) .   a 11 � � , then A ′ = A T = A = a 11 a 12 a 13 a 12   . a 13   a 11  , then A ′ = A T = � � A = a 12 a 11 a 12 a 13  . a 13   � a 11 a 11 a 12 � a 21 a 31  , then A ′ = A T = A = a 21 a 22  a 12 a 22 a 32 a 31 a 32 19 / 32

  20.   0 0  is a zero matrix. Zero matrix: All the entries are 0 . A = 0 0  0 0 Identity matrix: A square matrix with 1s on the diagonal elements and 0 on the off-diagonal elements is called the identity matrix.   1 0 0  is a ( 3 × 1 ) identity matrix. We always call an I = 0 1 0  0 0 1 identity matrix I . 20 / 32

Download Presentation
Download Policy: The content available on the website is offered to you 'AS IS' for your personal information and use only. It cannot be commercialized, licensed, or distributed on other websites without prior consent from the author. To download a presentation, simply click this link. If you encounter any difficulties during the download process, it's possible that the publisher has removed the file from their server.

Recommend


More recommend