for matrix a p p with real eigenvalues define f a the
play

For matrix A ( p p ) with real eigenvalues, define F A , the - PDF document

For matrix A ( p p ) with real eigenvalues, define F A , the empirical distribution function of the eigenvalues of A , to be F A ( x ) (1 /p ) (number of eigenvalues of A x ) . For and p.d.f. G the Stieltjes transform of G is defined


  1. For matrix A ( p × p ) with real eigenvalues, define F A , the empirical distribution function of the eigenvalues of A , to be F A ( x ) ≡ (1 /p ) · (number of eigenvalues of A ≤ x ) . For and p.d.f. G the Stieltjes transform of G is defined as 1 � z ∈ C + ≡ { z ∈ C : ℑ z > 0 } . m G ( z ) ≡ λ − z dG ( λ ) , Inversion formula � b G { [ a, b ] } = (1 /π ) lim ℑ m G ( ξ + iη ) dξ η → 0 + a ( a, b continuity points of G). Notice m F A ( z ) = (1 /p ) tr ( A − zI ) − 1 . 1

  2. Theorem [S. (1995)]. Assume a) For n = 1 , 2 , . . . X n = ( X n ij ), n × N , X n ij ∈ C , i.d. for all n, i, j, 1 1 | 2 = 1. independent across i, j for each n , E | X 1 1 1 − E X 1 b) N = N ( n ) with n/N → c > 0 as n → ∞ . c) T n n × n random Hermitian nonnegative definite, with F T n con- verging almost surely in distribution to a p.d.f. H on [0 , ∞ ) as n → ∞ . d) X n and T n are independent. Let T 1 / 2 be the Hermitian nonnegative square root of T n , and n let B n = (1 /N ) T 1 / 2 n T 1 / 2 (obviously F B n = F (1 /N ) X n X ∗ n T n ). X n X ∗ n n Then, almost surely, F B n converges in distribution, as n → ∞ , to a (nonrandom) p.d.f. F , whose Stieltjes transform m ( z ) ( z ∈ C + ) satisfies � 1 ( ∗ ) m = t (1 − c − czm ) − z dH ( t ) , in the sense that, for each z ∈ C + , m = m ( z ) is the unique solution to ( ∗ ) in { m ∈ C : − 1 − c + cm ∈ C + } . z 2

  3. We have F (1 /N ) X ∗ T X = (1 − n N ) I [0 , ∞ ) + n N F (1 /N ) XX ∗ T a.s. − → (1 − c ) I [0 , ∞ ) + cF ≡ F. Notice m F and m F satisfy 1 − c + 1 � 1 c m F ( z ) = m F ( z ) = − zm F t − z dH ( t ) . cz Therefore, m = m F solves z = − 1 t � m + c 1 + tmdH ( t ) . 3

  4. Facts on F: 1. The endpoints of the connected components (away from 0) of the support of F are given by the extrema of � f ( m ) = − 1 t m + c 1 + tmdH ( t ) m ∈ C [Marˇ cenko and Pastur (1967), S. and Choi (1995)]. 2. F has a continuous density away from the origin given by 1 cπ ℑ m ( x ) 0 < x ∈ support of F where m ( x ) = z ∈ C + → x m F ( z ) lim solves � x = − 1 t m + c 1 + tmdH ( t ) . (S. and Choi 1995). 3. F ′ is analytic inside its support, and when H is discrete, has infinite slopes at boundaries of its support [S. and Choi (1995)]. 4. c and F uniquely determine H . D a.s. 5. F − → H as c → 0 (complements B n − → T n as N → ∞ , n fixed). 4

  5. 0.5 0.4 0.3 0,2 0 . 1 0.0; f '1/3 ,'

  6. o I P I I I o + (j UI N O r o [ " " N I J + ( d = ll I \ ) . t s o + o O

  7. I O O r r) tl . N F \ A L i l tl N ' t s o F o o

  8. T n = I n = ⇒ F = F c , where, for 0 < c ≤ 1, F ′ c ( x ) = f c ( x ) = 1 � ( x − b 1 )( b 2 − x ) b 1 < x < b 2 , 2 πcx 0 otherwise, where b 1 = (1 − √ c ) 2 and b 2 = (1 + √ c ) 2 , and for 1 < c < ∞ , � x F c ( x ) = (1 − (1 /c )) I [0 , ∞ ) ( x ) + f c ( t ) dt. b 1 Marˇ cenko and Pastur (1967) Grenander and S. (1977) 8

  9. Let, for any d > 0 and d.f. G , F d,G denote the limiting spectral d.f. of (1 /N ) X ∗ TX corresponding to limiting ratio d and limiting F T n G . Theorem [Bai and S. (1998)]. Assume: a) X ij , i, j = 1 , 2 , ... are i.i.d. random variables in C with E X 11 = 0, E | X 11 | 2 = 1, and E | X 11 | 4 < ∞ . b) N = N ( n ) with c n = n/N → c > 0 as n → ∞ . c) For each n T n is an n × n Hermitian nonnegative definite satisfying D H n ≡ F T n − → H , a p.d.f. d) � T n � , the spectral norm of T n is bounded in n . e) B n = (1 /N ) T 1 / 2 n T 1 / 2 , T 1 / 2 X n X ∗ any Hermitian square root of n n n T n , B n = (1 /N ) X ∗ n T n X n , where X n = ( X ij ), i = 1 , 2 , . . . , n , j = 1 , 2 , . . . , N . f) The interval [ a, b ] with a > 0 lies in an open interval outside the support of F c n ,H n for all large n . Then P ( no eigenvalue of B n appears in [ a, b ] for all large n ) = 1 . 9

  10. Theorem [Bai and S. (1999)]. Assume (a)–(f) of the previous the- orem. 1) If c [1 − H (0)] > 1, then x 0 , the smallest value in the support of F c,H , is positive, and with probability one λ B n → x 0 as n → ∞ . N The number x 0 is the maximum value of the function z ( m ) = − 1 � t m + c 1 + tmdH ( t ) for m ∈ R + . 2) If c [1 − H (0)] ≤ 1, or c [1 − H (0)] > 1 but [ a, b ] is not contained in [0 , x 0 ] then m F c,H ( b ) < 0. Let for large n integer i n ≥ 0 be such that λ T n λ T n i n > − 1 /m F c,H ( b ) and i n +1 < − 1 /m F c,H ( a ) (eigenvalues arranged in non-increasing order). Then P ( λ B n λ B n i n > b i n +1 < a for all large n ) = 1 . and 10

  11. Theorem (Baik and S. (2006)) Assume the conditions in Bai and S. (1998). Suppose a fixed number of the eigenvalues of T n are different than 1, positive, and remain the same value for all n large. n largest eigenvalue, λ T n Assume the i th i n , is one of these numbers, say α . Then with probability one  if | α − 1 | > √ c cα α +   α − 1 (1 + √ c ) 2 if 1 < α ≤ 1 + √ c λ B n i n →  (1 − √ c ) 2 if 1 − √ c ≤ α < 1 .  11

  12. Theorem (Baik and S.). Assume when X 1 1 is complex, E X 2 1 1 = 0. Suppose α > 0 is an eigenvalue of T n with multiplicity k satisfying | α − 1 | > √ c ( ∗∗ ) and eigenspace formed from k elements of the canonical basis set B n = { (1 , 0 , . . . , 0) ∗ , (0 , 1 , . . . , 0) ∗ , . . . , (0 , . . . , 1) ∗ } in R n . Let λ i , i = 1 , . . . , k be the eigenvalues of B n corresponding to α . Let λ n = α + n α α − 1 . N Then √ n ( λ i − λ n ) i = 1 , . . . , k converge weakly to the eigenvalues of a mean zero Gaussian k × k Hermitian matrix containing inde- pendent random variables on and above the diagonal. The diagonal entries have variance = ( E | x 1 | 4 − 1 − t/ 2) α 2 (( α − 1) 2 − c ) 2 + ( t/ 2) α 2 (( α − 1) 2 − c ) ( α − 1) 4 ( α − 1) 2 where t = 4 when X 1 1 is real, t = 2 when X 1 1 is complex. In the complex case the real and imaginary parts of the off-diagonal entries are i.i.d. The real part of the off-diagonal elements (in either case) has variance ( t/ 4) α 2 (( α − 1) 2 − c ) . ( α − 1) 2 Moreover, the weak limit of eigenvalues corresponding to different positive α ’s satisfying ( ∗∗ ) with eigenvectors in B n are independent. 12

  13. Theorem (Rao and S.). Assume the conditions in Bai and S. (1998), and additionally a) There are r positive eigenvalues of T n all converging uniformly to t ′ , a positive number. Denote by ˆ H n the e.d.f. of the n − r other eigenvalues of T n . b) There exists positive t a < t b contained in an interval ( α, β ) with α > 0 which is outside the support of ˆ H n for all large n , such that for these n � λ 2 n ( λ − t ) 2 d ˆ H n ( λ ) ≤ 1 N for t = t a , t b . c) t ′ ∈ ( t a , t b ). Suppose λ T n i n , . . . , λ T n i n + r − 1 are the eigenvalues stated in a). Then, with probability one n →∞ λ B n n →∞ λ B n i n + r − 1 = z ( − 1 /t ′ ) lim i n = · · · = lim � � � λ = t ′ 1 + c t ′ − λdH ( λ ) . 13

  14. Application to array signal processing : Consider m i.i.d. samples of an n dimensional Gaussian (real or complex) distributed random vectors x 1 , . . . , x N , modeling the recordings ( N “snapshots”) taken from a bank of n antennas due to signals emitting from an unknown number, k , of sources, and through a noise-filled environment. It is typically assumed that x 1 is mean zero, with covariance matrix R = Ψ + Σ, where Ψ is the covariance matrix attributed to the signals and would be of rank k , and Σ is the covariance of the additive noise, assumed to be of full rank. Then the matrix R Σ = Σ − 1 R = Σ − 1 Ψ + I would have n − k eigenvalues, attributed to the noise, equal to 1, the remaining k eigenvalues all strictly greater than 1. 14

  15. When it is possible to sample the purely additive noise portion along the antennas, say z 1 , . . . , z N ′ with N ′ > n , then one would Σ − 1 � estimate R Σ with � Σ ≡ � R � R , where N ′ N � � R = 1 Σ = 1 � � x i x ∗ Σ z j z ∗ and Σ j , i N N ′ i =1 j =1 and seek to identify eigenvalues of � R � Σ which are near one, the number of the remaining eigenvalues (presumably all greater than these) would be an estimate of k , known in the literature as the detection problem. 15

  16. Theorem (Rao and S.) Denote the eigenvalues of R Σ by λ 1 ≥ λ 2 > . . . ≥ λ k > λ k +1 = . . . λ n = 1. Let l j denote the j -th largest eigenvalue of � R � Σ . Then as n, N ( n ) , N ′ ( n ) → ∞ , c n = n/N → c > N ′ = n/N ′ → c 1 < 1, with probability 1, l j converges to: 0, c 1 �   2 − 2 c 1 λ j 2 − 2 c 1 λ j + λ j 2 − 2 λ j +1 c 12 λ j − c 1 λ j − λ j +1+   λ j 1 − c − c 2 c 1 λ j if λ j > τ ( c, c 1 ) , � � 2 � 1 + 1 − (1 − c )(1 − c 1 ) 1 − c 1 if λ j ≤ τ ( c, c 1 ) for j = 1 , . . . , k . The threshold τ ( c, c 1 ) = √ c + c 1 − c 1 c −√ c + c 1 − c 1 c − 1) c − c 12 − 2 c 1 √ c + c 1 − c 1 c − c 1 ( c 12 + c 1 . (( c 1 − 1) c − c 1 ) (1 − c 1 ) 2 16

Download Presentation
Download Policy: The content available on the website is offered to you 'AS IS' for your personal information and use only. It cannot be commercialized, licensed, or distributed on other websites without prior consent from the author. To download a presentation, simply click this link. If you encounter any difficulties during the download process, it's possible that the publisher has removed the file from their server.

Recommend


More recommend