A useful variant of the Davis–Kahan theorem for statisticians
Abstract The Davis–Kahan theorem is used in the analysis of many statistical procedures to
bound the distance between subspaces spanned by population eigenvectors and their …
bound the distance between subspaces spanned by population eigenvectors and their …
The limiting distributions of eigenvalues of sample correlation matrices
T Jiang - Sankhyā: The Indian Journal of Statistics, 2004 - JSTOR
Let be an n by p data matrix, where the n rows form a random sample of size n from a certain
p-dimensional population distribution. Let be the p× p sample correlation coefficient matrix …
p-dimensional population distribution. Let be the p× p sample correlation coefficient matrix …
Exact separation of eigenvalues of large dimensional sample covariance matrices
ZD Bai, JW Silverstein - Annals of probability, 1999 - JSTOR
Let Bn=(1/N) T1/2 n Xn X* n T1/2 n where Xn is nx N with iid complex standardized entries
having finite fourth moment, and T1/2 n is a Hermitian square root of the nonnegative …
having finite fourth moment, and T1/2 n is a Hermitian square root of the nonnegative …
[PDF][PDF] The eigen-decomposition: Eigenvalues and eigenvectors
H Abdi - Encyclopedia of measurement and statistics, 2007 - malabdali.com
Eigenvectors and eigenvalues are numbers and vectors associated to square matrices, and
together they provide the eigen-decomposition of a matrix which analyzes the structure of …
together they provide the eigen-decomposition of a matrix which analyzes the structure of …
A simple rule for the selection of principal components
A vast literature has been devoted to the assessment of the proper number of eigenvalues
that have to be retained in Principal Components Analysis. Most of the publications are …
that have to be retained in Principal Components Analysis. Most of the publications are …
Combining eigenvalues and variation of eigenvectors for order determination
W Luo, B Li - Biometrika, 2016 - academic.oup.com
In applying statistical methods such as principal component analysis, canonical correlation
analysis, and sufficient dimension reduction, we need to determine how many eigenvectors …
analysis, and sufficient dimension reduction, we need to determine how many eigenvectors …
Defining probability density for a distribution of random functions
A Delaigle, P Hall - The Annals of Statistics, 2010 - JSTOR
The notion of probability density for a random function is not as straight-forward as in finite-
dimensional cases. While a probability density function generally does not exist for …
dimensional cases. While a probability density function generally does not exist for …
Disco analysis: A nonparametric extension of analysis of variance
ML Rizzo, GJ Székely - 2010 - projecteuclid.org
In classical analysis of variance, dispersion is measured by considering squared distances
of sample elements from the sample mean. We consider a measure of dispersion for …
of sample elements from the sample mean. We consider a measure of dispersion for …
An Eigenvector Perturbation Bound and Its Application
In statistics and machine learning, we are interested in the eigenvectors (or singular vectors)
of certain matrices (eg covariance matrices, data matrices, etc). However, those matrices are …
of certain matrices (eg covariance matrices, data matrices, etc). However, those matrices are …
Distinctness of the eigenvalues of a quadratic form in a multivariate sample
M Okamoto - The Annals of Statistics, 1973 - JSTOR
This paper shows that a quadratic form in a multivariate sample has a certain rank and its
nonzero eigenvalues are distinct with probability one under the assumption that the matrix …
nonzero eigenvalues are distinct with probability one under the assumption that the matrix …