In this expository note, we give a modern proof of hansonwright inequality for quadratic forms in sub gaussian random variables. The distribution of a gaussian process is the joint distribution of all those. A tail inequality for quadratic forms of subgaussian random vectors daniel hsu sham m. Linear transformations and gaussian random vectors. In the case of discrete functions, a gaussian process is simply a different interpretation of a multivariate normal distribution. The hansonwright inequality is a general concentration result for quadratic forms in subgaussian random variables. Informally, the tails of a sub gaussian distribution are dominated by i. On the estimation of the mean of a random vector joly, emilien, lugosi, gabor, and imbuzeiro oliveira, roberto, electronic journal of statistics, 2017. Feb 15, 2016 a random variable is subgaussian if its subgaussian norm. Sub gaussian estimators of the mean of a random matrix with heavytailed entries minsker, stanislav, the. Estimation of the covariance matrix has attracted a lot of attention of the statistical research community over the years, partially due to important applications such as principal component analysis. In probability, gaussian random variables are the easiest and most commonly used distribution encountered. Two examples are given to illustrate these results. Sub gaussian estimators of the mean of a random matrix with heavytailed entries minsker, stanislav, the annals of statistics, 2018 sub gaussian mean estimators devroye, luc, lerasle, matthieu, lugosi, gabor, and oliveira, roberto i.
Matrix decompositions using subgaussian random matrices. Kakadey tong zhangz abstract this article proves an exponential probability tail inequality for positive semide. Picking a random vector from spherical gaussian distribution. We deduce a useful concentration inequality for subgaussian random vectors. Multivariate gaussian random vectors part 1 definition. Jordan oncerf and thomas sibutpinote 1 subgaussian random variables in probabilit,y gaussian random ariablevs are the easiest and most commonly used distribution encountered. Random vectors and multivariate normal distributions 3. On simulating exchangeable subgaussian random vectors. Do october 10, 2008 a vectorvalued random variable x x1 xn t is said to have a multivariate normal or gaussian distribution with mean. Note that we are following the terminology of 5 in calling a random variable pregaussian when it has a subexponential tail decay. A nice reference on subgaussian random variables is rig15, which shows they have many useful properties similar to gaussian distributions, and we recall a few that will interest us bellow.
Supergaussian directions of random vectors boaz klartag abstract we establish the following universality property in high dimensions. Given a symmetric, positive semide nite matrix, is it the covariance matrix of some random vector. Subgaussian estimators of the mean of a random matrix with heavytailed entries stanislav minsker email. Tel aviv university, 2005 gaussian measures and gaussian processes 45 3b estimating the norm let m be a random n nmatrix distributed according to 3a1. In signal pro cessing x often used to represen t a set of n samples random signal x a pro cess. Widesense stationary gaussian processes are strictly stationary. This book places particular emphasis on random vectors, random matrices, and random projections. Informally, the tails of a subgaussian distribution are dominated by i. Chapter 3 random vectors and multivariate normal distributions. Subgaussian estimators of the mean of a random matrix with.
Sub gaussian variables are an important class of random variables that have strong tail decay properties. Highdimensional probability is an area of probability theory that studies random objects in rn where the dimension ncan be very large. Ourgoalinthissectionistodevelopanalyticalresultsfortheprobability distribution function pdf ofatransformedrandomvectory inrn. However, when the distribution is not necessarily subgaussian and is possibly heavytailed, one cannot expect such a subgaussian behavior of the sample mean. Certain characterizations for an exchangeable sub gaussian random vector are given and a method together with an splus function for simulating such a vector are introduced. In probability theory, a sub gaussian distribution is a probability distribution with strong tail decay. It is nonzeromean but still unit variance gaussian vector. However, when the distribution is not necessarily sub gaussian and is possibly heavytailed, one cannot expect such a sub gaussian behavior of the sample mean. The bound is analogous to one that holds when the vector has independent gaussian entries. Subgaussian estimators of the mean of a random matrix. Then, you generate random vectors coordinates by sampling each of the distributions. Overview of the proposed global gaussian distribution embedding network g.
It teaches basic theoretical skills for the analysis of these objects, which include. Effectively, the edited code below represents coordinates of 10 twodimensional. Oct 07, 2009 the definition of a multivariate gaussian random vector is presented and compared to the gaussian pdf for a single random variable as weve studied in past lectures. Formally, the probability distribution of a random variable x is called sub gaussian if there are positive. If the random vector x has probability density f x. In this expository note, we give a modern proof of hansonwright inequality for quadratic forms in subgaussian random variables. Probabilit y of random v ectors multiple random v ariables eac h outcome of a random exp erimen tma y need to b e describ ed b y a set of n 1 random v ariables f x 1x n g,orinv ector form. If is a random vector such that its components are independent and subgaussian, and is some deterministic matrix, then the hansonwright inequality tells us how quickly the quadratic form concentrates around its expectation. Whereas the multivariate normal distribution models random vectors, gaussian processes allow us to define distributions over functions and deformation fields. The distribution of mx does not depend on the choice of a unit vector x 2 rn due to the oninvariance and is equal to n 1 p n. A ndimensional complex random vector, is a complex standard normal random vector or complex standard gaussian random vector if its components are independent and all of them are standard complex normal random variables as defined above.
There is a proof for the bivariate case on the first page of this. Johnsonlindenstrauss theory 1 subgaussian random variables. If every pair of random variables in the random vector x have the same correlation. If is the covariance matrix of a random vector, then for any constant vector awe have at a 0.
This class contains, for example, all the bounded random variables and all the normal variables. We introduce a new estimator that achieves a purely sub gaussian performance under the only. In particular, any rv with such a finite norm has a tail bound that decays as fast as the one of a gaussian rv, i. For such large n, a question to ask would be whether a. I was recently reading a research paper on probabilistic matrix factorization and the authors were picking a random vector from a spherical gaussian distribution ui. Transformation of gaussian random vectors considerthecaseofnvariategaussianrandomvectorwithmeanvectormx, covariance matrixcx andpdfgivenby. A scatter matrix estimate based on the zonotope koshevoy, gleb a. Intuitively, a random variable is called subgaussian when it is subordinate to a gaussian random variable, in a sense that will be made precise. A tail inequality for quadratic forms of subgaussian. Subgaussian random variables and processes are considered.
Then, the random vector x is sub gaussian with variance proxy. A random variable is subgaussian if its subgaussian norm. Where lambda is a regularization parameter and ik is kth dimensional identity matrix. In this section, we introduce sub gaussian random variables and discuss some of their properties. Global gaussian distribution embedding network and its. A subgaussian distribution is any probability distribution that has tails bounded by a gaussian and has a mean of zero. In this section, we introduce subgaussian random variables and discuss some of their properties. In probability theory, a subgaussian distribution is a probability distribution with strong tail decay. Subgaussian estimators of the mean of a random vector article in the annals of statistics 472 february 2017 with 59 reads how we measure reads.
Certain characterizations for an exchangeable subgaussian random vector are given and a method together with an splus function for simulating such a vector are introduced. A note on subgaussian random variables cryptology eprint. If it is not zero mean, we can have noncentral chi distribution. The intuitive idea here is that gaussian rvs arise in practice because of the addition of large st m can be approximated by a gaussian rv. Formally, the probability distribution of a random variable x is called sub gaussian if there are positive constants c, v such that for every t 0. Probabilit y of random v ectors harvey mudd college. A ndimensional complex random vector, is a complex standard normal random vector or complex standard gaussian random vector if its components are independent and all of them are standard complex normal random variables as defined above p. Transformation of random vectors university of new mexico. These random variables whose exact definition is given below are said to be subgaussian. Ir has gaussian distribution iff it has a density p with.
The partition of a gaussian pdf suppose we partition the vector x. Chapter 2 sub gaussian random variables sources for this chapter, philippe rigollet and janchristian hutter lectures notes on high dimensionalstatisticschapter1. Quantized subgaussian random matrices are still rip. In this case, i think, youd need n normal distributions, each corresponding to an univariate distribution along one of the coordinates. A tail inequality for quadratic forms of subgaussian random. Then, the random vector x is subgaussian with variance proxy.
Subgaussian estimators of the mean of a random vector. Subgaussian estimators of the mean of a random vector gabor lugosi. The definition of a multivariate gaussian random vector is presented and compared to the gaussian pdf for a single random variable as weve studied in past lectures. The transpose at of an by m matrix a is an m by matrix 3 with. In probability theory and statistics, a gaussian process is a stochastic process a collection of random variables indexed by time or space, such that every finite collection of those random variables has a multivariate normal distribution, i. A traditional method for simulating a subgaussian random vector is by using 1, which we call it method 1 m1. Subgaussian estimators of the mean of a random matrix with heavytailed entries minsker, stanislav, the. We introduce a new estimator that achieves a purely subgaussian performance under the only. Gaussian random vectors october 11, 2011 140 the weak law of large numbers the central limit theorem covariance matrices the multidimensional gaussian law multidimensional gaussian density marginal distributions eigenvalues of the covariance matrix uncorrelation and independence linear combinations conditional densities 240 the weak law of. The standard benchmark hpl highperformance linpack chooses a to be a random matrix with elements from a uniform distribution on. Joint distribution of subset of jointly gaussian random.
Properties of gaussian random process the mean and autocorrelation functions completely characterize a gaussian random process. Gaussian random vectors october 11, 2011 140 the weak law of large numbers the central limit theorem covariance matrices the multidimensional gaussian law multidimensional gaussian density marginal distributions eigenvalues of the covariance matrix uncorrelation and independence. Thus, when is it not reasonable to assume a subgaussian distribution and heavy tails may be a concern, the sample mean is a risky choice. Four lectures on probabilistic methods for data science. Norms of subexponential random vectors sciencedirect. Sub gaussian estimators of the mean of a random matrix with heavytailed entries stanislav minsker email. The set of subgaussian random variables includes for instance the gaussian, the bernoulli and the bounded rvs, as. That is, satis es the property of being a positive semide nite matrix. Subgaussian variables are an important class of random variables that have strong tail decay properties. I just realized you were, probably, talking about multivariate gaussian distribution. If the random variable x has the gaussian distribution n02, then for each p0 one has ejxjp r 2p. Then, you generate random vector s coordinates by sampling each of the distributions. My guess is that the pdf is also a gaussian with the corresponding entries of the mean vector and covariance matrix, but i dont have a real proof of this. In fact, if the random variable xis subgaussian, then its absolute moments are bounded above by an expression involving the subgaussian parameter and the gamma function, somewhat similar to the right hand side of the.
1038 825 946 783 1216 463 1126 1408 15 1489 280 303 1354 869 924 1230 1339 414 1016 939 690 233 422 276 45 1501 735 305 409 279 238 940 1476 999 160 366 568 1228 839 1053 1446 601 345 1332 858