Precision (statistics)


In statistics, precision is the reciprocal of the variance, and the precision matrix is the matrix inverse of the covariance matrix. Thus, if we are considering a single random variable in isolation, its precision is the inverse of its variance: p=1/σ². Some particular statistical models define the term precision differently.
One particular use of the precision matrix is in the context of Bayesian analysis of the multivariate normal distribution: for example, Bernardo & Smith prefer to parameterise the multivariate normal distribution in terms of the precision matrix, rather than the covariance matrix, because of certain simplifications that then arise. For instance, if both the prior and the likelihood have Gaussian form, and the precision matrix of both of these exist, then the precision matrix of the posterior will simply be the sum of the precision matrices of the prior and the likelihood.
As the inverse of a Hermitian matrix, the precision matrix of real-valued random variables, if it exists, is positive definite and symmetrical.
Another reason the precision matrix may be useful is that if two dimensions i and j of a multivariate normal are conditionally independent, then the ij and ji elements of the precision matrix are 0. This means that precision matrices tend to be sparse when many of the dimensions are conditionally independent, which can lead to computational efficiencies when working with them. It also means that precision matrices are closely related to the idea of partial correlation.

History

The term precision in this sense first appeared in the works of GaussTheoria motus corporum coelestium in sectionibus conicis solem ambientium”. Gauss's definition differs from the modern one by a factor of. He writes, for the density function of a normal random variable with precision h,
Later Whittaker & Robinson “Calculus of observations” called this quantity the modulus, but this term has dropped out of use.