<< Chapter < Page Chapter >> Page >

Example

Reading the number shown on a die is a random variable. The number shown is in R .

Calculating the displacements of the nodes is a random variable. The actual displacements are in R 14

Definition 3 Probability Distribution

A function on a set B R n describing the probability that X ( ω ) B .

Example

The distribution, the log-normal distribution, and the Poisson distribution are all well-known distributions.

Definition 4 Expectation, Variance, Covariance

The expectation of a distribution is the “center of mass”:

E X = R n x π x d x = x ¯ = μ = x 0 .

The empirical expectation of a sample x 1 , x 2 , , x N is the mean:

x 0 ^ = 1 N j = 1 N x j .

The variance of a distribution is the expectation of the squared difference from the expectation of the distribution, var X = σ 2 = E X - x ¯ 2 = R X - x ¯ 2 π x d x . This is only for a one-dimensional distribution.

The covariance of two single-variable distributions is the expectation of the deviation from the expectation of one, times the expectation of the deviation from the expectation of the other, cov ( X , Y ) = E ( X - x ¯ ) ( Y - y ¯ ) = E X Y - E X E Y .

The covariance matrix of a distribution has, for its ( i , j ) entry the covariance of the i and j components of the distribution. The empirical covariance matrix of a sample x 1 , x 2 , , x N is the mean of deviation from the mean:

Γ ^ = 1 N j = 1 N ( x j - x 0 ^ ) ( x j - x 0 ^ ) T .

Definition 5 Normal (Gaussian) Distribution

In one dimension, the normal distribution has probability distribution function

π x = 1 σ 2 π exp - x - μ 2 2 σ 2 .

In n dimensions, the normal distribution has probability distribution function

π x = 1 2 π n det ( Γ ) 1 / 2 exp - 1 2 x - x 0 T Γ - 1 x - x 0 .

Definition 6 Maximum Likelihood Estimate

Given a sample from a known distribution depending on unknown parameter θ , the value of θ that maximizes the probability that the distribution returns the given sample.

Definition 7 Equiprobability Curve

Given an n -dimensional distribution, the locus of points in n -dimensions with equivalent probability.

Example

The equiprobability curves for a normal n -dimensional distribution are n -dimensional ellipses.

Normality

Because we wish to use statitsical inference to solve this problem, we begin by looking at the probability distribution of the displacements.

Calvetti and Somersalo [link] present a method for studying the normality of a 2-dimensional sample S , by using the equiprobability curves, which they also call credibility curves. Let π , corresponding to a random variable X , be a probability density. Let

B α = x R 2 | π ( x ) α , α > 0 ;

that is, the set of all points whose probability is greater than or equal to some positive number α . This corresponds to either the empty set (for α > 1 ) or the interior of an equiprobability curve; if π is normally distributed and α 1 , then B α is the interior of an ellipse. We calculate the probability that X is in B α , the credibility:

p = P X B α = B α π x d x , 0 < p < 1 .

We would like to find an explicit expression relating α and p , α = α ( p ) . If S is normally distributed, then the number of points inside B α ( p ) is about p n .

The immediate goal, then is to calculate the number of points in S that lie within B α for some α . We will evaluate the integral p = B α π x d x , but first we will need a change of variable. Let U D - 1 U T be the eigenvalue of decomposition of Γ ^ - 1 (the inverse of the empirical covariance matrix), where

Get Jobilize Job Search Mobile App in your pocket Now!

Get it on Google Play Download on the App Store Now




Source:  OpenStax, The art of the pfug. OpenStax CNX. Jun 05, 2013 Download for free at http://cnx.org/content/col10523/1.34
Google Play and the Google Play logo are trademarks of Google Inc.

Notification Switch

Would you like to follow the 'The art of the pfug' conversation and receive update notifications?

Ask