<< Chapter < Page Chapter >> Page >
This course is a short series of lectures on Introductory Statistics. Topics covered are listed in the Table of Contents. The notes were prepared by EwaPaszek and Marek Kimmel. The development of this course has been supported by NSF 0203396 grant.

The mean, variance, and standard deviation

Mean and variance

Certain mathematical expectations are so important that they have special names. In this section we consider two of them: the mean and the variance.

Mean Value

If X is a random variable with p.d.f. f ( x ) of the discrete type and space R = ( b 1 , b 2 , b 3 ,... ) , then E ( X ) = R x f ( x ) = b 1 f ( b 1 ) + b 2 f ( b 2 ) + b 3 f ( b 3 ) + ... is the weighted average of the numbers belonging to R , where the weights are given by the p.d.f. f ( x ) .

We call E ( X ) the mean of X (or the mean of the distribution ) and denote it by μ . That is, μ = E ( X ) .

In mechanics, the weighted average of the points b 1 , b 2 , b 3 ,... in one-dimensional space is called the centroid of the system. Those without the mechanics background can think of the centroid as being the point of balance for the system in which the weights f ( b 1 ) , f ( b 2 ) , f ( b 3 ) ,... are places upon the points b 1 , b 2 , b 3 ,... .

Let X have the p.d.f.

f ( x ) = { 1 8 , x = 0,3, 3 8 , x = 1,2.

The mean of X is

μ = E [ X = 0 ( 1 8 ) + 1 ( 3 8 ) + 2 ( 3 8 ) + 3 ( 1 8 ) = 3 2 .

Got questions? Get instant answers now!

The example below shows that if the outcomes of X are equally likely (i.e., each of the outcomes has the same probability), then the mean of X is the arithmetic average of these outcomes.

Roll a fair die and let X denote the outcome. Thus X has the p.d.f. f ( x ) = 1 6 , x = 1 , 2 , 3 , 4 , 5 , 6. Then,

E ( X ) = x = 1 6 x ( 1 6 ) = 1 + 2 + 3 + 4 + 5 + 6 6 = 7 2 ,

which is the arithmetic average of the first six positive integers.

Got questions? Get instant answers now!

Variance

It was denoted that the mean μ = E ( X ) is the centroid of a system of weights of measure of the central location of the probability distribution of X . A measure of the dispersion or spread of a distribution is defined as follows:

If u ( x ) = ( x μ ) 2 and E [ ( X μ ) 2 ] exists, the variance , frequently denoted by σ 2 or V a r ( X ) , of a random variable X of the discrete type (or variance of the distribution) is defined by

σ 2 = E [ ( X μ ) 2 ] = R ( x μ ) 2 f ( x ) .

The positive square root of the variance is called the standard deviation of X and is denoted by

σ = V a r ( X ) = E [ ( X μ ) 2 ] .

Let the p.d.f. of X by defined by f ( x ) = x 6 , x = 1 , 2 , 3.

The mean of X is

μ = E ( X ) = 1 ( 1 6 ) + 2 ( 2 6 ) + 3 ( 3 6 ) = 7 3 .

To find the variance and standard deviation of X we first find

E ( X 2 ) = 1 2 ( 1 6 ) + 2 2 ( 2 6 ) + 3 2 ( 3 6 ) = 36 6 = 6.

Thus the variance of X is σ 2 = E ( X 2 ) μ 2 = 6 ( 7 3 ) 2 = 5 9 ,

and the standard deviation of X is σ = 5 9 = 0.745.

Got questions? Get instant answers now!

Let X be a random variable with mean μ x and variance σ x 2 . Of course, Y = a X + b , where a and b are constants, is a random variable, too. The mean of Y is

μ Y = E ( Y ) = E ( a X + b ) = a E ( X ) + b = a μ X + b .  

Moreover, the variance of Y is

σ Y 2 = E [ ( Y μ Y ) 2 ] = E [ ( a X + b a μ X b ) 2 ] = E [ a 2 ( X μ X ) 2 ] = a 2 σ X 2 .

Got questions? Get instant answers now!

Moments of the distribution

Let r be a positive integer. If E ( X r ) = R x r f ( x ) exists, it is called the r th moment of the distribution about the origin. The expression moment has its origin in the study of mechanics.

In addition, the expectation E [ ( X b ) r ] = R x r f ( x ) is called the r th moment of the distribution about b . For a given positive integer r.

E [ ( X ) r ] = E [ X ( X 1 ) ( X 2 ) ( X r + 1 ) ] is called the r th factorial moment .

The second factorial moment is equal to the difference of the second and first moments: E [ X ( X 1 ) ] = E ( X 2 ) E ( X ) .

There is another formula that can be used for computing the variance that uses the second factorial moment and sometimes simplifies the calculations.

First find the values of E ( X ) and E [ X ( X 1 ) ] . Then σ 2 = E [ X ( X 1 ) ] + E ( X ) [ E ( X ) ] 2 , since using the distributive property of E , this becomes σ 2 = E ( X 2 ) E ( X ) + E ( X ) [ E ( X ) ] 2 = E ( X 2 ) μ 2 .

Let continue with example 4 , it can be find that

E [ X ( X 1 ) ] = 1 ( 0 ) ( 1 6 ) + 2 ( 1 ) ( 2 6 ) + 3 ( 2 ) ( 3 6 ) = 22 6 .

Thus σ 2 = E [ X ( X 1 ) ] + E ( X ) [ E ( X ) ] 2 = 22 6 + 7 3 ( 7 3 ) 2 = 5 9 .

Got questions? Get instant answers now!
Recall the empirical distribution is defined by placing the weight (probability) of 1/ n on each of n observations x 1 , x 2 ,..., x n . Then the mean of this empirical distribution is i = 1 n x i 1 n = i = 1 n x i n = x ¯ .

The symbol x ¯ represents the mean of the empirical distribution . It is seen that x ¯ is usually close in value to μ = E ( X ) ; thus, when μ is unknown, x ¯ will be used to estimate μ .

Similarly, the variance of the empirical distribution can be computed. Let v denote this variance so that it is equal to

v = i = 1 n ( x i x ¯ ) 2 1 n = i = 1 n x i 2 1 n x ¯ 2 = 1 n i = 1 n x i 2 x ¯ 2 .

This last statement is true because, in general, σ 2 = E ( X 2 ) μ 2 .

There is a relationship between the sample variance s 2 and variance v of the empirical distribution, namely s 2 = n s / ( n 1 ) . Of course, with large n , the difference between s 2 and v is very small. Usually, we use s 2 to estimate σ 2 when σ 2 is unknown.
BERNOULLI TRIALS and BINOMIAL DISTRIBUTION

Get Jobilize Job Search Mobile App in your pocket Now!

Get it on Google Play Download on the App Store Now




Source:  OpenStax, Introduction to statistics. OpenStax CNX. Oct 09, 2007 Download for free at http://cnx.org/content/col10343/1.3
Google Play and the Google Play logo are trademarks of Google Inc.

Notification Switch

Would you like to follow the 'Introduction to statistics' conversation and receive update notifications?

Ask