<< Chapter < Page Chapter >> Page >

Statistical analysis is fundamentally an inversion process. The objective is to the "causes"--parameters of the probabilistic data generationmodel--from the "effects"--observations. This can be seen in our interpretation of the likelihood function.

Given a parameter , observations are generated according to p x The likelihood function has the same form as the conditional density function above l | x p x except now x is given (we take measurements) and is the variable. The likelihood function essentially inverts the role of observation(effect) and parameter (cause).

Unfortunately, the likelihood function does not provide a formal framework for the desired inversion.

One problem is that the parameter is supposed to be a fixed and deterministic quantity while the observation x is the realization of a random process. So their role aren't really interchangeable in thissetting.

Moreover, while it is tempting to interpret the likelihood l | x as a density function for , this is not always possible; for example, often l | x

Another problematic issue is the mathematical formalization of statements like: "Based on the measurements x , I am 95% confident that falls in a certain range."

Suppose you toss a coin 10 times and each time it comes up "heads." It might be reasonable to say that we are99% sure that the coin is unfair, biased towards heads.

Formally: H 0 : prob heads 0.5 x N x x 1 N x which is the binomial likelihood. p x 0.5 ? The problem with this is that p x H 0 implies that is a random , not deterministic, quantity. So, while "confidence" statements are very reasonable and in fact a normal part of "everyday thinking," this idea can not besupported from the classical perspective.

All of these "deficiencies" can be circumvented by a change in how we view the parameter .

If we view as the realization of a random variable with density p , then Bayes Rule (Bayes, 1763) shows that p x p x p p x p Thus, from this perspective we obtain a well-defined inversion: Given x , the parameter is distributing according to p x .

From here, confidence measures such as p x H 0 are perfectly legitimate quantities to ask for.

Bayesian statistical model
A statistical model compose of a data generation model, p x , and a prior distribution on the parameters, p .

The prior distriubtion (or prior for short) models the uncertainty in the parameter. More specifically, p models our knowledge--or lack thereof--prior to collecting data.

Notice that p x p x p p x p x p since the data x are known , p x is just a constant. Hence, p x is proportional to the likelihood function multiplied by the prior.

Bayesian analysis has some significant advantages over classical statistical analysis:

  • properly inverts the relationship between causes and effects
  • permits meaningful assessments in confidence regions
  • enables the incorporation of prior knowledge into the analysis (which could come from previous experiments, forexample)
  • leads to more accurate estimators (provided the prior knowledge is accurate)
  • obeys the Likelihood and Sufficiency principles

Get Jobilize Job Search Mobile App in your pocket Now!

Get it on Google Play Download on the App Store Now




Source:  OpenStax, Statistical signal processing. OpenStax CNX. Jun 14, 2004 Download for free at http://cnx.org/content/col10232/1.1
Google Play and the Google Play logo are trademarks of Google Inc.

Notification Switch

Would you like to follow the 'Statistical signal processing' conversation and receive update notifications?

Ask