<< Chapter < Page Chapter >> Page >

We derived the minimum mean-squared error estimator in the previous section with no constraint on the form of the estimator. Depending on the problem, thecomputations could be a linear function of the observations (which is always the case in Gaussian problems) ornonlinear. Deriving this estimator is often difficult, which limits its application. We consider here a variation of MMSEestimation by constraining the estimator to be linear while minimizing the mean-squared estimation error. Such linear estimators may not be optimum; the conditional expected value may be nonlinear and it always has the smallest mean-squared error. Despite this occasionalperformance deficit, linear estimators have well-understood properties, they interact will with other signal processingalgorithms because of linearity, and they can always be derived, no matter what the problem.

Let the parameter estimate r be expressed as r where is a linear operator: a 1 r 1 a 2 r 2 a 1 r 1 a 2 r 2 where a 1 , a 2 are scalars. Although all estimators of this form are obviously linear, the term linear estimator denotes that member of this family that minimizes the mean-squarederror.

r LIN r

Because of the transformation's linearity, the theory of linear vector spaces can be fruitfully used to derive the estimator andto specify its properties. One result of that theoretical framework is the well-known Orthogonality Principle (Papoulis, pp. 407-414) The linear estimator is that particular linear transformation that yieldsan estimation error orthogonal to all linear transformations of the data. The orthogonality of the error to all linear transformations is termed the universality constraint. This principle provides us not only with a formal definition of the linear estimator butalso with the mechanism to derive it. To demonstrate this intriguing result, let denote the absract inner product between two vectors and the associated norm.

x 2 x x
For example, if x and y are each column matrices having only one column,
There is a confusion as to what a vector it. "Matricies having one column" are colloquially termedvectors as are the field quantities such as electric and magnetic fields. "Vectors" and their associated inner productsare taken to be much more general mathematical objects than these. Hence the prose in this section is rathercontorted.
their inner product might be defined as x x x y . Thus, the linear estimator as defined by the Orthogonality Principle must satisfy
for all linear transformations LIN r r 0
To see that this principle produces the MMSE linear estimator, we express the mean-squared estimation error 2 for any choice of linear estimator as
2 LIN LIN 2 LIN 2 LIN 2 2 LIN LIN
As LIN is the difference of two linear transformations, it too is linear and is orthogonal to the estimation error resultingfrom LIN . As a result, the last term is zero and the mean-squared estimation error is the sum of two squared norms,each of which is, of course, nonnegative. Only the second norm varies with estimator choice; we minimize the mean-squaredestimation error by choosing the estimator to be the estimator LIN , which sets the second term to zero.

The estimation error for the minimum mean-squared linear estimator can be calculated to some degree without knowledge ofthe form of the estimator. The mean-squared estimation error is given by

LIN 2 LIN LIN LIN LIN LIN
The first term is zero because of the Orthogonality Principle. Rewriting the second term yields a general expressionfor the MMSE linear estimator's mean-squared error.
2 2 LIN
This error is the difference of two terms. The first, the mean-squared value of the parameter, represents the largestvalue that the estimation error can be for any reasonable estimator. That error can be obtained by the estimator thatignores the data and has a value of zero. The second term reduces this maximum error and represents the degree to whichthe estimate and the parameter agree on the average.

Note that the definition of the minimum mean-squared error linear estimator makes no explicit assumptions about the parameter estimation problem beingsolved. This property makes this kind of estimator attractive in many applications where neither the a priori density of the parameter vector nor the density of the observations is known precisely. Linear transformations,however, are homogeneous: A zero-values input yields a zero output. Thus, the linear estimator is especially pertinent tothose problems where the expected value of the parameter is zero. If the expected value is nonzero, the linear estimatorwould not necessarily yield the best result (See this problem )

Express the first example in vector notation so that the observation vector is written as r A n where the matrix A has the form A 1 1 . The expected value of the parameter is zero. The linear estimator has the form LIN L r , where L is a 1 L matrix. The orthogonality Principle states that the linear estimator satisfies for all 1 L matricies M L r M r 0 To use the Orthogonality Principle to derive an equation implicitly specifying the linear estimator, the "for alllinear transformations" phrase must be interpreted. Usually the quantity specifying the linear transformation must beremoved from the constraining inner product by imposing a very stringent but equivalent condition. In this example, thisphrase becomes one about matrices. The elements of the matrix M can be such that each element of the observation vector multiplies each elementof the estimation error. Thus, in this problem the Othogonality Principle means that the expected value of thematrix consisting of all pairwise priducts of these elements must be zero. L r r 0 Thus, two terms must equal each other: L r r r . The second term equals 2 A as the additive noise and the parameter are assumed to be statistically independent quantities. The quantity r r in the first term is the correlation matrix of the observations, which is given by A A 2 K n . Here, K n is the noise covariance matrix, and 2 is the parameter's variance. The quantity A A is a L L matrix with each element equaling 1. The noise vector has independent components; the covariance matrix thusequals n 2 I . The equation that L must satisfy is therefore given by L 1 L L n 2 2 2 2 2 n 2 2 2 2 2 n 2 2 2 2 The components of L are equal and are given by L i 2 n 2 L 2 . Thus, the minimum mean-squared error linear estimator has the form LIN r 2 2 n 2 L 1 L l l r l

Note that this result equals the minimum mean-squared error estimate derived earlier under the condition that 0 . Mean-squared error, linear estimators, and Gaussian problems are intimately related to each other. The linearminimum mean-squared error solution to a problem is optimal if the underlying distributions are Gaussian.

Get Jobilize Job Search Mobile App in your pocket Now!

Get it on Google Play Download on the App Store Now




Source:  OpenStax, Estimation theory. OpenStax CNX. May 14, 2006 Download for free at http://cnx.org/content/col10352/1.2
Google Play and the Google Play logo are trademarks of Google Inc.

Notification Switch

Would you like to follow the 'Estimation theory' conversation and receive update notifications?

Ask