Introduction
The minimum variance unbiased estimator (MVUE) has the lowest variance among all unbiased estimators. In general, there is no method to find the MVUE; however, for a special class of estimation problems, we can find these estimators. This module concerns a special class of signal models for which a minimum variance unbiased estimator (MVUE) is available. This special class consists of estimation problems with linear signal models and additive Gaussian noise.
The module begins by introducing the linear signal model and derives the MVUE for colored Gaussian noise. Next, we assume white Gaussian noise, which can be considered a special case of colored noise. When the information on noise distribution is not complete, an interesting approach is to approximate the noise as Gaussian by considering only the expected value and variance for the noise. Such estimators are called best linear unbiased estimators (BLUE), covered at the end of the module.
Linear Signal Model
A linear signal model is any model for which the signal s[n,θ] can be expressed as a linear combination of the weighted parameters, i.e., For multiple signal values, the model can be expressed as a matrix-vector product of the form where is the signal vector of length , is the so-called observation matrix of size , and is the parameter vector of length .
Example:
Assume our signal model is a polynomial of degree with coefficients , i.e., and suppose observations for are given. The corresponding observation matrix is
MVUE for Linear Models and Additive Gaussian Noise
To find the MVUE for linear signals in Gaussian noise, we will take advantage of the equality constraint of the CRLB. Therefore, we need to show that we can express the derivative of the log-likelihood function in the form see module on the CRLB. To obtain the derivative of the log-likelihood function, we need first to determine the probability density function of the observations. For the linear signal model embedded in additive Gaussian noise, our observations are modeled as where is zero mean noise with covariance matrix , i.e., . Thus, the PDF of the observation vector is
The corresponding log-likelihood function is which after expanding the second term becomes
The gradient of the log-likelihood function is
How to compute the gradient.
The gradient of the log-likelihood function is i.e., it is a vector containing all the partial derivatives of the log-likelihood function. To evaluate the partial derivatives, we first use the linearity property of the differential operator. The single expressions in the log-likelihood function are of the form , , and . The partial derivatives and the gradient of these expressions can be obtained as follows:
- The expression is Taking the derivative with respect to one specific element is since only the addend is a function of with derivative . All other addends of the sum are constants and vanish.
- Since is true for real-valued vectors, we also have
- The expression can also be expressed as which after rearranging the summations is Taking the derivative with respect to , we obtain Note that the two summations miss the summation index and , respectively. However, we have additionally to the two sums the expression which is exactly the missing addend in the sums. Thus, we can add distribute over the two sums and get The above expression can equivalently be expressed as an inner product between the th row of , denoted by , and and the inner product between the th column of , denoted by , and , i.e.,
Combining the above results, we obtain and For symmetric matrices, i.e., , the latter expression simplifies to
Note that the covariance matrix is symmetric and so is its inverse. Using , , and , we obtain
If is invertible, we further have By comparing with , we recognize the Fisher information matrix and the estimator We found an efficient estimator, and thus, an MVUE. In the module on the MLE, we also stated that if an efficient estimator exits, it is the MLE, which is easily verified by equating with zero and solving for .
MVUE for Linear Models in White Additive Gaussian Noise
So far, we have assumed that our noise is colored, i.e., that our noise sample are correlated. For the case of additive white Gaussian noise we can reuse the above result. Therefore, we assume that our individual noise samples are IID with zero mean and variance . In this case, the covariance matrix is a diagonal matrix with elements on the main diagonal, i.e., Substitung in yields where we recognize the estimator and the Fisher information matrix
Example
Assume we have a signal composed of sinusoidal signals embedded in white Gaussian noise with zero mean and variance , i.e., where and are the unknown amplitude and phase, respectively. The frequency parameters to are assumed to be known, different, and between and . The signal is linear in the parameter , but it is not linear in the unknown phase . However, we can use the trigonometric relations and to express as The transformed model is now linear in the unknown parameters and . Due to the invariance of the MLE, we can find the maximum likelihood estimate for the amplitudes and and obtain the maximum likelihood estimate for and by using relations and The signal model in for multiple observation can be expressed as The observation matrix is with columns and respectively. The parameter vector for this model is
Note that the frequencies are multiple of the fundamental frequency , and thus, the columns of the observation matrix are orthogonal. The matrix is i.e., a diagonal matrix with entries on the main diagonal. The maximum likelihood estimate is The estimate of the parameter and are and Suppose that the frequency takes all values of from 0 to , then and are the discrete Fourier transform for real signals.
Best Linear Unbiased Estimator (BLUE)
So far, we have derived the MVUE for linear signal models with AWGN. In many applications, it is not possible to find an MVUE estimator using techniques presented up to this point. A popular choice is to restrict the estimator to be linear in the observation , i.e., of the form Here, we assume that observations are available and that we seek to estimate unknown parameter with . Thus, is of size . Depending on the choice of the 's coefficients, different estimators can be obtained. Among all linear estimators, we seek an estimator that is unbiased and has minimum variance. The estimator which fulfills these properties is called the best linear unbiased estimator (BLUE). The BLUE, as we will see, is also applicable, if limited knowledge about the PDF of the observation is available.
To find the BLUE, we have to start from the assumptions at hand; that the estimator has zero bias and minimum variance. The zero bias condition is formulated as where is the matrix that consists of the coefficients . The constraint of unbiasedness requires a linear relation between the parameter vector and the expected value of the data such that Substituting in yields
Thus, for the linear estimator to be unbiased, we require that where is the identity matrix.
The variance of the th estimate is where is the th row of . This can be equivalently expressed as Here, is the covariance matrix for the data, which is equivalent to the covariance matrix for the zero-mean noise due to the linear signal model.
The BLUE is found by minimizing the variance for all estimates with the constraint . Thus, the optimization problem is where is the th column of . The minimum can be found using Lagrange multipliers, which results in The covariance matrix for the estimate is
Note that the BLUE is the MVUE for the linear model in . This follows directly from the unbiased constraint