<< Chapter < Page | Chapter >> Page > |
Adaptive Equalization
Another type of equalization, capable of tracking a slowly time-varying channel response, is known as adaptive equalization. It can be implemented to perform tap-weight adjustments periodically or continually. Periodic adjustments are accomplished by periodically transmitting a preamble or short training sequence of digital data known by the receiver. Continual adjustment are accomplished by replacing the known training sequence with a sequence of data symbols estimated from the equalizer output and treated as known data. When performed continually and automatically in this way, the adaptive procedure is referred to as decision directed.
If the probability of error exceeds one percent, the decision directed equalizer might not converge. A common solution to this problem is to initialize the equalizer with an alternate process, such as a preamble to provide good channel-error performance, and then switch to decision-directed mode.
The simultaneous equations described in equation (4) of module “Transversal Equalizer” do not include the effects of channel noise. To obtain stable solution to the filter weights, it is necessary that the data be averaged to obtain the stable signal statistic, or the noisy solution obtained from the noisy data must be averaged. The most robust algorithm that average noisy solution is the least-mean-square (LMS) algorithm. Each iteration of this algorithm uses a noisy estimate of the error gradient to adjust the weights in the direction to reduce the average mean-square error.
The noisy gradient is simply the product of an error scalar and the data vector .
(1)
Where and are the desired output signal (a sample free of ISI) and the estimate at time k.
(2)
Where is the transpose of the weight vector at time k.
Iterative process that updates the set of weights is obtained as follows:
(3)
Where is the vector of filter weights at time k, and is a small term that limits the coefficient step size and thus controls the rate of convergence of the algorithm as well as the variance of the steady state solution. Stability is assured if the parameter is smaller than the reciprocal of the energy of the data in the filter. Thus, while we want the convergence parameter to be large for fast convergence but not so large as to be unstable, we also want it to be small enough for low variance.
Notification Switch
Would you like to follow the 'Principles of digital communications' conversation and receive update notifications?