WO2002023842A1 - Apparatus and method for using adaptive algorithms to exploit sparsity in target weight vectors in an adaptive channel equalizer - Google Patents

Apparatus and method for using adaptive algorithms to exploit sparsity in target weight vectors in an adaptive channel equalizer Download PDF

Info

Publication number
WO2002023842A1
WO2002023842A1 PCT/US2001/028758 US0128758W WO0223842A1 WO 2002023842 A1 WO2002023842 A1 WO 2002023842A1 US 0128758 W US0128758 W US 0128758W WO 0223842 A1 WO0223842 A1 WO 0223842A1
Authority
WO
WIPO (PCT)
Prior art keywords
time
algorithm
weight vector
received signal
sparse
Prior art date
Application number
PCT/US2001/028758
Other languages
French (fr)
Inventor
William A. Sethares
Robert C. Williamson
Richard K. Martin
Original Assignee
Fox Digital
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Fox Digital filed Critical Fox Digital
Priority to JP2002527158A priority Critical patent/JP2004509521A/en
Priority to EP01970985A priority patent/EP1319289A1/en
Priority to KR1020027006069A priority patent/KR20020096041A/en
Publication of WO2002023842A1 publication Critical patent/WO2002023842A1/en

Links

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N7/00Television systems
    • H04N7/015High-definition television systems
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N5/00Details of television systems
    • H04N5/14Picture signal circuitry for video frequency region
    • H04N5/21Circuitry for suppressing or minimising disturbance, e.g. moiré or halo
    • H04N5/211Ghost signal cancellation
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L25/00Baseband systems
    • H04L25/02Details ; arrangements for supplying electrical power along data transmission lines
    • H04L25/03Shaping networks in transmitter or receiver, e.g. adaptive shaping networks
    • H04L25/03006Arrangements for removing intersymbol interference
    • H04L25/03012Arrangements for removing intersymbol interference operating in the time domain
    • H04L25/03019Arrangements for removing intersymbol interference operating in the time domain adaptive, i.e. capable of adjustment during data reception
    • H04L25/03057Arrangements for removing intersymbol interference operating in the time domain adaptive, i.e. capable of adjustment during data reception with a recursive structure
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L25/00Baseband systems
    • H04L25/02Details ; arrangements for supplying electrical power along data transmission lines
    • H04L25/03Shaping networks in transmitter or receiver, e.g. adaptive shaping networks
    • H04L25/03006Arrangements for removing intersymbol interference
    • H04L25/03012Arrangements for removing intersymbol interference operating in the time domain
    • H04L25/03019Arrangements for removing intersymbol interference operating in the time domain adaptive, i.e. capable of adjustment during data reception
    • H04L25/03057Arrangements for removing intersymbol interference operating in the time domain adaptive, i.e. capable of adjustment during data reception with a recursive structure
    • H04L25/0307Arrangements for removing intersymbol interference operating in the time domain adaptive, i.e. capable of adjustment during data reception with a recursive structure using blind adaptation
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L25/00Baseband systems
    • H04L25/02Details ; arrangements for supplying electrical power along data transmission lines
    • H04L25/03Shaping networks in transmitter or receiver, e.g. adaptive shaping networks
    • H04L25/03987Equalisation for sparse channels
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/41Structure of client; Structure of client peripherals
    • H04N21/426Internal components of the client ; Characteristics thereof
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N5/00Details of television systems
    • H04N5/14Picture signal circuitry for video frequency region
    • H04N5/21Circuitry for suppressing or minimising disturbance, e.g. moiré or halo
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L25/00Baseband systems
    • H04L25/02Details ; arrangements for supplying electrical power along data transmission lines
    • H04L25/03Shaping networks in transmitter or receiver, e.g. adaptive shaping networks
    • H04L25/03006Arrangements for removing intersymbol interference
    • H04L2025/0335Arrangements for removing intersymbol interference characterised by the type of transmission
    • H04L2025/03375Passband transmission
    • H04L2025/03382Single of vestigal sideband
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L25/00Baseband systems
    • H04L25/02Details ; arrangements for supplying electrical power along data transmission lines
    • H04L25/03Shaping networks in transmitter or receiver, e.g. adaptive shaping networks
    • H04L25/03006Arrangements for removing intersymbol interference
    • H04L2025/03433Arrangements for removing intersymbol interference characterised by equaliser structure
    • H04L2025/03439Fixed structures
    • H04L2025/03445Time domain
    • H04L2025/03471Tapped delay lines
    • H04L2025/03484Tapped delay lines time-recursive
    • H04L2025/03503Tapped delay lines time-recursive as a combination of feedback and prediction filters
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L25/00Baseband systems
    • H04L25/02Details ; arrangements for supplying electrical power along data transmission lines
    • H04L25/03Shaping networks in transmitter or receiver, e.g. adaptive shaping networks
    • H04L25/03006Arrangements for removing intersymbol interference
    • H04L2025/03592Adaptation methods
    • H04L2025/03598Algorithms
    • H04L2025/03611Iterative algorithms
    • H04L2025/03617Time recursive algorithms
    • H04L2025/0363Feature restoration, e.g. constant modulus
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L25/00Baseband systems
    • H04L25/02Details ; arrangements for supplying electrical power along data transmission lines
    • H04L25/03Shaping networks in transmitter or receiver, e.g. adaptive shaping networks
    • H04L25/03006Arrangements for removing intersymbol interference
    • H04L2025/03592Adaptation methods
    • H04L2025/03598Algorithms
    • H04L2025/03681Control of adaptation
    • H04L2025/03687Control of adaptation of step size

Definitions

  • the present invention is directed, in general, to adaptive channel equalizers in digital communication devices and, more specifically, to an apparatus and method for using algorithms to exploit sparsity in target weight vectors in an adaptive channel equalizer.
  • Grand Alliance is a group of television manufacturing and research organizations in the television industry. After years of cooperative effort the Grand
  • HDTV High Definition
  • the ATSC Standard for HDTV transmission over terrestrial broadcast channels uses a signal that consists of a sequence of twelve (12) independent time-multiplexed trellis-coded data streams modulated as an eight (8) level vestigial sideband (VSB) symbol stream with a rate of 10.76 MHz.
  • This signal is converted to a six (6) MHz frequency band that corresponds to a standard VHF or UHF terrestrial television channel, over which the signal is then broadcast.
  • the ATSC Standard calls for two (2) bit data symbols of the HDTV signal to be trellis encoded in accordance with an eight (8) level (i.e., a three (3) bit) one dimensional constellation. One bit of each data symbol is pre-coded, and the other is subjected to a 1/2 encoding rate that produces two coded bits in accordance with a four (4) state trellis code.
  • twelve (12) identical encoders and pre-coders operate successively on every twelve successive data symbols. Symbols 0, 12, 24, 36, . . . are encoded as one series. Symbols 1, 13, 25, 37, . . . as a second series. Symbols 2, 14, 26, 38, . . . as a third series.
  • the ATSC Standard requires twelve (12) trellis decoders in the HDTV receiver for the twelve (12) series of time division interleaved data symbols in the signal.
  • Each trellis decoder in the HDTV receiver decodes every twelfth (12th) data symbol in the stream of coded data symbols.
  • trellis decoders are used to retrieve the original digital data that was trellis encoded just before being converted to 8-VSB symbols, modulated and broadcast.
  • the use of trellis coding provides an improvement in the signal to noise ratio of the received signal, and the time multiplexing of twelve (12) independent streams reduces the possibility of co-channel interference from an analog NTSC broadcast signal residing on the same frequency.
  • NTSC National Television Standards Committee.
  • Each of the trellis decoders for the four (4) state trellis code operates in accordance with the well-known Viterbi decoding algorithm.
  • Each of the decoders comprises a branch metric generator unit, an add-compare-select unit, and a path-memory unit. See, for example, "Trellis-coded Modulation With Redundant Signal Set, Part I, Introduction; Part II, State of the Art," by G. Ungerboeck, IEEE Communications Magazine, Vol. 25, pp. 5-21, February 1987.
  • the transmitted signal is also subject to deterministic channel distortions and distortions caused by multipath interference.
  • an adaptive channel equalizer is generally used in front of the trellis decoders to compensate for these effects.
  • the goal is to create a symbol stream that resembles, as much as possible, the symbol stream that was created by the twelve (12) trellis encoders at the transmitter.
  • One commonly used equalizer architecture makes use of a second equalizer known as a decision feedback equalizer (DFE) .
  • DFE decision feedback equalizer
  • FE forward equalizer
  • the input to the DFE is an estimate of the original transmitted value of the current output symbol of the complete equalizer (FE and DFE) .
  • the output of the decision feedback equalizer (DFE) is subsequently added to the output of the forward equalizer (FE) to generate the output symbol.
  • this estimate of the output symbol is obtained by simply "slicing" the equalizer output.
  • the term “slicing” refers to the process of taking the allowed symbol value (of the eight (8) levels specified by the 8- VSB ATSC Standard) that is nearest to that of the actual output.
  • DFE decision feedback equalizer
  • the HDTV signal is decoded in a trellis decoder that uses the Viterbi algorithm to decode the symbol stream based on the 1/2 rate trellis coding performed in the transmitter.
  • a trellis decoder that uses the Viterbi algorithm to decode the symbol stream based on the 1/2 rate trellis coding performed in the transmitter.
  • the ATSC Standard specifies that twelve (12) trellis encoders and decoders are used in parallel in a time multiplexed fashion. Trellis decoding is then followed by byte de- interleaving and Reed Solomon decoding to further correct transmission errors in the signal.
  • LMS least mean square
  • C n+ ⁇ C n + ⁇ E d n (2)
  • C n is the coefficient vector at time n
  • is an adaptation speed constant
  • d n is the data vector in the filter at time n.
  • E is the error calculated from the output of the equalizer.
  • E can be calculated in a decision directed fashion using the known training sequence embedded in the data stream. Alternatively, E can be calculated in a blind fashion using a Constant Modulus Algorithm (CMA) or a Decision Directed (DD) algorithm.
  • CMA Constant Modulus Algorithm
  • DD Decision Directed
  • Transmission channels are often sparse.
  • Optimally designed adaptive equalizers reflect the sparsity in the transmission channels.
  • Typical prior art equalization methods such as the Least Mean Square (LMS) algorithm or the Constant Modulus Algorithm (CMA) or the Decision Directed (DD) algorithm do not exploit known "a priori" information for sparse transmission channels.
  • LMS Least Mean Square
  • CMA Constant Modulus Algorithm
  • DD Decision Directed
  • Typical prior art approaches to exploiting sparsity attempt to reduce complexity. Complexity reduction is accomplished by only updating a subset of the channel model or equalizer taps.
  • Recently a prior art “exponentiated gradient” (EG) algorithm has been shown to have better performance than typical gradient methods when the target weight vector is sparse. See the article entitled “Exponentiated Gradient Versus Gradient Descent for Linear Predictors” by J.
  • the EG algorithm has several limitations that make it inappropriate for applications in communications and signal processing. For example, the EG algorithm does not decrease the required complexity.
  • the present invention uses information concerning known sparsities in a communication channel to calculate adaptive equalizer coefficients to increase a coefficient convergence rate for equalizer coefficients. It is an object of the present invention to provide an apparatus and method for designing efficient adaptive channel equalizers.
  • the term “controller,” “processor,” or “apparatus” means any device, system or part thereof that controls at least one operation, such a device may be implemented in hardware, firmware or software, or some combination of at least two of the same.
  • controllers may be centralized or distributed, whether locally or remotely.
  • a controller may comprise one or more data processors, and associated input/output devices and memory, that execute one or more application programs and/or an operating system program.
  • application programs and/or an operating system program.
  • FIGURE 1 illustrates a block diagram of an exemplary high definition television (HDTV) transmitter
  • FIGURE 2 illustrates a block diagram of an exemplary high definition television (HDTV) receiver
  • FIGURE 3 illustrates a block diagram of an adaptive channel equalizer comprising a forward equalizer (FE) filter and a decision feedback equalizer (DFE) filter;
  • FE forward equalizer
  • DFE decision feedback equalizer
  • FIGURE 4 illustrates a block diagram of an adaptive finite impulse response (FIR) filter for use in an adaptive channel equalizer
  • FIGURE 5 illustrates a block diagram showing the connection of a forward equalizer (FE) filter to a trellis decoder unit and the connection of the trellis decoder unit to a decision feedback equalizer (DFE) filter;
  • FE forward equalizer
  • DFE decision feedback equalizer
  • FIGURE 6 illustrates a graph showing the Mean Squared Error (MSE) versus time for a Least Mean Squares (LMS) algorithm and for a Sparse Least Mean Squares (LMS) algorithm
  • FIGURE 7 illustrates- a graph showing the tracking ability of a Least Mean Squares (LMS) algorithm and a Sparse Least Mean Squares (LMS) algorithm.
  • FIGURE 1 illustrates a block diagram of an exemplary high definition television (HDTV) transmitter 100.
  • MPEG compatible data packets are encoded for forward error correction (FEC) by a Reed Solomon (RS) encoder 110.
  • FEC forward error correction
  • RS Reed Solomon
  • Trellis encoder unit 130 produces a stream of data symbols representing three (3) bits for each symbol. One of the three bits is pre- coded and the other two bits are produced by a four (4) state trellis encoding.
  • Trellis encoder unit 130 comprises twelve (12) parallel trellis encoder and pre-coder units to provide twelve interleaved coded data sequences.
  • the encoded three (3) bits of each trellis encoder and pre-coder unit are combined with field and segment synchronization bit sequences in multiplexer 140.
  • a pilot signal is inserted by pilot insertion unit 150.
  • the data stream is then subjected to vestigial sideband (VSB) suppressed carrier eight (8) level modulation by VSB modulator 160.
  • VSB modulator 160 vestigial sideband
  • the data stream is then finally up-converted to a radio frequency by radio frequency (RF) by converter 170.
  • RF radio frequency
  • FIGURE 2 illustrates a block diagram of an exemplary high definition television (HDTV) receiver 200.
  • the received RF signal is down-converted to an intermediate frequency (IF) by tuner 210.
  • the signal is then filtered and converted to digital form by IF filter and detector 220.
  • the detected signal is then in the form of a stream of data symbols that each signify a level in an eight (8) level constellation.
  • the signal is then filtered by NTSC rejection filter 230 and subjected to equalization and phase tracking by equalizer and phase tracker unit 240.
  • the recovered encoded data symbols are then subjected to trellis decoding by trellis decoder unit 250.
  • the decoded data symbols are then further de-interleaved by data de- interleaver 260.
  • the data symbols are then subjected to Reed-Solomon decoding, by Reed Solomon decoder 270. This recovers the MPEG compatible data packets transmitted by transmitter 100.
  • FIGURE 2 A schematic representation of computer diskette 280 is also shown in FIGURE 2.
  • computer diskette 280 may be inserted into a computer disk drive (not shown) in television receiver 200.
  • the computer disk drive is capable of receiving signal information relating to target weight vectors in an adaptive channel equalizer of television receiver 200 and writing the signal information to computer diskette 280.
  • computer diskette 280 contains computer executable method steps for implementing the method of the present invention.
  • Computer diskette 280 may be considered as a representation of any type of medium that is capable of storing and sending computerized data and instruction.
  • FIGURE 3 illustrates a block diagram of a prior art adaptive channel equalizer 300 for use in equalizer and phase tracker unit 240.
  • Prior art adaptive channel equalizer unit 300 comprises a forward equalizer (FE) filter 310 and a decision feedback equalizer (DFE) filter 320.
  • the output from forward equalizer (FE) filter 310 is added to the output from decision feedback equalizer (DFE) filter 320 in adder unit 330 to form the output of adaptive channel equalizer unit 300.
  • Forward equalizer (FE) filter 310 accepts the uncompensated channel symbol data as its input.
  • decision feedback equalizer (DFE) filter 320 requires for its input an "estimate" of the symbol that was transmitted over the channel before the symbol was corrupted by noise.
  • DFE filter 320 can receive an estimate of the output symbol by simply "slicing" the equalizer output.
  • the term “slicing” refers to the process of taking the allowed symbol value (of the eight (8) levels ' specified by the 8-VSB ATSC Standard) that is nearest to that of the actual output.
  • level slicer 340 provides the "sliced" symbols to DFE filter 320 through multiplexer 350. This method of providing estimates of the output symbols to DFE filter 320 can suffer from error propagation caused by slicing errors.
  • DFE filter 320 can be adapted either in a "trained mode” or in a "blind” mode.
  • DFE filter 320 receives a "training sequence" of known symbols (through multiplexer 350) at a certain known time.
  • DFE filter 320 compares the known training sequence with the "equalizer error for trained adaptation.”
  • the equalizer error for trained adaptation is obtained by subtracting the equalizer output from the known training sequence.
  • DFE filter 320 then adjusts its operation to cause the equalizer output to match the known sequence of training signals.
  • DFE filter 320 can operate in a "blind mode.”
  • DFE filter 320 receives an "equalizer error for blind adaptation" from blind error unit 360.
  • Blind error unit 360 compares the equalizer output with an expected statistical distribution of the data to generate the equalizer error blind adaptation.
  • DFE filter 320 then adjusts its operation to cause the equalizer output to match the expected statistical distribution of the data.
  • FIGURE 4 Illustrates a conventional adaptive finite impulse response (FIR) filter 400 for use in forward equalizer (FE) filter 310 and in decision feedback equalizer (DFE) filter 320.
  • the coefficients of FIR filter 400 are computed to compensate as much as possible for channel distortions.
  • the length of FIR filter 400 corresponds to the maximum impai ⁇ ment delay that FIR filter 400 is designed to correct for.
  • FIR filter 400 comprises a number of filter tap cells 410 (also referred to as "filter taps"). Each filter tap 410 comprises a data storage register 420, a coefficient storage register 430, and multiplier 440. The output of multiplier 440 is input to an adder unit 450. Adder unit 450 sums all of the weighted tap values to generate a filter output. Filter tap 410 also comprises a coefficient adaptation unit 460 that computes the updated filter coefficients. The coefficient adaptation unit 460 has the following inputs: (1) the current coefficient value, (2) the data tap value, and (3) a measure of the equalizer error (i.e., the difference between the expected signal value and the actual output signal value) . The coefficient adaptation unit 460 operates only when the adaptation process is being performed.
  • a commonly used method of computing the filter coefficients uses the well known least mean square (LMS) algorithm.
  • LMS least mean square
  • the LMS algorithm is a successive approximation technique that uses the current coefficient and data tap values as well as the equalizer error to compute the new coefficient value.
  • the LMS algorithm repeats the procedure until each filter coefficient converges to the desired optimum value.
  • FIGURE 5 illustrates a block diagram showing the connection of forward equalizer (FE) filter 310 to trellis decoder unit 250 and the connection of trellis decoder unit 250 to decision feedback equalizer (DFE) filter 320.
  • FE forward equalizer
  • the output from forward equalizer (FE) filter 310 is added to the output from decision feedback equalizer (DFE) filter 320 in adder unit 330 to form the input to trellis decoder unit 250.
  • the path memory outputs from trellis decoder unit 250 are fed back to decision feedback equalizer (DFE) filter 320.
  • Information from the path memory outputs is used to reduce errors in decision feedback equalizer (DFE) filter 320.
  • the apparatus and method of the present invention provides improved algorithms for computing adaptive equalizer coefficients.
  • the improved algorithms of the present invention may be used in a variety of different equalizer circuits. In the exemplary equalizer circuit illustrated in FIGURE 5, the improved algorithms of the present invention may be used to calculate the adaptive equalizer coefficients that are provided to decision feedback equalizer (DFE) 320. It is understood that the equalizer circuit illustrated in FIGURE 5 is shown as an example.
  • the improved algorithms of the present invention are not limited to use only in the equalizer circuit shown in FIGURE 5.
  • the algorithms of the present invention operate in equalizer circuits that are designed for sparse transmission channels.
  • Prior art equalizer circuits use a class of algorithms that are designed to exploit sparsity in target weight vectors.
  • the prior art algorithms are designed with the assumptions that (1) a training sequence is available, and (2) the weights are positive, and (3) the weights sum to unity.
  • the improved algorithms of the present invention do not make these assumptions. Therefore, the improved algorithms of the present invention are better able to exploit known sparsities in target weight vectors in an adaptive channel equalizer.
  • EG exponentiated gradient
  • the EG algorithm has recently been shown to have better performance than typical gradient methods when the target weight vector is sparse. See the article entitled “Exponentiated Gradient Versus Gradient Descent for Linear Predictors” by J. Kivinen and M. K. Warmuth in Information and Computation, 132(1), pp. 1-64, 1997.
  • the EG algorithm has several limitations that make it inappropriate for applications in communications and signal processing. For example, the EG algorithm do'es not decrease the required complexity.
  • the EG algorithm also makes the assumptions (1) through (3) mentioned above.
  • the preferential metric is a family of functions ⁇ ⁇ w') that represent the known "a priori" knowledge (the Bayesian "prior") of the i th parameter (or weight) w' .
  • the "natural gradient” (NG) algorithm is then shown to be:
  • Equation (1) the updates of the NG algorithm set forth in Equation (1) can be quite complicated due to the presence of the nonlinearities ⁇ and ⁇ "1 .
  • a more useful method can be derived as a first order approximation to Equation (1) by using the Taylor series, which results in the new update equation:
  • Equation (2) The method represented by Equation (2) is referred to as the "approximate natural gradient” (ANG) method. More detailed information on the derivation of the ANG method is provided in an article entitled “Exploiting Sparsity in Equalization Algorithms,” by R. K. Martin, W. A. Sethares, and R. C. Williamson, Cornell University Technical Report XX-2000.
  • the present invention comprises a new series of algorithms based on different choices of cost functions and priors .
  • a practical method of choosing appropriate priors for particular applications (such as an 8-VSB HDTV equalizer) will be discussed later.
  • Prior information about a target weight vector may arise from first principles. For instance, a channel may be known to be sparse because it consists of a number of discrete reflections well separated in space or time. Prior knowledge may also arise as a result of a statistical study showing, for example, that a particular equalizer often converges to a configuration in which most taps are near zero. However it arises, it will be necessary to model the "a priori" information in simple functional form in order to utilize it in algorithm design.
  • the strategy is to create a histogram of all of the equalizer coefficients of all of the gathered channel data. In accordance with the expectations of sparsity, this has the form of a small number of large taps and a large number of small taps. It is then necessary to fit a parameterized curve to the histogram. This function will be a good candidate for the prior ⁇ .
  • Equation (3) ⁇ (w) is large for small w and small for large w.
  • the term a quantifies how the histograms decay. Observe that the absolute values allow w to take on both positive and negative values, while the ⁇ term insures that the algorithm does not become trapped at or near zero.
  • ⁇ , ⁇ , and ⁇ (y,y) are an alternate set of priors, parameterizations and cost function for the same ANG algorithm.
  • This prior will be called the "true" prior because it represents the prior beliefs without the confounding influence of the reparameterization function.
  • can be used to give the cost function over which the algorithm is evolving under the standard Euclidean gradient. These are useful because sometimes it is easier to understand the behavior of an algorithm from the point of view of priors, while sometimes it is easier from the perspective of the cost function.
  • the requirement that the expression ⁇ 2 divided by ⁇ 2 be equal allows translation of the same algorithm to both frameworks .
  • Equation (9) where c is a constant appropriate for the given constellation.
  • the most useful aspect of Equation (9) is that it leads to algorithms that are independent of y k and that can be used to adapt the weights even in the absence of a training signal. Such algorithms are referred to as "blind” algorithms.
  • This section will show how the NG and ANG strategies, in combination with Equation (9) , can be used to derive blind algorithms in which the updates are explicitly designed to exploit prior knowledge about the weights .
  • Equation (2) leads to an ANG algorithm that is of exactly the same form.
  • Equation (10) is a standard CMA algorithm.
  • All of the algorithms can also be generalized readily to the complex case. All of the algorithms may be designed so that different weights have different priors and different update rules matched to those priors .
  • Equation (16) is a standard DD algorithm.
  • Equation (18) corresponds to a Euclidean descent on a cost function defined by Equation (15) with the reparameterization defined by Equation (6).
  • Equation (6) the reparameterization defined by Equation (6).
  • Other priors lead in analogous fashion to other update rules and to other modifications.
  • the exponential power of two (2) in Equation (15) may be replaced with some other exponential power.
  • FIGURE 6 sets forth experimental performance curves in several system identifications scenarios that compare the performance of the LMS algorithm to that of the Sparse LMS algorithm.
  • the comparison was made in terms of Mean Squared Error (MSE) , convergence rates, and tracking ability.
  • MSE Mean Squared Error
  • the stepsizes were chosen to equate the MSE after convergence.
  • the convergence rates of the two algorithms were compared via simulations.
  • the sparsity is exploited in a completely known setting in order to give an idea as to the kinds of improvements that are possible.
  • the simulations were run in a sparse environment.
  • the first channel had ten taps, with non-zero taps of values [0.1, 1.0, -0.5, 0.1] located in positions [1, 3, 4, 8].
  • the second channel had one hundred (100) taps, with not- zero taps of the same values located in positions [1, 30, 35, 85] .
  • Zero initializations were used.
  • the noise power was (0.0215) 2 and the value of ⁇ for Sparse LMS was one sixteenth ( 1 / 16 ) .
  • FIGURE 6 shows the MSE versus time for both the LMS algorithm and for the Sparse LMS algorithm.
  • the MSE was computed by taking the ensemble average of each algorithm over one hundred (100) runs. It is clear that if the environment is sufficiently sparse, the sparse version of LMS achieves a much faster convergence rate.
  • FIGURE 7 shows the tracking ability of the two algorithms.
  • the simulation was run on a twenty (20) tap channel with two (2) taps set to vary sinusoidally in time.
  • the actual channel consisted of the first channel from FIGURE 6 with ten (10) zeros appended. Then taps fifteen (15) and sixteen (16) were set to vary as 1 + 0.2 sin ( ⁇ / 256 k ) and 0.2 sin ( ⁇ / 256 k ), respectively, where k 2,7
  • the upper plot in FIGURE 7 shows the values of the actual taps and both estimates as both a small and a large tap fluctuate.
  • the Sparse LMS is better at tracking large taps, but not small taps. This is because its effective stepsize becomes small near zero.
  • the lower plot shows the MSE for both algorithms when only the larger tap is fluctuating. Since Sparse LMS is better at tracking the large tap, the MSE does not suffer as much as for regular LMS.

Landscapes

  • Engineering & Computer Science (AREA)
  • Signal Processing (AREA)
  • Multimedia (AREA)
  • Power Engineering (AREA)
  • Computer Networks & Wireless Communication (AREA)
  • Cable Transmission Systems, Equalization Of Radio And Reduction Of Echo (AREA)
  • Filters That Use Time-Delay Elements (AREA)

Abstract

An apparatus (320) and method is disclosed for using adaptive algorithms to exploit sparsity in target weight vectors in an adaptive channel equalizer (300). An adaptive algorithm comprises a selected value of a prior and a selected value of a cost function. The present invention comprises algorithms adapted for calculating adaptive equalizer coefficients for sparse transmission channels. The present invention provides sparse algorithms in the form of a Sparse Least Mean Squares (LMS) algorithm and a Sparse Constant Modulus Algorithm (CMA) and a Sparse Decision Directed (DD) algorithm.

Description

APPARATUS AND METHOD FOR USING ADAPTIVE ALGORITHMS
TO EXPLOIT SPARSITY IN TARGET WEIGHT VECTORS
IN AN ADAPTIVE CHANNEL EQUALIZER
REFERENCE TO PROVISIONAL APPLICATIONS
This patent application refers to and claims the priority and benefit of Provisional Patent Application
Serial No. 60/231,610 filed September 11, 2000 and
Provisional Patent Application Serial No. 60/262,506 filed January 18, 2001.
TECHNICAL FIELD OF THE INVENTION
The present invention is directed, in general, to adaptive channel equalizers in digital communication devices and, more specifically, to an apparatus and method for using algorithms to exploit sparsity in target weight vectors in an adaptive channel equalizer.
BACKGROUND OF THE INVENTION
The Digital High Definition Television (HDTV) Grand
Alliance (Grand Alliance) is a group of television manufacturing and research organizations in the television industry. After years of cooperative effort the Grand
Alliance developed and proposed a standard for digital HDTV systems. The Grand Alliance standard has been adopted
(with a few changes) by the Federal Communication Commission (FCC) as an official broadcasting standard for
HDTV. The standard is known as the Advanced Television
Systems Committee Digital Television Standard (the ATSC
Standard" ) .
The ATSC Standard for HDTV transmission over terrestrial broadcast channels uses a signal that consists of a sequence of twelve (12) independent time-multiplexed trellis-coded data streams modulated as an eight (8) level vestigial sideband (VSB) symbol stream with a rate of 10.76 MHz. This signal is converted to a six (6) MHz frequency band that corresponds to a standard VHF or UHF terrestrial television channel, over which the signal is then broadcast.
The ATSC Standard calls for two (2) bit data symbols of the HDTV signal to be trellis encoded in accordance with an eight (8) level (i.e., a three (3) bit) one dimensional constellation. One bit of each data symbol is pre-coded, and the other is subjected to a 1/2 encoding rate that produces two coded bits in accordance with a four (4) state trellis code. For purposes of interleaving, twelve (12) identical encoders and pre-coders operate successively on every twelve successive data symbols. Symbols 0, 12, 24, 36, . . . are encoded as one series. Symbols 1, 13, 25, 37, . . . as a second series. Symbols 2, 14, 26, 38, . . . as a third series. And so on for a total of twelve (12) series. Therefore, the ATSC Standard requires twelve (12) trellis decoders in the HDTV receiver for the twelve (12) series of time division interleaved data symbols in the signal. Each trellis decoder in the HDTV receiver decodes every twelfth (12th) data symbol in the stream of coded data symbols.
In an ATSC Standard receiver trellis decoders are used to retrieve the original digital data that was trellis encoded just before being converted to 8-VSB symbols, modulated and broadcast. The use of trellis coding provides an improvement in the signal to noise ratio of the received signal, and the time multiplexing of twelve (12) independent streams reduces the possibility of co-channel interference from an analog NTSC broadcast signal residing on the same frequency. The abbreviation NTSC stands for National Television Standards Committee.
Each of the trellis decoders for the four (4) state trellis code operates in accordance with the well-known Viterbi decoding algorithm. Each of the decoders comprises a branch metric generator unit, an add-compare-select unit, and a path-memory unit. See, for example, "Trellis-coded Modulation With Redundant Signal Set, Part I, Introduction; Part II, State of the Art," by G. Ungerboeck, IEEE Communications Magazine, Vol. 25, pp. 5-21, February 1987. In addition to being corrupted by noise, the transmitted signal is also subject to deterministic channel distortions and distortions caused by multipath interference. Consequently, an adaptive channel equalizer is generally used in front of the trellis decoders to compensate for these effects. The goal is to create a symbol stream that resembles, as much as possible, the symbol stream that was created by the twelve (12) trellis encoders at the transmitter.
One commonly used equalizer architecture makes use of a second equalizer known as a decision feedback equalizer (DFE) . In this architecture, a conventional, or forward equalizer (FE) is supplemented by a DFE. The input to the DFE is an estimate of the original transmitted value of the current output symbol of the complete equalizer (FE and DFE) . The output of the decision feedback equalizer (DFE) is subsequently added to the output of the forward equalizer (FE) to generate the output symbol. In a typical implementation, this estimate of the output symbol is obtained by simply "slicing" the equalizer output. The term "slicing" refers to the process of taking the allowed symbol value (of the eight (8) levels specified by the 8- VSB ATSC Standard) that is nearest to that of the actual output. Using the "sliced" symbols in a decision feedback equalizer (DFE) gives a near optimum error rate performance with low complexity. This approach, however, can suffer from error propagation caused by slicing errors. Because the typical symbol error rate after the equalizer for the HDTV signal can be up to twenty percent (20%) , this can be a serious problem if the number of DFE filter taps is large .
After the equalizer, the HDTV signal is decoded in a trellis decoder that uses the Viterbi algorithm to decode the symbol stream based on the 1/2 rate trellis coding performed in the transmitter. As previously mentioned, the ATSC Standard specifies that twelve (12) trellis encoders and decoders are used in parallel in a time multiplexed fashion. Trellis decoding is then followed by byte de- interleaving and Reed Solomon decoding to further correct transmission errors in the signal.
Various prior art algorithms exist for computing the filter coefficients for adaptive equalizers. One commonly used method uses the well known least mean square (LMS) algorithm. The LMS algorithm is a successive approximation technique that uses the current coefficient and data tap values as well as the equalizer error to compute the new coefficient value. The LMS algorithm repeats the procedure until each filter coefficient converges to the desired optimum value.
In a typical LMS algorithm the coefficient vector is determined using the following formula:
Cn+ι = Cn + μ E dn (2) where Cn is the coefficient vector at time n, μ is an adaptation speed constant, and dn is the data vector in the filter at time n. E is the error calculated from the output of the equalizer. E can be calculated in a decision directed fashion using the known training sequence embedded in the data stream. Alternatively, E can be calculated in a blind fashion using a Constant Modulus Algorithm (CMA) or a Decision Directed (DD) algorithm.
Transmission channels are often sparse. Optimally designed adaptive equalizers reflect the sparsity in the transmission channels. Typical prior art equalization methods such as the Least Mean Square (LMS) algorithm or the Constant Modulus Algorithm (CMA) or the Decision Directed (DD) algorithm do not exploit known "a priori" information for sparse transmission channels. Typical prior art approaches to exploiting sparsity attempt to reduce complexity. Complexity reduction is accomplished by only updating a subset of the channel model or equalizer taps. Recently a prior art "exponentiated gradient" (EG) algorithm has been shown to have better performance than typical gradient methods when the target weight vector is sparse. See the article entitled "Exponentiated Gradient Versus Gradient Descent for Linear Predictors" by J. Kivinen and M. K. Warmuth in Information and Computation, 132(1), pp. 1-64, 1997. Unfortunately, the EG algorithm has several limitations that make it inappropriate for applications in communications and signal processing. For example, the EG algorithm does not decrease the required complexity.
There is therefore a need in the art for an apparatus and method for providing adaptive equalization algorithms capable of exploiting sparsity in target weight vectors in adaptive channel equalizers in general and in ATSC VSB receivers in particular. SUMMARY OF THE INVENTION To address the deficiencies of the prior art mentioned above, the apparatus and method of the present invention uses improved algorithms to exploit sparsity in target weight vectors in an adaptive channel equalizer.
The present invention uses information concerning known sparsities in a communication channel to calculate adaptive equalizer coefficients to increase a coefficient convergence rate for equalizer coefficients. It is an object of the present invention to provide an apparatus and method for designing efficient adaptive channel equalizers.
It is another object of the present invention to provide an apparatus and method for using improved algorithms that are capable of exploiting sparsity in target weight vectors in an adaptive channel equalizer.
It is also an object of the present invention to provide an apparatus and method for providing a version of a least mean squares algorithm that is capable of exploiting sparsity in target weight vectors in an adaptive channel equalizer.
It is another object of the present invention to provide an apparatus and method for providing a version of a constant modulus algorithm that is capable of exploiting sparsity in target weight vectors in an adaptive channel equalizer.
It is also an object of the present invention to provide an apparatus and method for providing a version of a decision directed algorithm that is capable of exploiting sparsity in target weight vectors in an adaptive channel equalizer .
The foregoing has outlined rather broadly the features and technical advantages of the present invention so that those skilled in the art may better understand the detailed description of the invention that follows. Additional features and advantages of the invention will be described hereinafter that form the subject of the claims of the invention. Those skilled in the art should appreciate that they may readily use the conception and the specific embodiment disclosed as a basis for modifying or designing other structures for carrying out the same purposes of the present invention. Those skilled in the art should also realize that such equivalent constructions do not depart from the spirit and scope of the invention in its broadest form.
Before undertaking the Detailed Description of the Invention, it may be advantageous to set forth definitions of certain words and phrases used throughout this patent document: the terms "include" and "comprise" and derivatives thereof, mean inclusion without limitation; the term "or," is inclusive, meaning and/or; the phrases "associated with" and "associated therewith," as well as derivatives thereof, may mean to include, be included within, interconnect with, contain, be contained within, connect to or with, couple to or with, be communicable with, cooperate with, interleave, juxtapose, be proximate to, be bound to or with, have, have a property of, or the like; and the term "controller," "processor," or "apparatus" means any device, system or part thereof that controls at least one operation, such a device may be implemented in hardware, firmware or software, or some combination of at least two of the same. It should be noted that the functionality associated with any particular controller' may be centralized or distributed, whether locally or remotely. In particular, a controller may comprise one or more data processors, and associated input/output devices and memory, that execute one or more application programs and/or an operating system program. Definitions for certain words and phrases are provided throughout this patent document. Those of ordinary skill in the art should understand that in many, if not most instances, such definitions apply to prior uses, as well as future uses, of such defined words and phrases.
BRIEF DESCRIPTION OF THE DRAWINGS
For a more complete understanding of the present invention, and the advantages thereof, reference is now made to the following descriptions taken in conjunction with the accompanying drawings, wherein like numbers designate like objects, and in which:
FIGURE 1 illustrates a block diagram of an exemplary high definition television (HDTV) transmitter;
FIGURE 2 illustrates a block diagram of an exemplary high definition television (HDTV) receiver;
FIGURE 3 illustrates a block diagram of an adaptive channel equalizer comprising a forward equalizer (FE) filter and a decision feedback equalizer (DFE) filter;
FIGURE 4 illustrates a block diagram of an adaptive finite impulse response (FIR) filter for use in an adaptive channel equalizer;
FIGURE 5 illustrates a block diagram showing the connection of a forward equalizer (FE) filter to a trellis decoder unit and the connection of the trellis decoder unit to a decision feedback equalizer (DFE) filter;
FIGURE 6 illustrates a graph showing the Mean Squared Error (MSE) versus time for a Least Mean Squares (LMS) algorithm and for a Sparse Least Mean Squares (LMS) algorithm; and FIGURE 7 illustrates- a graph showing the tracking ability of a Least Mean Squares (LMS) algorithm and a Sparse Least Mean Squares (LMS) algorithm. DETAILED DESCRIPTION OF THE INVENTION FIGURES 1 through 7, discussed below, and the various embodiments used to describe the principles of the present invention in this patent document are by way of illustration only and should not be construed in any way to limit the scope of the invention. In the description of the exemplary embodiment that follows, the present invention is integrated into, or is used in connection, with, a high definition television system. Those skilled in the art will recognize that the exemplary embodiment of the present invention may easily be modified for use in other similar types of systems to use an improved algorithm of the present invention to exploit sparsity in target weight vectors in an adaptive channel equalizer. FIGURE 1 illustrates a block diagram of an exemplary high definition television (HDTV) transmitter 100. MPEG compatible data packets are encoded for forward error correction (FEC) by a Reed Solomon (RS) encoder 110. The data packets in successive segments of each data field are then interleaved by data interleaver 120, and the interleaved data packets are then further interleaved and encoded by trellis encoder unit 130. Trellis encoder unit 130 produces a stream of data symbols representing three (3) bits for each symbol. One of the three bits is pre- coded and the other two bits are produced by a four (4) state trellis encoding.
Trellis encoder unit 130 comprises twelve (12) parallel trellis encoder and pre-coder units to provide twelve interleaved coded data sequences. The encoded three (3) bits of each trellis encoder and pre-coder unit are combined with field and segment synchronization bit sequences in multiplexer 140. A pilot signal is inserted by pilot insertion unit 150. The data stream is then subjected to vestigial sideband (VSB) suppressed carrier eight (8) level modulation by VSB modulator 160. The data stream is then finally up-converted to a radio frequency by radio frequency (RF) by converter 170.
FIGURE 2 illustrates a block diagram of an exemplary high definition television (HDTV) receiver 200. The received RF signal is down-converted to an intermediate frequency (IF) by tuner 210. The signal is then filtered and converted to digital form by IF filter and detector 220. The detected signal is then in the form of a stream of data symbols that each signify a level in an eight (8) level constellation. The signal is then filtered by NTSC rejection filter 230 and subjected to equalization and phase tracking by equalizer and phase tracker unit 240. The recovered encoded data symbols are then subjected to trellis decoding by trellis decoder unit 250. The decoded data symbols are then further de-interleaved by data de- interleaver 260. The data symbols are then subjected to Reed-Solomon decoding, by Reed Solomon decoder 270. This recovers the MPEG compatible data packets transmitted by transmitter 100.
A schematic representation of computer diskette 280 is also shown in FIGURE 2. In an alternate advantageous embodiment of the present invention, computer diskette 280 may be inserted into a computer disk drive (not shown) in television receiver 200. The computer disk drive is capable of receiving signal information relating to target weight vectors in an adaptive channel equalizer of television receiver 200 and writing the signal information to computer diskette 280. In another alternate advantageous embodiment of the present invention, computer diskette 280 contains computer executable method steps for implementing the method of the present invention. Computer diskette 280 may be considered as a representation of any type of medium that is capable of storing and sending computerized data and instruction.
FIGURE 3 illustrates a block diagram of a prior art adaptive channel equalizer 300 for use in equalizer and phase tracker unit 240. Prior art adaptive channel equalizer unit 300 comprises a forward equalizer (FE) filter 310 and a decision feedback equalizer (DFE) filter 320. The output from forward equalizer (FE) filter 310 is added to the output from decision feedback equalizer (DFE) filter 320 in adder unit 330 to form the output of adaptive channel equalizer unit 300. Forward equalizer (FE) filter 310 accepts the uncompensated channel symbol data as its input. In contrast, decision feedback equalizer (DFE) filter 320 requires for its input an "estimate" of the symbol that was transmitted over the channel before the symbol was corrupted by noise.
As is well known, DFE filter 320 can receive an estimate of the output symbol by simply "slicing" the equalizer output. The term "slicing" refers to the process of taking the allowed symbol value (of the eight (8) levels ' specified by the 8-VSB ATSC Standard) that is nearest to that of the actual output. In the embodiment shown in FIGURE 3, level slicer 340 provides the "sliced" symbols to DFE filter 320 through multiplexer 350. This method of providing estimates of the output symbols to DFE filter 320 can suffer from error propagation caused by slicing errors.
As is also well known, DFE filter 320 can be adapted either in a "trained mode" or in a "blind" mode. In the "trained mode" DFE filter 320 receives a "training sequence" of known symbols (through multiplexer 350) at a certain known time. DFE filter 320 compares the known training sequence with the "equalizer error for trained adaptation." The equalizer error for trained adaptation is obtained by subtracting the equalizer output from the known training sequence. DFE filter 320 then adjusts its operation to cause the equalizer output to match the known sequence of training signals. Alternatively, DFE filter 320 can operate in a "blind mode." In the "blind mode" DFE filter 320 receives an "equalizer error for blind adaptation" from blind error unit 360. Blind error unit 360 compares the equalizer output with an expected statistical distribution of the data to generate the equalizer error blind adaptation. DFE filter 320 then adjusts its operation to cause the equalizer output to match the expected statistical distribution of the data.
FIGURE 4 Illustrates a conventional adaptive finite impulse response (FIR) filter 400 for use in forward equalizer (FE) filter 310 and in decision feedback equalizer (DFE) filter 320. The coefficients of FIR filter 400 are computed to compensate as much as possible for channel distortions. The length of FIR filter 400 corresponds to the maximum impai±ment delay that FIR filter 400 is designed to correct for.
FIR filter 400 comprises a number of filter tap cells 410 (also referred to as "filter taps"). Each filter tap 410 comprises a data storage register 420, a coefficient storage register 430, and multiplier 440. The output of multiplier 440 is input to an adder unit 450. Adder unit 450 sums all of the weighted tap values to generate a filter output. Filter tap 410 also comprises a coefficient adaptation unit 460 that computes the updated filter coefficients. The coefficient adaptation unit 460 has the following inputs: (1) the current coefficient value, (2) the data tap value, and (3) a measure of the equalizer error (i.e., the difference between the expected signal value and the actual output signal value) . The coefficient adaptation unit 460 operates only when the adaptation process is being performed. A commonly used method of computing the filter coefficients uses the well known least mean square (LMS) algorithm. The LMS algorithm is a successive approximation technique that uses the current coefficient and data tap values as well as the equalizer error to compute the new coefficient value. The LMS algorithm repeats the procedure until each filter coefficient converges to the desired optimum value.
In a typical LMS algorithm the coefficient vector is determined using the following formula: Cn+ι = Cn + μ E -dn
(2) where Cn is the coefficient vector at time n, μ is an adaptation speed constant, and dn is the data vector in the filter at time n. E is the error calculated from the output of the equalizer. E can be calculated in a decision directed fashion using the known training sequence embedded in the data stream. Alternatively, E can be calculated in a blind fashion using a Constant Modulus Algorithm (CMA) or a Decision Directed (DD) algorithm. FIGURE 5 illustrates a block diagram showing the connection of forward equalizer (FE) filter 310 to trellis decoder unit 250 and the connection of trellis decoder unit 250 to decision feedback equalizer (DFE) filter 320. The output from forward equalizer (FE) filter 310 is added to the output from decision feedback equalizer (DFE) filter 320 in adder unit 330 to form the input to trellis decoder unit 250. The path memory outputs from trellis decoder unit 250 are fed back to decision feedback equalizer (DFE) filter 320. Information from the path memory outputs is used to reduce errors in decision feedback equalizer (DFE) filter 320. The apparatus and method of the present invention provides improved algorithms for computing adaptive equalizer coefficients. The improved algorithms of the present invention may be used in a variety of different equalizer circuits. In the exemplary equalizer circuit illustrated in FIGURE 5, the improved algorithms of the present invention may be used to calculate the adaptive equalizer coefficients that are provided to decision feedback equalizer (DFE) 320. It is understood that the equalizer circuit illustrated in FIGURE 5 is shown as an example. The improved algorithms of the present invention are not limited to use only in the equalizer circuit shown in FIGURE 5.
The algorithms of the present invention operate in equalizer circuits that are designed for sparse transmission channels. Prior art equalizer circuits use a class of algorithms that are designed to exploit sparsity in target weight vectors. However, the prior art algorithms are designed with the assumptions that (1) a training sequence is available, and (2) the weights are positive, and (3) the weights sum to unity.
The improved algorithms of the present invention do not make these assumptions. Therefore, the improved algorithms of the present invention are better able to exploit known sparsities in target weight vectors in an adaptive channel equalizer.
One prior art algorithm is an "exponentiated gradient" (EG) algorithm. The EG algorithm has recently been shown to have better performance than typical gradient methods when the target weight vector is sparse. See the article entitled "Exponentiated Gradient Versus Gradient Descent for Linear Predictors" by J. Kivinen and M. K. Warmuth in Information and Computation, 132(1), pp. 1-64, 1997. Unfortunately, the EG algorithm has several limitations that make it inappropriate for applications in communications and signal processing. For example, the EG algorithm do'es not decrease the required complexity. In addition, the EG algorithm also makes the assumptions (1) through (3) mentioned above.
The background of the improved algorithms of the present invention may be understood by reviewing the work of R. E. Mahony and R. C. Williamson, published in an article entitled "Riemannian Structure of Some New Gradient
Descent Learning Algorithms", Proceedings of IEEE 2000 Symposium on Adaptive Systems for Signal Processing, Communication and Control, pp. 197-202, 2000. This paper provides a general discussion concerning methods for encoding prior knowledge into learning algorithms using a geometric "preferential structure." The main concept involves defining a metric so that the algorithm evolves over an error surface that is shaped to incorporate the known prior information. For instance, if the ith component is known to be reliable while the jth component is not, then the algorithm should take larger steps in the jth direction. This is accomplished by warping the underlying space, in effect augmenting or diminishing the stepsizes of the algorithm in the desired direction.
Mathematically, the preferential metric is a family of functions φ ι w') that represent the known "a priori" knowledge (the Bayesian "prior") of the ith parameter (or weight) w' . The "natural gradient" (NG) algorithm is then shown to be:
dL 1
*L= ®?M) -
H ΦM) (1)
where L represents the "cost function" that is to be minimized, and where Φ is the indefinite integral of φ .
Unfortunately, the updates of the NG algorithm set forth in Equation (1) can be quite complicated due to the presence of the nonlinearities Φ and Φ"1. A more useful method can be derived as a first order approximation to Equation (1) by using the Taylor series, which results in the new update equation:
dL 1
Figure imgf000018_0001
(2)
The method represented by Equation (2) is referred to as the "approximate natural gradient" (ANG) method. More detailed information on the derivation of the ANG method is provided in an article entitled "Exploiting Sparsity in Equalization Algorithms," by R. K. Martin, W. A. Sethares, and R. C. Williamson, Cornell University Technical Report XX-2000.
By making suitable choices for the cost function L and the priors φ , it is possible to create adaptive methods specifically designed to exploit the sparsity and information structure of the weights. The present invention comprises a new series of algorithms based on different choices of cost functions and priors . A practical method of choosing appropriate priors for particular applications (such as an 8-VSB HDTV equalizer) will be discussed later.
Prior information about a target weight vector may arise from first principles. For instance, a channel may be known to be sparse because it consists of a number of discrete reflections well separated in space or time. Prior knowledge may also arise as a result of a statistical study showing, for example, that a particular equalizer often converges to a configuration in which most taps are near zero. However it arises, it will be necessary to model the "a priori" information in simple functional form in order to utilize it in algorithm design.
Suppose that a database of target weight vectors is available. The strategy is to create a histogram of all of the equalizer coefficients of all of the gathered channel data. In accordance with the expectations of sparsity, this has the form of a small number of large taps and a large number of small taps. It is then necessary to fit a parameterized curve to the histogram. This function will be a good candidate for the prior φ .
Suppose, for instance, that a prior form of:
Figure imgf000020_0001
appears to fit well. In Equation (3), φ(w) is large for small w and small for large w. The term a quantifies how the histograms decay. Observe that the absolute values allow w to take on both positive and negative values, while the ε term insures that the algorithm does not become trapped at or near zero.
The following relationship between the priors φ and the cost function L may be demonstrated.
Proposition One. Let φ , γ , and L ,y) represent the priors, parameterizations and cost function of an ANG algorithm as set forth in Equation (2) wherein y is parameterized as:
Figure imgf000020_0002
where xk is the measured input data, and the function γ(wk l ) is invertible (at least over its domain) . If there are functions γ and φ with: τ-2 ϊ 7 φ 2 φ (5)
then γ , φ , and ∑ (y,y) are an alternate set of priors, parameterizations and cost function for the same ANG algorithm.
In particular, the ANG algorithm may be derivable from an alternative prior φ using the standard parameterization γ(w) = w . This prior will be called the "true" prior because it represents the prior beliefs without the confounding influence of the reparameterization function. Alternatively, the ANG algorithm may be derivable from a reparameterization using the standard prior φ = 1 (which corresponds to a belief that all values of the parameter are equally likely) . In this case, γ can be used to give the cost function over which the algorithm is evolving under the standard Euclidean gradient. These are useful because sometimes it is easier to understand the behavior of an algorithm from the point of view of priors, while sometimes it is easier from the perspective of the cost function. The requirement that the expression γ 2 divided by φ 2 be equal allows translation of the same algorithm to both frameworks .
Therefore the prior function in Equation (3) with γ (w) = w and = 1 corresponds to a cost function L with parameterization
γ{w) = - sgn(w) W[ + Λ/6" W (6)
and a prior of φ (w) = 1 . Either point of view is valid and both lead to the same update rule. The first point of view is needed to easily determine what the prior ought to be, while the latter is needed to easily write down the appropriate ANG algorithm.
This section derives and describes the improved algorithms of the present invention. In most cases, the ANG algorithm set forth in Equation (2), and the cost function for the equivalent Euclidean gradient (corresponding to φ = 1), and the true prior (corresponding to γ (w) = w ) are given. The new algorithms include a version of LMS specifically designed for sparse applications, as well as blind CMA and DD algorithms that similarly exploit prior information about the target weight vector.
ALGORITHM ONE. SPARSE LMS. Using the standard MSE cost function:
Figure imgf000022_0001
(7) with the parameterization set forth in Equation (6), and with prior φ (w) = 1 gives the ANG algorithm:
wi+ι = + μ (yk - yk) 4 (|wi| + * )
(8)
When =1 the ANG algorithm in Equation (8) will also be obtained using the prior set forth in Equation (3) and the cost function γ (w) = w . Observe that the only difference between the ANG algorithm in Equation (8) and the standard LMS is the presence of the term wk' multiplying the stepsize. In essence, when the current estimate of wk' is small, the update is small, and when the current estimate of wk' is large, the update is large. This explains in a simple, intuitive way, why the algorithm will tend to perform better in sparse environments. It takes large steps when the data appears unreliable (i.e., when it differs from the prior) and it takes small steps when the data appears to be reliable (i.e., when it reinforces the prior) . All of the above derivations occur for individual weights. Hence, if different priors are available for different weights, then different algorithm updates can be used. This might be useful, for example, in an equalizer design problem where the center taps are likely to be large while the tails of the equalizer are likely to be small and sparse.
ALGORITHM TWO. BLIND ADAPTATION WITH SPARSE CMA In many communications problems the transmitted signal consists of elements taken from a finite alphabet. This property can be exploited to direct the adaptation. One popular method is the Constant Modulus Algorithm (CMA) with cost function
Figure imgf000023_0001
(9) where c is a constant appropriate for the given constellation. The most useful aspect of Equation (9) is that it leads to algorithms that are independent of yk and that can be used to adapt the weights even in the absence of a training signal. Such algorithms are referred to as "blind" algorithms. This section will show how the NG and ANG strategies, in combination with Equation (9) , can be used to derive blind algorithms in which the updates are explicitly designed to exploit prior knowledge about the weights . The simplest case is the uniform prior φ (w) = 1 which corresponds to Φ (w) = 1 and Φ"1 (z) = z . This leads to the NG algorithm:
Figure imgf000024_0001
(10)
Using Equation (2) leads to an ANG algorithm that is of exactly the same form. In fact, Equation (10) is a standard CMA algorithm.
Suppose, however, that prior knowledge suggests a prior as set forth in Equation (3) . Then the ANG algorithm becomes:
w k, +1 W μy y k ~c) OI +ε )
( 11 )
Of course, other priors will lead in similar fashion to other update rules.
Variations on L(ylc,yk) may also be made. One such variation generalizes Equation (9) to:
Figure imgf000024_0002
(12) where P and q take on various values. This can be combined with prior information leading to a generalized CMA for sparsity in its NG form or in its simpler ANG form. The simpler ANG form is:
Figure imgf000024_0003
( a + ε
( 13 ) with the true prior:
φ(w) = w\ + ε
(14) All of the algorithms can also be generalized readily to the complex case. All of the algorithms may be designed so that different weights have different priors and different update rules matched to those priors .
ALGORITHM THREE. BLIND ADAPTATION WITH SPARSE DD Another important class of algorithms are "decision directed", blind algorithms designed for use with a finite alphabet. The DD algorithm can be viewed as Euclidean descent over the cost function
Figure imgf000025_0001
(15) where the function Q quantizes the argument to the nearest symbol in the alphabet. In such a derivation it is necessary to ignore the fact that the function Q is discontinuous and to formally replace its derivative with zero. The simplest case is the uniform prior φ (w) = 1 and standard parameterization φ (w) = w . This leads to the NG algorithm:
Figure imgf000025_0002
(16)
Using Equation (2) leads to an ANG algorithm that has the same form. Equation (16) is a standard DD algorithm. For the fractional power priors
Figure imgf000026_0001
(17) and the standard parameterization, the ANG algorithm is
Figure imgf000026_0002
(is;
Equation (18) corresponds to a Euclidean descent on a cost function defined by Equation (15) with the reparameterization defined by Equation (6). Other priors lead in analogous fashion to other update rules and to other modifications. For example, the exponential power of two (2) in Equation (15) may be replaced with some other exponential power. ALGORITHM PERFORMANCE.
A fair amount of theoretical evidence exists that this kind of exploitation of prior information leads to more rapid convergence. See, for example, R. E. Mahony and R. C. Williamson, "Riemannian Structure of Some New Gradient Descent Learning Algorithms, Proceedings of IEEE 2000 Symposium on Adaptive Systems for Signal Processing, Communication and Control, pp. 197-202, 2000, and S. I. Hill, and R. C. Williamson, "Convergence of Exponentiated Gradient Algorithms," submitted September 1999 for publication in IEEE Transactions on Signal Processing.
FIGURE 6 sets forth experimental performance curves in several system identifications scenarios that compare the performance of the LMS algorithm to that of the Sparse LMS algorithm. The comparison was made in terms of Mean Squared Error (MSE) , convergence rates, and tracking ability. To ensure that the comparisons were fair, the stepsizes were chosen to equate the MSE after convergence. The convergence rates of the two algorithms were compared via simulations. The sparsity is exploited in a completely known setting in order to give an idea as to the kinds of improvements that are possible.
The simulations were run in a sparse environment. The first channel had ten taps, with non-zero taps of values [0.1, 1.0, -0.5, 0.1] located in positions [1, 3, 4, 8]. The second channel had one hundred (100) taps, with not- zero taps of the same values located in positions [1, 30, 35, 85] . Zero initializations were used. The stepsizes were chosen so that the MSE at convergence would be the same, leading to μ = 0.0050, μSparse = 0.0215 for channel No. 1 and μsparse = 0.0629 for channel No. 2. The noise power was (0.0215)2 and the value of ε for Sparse LMS was one sixteenth ( 1 / 16 ) . FIGURE 6 shows the MSE versus time for both the LMS algorithm and for the Sparse LMS algorithm. The MSE was computed by taking the ensemble average of each algorithm over one hundred (100) runs. It is clear that if the environment is sufficiently sparse, the sparse version of LMS achieves a much faster convergence rate.
FIGURE 7 shows the tracking ability of the two algorithms. The simulation was run on a twenty (20) tap channel with two (2) taps set to vary sinusoidally in time.
The actual channel consisted of the first channel from FIGURE 6 with ten (10) zeros appended. Then taps fifteen (15) and sixteen (16) were set to vary as 1 + 0.2 sin ( π / 256 k ) and 0.2 sin ( π / 256 k ), respectively, where k 2,7
is the iteration number. Again the stepsizes were chosen to equalize the MSE.
The upper plot in FIGURE 7 shows the values of the actual taps and both estimates as both a small and a large tap fluctuate. The Sparse LMS is better at tracking large taps, but not small taps. This is because its effective stepsize becomes small near zero. The lower plot shows the MSE for both algorithms when only the larger tap is fluctuating. Since Sparse LMS is better at tracking the large tap, the MSE does not suffer as much as for regular LMS.
The LMS algorithm and a wide range of variants have been derived through the framework of the Natural Gradient class of algorithms. In a similar fashion, other algorithms (such as CMA and DD-LMS) and their variants have been derived. Using the concepts of reparameterization and priors, it has been shown how to exploit prior knowledge of the system' s distribution, with particular attention to the case of a sparse distribution of taps. In cases where prior knowledge exists (such as in the 8-VSB equalizer problem) there is compelling evidence for the adoption of algorithms specially designed to exploit that knowledge.
While the present invention has been described in detail with respect to certain embodiments thereof, those skilled in the art should understand that they can make various changes, substitutions modifications, alterations, and adaptations in the present invention without departing from the concept and scope of the invention in its broadest form.

Claims

WHAT IS CLAIMED IS:
1. An apparatus (240) for calculating adaptive equalizer coefficients that uses an algorithm that is capable of exploiting sparsities in target weight vectors in an adaptive channel equalizer (240) to increase a convergence rate of at least one adaptive equalizer coefficient .
2. The apparatus (240) as claimed in Claim 1 wherein said algorithm is capable of utilizing one of: a selected value of a prior and a selected value of a cost function.
3. The apparatus (240) as claimed in Claim 1 wherein said algorithm is a sparse least mean squares algorithm of the form:
W k + 1 = W, + μ (yk - yk) 4 ( wi + £ ) where wk' is the i weight vector at time k, and where w[ th k+1 is the i weight vector at time k+1, and where μ is the stepsize, and where (yk - yk ) is a measured error at time k, and where xk' is the ith value of received signal at time k, and where ( wk' + ε ) is a measure of sparsity contribution specified by a cost function.
4. The apparatus (240) as claimed in Claim 1 wherein said algorithm is a sparse constant modulus algorithm of the form:
Figure imgf000029_0001
where w). is the i weight vector at time k, and where wk'+1 is the ith weight vector at time k+1, and where μ is the stepsize, and where yk is a received signal at time k after passing through filter w , and where (yk 2 - c ) is a calculated error at time k, and where xk' is the ith value of
received signal at time k, and where ( wk + εw ) is a measure of sparsity contribution specified by a cost function.
5. The apparatus (240) as claimed in Claim 1 wherein said algorithm is a sparse decision directed algorithm of the form:
wi+ι = wk' ~ (20>*) -;?* )* ( f* +*) where w). is the ith weight vector at time k, and where wω' is the ith weight vector at time k+1, and where § is the stepsize, and where yk is a received signal at time k after passing through filter w , and where Q(yk ) is a quantization of yk , and where ( Q(yk) - yk ) is a calculated error at time k, and where xk' is the ith value of received signal at time
k, and where ( wk + ε ) is a measure of sparsity contribution specified by a cost function.
6. A television receiver (200) capable of calculating adaptive equalizer coefficients that uses an algorithm that is capable of exploiting sparsities in target weight vectors in an adaptive channel equalizer (240) to increase a convergence rate of at least one adaptive equalizer coefficient.
7. The television receiver (200) as claimed in Claim 6 wherein said algorithm is capable of utilizing one of: a selected value of a prior and a selected value of a cost function.
8 . The television receiver (200 ) as claimed in Claim 6 wherein said algorithm is a sparse least mean squares algorithm of the form :
W k + 1 + (yk - yk) 4 Ol + ε ) where wk' is the ith weight vector at time k, and where w 4.+1 is the ith weight vector at time k+1, and where μ is the stepsize, and where (yk - yk ) is a measured error at time k, and where xk' is the ith value of received signal at time k, and where ( wk + ε ) is a measure of sparsity contribution specified by a cost function.
9. The television receiver (200) as claimed in Claim 6 wherein said algorithm is a sparse constant modulus algorithm of the form: a
W k, +1 = "wi- JXΛ -c) 4 (Wk +sw) where wk' is the i weight vector at time k, and where w k+1 is the ith weight vector at time k+1, and where μ is the stepsize, and where yk is a received signal at time k after passing through filter w , and where
Figure imgf000031_0001
- c ) is a calculated error at time k, and where xk' is the ith value of
received signal at time k, and where ( Wk\ + εw ) is a measure of sparsity contribution specified by a cost function.
10. The television receiver (200) as claimed in Claim 6 wherein said algorithm is a sparse decision directed algorithm of the form:
Figure imgf000032_0001
where wk l is the ith weight vector at time k, and where wk l +1 is the ith weight vector at time k+1, and where μ is the stepsize, and where yk is a received signal at time k after passing through filter w , and where Q(yk ) is a quantization of yk , and where ( Q(yk) - yk ) is a calculated error at time k, and where xk' is the ith value of received signal at time a k, and where ( wk + ε ) is a measure of sparsity contribution specified by a cost function.
11. A method for calculating adaptive equalizer coefficients comprising the steps of: calculating at least one adaptive equalizer coefficient with an algorithm that is capable of exploiting sparsities in target weight vectors in an adaptive channel equalizer (240) ; and increasing a convergence rate of at least one adaptive equalizer coefficient.
12. The method as claimed in Claim 11 further comprising the step o : providing said algorithm with one of : a selected value of a prior and a selected value of a cost function.
13. The method as claimed in Claim 11 wherein said algorithm is a sparse least mean squares algorithm of the form:
w ι = + μ (y k - y k ) 4 (| | + * ) where wk' is the ith weight vector at time k, and where w k+1 is the ith weight vector at time k+1, and where μ is the stepsize, and where (yk — yk ) is a measured error at time k, and where xk' is the ith value of received signal at time k, and where ( wk' + ε ) is a measure of sparsity contribution specified by a cost function.
14. The method as claimed in Claim 11 wherein said algorithm is a sparse constant modulus algorithm of the form:
Figure imgf000033_0001
where wk' is the ith weight vector at time k, and where wk'+1 is the ith weight vector at time k+1, and where μ is the stepsize, and where yk is a received signal at time k after passing through filter w , and where
Figure imgf000033_0002
- c ) is a calculated error at time k, and where xk' is the ith value of
received signal at time k, and where ( wk + εw ) is a measure of sparsity contribution specified by a cost function.
15. The method as claimed in Claim 11 wherein said algorithm is a sparse decision directed algorithm of the form:
X+ι = ~ μ (Q(yk) -yk)x[ (|w* |α + ε) where wk' is the ith weight vector at time k, and where wk'+l is the ith weight vector at time k+1, and where μ is the stepsize, and where yk is a received signal at time k after passing through filter w , and where Q(yk ) is a quantization of yk , and where ( Q(yk) - yk ) is a calculated error at time k, and where xk' is the ith value of received signal at time
k, and where ( wk + ε ) is a measure of sparsity contribution specified by a cost function.
16. Computer executable process steps, stored on a computer readable storage medium (280), for calculating adaptive equalizer coefficients comprising the steps of: calculating at least one adaptive equalizer coefficient with an algorithm that is capable of exploiting sparsities in target weight vectors in an adaptive channel equalizer (240) ; and increasing a convergence rate of at least one adaptive equalizer coefficient.
17. The computer executable process steps, stored on a computer readable storage medium (280) , as claimed in
Claim 16 further comprising the step of: providing said algorithm with one of: a selected value of a prior and a selected value of a cost function.
18. The computer executable process steps, stored on a computer readable storage medium (280) , as claimed in Claim 16 wherein said algorithm is a sparse least mean squares algorithm of the form:
Figure imgf000035_0001
where wk' is the i weight vector at time k, and where w k.+1 is the i weight vector at time k+1, and where μ is the stepsize, and where (yk - yk ) is a measured error at time k, and where xk' is the ith value of received signal at time k, and where ( wk' + ε ) is a measure of sparsity contribution specified by a cost function.
19. The computer executable process steps, stored on a computer readable storage medium (280) , as claimed in Claim 16 wherein said algorithm is a sparse constant modulus algorithm of the form:
+ι = **i -i"λ (?t -c) x k' (|w(|a +£w) where wk' is the ith weight vector at time k, and where wM' is the ith weight vector at time k+1, and where μ is the stepsize, and where yk is a received signal at time k after passing through filter w , and where (yk 2 - c ) is a calculated error at time k, and where xk' is the ith value of
received signal at time k, and where ( wk' + εw ) is a measure of sparsity contribution specified by a cost function.
20. The computer executable process steps, stored on a computer readable storage medium (280), as claimed in Claim 16 wherein said algorithm is a sparse decision directed algorithm of the form:
wi+ι
Figure imgf000036_0001
where wk' is the ith weight vector at time k, and where wk'+l is the ith weight vector at time k+1, and where μ is the stepsize, and where yk is a received signal at time k after passing through filter w , and where Q(yk ) is a quantization of yk , and where ( Q(yk) — yk ) is a calculated error at time k, and where xk' is the ith value of received
signal at time k, and where ( w'Λ + ε ) is a measure of sparsity contribution specified by a cost function.
PCT/US2001/028758 2000-09-11 2001-09-10 Apparatus and method for using adaptive algorithms to exploit sparsity in target weight vectors in an adaptive channel equalizer WO2002023842A1 (en)

Priority Applications (3)

Application Number Priority Date Filing Date Title
JP2002527158A JP2004509521A (en) 2000-09-11 2001-09-10 Apparatus and method for utilizing the weight vector sparsity of an adaptive channel equalizer using an adaptive algorithm
EP01970985A EP1319289A1 (en) 2000-09-11 2001-09-10 Apparatus and method for using adaptive algorithms to exploit sparsity in target weight vectors in an adaptive channel equalizer
KR1020027006069A KR20020096041A (en) 2000-09-11 2001-09-10 Apparatus and method for using adaptive algorithms to exploit sparsity in target weight vectors in an adaptive channel equalizer

Applications Claiming Priority (4)

Application Number Priority Date Filing Date Title
US23161000P 2000-09-11 2000-09-11
US60/231,610 2000-09-11
US26250601P 2001-01-18 2001-01-18
US60/262,506 2001-01-18

Publications (1)

Publication Number Publication Date
WO2002023842A1 true WO2002023842A1 (en) 2002-03-21

Family

ID=26925270

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/US2001/028758 WO2002023842A1 (en) 2000-09-11 2001-09-10 Apparatus and method for using adaptive algorithms to exploit sparsity in target weight vectors in an adaptive channel equalizer

Country Status (6)

Country Link
US (1) US7061977B2 (en)
EP (1) EP1319289A1 (en)
JP (1) JP2004509521A (en)
KR (1) KR20020096041A (en)
CN (1) CN1395780A (en)
WO (1) WO2002023842A1 (en)

Cited By (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2004080789A (en) * 2002-08-16 2004-03-11 Samsung Electronics Co Ltd Equalizer and equalization method to be used for high definition tv
JP2004302470A (en) * 2003-03-31 2004-10-28 Microsoft Corp Method of noise estimation using incremental bayes learning
EP1492289A2 (en) * 2003-06-25 2004-12-29 Texas Instruments Incorporated Decision feedback equalization for high speed serial links
RU2477551C1 (en) * 2011-11-01 2013-03-10 Общество с ограниченной ответственностью "Специальный Технологический Центр" Method for multichannel adaptive reception of radio signals and apparatus for realising said method

Families Citing this family (28)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US7027500B1 (en) * 2000-12-12 2006-04-11 Ati Research, Inc. Linear prediction based initialization of a single-axis blind equalizer for VSB signals
US7321642B2 (en) * 2002-04-05 2008-01-22 Micronas Semiconductors, Inc. Synchronization symbol re-insertion for a decision feedback equalizer combined with a trellis decoder
US7385913B2 (en) * 2002-04-24 2008-06-10 Motorola, Inc. Method and apparatus for compensating for variations in a receive portion of a wireless communication device
US7450518B2 (en) * 2003-03-12 2008-11-11 Broadcom Corporation Sparse channel dual-error tracking adaptive filter/equalizer
US7739320B2 (en) * 2004-05-19 2010-06-15 Panasonic Corporation Waveform equalizer, waveform equalization method, and integrated circuit
GB2425011A (en) * 2005-04-07 2006-10-11 Ely Jay Malkin Encoding video data using a transformation function
CN100393115C (en) * 2005-08-05 2008-06-04 上海明波通信技术有限公司 Channel self-adapting equilibrium method for digital television
KR100703124B1 (en) 2006-06-16 2007-04-09 한국산업기술대학교산학협력단 Power amp
CN100433795C (en) * 2006-09-01 2008-11-12 上海大学 Method for image noise reduction based on transforming domain mathematics morphology
US7616685B2 (en) * 2007-01-19 2009-11-10 Techwell, Inc. Method for channel tracking in an LMS adaptive equalizer for 8VSB
US8385397B2 (en) * 2007-01-19 2013-02-26 Techwell Llc Method for determining the step size for an LMS adaptive equalizer for 8VSB
EP1956782A1 (en) * 2007-02-08 2008-08-13 Abb Research Ltd. Method and apparatus for adaptive blind equalization
US8144759B2 (en) * 2007-05-04 2012-03-27 University Of Central Florida Research Foundation, Inc. Adaptive methods employing optimal convergence factors for processing complex signals and systems
KR100864008B1 (en) * 2007-05-10 2008-10-16 주식회사 대우일렉트로닉스 Adaptive crosstalk cancellation method for 3d audio
WO2008144004A1 (en) * 2007-05-16 2008-11-27 Thomson Licensing Apparatus and method for encoding and decoding signals
MX2010004146A (en) * 2007-10-15 2010-05-17 Thomson Licensing Apparatus and method for encoding and decoding signals.
KR101513283B1 (en) * 2007-10-15 2015-04-17 톰슨 라이센싱 Apparatus and method for communicating burst mode activity
WO2009079815A1 (en) * 2007-12-10 2009-07-02 Zte Corporation A method for optimizing the radio frequency performance of the terminal
CN101478349B (en) * 2009-01-20 2013-05-08 南京信息工程大学 Dual mode variant step blind balance method based on orthogonal wavelet packet transform
EP2466580A1 (en) * 2010-12-14 2012-06-20 Fraunhofer-Gesellschaft zur Förderung der Angewandten Forschung e.V. Encoder and method for predictively encoding, decoder and method for decoding, system and method for predictively encoding and decoding and predictively encoded information signal
WO2015013058A1 (en) * 2013-07-24 2015-01-29 Mh Acoustics, Llc Adaptive beamforming for eigenbeamforming microphone arrays
US9369213B1 (en) 2015-03-09 2016-06-14 Alcatel Lucent Demultiplexing processing for a receiver
US10129053B2 (en) * 2015-05-28 2018-11-13 Multiphy Ltd. Steepest descent FFE computation and tracking
CN106253878B (en) * 2016-07-19 2018-10-12 重庆邮电大学 A kind of adaptive combined filtering method under impact noise
US10902837B2 (en) 2017-08-25 2021-01-26 The Regents Of The University Of California Sparsity-aware adaptive feedback cancellation
US11696083B2 (en) 2020-10-21 2023-07-04 Mh Acoustics, Llc In-situ calibration of microphone arrays
US11223447B1 (en) * 2021-03-31 2022-01-11 Seagate Technology Llc Multiple detector data channel and data detection utilizing different cost functions
CN113347230B (en) * 2021-05-13 2022-09-06 长沙星融元数据技术有限公司 Load balancing method, device, equipment and medium based on programmable switch

Citations (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
EP0717559A2 (en) * 1994-12-14 1996-06-19 Daewoo Electronics Co., Ltd Equalization apparatus with effective coefficient updating operation
EP1014635A2 (en) * 1998-12-22 2000-06-28 Thomson Consumer Electronics, Inc. DFE with blind, decision directed and data directed modes

Family Cites Families (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US4027257A (en) * 1976-06-01 1977-05-31 Xerox Corporation Frequency domain automatic equalizer having logic circuitry
US4141072A (en) * 1976-12-28 1979-02-20 Xerox Corporation Frequency domain automatic equalizer using minimum mean square error correction criteria
US5526378A (en) * 1994-12-14 1996-06-11 Thomson Consumer Electronics, Inc. Blind multipath correction for digital communication channel
US6426972B1 (en) * 1998-06-19 2002-07-30 Nxtwave Communications Reduced complexity equalizer for multi mode signaling
US6477200B1 (en) * 1998-11-09 2002-11-05 Broadcom Corporation Multi-pair gigabit ethernet transceiver
US7075967B2 (en) * 2001-01-19 2006-07-11 Raze Technologies, Inc. Wireless communication system using block filtering and fast equalization-demodulation and method of operation

Patent Citations (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
EP0717559A2 (en) * 1994-12-14 1996-06-19 Daewoo Electronics Co., Ltd Equalization apparatus with effective coefficient updating operation
EP1014635A2 (en) * 1998-12-22 2000-06-28 Thomson Consumer Electronics, Inc. DFE with blind, decision directed and data directed modes

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
HILL S I ET AL: "An analysis of the exponentiated gradient descent algorithm", ISSPA '99. PROCEEDINGS OF THE FIFTH INTERNATIONAL SYMPOSIUM ON SIGNAL PROCESSING AND ITS APPLICATIONS (IEEE CAT. NO.99EX359), PROCEEDINGS OF FIFTH INTERNATIONAL SYMPOSIUM ON SIGNAL PROCESSING AND ITS APPLICATIONS, BRISBANE, QLD., AUSTRALIA, 22-25 AUG, 1999, Brisbane, Qld., Australia, Queensland Univ. Technol, Australia, pages 379 - 382 vol.1, XP002185290, ISBN: 1-86435-451-8 *

Cited By (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2004080789A (en) * 2002-08-16 2004-03-11 Samsung Electronics Co Ltd Equalizer and equalization method to be used for high definition tv
JP4638134B2 (en) * 2002-08-16 2011-02-23 三星電子株式会社 Equalizer and equalization method used for high-definition TV
JP2004302470A (en) * 2003-03-31 2004-10-28 Microsoft Corp Method of noise estimation using incremental bayes learning
EP1492289A2 (en) * 2003-06-25 2004-12-29 Texas Instruments Incorporated Decision feedback equalization for high speed serial links
EP1492289A3 (en) * 2003-06-25 2006-12-20 Texas Instruments Incorporated Decision feedback equalization for high speed serial links
RU2477551C1 (en) * 2011-11-01 2013-03-10 Общество с ограниченной ответственностью "Специальный Технологический Центр" Method for multichannel adaptive reception of radio signals and apparatus for realising said method

Also Published As

Publication number Publication date
US7061977B2 (en) 2006-06-13
CN1395780A (en) 2003-02-05
JP2004509521A (en) 2004-03-25
EP1319289A1 (en) 2003-06-18
US20020054634A1 (en) 2002-05-09
KR20020096041A (en) 2002-12-28

Similar Documents

Publication Publication Date Title
US7061977B2 (en) Apparatus and method for using adaptive algorithms to exploit sparsity in target weight vectors in an adaptive channel equalizer
US6734920B2 (en) System and method for reducing error propagation in a decision feedback equalizer of ATSC VSB receiver
EP1449337B1 (en) Apparatus and method for constraining the value of feedback filter tap coefficients in a decision feedback equalizer
JP4741254B2 (en) Decision feedback equalizer and feedback filter coefficient update method
JP4063677B2 (en) Two-stage equalizer for trellis coded systems
US6647071B2 (en) Method and apparatus for equalization and tracking of coded digital communications signals
EP1386460A1 (en) Generation of decision feedback equalizer data using trellis decoder traceback output in an atsc hdtv receiver
KR100556401B1 (en) Equalizer in VSB receiver
WO2009009536A1 (en) Method for channel tracking in an lms adaptive equalizer for 8vsb
US8194800B2 (en) Reducing equalizer error propagation with a low complexity soft output viterbi decoder
US7218673B2 (en) Channel equalizer of single carrier receiver and equalizing method thereof
US8385397B2 (en) Method for determining the step size for an LMS adaptive equalizer for 8VSB
WO2007123863A2 (en) Dual pdfe system with forward-backward viterbi
KR100723479B1 (en) Decision-feedback equalizer and method in the digital television receiver
US8358683B2 (en) Channel equalizer
EP1745623A1 (en) Constellation location dependent step sizes for equalizer error signals
Georgoulakis et al. An efficient decision feedback equalizer for the ATSC DTV receiver

Legal Events

Date Code Title Description
AK Designated states

Kind code of ref document: A1

Designated state(s): CN JP KR

AL Designated countries for regional patents

Kind code of ref document: A1

Designated state(s): AT BE CH CY DE DK ES FI FR GB GR IE IT LU MC NL PT SE TR

WWE Wipo information: entry into national phase

Ref document number: 1020027006069

Country of ref document: KR

ENP Entry into the national phase

Ref country code: JP

Ref document number: 2002 527158

Kind code of ref document: A

Format of ref document f/p: F

WWE Wipo information: entry into national phase

Ref document number: 018036546

Country of ref document: CN

WWE Wipo information: entry into national phase

Ref document number: 2001970985

Country of ref document: EP

121 Ep: the epo has been informed by wipo that ep was designated in this application
WWP Wipo information: published in national office

Ref document number: 1020027006069

Country of ref document: KR

WWP Wipo information: published in national office

Ref document number: 2001970985

Country of ref document: EP

WWW Wipo information: withdrawn in national office

Ref document number: 2001970985

Country of ref document: EP