EP1515470B1 - Détection en pipeline turbo d'utilisateurs multiples - Google Patents

Détection en pipeline turbo d'utilisateurs multiples Download PDF

Info

Publication number
EP1515470B1
EP1515470B1 EP04077495.2A EP04077495A EP1515470B1 EP 1515470 B1 EP1515470 B1 EP 1515470B1 EP 04077495 A EP04077495 A EP 04077495A EP 1515470 B1 EP1515470 B1 EP 1515470B1
Authority
EP
European Patent Office
Prior art keywords
data
mud
processing
user
blocks
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Expired - Lifetime
Application number
EP04077495.2A
Other languages
German (de)
English (en)
Other versions
EP1515470A3 (fr
EP1515470A2 (fr
Inventor
Dianne E. Egnor
Robert Bruce Macleod
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Collision Communications Inc
Original Assignee
Collision Communications Inc
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Collision Communications Inc filed Critical Collision Communications Inc
Publication of EP1515470A2 publication Critical patent/EP1515470A2/fr
Publication of EP1515470A3 publication Critical patent/EP1515470A3/fr
Application granted granted Critical
Publication of EP1515470B1 publication Critical patent/EP1515470B1/fr
Anticipated expiration legal-status Critical
Expired - Lifetime legal-status Critical Current

Links

Images

Classifications

    • HELECTRICITY
    • H03ELECTRONIC CIRCUITRY
    • H03MCODING; DECODING; CODE CONVERSION IN GENERAL
    • H03M13/00Coding, decoding or code conversion, for error detection or error correction; Coding theory basic assumptions; Coding bounds; Error probability evaluation methods; Channel models; Simulation or testing of codes
    • H03M13/29Coding, decoding or code conversion, for error detection or error correction; Coding theory basic assumptions; Coding bounds; Error probability evaluation methods; Channel models; Simulation or testing of codes combining two or more codes or code structures, e.g. product codes, generalised product codes, concatenated codes, inner and outer codes
    • H03M13/2957Turbo codes and decoding
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04BTRANSMISSION
    • H04B1/00Details of transmission systems, not covered by a single one of groups H04B3/00 - H04B13/00; Details of transmission systems not characterised by the medium used for transmission
    • H04B1/69Spread spectrum techniques
    • H04B1/707Spread spectrum techniques using direct sequence modulation
    • H04B1/7097Interference-related aspects
    • H04B1/7103Interference-related aspects the interference being multiple access interference
    • H04B1/7105Joint detection techniques, e.g. linear detectors
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04BTRANSMISSION
    • H04B7/00Radio transmission systems, i.e. using radiation field
    • H04B7/02Diversity systems; Multi-antenna system, i.e. transmission or reception using multiple antennas
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L1/00Arrangements for detecting or preventing errors in the information received
    • H04L1/004Arrangements for detecting or preventing errors in the information received by using forward error control
    • H04L1/0045Arrangements at the receiver end
    • H04L1/0047Decoding adapted to other signal detection operation
    • H04L1/0048Decoding adapted to other signal detection operation in conjunction with detection of multiuser or interfering signals, e.g. iteration between CDMA or MIMO detector and FEC decoder
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L1/00Arrangements for detecting or preventing errors in the information received
    • H04L1/004Arrangements for detecting or preventing errors in the information received by using forward error control
    • H04L1/0045Arrangements at the receiver end
    • H04L1/0047Decoding adapted to other signal detection operation
    • H04L1/005Iterative decoding, including iteration between signal detection and decoding operation
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L1/00Arrangements for detecting or preventing errors in the information received
    • H04L1/004Arrangements for detecting or preventing errors in the information received by using forward error control
    • H04L1/0045Arrangements at the receiver end
    • H04L1/0052Realisations of complexity reduction techniques, e.g. pipelining or use of look-up tables
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L1/00Arrangements for detecting or preventing errors in the information received
    • H04L1/004Arrangements for detecting or preventing errors in the information received by using forward error control
    • H04L1/0045Arrangements at the receiver end
    • H04L1/0054Maximum-likelihood or sequential decoding, e.g. Viterbi, Fano, ZJ algorithms
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L1/00Arrangements for detecting or preventing errors in the information received
    • H04L1/004Arrangements for detecting or preventing errors in the information received by using forward error control
    • H04L1/0045Arrangements at the receiver end
    • H04L1/0055MAP-decoding
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L25/00Baseband systems
    • H04L25/02Details ; arrangements for supplying electrical power along data transmission lines
    • H04L25/0202Channel estimation
    • H04L25/0204Channel estimation of multiple channels
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L25/00Baseband systems
    • H04L25/02Details ; arrangements for supplying electrical power along data transmission lines
    • H04L25/0202Channel estimation
    • H04L25/021Estimation of channel covariance
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L25/00Baseband systems
    • H04L25/02Details ; arrangements for supplying electrical power along data transmission lines
    • H04L25/03Shaping networks in transmitter or receiver, e.g. adaptive shaping networks
    • H04L25/03006Arrangements for removing intersymbol interference
    • H04L25/03171Arrangements involving maximum a posteriori probability [MAP] detection
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L25/00Baseband systems
    • H04L25/02Details ; arrangements for supplying electrical power along data transmission lines
    • H04L25/03Shaping networks in transmitter or receiver, e.g. adaptive shaping networks
    • H04L25/03006Arrangements for removing intersymbol interference
    • H04L25/03178Arrangements involving sequence estimation techniques
    • H04L25/03248Arrangements for operating in conjunction with other apparatus
    • H04L25/0328Arrangements for operating in conjunction with other apparatus with interference cancellation circuitry
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L25/00Baseband systems
    • H04L25/02Details ; arrangements for supplying electrical power along data transmission lines
    • H04L25/03Shaping networks in transmitter or receiver, e.g. adaptive shaping networks
    • H04L25/03006Arrangements for removing intersymbol interference
    • H04L25/03178Arrangements involving sequence estimation techniques
    • H04L25/03203Trellis search techniques
    • H04L25/03216Trellis search techniques using the M-algorithm
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L25/00Baseband systems
    • H04L25/02Details ; arrangements for supplying electrical power along data transmission lines
    • H04L25/03Shaping networks in transmitter or receiver, e.g. adaptive shaping networks
    • H04L25/03006Arrangements for removing intersymbol interference
    • H04L25/03178Arrangements involving sequence estimation techniques
    • H04L25/03248Arrangements for operating in conjunction with other apparatus
    • H04L25/03299Arrangements for operating in conjunction with other apparatus with noise-whitening circuitry

Definitions

  • This present invention relates to advanced receiver techniques for processing signals in a multi-user environment, and more particularly to processing multiple user signals with a pipelining approach for Turbo Multiuser Detection (Turbo MUD) applications.
  • CDMA Code Division Multiple Access
  • spread spectrum communications has been around for many years, there is an increasing interest in using spread spectrum systems in commercial applications to allow superior quality performance and a greater number of users within a given bandwidth.
  • the digital format of CDMA architecture allows complex processing and high-level algorithms for transmission and reception.
  • each signal is transmitted using spread spectrum techniques.
  • the transmitted informational data stream is impressed upon a much higher rate data stream termed a signature sequence.
  • the bit stream of the signature sequence data is typically binary, and can be generated using a pseudo-noise (PN) process that appears random, but can be replicated by an authorized receiver.
  • PN pseudo-noise
  • the informational data stream and the high bit rate signature sequence stream are combined by multiplying the two bit streams together, assuming the binary values of the two bit streams are represented by +1 or -1. This combination of the higher bit rate signal with the lower bit rate data stream is called spreading the informational data stream signal.
  • Each informational data stream or channel is allocated a unique signature sequence.
  • a plurality of spread information signals each modulate a radio frequency (RF) carrier and may be jointly received as a composite signal at the single receiver.
  • Each of the spread signals overlaps all of the other spread signals, as well as noise-related signals, in both frequency and time.
  • the receiver correlates the composite signal with one of the unique time, frequency, or code sequences, and the corresponding information signal is isolated and processed. This is generally termed matched filtering.
  • the FDMA, TDMA, or CDMA signals have been designed to separate in this fashion, and the desired information is retrieved.
  • the waveform appearing at the receiver is not processed by standard single-user processing, or a combination of several single-user processors.
  • the joint detection processing more commonly referred to as Multiuser Detection (MUD)
  • MUD Multiuser Detection
  • MUD is employed by replacing the matched filter with a decorrelator.
  • the matched filter taps for each user is collected into a signature matrix (one column for each user), sometimes called the 'S' matrix.
  • Maximum Likelihood Sequence Estimation determines the most likely set of transmitted information bits for a plurality of digital signals without multipath time dispersion.
  • the maximum likelihood joint demodulator is capable, in theory, of accommodating the largest number of interfering signals, but has a prohibitive computational complexity that makes it commercially impractical.
  • the decorrelation receiver is another, less computationally complex receiver processing approach that zeroes out the interfering signals, or decorrelates the different signals.
  • the decorrelator, as well as virtually every other lower complexity joint demodulator is generally non-operational when the number of signals is over a set threshold which falls significantly short of the theoretical maximum.
  • turbo-decoding refers to iteratively passing information in the form of temporary or provisional bit decisions between a MUD processing block, and a decoding processing block, wherein for a turbo MUD the data is coded at the source.
  • the Viterbi algorithm is a method for decoding convolutional codes, and is a maximum-likelihood decoding method that minimizes the sequence-error probability for convolutional codes.
  • Bahl, Cocke, Jelinek, and Raviv (BCJR) published a forward-backward a posteriori probability (APP) decoding algorithm, the BCJR algorithm.
  • APP forward-backward a posteriori probability
  • the BCJR algorithm has seen an increased activity in coding field in relation to turbo decoding because it computes and stores likelihoods at each step, and can therefore provides more precise information to the MUD.
  • a MUD algorithm within the TurboMUD system determines discrete estimates of the transmitted channel symbols, with the estimates then provided to a bank of single-user decoders (one decoder for each user) to recover the input bit streams of all transmitted signals.
  • Two general types of multi-user detectors within the TurboMUD system are possible, namely those that provide hard outputs, which are discrete values, and those that provide soft outputs, which indicate both the discrete estimate and the probability that the estimate is correct.
  • MAP decoders In a Turbo MUD system, decoding and confidence information is passed between the multiuser detector and decoder components.
  • Maximum a posteriori (MAP) decoders or approximations of MAP decoders are well known to those in the art and are used for both the MUD and single user (SU) decoders so that soft output information is available if desired.
  • the filtered signal is typically then passed into a MUD detector.
  • the MUD detector is a full-complexity MAP detector.
  • Suboptimal reduced complexity MAP-based approaches are known in the relevant art.
  • the MUD detector passes soft decisions in the form of reliability, or confidence, measures to the single user decoders.
  • the reliability measures are presented with one associated with each symbol of each user to the bank of decoders. If the signals were transmitted with interleaving, the reliability measures from the MUD are first passed through a deinterleaver and passed on in shuffled form to the decoder. Shuffling refers to same values but changes the placement or presentation of the values.
  • MAP decoding is known in the art and further described in C. Schlegel, Trellis Coding, IEEE Press, 1997 ; Robertson, Villebrun and Hoeher, "A Comparison of Optimal and Sub-Optimal MAP Decoding Algorithms Operation in the Log Domain," ICC95 ; Hagenauer, and Hoeher, "A Viterbi Algorithm with Soft-Decision Outputs and its Applications," Globecom 89 ; Pottie and Taylor, “A Comparison of Reduced complexity Decoding Algorithms for Trellis Codes," J Sel.Areas in Comm Dec 1989 .
  • Turbo MUD approaches are described in, for example, Alexander, Reed, Asenstorfer, and Schlegel, "Iterative Multiuser Interference Reduction: Turbo CDMA,” Trans on Comm, July 1999 ; Poor, “Turbo Multiuser Detection: An Overview,” ISSSTA 2000 ; and Wang and Poor, “Iterative (Turbo) Soft Interference Cancellation and Decoding for Coded CDMA", Trans on Comm, July 1999 .
  • the turbo MUD generally outperforms a simple MUD/decode processor because the processing blocks in the turbo configuration are allowed to share information in the form of the provisional bit decisions, rather than each processor operating autonomously.
  • the MUD algorithm within a Turbo MUD system computes estimates of the transmitted channel symbols and passes them to a bank of single-user decoders (one decoder for each user). This process proceeds for several iterations, eventually recovering the input bit streams of all transmitted signals.
  • the state of the art suggests linear algorithms for the MUD processing block rather than optimal MUD or the T and M algorithm counterparts. It is well known that the reduction in performance of the linear MUDs can be offset by the gains due to the iteration in the turbo MUD. For example, it is not uncommon that five iterations in a turbo MUD between a linear MUD and a decoder can obtain optimal MUD like performance with fewer computations, even though the processing is repeated five times.
  • the MUD outputs a bit (or symbol) stream to a bank of error correcting decoders.
  • the decoders provide soft output streams of values associated with the bit streams of each interfering user's signal present in the received signal.
  • MUD can employ any of the soft decision input soft decision output multiuser detector such as a linear-based soft decision MUD known in the art such as Chan and Wornell, 'A Class of Asymptotically Optimum Iterated-Decision Multiuser Detectors' , and Wang and Poor, 'Iterative (Turbo) Soft Interference Cancellation and Decoding for Coded CDMA', IEEE Trans.
  • the system can also provide hard decisions with the MUD and bank of decoders described in the art for a reduced multi-access interference linear multiuser detector and hard decision error correction decoder.
  • the invention is devised in the light of the problems of the prior art described herein. Accordingly it is a general object of the present invention to provide a novel and useful technique that can solve the problems described herein.
  • a summary objective of the invention is to provide a more efficient Turbo MUD on frame asynchronous multiuser signals by using a particular ordering (pipelining) of the processing elements of the Turbo MUD algorithm.
  • a further object is an optimal pipelining of a MUD turbo processing for decoding multiple access interferences of signals that have been encoded by forward error correction methods.
  • state of the art TurboMUD is not well-suited for frame-asynchronous operation as there are inherent delays that contribute to inefficiency.
  • the Turbo MUD systems pass blocks of soft data between a MUD module and a decoding module to accomplish the demodulation. It is typical that upon transmission the data has been coded into a frame of symbols and thus cannot be retrieved at the receiver until an entire frame is available to the decoder module of the receiver.
  • the state of the art Turbo MUD typically requires that the blocks of data passed around in a turbo decoder correspond exactly to the size of a coded frame of symbols. Such a configuration is possible when the reception of each user's frame-data can be precisely lined up with the blocking structure established in the turbo loop for a frame-synchronous situation.
  • An object of the present invention is to reduce complex processing for MUD applications such as IS95 CDMA and provide improved processing times.
  • the processing delays of the state of the art which correspond to the delays waiting for each user to complete a frame of MUD processed data before passing along the data to the decoder, are reduced with the pipelined approach of the present invention.
  • the subsequent MUD processing blocks are not delayed but are executed immediately by parallel processing elements.
  • One aspect of the invention is an apparatus for processing asynchronous data in a multiple access system as defined in claim 1.
  • a variation of the present invention includes the apparatus wherein each of the received signals are respectively coupled one of the multiuser detector processors, and wherein each of the multiuser detector processors are respectively coupled one of the decoders.
  • the decoders can use algorithms selected from the group of algorithms consisting of Viterbi algorithm and Bahl, Cocke, Jelinek, and Raviv (BCJR) algorithm.
  • algorithms include those selected from the group consisting of M-algorithm, T-algorithm, FANO, or reduced state Viterbi, maximum a posteriori (MAP) decoders and soft-output Viterbi algorithm (SOVA) decoders.
  • Additional elements can include an interleaver coupled between the multiuser detector processors and the decoders and a deinterleaver coupled between the decoders and the multiuser detector processors.
  • the invention is a digital processing system performing Turbo MUD processing on multiple access communications, which comprises a parameter estimation unit coupled to a plurality of received user signals, wherein each of the user signals are a plurality of blocks of data.
  • a multiuser detector coupled to the parameter estimation unit, wherein the multiuser detector processes the blocks of data for each of the user signals in a partial manner with a processing window defined by frame boundaries.
  • the multiuser detector outputs processed blocks of data, and there is a bank of decoders coupled to the multiuser detector.
  • the decoders process all of the processed blocks of data for one of the user signals as soon as one of the blocks of data for one of the user signals has been processed by the multiuser detector in its entirety.
  • the decoders produce improved processed blocks of data.
  • the iterative processing aspects of the present invention are realized when the improved processed blocks of data are fed back to the multiuser detector.
  • a further variation of the present invention includes where the processing window is edge triggered or commences processing on a central bit.
  • Another element applicable to the invention further comprises a synchronization unit coupled to the multiuser detector determining which of the processed blocks of data to decode.
  • the subject of the invention disclosed in this application does not require that the signals correspond to any particular multiple access (MA) scheme or even that they are all of the same type, or come from a wireless system.
  • the present invention operates in the same manner on any set of digitally modulated interfering signals to include cellular CDMA systems, TDMA systems, FDMA systems, storage medium, wired MA systems such a cable modems, wireless local area network systems, or yet undetermined systems.
  • each signal source produces a signal with the information digitally modulated using a signature pulse or finite duration signal of some sort.
  • CDMA is described for illustrative purposes to explain the invention, the specific example of CDMA is merely for ease of understanding.
  • the present invention is directed to any other form of digital communication or signal storage methods and the terminology used in the CDMA discussions such as “signature sequence”, “signature signal”, and “signature pulse” shall be deemed equivalent.
  • the wording of "CDMA signal” and "digitally modulated signal” shall also be considered as equivalent with respect to the capabilities of the processing of bits.
  • the present invention is not limited to CDMA and is equally applicable to any digitized communication protocol.
  • a further feature of the present invention is that it works equally well using mixed rate communication systems such as IS95, wherein the user chooses the transmission rate.
  • the parameter estimator that handles the differing transmission rates passes along the information to the present system.
  • Pipelined Turbo-MUD enable implementations of advanced receiver processing providing high quality real-time processing for multiple access systems.
  • the preferred embodiment is an illustration of the digital processing technique that is applicable to many variations and applications all within the scope of the invention.
  • the turboMUD approach in general, repeats the processing steps of decoding and multiuser detection several times before an output is produced.
  • the convolutional code imparted on the data at the source(s) provides a certain data dependency in time, while the multi-user reception provides a certain data dependency amongst users.
  • the present invention provides for computational efficiency by interrupting the processing of the MUD at each frame boundary to perform decoding before re-establishing the frame boundaries and continuing processing. Each of these interruptions spawns a new processing thread in the receiver hardware which allows more effective use of the MUD processing threads as they do not have the delays found in other processing schemes.
  • transmitter section 5 and receiver section 10 are depicted with MUD processing.
  • the interleaver 25 performs the interleaving of the data, which is then symbol mapped by the symbol mapper 30, such as BPSK symbol mapped with data symbols having a duration T.
  • Each data symbol is then modulated by a spreading waveform from the spreader 35 , and the signal data s k (t) is output from the transmitter end with a number of data symbols per user per frame.
  • the data is generally transmitted by antennas 40 .
  • the transmission on hard-wire or fixed wire systems is also known in the art, however the present example is directed towards the wireless communications systems.
  • Noise n(t) such as white Gaussian noise, is an inherent component of most transmission signals and some random noise components are generally present in transmitted data signals r(t).
  • antennas 45 receive the transmitted signals r(t) as well as the various interfering signals, such as n(t).
  • a MUD section 50 that interacts with the K number of channel decoders 65 for the received signal r(t) for each user, each channel passing through a de-interleaver 60 prior to being decoded.
  • the multiuser detector 50 is generally a soft input, soft output (SISO) detector, receiving and outputting soft values.
  • the MUD section 50 utilizes a priori information and delivers values such as a posteriori log-likelihood ratio (LLR) data of a transmitted '+1' or '-1' for each bit of every user.
  • LLR posteriori log-likelihood ratio
  • Certain a priori information is computed by the channel decoder 65 from a previous iteration which is interleaved in the interleaver 75 and fed back to the MUD section 50.
  • the de-interleaver 60 and interleaver 75 perform the de-interleaving and interleaving functions respectfully for the encoding schema.
  • the channel decoders 65 produce the output data stream (dk)' representing the best estimate of the transmitted signal (dk).
  • the wireless communication application is further presented in Figure 2 , wherein a number of users (1-K) generate signals that are sent by transmitters 100 into free space. There is normally a noise component n(t) 145 that is introduced from the environment of a random nature in the received signal. While any noise that has a repeatable or non-random nature can be eliminated through processing, random noise elements are reduced in other manners such as filtering but are inherent conditions.
  • the various signals are received at antennas (1-p) 140 wherein there is one signal for each polarization feed.
  • the signals represent directly received signals 110, as well as multi-path signals from the same user 130, and interfering signals 120 from other users.
  • the plurality of signals from each antenna 140 is processed in a RF front end unit 150.
  • the front end unit 150 downconverts the higher frequency signals into baseband signals and provide processing and filtering as is known in the art.
  • the baseband signals are also digitized by analog to digital converters (ADC).
  • ADC analog to digital converters
  • the front end 150 cooperates with the parameter estimation unit 160 to retrieve needed information for the signals such as relative received timing offsets, carrier phase, frequency offsets, received amplitudes, and multipath structure for each of the interfering signals present in the received signal.
  • the MUD element 170 consists of functional blocks that process the digital data and extract the user symbol streams.
  • the pre-processor 175 converts the baseband digital data into the proper format for further processing according to the desired detection scheme.
  • the format is often one measurement per 'dimension' per symbol.
  • An example of this processing is a bank of matched filters.
  • the multi-user detection stage 180 as detailed herein cooperates with the error correction decoding (ECD) 185 for iterations of the processing of the Turbo-Mud.
  • ECD error correction decoding
  • An input signal of raw non-manipulated data at the receiver is comprised of the aggregate of many signals from many different transmitters, where each signal is assigned a channel (frequency, timeslot, and/or spreading code) from a finite set of channels.
  • the interference from these various users generally requires complex processing and associated processing time.
  • the aggregate signal is collected at the receiver, down-converted and digitized. This digital input signal or raw input data is then input to the turbo multiuser detector (MUD) system 170.
  • MOD turbo multiuser detector
  • the basic Turbo MUD 170 assumes knowledge of various parameters such as relative received timing offsets, carrier phase, frequency offsets, received amplitudes, and multipath structure for each of the interfering signals present in the received signal.
  • This input data represents raw data after some front end processing such as downconversion, amplification, and analog-to-digital conversion. All MUD systems 170 need some raw data parameters in order to establish accurate decision trees for processing.
  • the input data represents a vector of data, transferred at some rate (e.g., the symbol rate), and this data is typically transmitted to a matched filter (not shown).
  • the same vector is passed on to the parameter estimation module 160.
  • the purpose of the parameter estimation module 160 is to estimate timing, signal amplitudes, phases, polarization, and identification of transmission channels.
  • a parameter estimation unit 160 processes the various parameters for the received raw data.
  • a parameter estimator as known in the art, is a broad description for a unit that provides information to the MUD and decoders such as convolutional code, signatures, multiplexing format. While the term parameter estimator is used herein, the term is intended to be interpreted in the broader sense as known in the joint demodulation field.
  • the multiuser detection unit 180 outputs a bit (or symbol) stream associated with each interfering signals present on the channel for one data block. De-interleavers and interleavers (not shown) are optional elements that are used if the transmitted signals are interleaved, such as the CDMA format and are known in the art.
  • the bit streams from the MUD 180 are passed to a bank of error correction decoders unit 185.
  • the iterative MUD or Turbo MUD can be structured as a hard output or soft output processing. In order to demonstrate the present variation, the soft output version is addressed herein, but it is within the scope of the present invention to utilize hard outputs.
  • the decoders 185 calculate conditional probabilities, one for each decoded symbol of each user, and output them as confidence values back to the MUD 180.
  • Soft outputs for each bit of each user from the bank of decoders 185 is fed back to the multiuser detector 180 for each iteration, one stream of outputs for each interfering user present in the received signal. These soft outputs are passed back to the MUD 180 to produce an improved stream of soft bit (or symbol) decisions that fed into the decoder for another iteration of improvement.
  • the information between the MUD 180 and the decoders 185 repeats in subsequent iterations until an asymptote is reached or the desired performance level is attained. At that point, estimates of the data sequences for all active users are output 190. Operation then commences for the next block of data, repeating the process described above.
  • the multiuser detector 180 takes these soft inputs along with the original raw input signal to calculate an improved, less corrupted bit stream for each user. This iterative process continues until a desired metric is reached or a fixed number is reached. At that point, estimates of the data sequences for all active users are output. Operation then commences for the next block of data, repeating the process described above.
  • the number of iterations for processing between the multiuser detector 180 and the decoders 185 can be set to a fixed counter or by checking if there were significant changes to the data from the last iteration. Once the data is no longer being altered or reaches a certain iteration counter limit, the data from the decoder 185 can be output as final estimates of what the user sent. A fixed number of iterations can be stored and used and processed by the decision block. Alternatively, the information between the multiuser detector 180 and the decoders 185 repeats in subsequent iterations until an asymptote is reached or the desired performance level is attained. A buffer (not shown) can store the previous values and compare them to the latter processed values during the iterative process.
  • the system 170 When processing is completed, the system 170 outputs the final stream of decisions or output data stream for each interfering user for the current data block. The process is repeated for all subsequent data blocks.
  • the interaction between the multiuser detection section 180 and the error correction decoding 185 is explained in further detail herein.
  • each block 200 consists of a series of transmitted symbols that can be of any length, but is typically several hundred symbols long.
  • Turbo MUD processing is accomplished by selecting a frame of data 210, which typically encompasses at least a full block of data for each user, wherein the frame of data 210 is processed by the MUD 205 and passed to the decoders.
  • the MUD 205 known in the art generally processes an entire data block 200 within a single frame of data 210.
  • frame 210 shows the frame boundaries which includes all of the blocks 'i' for each of the Users 1 - k. Only complete blocks 200 of data can be decoded, so the MUD 205 processing is performed for all of the symbols for a given frame 210 before decoding the results.
  • MUD processing requires that the processing window 210 includes frame 'i' for user 1 and frame 'i' for user 2, etc. Therefore the processing window is at least as wide as any single block 200 of data.
  • the step is repeated for a number of iterations for each user before proceeding to the next frame of data, including blocks 'i+1' for example.
  • This single frame processing a chunk of data in a serial fashion for each user and then decoding that single frame before proceeding to the next frame is inherently inefficient.
  • This prior process includes waiting for other processing to finish, incurring idle time of the MUD processing that lowers computational efficiency.
  • Figure 4 illustrates a presentation of the pipelined Turbo Mud processing for the frame asynchronous processing according to the present invention.
  • the processing is executed in a different manner than processing known in the art in which the MUD processors process whole frames and wait for all frames to be completed.
  • the prior schemes do not allow for partial frame processing, wherein the asynchronous nature of the various frames from multiple users results in certain users not being fully encompassed by the MUD processing window.
  • partial blocks are processed for all users thereby reducing idle time in the MUD processors. Once an entire block has been processed, the data in that block is updated or otherwise replaced and all blocks for that user are decoded.
  • the present invention processes multiple blocks of data in parallel, wherein each is spaced approximately one block apart, with the boundaries extending to the next block transition. The processing is based upon time, wherein each of the plurality of MUD processors are operating on older and older data with respect to the time element, until the MUD processor shown as MUD N_turbo-1 is executed and the results decoded.
  • the 'Data In' in this embodiment represents waveforms from multiple users transmitting signals which are received and processed by the receiver units.
  • the data is divided in blocks 310 for each user, the blocks 310 shown for illustrative purposes are represented as i-1, i, and i+1 for each user and represent three blocks of data wherein block i-1 represents the oldest data.
  • the blocks 310 for subsequent users are not lined up or entirely within the same boundary frame 300 subject to processing by a single MUD processor 305, but rather have some delay element.
  • MUD processors 305 As shown in the upper portion of Figure 4 , MUD processors 305, including MUD 0 , MUD 1 and MUD 2 , has a 'window' or frame region 300 in which it processes data.
  • the MUD window 300 processes small chunks of data.
  • the bits generally represent original processed bits as well as those that have undergone further processing and have different levels of maturity.
  • the size of the MUD processor window 300 or sub-window can be any size such as 1 bit wide, an entire block width, or any size that allows for efficient processing for a given application.
  • the processing for the window 300 can be edge triggered or commence processing on a central bit and encompass a certain number of bits on either side of the central bit.
  • each of the bit outputs from the MUD can be considered to be flagged with the user and location in time, both absolute and relative to the frame edge.
  • the 'trigger' can be activated when the flag has a particular value.
  • the MUD processors 305 process the partial frame data within the boundaries 300 at the same time, regardless of the asynchronous framing aspects of the data as it streams through the processing.
  • MUD 0 sees the most recent data and tries to process the first portion of block i+1 for User 1 and simultaneously processing the last portion of block i of User 2.
  • MUD processors 305 simultaneously processing data blocks 310 as the streams of data process through the system.
  • the data must be decoded for that User.
  • the data is 'replaced' or refreshed so that that next MUD processor sees a different set of data.
  • FIG. 4 graphically illustrates a snapshot in time for a plurality of MUD processors 305 processing partial blocks of data 310 within the processing window 300.
  • the MUDs 305 process the block data 310 within the processing window 300, including the data from multiple Users, herein shown as User 1, User 2 and User k.
  • the size of the processing window 300 is determined by the frame boundaries as established by the parameter estimator (not shown), wherein the processing window 300 commences at the beginning of the frame of data from User 1 up until the beginning of the frame of data of User 2, and so on.
  • the MUD 0 processes the end of the frame from User 2 as well as the beginning portion of the frame of User 1.
  • MUD 0 As soon as MUD 0 has processed the last bit from User 2 block i, MUD 0 must now pass on the data for decoding all the blocks for User 2 and replace the data before processing the data within those frames in the next MUD iteration. Therefore, once the MUD's 305 are finished processing User 2 frame, in the next instant, all User 2 frames are decoded 315.
  • the processing can be done sequentially or in parallel employing the processing scheme of the present invention.
  • the lower portion of Figure 4 illustrates a subsequent processing, wherein the MUD processors 320 are processing the blocks of data 310, including the refreshed data for the blocks of User 2.
  • the processing window 325 in this instant begins at the beginning of the block from User 2 and extends to the beginning of the block of the next User who will have a block completed by this frame, in this case it happens to be user 1.
  • MUD 0 is processing the last portion of block i+1 for User 1
  • MUD 1 is processing the last portion of block i for User 1
  • MUD 2 is processing the last bit of block i-1 for User 1.
  • the data from User 1 is decoded 330 and the data for User 1 is refreshed or replaced.
  • block i+1 will be processed in turn by MUDs 1, 2 and so on.
  • the block i+1 data Prior to reaching each of these MUD processors 320, the block i+1 data will have passed through several MUDs and decodings, each time being replaced with the newly computed values. In this fashion the iterative or turbo MUD processing is accomplished. The processing continues as noted herein, and the decoding process is further illustrated in Figure 5 .
  • this invention retains its properties of not doing extra or duplicate computation and of producing results with the minimum lag or computational delay.
  • single user perspective of the processing of the present system is shown illustrating the interaction between the MUD processors and the decoders. This single user perspective is employed to explain the principles of operation, however it should be understood that multiple users would have a similar configuration and are coupled to each other and the processing is synchronized among the users.
  • the 'Data In' represents blocks of data that are partitioned according to the code scheme used for the transmission/reception.
  • the blocks of data 400 for each user are asynchronously delayed with respect to each other by some time interval as described herein.
  • a plurality of MUD processors 410 are coupled to the incoming frames of data 400.
  • the MUD processors 410 are coupled to some form of parameter estimator (PE) 420 to obtain signal information for processing the data, as well as an interconnection to a priori information 405.
  • PE parameter estimator
  • the MUD processors 410 are also coupled to decoders 430, such as the BCJR decoders, used in this example for decoding the completed frames of data.
  • the first block 400 of data, block 0, is communicated to MUD 0 along with the a priori information 405 and parameter estimation 420 values as is known in the art.
  • the MUD 0 processes the partial frame of data within the processing window as detailed in the Figure 4 description.
  • the output from the MUD 0 is an input to a decoding algorithm such as BCJR.
  • the BCJR processes the data and provides the input to the next stage, MUD 1 .
  • the data input values, delayed by one frame, are input the MUD 1 along with the respective apriori and parameter estimator (PE) information.
  • PE parameter estimator
  • the output of the MUD 1 stage is input to the next parallel processed data frame and continues until the last stage at which point the output from the BCJR is the output values of the system, Data Out.
  • the Data Out is the symbol stream of improved and processed data.
  • the information passed from the MUD processors 410 to the decoders 430 encompass the decisions made as part of the MUD processing in establishing the most probable conditional values for the data in the frame.
  • the decoders 430 can be thought of as a storage register that holds the data that has been decoded at least once, wherein the MUD processors 410 are not ready to pass all Users through the decoding stage at once. Once a User has a complete block processed by the MUD that User is selected and all blocks for that User are decoded.
  • the decoder 430 provides improved data to each subsequent MUD in the User chain.
  • MUD 0 works on data with no significant a priori information from the convolutional code, but with symbol probabilities from some a priori information (e.g. determined by the channel estimates) or assumed equivalent.
  • MUD 1 works on data with a priori information from at least one pass through the convolutional decoders.
  • MUD # N turbo -1 works with data having a priori information and passed through the convolutional decoder N turbo times.
  • the blocks of data are replaced so that the MUD processing that data sees 'new' data.
  • the old data is generally not 'lost' and can be used for subsequent processing.
  • the conditional/decision section 440 links to each user in the system and provides the synchronization to switch to whichever User has a full frame of data in the MUD processor 410. Once the conditional/decision or synchronization section locates which User has completed processing of a complete frame, that User is selected for processing. The User's data is decoded and passed to the BCJR decoder 430 for that User. The MUD 410 moves to the next frame boundary and the processing continues by propagating down through the multiple layers and through all Users.
  • the MUD detector can pass soft decisions in the form of reliability, or confidence, measures to the bank of decoders.
  • the reliability measures are presented with one associated with each symbol of each user to the bank of decoders. If the signals were transmitted with interleaving, the reliability measures from the MUD are first passed through a deinterleaver (not shown) and passed on in shuffled form to the decoder. Shuffling refers to same values but changes the placement or presentation of the values.
  • multiple MUDs are running wherein the processing of the MUD is interrupted at each frame boundary of the next user. Once an entire block is processed by the MUD, the blocks for that user are decoded. The frame of data is 'replaced' and the processing goes to the next frame boundary for MUD processing. It should be understood that the old data that is replaced by new processed data does not generally result in the old data being lost, as the a priori information of the bits in the block are replaced with more mature processed data.
  • CPUs computer processing units
  • processing threads on a multiprocessor system can be individually assigned to the MUD and decoding processing tasks.
  • data latency and processor idle time are minimized to the lowest possible values.
  • each of the MUD processing blocks is continually processing the next symbol in its data path, possibly delayed, but still immediately available to the processor in real-time.
  • the same processor would be computing MUD 0 , MUD 1 , and so forth. This processor would therefore have to wait for the results of previous MUD and BCJR decode operations, and would be in an idle state for a significant portion of its time.
  • One application which shows a non-CDMA environment is to the application involving GSM, which is a narrow band TDMA system.
  • the user communicates over a timeslot and when the time slot is filled, another user has to wait until an open slot is available.
  • the present invention allows reassignment of the timeslot so that signals from a second user can overlay a first user.
  • the only distinguishing characteristics would be some phase and power differences that can be employed as described herein to differentiate user 1 from user 2.
  • Another application of the invention is to allow for multi-user detection for a variety of communications formats and not solely limited to CDMA. Any multiuser system that has user data coded at the source may be processed by the turbo-mud algorithms outlined in this application.
  • the format of the data is not limited to CDMA.
  • the subject system has been described in terms of a wireless communications network, it has application to any situation in which digitally encoded data are processed in a TurboMUD processing.
  • the reference to users and user signals has been present to facilitate ease of understanding in one embodiment for communications.
  • the subject system has application to any digital processing application such as cable networks and storage mediums.
  • the MUD processing with respect to the storage devices refer to the plurality of signals received when the optical head picks up the signals of the adjacent tracks of the storage mediums. The tight spacing between the tracks creates a multiple user detection problem involving the processing of the desired track signal from the other received tracks.

Landscapes

  • Engineering & Computer Science (AREA)
  • Computer Networks & Wireless Communication (AREA)
  • Signal Processing (AREA)
  • Power Engineering (AREA)
  • Physics & Mathematics (AREA)
  • Probability & Statistics with Applications (AREA)
  • Artificial Intelligence (AREA)
  • Theoretical Computer Science (AREA)
  • Error Detection And Correction (AREA)
  • Image Analysis (AREA)
  • Image Processing (AREA)
  • Train Traffic Observation, Control, And Security (AREA)

Claims (9)

  1. Un appareil de traitement de données asynchrones dans un système à accès multiples, l'appareil comprenant :
    une pluralité d'antennes (140) destinées à la réception d'une pluralité de signaux (110, 120, 130) provenant d'une pluralité d'utilisateurs (100), lesdits signaux (110, 120, 130) étant divisés en blocs de données (310) pour chacun desdits utilisateurs (100), chaque bloc de données étant partitionné conformément à un schéma de code utilisé pour l'émission et la réception dans le système à accès multiples, et des blocs de données pour un utilisateur de ladite pluralité d'utilisateurs étant asynchrones par rapport à des blocs de données pour un autre utilisateur de ladite pluralité d'utilisateurs,
    une pluralité de processeurs détecteurs multi-utilisateurs (180; 305), MUD0-MUDNturbo-1, couplés à ladite pluralité d'antennes (140) destinés à la réception de ladite pluralité de signaux (110, 120, 130; 310), chaque processeur détecteur multi-utilisateur (180; 305) traitant des parties desdits blocs de données asynchrones (310) pour chacun desdits utilisateurs (100) à l'intérieur d'une fenêtre de traitement (300),
    une pluralité de décodeurs (185; 315) couplés auxdits processeurs détecteurs multi-utilisateurs (180; 305) destinés à la réception desdites parties traitées de blocs de données asynchrones et à la production de blocs de données améliorés, chaque décodeur (185; 315) décodant la totalité desdits blocs de données (310) pour un desdits utilisateurs (100) une fois que les processeurs détecteurs multi-utilisateurs (180; 305) ont traités un bloc entier de données pour ledit utilisateur (100), et
    un estimateur de paramètre (160) couplé à ladite pluralité d'antennes (140) et à ladite pluralité de processeurs détecteurs multi-utilisateurs (180; 305) destiné à la détermination d'une taille de ladite fenêtre de traitement (300) où (i) la pluralité de décodeurs et la pluralité de processeurs détecteurs multi-utilisateurs (180; 305), MUD0-MUDNturbo-1, sont agencées en une configuration en pipeline de sorte que chaque processeur détecteur multi-utilisateur produise vers un décodeur, le décodeur étant agencé de façon à produire vers le processeur détecteur multi-utilisateur suivant dans le pipeline et où, (ii) MUD0 est agencé de façon à traiter des parties desdits blocs de données asynchrones pour chacun desdits utilisateurs à l'intérieur d'une fenêtre de traitement et MUD1 à MUDNturbo-1 sont agencés de façon à traiter des parties desdits blocs de données asynchrones pour chacun desdits utilisateurs à l'intérieur d'une fenêtre de traitement et la sortie du décodeur associée au processeur détecteur multi-utilisateur précédent dans le pipeline.
  2. L'appareil selon la Revendication 1, où chaque décodeur (315, 330) utilise des algorithmes sélectionnés dans le groupe d'algorithmes se composant de : algorithme de Viterbi et algorithme de Bahl, Cooke, Jelinek et Raviv (BCJR).
  3. L'appareil selon la Revendication 1 ou 2, où chaque processeur détecteur multi-utilisateur (305, 320) utilise des algorithmes sélectionnés dans le groupe se composant de : algorithme M, algorithme T, FANO ou Viterbi à état réduit, décodeurs maximum a posteriori (MAP) et décodeurs d'algorithme de Viterbi à sortie logicielle (SOVA).
  4. L'appareil selon l'une quelconque des Revendications précédentes, comprenant en outre au moins un entrelaceur couplé entre ladite pluralité de processeurs détecteurs multi-utilisateurs (180; 305) et ladite pluralité de décodeurs (315, 330), et au moins un désentrelaceur couplé entre ladite pluralité de décodeurs (315, 330) et ladite pluralité de processeurs détecteurs multi-utilisateurs (180; 305).
  5. L'appareil selon l'une quelconque des Revendications précédentes, comprenant en outre une unité de synchronisation couplée à ladite pluralité de processeurs détecteurs multi-utilisateurs (185; 305) destinée à la détermination du bloc desdits blocs de données traités à décoder.
  6. L'appareil selon l'une quelconque des Revendications précédentes, où ladite fenêtre de traitement (300) est définie par des limites de trame.
  7. L'appareil selon l'une quelconque des Revendications 1 à 5, où ladite fenêtre de traitement (300) est déclenchée par un front d'impulsion.
  8. L'appareil selon l'une quelconque des Revendications 1 to 5, où ladite fenêtre de traitement est déclenchée de façon à commencer le traitement sur un bit central.
  9. Un système de traitement numérique exécutant un traitement Turbo MUD sur des communications à accès multiples comprenant un appareil selon l'une quelconque des Revendications 1 à 8.
EP04077495.2A 2003-09-10 2004-09-08 Détection en pipeline turbo d'utilisateurs multiples Expired - Lifetime EP1515470B1 (fr)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
US659567 2003-09-10
US10/659,567 US7236546B2 (en) 2003-09-10 2003-09-10 Pipelined turbo multiuser detection

Publications (3)

Publication Number Publication Date
EP1515470A2 EP1515470A2 (fr) 2005-03-16
EP1515470A3 EP1515470A3 (fr) 2007-04-11
EP1515470B1 true EP1515470B1 (fr) 2013-08-21

Family

ID=34136759

Family Applications (1)

Application Number Title Priority Date Filing Date
EP04077495.2A Expired - Lifetime EP1515470B1 (fr) 2003-09-10 2004-09-08 Détection en pipeline turbo d'utilisateurs multiples

Country Status (2)

Country Link
US (1) US7236546B2 (fr)
EP (1) EP1515470B1 (fr)

Families Citing this family (20)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US6967598B2 (en) * 2004-02-20 2005-11-22 Bae Systems Information And Electronic Systems Integration Inc Reduced complexity multi-turbo multi-user detector
US9071279B2 (en) * 2004-05-18 2015-06-30 Nxp, B.V. Turbo decoder input reordering
US20050273520A1 (en) * 2004-05-21 2005-12-08 Bea Systems, Inc. Service oriented architecture with file transport protocol
US20060080419A1 (en) * 2004-05-21 2006-04-13 Bea Systems, Inc. Reliable updating for a service oriented architecture
US20060005063A1 (en) * 2004-05-21 2006-01-05 Bea Systems, Inc. Error handling for a service oriented architecture
US20050273517A1 (en) * 2004-05-21 2005-12-08 Bea Systems, Inc. Service oriented architecture with credential management
US7653008B2 (en) 2004-05-21 2010-01-26 Bea Systems, Inc. Dynamically configurable service oriented architecture
US20060031481A1 (en) * 2004-05-21 2006-02-09 Bea Systems, Inc. Service oriented architecture with monitoring
US20050278335A1 (en) * 2004-05-21 2005-12-15 Bea Systems, Inc. Service oriented architecture with alerts
EP1849252B1 (fr) * 2005-02-14 2016-04-13 ViaSat, Inc. Decodage a correction aval des erreurs integre et reception en diversite iterative
US7466771B2 (en) * 2005-02-14 2008-12-16 Viasat Inc. Block edge effects in iterative diversity reception
US7693239B2 (en) * 2006-02-08 2010-04-06 Harris Corporation Apparatus for decoding convolutional codes and associated method
US7965803B2 (en) * 2006-12-14 2011-06-21 The Trustees Of Columbia University In The City Of New York Methods and systems for providing feedback for beamforming
US7924951B2 (en) * 2006-12-14 2011-04-12 The Trustees Of Columbia University In The City Of New York Methods and systems for digital wireless communication
US7925964B2 (en) * 2006-12-22 2011-04-12 Intel Corporation High-throughput memory-efficient BI-SOVA decoder architecture
US8996394B2 (en) * 2007-05-18 2015-03-31 Oracle International Corporation System and method for enabling decision activities in a process management and design environment
WO2008157724A1 (fr) * 2007-06-19 2008-12-24 The Trustees Of Columbia University In The City Of New York Procédés et systèmes pour fournir une rétroaction à la formation de faisceau et à la commande de puissance
US8185916B2 (en) 2007-06-28 2012-05-22 Oracle International Corporation System and method for integrating a business process management system with an enterprise service bus
KR101651568B1 (ko) * 2009-10-27 2016-09-06 삼성전자주식회사 3차원 공간 인터페이스 장치 및 방법
US9712355B1 (en) * 2016-01-11 2017-07-18 Qualcomm Incorporated Spreading techniques for frequency-shift keying modulation

Family Cites Families (9)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US6304995B1 (en) 1999-01-26 2001-10-16 Trw Inc. Pipelined architecture to decode parallel and serial concatenated codes
CN1201499C (zh) 1999-09-21 2005-05-11 美商内数位科技公司 可变扩展因数多使用者检测器
US7099401B2 (en) * 1999-12-15 2006-08-29 Paradyne Corporation Discrete multitone interleaver
US6975666B2 (en) * 1999-12-23 2005-12-13 Institut National De La Recherche Scientifique Interference suppression in CDMA systems
US7139306B2 (en) * 2001-03-14 2006-11-21 Mercury Computer Systems, Inc. Wireless communication systems and methods for long-code communications for regenerative multiple user detection involving pre-maximal combination matched filter outputs
US6873646B2 (en) * 2001-07-12 2005-03-29 Industrial Technology Research Institute Perturbation apparatus and method for user detection in a multiple-access communication system
US7317750B2 (en) * 2002-10-31 2008-01-08 Lot 41 Acquisition Foundation, Llc Orthogonal superposition coding for direct-sequence communications
US7099378B2 (en) * 2003-01-30 2006-08-29 The Mitre Corporation Sub-symbol parallel interference cancellation
US6999498B2 (en) * 2003-04-24 2006-02-14 Bae Systems Information And Electronic Systems Integration Inc. Multiuser detection aided multiple access differential frequency-hopped spread spectrum

Also Published As

Publication number Publication date
EP1515470A3 (fr) 2007-04-11
EP1515470A2 (fr) 2005-03-16
US20050053173A1 (en) 2005-03-10
US7236546B2 (en) 2007-06-26

Similar Documents

Publication Publication Date Title
EP1515470B1 (fr) Détection en pipeline turbo d'utilisateurs multiples
US6704376B2 (en) Power and confidence ordered low complexity soft turbomud with voting system
EP1566911B1 (fr) Détecteur multi-utilisateur multi-turbo
EP1521387B1 (fr) Détecteur multi-utilisateur multi-turbo
EP1264456B1 (fr) Procede et appareil assurant a la fois l'elimination de l'interference fondee sur la decision ponderee et le decodage
US20050018793A1 (en) Hybrid turbo-mud for multiple access systems
WO2004095713A2 (fr) Estimateur de symboles, d'amplitude et de debit
US8406354B2 (en) Interference cancellation in a multi-user receiver
WO2000033527A1 (fr) Systemes et procedes permettant de recevoir un signal module contenant des bits codes et decodes au moyen d'une demodulation multipasse
Leung et al. Interleaving-based multiple access and iterative chip-by-chip multiuser detection
US8413021B2 (en) Efficient soft value generation for coded bits in a turbo decoder
US20040022336A1 (en) Turbo decoder with partial interference cancellation
Tonello Iterative MAP detection of coded M-DPSK signals in fading channels with application to IS-136 TDMA
EP1698057B1 (fr) Procédé et appareil pour détecteur multi-utilisateur turbo désorganisé de façon aléatoire

Legal Events

Date Code Title Description
PUAI Public reference made under article 153(3) epc to a published international application that has entered the european phase

Free format text: ORIGINAL CODE: 0009012

AK Designated contracting states

Kind code of ref document: A2

Designated state(s): AT BE BG CH CY CZ DE DK EE ES FI FR GB GR HU IE IT LI LU MC NL PL PT RO SE SI SK TR

AX Request for extension of the european patent

Extension state: AL HR LT LV MK

RIC1 Information provided on ipc code assigned before grant

Ipc: H04L 1/00 20060101AFI20041220BHEP

Ipc: H04L 25/03 20060101ALI20061219BHEP

Ipc: H04B 1/707 20060101ALI20061219BHEP

PUAL Search report despatched

Free format text: ORIGINAL CODE: 0009013

AK Designated contracting states

Kind code of ref document: A3

Designated state(s): AT BE BG CH CY CZ DE DK EE ES FI FR GB GR HU IE IT LI LU MC NL PL PT RO SE SI SK TR

AX Request for extension of the european patent

Extension state: AL HR LT LV MK

17P Request for examination filed

Effective date: 20070424

17Q First examination report despatched

Effective date: 20070605

AKX Designation fees paid

Designated state(s): AT BE BG CH CY CZ DE DK EE ES FI FR GB GR HU IE IT LI LU MC NL PL PT RO SE SI SK TR

RAP1 Party data changed (applicant data changed or rights of an application transferred)

Owner name: COLLISION TECHNOLOGY LLC

RAP1 Party data changed (applicant data changed or rights of an application transferred)

Owner name: COLLISION COMMUNICATIONS, INC.

GRAP Despatch of communication of intention to grant a patent

Free format text: ORIGINAL CODE: EPIDOSNIGR1

RIC1 Information provided on ipc code assigned before grant

Ipc: H04L 1/00 20060101AFI20130204BHEP

Ipc: H03M 13/29 20060101ALI20130204BHEP

Ipc: H04B 1/707 20110101ALI20130204BHEP

Ipc: H04B 1/7105 20110101ALI20130204BHEP

Ipc: H04L 25/03 20060101ALI20130204BHEP

Ipc: H04L 25/02 20060101ALI20130204BHEP

RIN1 Information on inventor provided before grant (corrected)

Inventor name: MACLEOD, ROBERT BRUCE

Inventor name: EGNOR, DIANNE E.

GRAS Grant fee paid

Free format text: ORIGINAL CODE: EPIDOSNIGR3

GRAA (expected) grant

Free format text: ORIGINAL CODE: 0009210

AK Designated contracting states

Kind code of ref document: B1

Designated state(s): AT BE BG CH CY CZ DE DK EE ES FI FR GB GR HU IE IT LI LU MC NL PL PT RO SE SI SK TR

REG Reference to a national code

Ref country code: GB

Ref legal event code: FG4D

REG Reference to a national code

Ref country code: CH

Ref legal event code: EP

REG Reference to a national code

Ref country code: AT

Ref legal event code: REF

Ref document number: 628618

Country of ref document: AT

Kind code of ref document: T

Effective date: 20130915

REG Reference to a national code

Ref country code: IE

Ref legal event code: FG4D

REG Reference to a national code

Ref country code: DE

Ref legal event code: R096

Ref document number: 602004043083

Country of ref document: DE

Effective date: 20131010

REG Reference to a national code

Ref country code: NL

Ref legal event code: VDEP

Effective date: 20130821

Ref country code: AT

Ref legal event code: MK05

Ref document number: 628618

Country of ref document: AT

Kind code of ref document: T

Effective date: 20130821

PG25 Lapsed in a contracting state [announced via postgrant information from national office to epo]

Ref country code: CY

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20130703

Ref country code: SE

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20130821

Ref country code: PT

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20131223

Ref country code: AT

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20130821

PG25 Lapsed in a contracting state [announced via postgrant information from national office to epo]

Ref country code: BE

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20130821

Ref country code: FI

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20130821

Ref country code: PL

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20130821

Ref country code: GR

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20131122

Ref country code: SI

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20130821

PG25 Lapsed in a contracting state [announced via postgrant information from national office to epo]

Ref country code: CY

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20130821

PG25 Lapsed in a contracting state [announced via postgrant information from national office to epo]

Ref country code: DK

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20130821

Ref country code: CZ

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20130821

Ref country code: SK

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20130821

Ref country code: EE

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20130821

Ref country code: RO

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20130821

Ref country code: NL

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20130821

REG Reference to a national code

Ref country code: CH

Ref legal event code: PL

PG25 Lapsed in a contracting state [announced via postgrant information from national office to epo]

Ref country code: MC

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20130821

Ref country code: ES

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20130821

Ref country code: IT

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20130821

PLBE No opposition filed within time limit

Free format text: ORIGINAL CODE: 0009261

STAA Information on the status of an ep patent application or granted ep patent

Free format text: STATUS: NO OPPOSITION FILED WITHIN TIME LIMIT

REG Reference to a national code

Ref country code: IE

Ref legal event code: MM4A

26N No opposition filed

Effective date: 20140522

PG25 Lapsed in a contracting state [announced via postgrant information from national office to epo]

Ref country code: LI

Free format text: LAPSE BECAUSE OF NON-PAYMENT OF DUE FEES

Effective date: 20130930

Ref country code: CH

Free format text: LAPSE BECAUSE OF NON-PAYMENT OF DUE FEES

Effective date: 20130930

Ref country code: IE

Free format text: LAPSE BECAUSE OF NON-PAYMENT OF DUE FEES

Effective date: 20130908

REG Reference to a national code

Ref country code: DE

Ref legal event code: R097

Ref document number: 602004043083

Country of ref document: DE

Effective date: 20140522

PG25 Lapsed in a contracting state [announced via postgrant information from national office to epo]

Ref country code: TR

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20130821

PG25 Lapsed in a contracting state [announced via postgrant information from national office to epo]

Ref country code: LU

Free format text: LAPSE BECAUSE OF NON-PAYMENT OF DUE FEES

Effective date: 20130908

Ref country code: BG

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20130821

Ref country code: HU

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT; INVALID AB INITIO

Effective date: 20040908

REG Reference to a national code

Ref country code: FR

Ref legal event code: PLFP

Year of fee payment: 12

REG Reference to a national code

Ref country code: FR

Ref legal event code: PLFP

Year of fee payment: 13

REG Reference to a national code

Ref country code: FR

Ref legal event code: PLFP

Year of fee payment: 14

REG Reference to a national code

Ref country code: FR

Ref legal event code: PLFP

Year of fee payment: 15

PGFP Annual fee paid to national office [announced via postgrant information from national office to epo]

Ref country code: FR

Payment date: 20180924

Year of fee payment: 15

Ref country code: DE

Payment date: 20180920

Year of fee payment: 15

PGFP Annual fee paid to national office [announced via postgrant information from national office to epo]

Ref country code: GB

Payment date: 20180919

Year of fee payment: 15

REG Reference to a national code

Ref country code: DE

Ref legal event code: R119

Ref document number: 602004043083

Country of ref document: DE

PG25 Lapsed in a contracting state [announced via postgrant information from national office to epo]

Ref country code: DE

Free format text: LAPSE BECAUSE OF NON-PAYMENT OF DUE FEES

Effective date: 20200401

GBPC Gb: european patent ceased through non-payment of renewal fee

Effective date: 20190908

PG25 Lapsed in a contracting state [announced via postgrant information from national office to epo]

Ref country code: FR

Free format text: LAPSE BECAUSE OF NON-PAYMENT OF DUE FEES

Effective date: 20190930

Ref country code: GB

Free format text: LAPSE BECAUSE OF NON-PAYMENT OF DUE FEES

Effective date: 20190908