EP3544005A1 - Audio encoder, audio decoder, audio encoding method and audio decoding method for dithered quantization for frequency-domain speech and audio coding - Google Patents

Audio encoder, audio decoder, audio encoding method and audio decoding method for dithered quantization for frequency-domain speech and audio coding Download PDF

Info

Publication number
EP3544005A1
EP3544005A1 EP18187597.2A EP18187597A EP3544005A1 EP 3544005 A1 EP3544005 A1 EP 3544005A1 EP 18187597 A EP18187597 A EP 18187597A EP 3544005 A1 EP3544005 A1 EP 3544005A1
Authority
EP
European Patent Office
Prior art keywords
spectral
samples
audio
sample
encoded
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
EP18187597.2A
Other languages
German (de)
French (fr)
Other versions
EP3544005B1 (en
Inventor
Johannes Fischer
Tom BÄCKSTRÖM
Sneha DAS
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Fraunhofer Gesellschaft zur Forderung der Angewandten Forschung eV
Original Assignee
Fraunhofer Gesellschaft zur Forderung der Angewandten Forschung eV
Friedrich Alexander Univeritaet Erlangen Nuernberg FAU
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Fraunhofer Gesellschaft zur Forderung der Angewandten Forschung eV, Friedrich Alexander Univeritaet Erlangen Nuernberg FAU filed Critical Fraunhofer Gesellschaft zur Forderung der Angewandten Forschung eV
Publication of EP3544005A1 publication Critical patent/EP3544005A1/en
Application granted granted Critical
Publication of EP3544005B1 publication Critical patent/EP3544005B1/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS OR SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING; SPEECH OR AUDIO CODING OR DECODING
    • G10L19/00Speech or audio signals analysis-synthesis techniques for redundancy reduction, e.g. in vocoders; Coding or decoding of speech or audio signals, using source filter models or psychoacoustic analysis
    • G10L19/04Speech or audio signals analysis-synthesis techniques for redundancy reduction, e.g. in vocoders; Coding or decoding of speech or audio signals, using source filter models or psychoacoustic analysis using predictive techniques
    • G10L19/16Vocoder architecture
    • G10L19/18Vocoders using multiple modes
    • G10L19/22Mode decision, i.e. based on audio signal content versus external parameters
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS OR SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING; SPEECH OR AUDIO CODING OR DECODING
    • G10L19/00Speech or audio signals analysis-synthesis techniques for redundancy reduction, e.g. in vocoders; Coding or decoding of speech or audio signals, using source filter models or psychoacoustic analysis
    • G10L19/0017Lossless audio signal coding; Perfect reconstruction of coded audio signal by transmission of coding error
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS OR SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING; SPEECH OR AUDIO CODING OR DECODING
    • G10L19/00Speech or audio signals analysis-synthesis techniques for redundancy reduction, e.g. in vocoders; Coding or decoding of speech or audio signals, using source filter models or psychoacoustic analysis
    • G10L19/02Speech or audio signals analysis-synthesis techniques for redundancy reduction, e.g. in vocoders; Coding or decoding of speech or audio signals, using source filter models or psychoacoustic analysis using spectral analysis, e.g. transform vocoders or subband vocoders
    • G10L19/032Quantisation or dequantisation of spectral components

Abstract

An audio encoder for encoding an audio signal, wherein the audio signal is represented in a spectral domain, is provided. The audio encoder comprises a spectral envelope encoder (110) configured for determining a spectral envelope of the audio signal and for encoding the spectral envelope. Moreover, the audio encoder comprises a spectral sample encoder (120) configured for encoding a plurality of spectral samples of the audio signal. The spectral sample encoder (120) is configured to estimate an estimated bitrate needed for encoding for each spectral sample of one or more spectral samples of the plurality of spectral samples depending on the spectral envelope. Moreover, the spectral sample encoder (120) is configured to encode each spectral sample of the plurality of spectral samples, depending on the estimated bitrate needed for encoding for the one or more spectral samples, according to a first coding rule or according to a second coding rule being different from the first coding rule.

Description

  • The present invention relates to audio encoding, audio processing and audio decoding, and in particular, to an audio encoder, an audio decoder, an audio encoding method and an audio decoding method for dithered quantization for frequency-domain speech and audio coding.
  • A common issue in coding speech and audio in the frequency domain, which appears with decreasing bitrate, is that quantization levels become increasingly sparse. With low accuracy, high-frequency components are typically quantized to zero, which leads to a muffled output signal and musical noise. Bandwidth extension and noise-filling methods attempt to treat the problem by inserting noise of similar energy as the original signal, at the cost of low signal to noise ratio.
  • State-of-the art codecs, such as 3GPP Enhanced Voice Services (EVS) and MPEG Unified speech and audio coding (USAC) use frequency domain coding in their intermediate and high bitrate ranges, but revert to time-domain coding at lower bitrates [1], [2], [4]. The reason is that the scalability of frequency-domain codecs in terms of coding efficiency at low bitrates remains a bottleneck even if they provide several other advantages such as low algorithmic complexity. A symptom of the issue is that frequency-domain codecs tend to quantize low-energy areas to zero, which further reduces their energy. This leads to a muffled character in the quantized output, since high-frequency components often have low energy and are thus zeroed (see Fig. 2).
  • Fig. 2 illustrates a mean energy of perceptually weighted and normalized MDCT-spectra over the TIMIT database, for original signal (thick line), conventional quantization (dotted), dithered (dashed), as well as dithered in combination with Wiener filtering (crosses) and energy matching (thin line). Quantization was scaled to match a bitrate of 13.2 kbit/s.
  • The problem of uniform quantization, in the conventional application, is that if the quantization bins are zero-centered, then the energy of the quantized signal decreases with decreasing accuracy. Alternatively, with off-center quantization the average energy can be retained, but limited in bit-rate to above 1 bit/sample, since the sign has to be transmitted. Moreover, at the extreme, at low bitrates, non-zero values can require so many bits to encode in the entropy coder, that one cannot ever afford to transmit them.
  • Entropy coding with uniform quantization therefore does not scale well to bitrates below 1 bit/sample.
  • This problem has been addressed in prior works primarily with two approaches. Firstly, one can encode high-frequency regions with bandwidth extension methods, where the objective is to retain the spectral magnitude envelope of the original signal, but sacrifice phase-information and other fine-structure, such that bitrate is limited [1]. Sometimes such methods also copy spectral structures from lower frequencies (copy-up) since the fine-structures are generally similar.
  • Secondly, with a method known as noise filling, one can insert noise in areas of the spectrum which have been quantized to zero such that absence of energy is avoided [64]. Both approaches thus aim to retain energy at a similar level as the original signal, but they do not optimize signal-to-noise ratio. A recent improvement, known as intelligent gap filling, combines these methods by using both noise filling and copy-up [65].
  • Classical dithering algorithms however also include methods which can retain the signal distribution without reduction in signal to noise ratio [66]. Common dithering methods such as Floyd-Steinberg, are based on error-diffusion or randomization of quantization levels, such that quantization errors can be diffused without loss in accuracy [67]. Alternatively, one can modify quantization bin locations to retain the probability distribution of the original signal even after quantization and coding [68] or use lattice quantization to pack quantization more densely [18].
  • These approaches however do not address the issue of very low bitrates, where one cannot afford to encode anything else than the most likely quantization bin. Algebraic coding can be used also at very low bitrates, but its output is also very sparse and it is not applicable on all bitrates [1], [69]. A further alternative would be vector coding, which provides optimal coding efficiency also at very low bitrates. However, vector coding approaches are not easily scalable across bitrates. Moreover, their computational complexity becomes prohibitively high at higher bitrates and if the vector length is high [1], [19]. Vector coding is thus also not a scalable approach.
  • It would therefore be highly be appreciated, if improved concepts to address the above-described problems would be provided.
  • The object of the present invention is to provide improved concepts for audio encoding and audio decoding. The object of the present invention is solved by an audio encoder according to claim 1, by an audio decoder according to claim 8, by a system according to claim 15, by a method according to claim 16, by a method according to claim 17 and by a computer program according to claim 18.
  • An audio encoder for encoding an audio signal, wherein the audio signal is represented in a spectral domain, is provided. The audio encoder comprises a spectral envelope encoder configured for determining a spectral envelope of the audio signal and for encoding the spectral envelope. Moreover, the audio encoder comprises a spectral sample encoder configured for encoding a plurality of spectral samples of the audio signal. The spectral sample encoder is configured to estimate an estimated bitrate needed for encoding for each spectral sample of one or more spectral samples of the plurality of spectral samples depending on the spectral envelope. Moreover, the spectral sample encoder is configured to encode each spectral sample of the plurality of spectral samples, depending on the estimated bitrate needed for encoding for the one or more spectral samples, according to a first coding rule or according to a second coding rule being different from the first coding rule.
  • Moreover, an audio decoder for decoding an encoded audio signal is provided. The audio decoder comprises an interface configured for receiving an encoded spectral envelope of the audio signal and configured for receiving an encoded plurality of spectral samples of the audio signal. Moreover, the audio decoder comprises a decoding unit configured for decoding the encoded audio signal by decoding the encoded spectral envelope and by decoding the encoded plurality of spectral samples. The decoding unit is configured to receive or to estimate an estimated bitrate needed for encoding for each spectral sample of one or more spectral samples of the encoded plurality of spectral samples. Moreover, the decoding unit is configured to decode each spectral sample of the encoded plurality of spectral samples, depending on the estimated bitrate needed for encoding for the one or more spectral samples of the encoded plurality of spectral samples, according to a first coding rule or according to a second coding rule being different from the first coding rule.
  • Furthermore, a method for encoding an audio signal, wherein the audio signal is represented in a spectral domain, is provided. The method comprises:
    • Determining a spectral envelope of the audio signal and for encoding the spectral envelope. And:
    • Encoding a plurality of spectral samples of the audio signal.
  • Encoding the plurality of spectral samples of the audio signal is conducted by estimating an estimated bitrate needed for encoding for each spectral sample of one or more spectral samples of the plurality of spectral samples depending on the spectral envelope. Moreover, encoding the plurality of spectral samples of the audio signal is conducted by encoding each spectral sample of the plurality of spectral samples, depending on the estimated bitrate needed for encoding for the one or more spectral samples, according to a first coding rule or according to a second coding rule being different from the first coding rule.
  • Furthermore, a method for decoding an encoded audio signal is provided. The method comprises:
    • Receiving an encoded spectral envelope of the audio signal and for receiving an encoded plurality of spectral samples of the audio signal. And:
    • Decoding the encoded audio signal by decoding the encoded spectral envelope and by decoding the encoded plurality of spectral samples.
  • Decoding the encoded audio signal is conducted by receiving or by estimating an estimated bitrate needed for encoding for each spectral sample of one or more spectral samples of the encoded plurality of spectral samples. Moreover, decoding the encoded audio signal is conducted by decoding each spectral sample of the encoded plurality of spectral samples, depending on the estimated bitrate needed for encoding for the one or more spectral samples of the encoded plurality of spectral samples, according to a first coding rule or according to a second coding rule being different from the first coding rule.
  • Moreover, computer programs are provided, wherein each of the computer programs is configured to implement one of the above-described methods when being executed on a computer or signal processor.
  • Embodiments are based on the following findings:
    In the prior art, frequency-domain codecs use uniform quantization and arithmetic coding to efficiently encode spectral components, and such entropy codec is near-optimal at high bitrates.
  • The combination of uniform quantization and arithmetic coding saturates with decreasing bitrate, such that quality drops significantly below the theoretical accuracy/bitrate limit.
  • At low bitrates (<1bit/sample) encoding efficiency decreases; the quantized spectrum becomes increasingly sparse which causes a muffled character to the output sound as well as musical noise. Several methods have been introduced to treat this problem, including noise filling and bandwidth extension.
  • Concepts which come close to the theoretical maximum efficiency, but with low computational complexity are provided
    Dithering methods have been used in other fields of signal processing, but not so much in speech and audio coding. Dithered coding concepts which use random rotations and 1-bit quantization are employed. In embodiments, such dithered coding concepts are applied in frequency-domain speech and audio coding.
  • The dithered coding concepts work well for low bitrates (<1bit/sample) whereas conventional entropy coding concepts work efficiently for high bitrates (>1bit/sample). In embodiments, a hybrid coder is provided, which may, for example, apply a threshold at 1bit/sample to choose between the two codecs. An estimate of the bit-consumption for individual samples may, for example, be extracted from the LPC-based arithmetic codec which is used in 3GPP EVS [48].
  • In embodiments, hybrid coding between uniform quantization with arithmetic coding (or other entropy codec, or other quantization) and dithered coding is conducted.
  • In particular, for dithered coding, randomization may, e.g., be based on orthonormal rotations and 1-bit coding (sign quantization), following [70].
  • In particular, 1-bit coding, where 1-bit coding (sign quantization) may e.g., be used on a select B samples of a vector length of N, to obtain a desired bit-consumption of B bits, which is equivalent with B/N bits per sample.
  • According to embodiments, hybrid coding is conducted such that the codec is chosen based on expected bit-rate of samples.
  • In particular, a bit-rate may, e.g., be estimated from an envelope model such as the LPC, following [48].
  • Optionally, one may, e.g., have multiple thresholds on the bit-rate to split the samples into categories of different accuracy, to allow for noise shaping.
  • In embodiments, dithered quantization and/or coding is conducted such that the output is post-filtered after reconstruction.
  • In particular, post-filtering based on a minimum mean square error minimization, or based on scaling output energy to match the (estimated) original energy or based on other scaling or optimization or a combination of these may, e.g., be conducted.
  • In particular, post-filtering may, e.g., be chosen to optimize a perceptual quality and/or a signal to noise ratio and/or another quality criteria.
  • According to embodiments, an audio encoder or a method of audio encoding or a computer program for implementing the method of audio coding using a hybrid coding as described in this application is provided.
  • An audio decoder or a method of audio decoding or a computer program for implementing the method of audio decoding using a hybrid decoding as described in this application is provided.
  • In particular embodiments, dithered quantization for frequency-domain speech and audio coding relating to a hybrid coding scheme for low-bitrate coding is employed, where low energy samples may, e.g., be quantized using dithering, instead of the conventional uniform quantizer. For dithering, one bit quantization in a randomized sub-space may, e.g., be applied.
  • In embodiments, a hybrid coder is provided, which applies a threshold at 1 bit/sample to choose between the two codecs. An estimate of the bit-consumption for individual samples may, e.g., be be extracted from the LPC-based arithmetic codec which is used in 3GPP EVS [48].
  • It is to be noted that described concepts equally relate to an apparatus for encoding and also apply for the apparatus for decoding where appropriate.
  • Some embodiments are based on using dithering instead of the conventional uniform quantizer. For dithering, one bit quantization in a randomized sub-space is applied.
  • In the following, embodiments of the present invention are described in more detail with reference to the figures, in which:
  • Fig. 1a
    illustrates an audio encoder according to an embodiment.
    Fig. 1b
    illustrates an audio decoder according to an embodiment.
    Fig. 1c
    illustrates a system according to an embodiment.
    Fig. 2
    illustrates a mean energy of perceptually weighted and normalized MDCT-spectra over the TIMIT database according to an embodiment.
    Fig. 3
    illustrates a diagram of the speech and audio encoder according to an embodiment.
    Fig. 4
    illustrates hybrid coding of spectral components according to an embodiment.
    Fig. 5
    illustrates collated histograms of K = 10000 vectors of unit variance Gaussian input, quantized by uniform quantization and entropy coding as well as the provided dithered coder (N = 32, _2), with 1 bit/sample according to an embodiment.
    Fig. 6
    depicts an illustration of performance for a typical speech spectrum at 13.2 kbits/s according to an embodiment.
    Fig. 7
    depicts results of a subjective MUSHRA listening test, comparing the provided 1 bit dithered quantizers with conventional arithmetic coding, as well as a synthetic dithering serving as an anchor according to an embodiment.
    Fig. 8
    illustrates differential scores of a subjective MUSHRA listening test, comparing the provided 1 bit dithered quantizers with conventional arithmetic coding, as well as a synthetic dithering serving as an anchor according to an embodiment.
    Fig. 9
    illustrates a histogram of relative differences in bit-consumptions when using the normal and Laplacian distributions.
    Fig. 10
    illustrates differential AB scores and their 95% confidence intervals of a comparison listening test.
    Fig. 11
    illustrates a flow diagram of the randomization process.
    Fig. 12
    illustrates a structure of an encoder and of a decoder of one node of the distributed speech and audio codec.
    Fig. 13
    illustrates an encoder/decoder structure of the distributed speech and audio codec with N encoder nodes and a single decoder node.
    Fig. 14
    depicts an illustration of histograms of the normalized covariance for different orthogonal matrices.
    Fig. 15
    depicts normalized histograms of the output samples after M consecutive randomizations.
    Fig. 16
    illustrates a convergence of distribution with increasing number of rotations to the normalized Gaussian and the normalized Laplacian.
    Fig. 17
    illustrates results of a MUSHRA test, given for different items and an average over items.
    Fig. 18
    illustrates the difference scores of the performed MUSHRA test of Fig. 17.
  • Embodiments of the present invention are based on particular combinations of first coding concepts and of second coding concepts.
  • For providing a better understanding of the embodiments of the present invention at first, the first coding concepts and the second coding concepts are described.
  • In the following, the first coding concepts are described. One concept or two or more concepts or all concepts of the first coding concepts may be referred to as a first coding rule.
  • The first coding concepts relate to arithmetic coding of speech and audio spectra using TCX based on linear predictive spectral envelopes.
  • Unified speech and audio codecs often use a frequency domain coding technique of the transform coded excitation (TCX) type. It is based on modeling the speech source with a linear predictor, spectral weighting by a perceptual model and entropy coding of the frequency components. While previous approaches have used neighbouring frequency components to form a probability model for the entropy coder of spectral components, it is proposed to use the magnitude of the linear predictor to estimate the variance of spectral components.
  • Since the linear predictor is transmitted in any case, this method does not require any additional side info. Subjective measurements show that the first coding concepts give a statistically significant improvement in perceptual quality when the bit-rate is held constant. Consequently, the first coding concepts have been adopted to the 3GPP Enhanced Voice Services speech coding standard.
  • At first, an introduction to the first coding concepts is provided.
  • Speech and audio coding technologies applied in modern standards such as MPEG USAC, G.718, AMR-WB+ and, importantly, the ETSI 3GPP Enhanced Voice Services, use multiple modes such as time-domain coding with ACELP and frequency-domain coding with TCX to gain efficient coding for as many signal types as possible [2], [57], [58], [59].
  • Generally, time-domain coding provides superior performance for signals with rapidly changing character and temporal events, such as spoken consonants, applause and percussive signals. Coding in the frequency-domain, on the other hand, is more effective for stationary signals such as harmonic music signals and sustained voiced speech sounds.
  • In the following, a focus is put on coding in the frequency-domain using models of the spectral envelope. Observe that there are two distinct types of spectral envelope models in classical literature and technologies; First of all, dedicated speech codecs are generally based on linear predictive models, which model the spectral energy envelope using an IIR filter. In contrast, classical audio codecs such as MP3 and the AAC family model the perceptual masking envelope [22], [60].
  • While these two envelopes do have many common features - their peaks and valleys are located at the same general frequency areas - the magnitude of peaks and valleys as well as the overall spectral tilt are very different. Roughly speaking, masking envelopes are much smoother and exhibit smaller variations in magnitude then the energy envelopes.
  • AAC-type codecs use the perceptual masking model to scale the spectrum such that the detrimental effect of quantization on spectral components has perceptually the same expected magnitude in every part of the spectrum [22]. To allow efficient coding of the perceptual spectrum, these codecs then apply entropy coding of the frequency components. For higher efficiency, the arithmetic coder can use the neighbouring spectral components to determine the probability distribution of the spectral components, such as in USAC [57], [61].
  • Speech codecs on the other hand use energy envelopes as a signal model and apply a perceptual weighting filter, much like the perceptual masking model, on top.
  • The assumptions rely on the fact that the spectral envelope, as described by the linear predictive model, provides information of the energy envelope of the spectrum. Since it thus describes the energy distribution of the spectral components, it can be used to describe their probability distributions. This distribution can, in turn, be used to design a highly efficient arithmetic coder for the spectral components. Since the linear predictive model is generally transmitted also for TCX frames, this spectral envelope information comes without additional side-information. In contrast to AAC-type codecs, an explicit source model may, e.g., be used in the form of the linear predictor, and in difference to TCX-type codecs, an adaptive probability distribution may, e.g., be used for the arithmetic codec derived from the magnitude of the linear predictor. A signal adaptive model of the probability distributions of spectral components is described based on the linear predictive model. The goal is to obtain a fixed bit-rate arithmetic coder applicable in speech and audio codecs which use linear predictive modeling. Moreover, the objective is to design a generic method which is efficient on a variety of bit-rates and bandwidths.
  • The encoder of the first coding concepts may, e.g., be configured to conduct three steps. First, the perceptually weighted linear predictor may, e.g., be used as a model for the shape of the perceptually weighted spectrum. Since this envelope does not contain information of the signal magnitude, the envelope is scaled such that the expected bit-consumption of a signal, whose variance follows the envelope, matches the desired bit-rate. Second, the actual perceptually weighted spectrum is scaled and quantized such that the bit-rate matches the desired bit-rate, when using the envelope model. Finally, one can then encode the spectrum with an arithmetic coder. The decoder can then apply the same scaling of the envelope to decode the spectral lines.
  • Now, a modeling of the probability distribution of the first coding concepts is described.
  • Let A k 1
    Figure imgb0001
    be the samples of the discrete Fourier transform of hat linear predictive model which describes the short-time temporal structure and thus the spectral envelope of a signal spectrum Sk . The filter residual can be obtained by multiplying Sk with Ak to obtain the residual Xk = AkSk . Given that Ak is an efficient model of Sk , then Xk will be the spectrum of a white-noise signal. It follows that the expected energy of every frequency component k is constant σ x 2 = ε X k 2 .
    Figure imgb0002
  • Conversely, the expectation of the energy of the perceptual weighted signal Sk is σ s , k 2 = ε S k 2 = σ x 2 A k 1 2 .
    Figure imgb0003
  • For perceptual weighting of quantization errors, prior to quantization, the spectrum is weighted by a perceptual masking model Wk . It follows that the expected energy of the weighted spectrum Yk = WkSk is σ y , k 2 = ε W k S k 2 = σ x 2 W k A k 1 2 .
    Figure imgb0004
  • This relation quantifies the relative energy of spectral components and can be used as an estimate of their relative variance in the design of models of the probability distribution of weighted spectral components.
  • A probability distribution model for the individual spectral components can then be chosen. The most obvious candidates are either the normal or the Laplacian distribution, since they are both simple to implement and commonly known to be fairly accurate models of speech and audio signals. To determine which distribution fits the approach better, the following experiment has been conducted.
  • As material for the test, the 16 critical items of speech, mixed and music material used in the standardization of MPEG USAC [57] were used. The material was resampled to 12.8kHz and windowed with sine-windows, and transformed to the frequency domain using the MDCT, with 256 frequency bands and full-overlap. Linear predictive models of order 16 were estimated for each frame using a Hamming window of length 30ms, centered to align with the MDCT windows. The spectrum Sk of each frame was whitened in the frequency domain with the corresponding envelope model Ak , and perceptually weighted with Wk , to obtain the perceptually white spectrum k = AkWkSl. The spectrum was then scaled by the standard deviation, to remove the effect of signal gain, whereby variance of the scaled signal was unity. The bit-consumption of both distributions was then estimated in each frame by b = k log 2 p X ˜ k .
    Figure imgb0005
    where the probability with the distributions is estimated p X ˜ k = 2 exp X ˜ k 2
    Figure imgb0006
    for the Laplacian distribution and p X ˜ k = 1 2 π exp X ˜ 2 2 2
    Figure imgb0007
    for the normal distribution.
  • Finally, the bit-consumption of all frames in the data for both the normal and Laplacian distributions were calculated.
  • The histogram of the relative difference in bit-consumption between the distributions is shown in Fig. 9.
  • Fig. 9 illustrates a histogram of relative differences in bit-consumptions bn and bl when using the normal and Laplacian distributions, respectively, as measured by r = bn□bl (bn+bl)=2 .
  • It can be observed that in a majority of cases, modeling with a normal distribution requires more bits than with a Laplacian. Moreover, importantly, encoding with the Laplacian never yields a large increase in bit-consumption, whereas it sometimes gives a big reduction in bit-rate. On average, the Laplace distribution gave a bit-consumption 6% smaller than the normal distribution. In addition, the highest gains where found for frames with stationary harmonic content, where the spectrum is sparse, which is also exactly the frames where TCX should be used. It is thus concluded that in an application, a Laplacian is better than a normal distribution.
  • Since the target is to design a codec with a fixed bit-rate, it is then necessary to scale the weighted envelope such that its expected bit-consumption matches the target bit-rate. A Laplacian random variable with variance σ 2 has a bit-consumption expectation of 1 + log2 (), where e is Euler's number. The variance for the magnitude of the weighted spectrum is then σ y , k 2 γ = γ 2 W k A k 1 2 ,
    Figure imgb0008
    where γ is the scaling coefficient. It follows that the expectation of bit-consumption of a spectrum with N coefficients is B = N + k = 0 N 1 log 2 eγσ y , k .
    Figure imgb0009
  • The above equation (1.5) can then be solved for: γ = 2 B 1 N e k = 0 N 1 σ y , k 1 / N .
    Figure imgb0010
  • It follows that when the spectral envelope is scaled by the given by equation (1.6), then the expected bit-rate is B for signals whose magnitude follows the envelope.
  • In the following, coding of spectral components of the first coding concepts is described.
  • The arithmetic coder of the first coding concepts is based on a Laplacian distribution, which is equivalent with a signed exponential distribution. To simplify the process, one can thus first encode the absolute magnitude of spectral lines and for non-zero values then also encode their sign with one bit.
  • The absolute magnitude can thus be encoded using an arithmetic coder with an exponential distribution. If a spectral component is quantized to an integer value q, then the original value has been in the interval q 1 2 , q + 1 2 ,
    Figure imgb0011
    whose probability is given by p Y ^ k = e λ q 1 2 e λ q + 1 2 = e + λ 2 e λ 2 e λq = C e λq ,
    Figure imgb0012
    where λ = 2 / σ k ,
    Figure imgb0013
    the scalar σk is the standard deviation of the kth frequency component and C = e + λ 2 e λ 2
    Figure imgb0014
    is a constant. Using this probability, a standard arithmetic coder for the spectral lines [62] can then be designed. With this arithmetic coder each frequency component can then be consecutively encoded from the low to high frequencies.
  • Since speech and audio spectra are often dominated by low-frequency content, at low bit-rates it is a common that large sections of the high-frequency spectra is sparse or zero.
  • To improve efficiency, trailing zeros can therefore be discarded. The decoder may, e.g., be set to decode the spectra until the maximal bit-consumption is reached and set any remaining spectral components to zero. The encoder thus encodes the spectrum up to the last non-zero frequency. Informal experiments showed that with this approach the number of encoded frequency components is often reduced by 30 to 40%.
  • To use the allocated bits efficiently, the spectrum must be scaled to obtain the highest accuracy which can be encoded with the available bits. As an initial guess for the scaling of the input spectrum Yk , one can scale it such that its energy matches the energy of the envelope that was scaled to the desired bit-rate. The optimal scaling can then be determined in a rate-loop implemented as a binomial search. Informal experiments showed that the best scaling can usually be found within 5 iterations.
  • In the following, implementation details of the first coding concepts are described.
  • The objective of the arithmetic coder is to encode the spectrum into the bit-stream as efficiently as possible. To encode and decode the bit-stream, the numerical operations must be implemented with fixed-point operations such that differences in numerical accuracy across different platforms will not change the outcome. A 14 bit integer representation stored in a 16 bit integer variable can be chosen, which allows for a sign bit and avoids problems in the last bit due to differences in rounding.
  • A related problem in implementation of the arithmetic coder is that when the standard deviation σk . is small compared to the quantized magnitude |k |, then the probability p(|k | will become so small that it is rounded to zero in the 14 bit representation. In other words, the numerical accuracy is exhausted and one would be unable to encode such values. To overcome this problem, the memory-lessness property of exponential distributions can be used. This principle states that p Y ^ k > r + s | Y ^ k > r = p Y ^ k > s .
    Figure imgb0015
  • Here, if r is the threshold above which the numerical representation saturates, then the probability p(|k | > r) can first be encoded and then, one can continue by encoding the probability p(|k | - r). By successively increasing r, it can be guaranteed that numerical saturation is avoided.
  • In the implementation of the rate-loop, the bit-stream should be output only for the final scaling coefficient. To reduce computational complexity, it is then sufficient to only estimate bit-consumption in the rate-loop and invoke the arithmetic coder only with the optimal scaling.
  • The bit-consumption of the spectrum for a given scaling coefficient γ can be efficiently estimated as follows. Let the quantized value be k = round(γYk ) whereby the total bit-consumption B(γ) can be calculated by B γ = k log 2 p Y ^ k = log 2 k p Y ^ k .
    Figure imgb0016
  • Instead of calculating the logarithm for every component, a logarithm of the product Π kp(k ) can thus be calculated, and thereby, computational complexity can be reduced. It should be noted, however, that the product can become a very large number, whereby the calculation of the product shall be implemented in a representation where it can be guaranteed that an overflow cannot occur.
  • Once the best possible scaling for quantization has been achieved, one can encode the spectral lines. However, since the rate-loop described above only approximates actual arithmetic coding, it may sometimes happen that the bit-budget is slightly exceeded. A safe-guard against this improbable event is to reduce the magnitude of the last encoded line until the actual bit-consumption remains within the budget.
  • Now, conducted experiments regarding the first coding concepts are described.
  • To determine the performance of the first coding concepts, a subjective AB comparison listening test [63] was performed. For this test, the codec was implemented according to the first coding concepts in a candidate version of the 3GPP Enhanced Voice Services speech and audio codec [2]. As comparison, an arithmetic coder derived from the MPEG Unified Speech and Audio Coder was used, which forms a probability model of spectral lines using neighbouring lines [57], [61]. The codecs were operating at a fixed bit-rate of 8 kbps in the wide-band mode. Bandwidth extension was encoded normally but its output was disabled to concentrate on differences of the core bandwidth. The double-blind AB comparison test was performed by 7 expert listeners in a silent room with high-quality headphones.
  • The test included 19 samples of mono speech, music, and mixed material. The differential results of the compare test are illustrated in Fig. 10, where a positive score indicates a subjective improvement of the codec of the first coding concepts over the reference codec using the arithmetic coder of MPEG USAC.
  • Fig. 10 illustrates differential AB scores and their 95% confidence intervals of a comparison listening test measuring the performance of the arithmetic coder in comparison to the coder from MPEG USAC at 8 kbps for wide-band signals.
  • These results show that the mean AB score indicates a statistically significant improvement of 0.48 points with 95% confidence. Moreover, no item had a statistically significant reduction in perceptual quality, but 6 out of 19 items had a statistically significant improvement. Note that it is remarkable that the improvement is statistically highly significant even with a limited number of listeners.
  • Now, a discussion and conclusions regarding the first coding concepts follow:
    The first coding concepts provide concepts for modeling the probability distribution of perceptually weighted frequency components of speech and audio signals using a model of the spectral envelope and the perceptual weighting function. Frequency-domain codecs based on the TCX concept model the spectral envelope using linear prediction, from which an estimate of the perceptual masking curve can be obtained. Since the linear predictor is transmitted in any case, the first coding concepts can be applied without transmission of additional side-information.
  • The first coding concepts use the spectral envelope model as a model of the speech source for construction of a probability model for the entropy coder. In contrast, conventional methods have used preceding frequency components to predict the magnitude of the current component [57], [61]. The conventional methods thus use an implicit source model, whereas the first coding concepts model the source explicitly.
  • It should be noted that in estimation of the bit-consumption of the spectral envelope, the theoretical entropy of Laplacian distributions may, e.g., be used, which is accurate only when the quantization accuracy is very high. The bias at lower bit-rates is due to fact that when a spectral line is quantized to zero, its sign does not need to be transmitted, whereby 1 bit is saved. When a significant part of the spectrum is quantized to zero, a rather large number of bits is saved, whereby the bit-consumption estimates are too high. Informal experiments, however, show that a more accurate estimate of bit-consumption increases complexity significantly, but that the impact on overall bit-consumption was marginal. Such more accurate and complex estimates of the bit-consumption of spectral envelopes could thus be avoided. Whereas here, only the spectral envelope associated with the linear predictor may, e.g., be used, it should be observed that speech codecs regularly use also other information which can be used to estimate spectral magnitude. Namely, long term prediction (LTP) is regularly used to model the fundamental frequency. The long term predictor can be used to model the comb-structure of spectra with a dominant fundamental frequency.
  • The presented results demonstrate that the first coding concepts improve perceptual quality at low bit-rates when the bit-consumption is kept constant. Specifically, subjective measurements with an AB test showed a statistically significant improvement. The coding scheme of the first coding concepts can thus be used to either increase quality at a given fixed bit-rate or decrease the bit-rate without losing perceptual quality. The presented approach is applicable in all speech and audio codecs which employ a frequency-domain coding of the TCX type, where a model of the spectral envelope is transmitted to the decoder. Such codecs include standards such as MPEG USAC, G.718 and AMR-WB+ [57], [58], [59]. In fact, the method has already been included in the ETSI 3GPP Enhanced Voice Services standard [2].
  • In the following, the second coding concepts are described. One concept or two or more concepts or all concepts of the second coding concepts may be referred to as a second coding rule.
  • The second coding concepts relate to fast randomization for distributed low-bitrate coding of speech and audio.
  • Efficient coding of speech and audio in a distributed system requires that quantization errors across nodes are uncorrelated. Yet, with conventional methods at low bitrates, quantization levels become increasingly sparse, which does not correspond to the distribution of the input signal and, importantly, also reduces coding efficiency in a distributed system.
  • The second coding concepts provide a distributed speech and audio codec design, which applies quantization in a randomized domain such that quantization errors are randomly rotated in the output domain. Similar to dithering, this ensures that quantization errors across nodes are uncorrelated and coding efficiency is retained.
  • The second coding concepts achieve fast randomization with a computational complexity of O N log N .
    Figure imgb0017
    The presented experiments demonstrate that the randomizations of the second coding concepts yield uncorrelated signals, that perceptual quality is competitive, and that the complexity of the second coding concepts is feasible for practical applications.
  • Now, an introduction to the second coding concepts is provided.
  • The second coding concepts describe an alternative method using dithered quantization, where the input signal is multiplied with a random rotation before quantization such that the quantization levels are obscured when the rotation is inverted for the output signal [70]. A similar approach is applied in the Opus codec [71], though only with Givens-rotations without permutations. A simple quantization such as 1 bit quantization is applied to obtain high performance at low complexity and very low bitrates [8]. The randomized quantization methods of the second coding concepts are unique in the way they allow quantization and coding of signals without a lower limit on bitrate, while simultaneously providing the best SNR per bit ratio. Concurrently, the second coding concepts provide the benefits of vector coding by joint processing of multiple samples, without significant penalty on complexity.
  • Digital compression of speech signals for transmission and storage applications, known as speech coding, is a classic topic within speech processing and modern speech coding standards achieve high efficiency in their respective application scenarios [1], [2], [3], [4], [5]. Though these standards are high-fidelity products, they are constrained to configurations with a single encoder. Designs which would allow using the microphones of multiple independent devices could improve signal quality, and moreover, it would allow a more natural interaction with the user-interface as the speaker would no more be constrained to a single device. If the codec can flexibly use all available hardware, then the user does not need to know which devices are recording, releasing mental capacity from attention to devices to the communication at hand.
  • Such an ideal user interface is possible only if devices cpoperate in the speech coding task. The aim is that, through cooperation, the acoustic signal should be flexibly captured and transmitted to one or several decoders or fusion centers. Clearly, one thus requires a distributed speech and audio codec. A distributed system however requires substantial modifications to existing codec designs; most notably, 1) the increase in algorithmic complexity due to added nodes becomes an issue and 2) concepts to ensure that each transmitted bit conveys unique information are needed. Specifically, conventional codec designs are based on an intelligent encoder and a simple decoder, whereby a majority of the computational complexity resides at the encoder. In a distributed system, the overall computational complexity increases linearly with both the encoder complexity as well as the number of nodes, whereby it is important to keep encoder complexity low to be able to use a large number of nodes. If one can move the main intelligence of the codec from the encoder to the decoder, then the overall complexity of the system would thus be much lower.
  • A majority of speech coding standards are based on the code-excited linear prediction (CELP) paradigm [1]. It is based on an analysis-by-synthesis loop, where the perceptual quality of a large number of different quantizations are evaluated to optimize output quality. While this approach provides the best quality for bitrate trade-off, its usefulness in distributed coding is limited by its computational complexity, rigid design and error propagation issues. Frequency domain methods, on the other hand, have not yet reached quite the same efficiency as CELP, but it is clear that coding in the frequency domain is computationally much simpler. Moreover, since most noise attenuation and spatial filtering methods are defined in the frequency domain [6], it will be straightforward to implement such methods if frequency-domain coding is used. Another issue is the amount of interaction between encoder nodes. Clearly communication between nodes requires some administration, whereby it would be beneficial to minimize or even avoid interaction between nodes if possible. If nodes only transmit data and interaction between nodes is avoided, then the overall system structure is simpler and the computational complexity required for said interaction is avoided. The question is thus whether interaction between nodes is required for coding efficiency. At high bit-rates (say 100 kbits s), very small differences in the signal, such as variations in delay, background noise or sensor noise, would be sufficient to make quantization noise between nodes uncorrelated [7], whereby each node will provide unique information. However, experience with lower bit-rates (such as 10 kbits s) has shown that low-energy areas of the signal are often quantized to zero, whereby quantization errors are perfectly correlated with the input signal [1]. Multiple nodes transmitting zeros would then convey no new information about the signal, whereby there is little advantage of using multiple devices.
  • An objective is to develop a distributed codec for speech and audio, where coding efficiency is optimized, but which can also be applied on any device, including simple mobile or even wearable devices with limited CPU and battery resources. For an overall design for such a method, see [8], [9]. The approach is based on randomizing the signal before quantization, such that quantization error expectations between devices are uncorrelated. It is assumed that the randomizer uses a random-number generator whose seed is communicated from the encoder to the decoder either offline or sufficiently seldom that it has a negligible effect on the bitrate. Overall, the randomizer in this context is similar to dithering and was inspired by the 1 bit quantization used in compressive sensing [10], [11].
  • Randomization has several distinct benefits in the codec of the second coding concepts:
    1. 1) In low-bitrate coding (below 10 kbits s), only a limited number of quantization levels are available which can be encoded with the available bits. With decreasing bitrate, the quantized signal distribution thus becomes increasingly sparse, granular and biased. By applying a randomization and its inverse before and after quantization, respectively, one can hide the undesirably sparse structure. Similarly as dithering, one can thus retain the signal distribution, without any penalty on the signal to noise ratio.
    2. 2) In perceptual audio coding a too low number of quantization levels for speech and audio signals leads to artifacts known as musical noise, where components which sporadically appear and disappear become coherent sound objects in their own right. A standard approach for avoiding musical noise in audio codecs is noise filling, a method similar to dithering, where noise is added to spectral areas quantized to zero [12]. In the described approach, by quantization in randomized domain, errors become incoherent and the reduction in SNR caused by noise filling can be avoided.
    3. 3) Randomization of the signal can also work as a component of encryption [13]. It provides diffusion in a similar way as the Hill cipher [14], that is, it distributes the contribution of the input vector evenly onto the bitstream.
    4. 4) In distributed coding,two alternative approaches can be applied. If nodes encode separate subspaces (or cosets in the vocabulary of distributed coding), then increasing the bitrate by 1 bit/sample yields a 6 dB improvement in quality. The downside in an ad-hoc network is that then the nodes have to negotiate which subspaces/cosets to transmit, which requires extract bandwidth, administration and increases the risk of eavesdropping. Moreover, spatio-temporal filtering such as beamforming is impossible if cosets do not overlap. On the other hand, if nodes are independent, then one can get a 3 dB improvement from a doubling of the number of nodes, as long as the quantization errors are uncorrelated [6]. Randomized quantization yields quantization errors which are uncorrelated, whereby in difference to conventional quantization, the 3 dB improvement when doubling the number of microphones is achieved.
    5. 5)When transmission errors corrupt some transmitted bits, conventional entropy coders (e.g. arithmetic coding) will loose all data after the first corrupted bit. With entropy coding according to the second coding concepts which is enabled by the randomizing scheme, there is no serial dependency of bits, whereby the signal can be reconstruct also when some bits are corrupted. The transmission errors will then be visible as noise in the reconstructed signal, which can be attenuated by conventional noise attenuation methods such as Wiener filtering [15], [16]. The details of these benefits are discussed in the following.
  • In comparison, conventional single-device quantization and coding methods all suffer from some constraints. Entropy coders such as Huffman or arithmetic coders with uniform quantization do not scale to very low bitrates (less than 2 bits/sample), since the output signal distribution becomes unnaturally sparse [1], [17]. Lattice quantization does reduce quantization error, but does not solve the issue of granularity at low bitrates and moreover, it does not easily lend itself to arbitrary probability distributions [1], [18]. Vector coding is optimal in accuracy and does not suffer much from sparsity, but computational complexity is high and it is challenging to encompass variable bitrates [19].
  • Moreover, achieving robustness to transmission errors is difficult with all of the above methods. Distributed source coding methods, on the other hand, do provide methods for optimal joint encoding [20], [21]. These methods make use of the correlation between the microphone signals by binning in order to reduce the rates using the results from distributed source coding. The most common way of implementing this is by using error-correcting codes, but in a practical setup due to complexity and other considerations, such implementations will be highly suboptimal, leading to higher complexity without significant gains. For these reasons, no focus is put on binning. Specifically, distributed source coding methods for ad-hoc networks are not available, which would include signal-adaptive perceptual modeling, which would solve the above mentioned problems with sparsity and which would simultaneously apply signal-adaptive source models. All of these properties are required features of a speech and audio codec to retain a competitive performance for the single-channel case.
  • The fields of speech and audio coding [1], [22], and distributed source coding e.g. [20], [21], are well-understood topics. Work on wireless acoustic sensor networks has however not yet made it to consumer products of the type discussed here [23]-[25]. Some works have addressed higher bitrates [26], or with the assumption that nodes are fully connected and co-operating [7], [27], though both approaches lack a perceptual model. Some early work do apply a perceptual model [28], [29], but do not include other elements of mainstream speech and audio codecs, such as source modeling. A somewhat similar problem is design of hands-free devices and the associated beamforming tasks [6], which can be applied in a distributed system [30], though methods usually require accurate synchronization of nodes to give competitive performance [31]. Similar methods have also been tried for hearing-aids, though distributed source coding gave there only a negligible improvement [32]. More generally, distributed processing of speech and audio can be also applied for classification and recognition tasks [33]-[35]. In comparison, the distributed scheme of the second coding concepts is now a complete system, with the exception of speech source modeling, may, e.g., be incorporated into the system, though this has not been studied intensely ([36], [37]).
  • Specifically, while a rudimentary entropy codec is not applied, in the current experiments, a model of the spectral magnitude envelope of harmonic structure nor spatio-temporal correlations is not included. While many of the prior works have admirably fine theoretical analyses, here, the design has no barriers against creating a practical system whose single-channel quality is near state-of-the-art while simultaneously providing a benefit when increasing the number of nodes. That is, by including the above mentioned source models, the single-channel performance should be similar to the performance of the TCX mode of the EVS standard [2]. It should be emphasized that not all details of best practices are included in lossy distributed source coding, since it has been opted to take incremental steps in order to retain the single-channel quality near the state-of-the-art.
  • The current contribution addresses the complexity bottleneck of the system, namely, the randomization and its inverse. The algorithmic complexity of generic quantization together with randomization is O N 2 .
    Figure imgb0018
    Moreover, at the decoder, the original approach required the inversion of an N × N matrix, which gives a complexity of O N 3
    Figure imgb0019
    using Gaussian elimination [38].
  • Algorithmic complexity shall be improved, and the randomization properties and coding efficiency shall be retained or improved as much as feasible. In addition to distributed coding, randomization and decorrelation are used also in many other fields of speech, audio and signal processing in general. For example, in upmixing of audio signals from a low to a higher number of channels, concepts for generating uncorrelated source signals are needed [39].
  • Randomization methods of the second coding concepts may find application in any such applications which require low-complexity methods for generation of uncorrelated signals. Notably, moreover, randomization can be used in single-channel codecs to diffuse unnaturally sparse quantization levels which appear at low bitrates.
  • In the following, randomized entropy coding of the second coding concepts is described.
  • The main objective of coding is to quantize and encode an input signal x R N × 1 ,
    Figure imgb0020
    with a given number of bits B, such that it can be decoded with highest accuracy possible. The objective of randomization, on the other hand, is to make sure that the resynthesized signal retains the continuous distribution of the original signal and that the quantization error is uncorrelated Gaussian noise. In other words, whereas quantization by construction yields a signal with a discrete distribution, a signal which follows a similar distribution as the original signal shall be obtained. Moreover, the aim is that if the signal is quantized and coded at multiple nodes, then the quantization errors of the outputs would be uncorrelated.
  • Fig. 11 illustrates a flow diagram of the randomization process, where P is a random (orthogonal) matrix and Q[·] is a quantizer.
  • It is needed to introduce randomness in the signal without reducing accuracy of the reconstruction. To achieve such randomness, three aspects of linear randomizing transforms shall be discussed (see Fig. 11). First, it is shown that orthonormal projections are optimal for the error criterion. Secondly, random permutations for diffusing information across the input vector are discussed. Finally, it is demonstrated that low-order random rotations can be used in block-matrices to diffuse quantization levels.
  • As objective design criteria for the randomization, the following concepts are used:
    1. 1) The accuracy of reconstruction is measured by the minimum mean square error min E[∥e2], where e = x - is the quantization error and is the quantized signal.
    2. 2) To measure the correlation between randomized vectors, the normalized covariance between the original x and its randomized counterpart Px is measured. If the randomization is effective, then the normalized covariance should behave like the normalized covariance x T y x y
      Figure imgb0021
      between two uncorrelated signals x and y. Specifically, the mean of the normalized covariance should be zero and its variance 1 N
      Figure imgb0022
      for Gaussian signals (see below for details).
    3. 3) The accuracy with which the distribution of the output signal follows the distribution of the input signal can be measured with the Kullback-Leibler (KL) divergence. However, since analytic analysis of divergences is difficult, the KL-divergence is applied only experimentally below.
    4. 4) Finally, algorithmic complexity is characterized with the Big-O notation.
  • Now, orthonormal randomization of the second coding concepts is considered.
  • To introduce randomness in the signal without compromising accuracy, the signal is multiplied with a random orthogonal matrix P before quantization. At the decoder, multiplication with the inverse PT is conducted (see Fig. 11). It is important that the transform is orthonormal, since it preserves signal energy, such that the transform provides perfect reconstruction and a minimal white noise gain. Specifically, let the output signal be = PT Q[Px] where Q[·] signifies quantization.
  • If the quantization is perfect, u = Q[u], then = PT Q[Px] = PT Px = x. In other words, the randomizing transform P does not corrupt the signal in any way; all information is retained and the signal can be perfectly reconstructed. Moreover, if the quantization error is v = u - , then the output error energy will be e 2 : = x x ^ 2 = x P T Q Px 2 = Px Q Px 2 = v 2 .
    Figure imgb0023
  • In other words, since for an orthogonal P: ∥e2 = ∥Pe2, it follows that quantization error energy in the transform domain is exactly equal to the error in the output domain. If the constraint would be relaxed from orthogonal matrices, and consider matrices P whose samples are uncorrelated and have unit variance, then E[PT P] = I would result, and the matrices would be orthogonal with respect to the expectation.
  • However, as is known from random matrix theory [40], the eigenvalue distribution of such matrices is significantly off unity. It follows that PT would not be an accurate inverse of the transform but one would have to use the actual inverse P -1 or a pseudo-inverse. Consequently, the inverse transform P -1 would emphasize the quantization errors corresponding to small eigenvalues, whereby the inverse transform would, on average, increase the error energy. Orthogonal random matrices are thus preferable to random matrices since they provide perfect reconstruction and unit white noise gain. Orthogonal matrices have also computational benefits with respect to noise attenuation at the decoder side, see below.
  • Now, random permutations of the second coding concepts is considered.
  • Permutations are computationally fast operations which correspond to orthogonal matrices, whereby their use in randomization is interesting. As a matter of definition, a permutation perfectly diffuses input information over the output vector if the output location of all input samples are uniformly distributed over the whole vector. Specifically, if an input sample ξh has an input location h ∈ [0, N - 1], then the probability that it will appear at location k after permutation is p k = 1 N ,
    Figure imgb0024
    whereby the input and output locations are not correlated. The covariance cP of the original signal x and the permuted signal c P = x T y = x T Px = k = 1 N ξ k ξ q k .
    Figure imgb0025
    where q(k) is the permutation function. If defining the set of fixed points of the permutation as S = {k|k = q(k)}, that is, this is the set of samples which do not move due to the permutation, whereby c P = k S ξ k 2 + k S ξ k ξ q k .
    Figure imgb0026
  • While the expectation of the latter sum is zero E[ξkξh ] = 0, for kh, the former sum has a non-negative expectation. It follows that the expectation of the covariance has a non-negative bias E[cp] ≥ 0, which is in contradiction with the objective. The set S is, however, small when N is large, whereby the deviation is not large. Nevertheless, for the sake of completeness, the problem can be remedied.
  • Let Λ± be a diagonal matrix whose diagonal elements are randomly chosen as ±1. Clearly this matrix is also orthogonal. One can thus apply a randomization y = Λ± Px, whereby the correlation is c P = x T y = x T Λ ± Px = k = 1 N ± ξ k ξ q k .
    Figure imgb0027
  • If both signs have equal probability, then clearly Eξkξ q(k)] = 0 and E[cP ] = 0 as required. The combination of random signs and permutations thus decorrelates the signal in the sense that the covariance has zero mean and one simultaneously achieves perfect diffusion. Moreover, multiplication by Λ± can be readily generalized to orthogonal block-matrices, which are discussed below.
  • Random permutations can be easily generated at algorithmic complexity O N log N
    Figure imgb0028
    [41]. A heuristic approach is for example to apply a sort algorithm, such as merge sort, on a vector of random uncorrelated samples. The sort operation then corresponds to a permutation, which is uniformly distributed over the length of the vector.
  • Now, blockwise random rotations are considered.
  • Multiplication with matrices has in general an algorithmic complexity of O N 2 .
    Figure imgb0029
    To reduce complexity, consider N × N random orthogonal block-diagonal matrix rotations B of the B = Q 1 0 Q 2 0 Q K .
    Figure imgb0030
    where K is the number of blocks and Qk are random orthogonal matrices of size Nk × Nk such that k = 1 K N k = N .
    Figure imgb0031
    The complexity of the transform is O k = 1 K N k 2 .
    Figure imgb0032
    Clearly the random sign operation Λ± is a block-matrix of this form with Nk = 1 and K = N.
  • Specifically, consider size 2 × 2 orthogonal matrices Q of the form Q = cos α sin α sin α cos α .
    Figure imgb0033
  • The related covariance is c Q : = x T Qx = ξ 1 ξ 2 T cos α sin α sin α cos α ξ 1 ξ 2 = ξ 1 2 + ξ 2 2 cos α .
    Figure imgb0034
  • If α is uniformly distributed on α ∈ [0, 2π], then clearly cQ has zero mean, E[cQ ] = ∥x2 E[cos α] = 0 as desired. Moreover, if y = Qx = [η 1, η 2] T , then the parameters ηk follow the arcsine-distribution on the range |ηk |2 ≤ 2∥x2.
  • In other words, by applying 2 × 2 blocks of random rotations on the quantized signal ^u, one can diffuse the quantization levels to make the distribution of the output less sparse. Note that since the outputs of the 2 × 2 transforms have symmetric distributions, signrandomization by Λ± becomes redundant and can be omitted. The 2 × 2 matrices were chosen since they are simple to implement and any larger orthonormal rotation can be implemented as a combination of 2 × 2 rotations.
  • Diffusion of the quantization levels is however not yet complete; the arcsine-distribution is less spiky than the distribution of the quantized signal, but it is still far from the normal distribution.
  • To obtain an output distribution which better resembles the normal distribution, a sequence of permutations Pk and block rotations Bk is therefore applied as P = k = 1 M B k P k .
    Figure imgb0035
    where each block rotation Bk is of the form in equation (2.5) with 2 × 2 rotations Qk of the form in equation (2.6). Each consecutive randomization BkPk will then further diffuse the output distribution. A number of rotations M are experimentally determined such that the output distribution is sufficiently diffused. However, the algorithmic complexity of applying the above rotation will in any case be O MN
    Figure imgb0036
    while generation of such permutations has complexity O N log N .
    Figure imgb0037
  • Now, conventional randomization methods with respect to the second coding concepts are considered.
  • Generating random orthogonal matrices, with uniformly distributed rotations is not as easy as it would seem. With 2 × 2 matrices such as those in equation (2.6) one can get uniform distribution if α is uniformly distributed on [0, 2π], however, with N > 2 such heuristic approaches are not as simple anymore. The reason is easy to understand in the 3 × 3 case, where uniform rotations along each axis would yield a higher concentration of points around the poles of the sphere.
  • In the general case, however, the problem is equivalent with choosing N random points on the unit N-sphere, such that the corresponding unit vectors are orthonormal. One can thus choose N random vectors of size N × 1 from a distribution with spherical symmetry, and orthogonalize the set of vectors. A numerically stable and efficient algorithm which generates such orthonormal vectors is the QR-algorithm [38]. Specifically, at first, an N × N matrix X with uncorrelated and normally distributed samples with zero mean and equal variance is generated.
  • Secondly, the QR-algorithm is applied to obtain an orthogonal matrix PQR . The overall algorithmic complexity of this approach is O N 2
    Figure imgb0038
    [38], [41].
  • A simplification of the QR-algorithm is to apply Householder transformations with each of the columns of X [42]. While this approach is efficient for small matrices, in informal experiments it is found that for large matrices, the simplification does not provide a uniform distribution. The subgroup algorithm presented in [43], [44] has also been tried. Though this algorithm is faster than the Householder-based algorithm by a factor of 2, unfortunately however, it suffers from the same issues as the Householder based algorithm. The QR-algorithm applied on random matrices thus remains the high-quality and -complexity method for reference.
  • Now, algorithmic complexity of the second coding concepts is considered.
  • In application of each of the orthogonal matrices of the second coding concepts, there are two sources of algorithmic complexity; one which emerges from generation of the matrix and a second which is related to application of the matrix and its inverse. Furthermore, one should evaluate storage requirements for the generated matrices. In each case, it shall be assumed that one has access to a pseudo-random number generator, which produces pseudo-random sequence of scalars ξk with independent and identically distributed values from the uniform distribution on ξk ∈ [0, 1].
  • The simplest case is generation of random signs for the matrix Λ±. For each of the N diagonal samples, a random sign with equal probability is needed, which can be obtained by thresholding ξk at 0.5. Given the sequence ξk the algorithmic complexity for generation and application is thus O N .
    Figure imgb0039
  • For the block rotations, the number of 2 × 2 blocks is N/2, whereby N/2 random scalars are needed to generate the matrix B. Application of B involves N/2 multiplications by a 2 × 2 matrix at complexity O 2 N ,
    Figure imgb0040
    as well as N/2 evaluations of cos α and sin α at O N ,
    Figure imgb0041
    though evaluation of trigonometric functions can have a high constant multiplier for the complexity.
  • Application of permutations is straightforward; it is essentially a mapping of sample indices, whereby it does not involve computations other than moving operations, O N .
    Figure imgb0042
    Here, the permutation indices and the permuted vector have to be stored, whereby the storage requirements are O 2 N .
    Figure imgb0043
    Generation of permutations can be applied by sorting a vector of random scalars ξk with, for example, the merge sort algorithm [41]. It requires also a storage of O 2 N ,
    Figure imgb0044
    but not at the same time as the application of the permutation, whereby it does not add to the overall storage requirements.
  • The algorithmic complexity for generating the permutation is O N log N
    Figure imgb0045
    [41]. To generate a random orthogonal matrix, the QR-algorithm can be applied with arbitrary accuracy with an algorithmic complexity of O N 2
    Figure imgb0046
    and storage O N 2
    Figure imgb0047
    [38]. Application of the randomization and its inverse are then simply multiplications by a matrix and its transpose, both at complexity O N 2 .
    Figure imgb0048
    It however requires N 2 random scalars as input, whereby also the complexity of generating pseudo-random numbers becomes an issue. Moreover, the random values at the input should have rotational symmetry, whereby the uniformly distributed scalars ξk variables are not sufficient. It is thus necessary to apply a transform such as the inverse cumulative normal distribution on ξk to obtain normally distributed variables, which comes at a considerable extra computational cost. Each of the above algorithms assume that one has access to a sequence of pseudo-random numbers ξk . If one chooses to generate the randomization on-line, then the complexity of generating pseudo-random numbers shall be considered. The algorithmic complexity of generating pseudo-random numbers is in general linear O N
    Figure imgb0049
    with the number N of scalars to be generated [45]. A commonly used generator is the Mersenne-twister, though there are also lower-complexity versions available [46], [47]. The trade-off is that if the random sequence is not generated on-line, then it needs to be stored. In any case, it is assumed that the seed of the random sequence is communicated either off-line, or sufficiently seldom such that it does not induce a significant penalty on the bit-rate.
  • In summary, algorithmic complexity of generating random matrices is O MN log N ,
    Figure imgb0050
    while their application has O MN ,
    Figure imgb0051
    where M is the number of iterations (typically M = 4) and N is the vector length. If the random coefficients are not generated on-line but stored, then a storage of 32 M N is necessary.
  • A typical speech and audio codec, such as the TCX mode of the EVS, would use a step of 20 ms between windows [1], [2], whereby the spectra would be of length N = 256 at a sampling rate of 12.8 kHz, N = 320 at 16 kHz or N = 882 at 44.1 kHz.
  • A typical frequency domain codec will have no components which require a complexity more than O N log N .
    Figure imgb0052
    Since the randomization of the second coding concepts is also O N log N ,
    Figure imgb0053
    in terms of algorithmic complexity, this corresponds to conventional TCX codecs. The complexity bottleneck thus returns to the rate-loop of the entropy codec [1], [48].
  • In the following, applications in distributed coding of speech and audio of the second coding concepts are described.
  • As a demonstration of the randomizer of the second coding concepts, the randomized quantizer was applied for coding of the fine-spectral structure in a distributed speech and audio codec. The overall structure is similar to that of the TCX mode in 3GPP EVS [2] and the implemented codec structure is illustrated in Fig. 12.
  • Fig. 12 illustrates a structure of the (a) encoder and (b) decoder of one node of the distributed speech and audio codec. Dashed boxes indicate modules which were not included in current experiments to facilitate isolated evaluation of randomization coefficients, while working memory must be always at least 2N coefficients.
  • First, the MDCT time-frequency transform and half-sine windowing on the input signal [22] was applied to obtain spectral representations xk of each time-frame at a node k. Here the window length was 20 ms with 10 ms overlap, and the sampling rate was 12.8 kHz.
  • The sampling rate was chosen to match the core-rate of EVS in wide-band mode [2]. In EVS, the remaining bandwidth is coded with bandwidth-extension methods, to obtain a total sampling rate of 16 kHz.
  • Then, the signal envelope and perceptual model is analyzed, using the LPC-based approach as in EVS [1], [2]. The signal is then perceptually weighted and multiplied with random rotation matrices to obtain randomized vectors. As a last step of the encoder, the signal is quantized and coded.
  • A fixed-bitrate entropy coder with 2 bits/sample was used as follows: the distribution was split into four quantization cells such that each cell had equal probability and each input sample was quantized to the nearest quantization cell. The quantization levels are thus fixed and the system does not require a rate-loop.
  • This corresponds to entropy coding the spectral fine structure alone with a bitrate of 25.6 kbits/s. Perceptual envelopes and signal gain are usually transmitted with a rate in the range 2-3 kbits/s, whereby the overall bitrate is approximately 28 kbits s. This does not strictly qualify as a low-bitrate codec, but on the other hand, an explicit source model or a rate-loop is not implemented, here. The inventors' experience with the EVS codec suggests that inclusion of a source model (such as a fundamental frequency model) and a proper rate-loop, would reduce bitrate to below 10 kbits s without reduction in perceptual quality, whereby this experiment is representative for low-bitrate coding. Conversely, the bitrate was chosen such that it roughly corresponds to the accuracy achieved in TCX in the EVS standard well below 10 kbit/s.
  • At the decoder, the operations are reversed, with the exception of the perceptual model estimation, which is assumed to be transmitted. One should focus on the performance of the randomizer, whereby in the current application, an explicit source model, rate-loop or quantization the perceptual model was not included. The quality of the output signal could be further enhanced with noise attenuation techniques such as Wiener filtering [6], [49]. Here, however, noise attenuation was omitted such that tuning parameters can be avoided and the comparison of different methods can be kept fair.
  • Fig. 13 illustrates an encoder/decoder structure of the distributed speech and audio codec with N encoder nodes and a single decoder node.
  • To demonstrate performance in a multi-node scenario, a distributed codec was implemented as illustrated in Fig. 13, where each individual encoder node follows the configuration of Fig. 12.
  • As described above, the decoder can then contain the inverse randomizations and a "merge & enhance" -block can implement Wiener filtering independently from the randomization.
  • As long as the quantization errors are orthonormal and quantization accuracy is uniform, one would not gain anything from joint processing of the randomization and enhancement, whereby independent blocks give optimal performance. It was decided to not implement Wiener filtering here, since it would present additional perceptual tuning factors, whereby the design of a fair comparison would be difficult. Instead merely the mean of the two channels was used and only the objective quality of the two-channel case was considered.
  • In the following, conducted experiments with respect to the second coding concepts are described.
  • To evaluate the randomization methods of the second coding concepts, objective experiments corresponding to the performance measures described above were performed as well as subjective perceptual experiments with the distributed speech and audio codec described above.
  • Now, statistical properties of the second coding concepts are described.
  • To quantify the influence that randomization has on the coding error magnitude, N × N matrices with N = 100 were created, such that 1) the matrices Po were orthogonal P o T P o = I
    Figure imgb0054
    and 2) the matrices Pr were orthogonal with respect to the expectation E P r T P r = I .
    Figure imgb0055
    Specifically, a random number generator was used to produce uncorrelated, normally distributed samples with zero mean and variance 1 N ,
    Figure imgb0056
    which form the entries of Pr . It follows that E P r T P r = I .
    Figure imgb0057
    By applying the QR-algorithm on Pr , a random orthogonal matrix is then obtained, which is defined as Po .
  • K = 1000 vectors x of length N are generated, whose samples are uncorrelated and follow the normal distribution.
  • Each vector was randomized with the two matrices uo = Pox and ur = Prx, quantized with the sign quantization ûo =sign(uo ) and r = sign(ur ) and the randomization was reverted by x ^ o = P o T u ^ o
    Figure imgb0058
    and x ^ r = P r T u ^ r .
    Figure imgb0059
    As reference, quantization without randomization is used as x ref = sign(x). Finally, each of the vectors xo .xr and xref were individually scaled for minimum error, and the quantization error energy for each vector was calculated.
  • The results of this experiment are listed in Table I. Table I:
    Method Orth Rand None
    SNR (dB) 4.52 2.25 4.52
  • Table I illustrates a signal-to-noise ratio (SNR) of sign quantization with orthonormal randomization (Orth), randomization with a random matrix (Rand), and without randomization (None).
  • Clearly randomization with the random matrix Pr yields a much lower SNR upon reconstruction, whereas randomization with the orthogonal matrix Po has no influence on accuracy. The results thus exactly follow the results above and randomization by orthogonal matrices should always be used.
  • To determine the efficiency of randomization in terms of decorrelation, normalized covariances λ = x T Ax x 2
    Figure imgb0060
    are then calculated for the different orthogonal matrices A.
  • Fig. 14 depicts an illustration of histograms of the normalized covariance λ = x T Ax x 2
    Figure imgb0061
    for different N × N orthogonal matrices: The random permutation matrix P, the random sign matrix Λ±, the combination Λ± P as well as the block-matrix (with random 2 × 2 rotations) in combination with permutation BP, evaluated over K = 10 000 matrices. The dashed line indicates the theoretical distribution of normalized covariance between random Gaussian vectors, scaled to fit to each histogram.
  • In particular, Fig. 14 illustrates the histogram of K = 1000 iterations of the normalized covariances for matrices of different sizes N = {4, 8, 256} as well as for random permutations P, random signs Λ± and random permutations with random signs PΛ± As a reference, the theoretical distribution illustrated with a dashed line is used (see below for details).
  • It was observed that the distribution of the random permutation P follows the theoretical distribution (dashed line) at higher N. However, it is biased to positive values especially at lower N.
  • The random sign Λ± yields a covariances whose variance (width of histogram) is higher than the theoretical distribution. Clearly neither method is sufficient alone in decorrelating the signal.
  • The combination of random signs and a permutation Λ± P performs much better in that the bias to positive values is removed and the variance is similar to the theoretical distribution. At N = 4, however, it was observed that the input signal x cannot be treated as a random variable anymore, but since samples of x frequently get multiplied with itself (though with random sign), peaks were obtained in the histogram corresponding to ±1 and 0. The situation is further improved by replacing the random signs with block-matrices B, where the 2 × 2 blocks are calculated with uniformly distributed angles a. The peaks at ±1 and 0 for N = 4 have almost completely disappeared and the histogram nicely follows the theoretical distribution. Overall, it was found that the decorrelation performance of randomization improves with increasing vector length, as well as when using a combination of at least two orthogonal matrices.
  • The third objective performance measure is the ability of randomization to diffuse the quantization levels in the output signal. Specifically, the aim is that the distribution of the output is transformed from a sparse distribution to something which resembles the input distribution. It has already been found that application of random permutations and block-matrix rotations is an effective combination, whereby it is an aim is to evaluate how many such pairs have to be applied to get proper diffusion. To that end it is defined P M = k = 1 M B k P k .
    Figure imgb0062
    where Bk and Pk are random block-matrix rotations and permutations, respectively. Fig. 15 illustrates the output histogram when applying sign-quantization and the inverse rotation PTM for different values of M. The use of sign quantization has been chosen here, since it is the worst-case in terms of sparsity.
  • In particular, Fig. 15 depicts normalized histograms of the output samples after M consecutive randomizations, as well as the theoretical distributions of normalized Gaussian (NNorm, dashed line) and Laplacian (NLap, gray line), for a vector of length N = 16.
  • It can be observed in Fig. 15 that the original quantized signal has a sparse distribution, where all samples are ±1, but each consecutive randomization makes it resemble more the normalized Gaussian distribution (dashed line). Note that the normalized Gaussian is here the reference distribution, since the normalized covariance has a limited range (see below for details).
  • At M = 4 iterations the histogram has already converged to a unimodal distribution and at M = 16 the histogram is very close to the normalized Gaussian distribution. The rate of convergence depends, however, on the vector length N.
  • As a final test of statistical properties, the rate of convergence to the normalized Gaussian distribution was tested with increasing number of iterations and different vector lengths N. As a measure of convergence, the Kullback-Leibler divergence between the normalized Gaussian distribution and the histogram of the output was used. As reference, randomization based on the QR-algorithm was used. The results are illustrated in Fig. 16(a).
  • Fig. 16 illustrates a convergence of distribution with increasing number of rotations M to (a) the normalized Gaussian and (b) the normalized Laplacian, as measured by the Kullback-Leibler divergence, for different vector lengths N. As a reference, randomization with the QR-algorithm is depicted with crosses "×," representing a high-complexity and high-performance target level.
  • It can be seen that convergence is faster for the shorter vectors, as is to be expected, since in a large space one needs more rotations to reach all possible dimensions. The performance of the QR algorithm is illustrated with crosses '×' and it can be seen that after 16 iterations, for all vector lengths N, the randomizers of the second coding concepts have more or less reached the diffusion of the QR algorithm. In fact, for N = 4 and N = 8, the performance saturates already after 5 and 7 iterations, respectively.
  • It is however clear that speech signals are not normally distributed, but it can often be assumed that spectral components follow the Laplace distribution [50], illustrated by the gray line in Fig. 15. Adding further rotations will not reduce the distance to a normalized Laplacian distribution, but it will saturate at some point, as is illustrated in Fig. 16(b). The divergence between the obtained histograms and the target distribution levels off after a few iterations. Moreover, when applying noise attenuation at the decoder, such as Wiener filtering, the distribution will be further modified. It can therefore be concluded that as few as M = 4 iterations should be sufficient to diffuse the quantization levels of a speech signal.
  • Now, applications in a speech and audio codec of the second coding concepts are described.
  • To evaluate the performance of randomization in a practical application, the speech and audio codec was calculated according to the second coding concepts, whose generic structure was described above as follows. As mentioned before, the LPC-based perceptual model was copied as-is from EVS [2]. The perceptually weighted frequency representation was then quantized with randomization using the QR-algorithm, the low-complexity randomization of the second coding concepts (equation (2.9)) with M = 4 iterations, as well as without randomization.
  • For the randomized signals, an assumption of Gaussian distribution and for the signal without randomization was used, and the Laplacian distribution was used. Previous experiments have shown that the Laplacian works best for speech signals [48], [50]. Above it has, however, been shown that randomized signals are closer to Gaussian, whereby the choice of distributions is well-warranted. Informal experiments confirmed these choices as the best in terms of perceptual SNR. Here, the perceptual SNR refers to the signal to noise ratio between the perceptually weighted original and quantized signals [22].
  • As test-material, 6 samples (3 male and 3 female) from the TIMIT corpus [51] were randomly chosen. For each coded sample, the window-wise perceptual SNR in decibel was calculated, and the mean perceptual SNRs of respective methods was calculated, which are listed in Table II. Table II:
    Method Perceptual SNR (dB)
    Single node Two nodes
    None 4.08 4.08
    QR 8.79 11.67
    Proposed botrule 6.42 8.02
  • Table II illustrates a perceptual SNRs of each evaluated method.
  • Even though all methods use entropy coding with the same bitrate, rather large differences in perceptual SNR are obtained. The QR method is over 4.8 dB better than no randomization, and the low-complexity of the second coding concepts falls in between the two. Informal experiments show that an increase in the number of iterations used for creating the randomization of the second coding concepts, will improve the SNR. The number of iterations is therefore directly proportional to complexity and SNR. It should be noted, however, that source modeling has not been applied explicitly here (such as that in [48]), which would most likely increase the performance of all methods, but especially the version without randomization. The obtained results should therefore be treated as provisional results until a source model has been implemented.
  • To evaluate the perceptual influence of the randomization on the quantization noise, a MUSHRA listening test was conducted [52] (Regarding MUSHRA, see also [56]). Thirteen subjects, aged between 22 and 53, were asked to evaluate the quality of the different approaches. Seven of the thirteen test persons referred to themselves as expert listeners.
  • As test items, the same six sentences of the TIMIT database were used as above. For each item, five conditions were used: no randomization, the fast randomization approach of the second coding concepts and as an upper bound, randomization using the QR approach, as well as a 3.5 kHz low pass signal as a lower anchor, and the hidden reference, in accordance with the MUSHRA standard.
  • The results of the listening test are presented in Fig. 17.
  • Fig. 17 illustrates the results of the MUSHRA test, given for the different items, and an average over all items. The reference was omitted as it was always rated to 100.
  • The results show that there is a clear trend that randomization improves the perceived quality, as both the QR and the fast randomization approach are rated higher than the approach without randomization.
  • Moreover, with the exception of item 3, the QR approach has high scores.. The coding quality of all methods is in the same range as the anchor, which is arguably low, even for a low bitrate codec. However, since the experiments did not include proper source modeling, the perceptual results overall should be treated as preliminary. In any case, for all items combined, the 95% confidence intervals do not overlap, and there is a clear difference between all three conditions under test, where the second coding concepts perform on average about 20 MUSHRA points better than the conventional, and the QR approach can improve the quality by approximately 15 points.
  • To determine whether there is a statistically significant difference between the ratings of the second coding concepts (Median = 42.5) and the lower anchor (Median = 40), hypotheses testing was applied. Since a Shapiro-Wilk test of the score differences (W = 0.917,p < 0.01) as well a visual inspection of Q-Q-plots indicated non-normally distributed data, a Wilcoxon signed rank test was performed which indicated no significant difference (V = 1469, p = 0.87) between the anchor and the second coding concepts.
  • However, it is unclear whether a comparison between the second coding concepts and lower anchor is relevant anyway, since the characteristics of the distortions in the two cases are very different, rendering a comparison difficult, and as a source model has not yet been included, the absolute quality level was rather arbitrarily chosen. The anchor thus serves only as way to roughly characterize the absolute quality level used in the experiment.
  • Fig. 18 illustrates the difference scores of the performed MUSHRA test, where the second coding concepts were used as a reference point. The lower anchor and the hidden reference were omitted.
  • The difference scores in Fig. 18 support the findings of the above analysis. Taking the second coding concepts as the reference point, the second coding concepts performed always significantly better than the conventional. Moreover, with the exception of item 3, the QR approach performed good. It is unclear why QR does not have an advantage for item 3, but it is suspected this is merely a statistical outlier.
  • In any case, the low-complexity second coding concepts are always better than no randomization, which was a target. This argument also validates the choice of not using source modeling; by source modeling, quantization accuracy can be improved, but the experiments show that the perceptual quality of a codec can be improved by randomization even with a fixed quantization accuracy.
  • Finally, to determine how well quantization errors are decorrelated, the randomization concepts were applied on two independent encoders (without difference in delay and without background noises) and took the mean of the outputs. In theory, taking the mean of two signals with uncorrelated noises should increase SNR by 3 dB. From Table II it can be seen that randomization with the QR-algorithm almost reaches this target level, with an improvement of 2.88 dB. The low-complexity randomizer of the second coding concepts achieves an improvement of 1.6 dB. It is thus again a compromise between complexity and quality, as the higher-complexity QR-method gives better SNR than the low-complexity randomizer of the second coding concepts. In a real-life scenario, one can expect to see higher numbers, since any differences in acoustic delay and background/sensor noises would further contribute to decorrelate the quantization errors.
  • In the following, conclusions with respect to the second coding concepts are presented.
  • Quality of speech and audio coding can be improved in terms of both signal quality and ease of interaction with the user-interface by including, in the coding process, all connected hardware which feature a microphone. For this purpose, a distributed speech and audio codec design may, e.g., be used which is based on randomization of the signal before quantization [8]. The complexity bottle-neck of the codec of the second coding concepts, that is, the randomizer, is considered.
  • The low-complexity randomizer of the second coding concepts may, e.g., be based on a sequence of random permutations and 2 × 2 block-rotations. Experiments show that by successive randomizations, high-accuracy decorrelation is obtained, such that the covariance of the original and the randomized signal behaves like uncorrelated signals, and such that the quantization levels of the output signal are diffused.
  • The randomization of the second coding concepts has multiple benefits for low-bitrate coding, distributed coding, perceptual performance, robustness and encryption. Experiments confirm these benefits by showing that randomization improves perceptual SNR and subjective quality. Though inclusion of a randomizer shows here an SNR improvement of 2.4 dB, this benefit is expected to be reduced when a proper source model is included. However, it is shown that if quantization errors are randomized, taking the mean of signals improves SNR as much as 2.8 dB, whereby one can always improve quality by adding more microphones.
  • The algorithmic complexity of the randomizer of the second coding concepts is O N log N ,
    Figure imgb0063
    where the main complexity is due to generation of random permutations. If the permutations are generated off-line the overall complexity is O MN
    Figure imgb0064
    where M is the number of iterations. Typically M = 4 is sufficient, whereby complexity is O N .
    Figure imgb0065
    Storage requirements are in all cases O N .
    Figure imgb0066
    It is believed that the complexity of the encoder therefore becomes viable even on low-performance nodes, such as wearable devices. Generation of the randomizer requires a sequence of pseudo-random numbers. It is assumed that the seed of the pseudo-random number generator is either known at both the encoder and decoder or seldom communicated as side-info.
  • Overall, here, the distributed speech and audio codec takes a large step forward to become a full system. Only source modeling was here omitted from a full system, due to space constraints. The randomizer is, however, a necessary and important part of the overall design, whereby finding a low-complexity solution was crucial.
  • In the following, a distribution of normalized generalized Gaussians of the second coding concepts are described.
  • When a signal which follows a Gaussian or Laplacian distribution is normalized by its norm, its range becomes limited.
  • Consequently, normalization of a signal changes its distribution and in the following, the form of such distributions shall be determined. In interest of generality, the generalized normal distribution is considered, which includes both the Gaussian and Laplacian distributions as special cases.
  • Suppose x is an N × 1 vector whose entries xk are uncorrelated and follow the generalized normal distribution with zero mean and equal variance σ 2 : f x k = p 2 b Γ 1 / p exp x k b p .
    Figure imgb0067
    where the scaling factor is b = σ 2 Γ 1 / p Γ 3 / p
    Figure imgb0068
    and where Γ(·) is the Gamma function [53].
  • By normalizing x with its L p norm x p ,
    Figure imgb0069
    a new random variable y = x x p
    Figure imgb0070
    is obtained, which is closely related to x but does not follow the generalized normal distribution. In particular, in difference to x, the entries yk of y have a limited range k = 1 N y k p = 1. whereby y k 1. + 1 .
    Figure imgb0071
  • To derive the marginal distributions of yk , one can start by considering the entries xk of x. Let γ k = 2 x k b p ,
    Figure imgb0072
    whereby one can find the distribution of yk by substitution of variables f γ k = 2 f x k dx k k = γ 1 p 1 e γ k 2 b Γ 1 / p x 2 2 p .
    Figure imgb0073
  • In other words, yk follows the Chi-squared distribution with 2 p
    Figure imgb0074
    degrees of freedom. Then, it is defined λ κ . = y k p = x k p x p p = x k p x k p + h k x h p .
    Figure imgb0075
  • Since the xk 's follow the generalized normal distribution, then |xk | p and ∑ hk |xh | p will follow the Chi-squared distribution with 2 p
    Figure imgb0076
    and N 1 2 p
    Figure imgb0077
    degrees of freedom, respectively. Ratios such as λk of Chi-squared distributed variables will follow the Beta-distribution with parameters α = 1 p
    Figure imgb0078
    and β = N 1 p ,
    Figure imgb0079
    or specifically [54], Section 4.2. f λ k = Γ N p Γ 1 p Γ B 1 p λ k 1 p 1 1 λ k N 1 p 1 .
    Figure imgb0080
  • Moreover, from equation (2.11) it follows that k = 1 N λ k = 1. and 0 λ k 1.
    Figure imgb0081
  • The joint distribution of the λk 's therefore follows the Dirichlet distribution with α k = 1 p ,
    Figure imgb0082
    [55].
  • One can then substitute λk = |yk | p to get the distribution of yk as f y k = 1 2 f λ k k dy k = Γ N p 1 y k p N 1 p 1 2 Γ 1 + 1 p Γ N 1 p .
    Figure imgb0083
  • This is the marginal distribution of the normalized Gaussian yk for any k. It can readily be seen that it is a scaled and translated version of a symmetric Beta distribution. It should be noted, however, that the entries yk are correlated with each other due to equation (2.15).
  • The distribution is symmetric around zero, whereby the mean is zero E[yk] = 0 and the variance is (omitting subscripts for brevity) E y 2 = 1 1 f y y 2 dy k = Γ N p Γ 3 P Γ 1 p Γ N + 2 P .
    Figure imgb0084
    where the substitution λ = |y| p was used. The integrand is similar to the Beta-distribution, whereby solution is simple. In particular, the variances E y 2 = { 1 N for p = 2 2 N + 1 N for p = 1.
    Figure imgb0085
    are obtained.
  • In the following, embodiments are described.
  • Fig. 1a illustrates an audio encoder for encoding an audio signal, wherein the audio signal is represented in a spectral domain, according to an embodiment.
  • The audio encoder comprises a spectral envelope encoder 110 configured for determining a spectral envelope of the audio signal and for encoding the spectral envelope.
  • Moreover, the audio encoder comprises a spectral sample encoder 120 configured for encoding a plurality of spectral samples of the audio signal.
  • The spectral sample encoder 120 is configured to estimate an estimated bitrate needed for encoding for each spectral sample of one or more spectral samples of the plurality of spectral samples depending on the spectral envelope.
  • Moreover, the spectral sample encoder 120 is configured to encode each spectral sample of the plurality of spectral samples, depending on the estimated bitrate needed for encoding for the one or more spectral samples, according to a first coding rule or according to a second coding rule being different from the first coding rule.
  • In some of the embodiments, it may not be necessary to estimate an estimated bitrate for each of the plurality of spectral samples. For example, if it is estimated that for one of the spectral samples, the estimated bitrate is not greater than a threshold value, than it may be concluded that following spectral samples are also not greater than a threshold value.
  • In some alternative embodiments, however, the estimated bitrate may, e.g., be determined for each of the plurality of spectral samples.
  • In an embodiment, if the estimated bitrate needed for encoding a spectral sample of the one or more spectral samples is greater than a threshold value, the spectral sample encoder 120 may, e.g., be configured to encode said spectral sample according to the first coding rule, and if said estimated bitrate needed for encoding said spectral samples is smaller than or equal to a threshold value, the spectral sample encoder 120 may, e.g., be configured to encode said spectral sample according to the second coding rule.
  • According to an embodiment, the threshold value may, e.g., be 1 bit/sample (1 bit/spectral sample).
  • In an embodiment, the spectral sample encoder 120 may, e.g., be configured to estimate the estimated bitrate needed for encoding for each spectral sample of the one or more spectral samples depending on an estimated variance of said spectral sample which depends on the spectral envelope.
  • According to an embodiment, the spectral sample encoder 120 may, e.g., be configured to estimate the estimated bitrate needed for encoding for each spectral sample of the one or more spectral samples depending on the equation b k = 1 2 log 4.1159 σ k 2
    Figure imgb0086
    wherein bk is a k th spectral samples of the one or more spectral samples, wherein σ k 2
    Figure imgb0087
    is an estimated variance of said spectral sample.
  • In an embodiment, the spectral sample encoder 120 may, e.g., be configured to encode the spectral samples that are encoded according to the second coding rule by quantizing said spectral samples that are encoded according to the second coding rule employing an orthogonal matrix. Moreover, the spectral sample encoder 120 may, e.g., be configured to encode the spectral samples that are encoded according to the first coding rule by quantizing said spectral samples that are encoded according to the first coding rule without employing the orthogonal matrix.
  • According to an embodiment, the spectral sample encoder 120 may, e.g., be configured to encode the spectral samples that are encoded according to the second coding rule using: Q B Ax .
    Figure imgb0088
    wherein x is a vector comprising the spectral samples, wherein QB [·] is a quantizer defined as Q B y : = γ sign y 0 sign y 1 sign y B 1 0 0
    Figure imgb0089
    wherein y 0 y 1, y B-1 are quantized values resulting from quantization, wherein B indicates a total bitrate, wherein γ is a scaling coefficient, wherein A is the orthogonal matrix.
  • Fig. 1b illustrates an audio decoder for decoding an encoded audio signal according to an embodiment.
  • The audio decoder comprises an interface 130 configured for receiving an encoded spectral envelope of the audio signal and configured for receiving an encoded plurality of spectral samples of the audio signal.
  • Moreover, the audio decoder comprises a decoding unit 140 configured for decoding the encoded audio signal by decoding the encoded spectral envelope and by decoding the encoded plurality of spectral samples.
  • The decoding unit 140 is configured to receive or to estimate an estimated bitrate needed for encoding for each spectral sample of one or more spectral samples of the encoded plurality of spectral samples.
  • Moreover, the decoding unit 140 is configured to decode each spectral sample of the encoded plurality of spectral samples, depending on the estimated bitrate needed for encoding for the one or more spectral samples of the encoded plurality of spectral samples, according to a first coding rule or according to a second coding rule being different from the first coding rule.
  • In some embodiments, the decoding unit 140 may, e.g., receive the estimated bitrate needed for encoding for each spectral sample of the one or more spectral samples from the audio encoder that encoded the plurality of spectral samples.
  • In some alternative embodiments, however, the decoding unit 140 may, e.g., estimate the estimated bitrate needed for encoding for each spectral sample of the one or more spectral samples depending on the spectral envelope in a same way as the audio encoded that encoded the spectral samples has estimated the estimated bitrate.
  • In an embodiment, if the estimated bitrate needed for encoding a spectral sample of the one or more spectral samples is greater than a threshold value, the decoding unit 140 may, e.g., be configured to decode said spectral sample according to the first coding rule, and if said estimated bitrate needed for encoding said spectral samples is smaller than or equal to a threshold value, the decoding unit 140 may, e.g., be configured to decode said spectral sample according to the second coding rule.
  • According to an embodiment, the threshold value may, e.g., be 1 bit/sample (1 bit/spectral sample).
  • In an embodiment, the decoding unit 140 may, e.g., be configured to estimate the estimated bitrate needed for encoding for each spectral sample of the one or more spectral samples depending on an estimated variance of said spectral sample which depends on the spectral envelope.
  • According to an embodiment, the decoding unit 140 may, e.g., be configured to estimate the estimated bitrate needed for encoding for each spectral sample of the one or more spectral samples depending on the equation b k = 1 2 log 2 4.1159 σ k 2
    Figure imgb0090
    wherein bk is a k th spectral samples of the one or more spectral samples, wherein σ k 2
    Figure imgb0091
    is an estimated variance of said spectral sample.
  • In an embodiment, the decoding unit 140 may, e.g., be configured to decode the spectral samples that are decoded according to the second coding rule by employing an orthogonal matrix. Moreover, the decoding unit 140 is configured to decode the spectral samples that are decoded according to the first coding rule without employing the orthogonal matrix.
  • According to an embodiment, if the estimated bitrate needed for encoding a spectral sample of the one or more spectral samples is smaller than another threshold value, the decoding unit 140 may, e.g., be configured to employ spectral noise shaping for decoding the encoded plurality of spectral samples.
  • Fig. 1c illustrates a system according to an embodiment. The system comprises an audio encoder 105 according to Fig. 1a and an audio decoder 125 according to Fig.1b.
  • The audio encoder 105 is configured to feed a encoded spectral envelope of an encoded audio signal and an encoded plurality of spectral samples of the encoded audio signal into the audio decoder.
  • The audio decoder 125 is configured to decode the encoded audio signal by decoding the encoded spectral envelope and by decoding the encoded plurality of spectral samples.
  • In the following, particular embodiments of the present invention are described.
  • Dithering methods however provide an alternative approach, where both accuracy and energy are retained. A hybrid coding approach is provided where low-energy samples are quantized using dithering, instead of the conventional uniform quantizer. For dithering, 1 bit quantization is applied in a randomized sub-space. It is moreover demonstrated that the output energy can be adjusted to the desired level using a scaling parameter. Objective measurements and listening tests demonstrate the advantages of the provided concepts.
  • In the following, an application of the provided randomization for dithered quantization in frequency-domain coding of speech and audio is provided, to allow coding at very low bitrates without excessive sparseness or low energy in the output. Perceptual listening tests demonstrate that the provided dithered quantizer gives the best performance.
  • In the following, quantization concepts of embodiments are described.
  • The performance of dithered quantization methods in comparison to conventional uniform quantization and in combination with entropy coding are considered. In the TCX mode of EVS [2], [48], entropy coding and uniform quantization is implemented assuming that the sample distribution is Laplacian, and the sample variance is estimated using the linear predictive envelope. The quantization accuracy is determined in a rate-loop such that the bit-budget is used as effectively as possible. In a vector of samples, trailing zeros are truncated. The scaling of the signal is determined after quantization, such that the output signal-to-noise ratio is optimized. This implementation of uniform quantization may, e.g., be used as a baseline system.
  • The above-described second coding concept comprises an approach for dithering and encoding data at low bit-rates (less than 1 bit/sample) based on random rotations and which is defined as follows [70]. Supposing there is a vector x R N × 1 ,
    Figure imgb0092
    which shall be encoded with B N bits. Using a random orthogonal matrix A, one can then quantize x ^ = A T Q B Ax .
    Figure imgb0093
    where QB [·] is a quantizer defined as Q B y : = γ sign y 0 sign y 1 sign y B 1 0 0
    Figure imgb0094
    and γ is a scaling coefficient. In other words, it uses a 1 bit-quantizer, where the B first samples are quantized with the sign of the input sample, for a total bitrate of B. The quantized then has an approximately normal distribution and the variance is E x ^ 2 = γ 2 B N .
    Figure imgb0095
    It has furthermore been shown that the orthogonal matrix A can be approximated by a low-order rotation such that the algorithmic complexity remains linear O N .
    Figure imgb0096
  • It can then readily been show that γ can be chosen according to a number of criteria, for example:
    1. 1. γ MSSE = σ 2 π
      Figure imgb0097
      is the minimum mean square error (MMSE) scaling for normal input of variance σ 2. This thus corresponds to Wiener filtering the quantized signal.
    2. 2. γ σ 2 = σ N B
      Figure imgb0098
      retains the variance σ 2 of the original signal. This corresponds to quantization on the surface of an N-dimensional hyper-sphere, which is normalized to original energy.
  • Clearly, γ can thus be tuned according to perceptual criteria, for a balance between accuracy and how well the quantizer retains the signal distribution and variance.
  • This approach to quantization provides error-diffusion similar to Floyd-Steinberg-type methods. However, in difference, error is not diffused forward to following components, but instead, it is diffused among the samples of a vector. In signal processing terms, Floyd-Steinberg-type methods are thus similar to infinite impulse responses (IIR) filters, whereas the proposed method is more like a finite impulse response (FIR) operation.
  • In the following speech and audio coding frameworks of embodiments are provided.
  • To evaluate the error characters of different quantizers, they should be implemented in a speech and audio codec which allows a fair comparison. This task is less straightforward than one might expect. The main issue is that codecs regularly use ad hoc tricks to overcome saturation effects of the arithmetic coder [1], [2], [48]. Namely, for example, high-frequency samples quantized to zero are typically truncated from the spectrum above the last non-zero sample. By omitting the transmission of zero samples one can save a considerable amount of bits, which can instead be used for coding low-frequency components. The performance of the arithmetic coder, in isolation, does therefore not accurately reflect the performance of the overall codec.
  • To obtain a fair comparison, a state-of-the-art baseline system is implemented following the simplified structure of the 3GPP Enhanced Voice Services (EVS) [2], [1], [48] (see Fig. 3).
  • Fig. 3 illustrates a diagram of the speech and audio encoder. The gray box is modified in by embodiments.
  • For frequency-domain coding, here the MDCT-transform with a window length of 30 ms, 50% overlap, a half-sine window and pre-emphasis with a filter P(z) =1 - 0.68z -1 is used. At a sampling rate of 16 kHz, the magnitude envelope is modeled with a linear predictive model of order M = 20, which is used as an estimate of the variance of each frequency component, and which is further fed into a conventional arithmetic coder with an assumption of a Laplacian distribution. Quantization in the perceptually weighted domain is applied as in [1]. It should be noted that a deadzone-quantizer was not implemented, even if it is known to improve signal-to-noise ratio [64], because it also amplifies the saturation effect at high frequencies. A deadzone-quantizer would therefore have unfairly penalized the baseline codec in terms low-rate performance.
  • Conventional codecs include noise-fill and bandwidth-extension methods to reduce the bitrate and to compensate for the energy-loss at high frequencies. To allow a straightforward and fair comparison between methods, bandwidth-extension was not included in the codec. The noise-fill algorithm is applied at frequencies above 1.6 kHz, on all spectral components which are quantized to zero, where noise with a random sign was added, and the magnitude was adjusted to match that obtained with the proposed dithering method with gain MMSE. The noise-fill used in EVS uses advanced signal analysis to fine-tune noise-filling, but this simplified method was chosen to make the test easy to reproduce. All parameters should anyway be tuned to the particular configuration of the final codec, whereby further perceptual tuning of parameters is not worthwhile for these experiments.
  • For the bitrate of the codec, it was assumed that spectral envelope, gain and other parameters are encoded with 2.6 kbits/s, whereby the remaining bits can be used for encoding the spectrum. Further, for simplicity and reproducability, any other parameters of the signal were not quantized. It should be noted, however, that bitrate calculations here are provided only to assist the reader in getting a realistic impression of performance, as the bitrate of side-information can vary in particular implementations of codecs.
  • In the following, a hybrid coder according to embodiments is described.
  • The combination of uniform quantization and arithmetic coder saturates at low bitrates, whereby it is proposed to replace the conventional approach by dithered coding for spectral samples whose bitrate is below 1 bit/sample. It is thus a hybrid entropy coder, which uses uniform quantization and arithmetic coding following [48] in high-energy areas of the spectrum and dithered coding at the low-energy areas.
  • The baseline entropy coder uses the linear predictive envelope to estimate the variance σ k 2
    Figure imgb0099
    of frequency components [48]. Note that this envelope has to be scaled such that the expected bitrate of a signal which follows that envelope, matches the target bitrate. Based on the variance σ k 2
    Figure imgb0100
    of the k th component, one can then estimate the expected bitrate of a sample as b k = 1 2 log 2 4.1159 σ k 2
    Figure imgb0101
    but limited to bk ≥ 0. For spectral components with bk > 1 quantization and arithmetic coding is used, for bk ≤ 1 dithered coding is applied (see Fig. 4). The bit-allocation between uniform and dithered quantization is derived directly from the expected bitrate bk .
  • Fig. 4 illustrates hybrid coding of spectral components.
  • All low-energy samples are thus collated into a vector x and are quantized with equation (3.1). Implicitly, it is thus assumed that vector x follows the normal distribution with uniform variance. To improve accuracy, embodiments further subdivide x according to their variance.
  • In [70], it was demonstrated that the randomization matrix A of sufficient quality can be readily generated with 4 iteration of N/2 random 2 × 2 rotations and length N permutations, when the bitrate is B = N. However, with B << N, a majority of samples are zeros, whereby the number of iterations were increased to 8 such that the output distribution remains normal.
  • Random rotations between the non-zero and zeroed values could be readily used to reduce the computational complexity without effect on the statistics of the output signal.
  • In the following, conducted experiments with respect to embodiments are described.
  • To evaluate the performance the provided hybrid codec, in comparison to uniform quantization, three types of experiments were performed. Firstly, the performance of dithering in isolation with synthetic input is considered. Secondly, speech from the TIMIT corpus was encoded and performance was evaluated by objective criteria. Finally, using samples from the TIMIT corpus, a MUSHRA subjective listening test was performed to determine perceptual preference among methods [52].
  • Output distribution of the provided dithered quantization (equation (3.1)) in comparison to uniform quantization is illustrated in Fig. 5.
  • Fig. 5 illustrates collated histograms of K = 10000 vectors of unit variance Gaussian input, quantized by uniform quantization and entropy coding as well as the provided dithered coder (N = 32, γσ 2), with 1 bit/sample.
  • Here, normally distributed K = 10000 vectors of length N = 32 were encoded with B = 32 bits, and the gain factor γσ 2 was used. It can readily be observed that uniform quantization is unable to retain the shape of the original distribution, whereas the distribution of the output of the provided dithered codec exactly matches that of the input.
  • Performance of the provided coder for a single frame of speech is illustrated in Fig. 6.
  • Fig. 6 depicts an illustration of performance for a typical speech spectrum at 13.2 kbits/s, wherein (a) depicts an input signal spectrum and its envelope, wherein (b) depicts the bit-rate estimated from the envelope and the threshold where quantizers are switched, and wherein (c) the quantized spectra with conventional uniform quantization and entropy coding in comparison to the provided, dithered coder.
  • The spectral magnitude envelope is estimated using linear predictive modelling in Fig. 6(a), the expected bit-rate for each frequency is estimated from the envelope using the method developed in [48] in Fig. 6(b) and a threshold is applied to determine the choice of quantizer. Finally, in Fig. 6(c), the quantized output of the conventional method is compared with the provided concepts, where the gain factor was _2 . It can be clearly seen that whereas for the conventional approach, all frequencies above 2 kHz are quantized to zero, the provided concepts retain the spectral shape also at the higher frequencies.
  • For objective evaluation of performance on real speech, the entire TIMIT database (training and evaluation) was encoded with different combinations of quantization and entropy coding [51]. Namely, it was applied: 1. uniform quantization with arithmetic coding following [48] (Conventional), 2. a dithering simulation by adding white noise to obtain same signal to noise ratio as the conventional approach (Dithering), 3. the provided hybrid codec using γMMSE (1 bit MMSE) and 4. Using γσ2 (1 bit EM). The mean output energy across frequencies for each method is illustrated in Fig. 2.
  • It can be seen that all modifications of conventional arithmetic coding bring the amount of energy closer to the original energy envelope. The dithering simulation saturates at the perceptual noise floor near -20 dB, which is higher than the original energy envelope. Informal listening confirms that such dithering has a noisy character, where the conventional is muffled. The two provided 1 bit dithering concepts are closer to the original energy envelope, such that the MMSE approach γMMSE is clearly below the original while the energy matching method γσ 2 approximates nicely the desired energy envelope.
  • The average signal to noise ratios (SNR) in the perceptual domain for the conventional and provided concepts are listed in Table III. Table III:
    1 bit MMSE 1 bit EM Conventional
    SNR (dB) 10.75 10.46 8.93
  • Table III illustrates a mean signal to noise ratio in the perceptually weighted domain for the conventional and the two provided concepts.
  • Clearly the 1 bit MMSE approach reaches the highest SNR, since it was designed to optimize SNR. However, as the conventional method is also designed to optimize SNR, it is surprising that such a large improvement of 1.81 dB is obtained. The energy-matching approach γσ2 looses slightly in SNR to the MMSE approach, but not the difference is not large 0.29 dB. A subjective listening test is needed to determine it is more important to preserve envelope shape or optimize SNR.
  • Finally, to determine subjective preference among methods, a MUSHRA listening test was performed [52]. In the test, 6 samples (3 male and 3 female) randomly chosen from the TIMIT corpus [51] were included. In addition to the above methods, Conventional, Dithered, 1 bit MMSE and 1 bit EM, also cases where the conventional uniform coder is enhanced by noise filling in post-processing were included. It was not included in the previous tests because it is a blind post-processing method in the sense that it adds noise without any transmitted information from the input signal, whereby it reduces accuracy even if it is designed to improve perceptual quality. In the listening test, 14 normal hearing subjects in the age-range 26 to 43 years attended. Fig. 7 illustrates the results.
  • Fig. 7 depicts results of a subjective MUSHRA listening test, comparing the provided 1 bit dithered quantizers with conventional arithmetic coding, as well as a synthetic dithering serving as an anchor.
  • It was observed for all items, that the provided dithered 1 bit quantizers have a higher mean than the other methods. Moreover, in the mean over all items (the "All" column), the provided dithered 1 bit quantizers have a statistically significant difference to the antecedent methods. Conventional arithmetic coding without noisefill also shows a statistically significant reduction in quality in comparison to all other methdods. To further determine whether listeners have a preference among the two provided dithered quantizers, the differential MUSHRA scores with noisefill as a reference (see Fig. 8) was calculated.
  • Fig. 8 illustrates differential scores of a subjective MUSHRA listening test, comparing the provided 1 bit dithered quantizers with conventional arithmetic coding, as well as a synthetic dithering serving as an anchor. Differences are calculated with the noisefill as reference.
  • However, the differential scores revealed no additional details.
  • Now, a discussion the provided concepts and conclusions follow.
  • The applicability of dithering in signal processing has been demonstrated., Conventional methods in coding, such as noisefill and bandwidth-extension do not attempt to optimize SNR. In contrast, embodiments may, e.g., use a recently developed method for dithering and coding which is applicable to very low bitrates [70]. The approach is based on a random rotation, sign-quantization in the randomized domain, and an inverse transform. It is proposed to apply it in combination with conventional uniform quantization and entropy coding, such that only frequency components where one can afford to use very little accuracy, are coded with the dithered quantizer.
  • By using dithering one can avoid the characteristic problem of conventional frequency-domain codecs, where higher frequencies are often quantized to zero such the output sounds muffled. In other words, the output is not unnaturally sparse.
  • Objective and subjective experiments demonstrate that the method gives a significant improvement in perceptual quality.
  • Although some aspects have been described in the context of an apparatus, it is clear that these aspects also represent a description of the corresponding method, where a block or device corresponds to a method step or a feature of a method step. Analogously, aspects described in the context of a method step also represent a description of a corresponding block or item or feature of a corresponding apparatus. Some or all of the method steps may be executed by (or using) a hardware apparatus, like for example, a microprocessor, a programmable computer or an electronic circuit. In some embodiments, one or more of the most important method steps may be executed by such an apparatus.
  • Depending on certain implementation requirements, embodiments of the invention can be implemented in hardware or in software or at least partially in hardware or at least partially in software. The implementation can be performed using a digital storage medium, for example a floppy disk, a DVD, a Blu-Ray, a CD, a ROM, a PROM, an EPROM, an EEPROM or a FLASH memory, having electronically readable control signals stored thereon, which cooperate (or are capable of cooperating) with a programmable computer system such that the respective method is performed. Therefore, the digital storage medium may be computer readable.
  • Some embodiments according to the invention comprise a data carrier having electronically readable control signals, which are capable of cooperating with a programmable computer system, such that one of the methods described herein is performed.
  • Generally, embodiments of the present invention can be implemented as a computer program product with a program code, the program code being operative for performing one of the methods when the computer program product runs on a computer. The program code may for example be stored on a machine readable carrier.
  • Other embodiments comprise the computer program for performing one of the methods described herein, stored on a machine readable carrier.
  • In other words, an embodiment of the inventive method is, therefore, a computer program having a program code for performing one of the methods described herein, when the computer program runs on a computer.
  • A further embodiment of the inventive methods is, therefore, a data carrier (or a digital storage medium, or a computer-readable medium) comprising, recorded thereon, the computer program for performing one of the methods described herein. The data carrier, the digital storage medium or the recorded medium are typically tangible and/or non-transitory.
  • A further embodiment of the inventive method is, therefore, a data stream or a sequence of signals representing the computer program for performing one of the methods described herein. The data stream or the sequence of signals may for example be configured to be transferred via a data communication connection, for example via the Internet.
  • A further embodiment comprises a processing means, for example a computer, or a programmable logic device, configured to or adapted to perform one of the methods described herein.
  • A further embodiment comprises a computer having installed thereon the computer program for performing one of the methods described herein.
  • A further embodiment according to the invention comprises an apparatus or a system configured to transfer (for example, electronically or optically) a computer program for performing one of the methods described herein to a receiver. The receiver may, for example, be a computer, a mobile device, a memory device or the like. The apparatus or system may, for example, comprise a file server for transferring the computer program to the receiver.
  • In some embodiments, a programmable logic device (for example a field programmable gate array) may be used to perform some or all of the functionalities of the methods described herein. In some embodiments, a field programmable gate array may cooperate with a microprocessor in order to perform one of the methods described herein. Generally, the methods are preferably performed by any hardware apparatus.
  • The apparatus described herein may be implemented using a hardware apparatus, or using a computer, or using a combination of a hardware apparatus and a computer.
  • The methods described herein may be performed using a hardware apparatus, or using a computer, or using a combination of a hardware apparatus and a computer.
  • The above described embodiments are merely illustrative for the principles of the present invention. It is understood that modifications and variations of the arrangements and the details described herein will be apparent to others skilled in the art. It is the intent, therefore, to be limited only by the scope of the impending patent claims and not by the specific details presented by way of description and explanation of the embodiments herein.
  • References
    1. [1] T. Bäckström, Speech Coding with Code-Excited Linear Prediction. New York, NY, USA: Springer, 2017.
    2. [2] TS 26.445, EVS Codec Detailed Algorithmic Description; 3GPP Technical Specification (Release 12), 3GPP, 2014.
    3. [3] TS 26.190, Adaptive Multi-Rate (AMR-WB) Speech Codec, 3GPP, 2007.
    4. [4] MPEG-D (MPEG Audio Technologies), Part 3: Unified Speech and Audio Coding, ISO/IEC 23003-3:2012, 2012.
    5. [5] M. Bosi et al., "ISO/IECMPEG-2 advanced audio coding," J. Audio Eng. Soc., vol. 45, no. 10, pp. 789-814, 1997.
    6. [6] J. Benesty, M. Sondhi, and Y. Huang, Springer Handbook of Speech Processing. New York, NY, USA: Springer, 2008.
    7. [7] A. Zahedi, J. Østergaard, S. H. Jensen, S. Bech, and P. Naylor, "Audio coding in wireless acoustic sensor networks," Signal Process., vol. 107, pp. 141-152, 2015.
    8. [8] T. Bäckström, F. Ghido, and J. Fischer, "Blind recovery of perceptual models in distributed speech and audio coding," in Proc. Interspeech, 2016, pp. 2483-2487.
    9. [9] T. Bäckström and J. Fischer, "Coding of parametric models with randomized quantization in a distributed speech and audio codec," in Proc. ITG Fachtagung Sprachkommunikation, 2016, pp. 1-5.
    10. [10] P. T. Boufounos and R. G. Baraniuk, "1-bit compressive sensing," in Proc. IEEE Inf. Sci. Syst. 42nd Ann. Conf., 2008, pp. 16-21.
    11. [11] A. Magnani, A. Ghosh, and R. M. Gray, "Optimal one-bit quantization," in Proc. IEEE Data Compression Conf., 2005, pp. 270-278.
    12. [12] M. M. Truman, G. A. Davidson, M. C. Fellers, M. S. Vinton, M. A. Watson, and C. Q. Robinson, "Audio coding system using spectral hole filling," U.S. Patent 7 447 631, Nov. 4, 2008 .
    13. [13] S. Vaudenay, "Decorrelation: a theory for block cipher security," J. Cryptol., vol. 16, no. 4, pp. 249-286, 2003.
    14. [14] S. Saeednia, "How to make the Hill cipher secure," Cryptologia, vol. 24, no. 4, pp. 353-360, 2000.
    15. [15] C.-C. Kuo and W. Thong, "Reduction of quantization error with adaptive Wiener filter in low bit rate coding," in Proc. 11th Eur. IEEE Signal Process. Conf., 2002, pp. 1-4.
    16. [16] G. E. Øien and T. A. Ramstad, "On the role of Wiener filtering in quantization and DPCM," in Proc. IEEE Norwegian Signal Process. Symp. Workshop, 2001.
    17. [17] J. Rissanen and G. G. Langdon, "Arithmetic coding," IBM J. Res.Develop., vol. 23, no. 2, pp. 149-162, 1979.
    18. [18] J. D. Gibson and K. Sayood, "Lattice quantization," Adv. Electron. Electron Phys., vol. 72, pp. 259-330, 1988.
    19. [19] A. Gersho and R. M. Gray, Vector Quantization and Signal Compression. New York, NY, USA: Springer, 1992.
    20. [20] Z. Xiong, A. D. Liveris, and S. Cheng, "Distributed source coding for sensor networks," IEEE Signal Process. Mag., vol. 21, no. 5, pp. 80-94, Sep. 2004.
    21. [21] Z. Xiong, A. D. Liveris, and Y. Yang, "Distributed source coding," in Handbook on Array Processing and Sensor Networks. New York, NY, USA: Wiley-IEEE Press, 2009, pp. 609-643.
    22. [22] M. Bosi and R. E. Goldberg, Introduction to Digital Audio Coding and Standards. Norwell, MA, USA: Kluwer, 2003.
    23. [23] A. Bertrand, "Applications and trends in wireless acoustic sensor networks: A signal processing perspective," in Proc. 18th IEEE Symp. Commun. Veh. Technol. Benelux, 2011, pp. 1-6.
    24. [24] I. F. Akyildiz, T. Melodia, and K. R. Chowdury, "Wireless multimedia sensor networks: A survey," IEEE Wireless Commun., vol. 14, no. 6, pp. 32-39, Dec. 2007.
    25. [25] B. Girod, A. M. Aaron, S. Rane, and D. Rebollo-Monedero, "Distributed video coding," Proc. IEEE, vol. 93, no. 1, pp. 71-83, Jan. 2005.
    26. [26] F. de la Hucha Arce, M. Moonen, M. Verhelst, and A. Bertrand, "Adaptive quantization for multi-channel Wiener filter-based speech enhancement in wireless acoustic sensor networks," Wireless Commun. Mobile Comput., vol. 2017, 2017, Art. no. 3173196.
    27. [27] A. Zahedi, J. Østergaard, S. H. Jensen, P. Naylor, and S. Bech, "Coding and enhancement in wireless acoustic sensor networks," in Proc. IEEE Data Compression Conf., 2015, pp. 293-302.
    28. [28] A. Majumdar, K. Ramchandran, and L. Kozintsev, "Distributed coding for wireless audio sensors," in Proc. IEEE Workshop Appl. Signal Process. Audio Acoust., 2003, pp. 209-212.
    29. [29] H. Dong, J. Lu, and Y. Sun, "Distributed audio coding in wireless sensor networks," in Proc. IEEE Int. Conf. Comput. Intell. Secur., 2006, vol. 2, pp. 1695-1699.
    30. [30] G. Barriac, R. Mudumbai, and U. Madhow, "Distributed beamforming for information transfer in sensor networks," in Proc. 3rd Int. Symp. Inf. Process. Sens. Netw., 2004, pp. 81-88.
    31. [31] R. Lienhart, I. Kozintsev, S. Wehr, and M. Yeung, "On the importance of exact synchronization for distributed audio signal processing," in Proc. IEEE Int. Conf. Acoust., Speech, Signal Process., 2003, vol. 4, p. IV-840-3.
    32. [32] O. Roy and M. Vetterli, "Rate-constrained collaborative noise reduction for wireless hearing aids," IEEE Trans. Signal Process., vol. 57, no. 2, pp. 645-657, Feb. 2009.
    33. [33] S. Bray and G. Tzanetakis, "Distributed audio feature extraction for music," in Proc. Int. Conf. Music Inf. Retrieval, 2005, pp. 434-437.
    34. [34] N. Rajput and A. A. Nanavati, "Distributed speech recognition," in Speech in Mobile and Pervasive Environments. New York, NY, USA: Wiley, 2012, pp. 99-114.
    35. [35] D. Pearce, "Distributed speech recognition standards," in Automatic Speech Recognition on Mobile Devices and Over Communication Networks. London, U.K.: Springer, 2008, pp. 87-106.
    36. [36] S. Korse, T. Jahnel, and T. Bäckström, "Entropy coding of spectral envelopes for speech and audio coding using distribution quantization," in Proc. Interspeech, 2016, pp. 2543-2547.
    37. [37] S. Das, A. Craciun, T. Jahnel, and T. Bäckström, "Spectral envelope statistics for source modelling in speech enhancement," in Proc. ITG Fachtagung Sprachkommunikation, 2016, pp. 1-5.
    38. [38] G. H. Golub and C. F. van Loan, Matrix Computations, 4th ed. Baltimore, MD, USA: John Hopkins Univ. Press, 2004.
    39. [39] V. Pulkki, "Spatial sound reproduction with directional audio coding," J. Audio Eng. Soc., vol. 55, no. 6, pp. 503-516, 2007.
    40. [40] A. Edelman and N. R. Rao, "Random matrix theory," Acta Numerica, vol. 14, pp. 233-297, 2005.
    41. [41] D. Knuth, The Art of Computer Programming. Reading, MA, USA: Addison-Wesley, 1998.
    42. [42] D. E. Knuth, "Volume 2: Seminumerical algorithms," in The Art of Computer Programming, 3rd ed. Reading, MA, USA: Addison-Wesley, 2007.
    43. [43] P. Diaconis and M. Shahshahani, "The subgroup algorithm for generating uniform random variables," Probab. Eng. Inf. Sci., vol. 1, no. 1, pp. 15-32, 1987.
    44. [44] G. W. Stewart, "The efficient generation of random orthogonal matrices with an application to condition estimators," SIAM J. Numer. Anal., vol. 17, no. 3, pp. 403-409, 1980.
    45. [45] P. L'Ecuyer, "Pseudorandom number generators," in, Encyclopedia of Quantitative Finance. New York, NY, USA: Wiley, 2010.
    46. [46] M. Matsumoto andT. Nishimura, "Mersenne twister:A623-dimensionally equidistributed uniform pseudo-random number generator," ACM Trans. Model. Comput. Simul., vol. 8, no. 1, pp. 3-30, 1998.
    47. [47] A. Jagannatam, "Mersenne twister-A pseudo random number generator and its variants," Dept. Elect. Comput. Eng., George Mason Univ., Fairfax, VA, USA, 2008.
    48. [48] T. Bäckström and C. R. Helmrich, "Arithmetic coding of speech and audio spectra using TCX based on linear predictive spectral envelopes," in Proc. IEEE Int. Conf. Acoust., Speech, Signal Process., Apr. 2015, pp. 5127-5131.
    49. [49] J. Fischer and T. Bäckström, "Wiener filtering in distributed speech and audio coding," IEEE Signal Process. Lett., 2017, submitted for publication.
    50. [50] T. Bäckström, "Estimation of the probability distribution of spectral fine structure in the speech source," in Proc. Interspeech, 2017, pp. 344-348.
    51. [51] J. S. Garofolo, et al., TIMIT: Acoustic-Phonetic Continuous Speech Corpus. Philadelphia, PA, USA: Linguistic Data Consortium, 1993.
    52. [52] Method for the Subjective Assessment of Intermediate Quality Levels of Coding Systems, ITU-R Recommendation BS.1534, 2003.
    53. [53] S. Nadarajah, "A generalized normal distribution," J. Appl. Statist., vol. 32, no. 7, pp. 685-694, 2005.
    54. [54] C. Walck, Handbook on Statistical Distributions for Experimentalists. Stockholm, Sweden: Univ. Stockholm, 2007.
    55. [55] A. Bela, A. Frigyik, and M. Gupta, "Introduction to the Dirichlet distribution and related processes," , Dept. Elect. Eng., Univ. Washington, Seattle, WA, USA, Tech. Rep. UWEETR-2010-0006, 2010.
    56. [56] M. Schoeffler, F.-R. Stöter, B. Edler, and J. Herre, "Towards the next generation of web-based experiments: A case study assessing basic audio quality following the ITU-R recommendation BS. 1534 (MUSHRA)," in Proc. 1st Web Audio Conf., Paris, France, 2015.
    57. [57] M. Neuendorf, M. Multrus, N. Rettelbach, G. Fuchs, J. Robilliard, J. Lecomte, S. Wilde, S. Bayer, S. Disch, C. Helmrich, R. Lefebvre, P. Gournay, B. Bessette, J. Lapierre, K. Kj¨orling, H. Purnhagen, L. Villemoes, W. Oomen, E. Schuijers, K. Kikuiri, T. Chinen, T. Norimatsu, K. S. Chong, E. Oh, M. Kim, S. Quackenbush, and B. Grill, "The ISO/MPEG unified speech and audio coding standard - consistent high quality for all content types and at all bit rates," Journal of the AES, vol. 61, no. 12, pp. 956-977, 2013.
    58. [58] ITU-T G.718, "Frame error robust narrow-band and wideband embedded variable bitrate coding of speech and audio from 8-32 kbit/s," 2008.
    59. [59] J. Mäkinen, B. Bessette, S. Bruhn, P. Ojala, R. Salami, and A. Taleb, "AMR-WB+: a new audio coding standard for 3rd generation mobile audio services," in Proc. ICASSP, 2005, vol. 2, pp. 1109-1112.
    60. [60] M. Bosi, K. Brandenburg, S. Quackenbush, L. Fielder, K. Akagiri, H. Fuchs, M. Dietz, J. Herre, G. Davidson, and Y. Oikawa, "ISO/IEC MPEG-2 Advanced audio coding," in 101 AES Convention, 2012.
    61. [61] G. Fuchs, M. Multrus, M. Neuendorf, and R. Geiger, "Mdct-based coder for highly adaptive speech and audio coding," in European Signal Processing Conference (EUSIPCO 2009), 2009, pp. 24-28.
    62. [62] I. H. Witten, R. M Neal, and J. G. Cleary, "Arithmetic coding for data compression," Communications of the ACM, vol. 30, no. 6, pp. 520-540, 1987.
    63. [63] ITU-T Recommendation P.800, Methods for subjective determination of transmission quality, 1996.
    64. [64] G. Fuchs, C. R. Helmrich, G. Marković, M. Neusinger, E. Ravelli, and T. Moriya, "Low delay LPC and MDCT-based audio coding in the EVS codec," in Proc. ICASSP. IEEE, 2015, pp. 5723-5727.
    65. [65] S Disch, A. Niedermeier, C. R. Helmrich, C. Neukam, K. Schmidt, R. Geiger, J. Lecomte, F. Ghido, F. Nagel, and B. Edler, "Intelligent gap filling in perceptual transform coding of audio," in Audio Engineering Society Convention 141. Audio Engineering Society, 2016.
    66. [66] J. Vanderkooy and S. P. Lipshitz, "Dither in digital audio," Journal of the Audio Engineering Society, vol. 35, no. 12, pp. 966-975, 1987.
    67. [67] R. W. Floyd and L. Steinberg, "An adaptive algorithm for spatial gray-scale," in Proc. Soc. Inf. Disp., 1976, vol. 17, pp. 75-77.
    68. [68] M Li, J. Klejsa, and W. B. Kleijn, "Distribution preserving quantization with dithering and transformation," IEEE Signal Processing Letters, vol. 17, no. 12, pp. 1014-1017, 2010.
    69. [69] T. Bäckström, "Enumerative algebraic coding for ACELP," in Proc. Interspeech, 2012.
    70. [70] T. Bäckström and J. Fischer, "Fast randomization for distributed low-bitrate coding of speech and audio," IEEE/ACM Trans. Audio, Speech, Lang. Process., vol. 26, no. 1, January 2018.
    71. [71] J.-M. Valin, G. Maxwell, T. B. Terriberry, and K. Vos, "High-quality, low-delay music coding in the OPUS codec," in Audio Engineering Society Convention 135. Audio Engineering Society, 2013.

Claims (18)

  1. An audio encoder for encoding an audio signal, wherein the audio signal is represented in a spectral domain, wherein the audio encoder comprises:
    a spectral envelope encoder (110) configured for determining a spectral envelope of the audio signal and for encoding the spectral envelope, and
    a spectral sample encoder (120) configured for encoding a plurality of spectral samples of the audio signal,
    wherein the spectral sample encoder (120) is configured to estimate an estimated bitrate needed for encoding for each spectral sample of one or more spectral samples of the plurality of spectral samples depending on the spectral envelope, and
    wherein the spectral sample encoder (120) is configured to encode each spectral sample of the plurality of spectral samples, depending on the estimated bitrate needed for encoding for the one or more spectral samples, according to a first coding rule or according to a second coding rule being different from the first coding rule.
  2. An audio encoder according to claim 1,
    wherein, if the estimated bitrate needed for encoding a spectral sample of the one or more spectral samples is greater than a threshold value, the spectral sample encoder (120) is configured to encode said spectral sample according to the first coding rule, and if said estimated bitrate needed for encoding said spectral samples is smaller than or equal to a threshold value, the spectral sample encoder (120) is configured to encode said spectral sample according to the second coding rule.
  3. An audio encoder according to claim 2,
    wherein the threshold value is 1 bit/sample.
  4. An audio encoder according to one of the preceding claims,
    wherein the spectral sample encoder (120) is configured to estimate the estimated bitrate needed for encoding for each spectral sample of the one or more spectral samples depending on an estimated variance of said spectral sample which depends on the spectral envelope.
  5. An audio encoder according to one of claims 1 to 3,
    wherein the spectral sample encoder (120) is configured to estimate the estimated bitrate needed for encoding for each spectral sample of the one or more spectral samples depending on the equation b k = 1 2 log 2 4.1159 σ k 2
    Figure imgb0102
    wherein bk is a k th spectral samples of the one or more spectral samples,
    wherein σ k 2
    Figure imgb0103
    is an estimated variance of said spectral sample.
  6. An audio encoder according to one of the preceding claims,
    wherein the spectral sample encoder (120) is configured to encode the spectral samples that are encoded according to the second coding rule by quantizing said spectral samples that are encoded according to the second coding rule employing an orthogonal matrix, and
    wherein the spectral sample encoder (120) is configured to encode the spectral samples that are encoded according to the first coding rule by quantizing said spectral samples that are encoded according to the first coding rule without employing the orthogonal matrix.
  7. An audio encoder according to 6,
    wherein the spectral sample encoder (120) is configured to encode the spectral samples that are encoded according to the second coding rule using: Q B Ax .
    Figure imgb0104
    wherein x is a vector comprising the spectral samples,
    wherein QB [·] is a quantizer defined as Q B y : = γ sign y 0 sign y 1 sign y B 1 0 0
    Figure imgb0105
    wherein y 0 y 1, y B-1 are quantized values resulting from quantization,
    wherein B indicates a total bitrate,
    wherein γ is a scaling coefficient,
    wherein A is the orthogonal matrix.
  8. An audio decoder for decoding an encoded audio signal, wherein the audio decoder comprises:
    an interface (130) configured for receiving an encoded spectral envelope of the audio signal and configured for receiving an encoded plurality of spectral samples of the audio signal, and
    a decoding unit (140) configured for decoding the encoded audio signal by decoding the encoded spectral envelope and by decoding the encoded plurality of spectral samples,
    wherein the decoding unit (140) is configured to receive or to estimate an estimated bitrate needed for encoding for each spectral sample of one or more spectral samples of the encoded plurality of spectral samples, and
    wherein the decoding unit (140) is configured to decode each spectral sample of the encoded plurality of spectral samples, depending on the estimated bitrate needed for encoding for the one or more spectral samples of the encoded plurality of spectral samples, according to a first coding rule or according to a second coding rule being different from the first coding rule.
  9. An audio decoder according to claim 8,
    wherein, if the estimated bitrate needed for encoding a spectral sample of the one or more spectral samples is greater than a threshold value, the decoding unit (140) is configured to decode said spectral sample according to the first coding rule, and if said estimated bitrate needed for encoding said spectral samples is smaller than or equal to a threshold value, the decoding unit (140) is configured to decode said spectral sample according to the second coding rule.
  10. An audio decoder according to claim 9,
    wherein the threshold value is 1 bit/sample.
  11. An audio decoder according to one of claims 8 to 10,
    wherein the decoding unit (140) is configured to estimate the estimated bitrate needed for encoding for each spectral sample of the one or more spectral samples depending on an estimated variance of said spectral sample which depends on the spectral envelope.
  12. An audio decoder according to one of claims 8 to 10,
    wherein the decoding unit (140) is configured to estimate the estimated bitrate needed for encoding for each spectral sample of the one or more spectral samples depending on the equation b k = 1 2 log 2 4.1159 σ k 2
    Figure imgb0106
    wherein bk is a k th spectral samples of the one or more spectral samples,
    wherein σ k 2
    Figure imgb0107
    is an estimated variance of said spectral sample.
  13. An audio decoder according to one of claims 8 to 12,
    wherein the decoding unit (140) is configured to decode the spectral samples that are decoded according to the second coding rule by employing an orthogonal matrix, and
    wherein the decoding unit (140) is configured to decode the spectral samples that are decoded according to the first coding rule without employing the orthogonal matrix.
  14. An audio decoder according to one of claims 8 to 13,
    wherein, if the estimated bitrate needed for encoding a spectral sample of the one or more spectral samples is smaller than another threshold value, the decoding unit (140) is configured to employ spectral noise shaping for decoding the encoded plurality of spectral samples.
  15. A system comprising:
    an audio encoder (105) according to one of claims 1 to 7, and
    an audio decoder (125) according to one of claims 8 to 14,
    wherein the audio encoder (105) is configured to feed a encoded spectral envelope of an encoded audio signal and an encoded plurality of spectral samples of the encoded audio signal into the audio decoder, and
    wherein the audio decoder (125) is configured to decode the encoded audio signal by decoding the encoded spectral envelope and by decoding the encoded plurality of spectral samples.
  16. A method for encoding an audio signal, wherein the audio signal is represented in a spectral domain, wherein the method comprises:
    determining a spectral envelope of the audio signal and for encoding the spectral envelope, and
    encoding a plurality of spectral samples of the audio signal,
    wherein encoding the plurality of spectral samples of the audio signal is conducted by estimating an estimated bitrate needed for encoding for each spectral sample of one or more spectral samples of the plurality of spectral samples depending on the spectral envelope, and
    wherein encoding the plurality of spectral samples of the audio signal is conducted by encoding each spectral sample of the plurality of spectral samples, depending on the estimated bitrate needed for encoding for the one or more spectral samples, according to a first coding rule or according to a second coding rule being different from the first coding rule.
  17. A method for decoding an encoded audio signal, wherein the method comprises:
    receiving an encoded spectral envelope of the audio signal and for receiving an encoded plurality of spectral samples of the audio signal, and
    decoding the encoded audio signal by decoding the encoded spectral envelope and by decoding the encoded plurality of spectral samples,
    wherein decoding the encoded audio signal is conducted by receiving or by estimating an estimated bitrate needed for encoding for each spectral sample of one or more spectral samples of the encoded plurality of spectral samples, and
    wherein decoding the encoded audio signal is conducted by decoding each spectral sample of the encoded plurality of spectral samples, depending on the estimated bitrate needed for encoding for the one or more spectral samples of the encoded plurality of spectral samples, according to a first coding rule or according to a second coding rule being different from the first coding rule.
  18. A computer program for implementing the method of claim 16 or 17 when being executed on a computer or signal processor.
EP18187597.2A 2018-03-22 2018-08-06 Audio coding with dithered quantization Active EP3544005B1 (en)

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
EP18163459 2018-03-22

Publications (2)

Publication Number Publication Date
EP3544005A1 true EP3544005A1 (en) 2019-09-25
EP3544005B1 EP3544005B1 (en) 2021-12-15

Family

ID=61837513

Family Applications (1)

Application Number Title Priority Date Filing Date
EP18187597.2A Active EP3544005B1 (en) 2018-03-22 2018-08-06 Audio coding with dithered quantization

Country Status (1)

Country Link
EP (1) EP3544005B1 (en)

Cited By (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN111681639A (en) * 2020-05-28 2020-09-18 上海墨百意信息科技有限公司 Multi-speaker voice synthesis method and device and computing equipment
US20220076365A1 (en) * 2019-11-26 2022-03-10 Ncr Corporation Frictionless security monitoring and management
CN116959459A (en) * 2023-09-19 2023-10-27 国网江西省电力有限公司信息通信分公司 Audio transmission method and system

Citations (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US7447631B2 (en) 2002-06-17 2008-11-04 Dolby Laboratories Licensing Corporation Audio coding system using spectral hole filling
WO2014161994A2 (en) * 2013-04-05 2014-10-09 Dolby International Ab Advanced quantizer

Patent Citations (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US7447631B2 (en) 2002-06-17 2008-11-04 Dolby Laboratories Licensing Corporation Audio coding system using spectral hole filling
WO2014161994A2 (en) * 2013-04-05 2014-10-09 Dolby International Ab Advanced quantizer

Non-Patent Citations (74)

* Cited by examiner, † Cited by third party
Title
"Frame error robust narrow-band and wideband embedded variable bit-rate coding of speech and audio from 8-32 kbit/s", ITU-T G.718, 2008
"Method for the Subjective Assessment of Intermediate Quality Levels of Coding Systems", ITU-R RECOMMENDATION BS.1534, 2003
"Methods for subjective determination of transmission quality", ITU-T RECOMMENDATION P.800, 1996
"Unified Speech and Audio Coding", MPEG-D (MPEG AUDIO TECHNOLOGIES, 2012
A. BELA; A. FRIGYIK; M. GUPTA: "Tech. Rep. UWEETR-2010-0006", 2010, DEPT. ELECT. ENG., UNIV. WASHINGTON, article "Introduction to the Dirichlet distribution and related processes"
A. BERTRAND: "Applications and trends in wireless acoustic sensor networks: A signal processing perspective", PROC. 18TH IEEE SYMP. COMMUN. VEH. TECHNOL. BENELUX, 2011, pages 1 - 6, XP032073747, DOI: doi:10.1109/SCVT.2011.6101302
A. EDELMAN; N. R. RAO: "Random matrix theory", ACTA NUMERICA, vol. 14, 2005, pages 233 - 297
A. GERSHO; R. M. GRAY: "Vector Quantization and Signal Compression", 1992, SPRINGER
A. JAGANNATAM: "Mersenne twister-A pseudo random number generator and its variants", 2008, DEPT. ELECT. COMPUT. ENG., GEORGE MASON UNIV.
A. MAGNANI; A. GHOSH; R. M. GRAY: "Optimal one-bit quantization", PROC. IEEE DATA COMPRESSION CONF., 2005, pages 270 - 278, XP010782778, DOI: doi:10.1109/DCC.2005.66
A. MAJUMDAR; K. RAMCHANDRAN; L. KOZINTSEV: "Distributed coding for wireless audio sensors", PROC. IEEE WORKSHOP APPL. SIGNAL PROCESS. AUDIO ACOUST., pages 209 - 212
A. ZAHEDI; J. 0STERGAARD; S. H. JENSEN; P. NAYLOR; S. BECH: "Coding and enhancement in wireless acoustic sensor networks", PROC. IEEE DATA COMPRESSION CONF., 2015, pages 293 - 302, XP032963977, DOI: doi:10.1109/DCC.2015.20
A. ZAHEDI; J. 0STERGAARD; S. H. JENSEN; S. BECH; P. NAYLOR: "Audio coding in wireless acoustic sensor networks", SIGNAL PROCESS., vol. 107, 2015, pages 141 - 152
B. GIROD; A. M. AARON; S. RANE; D. REBOLLO-MONEDERO: "Distributed video coding", PROC. IEEE, vol. 93, no. 1, January 2005 (2005-01-01), pages 71 - 83, XP011123854, DOI: doi:10.1109/JPROC.2004.839619
BACKSTROM TOM ET AL: "Arithmetic coding of speech and audio spectra using tcx based on linear predictive spectral envelopes", 2015 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP), IEEE, 19 April 2015 (2015-04-19), pages 5127 - 5131, XP033064629, DOI: 10.1109/ICASSP.2015.7178948 *
BOUFOUNOS P T ET AL: "1-Bit compressive sensing", INFORMATION SCIENCES AND SYSTEMS, 2008. CISS 2008. 42ND ANNUAL CONFERENCE ON, IEEE, PISCATAWAY, NJ, USA, 19 March 2008 (2008-03-19), pages 16 - 21, XP031282831, ISBN: 978-1-4244-2246-3 *
C. WALCK: "Handbook on Statistical Distributions for Experimentalists", 2007, UNIV. STOCKHOLM
C.-C. KUO; W. THONG: "Reduction of quantization error with adaptive Wiener filter in low bit rate coding", PROC. 11TH EUR. IEEE SIGNAL PROCESS. CONF., 2002, pages 1 - 4, XP032754048
D. E. KNUTH: "The Art of Computer Programming", vol. 2, 2007, ADDISON-WESLEY, article "Seminumerical algorithms"
D. KNUTH: "The Art of Computer Programming", 1998, ADDISON-WESLEY
D. PEARCE: "Automatic Speech Recognition on Mobile Devices and Over Communication Networks", 2008, SPRINGER, article "Distributed speech recognition standards", pages: 87 - 106
F. DE LA HUCHA ARCE; M. MOONEN; M. VERHELST; A. BERTRAND: "Adaptive quantization for multi-channel Wiener filter-based speech enhancement in wireless acoustic sensor networks", WIRELESS COMMUN. MOBILE COMPUT., vol. 2017, 2017
G. BARRIAC; R. MUDUMBAI; U. MADHOW: "Distributed beamforming for information transfer in sensor networks", PROC. 3RD INT. SYMP. INF. PROCESS. SENS. NETW., 2004, pages 81 - 88, XP058250798, DOI: doi:10.1145/984622.984635
G. E. 0IEN; T. A. RAMSTAD: "On the role of Wiener filtering in quantization and DPCM", PROC. IEEE NORWEGIAN SIGNAL PROCESS. SYMP. WORKSHOP, 2001
G. FUCHS; C. R. HELMRICH; G. MARKOVIC; M. NEUSINGER; E. RAVELLI; T. MORIYA: "Proc. ICASSP", 2015, IEEE, article "Low delay LPC and MDCT-based audio coding in the EVS codec", pages: 5723 - 5727
G. FUCHS; M. MULTRUS; M. NEUENDORF; R. GEIGER: "Mdct-based coder for highly adaptive speech and audio coding", EUROPEAN SIGNAL PROCESSING CONFERENCE (EUSIPCO 2009, 2009, pages 24 - 28
G. H. GOLUB; C. F. VAN LOAN: "Matrix Computations", 2004, JOHN HOPKINS UNIV. PRESS
G. W. STEWART: "The efficient generation of random orthogonal matrices with an application to condition estimators", SIAM J. NUMER. ANAL., vol. 17, no. 3, 1980, pages 403 - 409
H. DONG; J. LU; Y. SUN: "Distributed audio coding in wireless sensor networks", PROC. IEEE INT. CONF. COMPUT. INTELL. SECUR., vol. 2, 2006, pages 1695 - 1699, XP031013101
I. F. AKYILDIZ; T. MELODIA; K. R. CHOWDURY: "Wireless multimedia sensor networks: A survey", IEEE WIRELESS COMMUN., vol. 14, no. 6, December 2007 (2007-12-01), pages 32 - 39, XP011199018
I. H. WITTEN; R. M NEAL; J. G. CLEARY: "Arithmetic coding for data compression", COMMUNICATIONS OF THE ACM, vol. 30, no. 6, 1987, pages 520 - 540, XP000615171, DOI: doi:10.1145/214762.214771
J. BENESTY; M. SONDHI; Y. HUANG: "Springer Handbook of Speech Processing", 2008, SPRINGER
J. D. GIBSON; K. SAYOOD: "Lattice quantization", ADV. ELECTRON. ELECTRON PHYS., vol. 72, 1988, pages 259 - 330
J. FISCHER; T. BACKSTROM: "Wiener filtering in distributed speech and audio coding", IEEE SIGNAL PROCESS. LETT., 2017
J. MAKINEN; B. BESSETTE; S. BRUHN; P. OJALA; R. SALAMI; A. TALEB: "AMR-WB+: a new audio coding standard for 3rd generation mobile audio services", PROC. ICASSP, vol. 2, 2005, pages 1109 - 1112, XP010790838, DOI: doi:10.1109/ICASSP.2005.1415603
J. RISSANEN; G. G. LANGDON: "Arithmetic coding", IBM J. RES.DEVELOP., vol. 23, no. 2, 1979, pages 149 - 162, XP000938669
J. S. GAROFOLO ET AL.: "TIMIT: Acoustic-Phonetic Continuous Speech Corpus", LINGUISTIC DATA CONSORTIUM, 1993
J. VANDERKOOY; S. P. LIPSHITZ: "Dither in digital audio", JOURNAL OF THE AUDIO ENGINEERING SOCIETY, vol. 35, no. 12, 1987, pages 966 - 975
J.-M. VALIN; G. MAXWELL; T. B. TERRIBERRY; K. VOS: "Audio Engineering Society Convention", vol. 135, 2013, AUDIO ENGINEERING SOCIETY, article "High-quality, low-delay music coding in the OPUS codec"
M LI; J. KLEJSA; W. B. KLEIJN: "Distribution preserving quantization with dithering and transformation", IEEE SIGNAL PROCESSING LETTERS, vol. 17, no. 12, 2010, pages 1014 - 1017, XP011320162
M. BOSI ET AL.: "ISO/IECMPEG-2 advanced audio coding", J. AUDIO ENG. SOC., vol. 45, no. 10, 1997, pages 789 - 814
M. BOSI; K. BRANDENBURG; S. QUACKENBUSH; L. FIELDER; K. AKAGIRI; H. FUCHS; M. DIETZ; J. HERRE; G. DAVIDSON; Y. OIKAWA: "ISO/IEC MPEG-2 Advanced audio coding", AES CONVENTION, vol. 101, 2012
M. BOSI; R. E. GOLDBERG: "Introduction to Digital Audio Coding and Standards", 2003, KLUWER
M. MATSUMOTO; T. NISHIMURA: "Mersenne twister:A623-dimensionally equidistributed uniform pseudo-random number generator", ACM TRANS. MODEL. COMPUT. SIMUL., vol. 8, no. 1, 1998, pages 3 - 30, XP055244849, DOI: doi:10.1145/272991.272995
M. NEUENDORF; M. MULTRUS; N. RETTELBACH; G. FUCHS; J. ROBILLIARD; J. LECOMTE; S. WILDE; S. BAYER; S. DISCH; C. HELMRICH: "The ISO/MPEG unified speech and audio coding standard - consistent high quality for all content types and at all bit rates", JOURNAL OF THE AES, vol. 61, no. 12, 2013, pages 956 - 977, XP040636948
M. SCHOEFFLER; F.-R. STOTER; B. EDLER; J. HERRE: "Towards the next generation of web-based experiments: A case study assessing basic audio quality following the ITU-R recommendation BS. 1534 (MUSHRA", PROC. 1ST WEB AUDIO CONF., 2015
N. RAJPUT; A. A. NANAVATI: "Speech in Mobile and Pervasive Environments", 2012, WILEY, article "Distributed speech recognition", pages: 99 - 114
O. ROY; M. VETTERLI: "Rate-constrained collaborative noise reduction for wireless hearing aids", IEEE TRANS. SIGNAL PROCESS., vol. 57, no. 2, February 2009 (2009-02-01), pages 645 - 657, XP011238019
P. DIACONIS; M. SHAHSHAHANI: "The subgroup algorithm for generating uniform random variables", PROBAB. ENG. INF. SCI., vol. 1, no. 1, 1987, pages 15 - 32
P. L'ECUYER: "Encyclopedia of Quantitative Finance", 2010, WILEY, article "Pseudorandom number generators"
P. T. BOUFOUNOS; R. G. BARANIUK: "1-bit compressive sensing", PROC. IEEE INF. SCI. SYST. 42ND ANN. CONF., 2008, pages 16 - 21, XP031282831
R. LIENHART; I. KOZINTSEV; S. WEHR; M. YEUNG: "On the importance of exact synchronization for distributed audio signal processing", PROC. IEEE INT. CONF. ACOUST., SPEECH, SIGNAL PROCESS., vol. 4, 2003, pages IV-840 - 3
R. W. FLOYD; L. STEINBERG: "An adaptive algorithm for spatial gray-scale", PROC. SOC. INF. DISP., vol. 17, 1976, pages 75 - 77
S DISCH; A. NIEDERMEIER; C. R. HELMRICH; C. NEUKAM; K. SCHMIDT; R. GEIGER; J. LECOMTE; F. GHIDO; F. NAGEL; B. EDLER: "Audio Engineering Society Convention", vol. 141, 2016, AUDIO ENGINEERING SOCIETY, article "Intelligent gap filling in perceptual transform coding of audio"
S. BRAY; G. TZANETAKIS: "Distributed audio feature extraction for music", PROC. INT. CONF. MUSIC INF. RETRIEVAL, 2005, pages 434 - 437
S. DAS; A. CRACIUN; T. JAHNEL; T. BACKSTROM: "Spectral envelope statistics for source modelling in speech enhancement", PROC. ITG FACHTAGUNG SPRACHKOMMUNIKATION, 2016, pages 1 - 5
S. KORSE; T. JAHNEL; T. BACKSTROM: "Entropy coding of spectral envelopes for speech and audio coding using distribution quantization", PROC. INTERSPEECH, 2016, pages 2543 - 2547
S. NADARAJAH: "A generalized normal distribution", J. APPL. STATIST., vol. 32, no. 7, 2005, pages 685 - 694, XP009156797, DOI: doi:10.1080/02664760500079464
S. SAEEDNIA: "How to make the Hill cipher secure", CRYPTOLOGIA, vol. 24, no. 4, 2000, pages 353 - 360, XP000975779
S. VAUDENAY: "Decorrelation: a theory for block cipher security", J. CRYPTOL., vol. 16, no. 4, 2003, pages 249 - 286
T. BACKSTROM: "Enumerative algebraic coding for ACELP", PROC. INTERSPEECH, 2012
T. BACKSTROM: "Estimation of the probability distribution of spectral fine structure in the speech source", PROC. INTERSPEECH, 2017, pages 344 - 348
T. BACKSTROM: "Speech Coding with Code-Excited Linear Prediction", 2017, SPRINGER
T. BACKSTROM; C. R. HELMRICH: "Arithmetic coding of speech and audio spectra using TCX based on linear predictive spectral envelopes", PROC. IEEE INT. CONF. ACOUST., SPEECH, SIGNAL PROCESS., April 2015 (2015-04-01), pages 5127 - 5131, XP033064629, DOI: doi:10.1109/ICASSP.2015.7178948
T. BACKSTROM; F. GHIDO; J. FISCHER: "Blind recovery of perceptual models in distributed speech and audio coding", PROC. INTERSPEECH, 2016, pages 2483 - 2487, XP055369017, DOI: doi:10.21437/Interspeech.2016-27
T. BACKSTROM; J. FISCHER: "Coding of parametric models with randomized quantization in a distributed speech and audio codec", PROC. ITG FACHTAGUNG SPRACHKOMMUNIKATION, 2016, pages 1 - 5
T. BACKSTROM; J. FISCHER: "Fast randomization for distributed low-bitrate coding of speech and audio", IEEE/ACM TRANS. AUDIO, SPEECH, LANG. PROCESS., vol. 26, no. 1, January 2018 (2018-01-01), XP058381855, DOI: doi:10.1109/TASLP.2017.2757601
TOM BÄCKSTRÖM ET AL: "Dithered Quantization for Frequency-Domain Speech and Audio Coding", INTERSPEECH 2018, 1 January 2018 (2018-01-01), ISCA, pages 3533 - 3537, XP055579878, DOI: 10.21437/Interspeech.2018-46 *
TOM BACKSTROM ET AL: "Fast Randomization for Distributed Low-Bitrate Coding of Speech and Audio", IEEE/ACM TRANSACTIONS ON AUDIO, SPEECH, AND LANGUAGE PROCESSING, IEEE, USA, vol. 26, no. 1, 1 January 2018 (2018-01-01), pages 19 - 30, XP058381855, ISSN: 2329-9290, DOI: 10.1109/TASLP.2017.2757601 *
TS 26.190, ADAPTIVE MULTI-RATE (AMR-WB) SPEECH CODEC, 3GPP, 2007
TS 26.445, EVS CODEC DETAILED ALGORITHMIC DESCRIPTION; 3GPP TECHNICAL SPECIFICATION (RELEASE 12), 3GPP, 2014
V. PULKKI: "Spatial sound reproduction with directional audio coding", J. AUDIO ENG. SOC., vol. 55, no. 6, 2007, pages 503 - 516
Z. XIONG; A. D. LIVERIS; S. CHENG: "Distributed source coding for sensor networks", IEEE SIGNAL PROCESS. MAG., vol. 21, no. 5, September 2004 (2004-09-01), pages 80 - 94, XP011118155, DOI: doi:10.1109/MSP.2004.1328091
Z. XIONG; A. D. LIVERIS; Y. YANG: "Handbook on Array Processing and Sensor Networks", 2009, WILEY-IEEE PRESS, article "Distributed source coding", pages: 609 - 643

Cited By (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20220076365A1 (en) * 2019-11-26 2022-03-10 Ncr Corporation Frictionless security monitoring and management
US11727520B2 (en) * 2019-11-26 2023-08-15 Ncr Corporation Frictionless security monitoring and management
CN111681639A (en) * 2020-05-28 2020-09-18 上海墨百意信息科技有限公司 Multi-speaker voice synthesis method and device and computing equipment
CN111681639B (en) * 2020-05-28 2023-05-30 上海墨百意信息科技有限公司 Multi-speaker voice synthesis method, device and computing equipment
CN116959459A (en) * 2023-09-19 2023-10-27 国网江西省电力有限公司信息通信分公司 Audio transmission method and system
CN116959459B (en) * 2023-09-19 2023-12-22 国网江西省电力有限公司信息通信分公司 Audio transmission method and system

Also Published As

Publication number Publication date
EP3544005B1 (en) 2021-12-15

Similar Documents

Publication Publication Date Title
US11416742B2 (en) Audio signal encoding method and apparatus and audio signal decoding method and apparatus using psychoacoustic-based weighted error function
Ozerov et al. Coding-based informed source separation: Nonnegative tensor factorization approach
US11114110B2 (en) Noise attenuation at a decoder
JP6545748B2 (en) Audio classification based on perceptual quality for low or medium bit rates
EP3544005B1 (en) Audio coding with dithered quantization
US10460738B2 (en) Encoding apparatus for processing an input signal and decoding apparatus for processing an encoded signal
Bäckström et al. Blind Recovery of Perceptual Models in Distributed Speech and Audio Coding.
Bäckström et al. Fast randomization for distributed low-bitrate coding of speech and audio
Van Kuyk et al. On the information rate of speech communication
Bäckström et al. Dithered quantization for frequency-domain speech and audio coding
RU2716911C2 (en) Method and apparatus for encoding multiple audio signals and a method and apparatus for decoding a mixture of multiple audio signals with improved separation
Sugiura et al. Optimal coding of generalized-Gaussian-distributed frequency spectra for low-delay audio coder with powered all-pole spectrum estimation
Keser et al. A subspace based progressive coding method for speech compression
Moriya et al. Progress in LPC-based frequency-domain audio coding
US11295750B2 (en) Apparatus and method for noise shaping using subspace projections for low-rate coding of speech and audio
EP3008726B1 (en) Apparatus and method for audio signal envelope encoding, processing and decoding by modelling a cumulative sum representation employing distribution quantization and coding
Sugiura et al. Resolution warped spectral representation for low-delay and low-bit-rate audio coder
Vafin et al. Rate-distortion optimized quantization in multistage audio coding
Liu et al. A High Fidelity and Low Complexity Neural Audio Coding
Zhu et al. GAN-Based Inter-Channel Amplitude Ratio Decoding in Multi-Channel Speech Coding
EP3629327A1 (en) Apparatus and method for noise shaping using subspace projections for low-rate coding of speech and audio
Freiwald et al. Loss Functions for Deep Monaural Speech Enhancement
Wang et al. Critical band subspace-based speech enhancement using SNR and auditory masking aware technique
Kim KLT-based adaptive entropy-constrained vector quantization for the speech signals

Legal Events

Date Code Title Description
PUAI Public reference made under article 153(3) epc to a published international application that has entered the european phase

Free format text: ORIGINAL CODE: 0009012

STAA Information on the status of an ep patent application or granted ep patent

Free format text: STATUS: THE APPLICATION HAS BEEN PUBLISHED

AK Designated contracting states

Kind code of ref document: A1

Designated state(s): AL AT BE BG CH CY CZ DE DK EE ES FI FR GB GR HR HU IE IS IT LI LT LU LV MC MK MT NL NO PL PT RO RS SE SI SK SM TR

AX Request for extension of the european patent

Extension state: BA ME

STAA Information on the status of an ep patent application or granted ep patent

Free format text: STATUS: REQUEST FOR EXAMINATION WAS MADE

17P Request for examination filed

Effective date: 20200323

RBV Designated contracting states (corrected)

Designated state(s): AL AT BE BG CH CY CZ DE DK EE ES FI FR GB GR HR HU IE IS IT LI LT LU LV MC MK MT NL NO PL PT RO RS SE SI SK SM TR

STAA Information on the status of an ep patent application or granted ep patent

Free format text: STATUS: EXAMINATION IS IN PROGRESS

RAP1 Party data changed (applicant data changed or rights of an application transferred)

Owner name: FRAUNHOFER-GESELLSCHAFT ZUR FOERDERUNG DER ANGEWANDTEN FORSCHUNG E.V.

17Q First examination report despatched

Effective date: 20200515

STAA Information on the status of an ep patent application or granted ep patent

Free format text: STATUS: EXAMINATION IS IN PROGRESS

GRAP Despatch of communication of intention to grant a patent

Free format text: ORIGINAL CODE: EPIDOSNIGR1

STAA Information on the status of an ep patent application or granted ep patent

Free format text: STATUS: GRANT OF PATENT IS INTENDED

RIC1 Information provided on ipc code assigned before grant

Ipc: G10L 19/22 20130101AFI20210608BHEP

Ipc: G10L 19/032 20130101ALN20210608BHEP

Ipc: G10L 19/00 20130101ALN20210608BHEP

INTG Intention to grant announced

Effective date: 20210624

GRAS Grant fee paid

Free format text: ORIGINAL CODE: EPIDOSNIGR3

GRAA (expected) grant

Free format text: ORIGINAL CODE: 0009210

STAA Information on the status of an ep patent application or granted ep patent

Free format text: STATUS: THE PATENT HAS BEEN GRANTED

AK Designated contracting states

Kind code of ref document: B1

Designated state(s): AL AT BE BG CH CY CZ DE DK EE ES FI FR GB GR HR HU IE IS IT LI LT LU LV MC MK MT NL NO PL PT RO RS SE SI SK SM TR

REG Reference to a national code

Ref country code: GB

Ref legal event code: FG4D

Ref country code: CH

Ref legal event code: EP

REG Reference to a national code

Ref country code: IE

Ref legal event code: FG4D

Ref country code: DE

Ref legal event code: R096

Ref document number: 602018028108

Country of ref document: DE

REG Reference to a national code

Ref country code: AT

Ref legal event code: REF

Ref document number: 1456090

Country of ref document: AT

Kind code of ref document: T

Effective date: 20220115

REG Reference to a national code

Ref country code: LT

Ref legal event code: MG9D

RAP4 Party data changed (patent owner data changed or rights of a patent transferred)

Owner name: FRAUNHOFER-GESELLSCHAFT ZUR FOERDERUNG DER ANGEWANDTEN FORSCHUNG E.V.

REG Reference to a national code

Ref country code: NL

Ref legal event code: MP

Effective date: 20211215

PG25 Lapsed in a contracting state [announced via postgrant information from national office to epo]

Ref country code: RS

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20211215

Ref country code: LT

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20211215

Ref country code: FI

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20211215

Ref country code: BG

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20220315

REG Reference to a national code

Ref country code: AT

Ref legal event code: MK05

Ref document number: 1456090

Country of ref document: AT

Kind code of ref document: T

Effective date: 20211215

PG25 Lapsed in a contracting state [announced via postgrant information from national office to epo]

Ref country code: SE

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20211215

Ref country code: NO

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20220315

Ref country code: LV

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20211215

Ref country code: HR

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20211215

Ref country code: GR

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20220316

PG25 Lapsed in a contracting state [announced via postgrant information from national office to epo]

Ref country code: NL

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20211215

PG25 Lapsed in a contracting state [announced via postgrant information from national office to epo]

Ref country code: SM

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20211215

Ref country code: SK

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20211215

Ref country code: RO

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20211215

Ref country code: PT

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20220418

Ref country code: ES

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20211215

Ref country code: EE

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20211215

Ref country code: CZ

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20211215

PG25 Lapsed in a contracting state [announced via postgrant information from national office to epo]

Ref country code: PL

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20211215

Ref country code: AT

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20211215

REG Reference to a national code

Ref country code: DE

Ref legal event code: R097

Ref document number: 602018028108

Country of ref document: DE

PG25 Lapsed in a contracting state [announced via postgrant information from national office to epo]

Ref country code: IS

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20220415

PLBE No opposition filed within time limit

Free format text: ORIGINAL CODE: 0009261

STAA Information on the status of an ep patent application or granted ep patent

Free format text: STATUS: NO OPPOSITION FILED WITHIN TIME LIMIT

PG25 Lapsed in a contracting state [announced via postgrant information from national office to epo]

Ref country code: DK

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20211215

Ref country code: AL

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20211215

26N No opposition filed

Effective date: 20220916

PG25 Lapsed in a contracting state [announced via postgrant information from national office to epo]

Ref country code: SI

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20211215

PG25 Lapsed in a contracting state [announced via postgrant information from national office to epo]

Ref country code: MC

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20211215

REG Reference to a national code

Ref country code: CH

Ref legal event code: PL

GBPC Gb: european patent ceased through non-payment of renewal fee

Effective date: 20220806

PG25 Lapsed in a contracting state [announced via postgrant information from national office to epo]

Ref country code: LU

Free format text: LAPSE BECAUSE OF NON-PAYMENT OF DUE FEES

Effective date: 20220806

Ref country code: LI

Free format text: LAPSE BECAUSE OF NON-PAYMENT OF DUE FEES

Effective date: 20220831

Ref country code: CH

Free format text: LAPSE BECAUSE OF NON-PAYMENT OF DUE FEES

Effective date: 20220831

REG Reference to a national code

Ref country code: BE

Ref legal event code: MM

Effective date: 20220831

PG25 Lapsed in a contracting state [announced via postgrant information from national office to epo]

Ref country code: IT

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20211215

P01 Opt-out of the competence of the unified patent court (upc) registered

Effective date: 20230517

PG25 Lapsed in a contracting state [announced via postgrant information from national office to epo]

Ref country code: IE

Free format text: LAPSE BECAUSE OF NON-PAYMENT OF DUE FEES

Effective date: 20220806

Ref country code: FR

Free format text: LAPSE BECAUSE OF NON-PAYMENT OF DUE FEES

Effective date: 20220831

PG25 Lapsed in a contracting state [announced via postgrant information from national office to epo]

Ref country code: BE

Free format text: LAPSE BECAUSE OF NON-PAYMENT OF DUE FEES

Effective date: 20220831

PG25 Lapsed in a contracting state [announced via postgrant information from national office to epo]

Ref country code: GB

Free format text: LAPSE BECAUSE OF NON-PAYMENT OF DUE FEES

Effective date: 20220806

PGFP Annual fee paid to national office [announced via postgrant information from national office to epo]

Ref country code: DE

Payment date: 20230822

Year of fee payment: 6

PG25 Lapsed in a contracting state [announced via postgrant information from national office to epo]

Ref country code: HU

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT; INVALID AB INITIO

Effective date: 20180806

PG25 Lapsed in a contracting state [announced via postgrant information from national office to epo]

Ref country code: CY

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20211215