EP2374123A1 - Verbesserte codierung von mehrkanaligen digitalen audiosignalen - Google Patents

Verbesserte codierung von mehrkanaligen digitalen audiosignalen

Info

Publication number
EP2374123A1
EP2374123A1 EP09803838A EP09803838A EP2374123A1 EP 2374123 A1 EP2374123 A1 EP 2374123A1 EP 09803838 A EP09803838 A EP 09803838A EP 09803838 A EP09803838 A EP 09803838A EP 2374123 A1 EP2374123 A1 EP 2374123A1
Authority
EP
European Patent Office
Prior art keywords
sources
coding
sound
information
directivity
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
EP09803838A
Other languages
English (en)
French (fr)
Other versions
EP2374123B1 (de
Inventor
Florent Jaillet
David Virette
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Orange SA
Original Assignee
France Telecom SA
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by France Telecom SA filed Critical France Telecom SA
Publication of EP2374123A1 publication Critical patent/EP2374123A1/de
Application granted granted Critical
Publication of EP2374123B1 publication Critical patent/EP2374123B1/de
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Classifications

    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
    • G10L19/00Speech or audio signals analysis-synthesis techniques for redundancy reduction, e.g. in vocoders; Coding or decoding of speech or audio signals, using source filter models or psychoacoustic analysis
    • G10L19/008Multichannel audio signal coding or decoding using interchannel correlation to reduce redundancy, e.g. joint-stereo, intensity-coding or matrixing

Definitions

  • the present invention relates to the field of coding / decoding multi-channel digital audio signals. More particularly, the present invention relates to the parametric encoding / decoding of multi-channel audio signals.
  • This type of coding / decoding is based on the extraction of spatialization parameters so that at decoding, the spatial perception of the listener can be reconstituted.
  • Such a coding technique is known as "Binaural Cue
  • Coding in English (BCC) which aims on the one hand to extract and then code the auditory spatialization indices and on the other hand to code a monophonic or stereophonic signal from a mastering of the original multichannel signal.
  • This parametric approach is a low rate coding.
  • the main advantage of this coding approach is to allow a better compression ratio than conventional multi-channel digital audio compression methods while ensuring the backward compatibility of the compressed format obtained with existing coding formats and broadcasting systems.
  • FIG. 1 describes such a coding / decoding system in which the coder 100 constructs a sum signal ("downmix" in English) S s by matrixing in
  • the 110 channels of the original multichannel signal S and provides via a parameter extraction module 120, a reduced set of parameters P which characterize the spatial content of the original multichannel signal.
  • the multichannel signal is reconstructed (S ') by a synthesis module 160 which takes into account both the sum signal and the transmitted parameters P.
  • the sum signal has a reduced number of channels. These channels can be encoded by a conventional audio encoder before transmission or storage.
  • the sum signal has two channels and is compatible with conventional stereo broadcasting. Before transmission or storage, this sum signal can thus be encoded by any conventional stereo encoder. The signal thus coded is then compatible with the devices comprising the corresponding decoder which reconstruct the sum signal while ignoring the spatial data.
  • This coding scheme is based on a tree structure that allows the processing of only a limited number of channels simultaneously.
  • this technique is satisfactory for coding and decoding signals of reduced complexity used in the audiovisual field, for example for 5.1 signals.
  • it does not make it possible to obtain a satisfactory quality for more complex multichannel signals, for example for signals originating from direct multichannel sound taps or even surround signals.
  • the present invention improves the situation.
  • the method proposes a method of encoding a multichannel audio signal representing a sound scene comprising a plurality of sound sources.
  • the method is such that it comprises a step of decomposing the multichannel signal into frequency bands and the following steps per frequency band:
  • the information being representative of the spatial distribution of the sound source in the sound scene
  • the source-related directivity information gives not only the direction of the source but also the shape, or spatial distribution, of the source, ie the interaction that this source can have with the others. sources of the sound stage.
  • Knowing this directivity information associated with the sum signal will allow the decoder to obtain a signal of better quality which takes into account interchannel redundancies in a global manner and the probable phase oppositions between channels.
  • the sum signal from the coding according to the invention can be decoded by a standard decoder as known in the state of the art, thus providing interoperability with existing decoders.
  • the method further comprises a step of encoding secondary sources among the unselected sources of the sound scene and inserting coding information of the secondary sources in the bit stream.
  • the coding of the secondary sources will thus make it possible to provide additional precision on the decoded signal, in particular for complex signals of the type, for example, ambiophonic ones.
  • the coding information of the secondary sources may be, for example, coded spectral envelopes or coded time envelopes which may constitute parametric representations of the secondary sources.
  • the coding of secondary sources comprises the following steps:
  • pseudo-sources representing at least part of the secondary sources, by decorrelation with at least one main source and / or at least one coded secondary source;
  • part of the secondary sources or diffuse sources can then be represented by pseudo-sources. In this case, it is then possible to code this representation without increasing the coding rate.
  • the coding of the directivity information is performed by a parametric representation method.
  • This method is of low complexity and adapts particularly to the case of synthetic sound stage representing an ideal coding situation.
  • These parametric representations may comprise, for example, arrival direction information, for the reconstruction of a simulating directivity a plane wave or directivity pattern selection indicia in a dictionary of directivity shapes.
  • the coding of the directivity information is performed by a principal component analysis method delivering basic directivity vectors associated with gains allowing the reconstruction of the initial directivities.
  • the coding of the directivity information is performed by a combination of a principal component analysis method and a parametric representation method.
  • the present invention also relates to a method for decoding a multichannel audio signal representing a sound scene comprising a plurality of sound sources, from a bit stream and a sum signal.
  • the method is such that it comprises the following steps:
  • the decoding method thus makes it possible to reconstruct the multichannel signal of high quality for a faithful reproduction of the spatialized sound taking into account interchannel redundancies globally and the probable phase oppositions between channels.
  • the method further comprises the following steps:
  • the method further comprises the following step:
  • decoding the secondary sources by using a source actually transmitted and a predefined decorrelator to reconstruct pseudosources representative of at least a part of the secondary sources.
  • the method further comprises the following steps:
  • decoding secondary sources by using the source and the decorrelator index to reconstruct pseudo-sources representative of at least a part of the secondary sources. This makes it possible to find pseudo-sources representing part of the original secondary sources without degrading the sound reproduction of the decoded sound scene.
  • the present invention also relates to an encoder of a multichannel audio signal representing a sound scene having a plurality of sound sources.
  • the encoder is such that it comprises:
  • a directivity information obtaining module capable of obtaining this information by sound source of the sound scene and by frequency band, the information being representative of the spatial distribution of the sound source in the sound scene;
  • a coding module for the directivity information and a module for forming a bit stream comprising the coded directional information, the bit stream being able to be transmitted parallel to the sum signal.
  • This decoder is such that it comprises:
  • a means of storage readable by a computer or a processor, whether or not integrated into the encoder, possibly removable, stores a computer program implementing an encoding method and / or a decoding method according to the invention.
  • FIG. 1 illustrates a coding / decoding system of the state of the art of standard MPEG Surround system type
  • FIG. 2 illustrates an encoder and a coding method according to one embodiment of the invention
  • FIG. 3a illustrates a first embodiment of the coding of the directivities according to the invention
  • FIG. 3b illustrates a second embodiment of the coding of the directivities according to the invention
  • FIG. 4 represents examples of directivities used by the invention
  • FIG. 5 illustrates a decoder and a decoding method according to one embodiment of the invention
  • FIG. 6 represents an alternative embodiment of an encoder and a coding method according to the invention.
  • FIG. 7 represents an alternative embodiment of a decoder and a decoding method according to the invention.
  • FIGS. 8a and 8b respectively represent an exemplary device comprising an encoder and an exemplary device comprising a decoder according to the invention.
  • FIG. 2 illustrates in the form of a block diagram, an encoder according to one embodiment of the invention as well as the steps of a coding method according to one embodiment of the invention.
  • the encoder thus illustrated comprises a time-frequency transform module 210 which receives as input an original multichannel signal representing a sound scene comprising a plurality of sound sources.
  • This module therefore performs a step T of calculating the time-frequency transform of the original multichannel signal.
  • This transform is carried out for example by a short-term Fourier transform.
  • each of the n x channels of the original signal is window on the current time frame, then the Fourier transform F of the window signal is calculated using a fast calculation algorithm on ⁇ FFT points.
  • ⁇ FFT X n x containing the coefficients of the original multichannel signal in the frequency space.
  • the subsequent processing by the encoder is done by frequency band. This is done by cutting the matrix of coefficients X into a set of sub-matrices X j each containing the frequency coefficients in the j th band.
  • the signal is thus obtained for a given frequency band Sg.
  • a module for obtaining directional information 220 makes it possible to determine, by an OBT step, on the one hand, the directivities associated with each of the sources of the sound scene and, on the other hand, to determine the sources of the sound scene for the given frequency band.
  • the directivities are vectors of the same dimension as the number n s of channels of the multichannel signal S m .
  • Each source is associated with a vector of directivity.
  • the directivity vector associated with a source corresponds to the weighting function to be applied to this source before playing it on a loudspeaker, so as to best reproduce a direction of arrival and a source width . It is easily understood that for a very large number of regularly spaced loudspeakers, the directivity vector will faithfully represent the radiation of a sound source.
  • the vector of directivity will be obtained by the application of an inverse spherical Fourier transform on the components of the ambiophonic orders.
  • the ambiophonic signals correspond to a decomposition into spherical harmonics, hence the direct correspondence with the directivity of the sources.
  • the set of directivity vectors therefore constitutes a large amount of data that would be too expensive to transmit directly for low coding rate applications.
  • two methods of representing the directivities can for example be used.
  • the Cod.Di coding module 230 for directivity information can thus implement one of the two methods described below or a combination of the two methods.
  • a first method is a parametric modeling method that makes it possible to exploit knowledge a priori on the signal format used. It consists of transmitting only a very small number of parameters and reconstructing the directivities according to known coding schemes.
  • the associated directivity is known as a function of the direction of arrival of the sound source.
  • a search for peaks in the directivity diagram (by analogy with sinusoidal analysis, as explained for example in the document "Computer modeling of musical sound (analysis, transformation, synthesis)" by Sylvain Marchand, PhD thesis, University Bordeaux 1 , allows to detect the direction of arrival relatively faithfully.Other methods like the "matching pursuit", as presented in S.
  • a parametric representation can also use a simple form dictionary to represent the directivities.
  • FIG. 4 gives some simple forms of directivity (in azimuth) that can be used.
  • directivity in azimuth
  • FIG. 4 gives some simple forms of directivity (in azimuth) that can be used.
  • the directivities one associates with an element of the dictionary, the corresponding azimuth and a gain allowing to play on the amplitude of this vector of directivity of the dictionary.
  • One can thus, from a dictionary of form of directivity, to deduce the best form or combination of forms that will best reconstruct the initial directivity.
  • the directivity coding module 230 comprises a parametric modeling module which outputs P directionality parameters. These parameters are then quantized by the quantization module 240.
  • This first method makes it possible to obtain a very good level of compression when the scene corresponds to an ideal coding. This will be particularly the case on synthetic soundtracks. However, for complex scenes or microphonic sound, it is necessary to use more generic coding models, involving the transmission of a greater amount of information.
  • the representation of the directivity information is in the form of a linear combination of a limited number of basic directivities.
  • This method is based on the fact that the set of directivities at a given moment generally has a reduced dimension. Indeed, only a small number of sources is active at a given moment and the directivity for each source varies little with the frequency. It is thus possible to represent all the directivities in a group of frequency bands from a very small number of well-chosen basic directivities.
  • the parameters transmitted are then the basic directivity vectors for the group of bands considered, and for each directivity to be coded, the coefficients to be applied to the basic directivities to reconstruct the directivity considered.
  • PCA Principal component analysis
  • LT. Jolliffe in "Principal Component Analysis", Springer, 2002.
  • the application of the principal component analysis to the coding of the directivities is carried out as follows: firstly, a matrix of the initial directivities Di is formed, whose number of rows corresponds to the total number of sources of the sound stage, and the number of columns is the number of channels of the original multichannel. Then, one carries out properly the principal component analysis which corresponds to the diagonalization of the covariance matrix, and which gives the matrix of the eigenvectors. Finally, we select the eigenvectors carrying the most important information and corresponding to the eigenvalues of higher value.
  • the number of eigenvectors to keep may be fixed or variable in time depending on the available flow.
  • This new base thus gives the matrix D B T.
  • the representation of the directivities is therefore made from basic directivity.
  • the matrix of directivities Di is written as the linear combination of these basic directivities.
  • Di G D D B
  • D B is the matrix of the basic directivities for all the bands
  • G D the matrix of the associated gains.
  • the number of rows of this matrix represents the total number of sources of the sound stage and the number of columns represents the number of basic directivity vectors.
  • basic directivities are sent by group of considered bands, in order to more accurately represent the directivities. It is possible, for example, to provide two basic directivity groups: one for low frequencies and one for high frequencies. The limit between these two groups can for example be chosen between 5 and 7 kHz.
  • the coding module 230 comprises a main component analysis module delivering basic directivity vectors D B and associated coefficients or gain vectors G D -
  • FIG. 3a illustrates in a detailed manner, the directivity coding block 230, in a first variant embodiment.
  • This coding mode uses the two diagrams of representation of the directivities.
  • a module 310 performs parametric modeling as previously explained to provide directional parameters (P).
  • a module 320 performs principal component analysis to provide both basic directivity vectors (D B ) and associated coefficients (G D ).
  • a selection module 330 selects frequency band per frequency band, the best mode of coding for the directivity by choosing the best compromise reconstruction of the directivities / flow.
  • the choice of the representation chosen (parametric representation or by linear combination of basic directivities) is done in order to optimize the efficiency of the compression.
  • a selection criterion is, for example, the minimization of the mean squared error.
  • a perceptual weighting may possibly be used for the choice of the directivity coding mode. This weighting is intended for example to promote the reconstruction of the directivities in the frontal area, for which the ear is more sensitive.
  • the error function to be minimized in the case of the ACP encoding model can be in the following form:
  • the directivity parameters from the selection module are then quantized by a step Q by the quantization module 240 of FIG. 2.
  • the two coding modes are cascades.
  • Figure 3b illustrates in detail this block of coding.
  • a parametric modeling module 340 performs a modeling for a certain number of directivities and outputs at the same time directivity parameters (P) for the modeled directivities and unmodelled directivities or residual directivities DiR .
  • D residual directivities
  • main component analysis module 350 which outputs basic directional vectors (D B ) and associated coefficients (G D ).
  • D B basic directional vectors
  • G D associated coefficients
  • the directivity parameters, the basic directivity vectors as well as the coefficients are provided at the input of the quantization module 240 of FIG. 2.
  • Quantization Q is performed by reducing the accuracy as a function of perception data and then applying entropy coding. Also, the possibility of exploiting the redundancy between frequency bands or between successive frames can reduce the flow. Intra-frame or inter-frame predictions on the parameters can therefore be used. In general, the standard methods of quantification can be used. On the other hand, the vectors to be quantified being orthonormed, this property can be exploited during the scalar quantization of the components of the vector. Indeed, for a vector of dimension N, only N-I components will have to be quantified, the last component being able to be recalculated.
  • a construction module of a bit stream 250 inserts this coded direction information in a bit stream Fb according to the step Con.Fb.
  • the encoder as described here further comprises a selection module 260 able to select in the step Select main sources (S p ⁇ n c ) among the sources of the sound scene to be encoded (S tot ).
  • a particular embodiment uses a principal component analysis method, ACP, in each frequency band in block 220 to extract all the sources of the sound scene (S tot ).
  • ACP principal component analysis method
  • the sources of greater importance are then selected by the module 260 to constitute the main sources (S p ⁇ nc ), which are then stamped in step M by the module 270 to construct a sum signal (S Sf ,) (or "downmix" in English).
  • the number of main sources (S pr j nc ) is chosen according to the number of channels of the sum signal. This number is chosen less than or equal to the number of channels. Preferably, a number of main sources is chosen equal to the number of channels of the sum signal.
  • the matrix M is then a predefined square matrix.
  • This sum signal per frequency band undergoes an inverse time-frequency transformation T 1 by the inverse transform module 290 in order to provide a time sum signal (S s ).
  • This sum signal is then encoded by a speech coder or an audio coder of the state of the art (for example: G.729.1 or MPEG-4 AAC).
  • Secondary sources (S sec ) can be coded by a coding module
  • bit stream construction module 250 280 and added to the bit stream in the bit stream construction module 250.
  • the coding module 280 which may in one embodiment be a short-term Fourier transform coding module. These sources can then be separately encoded using the aforementioned audio or speech coders.
  • the coefficients of the transform of these secondary sources can be coded directly only in the bands considered to be important.
  • Secondary sources can be encoded by parametric representations; these representations may be in the form of spectral envelope or temporal envelope.
  • This method of encoding a multichannel signal as described is particularly interesting in that the analysis is made on windows that can be of short length.
  • this coding model causes a low algorithmic delay allowing its use in applications where control of the delay is important.
  • the encoder as described implements an additional preprocessing step P by a preprocessing module 215.
  • This module performs a basic changeover step in order to express the sound scene using the flat wave decomposition of the acoustic field.
  • the original surround signal is seen as the angular Fourier transform of a sound field.
  • the first plane wave decomposition operation therefore corresponds to taking the omnidirectional component of the ambiophonic signal as representing the zero angular frequency (this component is therefore a real component).
  • the following surround components order 1, 2, 3, etc. are combined to obtain the complex coefficients of the angular Fourier transform.
  • the first component represents the real part
  • the second component represents the imaginary part.
  • O For a two-dimensional representation, for an order O, we obtain O + 1 complex components.
  • a Short Term Fourier Transform (on the time dimension) is then applied to obtain the Fourier transforms (in the frequency domain) of each angular harmonic. This step then integrates the transformation step T of the module 210. the complete angular transform by recreating the harmonics of negative frequencies by Hermitian symmetry.
  • an inverse Fourier transform is carried out on the dimension of the angular frequencies to pass in the domain of the directivities.
  • This pre-processing step allows the coder to work in a signal space whose physical and perceptual interpretation is simplified, which makes it possible to more effectively exploit knowledge of spatial auditory perception and thus improve coding performance.
  • the encoding of the surround signals remains possible without this pre-processing step. For non-surround signals, this step is not necessary. For these signals, the knowledge of the recording system or reproduction associated with the signal makes it possible to directly interpret the signals as a plane wave decomposition of the acoustic field.
  • Figure 5 now describes a decoder and a decoding method in one embodiment of the invention.
  • This decoder receives as input the bit stream F b as constructed by the encoder described above as well as the sum signal S 5 .
  • the first decoding step consists of realizing the time-frequency transform T of the sum signal S s by the transform module 510 to obtain a sum signal per frequency band, S sf,.
  • This transform is carried out using, for example, the short-term Fourier transform. It should be noted that other transforms or filterbanks may also be used, including non-uniform filterbanks according to a perception scale (e.g. Bark). It may be noted that in order to avoid discontinuities during the reconstruction of the signal from this transform, a recovery addition method is used. For the time frame considered, the step of calculating the transform of
  • Fourier in the short term is to window each of the n f channels of the sum signal S s using a window w of length greater than the time frame, and then to calculate the Fourier transform of the window signal with the help of a fast calculation algorithm on npFT points.
  • a complex matrix F of size npF T xn f containing the coefficients of the sum signal in the frequency space is thus obtained.
  • the entire processing is done in frequency bands.
  • the matrix coefficients F is cut into a plurality of submatrices F j each containing the frequency coefficients in the j th band.
  • Different choices for the frequency division of the bands are possible.
  • symmetrical bands with respect to the zero frequency in the Fourier transform are chosen in the short term.
  • the description of the decoding steps performed by the decoder will be made for a given frequency band. The steps are of course carried out for each of the frequency bands to be processed.
  • the module 520 performs a dematrix N of the frequency coefficients of the signal transform sum of the frequency band considered so as to find the main sources of the sound scene. More specifically, the S p ⁇ ⁇ c matrix of frequency coefficients for the current frequency band of the n p ⁇ nc main sources is obtained according to the relation:
  • N is of dimension n f xn p ⁇ nc and B is a matrix of dimension n bm xn f where n bm is the number of components (or bins) frequency retained in the frequency band considered.
  • N I.
  • the number of rows of the matrix N corresponds to the number of channels of the sum signal, and the number of columns corresponds to the number of main sources transmitted.
  • the dimensions are inverted, I being an identity matrix of dimensions n p ⁇ nc xn p ⁇ nc .
  • the lines of B are the frequency components in the current frequency band, the columns correspond to the channels of the sum signal.
  • the lines of S p ⁇ n c are the frequency components in the current frequency band, and each column corresponds to a main source.
  • the number of main sources n p ⁇ nc is preferably less than or equal to the number n f of channels of the sum signal to ensure that the operation is invertible, and may possibly be different for each frequency band.
  • the number of sources to be reconstructed in the current frequency band to obtain a satisfactory reconstruction of the scene is greater than the number of channels of the sum signal.
  • additional or secondary sources are coded and then decoded from the bitstream for the current band by the module 550 for decoding the bitstream.
  • This decoding module decodes the information contained in the bit stream and in particular, the directional information and, where appropriate, the secondary sources.
  • the decoding of the secondary sources is carried out by the inverse operations that those which were carried out with the coding. Whatever the coding method that has been chosen for the secondary sources, if reconstruction data or coding information of the secondary sources has been transmitted in the bit stream for the current band, the corresponding data is decoded to reconstruct the dry matrix S frequency coefficients in the current band of the n sec secondary sources.
  • the shape of the dry matrix S is similar to the matrix S pnnc , that is, the lines are the frequency components in the current frequency band, and each column corresponds to a secondary source.
  • Ssupp according to S relation (S pntlL S mpp j therefore S is a matrix of dimension n b, n tot xn - Also, the form is identical to S p ⁇ matrices n c and S supp: lines are the frequency components in the current frequency band, each column is a source, with n tot sources in total.
  • the directivity information is extracted from the bit stream at the Decod step. Fb by the module 550.
  • the possible outputs of this decoding module of the bitstream depend on the coding methods of the directivities used in the coding. They can be in the form of basic directivity vectors D B and associated coefficients G D and / or modeling parameters P.
  • This data is then transmitted to a directional information reconstruction module 560 which performs the decoding of the directional information by reverse operations from those performed in the coding.
  • the number of directivities to be reconstructed is equal to the number n tot of sources in the frequency band considered, each source being associated with a directional vector.
  • the matrix of directivities Di is written as the linear combination of these basic directivities.
  • Di G D D B
  • D B is the matrix of the basic directivities for all the bands
  • G D the matrix of the associated gains.
  • This gain matrix has a number of lines equal to the total number of sources n tot , and a number of columns equal to the number of basic directivity vectors.
  • basic directivities are decoded by group of frequency bands considered, in order to more accurately represent the directivities.
  • group of frequency bands considered in order to more accurately represent the directivities.
  • a vector of gains associated with the basic directivities is then decoded for each band.
  • Y SD T , where Y is the reconstructed signal in the band.
  • the rows of the matrix Y are the frequency components in the current frequency band, and each column corresponds to a channel of the multichannel signal to be reconstructed.
  • the corresponding time signals are then obtained by inverse Fourier transform T ', using a fast algorithm implemented by the inverse transform module 540. This gives the multichannel signal S m on the current time frame.
  • the different time frames are then combined by conventional overlap-add (or overlap-add) method to reconstruct the complete multichannel signal.
  • temporal or frequency smoothing of the parameters can be used both for analysis and synthesis to ensure smooth transitions in the sound scene.
  • a sign of sudden change of the sound stage may be reserved in the bit stream to avoid smoothing the decoder in the case of detection of a rapid change in the composition of the sound stage.
  • conventional methods of adapting the resolution of the time-frequency analysis can be used (change in the size of the analysis and synthesis windows over time).
  • a base change module can perform a pre-processing P "1 to obtain a plane wave decomposition of the signals, a base change module 570 performs the reverse operation from the signals. in plane waves to find the original multichannel signal.
  • the coding of the embodiment described with reference to FIG. 2 makes it possible to obtain efficient compression when the complexity of the scene remains limited.
  • the complexity of the scene is greater, ie when the scene contains a high number of active sources in a frequency band, or important diffuse components, a large number of sources and associated directivity becomes necessary for to obtain a good quality of restitution of the scene. The effectiveness of compression is then reduced.
  • the encoder as represented in FIG. 6 comprises the modules 215, 210, 220, 230, 240 as described with reference to FIG. 2.
  • This encoder also includes the modules 260, 270 and 290 as described with reference to FIG.
  • This encoder comprises a coding module of the secondary sources 620, which differs from the module 280 of Figure 2 in the case where the number of secondary sources is important.
  • this coding module 620 a method of parametric coding secondary sources is implemented by this coding module 620.
  • the field is perceptibly comparable to a diffuse field, and the representation of the field by one or more statistical characteristics of the field is sufficient to reconstruct a perceptually equivalent field.
  • the spatially diffuse components of the sound scene can be perceptively reconstructed from the simple knowledge of the corresponding directivity, and by controlling the coherence of the created field. This can be done by using pseudo-sources constructed by decorrelation, from a limited number of transmitted sources and by using the diffuse component directivity estimated on the original multichannel signal. The objective is then to reconstruct a sound field statistically and perceptually equivalent to the original, even if it consists of signals whose waveforms are different.
  • a number of secondary sources are not transmitted and are replaced by pseudo-sources obtained by decorrelation of the transmitted sources, or by any other artificial source decorrelated sources transmitted. This avoids the transmission of data corresponding to these sources and significantly improves coding efficiency.
  • a source to be transmitted to the decoder and a predefined decorrelator known from both the encoder and the decoder, to be applied to the transmitted source to select pseudo-sources for the decoder, are chosen.
  • a parametric representation of the secondary sources is obtained by the coding module of the secondary sources 620 and is also transmitted to the construction module of the bitstream.
  • This parametric representation of secondary sources or diffuse sources is effected for example by a spectral envelope.
  • a time envelope can also be used.
  • the pseudo-sources are calculated by a decorrelation module 630 which calculates the decorrelated sources from at least one main source or with at least one coded secondary source to be transmitted.
  • decorrelators and several initial sources can be used, and one can select the initial source associated with a type of decorrelator giving the best reconstruction result.
  • These decorrelation data such as the index of the correlator used and the choice data of the initial source as the index of the source, are then transmitted to the building module of the bit stream to be inserted. The number of sources to transmit is reduced while maintaining a good perceptive quality of the reconstructed signal.
  • FIG. 7 represents a decoder and a decoding method adapted to the coding according to the variant embodiment described in FIG. 6.
  • This decoder comprises the modules 510, 520, 530, 540, 570, 560 as described with reference to FIG. 5. This decoder differs from that described in FIG. information decoded by the decoding module of the bit stream 720 and the decorrelation calculation block 710.
  • the module 720 obtains, in addition to directional information from sources of the sound scene and, if appropriate, decoded secondary sources, parametric representation data of certain secondary sources or diffuse sources and possibly information on the decorrelator and the transmitted sources. to use to rebuild the pseudo-sources.
  • the latter information is then used by the decorrelation module 710 which makes it possible to reconstruct the secondary pseudo-sources which will be combined with the main sources and the other potential secondary sources in the spatialization module as described with reference to FIG.
  • the encoders and decoders as described with reference to FIGS. 2, 6 and 5, 7 can be integrated into a multimedia equipment of the living room decoder type, a computer or even communication equipment such as a mobile telephone or personal electronic organizer.
  • FIG. 8a represents an example of such multimedia equipment or coding device comprising an encoder according to the invention.
  • This device comprises a PROC processor cooperating with a memory block BM having a storage and / or working memory MEM.
  • the memory block may advantageously comprise a computer program comprising code instructions for implementing the steps of the coding method within the meaning of the invention, when these instructions are executed by the processor PROC, and in particular the steps of
  • the information being representative of the spatial distribution of the sound source in the sound scene
  • bit stream comprising the coded directional information
  • the description of FIG. 2 repeats the steps of an algorithm of such a computer program.
  • the computer program can also be stored on a memory medium readable by a reader of the device or downloadable in the memory space of the equipment.
  • the device comprises an input module adapted to receive a multichannel signal representing a sound scene, either by a communication network, or by reading a content stored on a storage medium.
  • This multimedia equipment may also include means for capturing such a multichannel signal.
  • the device comprises an output module able to transmit a bit stream
  • FIG. 8b illustrates an example of multimedia equipment or decoding device comprising a decoder according to the invention.
  • This device comprises a PROC processor cooperating with a memory block BM having a storage and / or working memory MEM.
  • the memory block may advantageously comprise a computer program comprising code instructions for implementing the steps of the decoding method in the sense of the invention, when these instructions are executed by the processor PROC, and in particular the steps of: - extraction in the bitstream and decoding information of directivities representative of the spatial distribution of the sources in the sound scene;
  • the computer program can also be stored on a memory medium readable by a reader of the device or downloadable in the memory space of the equipment.
  • the device comprises an input module adapted to receive a bit stream Fb and a sum signal S s coming for example from a communication network. These input signals can come from a reading on a storage medium.
  • the device comprises an output module capable of transmitting a multichannel signal decoded by the decoding method implemented by the equipment.
  • This multimedia equipment may also include speaker-type reproduction means or communication means capable of transmitting this multi-channel signal.
  • Such multimedia equipment may include both the encoder and the decoder according to the invention.
  • the input signal then being the original multichannel signal and the output signal, the decoded multichannel signal.

Landscapes

  • Engineering & Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • Mathematical Physics (AREA)
  • Computational Linguistics (AREA)
  • Signal Processing (AREA)
  • Health & Medical Sciences (AREA)
  • Audiology, Speech & Language Pathology (AREA)
  • Human Computer Interaction (AREA)
  • Acoustics & Sound (AREA)
  • Multimedia (AREA)
  • Compression, Expansion, Code Conversion, And Decoders (AREA)
EP09803838.3A 2008-12-15 2009-12-11 Verbesserte codierung von mehrkanaligen digitalen audiosignalen Active EP2374123B1 (de)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
FR0858560 2008-12-15
PCT/FR2009/052491 WO2010070225A1 (fr) 2008-12-15 2009-12-11 Codage perfectionne de signaux audionumeriques multicanaux

Publications (2)

Publication Number Publication Date
EP2374123A1 true EP2374123A1 (de) 2011-10-12
EP2374123B1 EP2374123B1 (de) 2019-04-10

Family

ID=40679401

Family Applications (1)

Application Number Title Priority Date Filing Date
EP09803838.3A Active EP2374123B1 (de) 2008-12-15 2009-12-11 Verbesserte codierung von mehrkanaligen digitalen audiosignalen

Country Status (4)

Country Link
US (1) US8964994B2 (de)
EP (1) EP2374123B1 (de)
ES (1) ES2733878T3 (de)
WO (1) WO2010070225A1 (de)

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN111653283A (zh) * 2020-06-28 2020-09-11 讯飞智元信息科技有限公司 一种跨场景声纹比对方法、装置、设备及存储介质

Families Citing this family (74)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
KR101697550B1 (ko) * 2010-09-16 2017-02-02 삼성전자주식회사 멀티채널 오디오 대역폭 확장 장치 및 방법
ES2643163T3 (es) * 2010-12-03 2017-11-21 Fraunhofer-Gesellschaft zur Förderung der angewandten Forschung e.V. Aparato y procedimiento para codificación de audio espacial basada en geometría
US9495591B2 (en) 2012-04-13 2016-11-15 Qualcomm Incorporated Object recognition using multi-modal matching scheme
US8935164B2 (en) * 2012-05-02 2015-01-13 Gentex Corporation Non-spatial speech detection system and method of using same
EP2665208A1 (de) 2012-05-14 2013-11-20 Thomson Licensing Verfahren und Vorrichtung zur Komprimierung und Dekomprimierung einer High Order Ambisonics-Signaldarstellung
US9288603B2 (en) 2012-07-15 2016-03-15 Qualcomm Incorporated Systems, methods, apparatus, and computer-readable media for backward-compatible audio coding
US9190065B2 (en) 2012-07-15 2015-11-17 Qualcomm Incorporated Systems, methods, apparatus, and computer-readable media for three-dimensional audio coding using basis function coefficients
US9761229B2 (en) * 2012-07-20 2017-09-12 Qualcomm Incorporated Systems, methods, apparatus, and computer-readable media for audio object clustering
US9516446B2 (en) 2012-07-20 2016-12-06 Qualcomm Incorporated Scalable downmix design for object-based surround codec with cluster analysis by synthesis
JP6133422B2 (ja) * 2012-08-03 2017-05-24 フラウンホーファー−ゲゼルシャフト・ツール・フェルデルング・デル・アンゲヴァンテン・フォルシュング・アインゲトラーゲネル・フェライン マルチチャネルをダウンミックス/アップミックスする場合のため一般化された空間オーディオオブジェクト符号化パラメトリック概念のデコーダおよび方法
US9396732B2 (en) 2012-10-18 2016-07-19 Google Inc. Hierarchical deccorelation of multichannel audio
US9913064B2 (en) 2013-02-07 2018-03-06 Qualcomm Incorporated Mapping virtual speakers to physical speakers
US10178489B2 (en) 2013-02-08 2019-01-08 Qualcomm Incorporated Signaling audio rendering information in a bitstream
US9609452B2 (en) 2013-02-08 2017-03-28 Qualcomm Incorporated Obtaining sparseness information for higher order ambisonic audio renderers
US9883310B2 (en) 2013-02-08 2018-01-30 Qualcomm Incorporated Obtaining symmetry information for higher order ambisonic audio renderers
US9466305B2 (en) 2013-05-29 2016-10-11 Qualcomm Incorporated Performing positional analysis to code spherical harmonic coefficients
US10499176B2 (en) 2013-05-29 2019-12-03 Qualcomm Incorporated Identifying codebooks to use when coding spatial components of a sound field
EP2860728A1 (de) * 2013-10-09 2015-04-15 Thomson Licensing Verfahren und Vorrichtung zur Codierung und Decodierung gerichteter Nebeninformationen
US9489955B2 (en) 2014-01-30 2016-11-08 Qualcomm Incorporated Indicating frame parameter reusability for coding vectors
US9922656B2 (en) 2014-01-30 2018-03-20 Qualcomm Incorporated Transitioning of ambient higher-order ambisonic coefficients
CN104882145B (zh) * 2014-02-28 2019-10-29 杜比实验室特许公司 使用音频对象的时间变化的音频对象聚类
US20150264483A1 (en) * 2014-03-14 2015-09-17 Qualcomm Incorporated Low frequency rendering of higher-order ambisonic audio data
US10412522B2 (en) 2014-03-21 2019-09-10 Qualcomm Incorporated Inserting audio channels into descriptions of soundfields
US9852737B2 (en) 2014-05-16 2017-12-26 Qualcomm Incorporated Coding vectors decomposed from higher-order ambisonics audio signals
US9620137B2 (en) 2014-05-16 2017-04-11 Qualcomm Incorporated Determining between scalar and vector quantization in higher order ambisonic coefficients
US10770087B2 (en) 2014-05-16 2020-09-08 Qualcomm Incorporated Selecting codebooks for coding vectors decomposed from higher-order ambisonic audio signals
US9838819B2 (en) * 2014-07-02 2017-12-05 Qualcomm Incorporated Reducing correlation between higher order ambisonic (HOA) background channels
CN105336332A (zh) * 2014-07-17 2016-02-17 杜比实验室特许公司 分解音频信号
US9747910B2 (en) 2014-09-26 2017-08-29 Qualcomm Incorporated Switching between predictive and non-predictive quantization techniques in a higher order ambisonics (HOA) framework
US9940937B2 (en) 2014-10-10 2018-04-10 Qualcomm Incorporated Screen related adaptation of HOA content
US10140996B2 (en) 2014-10-10 2018-11-27 Qualcomm Incorporated Signaling layers for scalable coding of higher order ambisonic audio data
US9984693B2 (en) 2014-10-10 2018-05-29 Qualcomm Incorporated Signaling channels for scalable coding of higher order ambisonic audio data
US9666192B2 (en) 2015-05-26 2017-05-30 Nuance Communications, Inc. Methods and apparatus for reducing latency in speech recognition applications
US10559303B2 (en) * 2015-05-26 2020-02-11 Nuance Communications, Inc. Methods and apparatus for reducing latency in speech recognition applications
US10249312B2 (en) 2015-10-08 2019-04-02 Qualcomm Incorporated Quantization of spatial vectors
US9961467B2 (en) 2015-10-08 2018-05-01 Qualcomm Incorporated Conversion from channel-based audio to HOA
US9961475B2 (en) 2015-10-08 2018-05-01 Qualcomm Incorporated Conversion from object-based audio to HOA
US10070094B2 (en) 2015-10-14 2018-09-04 Qualcomm Incorporated Screen related adaptation of higher order ambisonic (HOA) content
US9959880B2 (en) 2015-10-14 2018-05-01 Qualcomm Incorporated Coding higher-order ambisonic coefficients during multiple transitions
US9832587B1 (en) 2016-09-08 2017-11-28 Qualcomm Incorporated Assisted near-distance communication using binaural cues
MX2019005147A (es) * 2016-11-08 2019-06-24 Fraunhofer Ges Forschung Aparato y metodo para codificar o decodificar una se?al multicanal usando una ganancia lateral y una ganancia residual.
US10659906B2 (en) 2017-01-13 2020-05-19 Qualcomm Incorporated Audio parallax for virtual reality, augmented reality, and mixed reality
US10405126B2 (en) 2017-06-30 2019-09-03 Qualcomm Incorporated Mixed-order ambisonics (MOA) audio data for computer-mediated reality systems
US11164606B2 (en) 2017-06-30 2021-11-02 Qualcomm Incorporated Audio-driven viewport selection
US10469968B2 (en) 2017-10-12 2019-11-05 Qualcomm Incorporated Rendering for computer-mediated reality systems
US11062713B2 (en) 2018-06-25 2021-07-13 Qualcomm Incorported Spatially formatted enhanced audio data for backward compatible audio bitstreams
US10999693B2 (en) 2018-06-25 2021-05-04 Qualcomm Incorporated Rendering different portions of audio data using different renderers
US11081116B2 (en) 2018-07-03 2021-08-03 Qualcomm Incorporated Embedding enhanced audio transports in backward compatible audio bitstreams
US10924876B2 (en) 2018-07-18 2021-02-16 Qualcomm Incorporated Interpolating audio streams
US11128976B2 (en) 2018-10-02 2021-09-21 Qualcomm Incorporated Representing occlusion when rendering for computer-mediated reality systems
US11798569B2 (en) 2018-10-02 2023-10-24 Qualcomm Incorporated Flexible rendering of audio data
US11019449B2 (en) 2018-10-06 2021-05-25 Qualcomm Incorporated Six degrees of freedom and three degrees of freedom backward compatibility
US10972853B2 (en) 2018-12-21 2021-04-06 Qualcomm Incorporated Signalling beam pattern with objects
US11184731B2 (en) 2019-03-20 2021-11-23 Qualcomm Incorporated Rendering metadata to control user movement based audio rendering
US11122386B2 (en) 2019-06-20 2021-09-14 Qualcomm Incorporated Audio rendering for low frequency effects
US11361776B2 (en) 2019-06-24 2022-06-14 Qualcomm Incorporated Coding scaled spatial components
US11538489B2 (en) 2019-06-24 2022-12-27 Qualcomm Incorporated Correlating scene-based audio data for psychoacoustic audio coding
US12073842B2 (en) 2019-06-24 2024-08-27 Qualcomm Incorporated Psychoacoustic audio coding of ambisonic audio data
US11432097B2 (en) 2019-07-03 2022-08-30 Qualcomm Incorporated User interface for controlling audio rendering for extended reality experiences
US11140503B2 (en) 2019-07-03 2021-10-05 Qualcomm Incorporated Timer-based access for audio streaming and rendering
US11937065B2 (en) 2019-07-03 2024-03-19 Qualcomm Incorporated Adjustment of parameter settings for extended reality experiences
US11429340B2 (en) 2019-07-03 2022-08-30 Qualcomm Incorporated Audio capture and rendering for extended reality experiences
US11580213B2 (en) 2019-07-03 2023-02-14 Qualcomm Incorporated Password-based authorization for audio rendering
US11354085B2 (en) 2019-07-03 2022-06-07 Qualcomm Incorporated Privacy zoning and authorization for audio rendering
US10972852B2 (en) 2019-07-03 2021-04-06 Qualcomm Incorporated Adapting audio streams for rendering
US11356793B2 (en) 2019-10-01 2022-06-07 Qualcomm Incorporated Controlling rendering of audio data
US11356796B2 (en) 2019-11-22 2022-06-07 Qualcomm Incorporated Priority-based soundfield coding for virtual reality audio
US11317236B2 (en) 2019-11-22 2022-04-26 Qualcomm Incorporated Soundfield adaptation for virtual reality audio
US11089428B2 (en) 2019-12-13 2021-08-10 Qualcomm Incorporated Selecting audio streams based on motion
US11967329B2 (en) 2020-02-20 2024-04-23 Qualcomm Incorporated Signaling for rendering tools
US11750998B2 (en) 2020-09-30 2023-09-05 Qualcomm Incorporated Controlling rendering of audio data
US11743670B2 (en) 2020-12-18 2023-08-29 Qualcomm Incorporated Correlation-based rendering with multiple distributed streams accounting for an occlusion for six degree of freedom applications
US11601776B2 (en) 2020-12-18 2023-03-07 Qualcomm Incorporated Smart hybrid rendering for augmented reality/virtual reality audio
EP4348637A1 (de) * 2021-05-27 2024-04-10 Fraunhofer-Gesellschaft zur Förderung der angewandten Forschung e.V. Audiorichtwirkungscodierung

Family Cites Families (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CA2992097C (en) * 2004-03-01 2018-09-11 Dolby Laboratories Licensing Corporation Reconstructing audio signals with multiple decorrelation techniques and differentially coded parameters
US20090299756A1 (en) * 2004-03-01 2009-12-03 Dolby Laboratories Licensing Corporation Ratio of speech to non-speech audio such as for elderly or hearing-impaired listeners
WO2007104882A1 (fr) * 2006-03-15 2007-09-20 France Telecom Dispositif et procede de codage par analyse en composante principale d'un signal audio multi-canal
US8379868B2 (en) * 2006-05-17 2013-02-19 Creative Technology Ltd Spatial audio coding based on universal spatial cues
US8712061B2 (en) * 2006-05-17 2014-04-29 Creative Technology Ltd Phase-amplitude 3-D stereo encoder and decoder
US20080004729A1 (en) * 2006-06-30 2008-01-03 Nokia Corporation Direct encoding into a directional audio coding format

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
See references of WO2010070225A1 *

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN111653283A (zh) * 2020-06-28 2020-09-11 讯飞智元信息科技有限公司 一种跨场景声纹比对方法、装置、设备及存储介质
CN111653283B (zh) * 2020-06-28 2024-03-01 讯飞智元信息科技有限公司 一种跨场景声纹比对方法、装置、设备及存储介质

Also Published As

Publication number Publication date
ES2733878T3 (es) 2019-12-03
US8964994B2 (en) 2015-02-24
US20110249821A1 (en) 2011-10-13
EP2374123B1 (de) 2019-04-10
WO2010070225A1 (fr) 2010-06-24

Similar Documents

Publication Publication Date Title
EP2374123B1 (de) Verbesserte codierung von mehrkanaligen digitalen audiosignalen
EP2374124B1 (de) Verwaltete codierung von mehrkanaligen digitalen audiosignalen
EP2002424B1 (de) Vorrichtung und verfahren zur skalierbaren kodierung eines mehrkanaligen audiosignals auf der basis einer hauptkomponentenanalyse
EP2539892B1 (de) Mehrkanal-audiostromkomprimierung
EP1600042B1 (de) Verfahren zum bearbeiten komprimierter audiodaten zur räumlichen wiedergabe
KR101854964B1 (ko) 구면 조화 계수들의 변환
EP2005420B1 (de) Einrichtung und verfahren zur codierung durch hauptkomponentenanalyse eines mehrkanaligen audiosignals
EP2304721B1 (de) Raumsynthese mehrkanaliger tonsignale
EP3427260B1 (de) Optimierte codierung und decodierung von verräumlichungsinformationen zur parametrischen codierung und decodierung eines mehrkanaligen audiosignals
EP2691952B1 (de) Zuweisung von bits anhand von subbändern zur quantifizierung von rauminformationsparametern für parametrische codierung
EP2143102B1 (de) Verfahren zur audiokodierung und -dekodierung, audiokodierer, audiodekodierer und zugehörige computerprogramme
EP2168121B1 (de) Quantifizierung nach linearer umwandlung durch kombination von audiosignalen einer klangszene und kodiergerät dafür
EP2656342A1 (de) Verbesserte stereoparametrische kodierung/dekodierung für gegenphasige kanäle
EP3391370A1 (de) Adaptive kanalreduktionsverarbeitung zur codierung eines mehrkanalaudiosignals
WO2011045548A1 (fr) Codage/decodage parametrique bas debit optimise
FR3049084A1 (de)
US9311925B2 (en) Method, apparatus and computer program for processing multi-channel signals
EP4042418B1 (de) Bestimmung von korrekturen zur anwendung auf ein mehrkanalaudiosignal, zugehörige codierung und decodierung
EP4172986A1 (de) Optimierte codierung einer für ein räumliches bild eines mehrkanaligen audiosignals repräsentativen information
WO2023232823A1 (fr) Titre: codage audio spatialisé avec adaptation d'un traitement de décorrélation

Legal Events

Date Code Title Description
PUAI Public reference made under article 153(3) epc to a published international application that has entered the european phase

Free format text: ORIGINAL CODE: 0009012

17P Request for examination filed

Effective date: 20110708

AK Designated contracting states

Kind code of ref document: A1

Designated state(s): AT BE BG CH CY CZ DE DK EE ES FI FR GB GR HR HU IE IS IT LI LT LU LV MC MK MT NL NO PL PT RO SE SI SK SM TR

DAX Request for extension of the european patent (deleted)
RIN1 Information on inventor provided before grant (corrected)

Inventor name: VIRETTE, DAVID

Inventor name: JAILLET, FLORENT

RAP1 Party data changed (applicant data changed or rights of an application transferred)

Owner name: ORANGE

STAA Information on the status of an ep patent application or granted ep patent

Free format text: STATUS: EXAMINATION IS IN PROGRESS

17Q First examination report despatched

Effective date: 20170804

REG Reference to a national code

Ref country code: DE

Ref legal event code: R079

Ref document number: 602009057861

Country of ref document: DE

Free format text: PREVIOUS MAIN CLASS: G10L0019000000

Ipc: G10L0019008000

GRAP Despatch of communication of intention to grant a patent

Free format text: ORIGINAL CODE: EPIDOSNIGR1

STAA Information on the status of an ep patent application or granted ep patent

Free format text: STATUS: GRANT OF PATENT IS INTENDED

RIC1 Information provided on ipc code assigned before grant

Ipc: G10L 19/008 20111012AFI20181029BHEP

INTG Intention to grant announced

Effective date: 20181114

RIN1 Information on inventor provided before grant (corrected)

Inventor name: VIRETTE, DAVID

Inventor name: JAILLET, FLORENT

RIC1 Information provided on ipc code assigned before grant

Ipc: G10L 19/008 20130101AFI20181029BHEP

RIC1 Information provided on ipc code assigned before grant

Ipc: G10L 19/008 20130101AFI20181029BHEP

GRAS Grant fee paid

Free format text: ORIGINAL CODE: EPIDOSNIGR3

GRAA (expected) grant

Free format text: ORIGINAL CODE: 0009210

STAA Information on the status of an ep patent application or granted ep patent

Free format text: STATUS: THE PATENT HAS BEEN GRANTED

AK Designated contracting states

Kind code of ref document: B1

Designated state(s): AT BE BG CH CY CZ DE DK EE ES FI FR GB GR HR HU IE IS IT LI LT LU LV MC MK MT NL NO PL PT RO SE SI SK SM TR

REG Reference to a national code

Ref country code: GB

Ref legal event code: FG4D

Free format text: NOT ENGLISH

REG Reference to a national code

Ref country code: CH

Ref legal event code: EP

Ref country code: AT

Ref legal event code: REF

Ref document number: 1119763

Country of ref document: AT

Kind code of ref document: T

Effective date: 20190415

REG Reference to a national code

Ref country code: DE

Ref legal event code: R096

Ref document number: 602009057861

Country of ref document: DE

REG Reference to a national code

Ref country code: IE

Ref legal event code: FG4D

Free format text: LANGUAGE OF EP DOCUMENT: FRENCH

REG Reference to a national code

Ref country code: NL

Ref legal event code: MP

Effective date: 20190410

REG Reference to a national code

Ref country code: LT

Ref legal event code: MG4D

REG Reference to a national code

Ref country code: AT

Ref legal event code: MK05

Ref document number: 1119763

Country of ref document: AT

Kind code of ref document: T

Effective date: 20190410

PG25 Lapsed in a contracting state [announced via postgrant information from national office to epo]

Ref country code: NL

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20190410

PG25 Lapsed in a contracting state [announced via postgrant information from national office to epo]

Ref country code: SE

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20190410

Ref country code: NO

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20190710

Ref country code: FI

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20190410

Ref country code: PT

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20190910

Ref country code: LT

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20190410

Ref country code: HR

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20190410

PG25 Lapsed in a contracting state [announced via postgrant information from national office to epo]

Ref country code: GR

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20190711

Ref country code: PL

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20190410

Ref country code: LV

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20190410

Ref country code: BG

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20190710

REG Reference to a national code

Ref country code: ES

Ref legal event code: FG2A

Ref document number: 2733878

Country of ref document: ES

Kind code of ref document: T3

Effective date: 20191203

PG25 Lapsed in a contracting state [announced via postgrant information from national office to epo]

Ref country code: AT

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20190410

Ref country code: IS

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20190810

REG Reference to a national code

Ref country code: DE

Ref legal event code: R097

Ref document number: 602009057861

Country of ref document: DE

PG25 Lapsed in a contracting state [announced via postgrant information from national office to epo]

Ref country code: DK

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20190410

Ref country code: EE

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20190410

Ref country code: RO

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20190410

Ref country code: CZ

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20190410

Ref country code: SK

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20190410

PLBE No opposition filed within time limit

Free format text: ORIGINAL CODE: 0009261

STAA Information on the status of an ep patent application or granted ep patent

Free format text: STATUS: NO OPPOSITION FILED WITHIN TIME LIMIT

PG25 Lapsed in a contracting state [announced via postgrant information from national office to epo]

Ref country code: SM

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20190410

26N No opposition filed

Effective date: 20200113

PG25 Lapsed in a contracting state [announced via postgrant information from national office to epo]

Ref country code: TR

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20190410

PG25 Lapsed in a contracting state [announced via postgrant information from national office to epo]

Ref country code: SI

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20190410

REG Reference to a national code

Ref country code: CH

Ref legal event code: PL

REG Reference to a national code

Ref country code: BE

Ref legal event code: MM

Effective date: 20191231

PG25 Lapsed in a contracting state [announced via postgrant information from national office to epo]

Ref country code: MC

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20190410

PG25 Lapsed in a contracting state [announced via postgrant information from national office to epo]

Ref country code: IE

Free format text: LAPSE BECAUSE OF NON-PAYMENT OF DUE FEES

Effective date: 20191211

Ref country code: LU

Free format text: LAPSE BECAUSE OF NON-PAYMENT OF DUE FEES

Effective date: 20191211

PG25 Lapsed in a contracting state [announced via postgrant information from national office to epo]

Ref country code: BE

Free format text: LAPSE BECAUSE OF NON-PAYMENT OF DUE FEES

Effective date: 20191231

Ref country code: CH

Free format text: LAPSE BECAUSE OF NON-PAYMENT OF DUE FEES

Effective date: 20191231

Ref country code: LI

Free format text: LAPSE BECAUSE OF NON-PAYMENT OF DUE FEES

Effective date: 20191231

PG25 Lapsed in a contracting state [announced via postgrant information from national office to epo]

Ref country code: CY

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20190410

PG25 Lapsed in a contracting state [announced via postgrant information from national office to epo]

Ref country code: MT

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20190410

Ref country code: HU

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT; INVALID AB INITIO

Effective date: 20091211

PG25 Lapsed in a contracting state [announced via postgrant information from national office to epo]

Ref country code: MK

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20190410

PGFP Annual fee paid to national office [announced via postgrant information from national office to epo]

Ref country code: GB

Payment date: 20231121

Year of fee payment: 15

PGFP Annual fee paid to national office [announced via postgrant information from national office to epo]

Ref country code: FR

Payment date: 20231122

Year of fee payment: 15

Ref country code: DE

Payment date: 20231121

Year of fee payment: 15

PGFP Annual fee paid to national office [announced via postgrant information from national office to epo]

Ref country code: ES

Payment date: 20240102

Year of fee payment: 15

PGFP Annual fee paid to national office [announced via postgrant information from national office to epo]

Ref country code: IT

Payment date: 20231121

Year of fee payment: 15