EP1563485B1 - Verfahren zur verarbeitung von audiodateien und erfassungsvorrichtung zur anwendung davon - Google Patents
Verfahren zur verarbeitung von audiodateien und erfassungsvorrichtung zur anwendung davon Download PDFInfo
- Publication number
- EP1563485B1 EP1563485B1 EP03782553A EP03782553A EP1563485B1 EP 1563485 B1 EP1563485 B1 EP 1563485B1 EP 03782553 A EP03782553 A EP 03782553A EP 03782553 A EP03782553 A EP 03782553A EP 1563485 B1 EP1563485 B1 EP 1563485B1
- Authority
- EP
- European Patent Office
- Prior art keywords
- distance
- sound
- components
- point
- signals
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Expired - Lifetime
Links
- 238000000034 method Methods 0.000 title claims description 51
- 238000012545 processing Methods 0.000 title claims description 23
- 230000006870 function Effects 0.000 claims abstract description 42
- 238000001914 filtration Methods 0.000 claims abstract description 20
- 239000011159 matrix material Substances 0.000 claims description 37
- 230000008447 perception Effects 0.000 claims description 32
- 230000000694 effects Effects 0.000 claims description 23
- 230000005669 field effect Effects 0.000 claims description 11
- 230000006978 adaptation Effects 0.000 claims description 9
- 230000001419 dependent effect Effects 0.000 claims description 5
- 210000005069 ears Anatomy 0.000 claims description 4
- 230000001902 propagating effect Effects 0.000 claims description 4
- 238000003672 processing method Methods 0.000 abstract description 4
- 230000013707 sensory perception of sound Effects 0.000 abstract description 3
- 238000009877 rendering Methods 0.000 description 41
- 230000015572 biosynthetic process Effects 0.000 description 14
- 238000003786 synthesis reaction Methods 0.000 description 14
- 238000004088 simulation Methods 0.000 description 11
- 230000005540 biological transmission Effects 0.000 description 8
- 239000002775 capsule Substances 0.000 description 8
- 230000005236 sound signal Effects 0.000 description 7
- 239000013598 vector Substances 0.000 description 6
- 230000003321 amplification Effects 0.000 description 5
- 238000003199 nucleic acid amplification method Methods 0.000 description 5
- 238000011282 treatment Methods 0.000 description 5
- 238000013459 approach Methods 0.000 description 4
- 230000008569 process Effects 0.000 description 4
- 238000001228 spectrum Methods 0.000 description 4
- 101100536354 Drosophila melanogaster tant gene Proteins 0.000 description 3
- 241000861223 Issus Species 0.000 description 3
- 230000001934 delay Effects 0.000 description 3
- 230000004044 response Effects 0.000 description 3
- 238000012546 transfer Methods 0.000 description 3
- 229920000297 Rayon Polymers 0.000 description 2
- 238000004364 calculation method Methods 0.000 description 2
- 230000007423 decrease Effects 0.000 description 2
- 238000011835 investigation Methods 0.000 description 2
- 230000004807 localization Effects 0.000 description 2
- 238000002156 mixing Methods 0.000 description 2
- 230000005855 radiation Effects 0.000 description 2
- 239000002964 rayon Substances 0.000 description 2
- 241001415961 Gaviidae Species 0.000 description 1
- 238000012550 audit Methods 0.000 description 1
- 230000002457 bidirectional effect Effects 0.000 description 1
- 238000012512 characterization method Methods 0.000 description 1
- 230000000295 complement effect Effects 0.000 description 1
- 230000006835 compression Effects 0.000 description 1
- 238000007906 compression Methods 0.000 description 1
- 238000012937 correction Methods 0.000 description 1
- 238000010586 diagram Methods 0.000 description 1
- 229940082150 encore Drugs 0.000 description 1
- 230000007613 environmental effect Effects 0.000 description 1
- 230000005284 excitation Effects 0.000 description 1
- 238000000605 extraction Methods 0.000 description 1
- 230000010354 integration Effects 0.000 description 1
- 238000007726 management method Methods 0.000 description 1
- 239000002245 particle Substances 0.000 description 1
- 230000010349 pulsation Effects 0.000 description 1
- 238000005070 sampling Methods 0.000 description 1
- 239000007787 solid Substances 0.000 description 1
- 230000002194 synthesizing effect Effects 0.000 description 1
- 230000009466 transformation Effects 0.000 description 1
- 238000000844 transformation Methods 0.000 description 1
- 238000011144 upstream manufacturing Methods 0.000 description 1
- 238000012800 visualization Methods 0.000 description 1
Images
Classifications
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10H—ELECTROPHONIC MUSICAL INSTRUMENTS; INSTRUMENTS IN WHICH THE TONES ARE GENERATED BY ELECTROMECHANICAL MEANS OR ELECTRONIC GENERATORS, OR IN WHICH THE TONES ARE SYNTHESISED FROM A DATA STORE
- G10H1/00—Details of electrophonic musical instruments
- G10H1/0091—Means for obtaining special acoustic effects
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10H—ELECTROPHONIC MUSICAL INSTRUMENTS; INSTRUMENTS IN WHICH THE TONES ARE GENERATED BY ELECTROMECHANICAL MEANS OR ELECTRONIC GENERATORS, OR IN WHICH THE TONES ARE SYNTHESISED FROM A DATA STORE
- G10H1/00—Details of electrophonic musical instruments
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04S—STEREOPHONIC SYSTEMS
- H04S2400/00—Details of stereophonic systems covered by H04S but not provided for in its groups
- H04S2400/15—Aspects of sound capture and related signal processing for recording or reproduction
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04S—STEREOPHONIC SYSTEMS
- H04S2420/00—Techniques used stereophonic systems covered by H04S but not provided for in its groups
- H04S2420/11—Application of ambisonics in stereophonic audio systems
Definitions
- the present invention relates to the processing of sound data.
- Techniques relating to the propagation of a sound wave in the three-dimensional space implement methods for processing the audio signal applied to the simulation of acoustic and psychoacoustic phenomena. .
- Such processing methods provide spatial encoding of the acoustic field, its transmission and spatial reproduction on a set of loudspeakers or on headphones of a stereo headset.
- a first category of treatments relates to processes for synthesizing the room effect, or more generally for environmental effects. From a description of one or more sound sources (transmitted signal, position, orientation, directivity, or other) and based on a room effect model (involving a room geometry, or an acoustic perception desired), a set of elementary acoustic phenomena (direct waves, reflected or diffracted), or a macroscopic acoustic phenomenon (reverberated and diffuse field), allowing to translate the spatial effect at the level of a listener located at a selected point of auditory perception, in the three-dimensional space.
- sound sources transmitted signal, position, orientation, directivity, or other
- a room effect model involving a room geometry, or an acoustic perception desired
- a set of elementary acoustic phenomena direct waves, reflected or diffracted
- a macroscopic acoustic phenomenon reverberated and diffuse field
- a set of signals typically associated with reflections ( "secondary" sources , active by re-emission of a received main wave, having a spatial position attribute) and / or associated with a late reverberation (decorrelated signals for a field) are then calculated. diffuse).
- a second category of processes relates to the positional or directional rendering of sound sources. These methods are applied to signals determined by a method of the first category described above (involving primary and secondary sources) depending on the spatial description (source position) associated with them.
- these methods according to this second category make it possible to obtain signals to be broadcast on loudspeakers or headphones, in order to finally give to a listener the auditory impression of sound sources placed at predetermined respective positions around the auditor.
- Processes according to this second category are called "creators of three-dimensional sound images", because of the distribution in the three-dimensional space of the feeling of the position of the sources by a listener.
- Methods according to the second category generally comprise a first stage of spatial encoding of the elementary acoustic events which produces a representation of the sound field in the three-dimensional space.
- this representation is transmitted or stored for a deferred use.
- the decoded signals are delivered on loudspeakers or headphones of a playback device.
- the present invention is rather in the second category mentioned above. It concerns in particular the spatial encoding of sound sources and a specification of the three-dimensional sound representation of these sources. It applies as well to an encoding of " virtual " sound sources (applications where sound sources are simulated such as games, a spatialized conference, or others), an "acoustic" encoding of a natural sound field, during sound recording by one or more three-dimensional microphone networks.
- virtual sound sources
- acoustic encoding of a natural sound field
- a similar acoustic encoding method is presented byJ. Chen et al: "Synthesis of 3D virtual auditory space via a spatial feature extraction and regularization model", Proceedings of the virtual reality annual international symposium, Seattle, Sept. 18-22, 1993, IEEE, New York, US, pages 188 -193.
- the ambisonic encoding which will be described in detail below, consists of representing signals relating to one or more sound waves in a base of spherical harmonics (in spherical coordinates involving in particular an elevation angle and an azimuthal angle, characterizing a direction of the sound or sounds).
- the components representing these signals and expressed in this basis of spherical harmonics are also a function, for the waves emitted in the near field, of a distance between the sound source emitting this field and a point corresponding to the origin of the harmonic base. spherical. More particularly, this dependence of distance is expressed as a function of the sound frequency, as will be seen below.
- this document presents a horizontal network of sensors, which assumes that the acoustic phenomena considered here propagate only in horizontal directions, which excludes any other direction of propagation and which, therefore, does not represent the physical reality of an ordinary acoustic field.
- An object of the present invention is to provide a method for processing, by encoding, transmission and reproduction, any type of sound field, in particular the effect of a sound source in the near field.
- Another object of the present invention is to provide a method for encoding virtual sources, not only in direction, but also in distance, and to define a decoding adaptable to any rendering device.
- Another object of the present invention is to provide a robust processing method for sounds of all sound frequencies (including low frequencies), especially for sound recording of natural acoustic fields using three-dimensional microphone networks.
- the data encoded and filtered in steps a) and b) are transmitted to the rendering device with a parameter representative of said second distance.
- the rendering device comprising means for reading a memory medium
- the encoded and filtered data is stored on a memory medium intended to be read by the rendering device. in steps a) and b) with a parameter representative of said second distance.
- an adaptation filter is applied to the coded and filtered data whose coefficients are a function of said second and third distances.
- the coefficients of a digital audio filter are defined from the root digital values of said power polynomials m.
- the aforementioned polynomials are Bessel polynomials.
- a microphone comprising an acoustic transducer array arranged substantially on the surface of a sphere whose center corresponds substantially to said reference point, to obtain said signals representative of at least one sound. propagating in three-dimensional space.
- a global filter is applied to step b) in order firstly to compensate for a near-field effect as a function of said second distance and secondly to equalize the signals from the transducers to compensate for a difference. directivity weighting of said transducers.
- a number of transducers function of a selected total number of components to represent the sound in said base of spherical harmonics.
- a total number of components in the base of spherical harmonics to obtain, at the restitution, a region of the space around the point of perception in which the sound reproduction is faithful and whose dimensions are increasing with the total number of components.
- a playback device having a number of speakers at least equal to said total number of components.
- the filtering performed by the processing unit consists, on the one hand, of equalizing, as a function of the radius of the sphere, the signals coming from the transducers to compensate for a directivity weighting of said transducers and, on the other hand, to compensating a near-field effect according to said reference distance.
- FIG. 1 represents by way of illustration a global system of sound spatialization.
- a simulation module of a virtual scene defines a sound object as a virtual source of a signal, for example monophonic, of position chosen in the three-dimensional space and which defines a direction of sound.
- specifications of the geometry of a virtual room can be provided to simulate reverberation of the sound.
- a processing module 11 applies a management of one or more of these sources with respect to a listener (definition of a virtual position of the sources with respect to this listener). It implements a room effect processor for simulating reverberations or the like by applying delays and / or routine filtering.
- the signals thus constructed are transmitted to a spatial encoding module 2a of the elementary contributions of the sources.
- a natural sound recording can be performed as part of a sound recording by one or more microphones arranged in a chosen manner with respect to real sources (module 1b).
- the signals picked up by the microphones are encoded by a module 2b.
- the acquired and encoded signals can be transformed according to an intermediate representation format (module 3b), before being mixed by the module 3 to the signals generated by the module 1a and encoded by the module 2a (from the virtual sources).
- the mixed signals are then transmitted or stored on a medium, for later retrieval (arrow TR). They are then applied to a decoding module 5, for the purpose of rendering on a reproduction device 6 comprising loudspeakers.
- the decoding step 5 may be preceded by a step of manipulation of the sound field, for example by rotation, by means of a processing module 4 provided upstream of the decoding module 5.
- the reproduction device may be in the form of a multiplicity of loudspeakers, arranged for example on the surface of a sphere in a three-dimensional configuration (periphery) to ensure, in the restitution, in particular a sense of direction sound in three-dimensional space.
- an auditor place generally in the center of the sphere formed by the network of speakers, this center corresponding to the auditory perception point cited above.
- the speakers of the playback device can be arranged in a plane (two-dimensional panoramic configuration), the speakers being arranged in particular on a circle and the listener usually placed in the center of this circle.
- the rendering device may be in the form of a "surround" type device (5.1).
- the rendering device can be in the form of a headset with two earphones for a binaural synthesis of the sound reproduced, which allows the listener to feel a direction of the sources in the three-dimensional space, as will be discussed in more detail below.
- a two-speaker reproduction device for a feeling in the three-dimensional space, can also be in the form of a transaural restitution device, with two loudspeakers arranged at a selected distance from a listener.
- a signal from a source 1 to N is transmitted to a spatial encoding module 2, as well as its position (real or virtual). Its position can be as well defined in terms of incidence (direction of the source as seen by the listener) and in terms of distance between this source and a listener.
- the plurality of signals thus encoded allows to obtain a multi-channel representation of a global sound field.
- the encoded signals are transmitted (arrow TR) to a sound reproduction device 6, for a sound reproduction in the three-dimensional space, as indicated above with reference to FIG.
- the set of weighting factors B mn ⁇ which are implicitly a function of the frequency, thus describe the pressure field in the zone considered. For this reason, these factors are called "spherical harmonic components" and represent a frequency expression of sound (or pressure field) in the spherical harmonics basis Y mn ⁇ .
- the spherical harmonics form an orthonormal basis where the scalar products between harmonic components and, generally between two functions F and G, are respectively defined by: ⁇ Y m not ⁇
- Y m ' not ' ⁇ ' > 4 ⁇ ⁇ mm ' ⁇ nn ' ⁇ ⁇ ' .
- BOY WUT 4 ⁇ 1 4 ⁇ ⁇ F ( ⁇ , ⁇ ) BOY WUT ( ⁇ , ⁇ ) , d ⁇ ( ⁇ , ⁇ )
- Spherical harmonics are bounded real functions, as shown in Figure 4, as a function of the order m and the indices n and ⁇ .
- the dark and light parts correspond respectively to the positive and negative values of spherical harmonic functions.
- the radial functions j m (kr) are spherical Bessel functions, the module of which is illustrated for some values of the order m in FIG.
- ambisonic representation can be given by a base of spherical harmonics as follows.
- the ambisonic components of the same order m finally express “derivatives” or “moments” of order m of the pressure field in the vicinity of the origin O (center of the sphere shown in FIG. 3).
- B 11 + 1 X
- B 11 - 1 Y
- an ambisonic system takes into account a subset of spherical harmonic components, as described above.
- a system of order M when this one takes into account ambisonic components of subscript m ⁇ M.
- the rendering device comprises loudspeakers disposed on the surface of a sphere (" periphery "), it is possible in principle to use as many harmonics as there are loudspeakers.
- the reference S designates the pressure signal carried by a plane wave and picked up at the point O corresponding to the center of the sphere of FIG. 3 (origin of the base in spherical coordinates).
- the incidence of the wave is described by the azimuth ⁇ and the elevation ⁇ .
- a filter is applied F m ( ⁇ / vs ) to "bend" the shape of the wave fronts, considering that a near field emits, as a first approximation, a spherical wave.
- this additional filter is of the "integrator" type, with an amplifying effect increasing and diverging (unbounded) as the sound frequencies decrease towards zero.
- a pre-compensation of the near field is introduced at the very stage of the encoding, this compensation involving filters of the analytical form.
- amplification F m ( ⁇ / vs ) ( ⁇ ) whose effect appears in FIG. 6 is compensated by attenuation of the filter applied as soon as encoding 1 F m ( R / vs ) ( ⁇ ) .
- the coefficients of this compensation filter 1 F m ( R / vs ) ( ⁇ ) are increasing with the frequency of the sound and, in particular, tend towards zero, for the low frequencies.
- this pre-compensation, performed as soon as encoding ensures that the data transmitted are not divergent for low frequencies.
- a pre-compensation is applied to the encoding, involving a filter of the type 1 F m ( R / vs ) ( ⁇ ) as indicated above, which allows, on the one hand, to transmit bounded signals, and, on the other hand, to choose the distance R, from the encoding, for the restitution of the sound from the loudspeakers HP i , as shown in FIG. 7.
- a virtual source placed at the distance p of the origin O was simulated at the time of acquisition (FIG.
- the pre-compensation of the near field of the loudspeakers (placed at the distance R), at the stage of the encoding, can be combined with a simulated near-field effect of a virtual source placed at a distance p.
- a total filter ultimately comes into play resulting, on the one hand, from the simulation of the near field, and, on the other hand, from the compensation of the near field, the coefficients of this filter being able to express itself.
- H m NFC ( ⁇ / vs , R / vs ) ( ⁇ ) F m ( ⁇ / vs ) ( ⁇ ) F m ( R / vs ) ( ⁇ )
- the total filter given by the relation [A11] is stable and constitutes the "distance encoding" part in the spatial ambisonic encoding according to the invention, as represented in FIG. 8.
- the coefficients of these filters correspond to the functions of FIG. monotonic transfer of the frequency, which tend towards the value 1 in high frequencies and towards the value (R / ⁇ ) m in low frequencies.
- the distance R between an auditory perception point and the speakers HP i is actually of the order of one or a few meters.
- steps a) and b) above can be brought together in one and the same global step, or even be interchanged (with distance encoding and compensation filtering, followed by direction encoding).
- the method according to the invention is therefore not limited to a successive implementation over time of steps a) and b).
- FIG. 11B shows the propagation of the initial sound wave from a near-field source situated at a distance p from a point in the acquisition space that corresponds, in the restitution space at point P of Figure 7 of auditory perception. Note in FIG. 11A that the listeners (symbolized by schematized heads) can locate the virtual source in the same geographical location located at the distance p from the perception point P in FIG. 11B.
- H m NFC ( ⁇ / vs , R / vs ) ( ⁇ ) F m ( ⁇ / vs ) ( ⁇ ) F m ( R / vs ) ( ⁇ )
- Table 1 values ⁇ i> R ⁇ / i> ⁇ sub> ⁇ i> e ⁇ / i> ⁇ /sub>[ ⁇ i> X ⁇ / i> ⁇ sub> ⁇ i> m, q ⁇ / i> ⁇ / sub>],
- the digital filters are thus implemented from the values of Table 1, by providing cascades of cells of order 2 (for m even), and an additional cell (for odd m), from the relationships [A14] given here. -before.
- Digital filters are thus produced in an infinite impulse response form, which is easily parameterizable as shown above. It should be noted that an implementation in a finite impulse response form can be envisaged and consists in calculating the complex spectrum of the transfer function from the analytic formula, then deduce a finite impulse response by inverse Fourier transform. A convolution operation is then applied for filtering.
- R is a reference distance with which a compensated near-field effect is associated and c is the speed of sound (typically 340 m / s in air).
- This modified ambisonic representation has the same scalability properties (schematically represented by transmitted data "surrounded" near the arrow TR of FIG. 1) and obeys the same transformations of rotation of the field (module 4 of FIG. usual ambisonic.
- the decoding operation is adaptable to any rendering device, of radius R 2 , different from the reference distance R above.
- filters of the type H m NFC ( ⁇ / vs , R / vs ) ( ⁇ ) as described above, but with distance parameters R and R 2 , instead of p and R.
- R / c is to be memorized (and / or transmitted) between encoding and decoding.
- the filtering module represented therein is provided, for example, in a processing unit of a rendering device.
- Ambisonic components received were pre-compensated for encoding for a reference distance R 1 as a second distance.
- the rendering device comprises a plurality of loudspeakers arranged at a third distance R 2 from an auditory perception point P, this third distance R 2 being different from the second aforementioned distance R 1 .
- the filtering module of FIG. 12, in the form H m NFC ( R 1 / vs , R two / vs ) ( ⁇ ) then adapts, upon reception of the data, the pre-compensation at the distance R 1 for a reproduction at the distance R 2 .
- the rendering device also receives the parameter R 1 / c.
- the invention also makes it possible to mix several ambisonic representations of sound fields (real and / or virtual sources), whose reference distances R are different (where appropriate with infinite reference distances and corresponding to distant sources).
- a pre-compensation of all these sources will be filtered at a smallest reference distance, before mixing the signals ambisic, which allows the restitution to obtain a correct definition of the sound relief.
- the distance encoding with near-field pre-compensation is advantageously applied in combination with the focus processing.
- the wave transmitted by each speaker is defined by a processing prior to "re-encoding" the ambisonic field in the center of the rendering device, as follows.
- the wave emitted by a loudspeaker of index i and incidence ( ⁇ i and ⁇ i ) is powered by a signal Si.
- This loudspeaker participates in the reconstruction of the component B mn ' , by his contribution S i .
- Y mn ⁇ ( ⁇ i , ⁇ i ) the wave emitted by a loudspeaker of index i and incidence ( ⁇ i and ⁇ i ) is powered by a signal Si.
- vs i [ Y 00 + 1 ( ⁇ i , ⁇ i ) Y 11 + 1 ( ⁇ i , ⁇ i ) Y 11 - 1 ( ⁇ i , ⁇ i ) ⁇ Y m not ⁇ ( ⁇ i , ⁇ i ) ⁇ ]
- the relation [B4] thus defines a re-encoding operation, prior to the restitution.
- decoding verifying different criteria by frequency bands is possible, which makes it possible to offer an optimized reproduction according to the listening conditions, in particular with regard to the positioning constraint at the center. O of the sphere of Figure 3, during the restitution.
- the mastering operation is preceded by a filtering operation that compensates for the near field on each component.
- B mn ⁇ and which can be implemented in digital form, as described above, with reference to relation [A14].
- the matrix C of "re-encoding" is specific to the rendering device. Its coefficients can be determined initially by parameterization and sound characterization of the restitution device reacting to a predetermined excitation.
- a listener having a two-headset headset of a binaural synthesis device is shown.
- the two ears of the listener are arranged at respective points O L (left ear) and O R (right ear) of the space.
- the center of the listener's head is located at point O and the radius of the listener's head is of value a.
- a sound source must be audibly perceived at a point M in the space, at a distance r from the center of the listener's head (and respectively at distances r R from the right ear and r L from the ear left).
- the direction of the source at the point M is defined by the vectors r ⁇ , r ⁇ R and r ⁇ The .
- binaural synthesis is defined as follows.
- Each listener has an ear shape of its own.
- the perception of a sound in the space by this listener is done by learning, from birth, according to the form of the ears (in particular the shape of the pavilions and the dimensions of the head) peculiar to this listener.
- the perception of sound in space is manifested inter alia by the fact that the sound reaches one ear, before the other ear, which results in a delay ⁇ between the signals to be emitted by each earphone of the device. restitution applying binaural synthesis.
- the playback device is initially set, for the same listener, by scanning a sound source around his head, at the same distance R from the center of his head. It will be understood that this distance R can be considered as a distance between a "restitution point" as stated above and a point of auditory perception (here the center O of the listener's head).
- the index L is associated with the signal to be restored by the earpiece attached to the left ear and the index R is associated with the signal to be restored by the earpiece attached to the right ear.
- a delay for each channel for producing a signal for a separate earphone is applied to the initial signal S.
- These delays ⁇ L and ⁇ R are a function of a maximum delay ⁇ MAX which corresponds here to the ratio a / c where a, as indicated previously, corresponds to the radius of the listener's head and c to the speed of sound.
- these delays are defined as a function of the difference in distance from the point O (center of the head) to the point M (position of the source whose sound is to be restored, in FIG. 13A) and of each ear at this point. M.
- respective gains g L and g R are also applied to each channel, which are a function of a ratio of the distances from the point O to the point M and from each ear to the point M.
- the respective modules applied to each channel 2 L and 2 R encode the signals of each channel, in an ambisonic representation, with near field pre-compensation NFC (for "Near Field Compensation") in the sense of the present invention.
- the signals coming from the source M are transmitted to the reproduction device comprising ambisonic decoding modules, for each channel, 5 L and 5 R.
- the reproduction device comprising ambisonic decoding modules, for each channel, 5 L and 5 R.
- an ambisonic encoding / decoding, with near-field compensation is applied for each channel (left listener, right listener) in the binaural synthesis restitution (here of type "B-FORMAT"), in split form.
- the near-field compensation is effected, for each channel, with the first distance p a distance r L and r R between each ear and the position M of the sound source to be restored.
- a microphone 141 comprises a plurality of transducer capsules capable of picking up acoustic pressures and reproducing electrical signals S l , ..., S N.
- Caps CAP i are arranged on a sphere of predetermined radius r (here, a rigid sphere, such as a ping-pong ball for example). The capsules are spaced with a regular pitch on the sphere. In practice, the number N of capsules is chosen according to the desired order M for the ambisonic representation.
- the pre-compensation of the near field can be applied not only for the virtual source simulation, as indicated above, but also to the acquisition and, more generally, by combining the pre-compensation of field close to all types of treatments involving an ambisonic representation.
- EQ m is an equalizer filter that compensates a weighting W m which is related to the directivity of the capsules and which further includes the diffraction by the rigid sphere.
- this equalization filter is not stable and we obtain an infinite gain at very low frequencies.
- the spherical harmonic components themselves, are not of finite amplitude when the sound field is not limited to propagation of plane waves, that is to say from distant sources, as we saw earlier.
- the signals S 1 to S N are recovered from the microphone 141. If necessary, a pre-equalization of these signals is applied by a processing module 142.
- the module 143 makes it possible to express these signals in the ambisonic context, under matrix form.
- the module 144 applies the filter of the relation [C7] to the components ambisonic expressed as a function of the radius r of the sphere of the microphone 141.
- the near-field compensation is performed for a reference distance R as a second distance.
- the signals encoded and thus filtered by the module 144 can be transmitted, if necessary, with the parameter representative of the reference distance R / c.
- near-field compensation within the meaning of the present invention can be applied to all types of processing involving an ambisonic representation.
- This near-field compensation makes it possible to apply the ambisonic representation to a multiplicity of sound contexts where the direction of a source and advantageously its distance must be taken into account.
- the possibility of the representation of sound phenomena of all types (near or far fields) in the ambisonic context is ensured by this pre-compensation, because of the limitation to finite real values of the ambison components.
- the near-field pre-compensation can be integrated, at the encoding, as much for a near source as for a distant source.
- the distance p expressed above will be considered infinite, without substantially modifying the expression of the filters H m given above.
- processing using room effect processors that typically provide decorrelated signals that can be used to model the late diffuse field (late reverberation) can be combined with near field pre-compensation.
- the various spherical harmonic components (with a chosen order M) can then be constructed by applying a gain correction for each ambisonic component and a field compensation close to the loudspeakers (with a reference distance R separating the loudspeakers the point of auditory perception as shown in Figure 7).
- the encoding principle in the sense of the present invention is generalizable to radiation models other than monopolar sources (real or virtual) and / or speakers.
- any form of radiation can be expressed by integration of a continuous distribution of point elementary sources.
- a decoding method has been described above in which a matrix system involving the ambison components is applied.
- it may be provided a generalized processing by fast Fourier transforms (circular or spherical) to limit the computing time and computing resources (in terms of memory) necessary for the decoding process.
- the pre-compensation encoding method may be coupled to a digital audio compression for quantizing and adjusting the gain for each frequency subband.
- the present invention applies to all types of sound spatialization systems, especially for "virtual reality” type applications (navigation in virtual scenes in three-dimensional space, cat-type conversations on the Internet), interface sonification, audio editing software for recording, mixing and restoring music, but also for acquiring, from use of three-dimensional microphones, for taking musical or cinematic sound, or for the transmission of sound environment on the Internet, for example for "Webcam” sound.
Landscapes
- Physics & Mathematics (AREA)
- Engineering & Computer Science (AREA)
- Acoustics & Sound (AREA)
- Multimedia (AREA)
- Stereophonic System (AREA)
- Measurement Of Mechanical Vibrations Or Ultrasonic Waves (AREA)
- Measurement Of Velocity Or Position Using Acoustic Or Ultrasonic Waves (AREA)
- Information Retrieval, Db Structures And Fs Structures Therefor (AREA)
Claims (22)
- Verfahren zur Verarbeitung von Audiodateien, bei dem:a) man Signale kodiert, die mindestens einen Ton darstellen, der sich im dreidimensionalen Raum ausbreitet und von einer Quelle kommt, die in einem ersten Abstand (p) von einem Bezugspunkt (O) gelegen ist, um eine Darstellung des Tons durch in einer Basis von sphärischen Harmonischen ausgedrückte Komponenten (Bmn σ) mit einem diesem Bezugspunkt (O) entsprechenden Ursprung zu erhalten,b) und man an diese Komponenten (Bmn σ) eine Kompensierung eines Nahfeldeffekts durch eine Filterung anlegt, die eine Funktion von einem zweiten Abstand (R) ist, der bei einer Wiedergabe des Tons durch eine Wiedergabevorrichtung im Wesentlichen einen Abstand zwischen einem Wiedergabepunkt (Hpi) und einem Hörwahrnehmungspunkt (P) definiert.
- Verfahren nach Anspruch 1, bei dem, wenn diese Quelle vom Bezugspunkt (O) entfernt ist,- man Komponenten von aufeinanderfolgenden Ordnungen m bei der Darstellung des Tons in dieser Basis von sphärischen Harmonischen erhält und- man ein Filter (1/Fm) anlegt, dessen jeweils an eine Komponente der Ordnung m angelegte Koeffizienten sich analytisch in der Form des Kehrwerts eines Polynoms der Potenz m ausdrückt, dessen Variable umgekehrt proportional zur Tonfrequenz und zum zweiten Abstand (R) ist, um einen Nahfeldeffekt auf Höhe der Wiedergabevorrichtung zu kompensieren.
- Verfahren nach Anspruch 1, bei dem, wenn diese Quelle eine in diesem ersten Abstand (ρ) vorgesehene virtuelle Quelle ist,- man Komponenten von aufeinanderfolgenden Ordnungen m bei der Darstellung des Tons in dieser Basis von sphärischen Harmonischen erhält und- man ein Globalfilter (Hm) anlegt, dessen jeweils an eine Komponente der Ordnung m angelegte Koeffizienten sich analytisch in der Form eines Bruchs ausdrücken, dessen- Zähler ein Polynom der Potenz m ist, dessen Variable umgekehrt proportional zur Tonfrequenz und zum ersten Abstand (p) ist, um einen Nahfeldeffekt der virtuellen Quelle zu simulieren, und- dessen Nenner ein Polynom der Potenz m ist, dessen Variable umgekehrt proportional zu der Tonfrequenz und zu dem zweiten Abstand (R) ist, um den Nahfeldeffekt der virtuellen Quelle in den niedrigen Tonfrequenzen zu kompensieren.
- Verfahren nach einem der vorhergehenden Ansprüche, bei dem man die in den Schritten a) und b) kodierten und gefilterten Daten mit einem diesen zweiten Abstand darstellenden Parameter (R/c) zur Wiedergabevorrichtung überträgt.
- Verfahren nach einem der Ansprüche 1 bis 3, bei dem man, wenn die Wiedergabevorrichtung Mittel zum Lesen eines Speicherträgers umfasst, auf einem Speicherträger, der dazu bestimmt ist, von der Wiedergabevorrichtung gelesen zu werden, die in den Schritten a) und b) kodierten und gefilterten Daten mit einem diesen zweiten Abstand darstellenden Parameter (R/c) speichert.
- Verfahren nach einem der Ansprüche 4 und 5, bei dem man vor einer Tonwiedergabe durch eine Wiedergabevorrichtung, die eine Vielzahl von Lautsprechern umfasst, die in einem dritten Abstand (R2) von diesem Hörwahrnehmungspunkt (P) angeordnet sind, an die kodierten und gefilterten Daten ein Anpassungsfilter (Hm (R1/c,R2/c)) anlegt, dessen Koeffizienten eine Funktion von dem zweiten Abstand (R1) und dem dritten Abstand (R2) sind.
- Verfahren nach Anspruch 6, bei dem die jeweils an eine Komponente der Ordnung m angelegten Koeffizienten des Anpassungsfilters (Hm (R1/c,R2/c)) sich analytisch in der Form eines Bruchs ausdrücken, dessen- Zähler ein Polynom der Potenz m ist, dessen Variable umgekehrt proportional zur Tonfrequenz und zum zweiten Abstand (R) ist, und- dessen Nenner ein Polynom der Potenz m ist, dessen Variable umgekehrt proportional zu der Tonfrequenz und zu dem dritten Abstand (R2) ist.
- Verfahren nach einem der Ansprüche 2, 3 und 7, bei dem man für die Durchführung des Schritts b)- für Komponenten geradzahliger Ordnung m audionumerische Filter in der Form einer Kaskade von Zellen der Ordnung zwei und- für Komponenten ungeradzahliger Ordnung m audionumerische Filter in der Form einer Kaskade von Zellen der Ordnung zwei und eine zusätzliche Zelle der Ordnung eins vorsieht.
- Verfahren nach Anspruch 8, bei dem die Koeffizienten eines audionumerischen Filters bei einer Komponente der Ordnung m ausgehend von den numerischen Werten der Wurzeln dieser Polynome der Potenz m definiert sind.
- Verfahren nach einem der Ansprüche 2, 3, 7, 8 und 9, bei dem diese Polynome Bessel-Polynome sind.
- Verfahren nach einem der Ansprüche 1, 2 und 4 bis 10, bei dem man ein Mikrophon vorsieht, das ein Netz von akustischen Wandlern aufweist, die im Wesentlichen auf der Oberfläche einer Kugel angeordnet sind, deren Mittelpunkt im Wesentlichen dem Bezugspunkt (O) entspricht, um die Signale zu erhalten, die mindestens einen sich im dreidimensionalen Raum ausbreitenden Ton darstellen.
- Verfahren nach Anspruch 11, bei dem man im Schritt b) ein Globalfilter anlegt, um einerseits einen Nahfeldeffekt in Abhängigkeit von diesem zweiten Abstand (R) zu kompensieren und andererseits die von den Wandlern kommenden Signale zu egalisieren, um eine Richtwirkungsgewichtung der Wandler zu kompensieren.
- Verfahren nach einem der Ansprüche 11 und 12, bei dem man eine Anzahl von Wandlern vorsieht, die von einer gewählten Gesamtzahl von Komponenten abhängt, um den Ton in der Basis der sphärischen Harmonischen darzustellen.
- Verfahren nach einem der vorhergehenden Ansprüche, bei dem man im Schritt a) eine Gesamtzahl von Komponenten in der Basis der sphärischen Harmonischen wählt, um bei der Wiedergabe einen Bereich des Raums um den Wahrnehmungspunkt (P) herum zu erhalten, in dem die Wiedergabe des Tons getreu ist und dessen Abmessungen mit der Gesamtzahl von Komponenten zunehmen.
- Verfahren nach Anspruch 14, bei dem man eine Wiedergabevorrichtung vorsieht, die eine Anzahl von Lautsprechern von mindestens gleich der Gesamtzahl der Komponenten vorsieht.
- Verfahren nach einem der Ansprüche 1 bis 5 und 8 bis 13, bei dem- man eine Wiedergabevorrichtung vorsieht, die mindestens einen ersten und einen zweiten Lautsprecher vorsieht, die in einem gewählten Abstand von einem Hörer angeordnet sind,- für diesen Hörer eine Information der Empfindung der Lage von Tonquellen, die in einem vorbestimmten Bezugsabstand (R) vom Hörer gelegen sind, im Raum erhält und- die Kompensierung des Schritts b) mit dem Bezugsabstand im Wesentlichen als zweiten Abstand anlegt.
- Verfahren nach einem der Ansprüche 1 bis 3 und 8 bis 13 in Kombination mit einem der Ansprüche 4 und 5, bei dem- man eine Wiedergabevorrichtung vorsieht, die mindestens einen ersten und einen zweiten Lautsprecher umfasst, die in einem gewählten Abstand von einem Hörer angeordnet sind,- man für diesen Hörer eine Information der Empfindung der Lage von Tonquellen, die in einem vorbestimmten Bezugsabstand (R2) vom Hörer gelegen sind, im Raum erhält und- man vor einer Tonwiedergabe durch die Wiedergabevorrichtung an die in den Schritten a) und b) kodierten und gefilterten Daten ein Anpassungsfilter (Hm (R/c,R2/c) anlegt, dessen Koeffizienten eine Funktion des zweiten Abstands (R) und im Wesentlichen des Bezugsabstands (R2) sind.
- Verfahren nach einem der Ansprüche 16 und 17, bei dem- die Wiedergabevorrichtung einen Kopfhörer mit zwei Hörern für die Ohren des Hörers umfasst und- man getrennt für jeden Hörer die Kodierung und Filterung der Schritte a) und b) für Signale, die jeweils zur Versorgung jedes Hörers bestimmt sind, anlegt, mit als erstem Abstand (p) jeweils einen Abstand (rR, rL) der jedes Ohr von einem Standort (M) einer wiederzugebenden Quelle trennt.
- Verfahren nach einem der vorhergehenden Ansprüche, bei dem man in den Schritten a) und b) ein Matrixsystem in Form bringt, das mindestens umfasst:- eine Matrix (B), die die Komponenten in der Basis der sphärischen Harmonischen umfasst, und- eine diagonale Matrix (Diag(1/Fm)), deren Koeffizienten Filterkoeffizienten des Schritts b) entsprechen, und man die Matrizes multiplizierten, um eine resultierende Matrix von kompensierten Komponenten (B̃) zu erhalten.
- Verfahren nach Anspruch 19, bei dem- die Wiedergabevorrichtung eine Vielzahl von Lautsprechern umfasst, die in einem gleichen Abstand (R) vom dem Hörwahrnehmungspunkt (P) angeordnet sind, und,- um diese in den Schritten a) und b) kodierten und gefilterten Daten zu dekodieren und Signale zu formen, die dafür ausgelegt sind, die Lautsprecher zu speisen,* man ein Matrixsystem bildet, das die resultierende Matrix (B̃) und eine vorbestimmte, der Wiedergabevorrichtung zugeordnete Dekodierungsmatrix (D) umfasst, und* man eine Matrix (S), die die Signale zur Versorgung der Lautsprecher darstellende Koeffizienten aufweist, durch Multiplikation der Matrix (B̃) der kompensierten Komponenten mit der Dekodierungsmatrix (D) erhält.
- Tonerfassungsvorrichtung, umfassend ein Mikrophon, das mit einem Netz von akustischen Wandlern versehen ist, die im Wesentlichen auf der Oberfläche einer Kugel angeordnet sind, dadurch gekennzeichnet, dass sie außerdem eine Verarbeitungseinheit aufweist, die dafür ausgelegt ist,- jeweils von einem Wandler ausgehende Signale zu empfangen,- an diese Signale eine Kodierung anzulegen, um eine Darstellung des Tons durch Komponenten (Bmn σ), die in einer Basis von sphärischen Harmonischen ausgedrückt sind, mit einem dem Mittelpunkt dieser Kugel (O) entsprechenden Ursprung zu erhalten,- und an diese Komponenten (Bmn σ) eine Filterung anzulegen, die eine Funktion einerseits von einem dem Radius der Kugel (r) entsprechenden Abstand und andererseits von einem Bezugsabstand (R) ist.
- Vorrichtung nach Anspruch 21, dadurch gekennzeichnet, dass die Filterung darin besteht, dass einerseits in Abhängigkeit von dem Radius der Kugel die von den Wandlern kommenden Signale egalisiert werden, um eine Richtwirkungsgewichtung der Wandler zu kompensieren, und andererseits ein Nahfeldeffekt in Abhängigkeit von einem gewählten Bezugsabstand (R) kompensiert wird, der bei einer Wiedergabe des Tons im Wesentlichen einen Abstand zwischen einem Wiedergabepunkt (HPi) und einem Hörwahrnehmungspunkt (P) definiert.
Applications Claiming Priority (3)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
FR0214444A FR2847376B1 (fr) | 2002-11-19 | 2002-11-19 | Procede de traitement de donnees sonores et dispositif d'acquisition sonore mettant en oeuvre ce procede |
FR0214444 | 2002-11-19 | ||
PCT/FR2003/003367 WO2004049299A1 (fr) | 2002-11-19 | 2003-11-13 | Procede de traitement de donnees sonores et dispositif d'acquisition sonore mettant en oeuvre ce procede |
Publications (2)
Publication Number | Publication Date |
---|---|
EP1563485A1 EP1563485A1 (de) | 2005-08-17 |
EP1563485B1 true EP1563485B1 (de) | 2006-03-29 |
Family
ID=32187712
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
EP03782553A Expired - Lifetime EP1563485B1 (de) | 2002-11-19 | 2003-11-13 | Verfahren zur verarbeitung von audiodateien und erfassungsvorrichtung zur anwendung davon |
Country Status (12)
Country | Link |
---|---|
US (1) | US7706543B2 (de) |
EP (1) | EP1563485B1 (de) |
JP (1) | JP4343845B2 (de) |
KR (1) | KR100964353B1 (de) |
CN (1) | CN1735922B (de) |
AT (1) | ATE322065T1 (de) |
AU (1) | AU2003290190A1 (de) |
DE (1) | DE60304358T2 (de) |
ES (1) | ES2261994T3 (de) |
FR (1) | FR2847376B1 (de) |
WO (1) | WO2004049299A1 (de) |
ZA (1) | ZA200503969B (de) |
Cited By (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US10595148B2 (en) | 2016-01-08 | 2020-03-17 | Sony Corporation | Sound processing apparatus and method, and program |
Families Citing this family (77)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
DE10328335B4 (de) * | 2003-06-24 | 2005-07-21 | Fraunhofer-Gesellschaft zur Förderung der angewandten Forschung e.V. | Wellenfeldsyntesevorrichtung und Verfahren zum Treiben eines Arrays von Lautsprechern |
US20050271216A1 (en) * | 2004-06-04 | 2005-12-08 | Khosrow Lashkari | Method and apparatus for loudspeaker equalization |
CN101263739B (zh) * | 2005-09-13 | 2012-06-20 | Srs实验室有限公司 | 用于音频处理的系统和方法 |
ES2335246T3 (es) * | 2006-03-13 | 2010-03-23 | France Telecom | Sintesis y especializacion sonora conjunta. |
FR2899424A1 (fr) * | 2006-03-28 | 2007-10-05 | France Telecom | Procede de synthese binaurale prenant en compte un effet de salle |
US8180067B2 (en) * | 2006-04-28 | 2012-05-15 | Harman International Industries, Incorporated | System for selectively extracting components of an audio input signal |
US7876903B2 (en) * | 2006-07-07 | 2011-01-25 | Harris Corporation | Method and apparatus for creating a multi-dimensional communication space for use in a binaural audio system |
US8036767B2 (en) * | 2006-09-20 | 2011-10-11 | Harman International Industries, Incorporated | System for extracting and changing the reverberant content of an audio input signal |
JP4949477B2 (ja) * | 2006-09-25 | 2012-06-06 | ドルビー ラボラトリーズ ライセンシング コーポレイション | 高次角度項による信号を抽出することでマルチチャンネルオーディオ再生システムの空間分解能を改善したサウンドフィールド |
DE102006053919A1 (de) * | 2006-10-11 | 2008-04-17 | Fraunhofer-Gesellschaft zur Förderung der angewandten Forschung e.V. | Vorrichtung und Verfahren zum Erzeugen einer Anzahl von Lautsprechersignalen für ein Lautsprecher-Array, das einen Wiedergaberaum definiert |
JP2008118559A (ja) * | 2006-11-07 | 2008-05-22 | Advanced Telecommunication Research Institute International | 3次元音場再生装置 |
JP4873316B2 (ja) * | 2007-03-09 | 2012-02-08 | 株式会社国際電気通信基礎技術研究所 | 音響空間共有装置 |
EP2094032A1 (de) * | 2008-02-19 | 2009-08-26 | Deutsche Thomson OHG | Audiosignal, Verfahren und Vorrichtung zu dessen Kodierung oder Übertragung sowie Verfahren und Vorrichtung zu dessen Verarbeitung |
WO2009109217A1 (en) * | 2008-03-03 | 2009-09-11 | Nokia Corporation | Apparatus for capturing and rendering a plurality of audio channels |
EP2154677B1 (de) | 2008-08-13 | 2013-07-03 | Fraunhofer-Gesellschaft zur Förderung der angewandten Forschung e.V. | Vorrichtung zur Bestimmung eines konvertierten Raumtonsignals |
EP2154910A1 (de) * | 2008-08-13 | 2010-02-17 | Fraunhofer-Gesellschaft zur Förderung der angewandten Forschung e.V. | Vorrichtung zum Mischen von Raumtonströmen |
GB0815362D0 (en) | 2008-08-22 | 2008-10-01 | Queen Mary & Westfield College | Music collection navigation |
US8819554B2 (en) * | 2008-12-23 | 2014-08-26 | At&T Intellectual Property I, L.P. | System and method for playing media |
EP2205007B1 (de) * | 2008-12-30 | 2019-01-09 | Dolby International AB | Verfahren und Vorrichtung zur Kodierung dreidimensionaler Hörbereiche und zur optimalen Rekonstruktion |
GB2476747B (en) * | 2009-02-04 | 2011-12-21 | Richard Furse | Sound system |
JP5340296B2 (ja) * | 2009-03-26 | 2013-11-13 | パナソニック株式会社 | 復号化装置、符号化復号化装置および復号化方法 |
EP2486737B1 (de) * | 2009-10-05 | 2016-05-11 | Harman International Industries, Incorporated | System zur räumlichen extraktion von tonsignalen |
BR112012024528B1 (pt) * | 2010-03-26 | 2021-05-11 | Dolby International Ab | método e dispositivo para decodificar uma representação para campo de som de áudio para reprodução de áudio e meio legível por computador |
JP5672741B2 (ja) * | 2010-03-31 | 2015-02-18 | ソニー株式会社 | 信号処理装置および方法、並びにプログラム |
US20110317522A1 (en) * | 2010-06-28 | 2011-12-29 | Microsoft Corporation | Sound source localization based on reflections and room estimation |
US9313599B2 (en) | 2010-11-19 | 2016-04-12 | Nokia Technologies Oy | Apparatus and method for multi-channel signal playback |
US9055371B2 (en) * | 2010-11-19 | 2015-06-09 | Nokia Technologies Oy | Controllable playback system offering hierarchical playback options |
US9456289B2 (en) | 2010-11-19 | 2016-09-27 | Nokia Technologies Oy | Converting multi-microphone captured signals to shifted signals useful for binaural signal processing and use thereof |
EP2541547A1 (de) | 2011-06-30 | 2013-01-02 | Thomson Licensing | Verfahren und Vorrichtung zum Ändern der relativen Standorte von Schallobjekten innerhalb einer Higher-Order-Ambisonics-Wiedergabe |
EP2777301B1 (de) * | 2011-11-10 | 2015-08-12 | SonicEmotion AG | Verfahren für praktische implementierungen einer schallfeldwiedergabe auf basis von flächenintegralen in drei dimensionen |
KR101282673B1 (ko) | 2011-12-09 | 2013-07-05 | 현대자동차주식회사 | 음원 위치 추정 방법 |
US8996296B2 (en) * | 2011-12-15 | 2015-03-31 | Qualcomm Incorporated | Navigational soundscaping |
CN106847737B (zh) | 2012-02-29 | 2020-11-13 | 应用材料公司 | 配置中的除污及剥除处理腔室 |
EP2645748A1 (de) | 2012-03-28 | 2013-10-02 | Thomson Licensing | Verfahren und Vorrichtung zum Decodieren von Stereolautsprechersignalen aus einem Ambisonics-Audiosignal höherer Ordnung |
CN108810744A (zh) | 2012-04-05 | 2018-11-13 | 诺基亚技术有限公司 | 柔性的空间音频捕捉设备 |
US9288603B2 (en) | 2012-07-15 | 2016-03-15 | Qualcomm Incorporated | Systems, methods, apparatus, and computer-readable media for backward-compatible audio coding |
EP2688066A1 (de) * | 2012-07-16 | 2014-01-22 | Thomson Licensing | Verfahren und Vorrichtung zur Codierung von Mehrkanal-HOA-Audiosignalen zur Rauschreduzierung sowie Verfahren und Vorrichtung zur Decodierung von Mehrkanal-HOA-Audiosignalen zur Rauschreduzierung |
US9473870B2 (en) | 2012-07-16 | 2016-10-18 | Qualcomm Incorporated | Loudspeaker position compensation with 3D-audio hierarchical coding |
US9516446B2 (en) | 2012-07-20 | 2016-12-06 | Qualcomm Incorporated | Scalable downmix design for object-based surround codec with cluster analysis by synthesis |
US9761229B2 (en) * | 2012-07-20 | 2017-09-12 | Qualcomm Incorporated | Systems, methods, apparatus, and computer-readable media for audio object clustering |
JP6167178B2 (ja) | 2012-08-31 | 2017-07-19 | ドルビー ラボラトリーズ ライセンシング コーポレイション | オブジェクトに基づくオーディオのための反射音レンダリング |
US9301069B2 (en) * | 2012-12-27 | 2016-03-29 | Avaya Inc. | Immersive 3D sound space for searching audio |
US9892743B2 (en) | 2012-12-27 | 2018-02-13 | Avaya Inc. | Security surveillance via three-dimensional audio space presentation |
US9838824B2 (en) | 2012-12-27 | 2017-12-05 | Avaya Inc. | Social media processing with three-dimensional audio |
US10203839B2 (en) | 2012-12-27 | 2019-02-12 | Avaya Inc. | Three-dimensional generalized space |
US9913064B2 (en) | 2013-02-07 | 2018-03-06 | Qualcomm Incorporated | Mapping virtual speakers to physical speakers |
US9685163B2 (en) * | 2013-03-01 | 2017-06-20 | Qualcomm Incorporated | Transforming spherical harmonic coefficients |
EP2982139A4 (de) | 2013-04-04 | 2016-11-23 | Nokia Technologies Oy | Audiovisuelle verarbeitungsvorrichtung |
EP2997573A4 (de) | 2013-05-17 | 2017-01-18 | Nokia Technologies OY | Räumliche objektorientierte audiovorrichtung |
US9420393B2 (en) | 2013-05-29 | 2016-08-16 | Qualcomm Incorporated | Binaural rendering of spherical harmonic coefficients |
US9769586B2 (en) | 2013-05-29 | 2017-09-19 | Qualcomm Incorporated | Performing order reduction with respect to higher order ambisonic coefficients |
EP2824661A1 (de) * | 2013-07-11 | 2015-01-14 | Thomson Licensing | Verfahren und Vorrichtung zur Erzeugung aus einer Koeffizientendomänenrepräsentation von HOA-Signalen eine gemischte Raum-/Koeffizientendomänenrepräsentation der besagten HOA-Signale |
DE102013013378A1 (de) * | 2013-08-10 | 2015-02-12 | Advanced Acoustic Sf Gmbh | Aufteilung virtueller Schallquellen |
CN105637901B (zh) | 2013-10-07 | 2018-01-23 | 杜比实验室特许公司 | 空间音频处理系统和方法 |
EP2866475A1 (de) * | 2013-10-23 | 2015-04-29 | Thomson Licensing | Verfahren und Vorrichtung zur Decodierung einer Audioschallfelddarstellung für Audiowiedergabe mittels 2D-Einstellungen |
US9922656B2 (en) | 2014-01-30 | 2018-03-20 | Qualcomm Incorporated | Transitioning of ambient higher-order ambisonic coefficients |
EP2930958A1 (de) * | 2014-04-07 | 2015-10-14 | Harman Becker Automotive Systems GmbH | Schallwellenfelderzeugung |
US10770087B2 (en) | 2014-05-16 | 2020-09-08 | Qualcomm Incorporated | Selecting codebooks for coding vectors decomposed from higher-order ambisonic audio signals |
JP6388551B2 (ja) * | 2015-02-27 | 2018-09-12 | アルパイン株式会社 | 複数領域音場再現システムおよび方法 |
DE102015008000A1 (de) * | 2015-06-24 | 2016-12-29 | Saalakustik.De Gmbh | Verfahren zur Schallwiedergabe in Reflexionsumgebungen, insbesondere in Hörräumen |
US10582329B2 (en) | 2016-01-08 | 2020-03-03 | Sony Corporation | Audio processing device and method |
EP3402221B1 (de) * | 2016-01-08 | 2020-04-08 | Sony Corporation | Audioverarbeitungsvorrichtung, -verfahren und -programm |
AU2017305249B2 (en) | 2016-08-01 | 2021-07-22 | Magic Leap, Inc. | Mixed reality system with spatialized audio |
US11032663B2 (en) * | 2016-09-29 | 2021-06-08 | The Trustees Of Princeton University | System and method for virtual navigation of sound fields through interpolation of signals from an array of microphone assemblies |
US20180124540A1 (en) * | 2016-10-31 | 2018-05-03 | Google Llc | Projection-based audio coding |
FR3060830A1 (fr) * | 2016-12-21 | 2018-06-22 | Orange | Traitement en sous-bandes d'un contenu ambisonique reel pour un decodage perfectionne |
US10182303B1 (en) * | 2017-07-12 | 2019-01-15 | Google Llc | Ambisonics sound field navigation using directional decomposition and path distance estimation |
US10764684B1 (en) | 2017-09-29 | 2020-09-01 | Katherine A. Franco | Binaural audio using an arbitrarily shaped microphone array |
US10721559B2 (en) | 2018-02-09 | 2020-07-21 | Dolby Laboratories Licensing Corporation | Methods, apparatus and systems for audio sound field capture |
CA3092756A1 (en) * | 2018-03-02 | 2019-09-06 | Wilfred Edwin Booij | Acoustic positioning transmitter and receiver system and method |
US10771913B2 (en) | 2018-05-11 | 2020-09-08 | Dts, Inc. | Determining sound locations in multi-channel audio |
CN110740416B (zh) * | 2019-09-27 | 2021-04-06 | 广州励丰文化科技股份有限公司 | 一种音频信号处理方法及装置 |
CN110740404B (zh) * | 2019-09-27 | 2020-12-25 | 广州励丰文化科技股份有限公司 | 一种音频相关性的处理方法及音频处理装置 |
CN115715470A (zh) | 2019-12-30 | 2023-02-24 | 卡姆希尔公司 | 用于提供空间化声场的方法 |
CN111537058B (zh) * | 2020-04-16 | 2022-04-29 | 哈尔滨工程大学 | 一种基于Helmholtz方程最小二乘法的声场分离方法 |
US11743670B2 (en) | 2020-12-18 | 2023-08-29 | Qualcomm Incorporated | Correlation-based rendering with multiple distributed streams accounting for an occlusion for six degree of freedom applications |
CN113791385A (zh) * | 2021-09-15 | 2021-12-14 | 张维翔 | 一种三维定位方法及系统 |
Family Cites Families (9)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JPS53114201U (de) * | 1977-02-18 | 1978-09-11 | ||
US4731848A (en) * | 1984-10-22 | 1988-03-15 | Northwestern University | Spatial reverberator |
JP2569872B2 (ja) * | 1990-03-02 | 1997-01-08 | ヤマハ株式会社 | 音場制御装置 |
JP3578783B2 (ja) * | 1993-09-24 | 2004-10-20 | ヤマハ株式会社 | 電子楽器の音像定位装置 |
US5745584A (en) * | 1993-12-14 | 1998-04-28 | Taylor Group Of Companies, Inc. | Sound bubble structures for sound reproducing arrays |
GB9726338D0 (en) * | 1997-12-13 | 1998-02-11 | Central Research Lab Ltd | A method of processing an audio signal |
US7231054B1 (en) * | 1999-09-24 | 2007-06-12 | Creative Technology Ltd | Method and apparatus for three-dimensional audio display |
US7340062B2 (en) * | 2000-03-14 | 2008-03-04 | Revit Lawrence J | Sound reproduction method and apparatus for assessing real-world performance of hearing and hearing aids |
AU2000280030A1 (en) * | 2000-04-19 | 2001-11-07 | Sonic Solutions | Multi-channel surround sound mastering and reproduction techniques that preservespatial harmonics in three dimensions |
-
2002
- 2002-11-19 FR FR0214444A patent/FR2847376B1/fr not_active Expired - Fee Related
-
2003
- 2003-11-13 DE DE60304358T patent/DE60304358T2/de not_active Expired - Lifetime
- 2003-11-13 ES ES03782553T patent/ES2261994T3/es not_active Expired - Lifetime
- 2003-11-13 EP EP03782553A patent/EP1563485B1/de not_active Expired - Lifetime
- 2003-11-13 CN CN2003801086029A patent/CN1735922B/zh not_active Expired - Lifetime
- 2003-11-13 WO PCT/FR2003/003367 patent/WO2004049299A1/fr active IP Right Grant
- 2003-11-13 US US10/535,524 patent/US7706543B2/en active Active
- 2003-11-13 AU AU2003290190A patent/AU2003290190A1/en not_active Abandoned
- 2003-11-13 JP JP2004554598A patent/JP4343845B2/ja not_active Expired - Lifetime
- 2003-11-13 KR KR1020057009105A patent/KR100964353B1/ko active IP Right Grant
- 2003-11-13 AT AT03782553T patent/ATE322065T1/de not_active IP Right Cessation
-
2005
- 2005-05-17 ZA ZA200503969A patent/ZA200503969B/en unknown
Cited By (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US10595148B2 (en) | 2016-01-08 | 2020-03-17 | Sony Corporation | Sound processing apparatus and method, and program |
Also Published As
Publication number | Publication date |
---|---|
AU2003290190A1 (en) | 2004-06-18 |
CN1735922B (zh) | 2010-05-12 |
KR20050083928A (ko) | 2005-08-26 |
US7706543B2 (en) | 2010-04-27 |
KR100964353B1 (ko) | 2010-06-17 |
CN1735922A (zh) | 2006-02-15 |
WO2004049299A1 (fr) | 2004-06-10 |
DE60304358T2 (de) | 2006-12-07 |
JP4343845B2 (ja) | 2009-10-14 |
EP1563485A1 (de) | 2005-08-17 |
DE60304358D1 (de) | 2006-05-18 |
ATE322065T1 (de) | 2006-04-15 |
FR2847376B1 (fr) | 2005-02-04 |
JP2006506918A (ja) | 2006-02-23 |
FR2847376A1 (fr) | 2004-05-21 |
US20060045275A1 (en) | 2006-03-02 |
ES2261994T3 (es) | 2006-11-16 |
BR0316718A (pt) | 2005-10-18 |
ZA200503969B (en) | 2006-09-27 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
EP1563485B1 (de) | Verfahren zur verarbeitung von audiodateien und erfassungsvorrichtung zur anwendung davon | |
EP1836876B1 (de) | Verfahren und vorrichtung zur individualisierung von hrtfs durch modellierung | |
EP1992198B1 (de) | Optimierung des binauralen raumklangeffektes durch mehrkanalkodierung | |
EP1586220B1 (de) | Verfahren und einrichtung zur steuerung einer wiedergabeeinheitdurch verwendung eines mehrkanalsignals | |
Ben-Hur et al. | Binaural reproduction based on bilateral ambisonics and ear-aligned HRTFs | |
EP1479266B1 (de) | Verfahren und vorrichtung zur steuerung einer anordnung zur wiedergabe eines schallfeldes | |
EP3475943B1 (de) | Verfahren zur umwandlung und stereophonen codierung eines dreidimensionalen audiosignals | |
EP3400599B1 (de) | Verbesserter ambisonic-codierer für eine tonquelle mit mehreren reflexionen | |
EP3025514B1 (de) | Klangverräumlichung mit raumwirkung | |
FR3065137A1 (fr) | Procede de spatialisation sonore | |
EP3384688B1 (de) | Aufeinanderfolgende dekompositionen von audiofiltern | |
EP4184505A1 (de) | Komplexitätsoptimierte klangraumisierung mit raumeffekt | |
Paulo et al. | Perceptual Comparative Tests Between the Multichannel 3D Capturing Systems Artificial Ears and the Ambisonic Concept | |
FR2866974A1 (fr) | Procede de traitement sonores, en particulier en contexte ambiophonique | |
EP3449643B1 (de) | Verfahren und system zum senden eines 360°-audiosignals | |
FR3040253B1 (fr) | Procede de mesure de filtres phrtf d'un auditeur, cabine pour la mise en oeuvre du procede, et procedes permettant d'aboutir a la restitution d'une bande sonore multicanal personnalisee | |
US20200186952A1 (en) | Method and system for processing an audio signal including ambisonic encoding | |
EP3484185A1 (de) | Modellierung einer menge von akustischen übertragungsfunktionen einer person, 3d-soundkarte und 3d-sound-reproduktionssystem | |
CN116261086A (zh) | 声音信号处理方法、装置、设备及存储介质 | |
BRPI0316718B1 (pt) | Processo de tratamento de dados sonoros e dispositivo de aquisição sonora, aplicando esse processo |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PUAI | Public reference made under article 153(3) epc to a published international application that has entered the european phase |
Free format text: ORIGINAL CODE: 0009012 |
|
17P | Request for examination filed |
Effective date: 20050530 |
|
AK | Designated contracting states |
Kind code of ref document: A1 Designated state(s): AT BE BG CH CY CZ DE DK EE ES FI FR GB GR HU IE IT LI LU MC NL PT RO SE SI SK TR |
|
AX | Request for extension of the european patent |
Extension state: AL LT LV MK |
|
GRAP | Despatch of communication of intention to grant a patent |
Free format text: ORIGINAL CODE: EPIDOSNIGR1 |
|
GRAS | Grant fee paid |
Free format text: ORIGINAL CODE: EPIDOSNIGR3 |
|
GRAA | (expected) grant |
Free format text: ORIGINAL CODE: 0009210 |
|
DAX | Request for extension of the european patent (deleted) | ||
AK | Designated contracting states |
Kind code of ref document: B1 Designated state(s): AT BE BG CH CY CZ DE DK EE ES FI FR GB GR HU IE IT LI LU MC NL PT RO SE SI SK TR |
|
PG25 | Lapsed in a contracting state [announced via postgrant information from national office to epo] |
Ref country code: RO Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT Effective date: 20060329 Ref country code: IT Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT;WARNING: LAPSES OF ITALIAN PATENTS WITH EFFECTIVE DATE BEFORE 2007 MAY HAVE OCCURRED AT ANY TIME BEFORE 2007. THE CORRECT EFFECTIVE DATE MAY BE DIFFERENT FROM THE ONE RECORDED. Effective date: 20060329 Ref country code: SK Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT Effective date: 20060329 Ref country code: AT Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT Effective date: 20060329 Ref country code: SI Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT Effective date: 20060329 Ref country code: IE Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT Effective date: 20060329 Ref country code: NL Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT Effective date: 20060329 |
|
REG | Reference to a national code |
Ref country code: GB Ref legal event code: FG4D Free format text: NOT ENGLISH |
|
REG | Reference to a national code |
Ref country code: CH Ref legal event code: EP |
|
REG | Reference to a national code |
Ref country code: IE Ref legal event code: FG4D Free format text: LANGUAGE OF EP DOCUMENT: FRENCH |
|
REF | Corresponds to: |
Ref document number: 60304358 Country of ref document: DE Date of ref document: 20060518 Kind code of ref document: P |
|
PG25 | Lapsed in a contracting state [announced via postgrant information from national office to epo] |
Ref country code: DK Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT Effective date: 20060629 Ref country code: SE Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT Effective date: 20060629 Ref country code: BG Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT Effective date: 20060629 |
|
GBT | Gb: translation of ep patent filed (gb section 77(6)(a)/1977) |
Effective date: 20060712 |
|
PG25 | Lapsed in a contracting state [announced via postgrant information from national office to epo] |
Ref country code: PT Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT Effective date: 20060829 |
|
NLV1 | Nl: lapsed or annulled due to failure to fulfill the requirements of art. 29p and 29m of the patents act | ||
REG | Reference to a national code |
Ref country code: IE Ref legal event code: FD4D |
|
REG | Reference to a national code |
Ref country code: ES Ref legal event code: FG2A Ref document number: 2261994 Country of ref document: ES Kind code of ref document: T3 |
|
PG25 | Lapsed in a contracting state [announced via postgrant information from national office to epo] |
Ref country code: MC Free format text: LAPSE BECAUSE OF NON-PAYMENT OF DUE FEES Effective date: 20061130 Ref country code: BE Free format text: LAPSE BECAUSE OF NON-PAYMENT OF DUE FEES Effective date: 20061130 |
|
PLBE | No opposition filed within time limit |
Free format text: ORIGINAL CODE: 0009261 |
|
STAA | Information on the status of an ep patent application or granted ep patent |
Free format text: STATUS: NO OPPOSITION FILED WITHIN TIME LIMIT |
|
26N | No opposition filed |
Effective date: 20070102 |
|
BERE | Be: lapsed |
Owner name: FRANCE TELECOM Effective date: 20061130 |
|
PG25 | Lapsed in a contracting state [announced via postgrant information from national office to epo] |
Ref country code: GR Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT Effective date: 20060630 Ref country code: CZ Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT Effective date: 20060329 |
|
PG25 | Lapsed in a contracting state [announced via postgrant information from national office to epo] |
Ref country code: EE Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT Effective date: 20060329 Ref country code: FI Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT Effective date: 20060329 |
|
PG25 | Lapsed in a contracting state [announced via postgrant information from national office to epo] |
Ref country code: CH Free format text: LAPSE BECAUSE OF NON-PAYMENT OF DUE FEES Effective date: 20071130 Ref country code: LU Free format text: LAPSE BECAUSE OF NON-PAYMENT OF DUE FEES Effective date: 20061113 Ref country code: LI Free format text: LAPSE BECAUSE OF NON-PAYMENT OF DUE FEES Effective date: 20071130 Ref country code: TR Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT Effective date: 20060329 Ref country code: HU Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT Effective date: 20060930 |
|
REG | Reference to a national code |
Ref country code: CH Ref legal event code: PL |
|
PG25 | Lapsed in a contracting state [announced via postgrant information from national office to epo] |
Ref country code: CY Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT Effective date: 20060329 |
|
PGFP | Annual fee paid to national office [announced via postgrant information from national office to epo] |
Ref country code: FR Payment date: 20101217 Year of fee payment: 8 |
|
REG | Reference to a national code |
Ref country code: FR Ref legal event code: ST Effective date: 20120731 |
|
PG25 | Lapsed in a contracting state [announced via postgrant information from national office to epo] |
Ref country code: FR Free format text: LAPSE BECAUSE OF NON-PAYMENT OF DUE FEES Effective date: 20111130 |
|
PGFP | Annual fee paid to national office [announced via postgrant information from national office to epo] |
Ref country code: IT Payment date: 20221020 Year of fee payment: 20 Ref country code: GB Payment date: 20221021 Year of fee payment: 20 Ref country code: ES Payment date: 20221201 Year of fee payment: 20 Ref country code: DE Payment date: 20221020 Year of fee payment: 20 |
|
REG | Reference to a national code |
Ref country code: DE Ref legal event code: R071 Ref document number: 60304358 Country of ref document: DE |
|
REG | Reference to a national code |
Ref country code: ES Ref legal event code: FD2A Effective date: 20231124 |
|
REG | Reference to a national code |
Ref country code: GB Ref legal event code: PE20 Expiry date: 20231112 |
|
PG25 | Lapsed in a contracting state [announced via postgrant information from national office to epo] |
Ref country code: GB Free format text: LAPSE BECAUSE OF EXPIRATION OF PROTECTION Effective date: 20231112 |
|
PG25 | Lapsed in a contracting state [announced via postgrant information from national office to epo] |
Ref country code: ES Free format text: LAPSE BECAUSE OF EXPIRATION OF PROTECTION Effective date: 20231114 |
|
PG25 | Lapsed in a contracting state [announced via postgrant information from national office to epo] |
Ref country code: GB Free format text: LAPSE BECAUSE OF EXPIRATION OF PROTECTION Effective date: 20231112 Ref country code: ES Free format text: LAPSE BECAUSE OF EXPIRATION OF PROTECTION Effective date: 20231114 |