EP3756363A1 - Dispositif et procédé pour matriçage audio spatial à base d'objet - Google Patents
Dispositif et procédé pour matriçage audio spatial à base d'objetInfo
- Publication number
- EP3756363A1 EP3756363A1 EP19710283.3A EP19710283A EP3756363A1 EP 3756363 A1 EP3756363 A1 EP 3756363A1 EP 19710283 A EP19710283 A EP 19710283A EP 3756363 A1 EP3756363 A1 EP 3756363A1
- Authority
- EP
- European Patent Office
- Prior art keywords
- audio
- audio objects
- processing object
- objects
- signal
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Pending
Links
- 238000000034 method Methods 0.000 title claims description 62
- 238000012545 processing Methods 0.000 claims abstract description 305
- 230000000694 effects Effects 0.000 claims abstract description 113
- 238000004590 computer program Methods 0.000 claims description 14
- 238000009877 rendering Methods 0.000 claims description 12
- 230000003321 amplification Effects 0.000 claims description 4
- 238000003199 nucleic acid amplification method Methods 0.000 claims description 4
- 230000008859 change Effects 0.000 description 15
- 230000008569 process Effects 0.000 description 13
- 230000005236 sound signal Effects 0.000 description 13
- 230000006835 compression Effects 0.000 description 11
- 238000007906 compression Methods 0.000 description 11
- 238000004519 manufacturing process Methods 0.000 description 11
- 238000004364 calculation method Methods 0.000 description 10
- 238000013459 approach Methods 0.000 description 7
- 230000000875 corresponding effect Effects 0.000 description 7
- 230000009466 transformation Effects 0.000 description 6
- 230000007704 transition Effects 0.000 description 6
- 230000008878 coupling Effects 0.000 description 5
- 238000010168 coupling process Methods 0.000 description 5
- 238000005859 coupling reaction Methods 0.000 description 5
- 238000003860 storage Methods 0.000 description 5
- 230000036962 time dependent Effects 0.000 description 5
- 210000003746 feather Anatomy 0.000 description 4
- 230000006870 function Effects 0.000 description 4
- 238000004458 analytical method Methods 0.000 description 3
- 239000011159 matrix material Substances 0.000 description 3
- 239000010981 turquoise Substances 0.000 description 3
- 101100058681 Drosophila melanogaster Btk29A gene Proteins 0.000 description 2
- 230000004888 barrier function Effects 0.000 description 2
- 238000005516 engineering process Methods 0.000 description 2
- 238000005562 fading Methods 0.000 description 2
- 239000003292 glue Substances 0.000 description 2
- 238000013507 mapping Methods 0.000 description 2
- 230000004048 modification Effects 0.000 description 2
- 238000012986 modification Methods 0.000 description 2
- 230000004044 response Effects 0.000 description 2
- 238000012800 visualization Methods 0.000 description 2
- 230000006978 adaptation Effects 0.000 description 1
- 230000005540 biological transmission Effects 0.000 description 1
- 238000006243 chemical reaction Methods 0.000 description 1
- 238000004891 communication Methods 0.000 description 1
- 230000001276 controlling effect Effects 0.000 description 1
- 230000003247 decreasing effect Effects 0.000 description 1
- 230000001419 dependent effect Effects 0.000 description 1
- 230000003993 interaction Effects 0.000 description 1
- 230000002452 interceptive effect Effects 0.000 description 1
- 230000003287 optical effect Effects 0.000 description 1
- 238000004091 panning Methods 0.000 description 1
- 230000002787 reinforcement Effects 0.000 description 1
- 230000000717 retained effect Effects 0.000 description 1
- 238000001228 spectrum Methods 0.000 description 1
- 230000000087 stabilizing effect Effects 0.000 description 1
- 230000002123 temporal effect Effects 0.000 description 1
- 230000001052 transient effect Effects 0.000 description 1
- 238000013519 translation Methods 0.000 description 1
Classifications
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04S—STEREOPHONIC SYSTEMS
- H04S7/00—Indicating arrangements; Control arrangements, e.g. balance control
- H04S7/30—Control circuits for electronic adaptation of the sound field
- H04S7/307—Frequency adjustment, e.g. tone control
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04S—STEREOPHONIC SYSTEMS
- H04S7/00—Indicating arrangements; Control arrangements, e.g. balance control
- H04S7/30—Control circuits for electronic adaptation of the sound field
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
- G10L19/00—Speech or audio signals analysis-synthesis techniques for redundancy reduction, e.g. in vocoders; Coding or decoding of speech or audio signals, using source filter models or psychoacoustic analysis
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
- G10L19/00—Speech or audio signals analysis-synthesis techniques for redundancy reduction, e.g. in vocoders; Coding or decoding of speech or audio signals, using source filter models or psychoacoustic analysis
- G10L19/008—Multichannel audio signal coding or decoding using interchannel correlation to reduce redundancy, e.g. joint-stereo, intensity-coding or matrixing
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04R—LOUDSPEAKERS, MICROPHONES, GRAMOPHONE PICK-UPS OR LIKE ACOUSTIC ELECTROMECHANICAL TRANSDUCERS; DEAF-AID SETS; PUBLIC ADDRESS SYSTEMS
- H04R3/00—Circuits for transducers, loudspeakers or microphones
- H04R3/04—Circuits for transducers, loudspeakers or microphones for correcting frequency response
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04R—LOUDSPEAKERS, MICROPHONES, GRAMOPHONE PICK-UPS OR LIKE ACOUSTIC ELECTROMECHANICAL TRANSDUCERS; DEAF-AID SETS; PUBLIC ADDRESS SYSTEMS
- H04R5/00—Stereophonic arrangements
- H04R5/04—Circuit arrangements, e.g. for selective connection of amplifier inputs/outputs to loudspeakers, for loudspeaker detection, or for adaptation of settings to personal preferences or hearing impairments
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04S—STEREOPHONIC SYSTEMS
- H04S2400/00—Details of stereophonic systems covered by H04S but not provided for in its groups
- H04S2400/03—Aspects of down-mixing multi-channel audio to configurations with lower numbers of playback channels, e.g. 7.1 -> 5.1
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04S—STEREOPHONIC SYSTEMS
- H04S2400/00—Details of stereophonic systems covered by H04S but not provided for in its groups
- H04S2400/11—Positioning of individual sound objects, e.g. moving airplane, within a sound field
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04S—STEREOPHONIC SYSTEMS
- H04S2400/00—Details of stereophonic systems covered by H04S but not provided for in its groups
- H04S2400/13—Aspects of volume control, not necessarily automatic, in stereophonic sound systems
Definitions
- the application relates to audio object processing, audio object encoding and audio object decoding and, more particularly, audio mastering for audio objects.
- Object-based spatial audio is an approach to interactive three-dimensional audio production. Not only does this concept change how content creators or authors interact with the audio, but also how it is stored and transmitted. To make this possible, a new process has to be established in the reproduction chain called "ren- dering". The rendering process generates speaker signals from an object-based scene description. Although recording and mixing have been explored in recent years, concepts for object-based mastering are almost absent. The main difference compared to channel-based audio mastering is that instead of adjusting the audio channels, the audio objects need to be changed. This requires a fundamentally new concept for mastering. The paper presents a new method for mastering object-based audio.
- An audio object may be considered as a virtual sound source consisting of an audio signal with additional metadata, e.g. B. position and gain, be available.
- additional metadata e.g. B. position and gain
- the audio rendering is the process of generating speaker or headphone signals based on other information, such as the position of speakers or the position of the listener in the virtual scene.
- the process of audio content creation can be divided into three main parts: recording, mixing and mastering. While all three steps in the past decade have been extensively covered for channel-based audio, object-based audio requires new workflows in future applications. So far, the recording step generally does not need to be changed, even if future techniques involve new possibilities [1], [2] could bring. It behaves in the mixing process slightly different as the Sound Engineer no longer creates a spatial mix by panning signals to dedicated speakers. Instead, all positions of audio objects are generated by a spatial authoring tool that allows the metadata portion of each audio object to be defined. A complete mastering process for audio objects has not yet been established [3]
- the mastering phase makes it useful for creators to produce audio in sub-optimal acoustic environments, as they can rely on a final examination of their mixing at the mastering stage. This lowers the barriers to accessing pro- professional content.
- the MEs themselves have been offered a wide range of mastering tools over the years that has drastically improved their ability to correct and improve. Nonetheless, the final content is usually limited to the playback device for which it was designed. This limitation is basically overcome by object-based Spatial Audio Production (OBAP).
- OBAP object-based Spatial Audio Production
- OBAP Unlike channel-based audio, OBAP relies on individual audio objects with metadata that includes their position in an artificial environment, also called a "scene.” Only at the final listening output does a dedicated rendering unit, the renderer, calculate the final loudspeaker signals in real time based on the loudspeaker equipment of the listener.
- OBAP provides each audio object and its metadata individually to the renderer, no direct channel-based adjustments are possible during production, and thus no existing mastering tools can be used for conventional rendering facilities. Meanwhile, OBAP requires that all final adjustments be made in the mix. While the requirement to realize overall sonic adjustments by manually treating each individual audio object is not only highly inefficient, this fact also places high demands on each creator's monitor and strictly limits the sonic quality of 3D object-based audio content to the acoustic Properties of the environment in which it was created.
- An apparatus according to claim 1, an encoder according to claim 14, a decoder according to claim 15, a system according to claim 17, a method according to claim 18 and a computer program according to claim 19 are provided.
- each audio object of the plurality of audio objects comprises an audio object signal and audio object metadata, wherein the audio object metadata includes a position of the audio object and a gain parameter of the audio object.
- the apparatus comprises: an interface for specifying at least one effect parameter of a processing object group of audio objects by a user, the processing object group of audio objects comprising two or more audio objects of the plurality of audio objects.
- the apparatus includes a processor unit configured to generate the processed signal such that the at least one effect parameter specified by the interface relates to the audio object signal or to the audio object metadata of each of the audio objects of the processing object group Audio objects is applied.
- One or more audio objects of the plurality of audio objects do not belong to the processing object group of audio objects.
- each audio object of the plurality of audio objects comprises an audio object signal and audio object metadata, the audio object metadata comprising a position of the audio object and a gain parameter of the audio object.
- the method comprises:
- a processor unit (120) such that the at least one effect parameter specified by the interface is applied to the audio object signal or to the audio object metadata of each of the audio objects of the processing object group of audio objects. Furthermore, a computer program with a program code for carrying out the method described above is provided.
- the provided audio mastering is based on a mastering of audio objects.
- the embodiments may be positioned anywhere in a scene and freely in real time. In embodiments, for example, the properties of all my audio objects are affected. In their function as artificial containers they can each contain an arbitrary number of audio objects. Each adaptation to a mastering object is converted in real time into individual adjustments to its audio objects.
- Such mastering objects are also referred to as processing objects.
- the user may use a mastering object to make mutual adjustments to multiple audio objects simultaneously.
- the set of target audio objects for a mastering object may be defined in numerous ways according to embodiments. From a spatial perspective, the user can specify a user-defined scope around the position of the mastering object. Alternatively, ös is possible to link individually selected audio objects with the mastering object, regardless of their position. The mastering object also takes into account potential changes in the position of audio objects over time.
- a second property of mastering objects may be their ability to compute how each audio object is individually influenced based on interaction models.
- a mastering object may take on any general mastering effect, such as equalizers and compressors. Effect plug-ins typically provide the user with numerous parameters, e.g. B. for frequency or gain control.
- B. for frequency or gain control.
- a new mastering effect is added to a mastering object, it is automatically copied to all audio objects of its target set. However, not all effect parameter values are transmitted unchanged.
- some parameters of the mastering effect may be weighted before being applied to a particular audio object. The weighting can be based on any metadata or a sound characteristic of the audio object.
- Fig. 1 shows an apparatus for generating a processed signal under
- Fig. 2 shows apparatus according to another embodiment, wherein the front direction is an encoder.
- Fig. 3 shows apparatus according to another embodiment, wherein the front direction is a decoder.
- FIG. 4 shows a system according to an embodiment.
- Fig. 5 shows a processing object with the area A and the fading area
- FIG. 6 shows a processing object having region A and object radii according to one embodiment.
- FIG. 7 shows a relative angle of audio objects to the processing object according to one embodiment.
- Fig. 8 shows an equalizer object with a new radial perimeter of a Ausry approximate shape
- FIG. 9 shows a signal flow of a compression of the signal from n sources according to an embodiment.
- FIG. 10 shows a scene transformation using a control panel M according to an embodiment.
- FIG. 11 shows the relationship of a processing object with which audio signal effects and metadata effects are effected, according to an embodiment.
- FIG. 12 shows the change of audio objects and audio signals to a user input according to an embodiment.
- FIG. 13 shows a processing object P0 4 with a rectangle M for the distortion of the
- Fig. 14 shows processing objects PCh and P0 2 with their respective overlapping two-dimensional catchment areas A and B according to an embodiment.
- Fig. 15 shows processing object P0 3 with rectangular, two-dimensional draw-in area C and the angles between P0 3 and the associated sources S 2 and S 3 according to one embodiment.
- FIG. 16 shows a possible schematic implementation of an equalizer effect applied to a processing object according to an embodiment.
- Fig. 17 shows the processing object P0 5 with a three-dimensional catchment area D and the respective distances dg, dg 2 and dg 3 to the over the catchment area associated sources Si, S 2 and S 3 according to a Ausry tion form.
- Fig. 18 shows a prototypical implementation of a processing object to which an equalizer has been applied according to an embodiment.
- FIG. 19 shows a processing object as in FIG. 18, only at a different position and without a transition surface according to an embodiment.
- FIG. 20 shows a processing object with a surface defined by its azimuth as the catchment area, such that the sources Src22 and Src4 are assigned to the processing object according to an embodiment
- Fig. 21 shows a processing object as in Fig. 20, but with additional transitional range, which can be controlled by the user via the 'feather' slider, in one embodiment.
- Fig. 22 shows several processing objects in the scene, with different ones
- Fig. 23 shows the red square on the right side of the image shows a processing object for horizontally distorting the position of audio objects according to an embodiment.
- Fig. 24 shows the scene after the user has warped the corners of the processing object. The position of all sources has changed according to the distortion according to one embodiment.
- FIG. 25 shows a possible visualization of the assignment of individual audio objects to a processing object according to an embodiment.
- each audio object of the plurality of audio objects comprises an audio object signal and audio object metadata
- the audio object metadata includes a position of the audio object and a gain parameter of the audio object include.
- the apparatus comprises: an interface for specifying at least one effect parameter of a processing object group of audio objects by a user, the processing object group of audio objects comprising two or more audio objects of the plurality of audio objects.
- the apparatus includes a processor unit 120 configured to generate the processed signal such that the at least one effect parameter specified by the interface 110 is applied to the audio object signal or to the audio object metadata of each of the audio objects of the processing object Group of audio objects is applied One or more audio objects of the plurality of audio objects do not belong to the processing object group of audio objects.
- the apparatus of FIG. 1 described above realizes an efficient form of audio mastering for audio objects.
- a group of two or more audio objects are now organized in a group of audio objects called a processing object group.
- a processing object group is a group of audio objects that are organized in that particular group, the processing object group.
- a user now has the option of specifying one or more (at least one) effect parameters by means of the interface 110.
- the processor unit 120 then causes the effect parameter to be applied to all two or more audio objects of the processing object group by a single input of the effect parameter.
- effect parameter is e.g. modifies a particular frequency range of the audio object signal of each of the audio objects of the processing object group.
- the gain parameter of the audio object metadata of each of the audio objects of the processing object group may be increased or decreased correspondingly, for example, depending on the effect parameter.
- the position of the audio object metadata of each of the audio objects of the processing object group may be changed accordingly depending on the effect parameter, for example.
- the effect parameter for example, it is conceivable that all audio objects of the processing object group are shifted by +2 along an x-coordinate axis, -3 along ay-coordinate axis and +4 along a z-coordinate axis.
- an effect parameter to the audio objects of the processing object group has a different effect for each audio object of the processing object group.
- an axis can be defined as an effect parameter.
- the position of all the audio objects of the processing object group is mirrored.
- the positional change of the audio objects of the processing object group then has a different effect for each audio object of the processing object group.
- the processor unit 120 may be e.g. be configured to apply at least one effect parameter specified by the interface to no audio object signal and no audio object metadata of the one or more audio objects not belonging to the processing object group of audio objects.
- the audio object mastering can be done either centrally on the encoder side. Or, on the decoder side, the end user as receiver of the audio object scenery can modify the audio objects themselves according to the invention.
- FIG. 1 An embodiment implementing audio object mastering according to the invention on the encoder side is shown in FIG.
- Fig. 2 shows apparatus according to another embodiment, wherein the apparatus is an encoder.
- the processor unit 120 is configured to generate a downmix signal using the audio object signals of the plurality of audio objects.
- the processor unit 120 is configured to generate a metadata signal using the audio object metadata of the plurality of audio objects.
- the processor unit 120 in Fig. 2 is configured to generate the downmix signal as the processed signal, wherein in the downmix signal at least one modified object signal is mixed for each audio object of the processing object group of audio objects, the processor unit 120 being formed is, for each audio object of the processing object group of audio objects, the modified object signal of this audio object by means of the application of the at least one effect Parameter specified by the interface 110 to generate the audio object signal of this audio object.
- the processor unit 120 of FIG. 2 is configured to generate the metadata signal as the processed signal, wherein the metadata signal comprises at least one modified position for each audio object of the processing object group of audio objects, wherein the processor unit 120 is configured, for each audio object of the processing object group of audio objects, to generate the modified position of that audio object by applying the at least one effect parameter specified by the interface 110 to the position of that audio object.
- the processor unit 120 of FIG. 2 is configured to generate as the processed signal the metadata signal, wherein the metadata signal comprises at least one modified gain parameter for each audio object of the processing object group of audio objects, wherein the processor unit 120 is configured, for each audio object of the processing object group of audio objects, generate the modified gain parameter of that audio object by applying the at least one effect parameter specified by the interface 110 to the gain parameter of that audio object.
- Fig. 3 shows apparatus according to another embodiment, wherein the apparatus is a decoder.
- the apparatus of Fig. 3 is formed out to receive a downmix signal in which the plurality of audio object signals of the plurality of audio objects are mixed.
- the device of FIG. 3 is configured to receive a metadata signal, wherein the metadata signal for each audio object of the plurality of audio objects comprises the audio object metadata of this audio object.
- the processor unit 120 of FIG. 3 is configured to reconstruct the plurality of audio object signals of the plurality of audio objects based on a downmix signal.
- processor unit 120 of FIG. 3 is configured to generate as the processed signal an audio output signal comprising one or more audio output channels.
- the processor unit 120 of FIG. 3 is configured to apply the at least one effect parameter specified by the interface 110 to the audio object signal of each of the audio objects of the processing object group of audio objects or to generate the processed signal the processed signal, the at least one effect parameter, by means of the interface 1 10 has been specified to apply to the position or gain parameter of the audio object metadata of each of the audio objects of the processing object group of audio objects.
- rendering on the decoder side is well known to those skilled in the art, for example from the SAOC Standard (Spatial Audio Object Coding), see [8].
- Decoder side can be specified by a user input via the interface 1 10, for example, one or more rendering parameters.
- the interface 110 of FIG. 3 may be further configured to specify one or more rendering parameters by the user.
- the processor unit 120 of FIG. 3 may be configured to generate the processed signal using the one or more rendering parameters depending on the position of each audio object of the processing object group of audio objects.
- FIG. 4 shows a system according to an embodiment comprising an encoder 200 and a decoder 300.
- the encoder 200 of FIG. 4 is configured to generate a downmix signal based on audio object signals of a plurality of audio objects and to generate a metadata signal based on audio object metadata of the plurality of audio objects, the audio object metadata being a Position of the audio object and a reinforcement parameter of the audio object.
- the decoder 400 of FIG. 4 is configured to generate an audio output signal comprising one or more audio output channels based on the downmix signal and based on the metadata signal.
- the encoder 200 of the system of FIG. 4 may be a device according to FIG.
- the decoder 300 of the system of FIG. 4 may be a device according to FIG. 3.
- the encoder 200 of the system of FIG. 4 may be a device according to FIG. 2, and the decoder 300 of the system of FIG. 4 may be an apparatus of FIG.
- the following embodiments can equally be implemented in a device of FIG. 1 and in an apparatus of FIG. 2 and in an apparatus of FIG. 3. Likewise, they can be implemented in an encoder 200 of the system of FIG. 4, as well as in a decoder 300 of the system of FIG. 4.
- the processor unit 120 may be e.g. be configured to generate the processed signal so that the at least one effect parameter specified by means of the interface 1 10 is applied to the audio object signal of each of the audio objects of the processing object group of audio objects.
- the processor unit 120 may be configured to apply the at least one effect parameter specified by the interface to no audio object signal of the one or more audio objects of the plurality of audio objects that do not belong to the processing object group of audio objects.
- Such an application of the effect parameter may now be, for example, that the application of the effect parameter to the audio object signal of each audio object of the processing object group e.g. modifies a particular frequency range of the audio object signal of each of the audio objects of the processing object group.
- the processor unit 120 may be e.g. be configured to generate the processed signal so that the at least one effect parameter specified by the interface 110 is applied to the gain parameter of the metadata of each of the audio objects of the processing object group of audio objects.
- the processor unit 120 can be designed, for example, to apply the at least one effect parameter specified by the interface to no amplification parameter of the audio object metadata of the one or more audio objects of the plurality of audio objects that areobjected to the processing object group of Audi do not belong.
- the amplification parameter of the audio object metadata of each of the audio objects of the processing object group may be increased correspondingly, for example, depending on the effect parameter.
- the processor unit 120 may, for example, be designed to generate the processed signal in such a way that the at least one effect parameter, which By means of the interface 1 10, to which position of the metadata of each of the audio objects of the processing object group of audio objects is applied.
- the processing unit 120 may be configured to apply the at least one effect parameter specified by the interface to no position of the audio object metadata of the one or more audio objects of the plurality of audio objects that does not belong to the processing object group of audio objects listen.
- the position of the audio object metadata of each of the audio objects of the processing object group may be changed correspondingly, for example, depending on the effect parameter.
- This can e.g. by specifying the corresponding x, y, and z coordinate values by which to move the position of each of the audio objects.
- a shift may be specified by a certain angle rotated around a defined midpoint, for example a user position, or, but, for example, it may be a doubling (or halving, for example) of the distance to a particular point as an effect - Provide parameters for the position of each audio object of the processing object group.
- interface 110 may be configured to specify at least one definition parameter of the processing object group of audio objects by the user.
- the processor unit 120 may be configured, for example, depending on the at least one definition parameter of the processing object group of audio objects specified by the interface 110, to determine which audio objects of the plurality of audio objects of the processing object group of Belong to audio objects.
- the at least one definition parameter of the processing object group of audio objects may include at least one position of a region of interest (where the position of the region of interest is, for example, the center or centroid of the region of interest).
- the region of interest of the processing object group can be assigned to audio objects.
- the processor unit 120 may be designed, for example, for each audio object of the plurality of audio objects depending on the position of the audio object metadata of this audio object and depending on the position of the region of interest determine if this audio object belongs to the processing object group of audio objects.
- the at least one definition parameter of the processing object group of audio objects may e.g. further comprises a radius of the region of interest, associated with the processing object group of audio objects.
- the processor unit 120 can be designed, for example, to decide for each audio object of the plurality of audio objects depending on the position of the audio object metadata of this audio object and depending on the position of the region of interest and depending on the radius of the region of interest This audio object belongs to the processing object group of Audiobjects.
- a user may specify a position of the processing object group and a radius of the processing object group.
- the position of the processing object group can specify a spatial center, and the radius of the processing object group then defines a circle together with the center of the processing object group. All audio objects with a position within the circle or on the circle can then be defined as audio objects of this group of processing objects; any audio objects with a position outside the circle are then not covered by the processing object group.
- the area within the circle line and on the circle line can then be understood as a "region of interest".
- the processor unit 120 may be e.g. be configured to determine a weighting factor for each of the audio objects of the processing object group of audio objects in dependence on a distance between the position of the audio object metadata of this audio object and the position of the area of interest.
- the processor unit 120 may be configured, for example, for each of the audio objects of the processing object group of audio objects, the weighting factor of this audio object together with the at least one effect parameter specified by means of the interface 110 on the audio object signal or on the gain parameter the audio object metadata of this audio object.
- influence of the effect parameter on the individual audio objects of the processing object group is individualized for each audio object by determining, in addition to effect parameters, an individual weighting factor for each audio object that is applied to the audio object.
- the at least one definition parameter of the processing object group of audio objects may include at least one angle specifying a direction from a defined user position in which a region of interest is associated with the processing object group of audio objects .
- the processor unit 120 may be configured, for example, for each audio object of the plurality of audio objects, depending on the position of the metadata of this audio object and in dependence on the angle specifying the direction from the defined user position in which the user of interest Range is to determine if this audio object belongs to the processing object group of audio objects.
- the processor unit 120 may be e.g. be configured to determine a weighting factor for each of the audio objects of the processing object group of audio objects, which depends on a difference of a first angle and a wide ren angle, wherein the first angle is the angle, the direction of the de-defined user position specified, in which the area of interest is located, and wherein the further angle depends on the defined user position and the position of the metadata of this audio object.
- the processor unit 120 may be formed, for example, for each of the audio objects of the processing object group of audio objects, the weighting factor of this audio object together with the at least one effect parameter specified by the interface 110 to the audio object signal or to the audio object Apply gain parameters to the audio object metadata of this audio object.
- the processing object group of audio objects may be a first processing object group of audio objects, e.g.
- one or more other processing object groups of audio objects may exist.
- each processing object group of the one or more further processing object groups of audio objects may comprise one or more audio objects of the plurality of audio objects, wherein at least one audio object of a processing object group of the one or more further processing object groups of audio objects does not contain an audio object of the first processing object Group of audio objects.
- the interface 1 10 may specify the one or more further processing object groups of audio objects for specifying. At least one further effect parameter for this processing object group of audio objects is formed by the user.
- the processor unit 120 may be configured to generate the processed signal such that for each processing object group of the one or more further processing object groups of audio objects the at least one further effect parameter of this processing object group specified by means of the interface 110 to the audio object signal or to the audio object metadata of each of the one or more audio objects of that processing object group, wherein one or more audio objects of the plurality of audio objects do not belong to that processing object group.
- the processor unit 120 may be configured, for example, to apply the at least one further effect parameter of this processing object group specified by means of the interface to no audio object signal and audio object metadata of the one or more audio objects that do not belong to this processing object group ,
- more than one processing object group may exist.
- one or more own effect parameters are determined.
- the interface 110 may be configured by the user to specify the one or more further processing object groups of one or more audio objects by the interface 110 for each processing object group the one or more further processing object groups of one or more audio objects is configured for specifying by the user at least one definition parameter of this processing object group.
- the processor unit 120 can be configured, for example, for each processing object group of the one or more further processing object groups of one or more audio objects in dependence on the at least one definition parameter of this processing object group which specifies the interface 1 10 to determine which audio objects of the plurality of audio objects belong to that processing object group.
- any types of global adjustments in OBAP are made possible by converting global adjustments to individual changes in the affected audio objects (e.g., by the processing unit 120).
- Spatial mastering for object-based audio production can be realized, for example, as follows, by realizing processing objects according to the invention.
- processing objects Processing Objects, POs
- These can be positioned anywhere in a scene and freely in real time just like ordinary audio objects.
- the user can apply any signal processing to the processing object (to the processing object group), for example equalizer (EQ) or compression.
- EQ equalizer
- the parameter settings of the processing object can be converted into object-specific settings.
- Various methods are presented for this calculation.
- FIG 5 shows a processing object with the area A and the fading area Af according to an embodiment.
- the user defines an area A and a blanking area Af around the processing object.
- the processing parameters of the processing object are divided into constant parameters and weighted parameters. Values of constant parameters are unchanged by all audio objects within A and /! / inherited. Weighted parameter values are only inherited by audio objects within A Audio objects within / J / are weighted by a distance factor The decision of which parameters are weighted and which are not, depends on the type of pa rameter. Given the user-defined value px t of such a weighted parameter for the processing object, for each audio object S , the parameter function p is defined as follows:
- FIG. 6 shows a processing object having region A and object radii according to one embodiment.
- a j is a constant for the closest possible distance to an audio object, and d, (t) is the distance from the audio object to the EQ object. Derived from the law of distance, the function has been changed to correctly handle any positive or negative EQ gain changes.
- Fig. 7 shows a relative angle of audio objects to the processing object according to one embodiment.
- Fig. 8 shows an equalizer object with a new radial radius according to an embodiment form.
- the distance d t in this context could simply be interpreted as the angle between the audio object and the EQ object, this would no longer justify applying the spacing law. Therefore, only the custom area is changed while maintaining the gain calculation as before.
- equalization is realized as the application.
- Equalization can be considered the most important tool in mastering, as the frequency response of a mix is the most critical factor for good translation across replay systems.
- dynamic control is realized as an application.
- FIG. 9 shows a signal flow of compression of the signal from n sources according to one embodiment.
- scene transformation is realized as an application.
- center / side processing is a commonly used technique for expanding or stabilizing the stereo image of a mix.
- center / side processing is a commonly used technique for expanding or stabilizing the stereo image of a mix.
- a similar option may be helpful if the mix was created in an acoustically critical environment with potentially asymmetric room or speaker characteristics. It could also provide new creative opportunities for the ME to enhance the impact of a mix.
- FIG. 10 shows a scene transformation using a control panel M according to an embodiment. Specifically, Fig. 10 shows a schematic conversion using a distortion range with user-settable edges C, to C £.
- a two-dimensional transformation of a scene in the horizontal plane can be realized using a homography transformation matrix H which maps each audio object at position p to a new position p r , see also [7]:
- dynamic equalizers are realized.
- Other embodiments realize multiband compression.
- Object-based sound adjustments are not limited to the introduced equalizer applications.
- Audio scenes describe the arrangement of audio objects on a time-dependent basis. Audio objects consist of audio signals and metadata. These metadata include, but are not limited to, Position in the room and volume. To edit the scene, the user has to change all the audio objects of a scene individually.
- processing object group and, on the other hand, "processing object”, it is to be noted that a processing object group is always defined for each processing object, which comprises audio objects.
- the processing object group is also referred to as the container of the processing object.
- a group of audio objects from the plurality of audio objects is defined corresponding processing object group comprises the group of audio objects thus specified.
- a processing object group is therefore a group of audio objects.
- Processing objects can be defined as objects that can change the properties of other audio objects.
- Processing objects are artificial containers to which any audio objects can be assigned, i. all its assigned audio objects are addressed via the container. Any number of effects affect the associated audio objects.
- processing objects provide the user with the ability to simultaneously manipulate multiple audio objects.
- a processing object includes, for example, position, assignment methods, containers, weighting methods, audio signal processing effects, and metadata effects.
- the position is a position of the processing object in a virtual scene.
- the mapping method assigns audio objects to the processing object (using their position if necessary).
- the container (or connections) is the set of all audio objects (or any additional other processing objects) associated with the processing object.
- Weighting methods are the algorithms for calculating the individual effect parameter values for the associated audio objects.
- Audio signal processing effects alter the audio component of audio objects (e.g., equalizer, dynamics).
- Metadata effects alter the metadata of audio objects and / or processing objects (e.g., positional distortion).
- the processing object group may be assigned the position described above, the mapping method, the container, weighting methods, audio signal processing effects, and metadata effects.
- the audio objects of the container of the processing object are the audio objects of the processing object group.
- FIG. 1 shows the relationship of a processing object with which audio signal effects and metadata effects are effected, according to one embodiment.
- Processing objects can be arbitrarily placed in a scene by the user, the position can be set constant or time-dependent over time.
- Processing objects can be assigned by the user with effects which change the audio signal and / or the metadata of audio objects. Examples of effects are equalization of the audio signal, processing the dynamics of the audio signal, or changing the position coordinates of audio objects.
- Processing objects can be populated with any number of effects in any order.
- Effects alter the audio signal and / or the metadata of the associated set of audio objects, either constant over time or time dependent.
- Effects have parameters for controlling signal and / or metadata processing. These parameters are divided into constant and weighted parameters by the user, or defined by type.
- the effects of a processing object are copied and applied to its associated audio objects.
- the values of constant parameters are adopted unchanged by each audio object.
- the values of weighted parameters are calculated individually for each audio object according to different weighting methods. The user can choose a weighting method for each effect, or enable or disable it for individual audio sources.
- the weighting procedures take into account individual metadata and / or
- Signal characteristics of individual audio objects This corresponds, for example, to the distance of an audio object to the processing object or the frequency spectrum of an audio object.
- the weighting methods may also take into account the listening position of the listener. Furthermore, the mentioned properties of audio objects for the weighting methods can also be combined with one another in order to produce individual ones
- the sound levels of audio objects can be added in the context of dynamic processing in order to individually derive a change in the volume for each audio object
- Effect parameters can be set constant over time or time-dependent.
- the weighting procedures take into account such temporal changes.
- Weighting methods may also process information that the audio renderer analyzes from the scene.
- the order of occupancy of the processing object with effects corresponds to the sequence of processing signals and / or metadata of each audio object, i. H. the data modified by a previous effect is used by the next effect as the basis for its calculation.
- the first effect works on the still unchanged data of an audio object.
- An explicitly newly developed effect is the change of the position of audio objects by means of homography ("distortion effect").
- the user is shown a rectangle with individually movable corners at the position of the processing object. If the user moves a corner, a transformation matrix for this distortion is calculated from the previous state of the rectangle and the newly distorted state. The matrix is then applied to all position coordinates of the audio objects associated with the processing object, so that their position changes according to the distortion.
- the assignment of audio sources to the processing objects can be done in various ways.
- the amount of associated audio objects may change over time depending on the nature of the assignment. This change is taken into account by all calculations.
- a catchment area can be defined around the position of processing objects All audio objects positioned within the catchment area form the assigned set of audio objects to which the effects of the processing object are applied.
- the catchment area can be any body (three-dimensional) or any shape (two-dimensional) that is defined by the user.
- the midpoint of the catchment area may or may not correspond to the position of the processing object. The user makes this determination.
- a three-dimensional catchment area Within a three-dimensional catchment area lies an audio object when its position lies within the three-dimensional body.
- an audio object lies when its position projected on the horizontal plane lies within the two-dimensional shape.
- the listening area may take on an unspecified overall size so that all the audio objects of a scene are in the catchment area.
- the catchment areas may be adapted to changes in scene properties (e.g., scene scaling).
- processing objects can be coupled to any selection of audio objects in a scene.
- the coupling can be defined by the user so that all selected audio objects form a set of audio objects to which the effects of the processing object are applied.
- the coupling may be defined by the user so that the processing object adjusts its position time-dependently to the position of the selected audio objects. This adjustment of the position may take into account the listener's listening position. The effects of the processing object do not necessarily have to be applied to the coupled audio objects
- the assignment can be made automatically based on criteria defined by the user. All the audio objects in a scene are continuously examined for the defined criterion (s) and assigned to the processing object when the criteria are met. net.
- the duration of the assignment may be limited to the time of fulfillment of the criteria or transitional periods may be defined. The transition periods determine how long one or more criteria must be continuously fulfilled by the audio object so that it is assigned to the processing object or how long one or more criteria must be continuously violated so that the assignment to the processing object is resolved again becomes.
- Processing objects can be deactivated by the user so that their properties are retained and continue to be displayed to the user, but no influencing of audio objects by the processing object takes place.
- any number of properties of a processing object can be coupled by the user with similar properties of any number of other processing objects. These features include parameters of effects.
- the coupling can be chosen absolutely or relatively by the user. With constant coupling, the modified property value of a processing object is copied exactly by all coupled processing objects. With relative coupling, the value of the change is offset against the property values of coupled processing objects. Processing objects can be duplicated. In this case, a second processing object is produced with identical properties of the original processing objects. The properties of the processing objects are then independent of each other.
- Properties of processing objects may e.g. be permanently inherited when copying, so that changes in the parents are automatically transferred to the children,
- FIG. 12 shows the change of audio objects and audio signals to an input of a user according to an embodiment.
- Another new application of processing objects is the intelligent parameter calculation by means of a scene analysis.
- the user defines effect parameters at a specific position via the processing object.
- the audio renderer does a predictive scene analysis to detect which audio sources influence the position of the processing object. Then, effects are applied to the selected audio sources, taking into account the scene analysis, so that the User-defined effect settings are best achieved at the position of the processing object.
- FIG. 13 shows processing object P0 4 with rectangle M for distortion of the corners C 1, C 2, C 3 and C 4 by the user.
- Fig. 13 shows schematically a possible Verzer tion towards M 'with the corners CY, C 2 ', C 3 'and C 4 ', and the corresponding effect on the sources S ,, S 2 , S 3 and S 4 with their new positions SG S 2 ', S 3 ' and S 4 '.
- Fig. 14 shows processing objects RO ⁇ and P0 2 with their respective, overlapping two-dimensional catchment areas A and B, and the distances ag r ag 2 and ag 3 and bs 3 , bg 4 and bg 6 from the respective processing object to those through the catchment areas associated sources Si, S 2 , S 3 , S 4 and S 6 -
- Fig. 15 shows processing object PO ; ⁇ with rectangular, two-dimensional Einzugsbe range C and the angles between P0 3 and the associated sources S, S 2 and S 3 for a possible weighting of parameters that includes the listening position of the listener.
- the angles can be determined by the difference of the azimuth of the individual sources and the azimuth a po of P0 3 .
- FIG. 16 shows a possible schematic implementation of an equalizer effect applied to a processing object. Using buttons like w next to each parameter, the weighting for the respective parameter can be activated. m, m 2 and m 3 provide options for the weighting method for the weighted parameters mentioned.
- FIG. 17 shows the processing object P0 5 with a three-dimensional catchment area D and the respective distances dg r dg 2 and dg 3 to the sources S 1 S 2 and S 3 assigned via the catchment area .
- FIG. 18 shows a prototype implementation of a processing object to which an equalizer has been applied.
- the turquoise object with the wave symbol on the right-hand side of the image shows the processing object in the audio scene, which the user can freely move with the mouse.
- the equalizer parameters as defined on the left side of the image are applied unchanged to the audio objects Src1, Src2 and Src3 Circular area, the transparent shading indicates the area in which all parameters except for the gain parameters are taken over unchanged from the sources.
- the gain parameters of the equalizer are weighted according to the distance of the source to the processing object.
- Source Src22 is not affected by the processing object.
- the user uses the “Area” slider to control the size of the radius of the circular area around the processing object. He uses the “feather” slider to control the size of the radius of the surrounding transition area.
- Fig. 19 shows a processing object as in Fig. 18, only at a different position and without a transition surface. All parameters of the Equalizer are taken over unchanged on the sources Src22 and Src4. The sources Src3, Src2, Src1 and Src24 are not affected by the processing object.
- Fig. 20 shows a processing object having a surface defined by its azimuth as a drawing region, so that the sources Src22 and Sre4 are assigned to the processing object.
- the top of the feed surface in the middle of the right-hand side of the image corresponds to the position of the listener / user.
- the area is moved according to the azimuth.
- the user determines the size of the angle of the feed surface.
- the change from a circular to angle-based feed surface is reached by the user via the lower selection field above the "Area” / "Feather” slider, now "radius". displays.
- Fig. 21 shows a processing object as in Fig. 20, but with additional transition area that can be controlled by the user via the "feather" slider.
- Fig. 22 shows several processing objects in the scene, with different catchment areas.
- the gray processing objects have been deactivated by the user, i. h They do not affect the audio objects in their catchment area.
- the left side of the screen always displays the equalizer parameters of the currently selected processing object. The selection is indicated by a thin, bright turquoise line around the object.
- Fig. 23 shows the red square on the right side of the image showing a processing object for horizontally distorting the position of audio objects.
- the user can drag the corners in any direction with the mouse to achieve a distortion of the scene
- Fig. 24 shows the scene after the user has moved the corners of the processing object. The position of all sources has changed according to the distortion.
- Fig. 25 shows a possible visualization of the assignment of individual audio objects to a processing object.
- aspects have been described in the context of a device, it should be understood that these aspects also constitute a description of the corresponding method, so that a block or device of a device is also to be understood as a corresponding method step or as a feature of a method step , Similarly, aspects described in connection with or as a method step also represent a description of a corresponding block or detail or feature of a corresponding device.
- Some or all of the method steps may be performed by a hardware device (or using hardware -Apparats), such as a microprocessor, a programmable coraputer or an electronic circuit can be performed. In some embodiments, some or more of the most important method steps may be performed by such an apparatus.
- embodiments of the invention may be implemented in hardware or in software, or at least partially in hardware, or at least partially in software.
- the implementation may be performed using a digital storage medium, such as a floppy disk, a DVD, a BluRay disc, a CD, a ROM, a PROM, an EPROM, an EEPROM or a FLASH memory, a hard disk or other magnetic or optical memory are stored on the electronically readable control signals that can cooperate with or cooperate with a programmable computer system such that the respective method is performed. Therefore, the digital storage medium can be computer readable.
- some embodiments according to the invention include a data carrier having electronically readable control signals capable of interacting with a programmable computer system such that one of the methods described herein is performed.
- embodiments of the present invention may be implemented as a computer program product having a program code, wherein the program code is effective to perform one of the methods when the computer program product runs on a computer.
- the program code can also be stored on a machine-readable carrier, for example.
- an embodiment of the method according to the invention is thus a computer program which has a program code for carrying out one of the methods described herein when the computer program runs on a computer.
- a further embodiment of the inventive method is thus a data carrier (or a digital storage medium or a computer-readable medium) on which the computer program for performing one of the methods described herein is recorded.
- the data carrier or the digital storage medium or the computer-readable medium are typically tangible and / or non-volatile.
- a further exemplary embodiment of the method according to the invention is thus a data stream or a sequence of signals which represents the computer program for performing one of the methods described herein.
- the data stream or the sequence of signals may be configured, for example, to be transferred via a data communication connection, for example via the Internet.
- Another embodiment includes a processing device, such as a computer or programmable logic device, configured or adapted to perform one of the methods described herein.
- a processing device such as a computer or programmable logic device, configured or adapted to perform one of the methods described herein.
- Another embodiment includes a computer on which the computer program is installed to perform one of the methods described herein.
- Another embodiment according to the invention comprises a device or system adapted to transmit a computer program for performing at least one of the methods described herein to a receiver.
- the transmission can be carried out, for example, electronically or optically.
- the receiver can be, for example, a computer, a mobile device, a storage device or a similar device. be direction.
- the device or system may include a file server for transmitting the computer program to the recipient.
- a programmable logic device eg, a field programmable gate array, an FPGA
- a field programmable gate array may cooperate with a microprocessor to perform any of the methods described herein.
- the methods are performed by any flardware device. This may be a universally applicable flardware such as a computer processor (CPU) or hardware specific to the process, such as an ASIC.
- SAOC Spatial Audio Object Coding
Landscapes
- Engineering & Computer Science (AREA)
- Physics & Mathematics (AREA)
- Signal Processing (AREA)
- Acoustics & Sound (AREA)
- Computational Linguistics (AREA)
- Health & Medical Sciences (AREA)
- Audiology, Speech & Language Pathology (AREA)
- Human Computer Interaction (AREA)
- Multimedia (AREA)
- Mathematical Physics (AREA)
- Stereophonic System (AREA)
Abstract
Applications Claiming Priority (3)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
DE102018202511 | 2018-02-19 | ||
DE102018206025.8A DE102018206025A1 (de) | 2018-02-19 | 2018-04-19 | Vorrichtung und Verfahren für objektbasiertes, räumliches Audio-Mastering |
PCT/EP2019/053961 WO2019158750A1 (fr) | 2018-02-19 | 2019-02-18 | Dispositif et procédé pour matriçage audio spatial à base d'objet |
Publications (1)
Publication Number | Publication Date |
---|---|
EP3756363A1 true EP3756363A1 (fr) | 2020-12-30 |
Family
ID=67482130
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
EP19710283.3A Pending EP3756363A1 (fr) | 2018-02-19 | 2019-02-18 | Dispositif et procédé pour matriçage audio spatial à base d'objet |
Country Status (8)
Country | Link |
---|---|
US (1) | US20200374649A1 (fr) |
EP (1) | EP3756363A1 (fr) |
JP (2) | JP7277475B2 (fr) |
KR (3) | KR102654354B1 (fr) |
CA (1) | CA3091529A1 (fr) |
DE (1) | DE102018206025A1 (fr) |
SG (1) | SG11202105234TA (fr) |
WO (1) | WO2019158750A1 (fr) |
Families Citing this family (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
GB2594265A (en) * | 2020-04-20 | 2021-10-27 | Nokia Technologies Oy | Apparatus, methods and computer programs for enabling rendering of spatial audio signals |
US11579838B2 (en) * | 2020-11-26 | 2023-02-14 | Verses, Inc. | Method for playing audio source using user interaction and a music application using the same |
EP4396810A1 (fr) * | 2021-09-03 | 2024-07-10 | Dolby Laboratories Licensing Corporation | Synthétiseur de musique à sortie de métadonnées spatiales |
DE112022004952T5 (de) * | 2021-10-15 | 2024-08-14 | Sony Group Corporation | Informationsverarbeitungseinrichtung, verfahren und programm |
WO2024038925A1 (fr) * | 2022-08-18 | 2024-02-22 | 엘지전자 주식회사 | Téléviseur et procédé de commande de téléviseur |
Family Cites Families (7)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
KR101111520B1 (ko) | 2006-12-07 | 2012-05-24 | 엘지전자 주식회사 | 오디오 처리 방법 및 장치 |
US20100223552A1 (en) * | 2009-03-02 | 2010-09-02 | Metcalf Randall B | Playback Device For Generating Sound Events |
DE102010030534A1 (de) * | 2010-06-25 | 2011-12-29 | Iosono Gmbh | Vorrichtung zum Veränderung einer Audio-Szene und Vorrichtung zum Erzeugen einer Richtungsfunktion |
KR102003191B1 (ko) * | 2011-07-01 | 2019-07-24 | 돌비 레버러토리즈 라이쎈싱 코오포레이션 | 적응형 오디오 신호 생성, 코딩 및 렌더링을 위한 시스템 및 방법 |
US9516446B2 (en) * | 2012-07-20 | 2016-12-06 | Qualcomm Incorporated | Scalable downmix design for object-based surround codec with cluster analysis by synthesis |
EP3264259A1 (fr) * | 2016-06-30 | 2018-01-03 | Nokia Technologies Oy | Manipulation de volume audio |
US9980078B2 (en) * | 2016-10-14 | 2018-05-22 | Nokia Technologies Oy | Audio object modification in free-viewpoint rendering |
-
2018
- 2018-04-19 DE DE102018206025.8A patent/DE102018206025A1/de active Pending
-
2019
- 2019-02-18 JP JP2020543912A patent/JP7277475B2/ja active Active
- 2019-02-18 SG SG11202105234TA patent/SG11202105234TA/en unknown
- 2019-02-18 CA CA3091529A patent/CA3091529A1/fr active Pending
- 2019-02-18 KR KR1020227026195A patent/KR102654354B1/ko active IP Right Grant
- 2019-02-18 EP EP19710283.3A patent/EP3756363A1/fr active Pending
- 2019-02-18 KR KR1020247010821A patent/KR20240052967A/ko active Search and Examination
- 2019-02-18 WO PCT/EP2019/053961 patent/WO2019158750A1/fr active Search and Examination
- 2019-02-18 KR KR1020207026941A patent/KR102427809B1/ko active IP Right Grant
-
2020
- 2020-08-11 US US16/990,835 patent/US20200374649A1/en active Pending
-
2023
- 2023-01-13 JP JP2023004163A patent/JP2023055736A/ja active Pending
Also Published As
Publication number | Publication date |
---|---|
JP7277475B2 (ja) | 2023-05-19 |
WO2019158750A1 (fr) | 2019-08-22 |
SG11202105234TA (en) | 2021-06-29 |
KR102654354B1 (ko) | 2024-04-03 |
KR20240052967A (ko) | 2024-04-23 |
US20200374649A1 (en) | 2020-11-26 |
JP2023055736A (ja) | 2023-04-18 |
CA3091529A1 (fr) | 2019-08-22 |
KR20220108218A (ko) | 2022-08-02 |
JP2021514164A (ja) | 2021-06-03 |
KR102427809B1 (ko) | 2022-08-01 |
KR20200120734A (ko) | 2020-10-21 |
DE102018206025A1 (de) | 2019-08-22 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
EP3756363A1 (fr) | Dispositif et procédé pour matriçage audio spatial à base d'objet | |
EP1652405B1 (fr) | Dispositif et procede de production, de mise en memoire ou de traitement d'une representation audio d'une scene audio | |
EP1637012B1 (fr) | Dispositif de synthese de champ electromagnetique et procede d'actionnement d'un reseau de haut-parleurs | |
EP2080411B1 (fr) | Dispositif et procédé pour produire un certain nombre de signaux de haut-parleur pour un réseau de haut-parleurs définissant un espace de restitution | |
EP1844627B1 (fr) | Dispositif et procédé pour simuler un système de synthèse de champ d'onde | |
EP1844628B1 (fr) | Procede et dispositif d'amorçage d'une installation de moteur de rendu de synthese de front d'onde avec objets audio | |
DE69839212T2 (de) | Raumklangwiedergabe | |
EP1671516B1 (fr) | Procede et dispositif de production d'un canal a frequences basses | |
DE10254404B4 (de) | Audiowiedergabesystem und Verfahren zum Wiedergeben eines Audiosignals | |
DE102010030534A1 (de) | Vorrichtung zum Veränderung einer Audio-Szene und Vorrichtung zum Erzeugen einer Richtungsfunktion | |
EP1525776B1 (fr) | Dispositif de correction de niveau dans un systeme de synthese de champ d'ondes | |
EP3020042B1 (fr) | Traitement de métadonnées à variation temporelle pour un ré-échantillonnage sans perte | |
EP1723825B1 (fr) | Dispositif et procede pour reguler un dispositif de rendu de synthese de champ electromagnetique | |
DE10321980B4 (de) | Vorrichtung und Verfahren zum Berechnen eines diskreten Werts einer Komponente in einem Lautsprechersignal | |
WO2007101498A1 (fr) | Dispositif et procédé de simulation de systèmes wfs et de compensation de propriétés wfs influençant le son | |
Breebaart et al. | Spatial coding of complex object-based program material | |
Hestermann et al. | An Approach for Mastering Audio Objects |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
STAA | Information on the status of an ep patent application or granted ep patent |
Free format text: STATUS: UNKNOWN |
|
STAA | Information on the status of an ep patent application or granted ep patent |
Free format text: STATUS: THE INTERNATIONAL PUBLICATION HAS BEEN MADE |
|
PUAI | Public reference made under article 153(3) epc to a published international application that has entered the european phase |
Free format text: ORIGINAL CODE: 0009012 |
|
STAA | Information on the status of an ep patent application or granted ep patent |
Free format text: STATUS: REQUEST FOR EXAMINATION WAS MADE |
|
17P | Request for examination filed |
Effective date: 20200730 |
|
AK | Designated contracting states |
Kind code of ref document: A1 Designated state(s): AL AT BE BG CH CY CZ DE DK EE ES FI FR GB GR HR HU IE IS IT LI LT LU LV MC MK MT NL NO PL PT RO RS SE SI SK SM TR |
|
AX | Request for extension of the european patent |
Extension state: BA ME |
|
DAV | Request for validation of the european patent (deleted) | ||
DAX | Request for extension of the european patent (deleted) | ||
REG | Reference to a national code |
Ref country code: HK Ref legal event code: DE Ref document number: 40043106 Country of ref document: HK |
|
RAP3 | Party data changed (applicant data changed or rights of an application transferred) |
Owner name: FRAUNHOFER-GESELLSCHAFT ZUR FOERDERUNG DER ANGEWANDTEN FORSCHUNG E.V. |
|
STAA | Information on the status of an ep patent application or granted ep patent |
Free format text: STATUS: EXAMINATION IS IN PROGRESS |
|
17Q | First examination report despatched |
Effective date: 20220607 |