WO2014199536A1 - オーディオ再生装置及びその方法 - Google Patents
オーディオ再生装置及びその方法 Download PDFInfo
- Publication number
- WO2014199536A1 WO2014199536A1 PCT/JP2014/000868 JP2014000868W WO2014199536A1 WO 2014199536 A1 WO2014199536 A1 WO 2014199536A1 JP 2014000868 W JP2014000868 W JP 2014000868W WO 2014199536 A1 WO2014199536 A1 WO 2014199536A1
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- position information
- audio
- reproduction
- coordinate
- axis
- Prior art date
Links
Images
Classifications
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04R—LOUDSPEAKERS, MICROPHONES, GRAMOPHONE PICK-UPS OR LIKE ACOUSTIC ELECTROMECHANICAL TRANSDUCERS; DEAF-AID SETS; PUBLIC ADDRESS SYSTEMS
- H04R1/00—Details of transducers, loudspeakers or microphones
- H04R1/20—Arrangements for obtaining desired frequency or directional characteristics
- H04R1/32—Arrangements for obtaining desired frequency or directional characteristics for obtaining desired directional characteristic only
- H04R1/40—Arrangements for obtaining desired frequency or directional characteristics for obtaining desired directional characteristic only by combining a number of identical transducers
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04R—LOUDSPEAKERS, MICROPHONES, GRAMOPHONE PICK-UPS OR LIKE ACOUSTIC ELECTROMECHANICAL TRANSDUCERS; DEAF-AID SETS; PUBLIC ADDRESS SYSTEMS
- H04R5/00—Stereophonic arrangements
- H04R5/04—Circuit arrangements, e.g. for selective connection of amplifier inputs/outputs to loudspeakers, for loudspeaker detection, or for adaptation of settings to personal preferences or hearing impairments
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04R—LOUDSPEAKERS, MICROPHONES, GRAMOPHONE PICK-UPS OR LIKE ACOUSTIC ELECTROMECHANICAL TRANSDUCERS; DEAF-AID SETS; PUBLIC ADDRESS SYSTEMS
- H04R1/00—Details of transducers, loudspeakers or microphones
- H04R1/20—Arrangements for obtaining desired frequency or directional characteristics
- H04R1/32—Arrangements for obtaining desired frequency or directional characteristics for obtaining desired directional characteristic only
- H04R1/40—Arrangements for obtaining desired frequency or directional characteristics for obtaining desired directional characteristic only by combining a number of identical transducers
- H04R1/403—Arrangements for obtaining desired frequency or directional characteristics for obtaining desired directional characteristic only by combining a number of identical transducers loud-speakers
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04R—LOUDSPEAKERS, MICROPHONES, GRAMOPHONE PICK-UPS OR LIKE ACOUSTIC ELECTROMECHANICAL TRANSDUCERS; DEAF-AID SETS; PUBLIC ADDRESS SYSTEMS
- H04R3/00—Circuits for transducers, loudspeakers or microphones
- H04R3/12—Circuits for transducers, loudspeakers or microphones for distributing signals to two or more loudspeakers
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04S—STEREOPHONIC SYSTEMS
- H04S7/00—Indicating arrangements; Control arrangements, e.g. balance control
- H04S7/30—Control circuits for electronic adaptation of the sound field
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04S—STEREOPHONIC SYSTEMS
- H04S7/00—Indicating arrangements; Control arrangements, e.g. balance control
- H04S7/30—Control circuits for electronic adaptation of the sound field
- H04S7/308—Electronic adaptation dependent on speaker or headphone connection
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04R—LOUDSPEAKERS, MICROPHONES, GRAMOPHONE PICK-UPS OR LIKE ACOUSTIC ELECTROMECHANICAL TRANSDUCERS; DEAF-AID SETS; PUBLIC ADDRESS SYSTEMS
- H04R2203/00—Details of circuits for transducers, loudspeakers or microphones covered by H04R3/00 but not provided for in any of its subgroups
- H04R2203/12—Beamforming aspects for stereophonic sound reproduction with loudspeaker arrays
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04S—STEREOPHONIC SYSTEMS
- H04S2400/00—Details of stereophonic systems covered by H04S but not provided for in its groups
- H04S2400/03—Aspects of down-mixing multi-channel audio to configurations with lower numbers of playback channels, e.g. 7.1 -> 5.1
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04S—STEREOPHONIC SYSTEMS
- H04S2400/00—Details of stereophonic systems covered by H04S but not provided for in its groups
- H04S2400/11—Positioning of individual sound objects, e.g. moving airplane, within a sound field
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04S—STEREOPHONIC SYSTEMS
- H04S2420/00—Techniques used stereophonic systems covered by H04S but not provided for in its groups
- H04S2420/13—Application of wave-field synthesis in stereophonic audio systems
Definitions
- the present disclosure relates to an apparatus and a method for reproducing an audio object using a speaker array.
- the present invention relates to an apparatus and method for reproducing an audio object including reproduction position information indicating a position of a sound image to be localized in a three-dimensional space.
- 5.1ch is a channel setting in which a front left / right channel, a front center channel, and a surround channel are arranged on the left and right.
- some recent Blu-ray (registered trademark) players have a 7.1 channel configuration in which left and right channels are added to the back surround.
- FIG. 14 shows a speaker arrangement in 22.2ch audio reproduction currently being researched and developed by NHK (Japan Broadcasting Corporation). Unlike the conventional arrangement of the speaker only on the two-dimensional plane (corresponding to the middle stage in FIG. 14), it has a three-dimensional configuration in which the speaker is also arranged on the foot (lower stage) and the ceiling (upper stage) (non-patent document). 1).
- Non-patent Document 2 the speaker has a 3D (three-dimensional) configuration that is also arranged on the ceiling.
- the content is encoded as an audio object.
- An audio object is an audio signal accompanied by reproduction position information indicating a position of a sound image to be localized in a three-dimensional space. For example, it is a signal encoded as a set of reproduction position information representing the position where the sound source (sound image) is localized on three axes (x, y, z) and the audio signal of the sound source.
- the reproduction position information may be vector information representing the transition direction.
- the reproduction position information is constant.
- HRTF Head Related Transfer Function
- HRTF is a transfer function that simulates the propagation characteristics of sound around a human head.
- the perception of which direction the sound is heard from is said to be affected by HRTF, and as shown in FIG. 15, mainly the difference in sound pressure between both ears and the time difference between sound waves reaching between both ears. Is influenced by.
- the direction in which sound is heard can be controlled by artificially controlling it by signal processing. Details are described in Non-Patent Document 3.
- clues related to localization in the front-rear and up-down directions are included in the amplitude spectrum of HRTF. Details are described in Patent Document 1.
- the basic operation principle of the beamform is as shown in FIG. Similar to the wavefront synthesis, the sound pressure level at a specific position can be made higher than the surrounding area by using a speaker array and appropriately controlling the sound pressure and phase in the beam form as well. As a result, the state as if the sound source exists at that position can be reproduced. Details are described in Non-Patent Document 5.
- HRTF wavefront synthesis
- beamform as methods for enhancing the sense of presence of sound even when speakers cannot be freely arranged.
- HRTF is an excellent method for controlling the direction in which sound is heard.
- wavefront synthesis and beamform can reproduce the actual physical wavefront, so that the distance between the listener and the sound source can be reproduced, but a sound source cannot be generated behind the listener. This is because the sound wave reaches the listener before the sound wave output from the speaker array forms a sound image.
- reproduction position information included in an audio object is expressed as three-dimensional spatial information.
- the signal processing reflecting the reproduction position information cannot be performed.
- the present disclosure has been made in view of such a conventional problem, and is an audio that can reproduce an audio object including three-dimensional reproduction position information with high presence even in a space where speakers cannot be freely arranged. It is an object of the present invention to provide a playback apparatus and method.
- an audio playback apparatus plays back an audio object including an audio signal and playback position information indicating a position in a three-dimensional space where a sound image of the audio signal is localized.
- An apparatus that converts at least one speaker array that converts an acoustic signal into acoustic vibration and the reproduction position information into corrected reproduction position information that is position information on a two-dimensional coordinate axis based on the position of the speaker array.
- a conversion unit that converts the signal, and a signal processing unit that performs a process of localizing a sound image of the audio signal included in the audio object according to the modified reproduction position information.
- the three-dimensional reproduction position information included in the audio object is converted into the corrected reproduction position information on the two-dimensional coordinate axis based on the position of the speaker array, and a sound image is generated according to the corrected reproduction position information after correction. Since the sound is localized, the audio object can be reproduced with a high sense of presence even when the speaker arrangement is limited.
- the arrangement direction of the speaker elements constituting the speaker array is the X axis
- the Y axis is the direction perpendicular to the X axis
- the installation surface that is the surface on which the speaker array is installed
- the corrected reproduction position information indicates a position on a coordinate axis composed of the X axis and the Y axis
- the modified reproduction position information may be a value corresponding to the x, y.
- the corrected reproduction position information becomes a value corresponding to the x, y, and thus a space in which speakers cannot be arranged in three dimensions. Even so, an audio object including three-dimensional reproduction position information can be reproduced with a high sense of presence.
- the back direction of the speaker array is a minus coordinate
- the front direction of the speaker array is a plus coordinate
- the center of the speaker array is shifted to the left and right.
- the value of the modified reproduction position information may be a value obtained by multiplying at least one of the x and y by a predetermined value.
- the value of the corrected reproduction position information is a value obtained by multiplying the above x and y by a predetermined value, it is possible to virtually change the perceived space.
- the x coordinate value of the modified reproduction position information may be limited to the width of the speaker array.
- the x-coordinate value of the corrected reproduction position information is a value limited by the width of the speaker array, signal processing suitable for the performance of the speaker array can be performed.
- the signal processing unit may be a beamform unit that forms a sound image at a position on the two-dimensional coordinate axis.
- the strong acoustic vibration is generated at the target position by the beamform unit, it is possible to generate a sound field as if a sound source exists there.
- the back direction of the speaker array is a negative coordinate
- the front direction of the speaker array is a positive coordinate
- the signal processing unit has a negative y coordinate value of the modified reproduction position information. If the value is, wavefront synthesis may be performed by signal processing using the Huygens theorem.
- the wavefront is synthesized by signal processing using the Huygens theorem, so that the target position of the sound image to be localized is the back of the speaker
- the modified playback position information includes a direction angle to a position indicated by the playback position information viewed from a position of a listener who listens to sound output from the speaker array, and the playback position information from the position of the listener.
- the position on the two-dimensional coordinate axis may be indicated by the distance to the position indicated by.
- the corrected reproduction position information is represented on the two-dimensional coordinate axis by the direction angle to the position indicated by the reproduction position information viewed from the listener's position and the distance from the listener's position to the position indicated by the reproduction position information. Therefore, it is possible to control which direction and distance the sound source exists when listening from the listener.
- the signal processing unit performs a process of localizing the sound image using an HRTF (Head Related Transfer Function), and the HRTF is set so that sound can be heard from a position indicated by the modified reproduction position information. Also good.
- HRTF Head Related Transfer Function
- the sound image is localized using the HRTF set so that the sound can be heard from the position indicated by the corrected reproduction position information, the direction to the sound source when listening from the listener is reflected. Can play.
- the signal processing unit may adjust the volume according to the distance between the position of the listener and the position indicated by the modified playback position information.
- the volume is adjusted according to the distance between the position of the listener and the position indicated by the modified playback position information, it is possible to reproduce the distance reflecting the distance to the sound source when the listener listens.
- the signal processing unit may change the signal processing method according to the position indicated by the modified reproduction position information.
- the signal processing method is changed according to the position indicated by the corrected reproduction position information, it is possible to select the optimum signal processing method according to the target reproduction position.
- the arrangement direction of the speaker elements constituting the speaker array is the X axis, the direction perpendicular to the X axis, and the installation surface, which is the surface where the speaker array is installed, is the Y axis, and
- the back direction of the speaker array is a negative coordinate
- the front direction of the speaker array is a positive coordinate.
- the signal processing unit performs wavefront synthesis by signal processing using the Huygens theorem, and receives the y coordinate value of the modified reproduction position information.
- HR F may be localizing a sound image at the signal processing using the.
- the y-coordinate value of the corrected playback position information is a negative value
- wavefront synthesis is performed by signal processing using Huygens' theorem
- the y-coordinate value of the corrected playback position information is calculated from the listener's position. If it is a positive value before, a sound image is generated by signal processing using a beamform, and if the y-coordinate value of the corrected reproduction position information is a positive value behind the listener's position, a signal using HRTF Since the sound image is localized in the process, an acoustic vibration is generated as if there is a sound source at the target position in front of the listener's position, and perceived from the direction perceptually behind the listener's position. You can play the sound.
- the audio reproduction device includes at least two speaker arrays, the at least two speaker arrays constitute at least two two-dimensional coordinates, and one speaker array of the at least two speaker arrays is provided.
- the arrangement direction of the speaker elements constituting the X-axis is orthogonal to the X-axis, and the horizontal direction to the installation surface, which is the surface on which the one speaker array is installed, is orthogonal to the Y-axis and the X-axis.
- the signal processing unit is responsive to the value of z
- the at least two speaker arrays may be controlled, and when the two two-dimensional coordinates are parallel, the signal processing unit may determine that the value of z is greater than a predetermined value.
- the volume of the speaker array constituting the upper two-dimensional coordinates with respect to the installation surface is increased and the value of z is smaller than a predetermined value
- the lower two-dimensional coordinates with respect to the installation surface The volume of the speaker array constituting the coordinates may be increased, and when the two two-dimensional coordinates are orthogonal, the signal processing unit has a value of z greater than a predetermined value.
- the volume of the speaker element above the predetermined position is increased, and the value of z Is smaller than a predetermined value, among the speaker elements constituting the speaker array constituting the two-dimensional coordinates perpendicular to the installation surface, the speaker elements below the predetermined position Increase volume Comb may be.
- the audio playback device is provided with at least two speaker arrays, and at least two speaker arrays are provided according to the value of z at the position (x, y, z) specified by the playback position information. Since it is controlled, the height information of the reproduction position information can be controlled, and the audio object including the three-dimensional reproduction position information is reproduced with a high sense of presence.
- the audio reproduction apparatus reproduces an audio object including an audio signal and reproduction position information indicating a position in a three-dimensional space where a sound image of the audio signal is localized, and the audio object includes the audio object at predetermined time intervals.
- the audio frame included in the audio object may be reproduced by using it as the reproduction position information of the audio frame lacking the reproduction position information.
- the reproduction position information included in the audio frame reproduced in the past is used as the reproduction position information of the audio frame. Even when the audio object is present, it is possible to reproduce a natural sound field, or to reduce the amount of information when the audio object is recorded or transmitted when the audio object is not moving.
- an audio object including three-dimensional reproduction position information can be reproduced with high presence even in a space where speakers cannot be freely arranged.
- FIG. 1 is a diagram illustrating a configuration of an audio reproduction device according to an embodiment.
- FIG. 2 is a diagram illustrating a configuration of an audio object.
- FIG. 3 is a diagram illustrating an example of the shape of the speaker array.
- FIG. 4A is a diagram illustrating a relationship between a speaker array and a two-dimensional coordinate axis.
- FIG. 4B is a diagram illustrating a relationship between another type of speaker array and a two-dimensional coordinate axis.
- FIG. 5 is a diagram showing the relationship between the three-dimensional reproduction position information and the corrected reproduction position information (x, y).
- FIG. 6 is a diagram showing the relationship between the three-dimensional reproduction position information and the corrected reproduction position information (direction, distance).
- FIG. 1 is a diagram illustrating a configuration of an audio reproduction device according to an embodiment.
- FIG. 2 is a diagram illustrating a configuration of an audio object.
- FIG. 3 is a diagram illustrating an example of the shape of the speaker array.
- FIG. 7 is a diagram illustrating the relationship between the corrected reproduction position information and the signal processing method.
- FIG. 8 is a flowchart showing the main operation of the audio playback apparatus of the present embodiment.
- FIG. 9 is a flowchart showing an operation relating to handling of playback position information included in an audio frame, among the operations of the audio playback device of the present embodiment.
- FIG. 10 is a diagram illustrating the relationship between the position of the audio object and the signal processing method.
- FIG. 11 is a diagram illustrating a signal processing method when an audio object passes overhead.
- FIG. 12 is a diagram showing a variation of the embodiment using two speaker arrays.
- FIG. 13 is a diagram showing a variation of the embodiment using three speaker arrays.
- FIG. 14 is a diagram illustrating an example of a 22.2 ch speaker arrangement in the prior art.
- FIG. 15 is a diagram showing the principle of HRTF in the prior art.
- FIG. 16 is a diagram showing the principles of wavefront synthesis and beamform in the prior art.
- FIG. 1 is a diagram showing a configuration of an audio playback device 110 according to the present embodiment.
- the audio reproduction device 110 is an audio reproduction device that reproduces an audio object including an audio signal (here, an audio encoded signal) and reproduction position information indicating a position in a three-dimensional space where a sound image of the audio signal is localized.
- an audio object separation unit 100 is a processing unit that separates reproduction position information and an audio encoded signal from an audio object composed of reproduction position information and an audio encoded signal.
- the setting unit 101 is a processing unit that virtually sets a two-dimensional coordinate axis (that is, a two-dimensional coordinate axis based on the position of the speaker array 106) according to the position where the speaker array 106 is installed.
- the conversion unit 102 is a processing unit that converts the reproduction position information separated by the audio object separation unit 100 into corrected reproduction position information that is position information (two-dimensional information) on the two-dimensional coordinate axis set by the setting unit 101. is there.
- the selection unit 103 receives the corrected reproduction position information generated by the conversion unit 102, the two-dimensional coordinate axis set by the setting unit 101, and the position of the listener who listens to the sound output from the speaker array 106 (or this audio). This is a processing unit that selects a signal processing method to be adopted by the signal processing unit 105 from the listening position planned by the playback apparatus 110.
- the decoding unit 104 is a processing unit that decodes the audio encoded signal separated by the audio object separation unit 100 and generates an audio signal (acoustic signal).
- the signal processing unit 105 is a processing unit that performs processing to localize the sound image of the audio signal obtained by decoding in the decoding unit 104 according to the modified reproduction position obtained by conversion in the conversion unit 102.
- processing is performed by the signal processing method selected by the selection unit 103.
- the speaker array 106 is at least one speaker array (a group of a plurality of speaker elements arranged in a line) that converts an output signal (acoustic signal) from the signal processing unit into acoustic vibration.
- the audio object separation unit 100, the setting unit 101, the conversion unit 102, the selection unit 103, the decoding unit 104, and the signal processing unit 105 are typically implemented in hardware by an electronic circuit such as a semiconductor integrated circuit. However, it may be realized as software by a program executed by a computer including a CPU, a ROM, a RAM, and the like.
- the audio object separation unit 100 separates an audio object composed of reproduction position information and an audio encoded signal into reproduction position information and an audio encoded signal.
- the audio object has, for example, a configuration as shown in FIG. That is, an audio object is a set of an audio encoded signal and reproduction position information indicating a position in a three-dimensional space where a sound image of the audio encoded signal is localized. Such information (audio encoded signal and reproduction position information) is encoded in units of audio frames at a predetermined time interval to constitute an audio object.
- the reproduction position information is three-dimensional information (information indicating a position in a three-dimensional space) on the premise that a speaker is also arranged on the ceiling.
- the playback position information does not necessarily have to be inserted in units of all audio frames, and audio frames that are missing are included in audio frames that have been played back by the audio object separation unit 100 in the past. Playback position information is used. Such reuse of reproduction position information can be realized by using a storage unit included in the audio reproduction apparatus 110.
- the audio object separation unit 100 extracts reproduction position information and an audio encoded signal from the audio object as shown in FIG.
- the setting unit 101 virtually sets a two-dimensional coordinate axis according to the position where the speaker array 106 is installed.
- An overview of the speaker array 106 is, for example, as shown in FIG. That is, a plurality of speaker elements are arranged.
- the setting unit 101 virtually sets a two-dimensional coordinate axis (a two-dimensional coordinate axis based on the position of the speaker array) according to the position where the speaker array 106 is installed.
- the setting unit 101 sets the direction in which the speaker elements constituting the speaker array 106 are arranged to be X-axis, orthogonal to the X-axis, and the direction parallel to the installation surface where the speaker array 106 is installed is Y.
- An XY plane as an axis is set as a two-dimensional coordinate axis.
- the back direction of the speaker array 106 is a negative coordinate
- the front direction of the speaker array 106 is a positive coordinate.
- the center of the speaker array 106 is shifted to the left and right, respectively.
- the speaker arrays do not need to be arranged in a straight line, and may be arranged in an arch shape, for example, as shown in FIG. 4B.
- each individual speaker unit (speaker element) is drawn so as to face the front, but this is not always necessary, and the angle is adjusted so that each individual speaker unit (speaker element) faces radially. May be arranged.
- the conversion unit 102 converts the three-dimensional reproduction position information into corrected reproduction position information that is two-dimensional information.
- the reproduction position information is originally two-dimensional composed of the X axis and the Y axis. It is mapped to a position in a three-dimensional coordinate having a Z axis perpendicular to the coordinates (that is, the installation surface).
- the position indicated by the reproduction position information after mapping is (x1, y1, z1).
- the conversion unit 102 converts this position information into two-dimensional position information and generates corrected reproduction position information.
- the conversion from the three-dimensional reproduction position information to the two-dimensional corrected reproduction position information is performed, for example, by a method as shown in FIG.
- the position indicated by the reproduction position information of the audio object 1 is (x1, y1, z1) like the audio object 1
- the position indicated by the corresponding corrected reproduction position information is (x1, y1).
- the position indicated by the modified playback position information corresponds to the position (x2, y2, z2) indicated by the playback position information, like the audio object 2, but is not necessarily the same position (x2, Y2) as the X coordinate value and the Y coordinate value. It may not be y2).
- At least one of the X coordinate value and the Y coordinate value is multiplied by one or more values ⁇ (predetermined value) like the position (x2, y2 * ⁇ ) indicated by the modified reproduction position information 2 shown in FIG.
- ⁇ predetermined value
- a larger acoustic space may be produced by setting a value larger than the value actually designated by the reproduction position information.
- the X-axis coordinate may be multiplied by a value ⁇ (predetermined value) smaller than 1 in accordance with restrictions on the width of the speaker array 106 (not shown in FIG. 5). That is, the X coordinate value may be limited to the width of the speaker array 106 (may be a value within the range of the width of the speaker array 106).
- the method shown in FIG. 6 may be used as another method for converting the three-dimensional reproduction position information into the two-dimensional corrected reproduction position information. That is, the corrected reproduction position information may be converted into information on the direction and distance of the audio object (position indicated by the reproduction position information) viewed from the listener. That is, the corrected playback position information includes the direction angle to the position indicated by the playback position information viewed from the position of the listener who listens to the sound output from the speaker array 106, and the position indicated by the playback position information from the position of the listener. It may be polar coordinates indicating the distance.
- the reproduction position information of the audio object 1 is (x1, y1, z1), the direction angle ⁇ 1 to the position (x1, y1, z1) viewed from the listener's position, and the listener Assuming that the distance r1 from the position to the position (x1, y1, z1), the corresponding corrected reproduction position information 1 is represented by ( ⁇ 1, r1 ′).
- r1 ' is a value determined depending on r1.
- the reproduction position information of the audio object 2 is (x2, y2, z2)
- the direction angle ⁇ 2 from the listener's position to the position (x2, y2, z2) Assuming that the distance r2 from the listener's position to the position (x2, y2, z2) is the corrected reproduction position information 2 corresponding to the distance r2, ( ⁇ 2, r2 ′) is represented.
- r2 ' is a value determined depending on r2. This (representation of the corrected reproduction position information in polar coordinates) is that when HRTF is used as the sound image localization method, the filter coefficient of HRTF is set based on the direction information from the listener, so that signal processing is facilitated.
- r1 ′ is determined according to r1, but is controlled so that ⁇ 1 is closer to 0 ° and closer to r1, and ⁇ 1 is closer to 90 ° and smaller than r1. Also good.
- the signal processing unit 105 may perform a process of setting a sound image using HRTF that is set so that sound can be heard from the direction indicated by the corrected reproduction position information. Thereby, it is possible to control in which direction and distance the sound source exists when listening from the listener. Furthermore, the signal processing unit 105 may adjust the volume according to the distance (r1 ', r2', etc.) between the position of the listener and the position indicated by the modified playback position information. Thereby, the reproduction reflecting the distance to the sound source when listening from the listener can be performed.
- the selection unit 103 selects the corrected reproduction position information generated by the conversion unit 102, the two-dimensional coordinate axis set by the setting unit 101, and the position of the listener (or the listening position determined in advance by the audio reproduction device 110).
- the signal processing method to be adopted by the signal processing unit 105 is selected. An example is shown in FIG. For example, for the audio object 1 (when the y coordinate value of the corrected playback position information is a positive value before the listener's position), a sound image is synthesized with the beam form at the position of the corrected playback position information 1. This is because when the sound source reproduction position is in front of the speaker array 106 and in front of the listener, a sound image can be formed by the beam form.
- wavefront synthesis is performed based on the Huygens principle using the position of the corrected playback position information 2 as a sound source. . This is because when the sound source reproduction position is behind the speaker array 106, an acoustic effect can be created as if the sound source exists there by wavefront synthesis.
- the audio object 3 when the y-coordinate value of the corrected playback position information is a positive value behind the listener's position, the sound from the direction ( ⁇ 1) indicated by the corrected playback position information 3 Sound image localization as if the sound is heard is realized using HRFT.
- the method of using HRFT is selected because the beamform and wavefront synthesis are not effective when the sound source playback position is behind the listener.
- HRTF the direction can be accurately reproduced, but the sense of distance cannot be reproduced. Therefore, the volume may be controlled according to the distance r1 to the sound source.
- the audio encoded signal separated by the audio object separation unit 100 is decoded by the decoding unit 104 into an audio PCM signal. This may be achieved by using a decoder in a codec system for audio encoded signals.
- the audio PCM signal decoded in this way is processed in the signal processing unit 105 by the signal processing method selected by the selection unit 103. That is, when the y-coordinate value of the corrected playback position information is a negative value, the signal processing unit 105 performs wavefront synthesis by signal processing using Huygens' theorem, and the y-coordinate value of the corrected playback position information is the listener's position. When the positive value is earlier, a sound image is generated by signal processing using the beamform, and when the y-coordinate value of the corrected reproduction position information is a positive value after the listener's position, HRTF is used. The sound image is localized by signal processing.
- the signal processing method is any one of beamform, wavefront synthesis, and HRTF.
- any signal processing method a more specific implementation method is conventionally known. What is necessary is just to use the signal processing system currently used.
- the speaker array 106 converts the output signal (acoustic signal) from the signal processing unit 105 into acoustic vibration.
- FIG. 8 is a flowchart showing main operations of the audio playback device 110 according to the present embodiment.
- the audio object separation unit 100 separates the three-dimensional reproduction position information and the audio encoded signal from the audio object (S10).
- the conversion unit 102 corrects the three-dimensional reproduction position information separated by the audio object separation unit 100 as position information (two-dimensional information) on a two-dimensional coordinate axis based on the position of the speaker array 106. Conversion into position information (S11).
- the selection unit 103 receives the corrected reproduction position information generated by the conversion unit 102, the two-dimensional coordinate axis set by the setting unit 101, and the position of the listener who listens to the sound output from the speaker array 106 (or The signal processing method to be adopted by the signal processing unit 105 is selected from the listening position planned by the audio playback device 110 (S12).
- the signal processing unit 105 localizes the sound image of the audio signal separated by the audio object separation unit 100 and decoded by the decoding unit 104 according to the modified reproduction position obtained by the conversion by the conversion unit 102. Processing is performed (S13). At this time, the signal processing unit 105 performs processing using the signal processing method selected by the selection unit 103.
- the three-dimensional reproduction position information included in the audio object is converted into the corrected reproduction position information on the two-dimensional coordinate axis based on the position of the speaker array, and the sound image is localized according to the corrected reproduction position information after correction. Therefore, even when the speaker arrangement is limited, the audio object is reproduced with a high sense of presence.
- steps S10 to S13 are shown as the main operation steps. However, as a minimum step, the conversion step S11 and the signal processing step S13 may be executed. By these two steps, the three-dimensional reproduction position information is converted into the corrected reproduction position information on the two-dimensional coordinate axis, so that an audio object including the three-dimensional reproduction position information can be obtained even in a space where speakers cannot be freely arranged. It can be played with high presence.
- FIG. 9 is a flowchart showing an operation relating to handling of playback position information included in an audio frame, among the operations of the audio playback device 110 of the present embodiment. Here, an operation related to reproduction position information performed for each audio frame included in the audio object is shown.
- the audio object separation unit 100 determines whether or not reproduction position information is missing in the audio frame to be processed (S20).
- the reproduction position information included in the audio frame reproduced in the past by the audio object separation unit 100 is processed. It is used as reproduction position information of the audio frame, and signal processing is performed by the signal processing unit 105 according to the reproduction position information (after being converted into two-dimensional corrected reproduction position information) (S21).
- the audio object separation unit 100 separates the reproduction position information included in the audio frame to be processed, and the reproduction position.
- the signal processing unit 105 performs signal processing (S22).
- the playback position information included in the audio frame played back in the past is used, so that natural sound field playback can be performed or the audio object moves. It is possible to reduce the amount of information when recording or transmitting the audio object when it is not.
- one method is selected from the three signal processing methods in accordance with the corrected reproduction position information.
- (A) of FIG. 10 is a diagram in which it is arranged.
- wavefront synthesis based on the Huygens principle is used, when the front of the listener is ahead of the listener, the beamform method is used, and when the listener is behind the listener, the method using HRTF is used.
- FIG. 10B shows a signal processing method in the vicinity of each boundary line when the audio object (the position indicated by the reproduction position information included in the audio object) moves with time.
- the signal processing unit 105 when the corrected reproduction position information is in the vicinity of the line of the speaker array, the signal processing unit 105 generates a signal obtained by mixing the output signal by the wavefront synthesis method and the output signal by the beamform method at a predetermined ratio. Similarly, in the vicinity of the listener, the signal processing unit 105 generates a signal obtained by mixing the output signal by the beamform method and the output signal by the HRTF method at a predetermined ratio.
- one of the three signal processing methods is selected according to the corrected playback position information.
- the HRTF method is selected regardless of the position of the corrected playback position information. May be.
- the sound image cannot be localized in front of the speaker, and in the beamform, the sound image cannot be localized behind the speaker or behind the listener. This is because any control is possible by simulating the phase difference information and level difference information between the ears, and the sound transfer characteristics around the head.
- FIG. 11 shows a locus of position information targeted by the HRTF when the audio object (the position indicated by the reproduction position information included in the audio object) passes over the listener's head. When the audio object (the position indicated by the playback position information included in the audio object) has reached the listener's head, control is performed so as to go around the head. By doing so, the presence in the vicinity of the overhead can be enhanced.
- Patent Document 1 a research result that a clue related to localization in the vertical direction is included in the amplitude spectrum of the acoustic transfer function around the head (Patent Document 1). ) May be used to add that element to the HRTF.
- FIG. 12 shows a variation using two speaker arrays 106a and 106b
- FIG. 13 shows a variation using three speaker arrays 106a to 106c.
- the audio playback device includes at least two speaker arrays, and the at least two speaker arrays constitute at least two two-dimensional coordinates and are specified by the playback position information.
- the signal processing unit 105 controls at least two speaker arrays in accordance with the value of z.
- the signal processing unit 105 determines that the XY plane (when the value of z is greater than (or greater than) a predetermined value)
- the XY plane When the volume of the speaker array constituting the upper two-dimensional coordinates with respect to the installation surface) is increased and the value of z is smaller than (or less than) a predetermined value, the XY plane ( The volume of the speaker array constituting the lower two-dimensional coordinates with respect to the installation surface) is increased.
- the signal processing unit 105 determines that the value of the z is greater than (or more than) a predetermined value, the XY plane (installation plane).
- the XY plane installation plane
- loudspeaker elements constituting a loudspeaker array constituting a two-dimensional coordinate perpendicular to the above the volume of the loudspeaker element above a predetermined position is increased, and the value of z is a predetermined value. Is smaller than (or less than) a speaker element constituting a speaker array constituting a two-dimensional coordinate perpendicular to the XY plane (installation plane) below a predetermined position. Increase the volume of the speaker element.
- the audio playback device 110 includes at least two speaker arrays
- the at least two speaker arrays are set according to the value of z at the position (x, y, z) specified by the playback position information. Therefore, the height information of the reproduction position information can be controlled, and the audio object including the three-dimensional reproduction position information is reproduced with a high sense of presence.
- the audio reproduction device 110 includes at least one speaker array 106 that converts an acoustic signal into acoustic vibration, and two-dimensional reproduction position information based on the position of the speaker array 106.
- a conversion unit 102 that converts position information (corrected playback position information) on a dimensional coordinate axis and a signal processing unit 105 that processes the sound image of the audio object according to the corrected playback position are provided.
- An audio object accompanied with reproduction position information can be reproduced with the best possible presence even in an environment where speakers cannot be freely arranged, such as a ceiling speaker cannot be installed.
- the audio playback device has been described based on the embodiment.
- the audio playback device according to the present disclosure is not limited to this embodiment. As long as it does not deviate from the gist of the present disclosure, it may be a form in which various modifications conceived by those skilled in the art have been made in the present embodiment, or a structure constructed by combining components in different embodiments.
- the setting unit 101 is provided in the present embodiment, it is needless to say that the setting unit 101 is not necessary when the installation position of the speaker array is determined in advance.
- the listener position information is input to the selection unit 103.
- the position of the listener is determined in advance, or the present apparatus assumes that the position of the listener is in advance. Needless to say, if the current position is fixed, it is not necessary (input of listener position information).
- the selection unit 103 is not necessary when the signal processing method is fixed (for example, when it is determined that processing is always performed by HRTF).
- the decoding unit 104 is provided.
- the audio encoded signal is a simple PCM signal, that is, when the audio signal included in the audio object is not encoded, decoding is performed. It goes without saying that the conversion unit 104 is unnecessary.
- the audio object separation unit 100 is provided. However, when an audio object having a structure in which an audio signal and reproduction position information are separated is input to the audio reproduction device 110, the audio object separation is performed. Needless to say, the portion 100 is unnecessary.
- the speaker array does not have to have speaker elements arranged in a straight line, and may be, for example, an arch (arc). Further, the distance between the speaker elements may not be constant. In the present disclosure, the shape of the speaker array is not limited.
- the audio playback device is an audio playback device including a speaker array, an audio object including 3D position information can be played with high presence even in a space where speakers cannot be arranged in 3D. It can be used for a wide range of equipment that plays audio signals.
Landscapes
- Physics & Mathematics (AREA)
- Engineering & Computer Science (AREA)
- Acoustics & Sound (AREA)
- Signal Processing (AREA)
- Health & Medical Sciences (AREA)
- Otolaryngology (AREA)
- General Health & Medical Sciences (AREA)
- Stereophonic System (AREA)
- Circuit For Audible Band Transducer (AREA)
Abstract
Description
101 設定部
102 変換部
103 選択部
104 復号化部
105 信号処理部
106、106a~106c スピーカアレー
110 オーディオ再生装置
Claims (17)
- オーディオ信号と前記オーディオ信号の音像を定位させる3次元空間における位置を示す再生位置情報とを含むオーディオオブジェクトを再生するオーディオ再生装置であって、
音響信号を音響振動に変換する少なくとも1体のスピーカアレーと、
前記再生位置情報を、前記スピーカアレーの位置を基準とした2次元座標軸上の位置情報である修正再生位置情報に変換する変換部と、
前記修正再生位置情報に応じて前記オーディオオブジェクトに含まれる前記オーディオ信号の音像を定位させる処理をする信号処理部と、
を備えるオーディオ再生装置。 - 前記スピーカアレーを構成するスピーカ素子の並び方向をX軸、前記X軸と直交し、かつ、前記スピーカアレーが設置されている面である設置面と水平な方向をY軸、及び、前記X軸と直交し、かつ、前記設置面と垂直な方向をZ軸としたとき、
前記修正再生位置情報は、前記X軸と前記Y軸とで構成された座標軸上の位置を指し、
前記再生位置情報で特定される位置が(x、y、z)であるとき、前記修正再生位置情報は前記x、yに応じた値である
請求項1記載のオーディオ再生装置。 - 前記2次元座標におけるY座標では、前記スピーカアレーの背面方向がマイナス座標、前記スピーカアレーの前面方向がプラス座標であり、前記2次元座標におけるX座標では、前記スピーカアレーの中央から左右にそれぞれマイナス座標、プラス座標であるとき、前記修正再生位置情報の値は、前記x、yの少なくとも一方に所定の値を乗じた値である
請求項2記載のオーディオ再生装置。 - 前記修正再生位置情報のx座標値は、前記スピーカアレーの幅に制限される
請求項2又は3記載のオーディオ再生装置。 - 前記信号処理部は、前記2次元座標軸上の位置に音像を構成するビームフォーム部である
請求項1から4のいずれか1項に記載のオーディオ再生装置。 - 前記2次元座標におけるY座標では、前記スピーカアレーの背面方向がマイナス座標、前記スピーカアレーの前面方向がプラス座標であり、
前記信号処理部は、前記修正再生位置情報のy座標値が負の値である場合、Huygensの定理を利用した信号処理で波面合成する
請求項2記載のオーディオ再生装置。 - 前記修正再生位置情報は、前記スピーカアレーから出力される音響を受聴する受聴者の位置から見た前記再生位置情報が示す位置への方向角と、前記受聴者の位置から前記再生位置情報が示す位置までの距離とによって前記2次元座標軸上の位置を示す
請求項1記載のオーディオ再生装置。 - 前記信号処理部は、HRTF(Head Related Transfer Function)を用いて前記音像を定位させる処理をし、
前記HRTFは、前記修正再生位置情報が示す位置の方向から音が聞こえるように設定されている
請求項7記載のオーディオ再生装置。 - 前記信号処理部は、前記受聴者の位置と前記修正再生位置情報が示す位置との距離に応じて音量を調整する
請求項8記載のオーディオ再生装置。 - 前記信号処理部は、前記修正再生位置情報が示す位置に応じて信号処理方式を変更する
請求項1記載のオーディオ再生装置。 - 前記スピーカアレーを構成するスピーカ素子の並び方向をX軸、前記X軸と直交し、かつ、前記スピーカアレーが設置されている面である設置面と水平な方向をY軸、及び、前記X軸と直交し、かつ、前記設置面と垂直な方向をZ軸とし、
前記Y軸における位置を示すY座標では、前記スピーカアレーの背面方向がマイナス座標、前記スピーカアレーの前面方向がプラス座標であり、
前記信号処理部は、
前記修正再生位置情報のy座標値が負の値である場合、Huygensの定理を利用した信号処理で波面合成し、
前記修正再生位置情報のy座標値が受聴者の位置より前の正の値である場合、ビームフォームを利用した信号処理で音像を生成し、
前記修正再生位置情報のy座標値が受聴者の位置より後ろの正の値である場合、HRTFを利用した信号処理で音像を定位させる
請求項10記載のオーディオ再生装置。 - 前記オーディオ再生装置は、少なくとも2体のスピーカアレーを備え、
前記少なくとも2体のスピーカアレーが少なくとも2つの2次元座標を構成し、
前記少なくとも2体のスピーカアレーのうちの一つのスピーカアレーを構成するスピーカ素子の並び方向をX軸、前記X軸と直交し、かつ、前記一つのスピーカアレーが設置されている面である設置面と水平な方向をY軸、及び、前記X軸と直交し、かつ、前記設置面と垂直な方向をZ軸とし、前記再生位置情報で特定される位置が(x、y、z)であるとき、
前記信号処理部は、前記zの値に応じて前記少なくとも2体のスピーカアレーを制御する
請求項1記載のオーディオ再生装置。 - 前記2つの2次元座標が平行しているとき、前記信号処理部は、
前記zの値が予め定められた値よりも大きい場合は、前記設置面に対して上側の2次元座標を構成しているスピーカアレーの音量を大きくし、
前記zの値が予め定められた値よりも小さい場合は、前記設置面に対して下側の2次元座標を構成しているスピーカアレーの音量を大きくする
請求項12記載のオーディオ再生装置。 - 前記2つの2次元座標が直交しているとき、前記信号処理部は、
前記zの値が予め定められた値よりも大きい場合は、前記設置面に対して垂直の2次元座標を構成しているスピーカアレーを構成するスピーカ素子のうち、予め定められた位置よりも上方のスピーカ素子の音量を大きくし、
前記zの値が予め定められた値よりも小さい場合は、前記設置面に対して垂直の2次元座標を構成しているスピーカアレーを構成するスピーカ素子のうち、予め定められた位置よりも下方のスピーカ素子の音量を大きくする
請求項12記載のオーディオ再生装置。 - オーディオ信号と前記オーディオ信号の音像を定位させる3次元空間における位置を示す再生位置情報とを含むオーディオオブジェクトを再生するオーディオ再生装置であって、
前記オーディオオブジェクトは、所定の時間間隔ごとの前記オーディオ信号と前記再生位置情報とを含むオーディオフレームから構成され、
前記オーディオ再生装置は、
前記再生位置情報が欠落している場合には、過去に再生したオーディオフレームに含まれていた再生位置情報を、前記再生位置情報が欠落したオーディオフレームの再生位置情報として用いることで、前記オーディオオブジェクトに含まれるオーディオフレームを再生する
オーディオ再生装置。 - オーディオ信号と前記オーディオ信号の音像を定位させる3次元空間における位置を示す再生位置情報とを含むオーディオオブジェクトを、スピーカアレーを用いて、再生するオーディオ再生方法であって、
前記再生位置情報を、前記スピーカアレーの位置を基準とした2次元座標軸上の位置情報である修正再生位置情報に変換する変換ステップと、
前記修正再生位置情報に応じて前記オーディオオブジェクトに含まれる前記オーディオ信号の音像を定位させる処理をする信号処理ステップとを含む
オーディオ再生方法。 - オーディオ信号と前記オーディオ信号の音像を定位させる3次元空間における位置を示す再生位置情報とを含むオーディオオブジェクトを、スピーカアレーを用いて、再生するオーディオ再生方法であって、
前記オーディオオブジェクトは、所定の時間間隔ごとの前記オーディオ信号と前記再生位置情報とを含むオーディオフレームから構成され、
前記オーディオ再生方法は、
前記再生位置情報が欠落している場合には、過去に再生したオーディオフレームに含まれていた再生位置情報を、前記再生位置情報が欠落したオーディオフレームの再生位置情報として用いることで、前記オーディオオブジェクトに含まれるオーディオフレームを再生するステップを含む
オーディオ再生方法。
Priority Applications (3)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201480032404.7A CN105264914B (zh) | 2013-06-10 | 2014-02-19 | 音频再生装置以及方法 |
JP2015522476A JP6022685B2 (ja) | 2013-06-10 | 2014-02-19 | オーディオ再生装置及びその方法 |
US14/961,739 US9788120B2 (en) | 2013-06-10 | 2015-12-07 | Audio playback device and audio playback method |
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
JP2013122254 | 2013-06-10 | ||
JP2013-122254 | 2013-06-10 |
Related Child Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US14/961,739 Continuation US9788120B2 (en) | 2013-06-10 | 2015-12-07 | Audio playback device and audio playback method |
Publications (1)
Publication Number | Publication Date |
---|---|
WO2014199536A1 true WO2014199536A1 (ja) | 2014-12-18 |
Family
ID=52021863
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
PCT/JP2014/000868 WO2014199536A1 (ja) | 2013-06-10 | 2014-02-19 | オーディオ再生装置及びその方法 |
Country Status (4)
Country | Link |
---|---|
US (1) | US9788120B2 (ja) |
JP (1) | JP6022685B2 (ja) |
CN (3) | CN105264914B (ja) |
WO (1) | WO2014199536A1 (ja) |
Cited By (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JPWO2019163013A1 (ja) * | 2018-02-21 | 2021-02-04 | 株式会社ソシオネクスト | 音声信号処理装置、音声調整方法及びプログラム |
JP2021048501A (ja) * | 2019-09-19 | 2021-03-25 | ソニー株式会社 | 信号処理装置、信号処理方法および信号処理システム |
WO2024014390A1 (ja) * | 2022-07-13 | 2024-01-18 | パナソニック インテレクチュアル プロパティ コーポレーション オブ アメリカ | 音響信号処理方法、情報生成方法、コンピュータプログラム、及び、音響信号処理装置 |
Families Citing this family (9)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN107979807A (zh) * | 2016-10-25 | 2018-05-01 | 北京酷我科技有限公司 | 一种模拟环绕立体声的方法及系统 |
US10531196B2 (en) * | 2017-06-02 | 2020-01-07 | Apple Inc. | Spatially ducking audio produced through a beamforming loudspeaker array |
CN108414072A (zh) * | 2017-11-07 | 2018-08-17 | 四川大学 | 一种真三维声音录制与播放系统 |
CN109286888B (zh) * | 2018-10-29 | 2021-01-29 | 中国传媒大学 | 一种音视频在线检测与虚拟声像生成方法及装置 |
JP2021153292A (ja) * | 2020-03-24 | 2021-09-30 | ヤマハ株式会社 | 情報処理方法および情報処理装置 |
CN111787460B (zh) * | 2020-06-23 | 2021-11-09 | 北京小米移动软件有限公司 | 设备控制方法及装置 |
JP7337234B2 (ja) | 2020-09-01 | 2023-09-01 | 株式会社Lixil | サッシ窓 |
CN113329319B (zh) * | 2021-05-27 | 2022-10-21 | 音王电声股份有限公司 | 一种扬声器阵列的沉浸声还音制式方法及其应用 |
GB2611547A (en) * | 2021-10-07 | 2023-04-12 | Nokia Technologies Oy | Apparatus, methods and computer programs for processing spatial audio |
Citations (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP2001197598A (ja) * | 2000-01-05 | 2001-07-19 | Mitsubishi Electric Corp | 映像音声再生装置 |
JP2006128818A (ja) * | 2004-10-26 | 2006-05-18 | Victor Co Of Japan Ltd | 立体映像・立体音響対応記録プログラム、再生プログラム、記録装置、再生装置及び記録メディア |
JP2011035784A (ja) * | 2009-08-04 | 2011-02-17 | Sharp Corp | 立体映像・立体音響記録再生装置・システム及び方法 |
JP2011066868A (ja) * | 2009-08-18 | 2011-03-31 | Victor Co Of Japan Ltd | オーディオ信号符号化方法、符号化装置、復号化方法及び復号化装置 |
Family Cites Families (13)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US6990205B1 (en) | 1998-05-20 | 2006-01-24 | Agere Systems, Inc. | Apparatus and method for producing virtual acoustic sound |
DE10328335B4 (de) * | 2003-06-24 | 2005-07-21 | Fraunhofer-Gesellschaft zur Förderung der angewandten Forschung e.V. | Wellenfeldsyntesevorrichtung und Verfahren zum Treiben eines Arrays von Lautsprechern |
DE10344638A1 (de) | 2003-08-04 | 2005-03-10 | Fraunhofer Ges Forschung | Vorrichtung und Verfahren zum Erzeugen, Speichern oder Bearbeiten einer Audiodarstellung einer Audioszene |
JP4551652B2 (ja) * | 2003-12-02 | 2010-09-29 | ソニー株式会社 | 音場再生装置及び音場空間再生システム |
US8005245B2 (en) | 2004-09-16 | 2011-08-23 | Panasonic Corporation | Sound image localization apparatus |
DE102005008369A1 (de) | 2005-02-23 | 2006-09-07 | Fraunhofer-Gesellschaft zur Förderung der angewandten Forschung e.V. | Vorrichtung und Verfahren zum Simulieren eines Wellenfeldsynthese-Systems |
DE102005008366A1 (de) | 2005-02-23 | 2006-08-24 | Fraunhofer-Gesellschaft zur Förderung der angewandten Forschung e.V. | Vorrichtung und Verfahren zum Ansteuern einer Wellenfeldsynthese-Renderer-Einrichtung mit Audioobjekten |
DE102005008333A1 (de) | 2005-02-23 | 2006-08-31 | Fraunhofer-Gesellschaft zur Förderung der angewandten Forschung e.V. | Vorrichtung und Verfahren zum Steuern einer Wellenfeldsynthese-Rendering-Einrichtung |
JP4449998B2 (ja) * | 2007-03-12 | 2010-04-14 | ヤマハ株式会社 | アレイスピーカ装置 |
JP2011124723A (ja) * | 2009-12-09 | 2011-06-23 | Sharp Corp | オーディオデータ処理装置、オーディオ装置、オーディオデータ処理方法、プログラム及び当該プログラムを記録した記録媒体 |
KR101268779B1 (ko) * | 2009-12-09 | 2013-05-29 | 한국전자통신연구원 | 라우드 스피커 어레이를 사용한 음장 재생 장치 및 방법 |
US9179236B2 (en) * | 2011-07-01 | 2015-11-03 | Dolby Laboratories Licensing Corporation | System and method for adaptive audio signal generation, coding and rendering |
TW201412092A (zh) * | 2012-09-05 | 2014-03-16 | Acer Inc | 多媒體處理系統及音訊信號處理方法 |
-
2014
- 2014-02-19 CN CN201480032404.7A patent/CN105264914B/zh active Active
- 2014-02-19 CN CN201710205756.3A patent/CN106961645B/zh active Active
- 2014-02-19 CN CN201710209373.3A patent/CN106961647B/zh active Active
- 2014-02-19 WO PCT/JP2014/000868 patent/WO2014199536A1/ja active Application Filing
- 2014-02-19 JP JP2015522476A patent/JP6022685B2/ja active Active
-
2015
- 2015-12-07 US US14/961,739 patent/US9788120B2/en active Active
Patent Citations (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP2001197598A (ja) * | 2000-01-05 | 2001-07-19 | Mitsubishi Electric Corp | 映像音声再生装置 |
JP2006128818A (ja) * | 2004-10-26 | 2006-05-18 | Victor Co Of Japan Ltd | 立体映像・立体音響対応記録プログラム、再生プログラム、記録装置、再生装置及び記録メディア |
JP2011035784A (ja) * | 2009-08-04 | 2011-02-17 | Sharp Corp | 立体映像・立体音響記録再生装置・システム及び方法 |
JP2011066868A (ja) * | 2009-08-18 | 2011-03-31 | Victor Co Of Japan Ltd | オーディオ信号符号化方法、符号化装置、復号化方法及び復号化装置 |
Cited By (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JPWO2019163013A1 (ja) * | 2018-02-21 | 2021-02-04 | 株式会社ソシオネクスト | 音声信号処理装置、音声調整方法及びプログラム |
JP7115535B2 (ja) | 2018-02-21 | 2022-08-09 | 株式会社ソシオネクスト | 音声信号処理装置、音声調整方法及びプログラム |
JP2021048501A (ja) * | 2019-09-19 | 2021-03-25 | ソニー株式会社 | 信号処理装置、信号処理方法および信号処理システム |
JP7456106B2 (ja) | 2019-09-19 | 2024-03-27 | ソニーグループ株式会社 | 信号処理装置、信号処理方法および信号処理システム |
US12063495B2 (en) | 2019-09-19 | 2024-08-13 | Sony Group Corporation | Signal processing apparatus, signal processing method, and signal processing system |
WO2024014390A1 (ja) * | 2022-07-13 | 2024-01-18 | パナソニック インテレクチュアル プロパティ コーポレーション オブ アメリカ | 音響信号処理方法、情報生成方法、コンピュータプログラム、及び、音響信号処理装置 |
Also Published As
Publication number | Publication date |
---|---|
JP6022685B2 (ja) | 2016-11-09 |
CN106961645A (zh) | 2017-07-18 |
CN105264914A (zh) | 2016-01-20 |
US20160088393A1 (en) | 2016-03-24 |
JPWO2014199536A1 (ja) | 2017-02-23 |
CN106961647B (zh) | 2018-12-14 |
CN105264914B (zh) | 2017-03-22 |
CN106961647A (zh) | 2017-07-18 |
US9788120B2 (en) | 2017-10-10 |
CN106961645B (zh) | 2019-04-02 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
JP6022685B2 (ja) | オーディオ再生装置及びその方法 | |
US10659899B2 (en) | Methods and systems for rendering audio based on priority | |
US10021507B2 (en) | Arrangement and method for reproducing audio data of an acoustic scene | |
EP2954702B1 (en) | Mapping virtual speakers to physical speakers | |
EP3282716B1 (en) | Rendering of audio objects with apparent size to arbitrary loudspeaker layouts | |
EP3028476B1 (en) | Panning of audio objects to arbitrary speaker layouts | |
JP5919201B2 (ja) | 音声を定位知覚する技術 | |
JP6515087B2 (ja) | オーディオ処理装置及び方法 | |
KR101268779B1 (ko) | 라우드 스피커 어레이를 사용한 음장 재생 장치 및 방법 | |
JP2015529415A (ja) | 多次元的パラメトリック音声のシステムおよび方法 | |
US9467792B2 (en) | Method for processing of sound signals | |
JP2006303658A (ja) | 再生装置および再生方法 | |
JP6663490B2 (ja) | スピーカシステム、音声信号レンダリング装置およびプログラム | |
JP2018110366A (ja) | 3dサウンド映像音響機器 | |
US10440495B2 (en) | Virtual localization of sound | |
US20140219458A1 (en) | Audio signal reproduction device and audio signal reproduction method | |
JP2013128314A (ja) | 波面合成信号変換装置および波面合成信号変換方法 |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
WWE | Wipo information: entry into national phase |
Ref document number: 201480032404.7 Country of ref document: CN |
|
121 | Ep: the epo has been informed by wipo that ep was designated in this application |
Ref document number: 14810259 Country of ref document: EP Kind code of ref document: A1 |
|
ENP | Entry into the national phase |
Ref document number: 2015522476 Country of ref document: JP Kind code of ref document: A |
|
NENP | Non-entry into the national phase |
Ref country code: DE |
|
122 | Ep: pct application non-entry in european phase |
Ref document number: 14810259 Country of ref document: EP Kind code of ref document: A1 |