WO2022054602A1 - 音響処理装置および方法、並びにプログラム - Google Patents
音響処理装置および方法、並びにプログラム Download PDFInfo
- Publication number
- WO2022054602A1 WO2022054602A1 PCT/JP2021/031449 JP2021031449W WO2022054602A1 WO 2022054602 A1 WO2022054602 A1 WO 2022054602A1 JP 2021031449 W JP2021031449 W JP 2021031449W WO 2022054602 A1 WO2022054602 A1 WO 2022054602A1
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- speaker
- processing unit
- rendering
- audio signal
- band
- Prior art date
Links
- 238000012545 processing Methods 0.000 title claims abstract description 442
- 238000000034 method Methods 0.000 title claims abstract description 110
- 238000009877 rendering Methods 0.000 claims abstract description 283
- 230000005236 sound signal Effects 0.000 claims abstract description 93
- 238000003672 processing method Methods 0.000 claims description 3
- 238000005516 engineering process Methods 0.000 abstract description 23
- 238000006243 chemical reaction Methods 0.000 description 93
- 230000003321 amplification Effects 0.000 description 84
- 238000003199 nucleic acid amplification method Methods 0.000 description 84
- 238000001914 filtration Methods 0.000 description 50
- 230000007274 generation of a signal involved in cell-cell signaling Effects 0.000 description 43
- 101100365087 Arabidopsis thaliana SCRA gene Proteins 0.000 description 35
- 230000004807 localization Effects 0.000 description 13
- 230000006870 function Effects 0.000 description 10
- 230000006866 deterioration Effects 0.000 description 9
- 238000007726 management method Methods 0.000 description 6
- 238000013461 design Methods 0.000 description 5
- 238000004091 panning Methods 0.000 description 5
- 238000004891 communication Methods 0.000 description 3
- 230000005540 biological transmission Effects 0.000 description 2
- 238000010586 diagram Methods 0.000 description 2
- 238000004458 analytical method Methods 0.000 description 1
- 230000006399 behavior Effects 0.000 description 1
- 230000000694 effects Effects 0.000 description 1
- 230000001771 impaired effect Effects 0.000 description 1
- 230000003287 optical effect Effects 0.000 description 1
- 230000002787 reinforcement Effects 0.000 description 1
- 239000004065 semiconductor Substances 0.000 description 1
- 230000008685 targeting Effects 0.000 description 1
- 230000001755 vocal effect Effects 0.000 description 1
Images
Classifications
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04R—LOUDSPEAKERS, MICROPHONES, GRAMOPHONE PICK-UPS OR LIKE ACOUSTIC ELECTROMECHANICAL TRANSDUCERS; DEAF-AID SETS; PUBLIC ADDRESS SYSTEMS
- H04R3/00—Circuits for transducers, loudspeakers or microphones
- H04R3/005—Circuits for transducers, loudspeakers or microphones for combining the signals of two or more microphones
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04R—LOUDSPEAKERS, MICROPHONES, GRAMOPHONE PICK-UPS OR LIKE ACOUSTIC ELECTROMECHANICAL TRANSDUCERS; DEAF-AID SETS; PUBLIC ADDRESS SYSTEMS
- H04R29/00—Monitoring arrangements; Testing arrangements
- H04R29/001—Monitoring arrangements; Testing arrangements for loudspeakers
- H04R29/002—Loudspeaker arrays
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04S—STEREOPHONIC SYSTEMS
- H04S7/00—Indicating arrangements; Control arrangements, e.g. balance control
- H04S7/30—Control circuits for electronic adaptation of the sound field
- H04S7/302—Electronic adaptation of stereophonic sound system to listener position or orientation
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04S—STEREOPHONIC SYSTEMS
- H04S7/00—Indicating arrangements; Control arrangements, e.g. balance control
- H04S7/30—Control circuits for electronic adaptation of the sound field
- H04S7/307—Frequency adjustment, e.g. tone control
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04R—LOUDSPEAKERS, MICROPHONES, GRAMOPHONE PICK-UPS OR LIKE ACOUSTIC ELECTROMECHANICAL TRANSDUCERS; DEAF-AID SETS; PUBLIC ADDRESS SYSTEMS
- H04R2430/00—Signal processing covered by H04R, not provided for in its groups
- H04R2430/20—Processing of the output signals of the acoustic transducers of an array for obtaining a desired directivity characteristic
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04R—LOUDSPEAKERS, MICROPHONES, GRAMOPHONE PICK-UPS OR LIKE ACOUSTIC ELECTROMECHANICAL TRANSDUCERS; DEAF-AID SETS; PUBLIC ADDRESS SYSTEMS
- H04R3/00—Circuits for transducers, loudspeakers or microphones
- H04R3/12—Circuits for transducers, loudspeakers or microphones for distributing signals to two or more loudspeakers
- H04R3/14—Cross-over networks
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04S—STEREOPHONIC SYSTEMS
- H04S2400/00—Details of stereophonic systems covered by H04S but not provided for in its groups
- H04S2400/07—Generation or adaptation of the Low Frequency Effect [LFE] channel, e.g. distribution or signal processing
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04S—STEREOPHONIC SYSTEMS
- H04S2400/00—Details of stereophonic systems covered by H04S but not provided for in its groups
- H04S2400/11—Positioning of individual sound objects, e.g. moving airplane, within a sound field
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04S—STEREOPHONIC SYSTEMS
- H04S2420/00—Techniques used stereophonic systems covered by H04S but not provided for in its groups
- H04S2420/03—Application of parametric coding in stereophonic audio systems
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04S—STEREOPHONIC SYSTEMS
- H04S2420/00—Techniques used stereophonic systems covered by H04S but not provided for in its groups
- H04S2420/07—Synergistic effects of band splitting and sub-band processing
Definitions
- the present technology relates to sound processing devices and methods, and programs, and in particular, to sound processing devices, methods, and programs that enable higher-quality audio reproduction.
- audio data is composed of a waveform signal (audio signal) with respect to an object and metadata showing localization information indicating the relative position of the object as seen from a predetermined reference viewing point (listening position). .. Then, based on the metadata, the waveform signal is rendered to a desired number of channels by, for example, VBAP (Vector Based Amplitude Panning), and reproduced (see, for example, Non-Patent Document 1 and Non-Patent Document 2).
- VBAP Vector Based Amplitude Panning
- in-vehicle audio is a use case in which many speakers can be arranged.
- In-vehicle audio is composed of a speaker layout in which a speaker with a low-frequency reproduction band generally called a woofer, a speaker with a mid-range reproduction band called a squawker, and a speaker with a high-frequency reproduction band called a tweeter are mixed.
- a speaker with a low-frequency reproduction band generally called a woofer
- a speaker with a mid-range reproduction band called a squawker a speaker with a high-frequency reproduction band called a tweeter are mixed.
- the playback band of the speaker used for playback differs depending on the localization position of the object.
- the sound quality may disappear depending on the frequency band or localization position of the object's sound. Deterioration will occur.
- This technology was made in view of such a situation, and makes it possible to perform audio reproduction with higher sound quality.
- the sound processing device of one aspect of the present technology is a first rendering processing unit that performs rendering processing based on an audio signal and generates a first output audio signal for outputting sound by a plurality of first speakers.
- a second rendering process that performs rendering processing based on the audio signal and generates a second output audio signal for outputting sound by a plurality of second speakers having a reproduction band different from that of the first speaker. It has a part.
- the sound processing method or program of one aspect of the present technology performs rendering processing based on an audio signal, generates a first output audio signal for outputting sound by a plurality of first speakers, and uses the audio signal as the first output audio signal. Based on the above, a rendering process is performed, and a step of generating a second output audio signal for outputting sound by a plurality of second speakers having a reproduction band different from that of the first speaker is included.
- rendering processing is performed based on an audio signal, a first output audio signal for outputting sound by a plurality of first speakers is generated, and rendering processing is performed based on the audio signal. Is performed, and a second output audio signal for outputting sound is generated by a plurality of second speakers having a reproduction band different from that of the first speaker.
- This technology achieves higher sound quality by performing rendering processing for each speaker layout consisting of speakers with the same playback band when playing object-based audio on a speaker system consisting of speakers with multiple playback bands that are different from each other. It plays audio.
- a plurality of speakers SP11-1 to SP11-18 surround the user U11 on the surface of the sphere P11 centered on the user U11 who is a listener of object-based audio. Be placed.
- the object-based audio is reproduced by using the speaker system consisting of these speakers SP11-1 to SP11-18.
- the speaker SP11-1 when it is not necessary to distinguish between the speaker SP11-1 and the speaker SP11-18, they are also simply referred to as the speaker SP11.
- a speaker group consisting of speakers SP11 having the same reproduction band, and more specifically, a three-dimensional arrangement of each speaker SP11 constituting the speaker group is referred to as one speaker layout.
- rendering processing is performed for each speaker layout constituting the speaker system, and a speaker reproduction signal for reproducing the sound of an object (audio object) in the speaker layout is generated.
- the rendering process may be any process such as VBAP or panning.
- the speaker reproduction signal of each speaker SP11 of that speaker layout is generated.
- one or more meshes are formed on the surface of the sphere P11 by all the speakers SP11 constituting the speaker layout.
- One mesh is a triangular area on the surface of the sphere P11 surrounded by the three speakers SP11 that make up the speaker layout.
- the object data of the object is supplied, and the object data is composed of an object signal which is an audio signal for reproducing the sound of the object and metadata which is information about the object.
- the metadata contains at least the position of the object, that is, the position information indicating the sound image localization position of the sound of the object.
- the position information of this object is, for example, coordinate information indicating the relative position of the object as seen from the position of the head of the user U11, which is a predetermined listening position.
- the position information is information indicating the relative position of the object with respect to the head position of the user U11.
- one mesh including the position indicated by the position information of the object (hereinafter, also referred to as the object position) is selected from the mesh formed by the speaker SP11 of the speaker layout.
- the selected mesh is referred to as a selection mesh.
- the VBAP gain is obtained for each speaker SP11 based on the positional relationship between the placement position of each speaker SP11 constituting the selection mesh and the object position, and the gain of the object signal is adjusted by the VBAP gain. It is regarded as a speaker reproduction signal.
- the signal obtained by adjusting the gain of the object signal based on the VBAP gain obtained for the speaker SP11 is the speaker reproduction signal of the speaker SP11.
- the speaker reproduction signals of the speakers SP11 other than the speakers SP11 constituting the selection mesh are regarded as zero signals.
- the VBAP gain of the speakers SP11 other than the speakers SP11 constituting the selection mesh is set to 0.
- panning can be used to generate a speaker reproduction signal for each speaker SP11 in the speaker layout.
- the gain for each speaker SP11 is obtained based on the positional relationship between each speaker SP11 in the speaker layout and the object in each direction such as the front-back direction, the left-right direction, and the up-down direction. Then, the gain of the object signal is adjusted by the obtained gain for each speaker SP11, and the speaker reproduction signal of each speaker SP11 is generated.
- the rendering process for each speaker layout may be any process such as VBAP or panning, but the case where VBAP is performed as the rendering process will be described below.
- rendering processing is performed for each of a plurality of speaker layouts having different reproduction bands from each other constituting the speaker system, and speaker reproduction signals of all the speakers SP11 constituting the speaker system are generated.
- a plurality of speaker layout configurations are prepared for each reproduction band, and rendering processing is performed for each reproduction band.
- a mesh is formed by all the speakers SP11 constituting the speaker system, and VBAP is performed as a rendering process.
- the speaker SP11-1, the speaker SP11-2, and the speaker SP11-5 are located. Plays the sound of the object.
- the speakers SP11-1, speaker SP11-2, and speaker SP11-5 are speakers whose reproduction band is low frequency, those speakers SP11 The sound of the object cannot be reproduced with sufficient sound pressure. Then, the sound quality of the object deteriorates, such as the sound of the object becoming quiet and inaudible.
- rendering processing is performed for each of a plurality of playback bands, so that the components of each frequency band are always reproduced by the speaker SP11 in the reproduction band including those frequency bands. Therefore, deterioration of sound quality due to the reproduction band of the speaker SP11 can be suppressed, and higher sound quality audio reproduction can be performed.
- the number of speakers SP11 constituting the speaker system, the reproduction band of each speaker SP11, and the arrangement position of the speaker SP11 in each reproduction band can be any number, reproduction band, or arrangement position. ..
- FIG. 2 is a diagram showing a configuration example of an embodiment of an audio reproduction system to which the present technology is applied.
- the audio reproduction system 11 shown in FIG. 2 has an audio processing device 21 and a speaker system 22, and reproduces object-based audio contents based on the supplied object data.
- the content consists of N objects, and the object data of those N objects is supplied, but the number of objects may be any number.
- the object data of one object includes an object signal for reproducing the sound of the object and the metadata of the object.
- the sound processing device 21 includes a reproduction signal generation unit 31, a D / A (Digital / Analog) conversion unit 32-1-1 to a D / A conversion unit 32-3-Nw, and an amplification unit 33-1-1 to an amplification unit. It has 33-3-Nw.
- the reproduction signal generation unit 31 performs rendering processing for each reproduction band to generate a speaker reproduction signal which is an output audio signal as an output.
- the reproduction signal generation unit 31 includes rendering processing unit 41-1 to rendering processing unit 41-3, HPF (High Pass Filter) 42-1 to HPF 42-Nt, BPF (Band Pass Filter) 43-1 to BPF 43-Ns, and It has LPF (Low Pass Filter) 44-1 to LPF44-Nw.
- HPF High Pass Filter
- BPF Band Pass Filter
- LPF Low Pass Filter
- the speaker system 22 includes speakers 51-1-1 to 51-1-Nt, speakers 51-2-1 to speaker 51-2-Ns, and speakers 51-3-1 to speaker 51-, which have different reproduction bands. It has 3-Nw.
- speaker 51-1-1 when it is not necessary to distinguish between the speaker 51-1-1 and the speaker 51-1-Nt, it is also simply referred to as a speaker 51-1.
- the speaker 51-2 when it is not necessary to particularly distinguish between the speaker 51-2-1 and the speaker 51-2-Ns, they are also simply referred to as the speaker 51-2, and the speaker 51-3-1 to the speaker 51-3-Nw are particularly referred to. When it is not necessary to distinguish between them, it is also simply referred to as a speaker 51-3.
- the speaker 51 constituting the speaker system 22 corresponds to the speaker SP11 shown in FIG.
- the rendering processing unit 41-1 to the rendering processing unit 41-3 perform rendering processing such as VBAP based on the object signal and metadata constituting the supplied object data, and generate a speaker reproduction signal of each speaker 51. ..
- the rendering processing unit 41-1 performs rendering processing for each of N objects, and generates each of the speaker reproduction signals having the speaker 51-1-1 to the speaker 51-1-Nt as output destinations for each object. do.
- the rendering processing unit 41-1 adds the speaker reproduction signals for each object generated for the same speaker 51-1 to obtain the final speaker reproduction signal of the speaker 51-1.
- the sound based on the speaker reproduction signal thus obtained includes the sound of each of N objects.
- the rendering processing unit 41-1 supplies the final speaker reproduction signal generated for the speakers 51-1-1 to the speaker 51-1-Nt to the HPF42-1 to HPF42-Nt.
- the rendering processing unit 41-2 also reproduces the sound of N objects having the speakers 51-2-1 to the speaker 51-2-Ns as the final output destinations in the same manner as the rendering processing unit 41-1.
- a speaker reproduction signal of each speaker 51-2 is generated and supplied to BPF43-1 to BPF43-Ns.
- the rendering processing unit 41-3 also reproduces the sound of N objects having the speakers 51-3-1 to the speaker 51-3-Nw as the final output destinations in the same manner as the rendering processing unit 41-1.
- a speaker reproduction signal of each speaker 51-3 is generated and supplied to LPF44-1 to LPF44-Nw.
- rendering processing unit 41-1 when it is not necessary to particularly distinguish between the rendering processing unit 41-1 and the rendering processing unit 41-3, it is also simply referred to as the rendering processing unit 41.
- HPF42-1 to HPF42-Nt are HPFs that pass at least the frequency band including the reproduction band of the speaker 51-1, that is, the high frequency component and block the mid-low frequency component.
- the HPF42-1 to HPF42-Nt perform filtering processing on the speaker reproduction signal supplied from the rendering processing unit 41-1, and the speaker reproduction signal containing only the high frequency component obtained as a result is a D / A conversion unit. It is supplied to 32-1-1 to D / A conversion unit 32-1-Nt.
- HPF42 performs a band limiting process according to the reproduction band of the speaker 51-1 called a filtering process by the HPF on the input speaker reproduction signal, and obtains a band-limited speaker reproduction signal (band limitation signal). It can be said that it functions as a band limiting processing unit to be generated.
- BPF43-1 to BPF43-Ns are BPFs that pass at least the frequency band including the reproduction band of the speaker 51-2, that is, the mid-range component, and block other components.
- BPF43-1 to BPF43-Ns perform filtering processing on the speaker reproduction signal supplied from the rendering processing unit 41-2, and the speaker reproduction signal containing only the mid-range component obtained as a result is a D / A conversion unit. It is supplied to 32-2-1 to D / A conversion unit 32-2-Ns.
- BPF43 performs a band limiting process according to the reproduction band of the speaker 51-2, which is a filtering process by the BPF, on the input speaker reproduction signal, and obtains a band-limited speaker reproduction signal (band limitation signal). It can be said that it functions as a band limiting processing unit to be generated.
- LPF44-1 to LPF44-Nw are LPFs that pass at least the frequency band including the reproduction band of the speaker 51-3, that is, the low frequency component, and block the mid-high frequency component.
- LPF44-1 to LPF44-Nw perform filtering processing on the speaker reproduction signal supplied from the rendering processing unit 41-3, and the speaker reproduction signal containing only the low-pass component obtained as a result is a D / A conversion unit. It is supplied to 32-3-1 to D / A conversion unit 32-3-Nw.
- LPF44 performs a band limiting process according to the reproduction band of the speaker 51-3, which is a filtering process by the LPF, on the input speaker reproduction signal, and obtains a band-limited speaker reproduction signal (band limitation signal). It can be said that it functions as a band limiting processing unit to be generated.
- the speaker reproduction signal of the above is supplied to the amplification unit 33-1-1 to the amplification unit 33-1-Nt.
- D / A conversion unit 32-1-1 when it is not necessary to distinguish between the D / A conversion unit 32-1-1 and the D / A conversion unit 32-1-Nt, they are also simply referred to as the D / A conversion unit 32-1.
- the D / A conversion unit 32-2-1 to the D / A conversion unit 32-2-Ns perform D / A conversion of the speaker reproduction signal supplied from the BPF43-1 to BPF43-Ns, and the analog obtained as a result.
- the speaker reproduction signal of the above is supplied to the amplification unit 33-2-1 to the amplification unit 33-2-Ns.
- D / A conversion unit 32-2-1 when it is not necessary to distinguish between the D / A conversion unit 32-2-1 and the D / A conversion unit 32-2-Ns, they are also simply referred to as the D / A conversion unit 32-2.
- the speaker reproduction signal of the above is supplied to the amplification unit 33-3-1 to the amplification unit 33-3-Nw.
- the D / A conversion unit 32-3-1 and the D / A conversion unit 32-3-Nw are also simply referred to as the D / A conversion unit 32-3.
- the D / A conversion unit 32 is also simply referred to as the D / A conversion unit 32.
- the amplification unit 33-1-1 to the amplification unit 33-1-Nt amplify the speaker reproduction signal supplied from the D / A conversion unit 32-1-1 to the D / A conversion unit 32-1-Nt, and the speaker is used. It is supplied to 51-1-1 to the speaker 51-1-Nt.
- the amplification unit 33-2-1 to the amplification unit 33-2-Ns amplifies the speaker reproduction signal supplied from the D / A conversion unit 32-2-1 to the D / A conversion unit 32-2-Ns, and the speaker is used. It is supplied to 51-2-1 to the speaker 51-2-Ns.
- the amplification unit 33-3-1 to the amplification unit 33-3-Nw amplifies the speaker reproduction signal supplied from the D / A conversion unit 32-3-1 to the D / A conversion unit 32-3-Nw, and the speaker is used. It is supplied to 51-3-1 to the speaker 51-3-Nw.
- amplification unit 33-1-1 when it is not necessary to distinguish the amplification unit 33-1-1 to the amplification unit 33-1-Nt, they are also simply referred to as the amplification unit 33-1 and the amplification units 33-2-1 to the amplification unit 33-2-Ns. Is not particularly necessary to be distinguished, it is also simply referred to as an amplification unit 33-2.
- amplification unit 33-3-1 when it is not necessary to particularly distinguish the amplification unit 33-3-1 to the amplification unit 33-3-Nw, they are also simply referred to as the amplification unit 33-3, and the amplification units 33-1 to the amplification unit 33-3 are particularly distinguished. When it is not necessary, it is also simply referred to as an amplification unit 33.
- the D / A conversion unit 32 and the amplification unit 33 may be provided outside the sound processing device 21.
- the speaker 51-1-1 to the speaker 51-1-Nt output sound based on the speaker reproduction signal supplied from the amplification unit 33-1-1 to the amplification unit 33-1-Nt.
- Each of the Nt speakers 51-1 constituting the speaker system 22 is a speaker called a tweeter, which mainly has a high band (high range) reproduction band.
- these Nt speakers 51-1 form one speaker layout for high bands.
- the speaker 51-2-1 to the speaker 51-2-Ns output sound based on the speaker reproduction signal supplied from the amplification unit 33-2-1 to the amplification unit 33-2-Ns.
- Each of the Ns speakers 51-2 constituting the speaker system 22 is a speaker called a squawker, which mainly has a reproduction band in the middle band (mid range).
- these Ns speakers 51-2 form one speaker layout for the middle band.
- the speaker 51-3-1 to the speaker 51-3-Nw output sound based on the speaker reproduction signal supplied from the amplification unit 33-3-1 to the amplification unit 33-3-Nw.
- Each of the Nw speakers 51-3 constituting the speaker system 22 is a speaker called a woofer, which mainly has a low band (low frequency) reproduction band.
- these Nw speakers 51-3 form one speaker layout for low band.
- the speaker system 22 is composed of a plurality of speakers 51 having different reproduction bands of high band, medium band, and low band. That is, a plurality of speakers 51 having different reproduction bands are mixedly arranged around the listener who listens to the content.
- the speaker system 22 including the speakers 51-1 to 51-3 is provided separately from the sound processing device 21
- the speaker system 22 is provided in the sound processing device 21 as a configuration. May be good. That is, the speaker system 22 may be included in the sound processing device 21.
- the rendering process is performed for each reproduction band of the speaker 51, that is, for each speaker layout of each reproduction band.
- the rendering processing unit 41-1 selects the above-mentioned selection mesh from the mesh formed by the Nt speakers 51-1. ..
- the rendering processing unit 41-2 selects the above-mentioned selection mesh from the mesh formed by the Ns speakers 51-2, and the rendering processing unit 41-3 selects the mesh formed by the Nw speakers 51-3.
- the above-mentioned selection mesh is selected from.
- the frequency characteristics of the HPF42, BPF43, and LPF44 that function as the band limiting processing unit, that is, the limiting band (passing band) are as shown in FIG. 3, for example.
- the horizontal axis indicates the frequency (Hz)
- the vertical axis indicates the sound pressure level (dB).
- the polygonal line L11 shows the frequency characteristic of HPF42
- the polygonal line L12 shows the frequency characteristic of BPF43
- the polygonal line L13 shows the frequency characteristic of LPF44.
- HPF42 high frequency pass filtering is performed to pass a component in a frequency band higher than other BPF43 and LPF44, that is, a high frequency component.
- mid-range pass filtering is performed to pass components in the frequency band higher than LPF44 and lower than HPF42, that is, mid-range components.
- LPF44 low-pass filtering is performed to pass components in a frequency band lower than those of other BPF43 and HPF42, that is, low-frequency components.
- the passbands of HPF42 and BPF43 cross over, and the passbands of BPF43 and LPF44 also cross over.
- the pass band of HPF42 and BPF43 and the pass band of BPF43 and LPF44 cross over
- the present invention is not limited to this.
- neither the pass band of HPF 42 and BPF 43 nor the pass band of BPF 43 and LPF 44 may cross over, and one of them may have a characteristic of crossing over.
- Nt HPF 42s have the same characteristics (frequency characteristics), but even if these Nt HPF 42s are filters (HPFs) having different characteristics from each other. good.
- the HPF 42 is not provided between the rendering processing unit 41-1 and the speaker 51-1 so that the speaker reproduction signal obtained by the rendering processing unit 41-1 is amplified by the D / A conversion unit 32-1. It may be supplied to the speaker 51-1 via the unit 33-1. That is, the sound based on the speaker reproduction signal may be reproduced by the speaker 51-1 without performing the filtering process (band limitation process) by the HPF42.
- Ns BPF43s have the same characteristics (frequency characteristics), but these BPF43s may have different characteristics from each other, and the rendering processing unit 41-2 and the speaker 51 may have different characteristics. BPF43 may not be provided between -2 and -2.
- Nw LPF44s have the same characteristics (frequency characteristics), but these LPF44s may have different characteristics from each other, and the rendering processing unit 41-3 and the speaker 51-3 may have different characteristics.
- the LPF 44 may not be provided between the and.
- step S11 the rendering processing unit 41-1 performs rendering processing on the high band speaker 51-1 based on the supplied N object data, and supplies the speaker reproduction signal obtained as a result to the HPF 42. ..
- rendering is performed on the speaker layout consisting of Nt speakers 51-1 to generate a speaker reproduction signal as an output audio signal.
- a mesh formed by Nt speakers 51-1 is used, and VBAP is performed as a rendering process.
- step S12 the HPF 42 performs filtering processing (band limitation processing) by the HPF on the speaker reproduction signal supplied from the rendering processing unit 41-1, and D / D / the speaker reproduction signal after the band limitation obtained as a result. It is supplied to the A conversion unit 32-1.
- the amplification unit 33-1 is supplied from the D / A conversion unit 32-1.
- the generated speaker reproduction signal is amplified and supplied to the speaker 51-1.
- step S13 the rendering processing unit 41-2 performs rendering processing on the speaker 51-2 for the middle band based on the supplied N object data, and supplies the speaker reproduction signal obtained as a result to the BPF 43. ..
- step S13 a mesh formed by Ns speakers 51-2 is used, and VBAP is performed as a rendering process.
- step S14 the BPF 43 performs filtering processing (band limitation processing) by BPF on the speaker reproduction signal supplied from the rendering processing unit 41-2, and D / D / the speaker reproduction signal after the band limitation obtained as a result. It is supplied to the A conversion unit 32-2.
- the generated speaker reproduction signal is amplified and supplied to the speaker 51-2.
- step S15 the rendering processing unit 41-3 performs rendering processing on the low-band speaker 51-3 based on the supplied N object data, and supplies the speaker reproduction signal obtained as a result to the LPF44. ..
- step S15 a mesh formed by Nw speakers 51-3 is used, and VBAP is performed as a rendering process.
- step S16 the LPF 44 performs filtering processing (band limitation processing) by the LPF on the speaker reproduction signal supplied from the rendering processing unit 41-3, and D / D / the speaker reproduction signal after the band limitation obtained as a result. It is supplied to the A conversion unit 32-3.
- the D / A conversion unit 32-3 D / A converts the speaker reproduction signal supplied from the LPF 44 and supplies it to the amplification unit 33-3, and the amplification unit 33-3 supplies the speaker reproduction signal from the D / A conversion unit 32-3.
- the generated speaker reproduction signal is amplified and supplied to the speaker 51-3.
- step S17 all the speakers 51 constituting the speaker system 22 output sound based on the speaker reproduction signal supplied from the amplification unit 33, and the reproduction process ends.
- the sound based on the speaker reproduction signal is output from all the speakers 51, the sound of N objects is reproduced in each reproduction band depending on the speaker layout of each reproduction band. Then, the sound image of each of those N objects is localized at the object position indicated by the position information included in the metadata of each object.
- the audio reproduction system 11 performs rendering processing for each reproduction band of the speaker 51, that is, for each speaker layout of each of the plurality of reproduction bands, and reproduces the content. By doing so, deterioration of sound quality due to the reproduction band of the speaker 51 can be suppressed, and audio reproduction with higher sound quality can be performed.
- speakers 51 having different reproduction bands are mixed.
- speaker layout configurations are prepared for each of a plurality of reproduction bands, and each object is rendered and reproduced for each reproduction band.
- the object is reproduced with an appropriate localization for each speaker layout of each reproduction band, and more appropriate rendering reproduction of the object-based audio is realized.
- This makes it possible to avoid deterioration of sound quality such as sound disappearing depending on the frequency band and localization position of the object, for example. That is, it is possible to reproduce audio with higher sound quality.
- the present invention is not limited to this, and for example, the object signal input to the rendering processing unit 41 may be subjected to band limiting filtering processing according to the target speaker layout.
- the audio reproduction system has, for example, the configuration shown in FIG. In FIG. 5, the same reference numerals are given to the portions corresponding to those in FIG. 2, and the description thereof will be omitted as appropriate.
- the audio reproduction system 81 shown in FIG. 5 has an audio processing device 91 and a speaker system 22.
- the sound processing device 91 includes a reproduction signal generation unit 101, a D / A conversion unit 32-1-1 to a D / A conversion unit 32-3-Nw, and an amplification unit 33-1-1 to an amplification unit 33-3. -Has Nw.
- the reproduction signal generation unit 101 includes HPF42-1 to HPF42-N, BPF43-1 to BPF43-N, LPF44-1 to LPF44-N, and a rendering processing unit 41-1 to a rendering processing unit 41-3. ..
- the configuration of the audio reproduction system 81 is different from the audio reproduction system 11 shown in FIG. 2 in that the acoustic processing apparatus 91 is provided in place of the acoustic processing apparatus 21, and is otherwise the same as the audio reproduction system 11. There is.
- the configuration of the sound processing device 91 is such that the reproduction signal generation unit 31 of the sound processing device 21 is replaced with the reproduction signal generation unit 101.
- HPF42 As described above, in the reproduction signal generation unit 31, HPF42, BPF43, and LPF44 are provided after the rendering processing unit 41.
- HPF42 HPF42, BPF43, and LPF44 are provided in front of the rendering processing unit 41.
- filtering processing is performed on the object signals of each of the N objects input to the rendering processing unit 41, so that N HPF42, BPF43, and LPF44 are each. It is provided one by one. That is, HPF42, BPF43, and LPF44 are provided for each object.
- each of HPF42-1 to HPF42-N performs filtering processing on each of the object signals of each of the supplied N object data, and renders the object signal containing only the high frequency component obtained as a result. It is supplied to the processing unit 41-1.
- the same filtering processing bandwidth limiting processing
- HPF42 in the reproduction signal generation unit 31 is performed.
- each of BPF43-1 to BPF43-N performs a filtering process on each of the object signals of each of the N supplied object data, and obtains an object signal containing only the mid-range component obtained as a result. It is supplied to the rendering processing unit 41-2.
- the same filtering processing band limiting processing
- BPF43 in the reproduction signal generation unit 31 is performed.
- Each of LPF44-1 to LPF44-N performs filtering processing on each of the object signals of each of the N supplied object data, and renders the object signal containing only the low frequency component obtained as a result in the rendering processing unit. Supply to 41-3.
- LPF44-1 to LPF44-N the same filtering processing (band limiting processing) as that of LPF44 in the reproduction signal generation unit 31 is performed.
- HPF42, BPF43, and LPF44 are provided for each speaker 51, whereas in the audio reproduction system 81, HPF42, BPF43, and LPF44 are provided for each object. There is.
- the audio reproduction system 81 is provided with N HPF42, BPF43, and LPF44 each.
- the N HPF 42s have the same frequency characteristics, but these N HPF 42s are filters (HPFs) having different characteristics from each other.
- the HPF 42 may not be provided in front of the rendering processing unit 41-1.
- the N BPF 43s have the same characteristics (frequency characteristics), but these BPF 43s may have different characteristics from each other, and may be in the preceding stage of the rendering processing unit 41-2.
- the BPF 43 may not be provided.
- N LPF 44s have the same characteristics (frequency characteristics), but these LPF 44s may have different characteristics from each other, and the LPF 44s are placed in front of the rendering processing unit 41-3. It may not be provided.
- each of HPF42-1 to HPF42-N performs filtering processing by HPF on each of the supplied object signals of N objects, and renders the resulting object signal after band limitation. It is supplied to the processing unit 41-1.
- step S42 the rendering processing unit 41-1 is used for high bandwidth based on the metadata of each of the N objects supplied and the signals of each of the N objects supplied from HPF42-1 to HPF42-N. Rendering is performed on the speaker 51-1.
- step S42 the same processing as in step S11 of FIG. 4 is performed.
- the rendering processing unit 41-1 supplies the speaker reproduction signal corresponding to each speaker 51-1 obtained by the rendering processing to the D / A conversion unit 32-1-1 to the D / A conversion unit 32-1-Nt. ..
- the amplification unit 33-1 is a D / A conversion unit.
- the speaker reproduction signal supplied from 32-1 is amplified and supplied to the speaker 51-1.
- each of BPF43-1 to BPF43-N performs filtering processing by BPF on each of the supplied object signals of N objects, and renders the resulting object signal after band limitation. It is supplied to the processing unit 41-2.
- step S44 the rendering processing unit 41-2 is used for the middle band based on the metadata of each of the N objects supplied and the signals of the N objects supplied from BPF43-1 to BPF43-N. Rendering is performed on the speaker 51-2.
- step S44 the same processing as in step S13 of FIG. 4 is performed.
- the rendering processing unit 41-2 supplies the speaker reproduction signal corresponding to each speaker 51-2 obtained by the rendering processing to the D / A conversion unit 32-2-1 to the D / A conversion unit 32-2-Ns. ..
- the speaker reproduction signal supplied from 32-2 is amplified and supplied to the speaker 51-2.
- each of LPF44-1 to LPF44-N performs filtering processing by LPF for each of the object signals of the supplied N objects, and renders the resulting object signal after band limitation. It is supplied to the processing unit 41-3.
- step S46 the rendering processing unit 41-3 is used for low bandwidth based on the metadata of each of the N objects supplied and the signals of each of the N objects supplied from LPF44-1 to LPF44-N. Rendering is performed on the speaker 51-3.
- step S46 the same processing as in step S15 of FIG. 4 is performed.
- the rendering processing unit 41-3 supplies the speaker reproduction signal corresponding to each speaker 51-3 obtained by the rendering processing to the D / A conversion unit 32-3-1 to the D / A conversion unit 32-3-Nw. ..
- the speaker reproduction signal supplied from 32-3 is amplified and supplied to the speaker 51-3.
- step S47 is performed and the reproduction process is terminated, but the process of step S47 is the same as the process of step S17 of FIG. Since there is, the explanation is omitted.
- the audio reproduction system 81 performs filtering processing for each object, and then performs rendering processing for each speaker layout of a plurality of reproduction bands to reproduce the content. By doing so, deterioration of sound quality due to the reproduction band of the speaker 51 can be suppressed, and audio reproduction with higher sound quality can be performed.
- the amount of processing is smaller than in the case of the audio playback system 11 especially when the number of objects constituting the content (number of objects N) is small. It can be suppressed.
- the processing amount of the filtering processing in HPF42, BPF43, and LPF44 is the same.
- the processing amount (processing number) of the filtering processing required in the audio reproduction system 81 is the number of objects N ⁇ 3.
- "3" is the number of rendering processing units 41.
- filtering processing is performed for the total number (number of times) of the total number of speakers 51 (Nt + Ns + Nw) constituting the speaker system 22.
- the number of processing of filtering processing is larger than that in the case of the audio reproduction system 11 by configuring the audio reproduction system 81. (Number of processing) can be reduced, and as a result, the overall processing amount can be suppressed to a small value.
- ⁇ Third embodiment> ⁇ Configuration example of audio playback system>
- whether the amount of processing is smaller when the filtering process is performed in the first stage or the second stage of the rendering process is determined by the number of objects N, the total number of speakers 51, and the number of types (reproduction bands) of the speakers 51 (rendering processing unit 41). It is determined by the number).
- the audio reproduction system is configured as shown in FIG. 7, for example.
- the same reference numerals are given to the portions corresponding to those in FIGS. 2 or 5, and the description thereof will be omitted as appropriate.
- the audio reproduction system 131 shown in FIG. 7 has an audio processing device 141 and a speaker system 22.
- the sound processing device 141 includes a selection unit 151, a reproduction signal generation unit 31, a reproduction signal generation unit 101, a D / A conversion unit 32-1-1 to a D / A conversion unit 32-3-Nw, and an amplification unit 33. It has 1-1 to the amplification unit 33-3-Nw.
- the reproduction signal generation unit 31 has the same configuration as in the case of FIG. 2, and the reproduction signal generation unit 101 has the same configuration as in the case of FIG.
- the object data of each of N objects is input to the selection unit 151.
- the selection unit 151 selects one of the reproduction signal generation unit 31 and the reproduction signal generation unit 101 as the output destination of the object data based on the number of objects N and the total number of speakers 51, and sends the selected output destination to the selected output destination. And output object data.
- the selection unit 151 causes the reproduction signal generation unit 31 to perform the rendering process and then the band limiting process, or causes the reproduction signal generation unit 101 to perform the band limiting process and then renders each object. Select whether to perform processing.
- either the reproduction signal generation unit 31 or the reproduction signal generation unit 101 generates a speaker reproduction signal based on the object data, and the speaker reproduction signal is supplied to the D / A conversion unit 32. Will be done.
- step S71 whether or not the selection unit 151 performs filtering processing prior to rendering processing based on the number N of supplied object data, the total number of speakers 51, and the number of reproduction bands (number of rendering processing units 41). Is determined. That is, the selection unit 151 selects the output destination of the supplied object data.
- the number of reproduction bands that is, the number of rendering processing units 41 is "3".
- the selection unit 151 determines that the filtering process is performed first.
- the selection unit 151 determines that the filtering process is performed after the rendering process when the number of objects N ⁇ 3 is equal to or greater than the total number of speakers 51 (Nt + Ns + Nw).
- step S71 If it is determined in step S71 that the filtering process is to be performed first, the selection unit 151 selects the reproduction signal generation unit 101 as the output destination of the supplied object data, and then the process proceeds to step S72.
- the selection unit 151 supplies the object signal of the supplied object data to the HPF42, BPF43, and LPF44 of the reproduction signal generation unit 101, and supplies the metadata of the object data to the rendering processing unit 41 of the reproduction signal generation unit 101. Supply.
- step S71 determines whether the filtering process will be performed later. If it is determined in step S71 that the filtering process will be performed later, the selection unit 151 selects the reproduction signal generation unit 31 as the output destination of the supplied object data, and then the process proceeds to step S78.
- the selection unit 151 supplies the supplied object data, that is, the object signal and the metadata, to the rendering processing unit 41 of the reproduction signal generation unit 31.
- step S84 is performed thereafter.
- step S84 all the speakers 51 constituting the speaker system 22 output sound based on the speaker reproduction signal supplied from the amplification unit 33, and the reproduction process ends.
- the audio reproduction system 131 selects the one of the reproduction signal generation unit 31 and the reproduction signal generation unit 101, which has a smaller processing amount, based on the number of objects N and the total number of speakers 51, and performs filtering processing. And render processing. That is, depending on the number of objects N and the total number of speakers 51, it is possible to switch between the reproduction signal generation unit 31 and the reproduction signal generation unit 101 to perform rendering processing and filtering processing.
- switching (selection) between the reproduction signal generation unit 31 and the reproduction signal generation unit 101 to perform the rendering process and the filtering process may be performed for each frame or the like.
- the reproduction signal generation unit 31 it is effective to limit the speaker reproduction signal according to the speaker layout for each reproduction band when the number of objects N is large.
- the reproduction signal generation unit 101 it is effective to limit the band of the object signal according to the speaker layout for each reproduction band when the number of objects N is small.
- the speaker layout for reproducing the sound of the object may be switched according to the contents of the object, that is, the characteristics of the object such as the sound source type of the object and the characteristics of the object signal.
- the audio reproduction system is configured as shown in FIG. 9, for example.
- the parts corresponding to the case in FIG. 2 are designated by the same reference numerals, and the description thereof will be omitted as appropriate.
- the audio reproduction system 181 shown in FIG. 9 has an audio processing device 191 and a speaker system 192.
- the sound processing device 191 includes a reproduction signal generation unit 201, a D / A conversion unit 32-1-1 to a D / A conversion unit 32-1-Nt, and a D / A conversion unit 32-3-1 to a D / A conversion unit. It has a 32-3-Nw, an amplification unit 33-1-1 to an amplification unit 33-1-Nt, and an amplification unit 33-3-1 to an amplification unit 33-3-Nw.
- the reproduction signal generation unit 201 has a determination unit 211, a switching unit 212, a rendering processing unit 41-1, and a rendering processing unit 41-3.
- the speaker system 192 has speakers 51-1-1 to 51-1-Nt, and speakers 51-3-1 to 51-3-Nw.
- the reproduction signal generation unit 201 is not provided with a filter that functions as a band limitation processing unit such as the HPF42.
- the speaker system 192 is provided with a speaker 51-1 which is a tweeter and a speaker 51-3 which is a woofer, but is not provided with a speaker 51-2 which is a squawker. Similar to the speaker system 22 described above, the speaker system 192 may be provided with a speaker 51-2 which is a squawker.
- Object data of each of N objects is supplied to the determination unit 211.
- the determination unit 211 determines which rendering processing unit 41 performs rendering processing, that is, which speaker layout is used for reproduction, for each object based on the object signal and metadata included in the supplied object data. Performs judgment processing.
- the rendering processing is performed only by the rendering processing unit 41-1, the rendering processing is performed only by the rendering processing unit 41-3, or the rendering processing unit 41-1 and the rendering processing unit 41. It is determined (determined) whether to perform the rendering process in both of -3. At this time, at least one of information about the object such as an object signal and metadata can be used to make a determination.
- the determination unit 211 supplies the supplied object data to the switching unit 212, controls the switching unit 212 based on the result of the determination processing, and transfers the object data to the rendering processing unit 41 according to the result of the determination processing. Supply.
- the determination process as a characteristic of the object, it may be determined for each object which reproduction band to render to the speaker layout based on the frequency characteristic of the object signal.
- the determination unit 211 performs frequency analysis on the supplied object signal by FFT (Fast Fourier Transform) or the like, and from the information indicating the frequency characteristics obtained as a result, to the speaker layout of which reproduction band. In other words, it is determined (determined) which rendering processing unit 41 performs the rendering processing.
- FFT Fast Fourier Transform
- the rendering process can be performed only by the rendering processing unit 41-3.
- each object is rendered by the rendering processing unit 41 corresponding to all the reproduction bands.
- the object signal contains only the low frequency component, the sound quality does not deteriorate even if the rendering process is performed only by the rendering processing unit 41-3.
- the object signal containing only the low frequency component is rendered by only the rendering processing unit 41-3 corresponding to the low band, so that the sound quality is not deteriorated.
- the amount of processing can be reduced.
- rendering processing can be performed by both the rendering processing unit 41-1 and the rendering processing unit 41-3.
- the metadata may contain information about the object.
- the metadata contains sound source type information indicating what type of sound source the object is, such as a musical instrument such as a guitar or vocals.
- the determination unit 211 determines (determines) which rendering processing unit 41 performs the rendering processing based on the sound source type information included in the metadata.
- the rendering processing unit 41-1 targeting the high band can perform rendering processing on the object. It should be noted that which rendering processing unit 41 renders which sound source type object may be predetermined. Further, the sound source type of the object may be specified from the file name of the object signal or the like.
- a content creator or the like specifies in advance which object should be rendered by which rendering processing unit 41, and the specified information indicating the specified result is included in the metadata as information about the object. You may.
- the determination unit 211 determines (determines) which rendering processing unit 41 renders the object based on the specified information included in the metadata.
- the designated information may be supplied to the determination unit 211 separately from the object data.
- the switching unit 212 switches the output destination of the object data supplied from the determination unit 211 for each object according to the control of the determination unit 211.
- the switching unit 212 either supplies the object data to the rendering processing unit 41-1 or the rendering processing unit 41-3 according to the control of the determination unit 211, or the rendering processing unit 41-1 and the rendering processing unit. Supply to 41-3.
- step S111 the determination unit 211 performs determination processing for each object based on the supplied object data.
- the determination process it is determined which reproduction band the rendering processing unit 41 corresponds to the rendering process based on at least the object signal and the metadata.
- the determination unit 211 supplies the supplied object data to the switching unit 212, and controls the output of the object data by the switching unit 212 based on the result of the determination process.
- step S112 the switching unit 212 supplies the object data supplied from the determination unit 211 according to the result of the determination process according to the control of the determination unit 211.
- the switching unit 212 transfers the object data supplied from the determination unit 211 to the rendering processing unit 41-1, the rendering processing unit 41-3, or the rendering processing unit 41-1 and the rendering processing unit 41-3 for each object. And supply.
- step S113 the rendering processing unit 41-1 performs rendering processing on the high-bandwidth speaker 51-1 based on the object data supplied from the switching unit 212, and D / A the speaker reproduction signal obtained as a result. It is supplied to the speaker 51-1 via the conversion unit 32-1 and the amplification unit 33-1.
- step S114 the rendering processing unit 41-3 performs rendering processing on the low-band speaker 51-3 based on the object data supplied from the switching unit 212, and D / A the speaker reproduction signal obtained as a result. It is supplied to the speaker 51-3 via the conversion unit 32-3 and the amplification unit 33-3.
- step S113 and step S114 the same processing as in step S11 and step S15 in FIG. 4 is performed.
- step S115 all the speakers 51 constituting the speaker system 192 output sound based on the speaker reproduction signal supplied from the amplification unit 33, and the reproduction process ends.
- the audio reproduction system 181 determines which reproduction band the rendering processing unit 41 corresponds to, based on at least one of the object signal and the information about the object such as metadata. Rendering is performed according to the determination result.
- the rendering process in the rendering processing unit 41 corresponding to the appropriate reproduction band can be selectively performed, and the audio reproduction with higher sound quality can be performed.
- the amount of processing due to multiple rendering processes can be increased as much as possible. It can be suppressed. That is, it is possible to omit the rendering process for the unnecessary reproduction band and reduce the processing amount.
- ⁇ Fifth Embodiment> ⁇ Configuration example of audio playback system>
- a subwoofer is added to reinforce the low frequency range during audio reproduction, and a method called bass management or bus management may be used.
- a low-frequency component signal is extracted from the playback signal of the main speaker by filtering processing, and the extracted signal is routed to one or more subwoofers. That is, the reproduction of the low frequency component is performed by one or a plurality of subwoofers.
- the same low frequency component is generally reproduced in all the subwoofers, so that the sense of localization of the object is impaired.
- the subwoofer that routes the low frequency component of which main speaker is divided for each subwoofer, and the low frequency component is reproduced according to the localization direction of the object can also be changed.
- the behavior of the entire system such as routing depends on the design, but the design becomes complicated and difficult.
- rendering processing is performed for each of a plurality of playback bands, and the content is played back in the speaker layout for each playback band, so that the object can be localized without the need for complicated design. It is possible to realize base management that can suppress the decrease in the.
- an audio signal of the LFE (Low Frequency Effect) channel for the subwoofer (hereinafter, also referred to as an LFE channel signal) may be prepared in advance.
- the gain of the LFE channel signal may be adjusted as appropriate and added to the speaker reproduction signal of the subwoofer.
- the audio reproduction system is as shown in FIG. 11, for example.
- the audio reproduction system 241 shown in FIG. 11 has an audio processing device 251 and a speaker system 252, and reproduces object-based audio contents based on the supplied object data.
- the content data consists of object data of N objects and channel-based LFE channel signals.
- the LFE channel signal is a channel-based audio signal, metadata including position information and the like is not supplied.
- the number of objects N can be any number.
- the sound processing device 251 includes a reproduction signal generation unit 261 and a D / A conversion unit 271-1 to D / A conversion unit 271-2-2-Nsw, and an amplification unit 272-1 to an amplification unit 272-2-Nsw. have.
- the reproduction signal generation unit 261 has a rendering processing unit 281-1, a rendering processing unit 281-2, HPF282-1 to HPF282-Nls, and LPF283-1 to LPF283-Nsw.
- the speaker system 252 has speakers 291-1 to 291-1-1Nls and speakers 291-2-1 to 291-2-2Nsw having different reproduction bands from each other.
- speaker 291-1 when it is not necessary to particularly distinguish between the speaker 291-1 to the speaker 291-1-Nls, it is also simply referred to as the speaker 291-1, and it is necessary to particularly distinguish the speaker 291-2-1 to the speaker 291-2-2Nsw. If there is no speaker, it is also simply referred to as a speaker 291-2.
- speaker 291 when it is not necessary to distinguish between the speaker 291-1 and the speaker 291-2, it is also simply referred to as a speaker 291.
- each of the Nls speakers 291-1 constituting the speaker system 252 is mainly a loudspeaker for a wide band having a wide band (broadband) from a relatively low band to a high range as a reproduction band. It is a speaker called.
- these Nls speakers 291-1 form one speaker layout for a wide band.
- each of the Nsw speakers 291-2 constituting the speaker system 252 is a speaker called a sub-woofer for low-frequency reinforcement having a low-frequency reproduction band of, for example, about 100 Hz or less.
- these Nsw speakers 291-2 form one speaker layout for low band.
- Object data of N objects constituting the content are supplied to the rendering processing unit 281-1 and the rendering processing unit 281-2, respectively.
- the rendering processing unit 281-1 and the rendering processing unit 281-2 perform rendering processing such as VBAP based on the object signal and metadata constituting the supplied object data. That is, in the rendering processing unit 281-1 and the rendering processing unit 281-2, the same processing as in the case of the rendering processing unit 41 is performed.
- each of the speaker reproduction signals having the speaker 291-1 to the speaker 291-1-Nls as the output destination is generated for each object. Then, the speaker reproduction signals for each object generated for the same speaker 291-1 are added to obtain the final speaker reproduction signal.
- the rendering process section 281-1 uses a mesh formed by Nls speakers 291-1.
- the rendering processing unit 281-1 supplies the final speaker reproduction signal generated for the speakers 291-1 to 291-1 to Nls to HPF282-1 to HPF282-Nls.
- the rendering processing unit 281-2 Similar to the rendering processing unit 281-1, the rendering processing unit 281-2 also reproduces the speaker of each speaker 291-2 with each of the speaker 291-2-1 to the speaker 291-2-2Nsw as the final output destination. Generate a signal.
- the rendering process section 281-2 uses a mesh formed by Nsw speakers 291-2.
- the LFE channel signal is supplied to the rendering processing unit 281-2.
- the LFE channel signal does not have localization information (position information), so that the rendering processing unit 281-2 distributes the LFE channel signal to all speakers 291-2 instead of rendering processing such as VBAP. It is output after being multiplied by a certain coefficient.
- the rendering processing unit 281-2 obtains the LFE channel signal by adjusting the gain of each speaker 291-2 with respect to the speaker reproduction signal corresponding to the speaker 291-2 obtained by the rendering processing by a predetermined coefficient.
- the generated signals are added together to obtain the final speaker reproduction signal.
- the coefficient used in the gain adjustment can be, for example, (1 / Nsw) 1/2 .
- the rendering processing unit 281-2 supplies the final speaker reproduction signal generated for the speakers 291-2-1 to 291-2-2Nsw to the LPF283-1 to LPF283-Nsw.
- rendering processing unit 281 when it is not necessary to distinguish between the rendering processing unit 281-1 and the rendering processing unit 281-2, it is also simply referred to as the rendering processing unit 281.
- HPF282-1 to HPF282-Nls are HPFs that pass a frequency band including at least the reproduction band of the speaker 291-1, that is, a frequency component of a relatively wide predetermined band.
- the HPF282-1 to HPF282-Nls perform filtering processing on the speaker reproduction signal supplied from the rendering processing unit 281-1, and D / A the speaker reproduction signal composed of the frequency components of the predetermined band obtained as a result. It is supplied to the conversion unit 271-1 to the D / A conversion unit 271-1-1Nls.
- HPF282 Similar to the HPF 42 shown in FIG. 2, the HPF 282 also functions as a band limiting processing unit that performs band limiting processing according to the reproduction band of the speaker 291-1.
- LPF283-1 to LPF283-Nsw are LPFs that pass frequency components in a frequency band including at least the reproduction band of the speaker 291-2, that is, a band of, for example, about 100 Hz or less.
- LPF283-1 to LPF283-Nsw perform filtering processing on the speaker reproduction signal supplied from the rendering processing unit 281-2, and D / A convert the speaker reproduction signal consisting of the frequency component of the low band obtained as a result. It is supplied to the unit 271-2-1 to the D / A conversion unit 271-2-2Nsw.
- LPF283 when it is not necessary to distinguish between LPF283-1 and LPF283-Nsw, they are also simply referred to as LPF283. Similar to the LPF44 shown in FIG. 2, the LPF283 also functions as a band limiting processing unit that performs band limiting processing according to the reproduction band of the speaker 291-2.
- the D / A conversion unit 271-1 to D / A conversion unit 271-1-1Nls D / A-convert the speaker reproduction signal supplied from HPF282-1 to HPF282-Nls, and the analog obtained as a result.
- the speaker reproduction signal of the above is supplied to the amplification unit 272-1-1 to the amplification unit 272-1-Nls.
- D / A conversion unit 271-1 when it is not necessary to distinguish D / A conversion unit 271-1 to D / A conversion unit 271-1-Nls, they are also simply referred to as D / A conversion unit 271-1.
- the speaker reproduction signal of the above is supplied to the amplification unit 272-2-1 to the amplification unit 272-2-Nsw.
- the D / A conversion unit 271-2-1 when it is not necessary to distinguish the D / A conversion unit 271-2-1 to the D / A conversion unit 271-2-Nsw, they are also simply referred to as the D / A conversion unit 271-2. Further, hereinafter, when it is not necessary to distinguish between the D / A conversion unit 271-1 and the D / A conversion unit 271-2, the D / A conversion unit 271 will be simply referred to as the D / A conversion unit 271.
- the amplification unit 272-1-1 to the amplification unit 272-1-Nls amplifies the speaker reproduction signal supplied from the D / A conversion unit 271-1-1 to the D / A conversion unit 271-1-1Nls, and the speaker is used. It is supplied to 291-1 to 291-1 to the speaker 291-1-Nls.
- the amplification unit 272-2-1 to the amplification unit 272-2-Nsw amplifies the speaker reproduction signal supplied from the D / A conversion unit 271-2-1 to the D / A conversion unit 271-2-2Nsw, and the speaker is used. It is supplied to 291-2-1 to the speaker 291-2-2Nsw.
- amplification unit 272-1-1 and the amplification unit 272-1-Nls are also simply referred to as the amplification unit 272-1, and the amplification unit 272-1 to the amplification unit 272-2.
- amplification unit 272-2 when it is not necessary to distinguish Nsw, it is simply referred to as amplification unit 272-2.
- amplification unit 272-1 when it is not necessary to distinguish between the amplification unit 272-1 and the amplification unit 272-2, they are also simply referred to as the amplification unit 272.
- the speaker 291-1 to speaker 291-1-Nls output sound based on the speaker reproduction signal supplied from the amplification unit 272-1-1 to the amplification unit 272-1-Nls.
- the speaker 291-2-1 to the speaker 291-2-Nsw output sound based on the speaker reproduction signal supplied from the amplification unit 272-2-1 to the amplification unit 272-2-Nsw.
- the speaker system 252 is composed of a plurality of speakers 291 having different reproduction bands from each other. That is, a plurality of speakers 291 having different reproduction bands are mixedly arranged around the listener who listens to the content.
- the speaker system 252 may be provided in the sound processing device 251.
- the frequency characteristics of the HPF282 and LPF283 that function as the band limiting processing unit are as shown in FIG. 12, for example.
- the horizontal axis indicates the frequency (Hz)
- the vertical axis indicates the sound pressure level (dB).
- the polygonal line L21 shows the frequency characteristic of HPF282
- the polygonal line L22 shows the frequency characteristic of LPF283.
- HPF282 high frequency pass filtering is performed to pass components in a frequency band higher than LPF283, that is, a wide frequency band of about 100 Hz or more.
- LPF283 low frequency pass filtering for passing a frequency band lower than the HPF282, that is, a component having a low frequency of about 100 Hz or less is performed.
- the passbands of HPF282 and LPF283 are crossed over, but the passbands of HPF282 and LPF283 may not be crossed over.
- the Nls HPF282 has the same characteristics (frequency characteristics), but even if these Nls HPF282s are filters (HPFs) having different characteristics from each other. good. Further, the HPF 282 may not be provided between the rendering processing unit 281-1 and the speaker 291-1.
- the Nsw LPF283s have the same characteristics (frequency characteristics), but these LPF283s may have different characteristics from each other, and the rendering processing unit 281-2 and the speaker 291 may be used.
- the LPF283 may not be provided between -2 and -2.
- step S141 the rendering processing unit 281-1 performs rendering processing on the wideband speaker 291-1 based on the supplied N object data, and supplies the speaker reproduction signal obtained as a result to the HPF282. For example, in step S141, the same processing as in step S11 of FIG. 4 is performed.
- step S142 the HPF 282 performs filtering processing (band limitation processing) by the HPF on the speaker reproduction signal supplied from the rendering processing unit 281-1.
- the HPF282 supplies the speaker reproduction signal after band limitation obtained by the filtering process to the speaker 291-1 via the D / A conversion unit 271-1 and the amplification unit 272-1.
- step S143 the rendering processing unit 281-2 performs rendering processing on the low-band speaker 291-2 based on the supplied N object data. For example, in step S143, the same processing as in step S15 of FIG. 4 is performed.
- step S144 the rendering processing unit 281-2 adjusts the gain of the supplied LFE channel signal according to a predetermined coefficient, adds it to the speaker reproduction signal, and supplies the final speaker reproduction signal obtained as a result to the LPF283.
- step S145 the LPF 283 performs filtering processing (band limitation processing) by the LPF on the speaker reproduction signal supplied from the rendering processing unit 281-2.
- the LPF283 supplies the band-limited speaker reproduction signal obtained by the filtering process to the speaker 291-2 via the D / A conversion unit 271-2 and the amplification unit 272-2.
- base management is realized by the processing of steps S143 and S144.
- the rendering processing unit 281-2 since the rendering processing unit 281-2 performs the rendering processing for the low band, it is possible to easily suppress the deterioration of the localization feeling of the object without requiring a complicated design.
- step S146 all the speakers 291 constituting the speaker system 252 output sound based on the speaker reproduction signal supplied from the amplification unit 272, and the reproduction process ends.
- the audio reproduction system 241 performs rendering processing for each reproduction band of the speaker 291, that is, for each speaker layout of a plurality of reproduction bands, and adjusts the gain of the LFE channel signal to adjust the gain of the LFE channel signal to obtain a low-band speaker reproduction signal. Add to.
- the optimum rendering according to the metadata of the object is realized.
- deterioration of sound quality due to the reproduction band of the speaker 291 is suppressed, and deterioration of the localization of the object is easily suppressed without requiring a complicated design, and higher sound quality audio reproduction is performed. Can be done.
- the series of processes described above can be executed by hardware or software.
- the programs constituting the software are installed on the computer.
- the computer includes a computer embedded in dedicated hardware and, for example, a general-purpose personal computer capable of executing various functions by installing various programs.
- FIG. 14 is a block diagram showing a configuration example of computer hardware that executes the above-mentioned series of processes programmatically.
- the CPU Central Processing Unit
- ROM Read Only Memory
- RAM Random Access Memory
- An input / output interface 505 is further connected to the bus 504.
- An input unit 506, an output unit 507, a recording unit 508, a communication unit 509, and a drive 510 are connected to the input / output interface 505.
- the input unit 506 includes a keyboard, a mouse, a microphone, an image pickup device, and the like.
- the output unit 507 includes a display, a speaker, and the like.
- the recording unit 508 includes a hard disk, a non-volatile memory, and the like.
- the communication unit 509 includes a network interface and the like.
- the drive 510 drives a removable recording medium 511 such as a magnetic disk, an optical disk, a magneto-optical disk, or a semiconductor memory.
- the CPU 501 loads the program recorded in the recording unit 508 into the RAM 503 via the input / output interface 505 and the bus 504 and executes the above-mentioned series. Is processed.
- the program executed by the computer (CPU501) can be recorded and provided on a removable recording medium 511 as a package medium or the like, for example.
- the program can also be provided via a wired or wireless transmission medium such as a local area network, the Internet, or digital satellite broadcasting.
- the program can be installed in the recording unit 508 via the input / output interface 505 by mounting the removable recording medium 511 in the drive 510. Further, the program can be received by the communication unit 509 and installed in the recording unit 508 via a wired or wireless transmission medium. In addition, the program can be pre-installed in the ROM 502 or the recording unit 508.
- the program executed by the computer may be a program in which processing is performed in chronological order according to the order described in the present specification, in parallel, or at a necessary timing such as when a call is made. It may be a program in which processing is performed.
- the embodiment of the present technology is not limited to the above-described embodiment, and various changes can be made without departing from the gist of the present technology.
- this technology can take a cloud computing configuration in which one function is shared by multiple devices via a network and processed jointly.
- each step described in the above flowchart can be executed by one device or shared by a plurality of devices.
- the plurality of processes included in the one step can be executed by one device or shared by a plurality of devices.
- this technology can also have the following configurations.
- a first rendering processing unit that performs rendering processing based on an audio signal and generates a first output audio signal for outputting sound by a plurality of first speakers, and a first rendering processing unit.
- a second rendering processing unit that performs rendering processing based on the audio signal and generates a second output audio signal for outputting sound by a plurality of second speakers having a reproduction band different from that of the first speaker.
- a sound processing device equipped with (2) A first band limiting processing unit that performs band limiting processing according to the reproduction band of the first speaker with respect to the first output audio signal.
- the audio processing apparatus according to (1) further comprising a second band limiting processing unit that performs band limiting processing on the second output audio signal according to the reproduction band of the second speaker.
- a third band limiting processing unit that performs band limiting processing according to the reproduction band of the first speaker with respect to the audio signal, and a third band limiting processing unit.
- a third rendering processing unit that generates A fourth band limiting processing unit that performs band limiting processing according to the reproduction band of the second speaker with respect to the audio signal, and a fourth band limiting processing unit.
- the fourth rendering processing unit that generates Whether the third band limiting processing unit and the fourth band limiting processing unit are to perform the band limiting processing, and the third rendering processing unit and the fourth rendering processing unit are to perform the rendering processing. Alternatively, whether the first rendering processing unit and the second rendering processing unit are to perform the rendering processing, and the first band limiting processing unit and the second band limiting processing unit are to perform the band limiting processing.
- the sound processing apparatus according to (2) further comprising a selection unit for selection.
- the selection unit makes the selection based on the number of the audio signals and the total number of the first speaker and the second speaker.
- a first band limiting processing unit that performs band limiting processing according to the reproduction band of the first speaker with respect to the audio signal, and a first band limiting processing unit. Further, the audio signal is further provided with a second band limiting processing unit that performs band limiting processing according to the reproduction band of the second speaker.
- the first rendering processing unit performs rendering processing based on the first band limiting signal obtained by the band limiting processing by the first band limiting processing unit.
- the sound processing apparatus according to (1), wherein the second rendering processing unit performs rendering processing based on a second band limiting signal obtained by the band limiting processing by the second band limiting processing unit.
- the rendering process based on the audio signal is performed by the first rendering processing unit, or the second rendering process is performed.
- (1), (2), or (5) further includes a determination unit for determining whether to have the rendering processing unit perform the processing, or both the first rendering processing unit and the second rendering processing unit.
- the sound processing device described in. (7) The acoustic processing device according to (6), wherein the determination unit makes the determination based on the frequency characteristics of the audio signal.
- the audio signal is an object signal of an audio object.
- Sound processing equipment (10) The acoustic processing apparatus according to (9), wherein the metadata includes position information indicating the position of the audio object.
- the position information is information indicating a relative position of the audio object with respect to a predetermined listening position.
- the second rendering processing unit adds the second output audio signal obtained by the rendering processing and the channel-based audio signal to obtain the final second output audio signal (9).
- (13) The acoustic processing apparatus according to (12), wherein the channel-based audio signal is an audio signal of an LFE channel.
- the sound processing apparatus according to any one of (1) to (13), wherein the first rendering processing unit and the second rendering processing unit perform processing using VBAP as rendering processing.
- the sound processing apparatus according to any one of (1) to (14), further comprising the plurality of the first speaker and the plurality of the second speakers.
- the sound processing device Rendering is performed based on the audio signal, and a first output audio signal for outputting sound by a plurality of first speakers is generated.
- An acoustic processing method that performs rendering processing based on the audio signal and generates a second output audio signal for outputting sound by a plurality of second speakers having a reproduction band different from that of the first speaker.
- Rendering is performed based on the audio signal, and a first output audio signal for outputting sound by a plurality of first speakers is generated.
- a computer performs a process including a step of performing rendering processing based on the audio signal and generating a second output audio signal for outputting sound by a plurality of second speakers having a reproduction band different from that of the first speaker. Program to be executed by.
- Audio playback system 21 Sound processing device, 22 Speaker system, 41-1 to 41-3, 41 Rendering processing unit, 42-1 to 42-Nt, 42 HPF, 43-1 to 43-Ns, 43 BPF, 44 -1 to 44-Nw, 44 LPF, 151 selection unit, 211 judgment unit
Landscapes
- Physics & Mathematics (AREA)
- Engineering & Computer Science (AREA)
- Acoustics & Sound (AREA)
- Signal Processing (AREA)
- Health & Medical Sciences (AREA)
- General Health & Medical Sciences (AREA)
- Otolaryngology (AREA)
- Circuit For Audible Band Transducer (AREA)
- Stereophonic System (AREA)
Abstract
Description
〈本技術について〉
本技術は、互いに異なる複数の再生帯域を有するスピーカからなるスピーカシステムでオブジェクトベースオーディオを再生する場合に、同じ再生帯域を有するスピーカからなるスピーカレイアウトごとにレンダリング処理を行うことで、より高音質なオーディオ再生を行うものである。
図2は、本技術を適用したオーディオ再生システムの一実施の形態の構成例を示す図である。
次に、オーディオ再生システム11の動作について説明する。すなわち、以下、図4のフローチャートを参照して、オーディオ再生システム11による再生処理について説明する。この再生処理は、コンテンツを構成するN個のオブジェクトのオブジェクトデータが各レンダリング処理部41に供給されると開始される。
〈オーディオ再生システムの構成例〉
なお、以上においてはレンダリング処理部41の出力に対して、対象となるスピーカレイアウトに応じた帯域制限のフィルタリング処理が行われる例について説明した。
次に、図6のフローチャートを参照して、オーディオ再生システム81による再生処理について説明する。
〈オーディオ再生システムの構成例〉
ところで、フィルタリング処理をレンダリング処理の前段と後段のどちらで行う方が処理量が少なくなるかはオブジェクト数Nや、スピーカ51の総数、スピーカ51の種類(再生帯域)の数(レンダリング処理部41の数)によって定まる。
次に、図8のフローチャートを参照して、オーディオ再生システム131による再生処理について説明する。この再生処理は、選択部151にコンテンツを構成するN個のオブジェクトのオブジェクトデータが供給されると開始される。
〈オーディオ再生システムの構成例〉
また、オブジェクトの内容、すなわちオブジェクトの音源種別やオブジェクト信号の特性など、オブジェクトが有する特徴によって、そのオブジェクトの音を再生するスピーカレイアウトを切り替えるようにしてもよい。
次に、図10のフローチャートを参照して、オーディオ再生システム181による再生処理について説明する。この再生処理は、判定部211にコンテンツを構成するN個のオブジェクトのオブジェクトデータが供給されると開始される。
〈オーディオ再生システムの構成例〉
ところで、オーディオ再生時の低域の補強のためにサブウーハが追加され、ベース・マネジメントやバス・マネジメントなどと呼ばれる手法が用いられることがある。
次に、図13のフローチャートを参照して、オーディオ再生システム241による再生処理について説明する。
ところで、上述した一連の処理は、ハードウェアにより実行することもできるし、ソフトウェアにより実行することもできる。一連の処理をソフトウェアにより実行する場合には、そのソフトウェアを構成するプログラムが、コンピュータにインストールされる。ここで、コンピュータには、専用のハードウェアに組み込まれているコンピュータや、各種のプログラムをインストールすることで、各種の機能を実行することが可能な、例えば汎用のパーソナルコンピュータなどが含まれる。
オーディオ信号に基づいてレンダリング処理を行い、複数の第1のスピーカにより音を出力するための第1の出力オーディオ信号を生成する第1のレンダリング処理部と、
前記オーディオ信号に基づいてレンダリング処理を行い、前記第1のスピーカとは再生帯域が異なる複数の第2のスピーカにより音を出力するための第2の出力オーディオ信号を生成する第2のレンダリング処理部と
を備える音響処理装置。
(2)
前記第1の出力オーディオ信号に対して、前記第1のスピーカの再生帯域に応じた帯域制限処理を行う第1の帯域制限処理部と、
前記第2の出力オーディオ信号に対して、前記第2のスピーカの再生帯域に応じた帯域制限処理を行う第2の帯域制限処理部と
をさらに備える(1)に記載の音響処理装置。
(3)
前記オーディオ信号に対して、前記第1のスピーカの再生帯域に応じた帯域制限処理を行う第3の帯域制限処理部と、
前記第3の帯域制限処理部による帯域制限処理により得られた第1の帯域制限信号に基づいてレンダリング処理を行い、複数の前記第1のスピーカにより音を出力するための第3の出力オーディオ信号を生成する第3のレンダリング処理部と、
前記オーディオ信号に対して、前記第2のスピーカの再生帯域に応じた帯域制限処理を行う第4の帯域制限処理部と、
前記第4の帯域制限処理部による帯域制限処理により得られた第2の帯域制限信号に基づいてレンダリング処理を行い、複数の前記第2のスピーカにより音を出力するための第4の出力オーディオ信号を生成する第4のレンダリング処理部と、
前記第3の帯域制限処理部および前記第4の帯域制限処理部に帯域制限処理を行わせるとともに、前記第3のレンダリング処理部および前記第4のレンダリング処理部にレンダリング処理を行わせるか、
または
前記第1のレンダリング処理部および前記第2のレンダリング処理部にレンダリング処理を行わせるとともに、前記第1の帯域制限処理部および前記第2の帯域制限処理部に帯域制限処理を行わせるか
を選択する選択部と
をさらに備える(2)に記載の音響処理装置。
(4)
前記選択部は、前記オーディオ信号の数と、前記第1のスピーカおよび前記第2のスピーカの総数とに基づいて前記選択を行う
(3)に記載の音響処理装置。
(5)
前記オーディオ信号に対して、前記第1のスピーカの再生帯域に応じた帯域制限処理を行う第1の帯域制限処理部と、
前記オーディオ信号に対して、前記第2のスピーカの再生帯域に応じた帯域制限処理を行う第2の帯域制限処理部と
をさらに備え、
前記第1のレンダリング処理部は、前記第1の帯域制限処理部による帯域制限処理により得られた第1の帯域制限信号に基づいてレンダリング処理を行い、
前記第2のレンダリング処理部は、前記第2の帯域制限処理部による帯域制限処理により得られた第2の帯域制限信号に基づいてレンダリング処理を行う
(1)に記載の音響処理装置。
(6)
前記オーディオ信号、および前記オーディオ信号に関する情報の少なくとも何れか一方に基づいて、前記オーディオ信号ごとに、前記オーディオ信号に基づくレンダリング処理を、前記第1のレンダリング処理部に行わせるか、前記第2のレンダリング処理部に行わせるか、または前記第1のレンダリング処理部と前記第2のレンダリング処理部の両方に行わせるかを判定する判定部をさらに備える
(1)、(2)、または(5)に記載の音響処理装置。
(7)
前記判定部は、前記オーディオ信号の周波数特性に基づいて前記判定を行う
(6)に記載の音響処理装置。
(8)
前記判定部は、前記オーディオ信号の音源種別を示す情報に基づいて前記判定を行う
(6)または(7)に記載の音響処理装置。
(9)
前記オーディオ信号は、オーディオオブジェクトのオブジェクト信号であり、
前記第1のレンダリング処理部および前記第2のレンダリング処理部は、前記オーディオ信号、および前記オーディオ信号のメタデータに基づいてレンダリング処理を行う
(1)乃至(8)の何れか一項に記載の音響処理装置。
(10)
前記メタデータには、前記オーディオオブジェクトの位置を示す位置情報が含まれている
(9)に記載の音響処理装置。
(11)
前記位置情報は、所定の受聴位置を基準とした前記オーディオオブジェクトの相対的な位置を示す情報である
(10)に記載の音響処理装置。
(12)
前記第2のレンダリング処理部は、レンダリング処理により得られた前記第2の出力オーディオ信号と、チャネルベースのオーディオ信号とを加算して、最終的な前記第2の出力オーディオ信号とする
(9)乃至(11)の何れか一項に記載の音響処理装置。
(13)
前記チャネルベースのオーディオ信号は、LFEチャネルのオーディオ信号である
(12)に記載の音響処理装置。
(14)
前記第1のレンダリング処理部および前記第2のレンダリング処理部は、レンダリング処理としてVBAPを用いた処理を行う
(1)乃至(13)の何れか一項に記載の音響処理装置。
(15)
前記複数の前記第1のスピーカおよび前記複数の前記第2のスピーカをさらに備える
(1)乃至(14)の何れか一項に記載の音響処理装置。
(16)
音響処理装置が、
オーディオ信号に基づいてレンダリング処理を行い、複数の第1のスピーカにより音を出力するための第1の出力オーディオ信号を生成し、
前記オーディオ信号に基づいてレンダリング処理を行い、前記第1のスピーカとは再生帯域が異なる複数の第2のスピーカにより音を出力するための第2の出力オーディオ信号を生成する
音響処理方法。
(17)
オーディオ信号に基づいてレンダリング処理を行い、複数の第1のスピーカにより音を出力するための第1の出力オーディオ信号を生成し、
前記オーディオ信号に基づいてレンダリング処理を行い、前記第1のスピーカとは再生帯域が異なる複数の第2のスピーカにより音を出力するための第2の出力オーディオ信号を生成する
ステップを含む処理をコンピュータに実行させるプログラム。
Claims (17)
- オーディオ信号に基づいてレンダリング処理を行い、複数の第1のスピーカにより音を出力するための第1の出力オーディオ信号を生成する第1のレンダリング処理部と、
前記オーディオ信号に基づいてレンダリング処理を行い、前記第1のスピーカとは再生帯域が異なる複数の第2のスピーカにより音を出力するための第2の出力オーディオ信号を生成する第2のレンダリング処理部と
を備える音響処理装置。 - 前記第1の出力オーディオ信号に対して、前記第1のスピーカの再生帯域に応じた帯域制限処理を行う第1の帯域制限処理部と、
前記第2の出力オーディオ信号に対して、前記第2のスピーカの再生帯域に応じた帯域制限処理を行う第2の帯域制限処理部と
をさらに備える請求項1に記載の音響処理装置。 - 前記オーディオ信号に対して、前記第1のスピーカの再生帯域に応じた帯域制限処理を行う第3の帯域制限処理部と、
前記第3の帯域制限処理部による帯域制限処理により得られた第1の帯域制限信号に基づいてレンダリング処理を行い、複数の前記第1のスピーカにより音を出力するための第3の出力オーディオ信号を生成する第3のレンダリング処理部と、
前記オーディオ信号に対して、前記第2のスピーカの再生帯域に応じた帯域制限処理を行う第4の帯域制限処理部と、
前記第4の帯域制限処理部による帯域制限処理により得られた第2の帯域制限信号に基づいてレンダリング処理を行い、複数の前記第2のスピーカにより音を出力するための第4の出力オーディオ信号を生成する第4のレンダリング処理部と、
前記第3の帯域制限処理部および前記第4の帯域制限処理部に帯域制限処理を行わせるとともに、前記第3のレンダリング処理部および前記第4のレンダリング処理部にレンダリング処理を行わせるか、
または
前記第1のレンダリング処理部および前記第2のレンダリング処理部にレンダリング処理を行わせるとともに、前記第1の帯域制限処理部および前記第2の帯域制限処理部に帯域制限処理を行わせるか
を選択する選択部と
をさらに備える請求項2に記載の音響処理装置。 - 前記選択部は、前記オーディオ信号の数と、前記第1のスピーカおよび前記第2のスピーカの総数とに基づいて前記選択を行う
請求項3に記載の音響処理装置。 - 前記オーディオ信号に対して、前記第1のスピーカの再生帯域に応じた帯域制限処理を行う第1の帯域制限処理部と、
前記オーディオ信号に対して、前記第2のスピーカの再生帯域に応じた帯域制限処理を行う第2の帯域制限処理部と
をさらに備え、
前記第1のレンダリング処理部は、前記第1の帯域制限処理部による帯域制限処理により得られた第1の帯域制限信号に基づいてレンダリング処理を行い、
前記第2のレンダリング処理部は、前記第2の帯域制限処理部による帯域制限処理により得られた第2の帯域制限信号に基づいてレンダリング処理を行う
請求項1に記載の音響処理装置。 - 前記オーディオ信号、および前記オーディオ信号に関する情報の少なくとも何れか一方に基づいて、前記オーディオ信号ごとに、前記オーディオ信号に基づくレンダリング処理を、前記第1のレンダリング処理部に行わせるか、前記第2のレンダリング処理部に行わせるか、または前記第1のレンダリング処理部と前記第2のレンダリング処理部の両方に行わせるかを判定する判定部をさらに備える
請求項1に記載の音響処理装置。 - 前記判定部は、前記オーディオ信号の周波数特性に基づいて前記判定を行う
請求項6に記載の音響処理装置。 - 前記判定部は、前記オーディオ信号の音源種別を示す情報に基づいて前記判定を行う
請求項6に記載の音響処理装置。 - 前記オーディオ信号は、オーディオオブジェクトのオブジェクト信号であり、
前記第1のレンダリング処理部および前記第2のレンダリング処理部は、前記オーディオ信号、および前記オーディオ信号のメタデータに基づいてレンダリング処理を行う
請求項1に記載の音響処理装置。 - 前記メタデータには、前記オーディオオブジェクトの位置を示す位置情報が含まれている
請求項9に記載の音響処理装置。 - 前記位置情報は、所定の受聴位置を基準とした前記オーディオオブジェクトの相対的な位置を示す情報である
請求項10に記載の音響処理装置。 - 前記第2のレンダリング処理部は、レンダリング処理により得られた前記第2の出力オーディオ信号と、チャネルベースのオーディオ信号とを加算して、最終的な前記第2の出力オーディオ信号とする
請求項9に記載の音響処理装置。 - 前記チャネルベースのオーディオ信号は、LFEチャネルのオーディオ信号である
請求項12に記載の音響処理装置。 - 前記第1のレンダリング処理部および前記第2のレンダリング処理部は、レンダリング処理としてVBAPを用いた処理を行う
請求項1に記載の音響処理装置。 - 前記複数の前記第1のスピーカおよび前記複数の前記第2のスピーカをさらに備える
請求項1に記載の音響処理装置。 - 音響処理装置が、
オーディオ信号に基づいてレンダリング処理を行い、複数の第1のスピーカにより音を出力するための第1の出力オーディオ信号を生成し、
前記オーディオ信号に基づいてレンダリング処理を行い、前記第1のスピーカとは再生帯域が異なる複数の第2のスピーカにより音を出力するための第2の出力オーディオ信号を生成する
音響処理方法。 - オーディオ信号に基づいてレンダリング処理を行い、複数の第1のスピーカにより音を出力するための第1の出力オーディオ信号を生成し、
前記オーディオ信号に基づいてレンダリング処理を行い、前記第1のスピーカとは再生帯域が異なる複数の第2のスピーカにより音を出力するための第2の出力オーディオ信号を生成する
処理をコンピュータに実行させるプログラム。
Priority Applications (7)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US18/023,882 US20230336913A1 (en) | 2020-09-09 | 2021-08-27 | Acoustic processing device, method, and program |
CN202180053759.4A CN116114267A (zh) | 2020-09-09 | 2021-08-27 | 声学处理装置、方法和程序 |
EP21866561.0A EP4213505A4 (en) | 2020-09-09 | 2021-08-27 | APPARATUS, METHOD AND PROGRAM FOR ACOUSTIC PROCESSING |
BR112023003964A BR112023003964A2 (pt) | 2020-09-09 | 2021-08-27 | Dispositivo e método de processamento acústico, e, programa |
KR1020237005842A KR20230062814A (ko) | 2020-09-09 | 2021-08-27 | 음향 처리 장치 및 방법, 그리고 프로그램 |
MX2023002587A MX2023002587A (es) | 2020-09-09 | 2021-08-27 | Dispositivo y metodo de procesamiento acustico y programa. |
JP2022547497A JPWO2022054602A1 (ja) | 2020-09-09 | 2021-08-27 |
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
JP2020151446 | 2020-09-09 | ||
JP2020-151446 | 2020-09-09 |
Publications (1)
Publication Number | Publication Date |
---|---|
WO2022054602A1 true WO2022054602A1 (ja) | 2022-03-17 |
Family
ID=80631626
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
PCT/JP2021/031449 WO2022054602A1 (ja) | 2020-09-09 | 2021-08-27 | 音響処理装置および方法、並びにプログラム |
Country Status (8)
Country | Link |
---|---|
US (1) | US20230336913A1 (ja) |
EP (1) | EP4213505A4 (ja) |
JP (1) | JPWO2022054602A1 (ja) |
KR (1) | KR20230062814A (ja) |
CN (1) | CN116114267A (ja) |
BR (1) | BR112023003964A2 (ja) |
MX (1) | MX2023002587A (ja) |
WO (1) | WO2022054602A1 (ja) |
Citations (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP2011529658A (ja) * | 2008-07-28 | 2011-12-08 | コーニンクレッカ フィリップス エレクトロニクス エヌ ヴィ | オーディオシステム及びその操作方法 |
US20160066118A1 (en) * | 2013-04-15 | 2016-03-03 | Intellectual Discovery Co., Ltd. | Audio signal processing method using generating virtual object |
JP2018527825A (ja) * | 2015-08-14 | 2018-09-20 | ディーティーエス・インコーポレイテッドDTS,Inc. | オブジェクトベースのオーディオのための低音管理 |
Family Cites Families (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
KR102356246B1 (ko) * | 2014-01-16 | 2022-02-08 | 소니그룹주식회사 | 음성 처리 장치 및 방법, 그리고 프로그램 |
US11477601B2 (en) * | 2018-10-16 | 2022-10-18 | Dolby Laboratories Licensing Corporation | Methods and devices for bass management |
-
2021
- 2021-08-27 CN CN202180053759.4A patent/CN116114267A/zh active Pending
- 2021-08-27 WO PCT/JP2021/031449 patent/WO2022054602A1/ja unknown
- 2021-08-27 US US18/023,882 patent/US20230336913A1/en active Pending
- 2021-08-27 JP JP2022547497A patent/JPWO2022054602A1/ja active Pending
- 2021-08-27 BR BR112023003964A patent/BR112023003964A2/pt unknown
- 2021-08-27 MX MX2023002587A patent/MX2023002587A/es unknown
- 2021-08-27 EP EP21866561.0A patent/EP4213505A4/en active Pending
- 2021-08-27 KR KR1020237005842A patent/KR20230062814A/ko unknown
Patent Citations (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP2011529658A (ja) * | 2008-07-28 | 2011-12-08 | コーニンクレッカ フィリップス エレクトロニクス エヌ ヴィ | オーディオシステム及びその操作方法 |
US20160066118A1 (en) * | 2013-04-15 | 2016-03-03 | Intellectual Discovery Co., Ltd. | Audio signal processing method using generating virtual object |
JP2018527825A (ja) * | 2015-08-14 | 2018-09-20 | ディーティーエス・インコーポレイテッドDTS,Inc. | オブジェクトベースのオーディオのための低音管理 |
Non-Patent Citations (2)
Title |
---|
See also references of EP4213505A4 |
VILLE PULKKI: "Virtual Sound Source Positioning Using Vector Base Amplitude Panning", JOURNAL OF AES, vol. 45, no. 6, 1997, pages 456 - 466 |
Also Published As
Publication number | Publication date |
---|---|
MX2023002587A (es) | 2023-03-22 |
EP4213505A4 (en) | 2024-03-06 |
BR112023003964A2 (pt) | 2023-04-11 |
EP4213505A1 (en) | 2023-07-19 |
CN116114267A (zh) | 2023-05-12 |
KR20230062814A (ko) | 2023-05-09 |
JPWO2022054602A1 (ja) | 2022-03-17 |
US20230336913A1 (en) | 2023-10-19 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US11785410B2 (en) | Reproduction apparatus and reproduction method | |
RU2682864C1 (ru) | Устройство и способ обработки аудиоданных, и его программа | |
TWI489887B (zh) | 用於喇叭或耳機播放之虛擬音訊處理技術 | |
JP5496235B2 (ja) | 多重オーディオチャンネル群の再現の向上 | |
JP6918777B2 (ja) | オブジェクトベースのオーディオのための低音管理 | |
GB2549532A (en) | Merging audio signals with spatial metadata | |
KR20140010468A (ko) | 오디오 신호의 공간 추출 시스템 | |
US20220167103A1 (en) | Computer system for realizing customized being-there in assocation with audio and method thereof | |
JP5931182B2 (ja) | 付加的な出力チャンネルを提供するためのステレオ出力信号を生成する装置、方法およびコンピュータプログラム | |
WO2022054602A1 (ja) | 音響処理装置および方法、並びにプログラム | |
WO2021140959A1 (ja) | 符号化装置および方法、復号装置および方法、並びにプログラム | |
JP2004511927A (ja) | パラメトリックスピーカー用電力アンプ | |
JP6699280B2 (ja) | 音響再生装置 | |
JP7517500B2 (ja) | 再生装置、再生方法、およびプログラム | |
US11924628B1 (en) | Virtual surround sound process for loudspeaker systems | |
JP2012049652A (ja) | マルチチャネルオーディオ再生装置およびマルチチャネルオーディオ再生方法 | |
JP6819236B2 (ja) | 音処理装置、音処理方法、及びプログラム | |
JP6834398B2 (ja) | 音処理装置、音処理方法、及びプログラム | |
JP2010118977A (ja) | 音像定位制御装置および音像定位制御方法 | |
KR20230119192A (ko) | 스테레오 헤드폰 심리음향 음 위치측정 시스템 및 이를사용한 스테레오 심리음향 음 신호를 재구성하기 위한 방법 | |
RU2384973C1 (ru) | Устройство и способ синтезирования трех выходных каналов, используя два входных канала | |
JP2006174078A (ja) | オーディオ信号処理方法及び装置 | |
JP2019201308A (ja) | 音響制御装置、方法及びプログラム | |
KR20110102719A (ko) | 오디오 업믹싱 장치 및 방법 |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
121 | Ep: the epo has been informed by wipo that ep was designated in this application |
Ref document number: 21866561 Country of ref document: EP Kind code of ref document: A1 |
|
ENP | Entry into the national phase |
Ref document number: 2022547497 Country of ref document: JP Kind code of ref document: A |
|
REG | Reference to national code |
Ref country code: BR Ref legal event code: B01A Ref document number: 112023003964 Country of ref document: BR |
|
ENP | Entry into the national phase |
Ref document number: 112023003964 Country of ref document: BR Kind code of ref document: A2 Effective date: 20230302 |
|
NENP | Non-entry into the national phase |
Ref country code: DE |
|
ENP | Entry into the national phase |
Ref document number: 2021866561 Country of ref document: EP Effective date: 20230411 |