CN112019994B - Method and device for constructing in-vehicle diffusion sound field environment based on virtual loudspeaker - Google Patents
Method and device for constructing in-vehicle diffusion sound field environment based on virtual loudspeaker Download PDFInfo
- Publication number
- CN112019994B CN112019994B CN202010805149.2A CN202010805149A CN112019994B CN 112019994 B CN112019994 B CN 112019994B CN 202010805149 A CN202010805149 A CN 202010805149A CN 112019994 B CN112019994 B CN 112019994B
- Authority
- CN
- China
- Prior art keywords
- information
- audio
- constructing
- hrtf
- field environment
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Active
Links
Images
Classifications
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04S—STEREOPHONIC SYSTEMS
- H04S7/00—Indicating arrangements; Control arrangements, e.g. balance control
- H04S7/30—Control circuits for electronic adaptation of the sound field
- H04S7/302—Electronic adaptation of stereophonic sound system to listener position or orientation
- H04S7/303—Tracking of listener position or orientation
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04S—STEREOPHONIC SYSTEMS
- H04S3/00—Systems employing more than two channels, e.g. quadraphonic
- H04S3/006—Systems employing more than two channels, e.g. quadraphonic in which a plurality of audio signals are transformed in a combination of audio signals and modulated signals, e.g. CD-4 systems
Landscapes
- Engineering & Computer Science (AREA)
- Physics & Mathematics (AREA)
- Acoustics & Sound (AREA)
- Signal Processing (AREA)
- Multimedia (AREA)
- Stereophonic System (AREA)
Abstract
The invention belongs to the technical field of digital signal processing, and discloses a method and a device for constructing a diffusion sound field environment in a vehicle based on virtual loudspeakers. The device comprises an audio decorrelation processing unit, an equalization processing unit, an HRTF filter unit and an audio gain superposition unit. The method solves the problem that the reality of the ASD sound is reduced due to the fact that passengers in the automobile perceive the ASD sound direction in practice, can achieve the surround effect of an acoustic field in the automobile on an acoustic space, and meets the set requirement of the quality of the sound in the automobile.
Description
Technical Field
The invention relates to the technical field of digital signal processing, in particular to a method and a device for constructing an in-vehicle diffuse sound field environment based on a virtual loudspeaker.
Background
With the development of the automobile NVH (Noise, Vibration, Harshness) industry, people have new requirements on the automobile sound environment, and the pursuit of sound quality by various large host factories and consumers brought by the popularization of electric vehicles. The ASD (Active Sound Design) meets the driving experience of consumers in a form of simulating engine Sound, solves the problem of excessive silence of the Sound environment in the electric vehicle and improves the comfort level in the vehicle.
In order to design a more realistic engine sound effect, in addition to the requirement of closing the fidelity of the sound, the design of the speakers in the vehicle is also needed, so that passengers cannot feel that the engine sound is emitted from the speakers in the vehicle. The prior art is usually realized by additionally increasing the number of loudspeakers, so that the production cost of the automobile is increased.
Disclosure of Invention
The embodiment of the application provides a method and a device for constructing an in-vehicle diffusion sound field environment based on a virtual loudspeaker, and solves the problem that in practice, the reality of ASD sound is reduced due to the fact that in-vehicle passengers perceive the ASD sound direction.
The embodiment of the application provides a method for constructing an in-vehicle diffusion sound field environment based on a virtual loudspeaker, which comprises the following steps:
performing decorrelation processing on input original audio information to obtain a plurality of different first audio information;
constructing a target diffusion sound field environment, obtaining first HRTF parameter information by using artificial false head measurement, and performing frequency equalization processing on the first HRTF parameter information to obtain second HRTF parameter information;
building an HRTF filter according to the second HRTF parameter information, and filtering all the first audio information by using the HRTF filter to obtain a plurality of left audio sub-information and a plurality of right audio sub-information;
superposing all the left audio sub-information to obtain left channel information, and outputting the left channel information to a first real loudspeaker; and superposing all the right audio sub-information to obtain right channel information, and outputting the right channel information to a second real loudspeaker.
Preferably, before the decorrelation processing, the method further includes: performing analog-to-digital conversion processing on the original audio information through an audio processing chip;
before outputting the left channel information to the first real speaker, the method further comprises: performing digital-to-analog conversion processing on the left channel information;
before outputting the right channel information to the second real speaker, the method further comprises: and D/A conversion processing is carried out on the right channel information.
Preferably, the decorrelation processing is performed by using a time delay method, and the original audio information is respectively subjected to time delay processing according to different time delay times to obtain a plurality of different first audio information, where the audio of the left and right channels in each first audio information is the same, and is represented as:
e0L(t)=e0R(t)=ea(t)
e1L(t)=e1R(t)=e0(t-τ1)
eyL(t)=eyR(t)=e0(t-τy)
in the formula, e0Representing the original audio information, τyRepresenting the delay interval corresponding to the y-th first audio information, eyL(t) denotes the audio of the left channel in the y-th first audio information, eyR(t) represents the audio of the right channel in the y-th first audio information.
Preferably, the specific implementation manner of obtaining the first HRTF parameter information by using artificial false head measurement in constructing the target diffuse sound field environment is as follows:
constructing a target diffusion sound field environment by arranging the first real loudspeaker and the second real loudspeaker, and recording sound information of each acquisition loudspeaker by using an artificial false head to obtain first HRTF parameter information, wherein the first HRTF parameter information is expressed as:
in the formula, sL(t)、sR(t) represents a time domain signal received by the artificial false head; e.g. of the typeL(t) an audio time domain signal representing the original left channel, corresponding to e0L(t)、……、eyL(t);eR(t) an audio time domain signal representing the original right channel, corresponding to e0R(t)、……、eyR(t);hL(t)、hR(t) denotes a binaural impulse response; sL(f)、SR(f)、EL(f)、ER(f)、All obtained by Fourier transform;is a virtual loudspeaker head transfer function, i.e. first HRTF parameter information.
Preferably, the frequency equalization processing on the first HRTF parameter information adopts the following formula:
in the formula, HL、HRRespectively represent first HRTF parameter information, H'L(θ,f)、H′R(θ, f) respectively represent the second HRTF parameter information, and W is an amplitude normalization constant.
Preferably, the specific implementation manner of building the HRTF filter is as follows: and calculating the IIR filter coefficient of the corresponding HRTF by using the CAPZ model, and constructing an IIR filter model in Sigma Studio by using the obtained IIR filter coefficient.
Preferably, the IIR filter coefficients are calculated by using the following formula:
for HRTF parameters of a space direction theta, a filter system function designed by a CAPZ model is adoptedComprises the following steps:
A(z)=1+a1z-1+…+aPz-P
B(z)=b0(θ)+b1(θ)z-1+…+bQ(θ)z-Q
wherein the P coefficients of the A (z) part are independent of the sound source direction and determineP common poles; q coefficients of B (z) are related to the direction of the sound source, and they determineQ zeros related to the sound source direction;representing a common transfer function independent of the direction of the sound source; b (z) represents a direction transfer function related to a direction;
in the formula (I), the compound is shown in the specification,is andthe impulse response of the corresponding filter; δ (n-q) is the unit impulse response, i.e., HRIR;
let it be assumed that the original HRIR for M spatial directions is known, denoted as h (θ)iN), i is 0, 1, …, M-1, HRIR length in each direction is N points, i.e. N is 0, 1, …, N-1; for the ith direction, the squared error of the impulse response of the filter from the original HRIR is:
the sum of the squared errors in all directions is:
in the formula, epsilonallContaining P coefficients a independent of the direction of the sound sourcepAnd M (Q +1) coefficients b related to the direction of the sound sourceq(θi) The total number of P + M (Q +1) waiting coefficients is;
define a column matrix (vector) of [ M (N + P) ]. times.1:
writing all pending coefficients as a column matrix x of [ P + M (Q +1) ]. times.1:
x=[a1,a2,…,aP,b0(θ0),b0(θ1),…,b0(θM-1),…,bQ(θ0),bQ(θ1),…,bQ(θM-1)]T
e=h1-[A]x
in the formula, h1Is [ M (N + P)]X 1 column matrix, [ A ]]Is [ M (N + P)]×[P+M(Q+1)]Matrix, from known h (θ)iN) obtaining;
the sum of squared errors can be written as:
εall=e+e=h1-[A]x+h1-[A]x
in the formula, the symbol "+" represents the transposed conjugate of the matrix;
selecting P + M (Q +1) coefficients such that ε of the above formulaallMinimum:
the solution that can be obtained for the CAPZ filter coefficients is:
x={[A]+[A]}-1[A]+h1
in the formula, x represents the calculated IIR filter coefficient.
On the other hand, the embodiment of the present application provides an apparatus for constructing an in-vehicle diffuse sound field environment based on virtual speakers, including:
the audio decorrelation processing unit is used for performing decorrelation processing on the original audio information;
the equalization processing unit is used for carrying out equalization processing on the first HRTF parameter information obtained by artificial false head measurement to obtain second HRTF parameter information;
the HRTF filter unit is used for building an HRTF filter according to the second HRTF parameter information and filtering all the first audio information;
the audio gain superposition unit is used for superposing all the left audio sub-information to obtain left channel information and superposing all the right audio sub-information to obtain right channel information;
the device for constructing the in-vehicle diffuse sound field environment based on the virtual loudspeaker is used for realizing the steps in the method for constructing the in-vehicle diffuse sound field environment based on the virtual loudspeaker.
Preferably, the apparatus for constructing an in-vehicle diffuse sound field environment based on virtual speakers further comprises:
the audio input unit is used for carrying out analog-to-digital conversion processing on the original audio information before decorrelation processing;
the audio output unit is used for performing digital-to-analog conversion processing on the left channel information before outputting the left channel information to a first real loudspeaker; and the processing module is used for performing digital-to-analog conversion processing on the right channel information before outputting the right channel information to a second real loudspeaker.
One or more technical solutions provided in the embodiments of the present application have at least the following technical effects or advantages:
in the embodiment of the application, decorrelation processing is performed on input original audio information to obtain a plurality of different first audio information; constructing a target diffuse sound field environment, obtaining first HRTF parameter information by using artificial false head measurement, and carrying out frequency equalization processing on the first HRTF parameter information to obtain second HRTF parameter information; building an HRTF filter according to the second HRTF parameter information, and filtering all the first audio information by using the HRTF filter to obtain a plurality of left audio sub-information and a plurality of right audio sub-information; superposing all the left audio sub-information to obtain left channel information, and outputting the left channel information to a first real loudspeaker; and superposing all the right audio sub-information to obtain right channel information, and outputting the right channel information to a second real loudspeaker. The invention leads in the audio and carries out decorrelation processing on the original audio to obtain a plurality of audio, frequency equalization is carried out according to the measured Head Transfer Functions (HRTF), HRTF filter design is carried out, the audio virtualizes sounds from different directions after passing through the HRTF filter, the obtained audio in a plurality of directions is added, and finally the audio is transmitted to a real loudspeaker. The device corresponding to the method comprises an audio decorrelation processing unit, an equalization processing unit, an HRTF filter unit and an audio gain superposition unit. The virtual loudspeaker is constructed to solve the problem that the false impression of sound is generated in an active sound generating system in a vehicle because a passenger accurately positions the loudspeaker generating sound from the ASD system, so that the authenticity of the designed ASD sound is reduced, and the driving experience of the passenger is influenced.
Drawings
In order to more clearly illustrate the technical solution in the present embodiment, the drawings needed to be used in the description of the embodiment will be briefly introduced below, and it is obvious that the drawings in the following description are one embodiment of the present invention, and it is obvious for those skilled in the art to obtain other drawings based on the drawings without creative efforts.
Fig. 1 is a flowchart of a method for constructing an in-vehicle diffuse sound field environment based on virtual speakers according to an embodiment of the present invention;
fig. 2 is a schematic model of a method for constructing an in-vehicle diffuse sound field environment based on virtual speakers according to an embodiment of the present invention;
fig. 3 is a structural diagram of a method for constructing an in-vehicle diffuse sound field environment based on virtual speakers according to an embodiment of the present invention;
fig. 4 is a schematic diagram of decorrelation processing in a method for constructing an in-vehicle diffuse sound field environment based on virtual speakers according to an embodiment of the present invention;
fig. 5 is a structural diagram model of an IIR filter in the method for constructing an in-vehicle diffuse sound field environment based on a virtual speaker according to the embodiment of the present invention.
Detailed Description
In order to better understand the technical solution, the technical solution will be described in detail with reference to the drawings and the specific embodiments.
The embodiment provides a method for constructing an in-vehicle diffuse sound field environment based on virtual speakers, as shown in fig. 1, fig. 2, and fig. 3, the method mainly includes the following steps:
The original audio e0And inputting the data into a DSP chip, and performing analog-to-digital conversion in the DSP chip.
In particular, the original audio e0And the audio signal is transmitted to an audio processing chip for the chip to read. The input audio may be single-channel audio or multi-channel audio.
One specific audio input method is: an audio input model is built in DSP processing software Sigma Studio, an audio connection line is used for connecting the DSP development board with an audio player, audio is input to a DSP chip, and data of each channel is separated so as to facilitate subsequent data processing.
And 2, performing decorrelation processing on the original audio information to obtain a plurality of different first audio information.
Specifically, the decorrelation processing is performed by using a time delay method, and the original audio information is respectively subjected to time delay processing according to different time delay times to obtain a plurality of different first audio information, where the audio of the left and right channels in each first audio information is the same, and is represented as:
eOL(t)=eOR(t)=e0(t)
e1L(t)=e1R(t)=e0(t-τ1)
eyL(t)=eyR(t)=e0(t-τy)
in the formula, e0Representing the original audio information, τyRepresenting the delay interval corresponding to the y-th first audio information, eyL(t) denotes the audio of the left channel in the y-th first audio information, eyR(t) denotes the right in the y-th first audio informationAudio of the soundtrack.
I.e. to the original audio e0Performing decorrelation processing to generate a plurality of audio e0、e1、……、ey。
Take the final generation of 3 audios as an example, for the original audio e0Performing decorrelation processing to generate a plurality of audio e0、e1、e2。
The audio decorrelation process includes: the delay time is set, and each path of audio is delayed to obtain a new audio, as shown in fig. 4.
The method has the main technical key points that the method comprises the steps of calculating delay time, building an audio delay processing model in DSP processing software Sigma Studio, carrying out delay processing on the audio to obtain a plurality of new audios, eliminating the relevance of signals to a certain extent and enhancing the surrounding sense of subjective hearing. Namely, the audio is delayed in different degrees, so that the correlation between the generated new audio is reduced, and the sound surrounding sense in subjective auditory sense is generated.
The time delay calculation combines the priority effect of sound, and the relative time delay tau between the multiple sound source signals exceeds a certain lower limit tauLBut not exceeding a certain upper limit τHWhen this occurs, a different spatial auditory effect of the synthesized sound image localization occurs, in which the audio of the left and right channels is the same.
e0L(t)=e0R(t)=e0(t)
e1L(t)=e1R(t)=e0(t-τ1)
e2L(t)=e2R(t)=e0(t-τ2)
And 3, constructing a target diffuse sound field environment, obtaining first HRTF parameter information by using artificial false head measurement, and performing frequency equalization processing on the first HRTF parameter information to obtain second HRTF parameter information.
The artificial false head is used for measuring the HRTF parameter information of a plurality of loudspeakers capable of recording a target diffusion sound field, and the HRTF parameters are subjected to equalization processing, so that the problems of frequency distortion, change of reproduced subjective timbre and the like are avoided.
Specifically, a target diffuse sound field environment is constructed by arranging a first real loudspeaker and a second real loudspeaker (for example, 2 real loudspeakers in front in fig. 2), HRTF parameter information of each loudspeaker (namely, the loudspeaker placed when a head-related transfer function is collected in a half-anechoic chamber) is recorded by using an artificial dummy head, and frequency equalization processing is performed on the HRTF parameters, so that the problems of reproduced sound image distortion and frequency distortion are avoided.
The HRTF parameter equalization processing includes: measurement of virtual speaker head transfer function (HRTF) parameters, HRTF parameter equalization.
Specifically, step 3 mainly comprises the following steps:
(1) an ideal diffusion sound field environment is constructed by arranging real loudspeakers, and sound information of each loudspeaker is recorded by utilizing an artificial false head to obtain HRTF parameters (namely, the HRTF parameters are obtained )。
In the formula, sL(t)、sR(t) time-domain signals received by the artificial dummy head, eL(t)、eR(t) is the original audioTime domain signal, hL(t)、hR(t) is the binaural impulse response, SL(f)、SR(f)、EL(f)、ER(f)、All the data are obtained by Fourier transform,is a head transfer function, i.e. first HRTF parameter information. In particular, eL(t) an audio time domain signal representing the original left channel, corresponding to e0L(t)、e1L(t)、e2L(t);eR(t) an audio time domain signal representing the original right channel, corresponding to e0R(t)、e1R(t)、e2R(t)。
(2) And (3) equalizing the first HRTF parameter information by using the sum of the HRTF functions of the left ear and the right ear:
in the formula, HL、HRRespectively represent first HRTF parameter information, H'L(θ,f)、H′R(θ, f) respectively represent the second HRTF parameter information, and W is an amplitude normalization constant. I.e. using H'L(θ,f)、H′R(θ, f) instead of the original HRTF parameters.
And 4, building an HRTF filter according to the second HRTF parameter information, and filtering all the first audio information by using the HRTF filter to obtain a plurality of left audio sub-information and a plurality of right audio sub-information.
Specifically, a Common-acoustic-polar and Zero (CAPZ) model is used, i.e., IIR filter coefficients of corresponding HRTFs are calculated by using a Common Pole model and a Pole model related to a direction, an IIR filter model is built inside a DSP by using Sigma Studio, and after audio passes through a filter, sound with spatial information is generated.
Referring to fig. 5, the specific process of calculating IIR filter coefficients mainly includes:
for HRTF parameters in the spatial direction theta, a filter system function designed by using a CAPZ model is as follows:
A(z)=1+a1z-1+…+aPz-P
B(z)=b0(θ)+b1(θ)z-1+…+bQ(θ)z-Q
wherein the P coefficients of the A (z) part are independent of the sound source direction and determineP common poles; q coefficients of B (z) are related to the direction of the sound source, and they determineQ zeros related to the sound source direction;representing a common transfer function independent of the direction of the sound source; b (z) represents a directional transfer function with respect to direction.
Wherein the content of the first and second substances,is andthe impulse response of the corresponding filter, δ (n-q), is the unit impulse response (HRIR).
Let us assume that the original HRIR for M spatial directions (M corresponding to the number of virtual loudspeakers) is known, denoted h (θ)iN), i is 0, 1, …, M-1, and the HRIR length in each direction is N points, i.e., N is 0, 1, …, N-1. For the ith direction, the squared error of the impulse response of the filter from the original HRIR is:
the sum of the squared errors in all directions is:
wherein epsilonallContaining P coefficients a independent of the direction of the sound sourcepAnd M (Q +1) coefficients b related to the direction of the sound sourceq(θi) There are P + M (Q +1) determinants.
Define a column matrix (vector) of [ M (N + P) ]. times.1:
writing all undetermined coefficients into a column matrix x of [ P + M (Q +1) ] × 1, i.e.:
x=[a1,a2,…,aP,b0(θ0),b0(θ1),…,b0(θM-1),…,bQ(θ0),bQ(θ1),…,bQ(θM-1)]T
e=h1-[A]x
wherein h is1Is [ M (N + P)]X 1 column matrix (vector), [ A ]]Is [ M (N + P)]×[P+M(Q+1)]Matrix, from known h (θ)iAnd n) obtaining.
And the sum of squared errors can be written as:
εall=e+e=h1[A]x+h1-[A]x
the symbol "+" represents the transposed conjugate of the matrix.
Selecting P + M (Q +1) coefficients such that ε of the above formulaallThe minimum, namely:
the solution that can be obtained for the CAPZ filter coefficients is:
x={[A]+[A]}-1[A]+h1
in the formula, x represents the calculated IIR filter coefficient.
And (4) constructing an IIR filter model in Sigma Studio software of the DSP by using the obtained filter coefficient.
Step 5, superposing all the left audio sub-information to obtain left channel information, and outputting the left channel information to a first real loudspeaker after performing digital-to-analog conversion processing on the left channel information; and superposing all the right audio sub-information to obtain right channel information, and outputting the right channel information to a second real loudspeaker after performing digital-to-analog conversion processing on the right channel information.
Namely, the filtered audio is superposed and then is transmitted to a corresponding loudspeaker through digital-to-analog conversion.
Specifically, the filtered left and right channel audio frequencies are respectively added and combined into a left channel audio frequency and a right channel audio frequency, and are sent to the corresponding real speakers after digital-to-analog conversion.
As shown in fig. 2, the invention constructs 5 virtual speakers, and produces sound through 2 actual speakers, thereby achieving the purpose of realizing the sound field diffusion effect of 5 speakers. The invention realizes the effect of playing a plurality of virtual loudspeakers through 2 actual loudspeakers, generates the sense of encirclement on subjective feeling, confuses the perception of passengers to the loudspeakers in the vehicle, and thus achieves the purpose of improving the ASD sound reality sense.
Furthermore, it is also possible to simulate 5 virtual speakers by the front 2 actual speakers and the rear 5 virtual speakers by the rear 2 actual speakers.
It should be noted that the basic solution of the present invention is to use only the front 2 actual speakers to construct the diffuse sound field, because most of the vehicles of ASD sound in the market today sound through the front 2 speakers in the vehicle. In specific application, whether 5 rear virtual speakers need to be constructed or not can be considered according to the actually constructed in-vehicle effect. If a rear virtual loudspeaker needs to be constructed, the loudspeaker at the rear in the vehicle is designed to generate a corresponding diffused sound field by the same method (namely, the steps 3 to 5 are repeated).
Corresponding to the method, the embodiment further provides an apparatus for constructing an in-vehicle diffuse sound field environment based on the virtual speaker, including:
the audio input unit is used for carrying out analog-to-digital conversion processing on the original audio information before decorrelation processing; i.e. audio input to the DSP chip.
The audio decorrelation processing unit is used for performing decorrelation processing on the original audio information; namely, a plurality of audios with low correlation are generated in the DSP chip, and a sound effect with good subjective auditory sense surrounding sense is generated.
The equalization processing unit is used for carrying out equalization processing on the first HRTF parameter information obtained by artificial false head measurement to obtain second HRTF parameter information; namely, the measured HRTF parameters are equalized, so that the problems of frequency distortion and sound color change are avoided.
The HRTF filter unit is used for building an HRTF filter according to the second HRTF parameter information and filtering all the first audio information; the filter coefficient is calculated by referring to the equalized HRTF parameters by using a CAPZ model, and an IIR filter is designed in a DSP chip, so that the problem of overlarge direct convolution calculation is solved.
The audio gain superposition unit is used for superposing all the left audio sub-information to obtain left channel information and superposing all the right audio sub-information to obtain right channel information; i.e. the filtered sounds are summed in the DSP chip.
The audio output unit is used for performing digital-to-analog conversion processing on the left channel information before outputting the left channel information to a first real loudspeaker; the digital-to-analog conversion processing is carried out on the right channel information before the right channel information is output to a second real loudspeaker; i.e. finally to the real speaker (actual speaker).
The device for constructing the in-vehicle diffuse sound field environment based on the virtual loudspeaker is used for realizing the steps in the method for constructing the in-vehicle diffuse sound field environment based on the virtual loudspeaker.
In addition, the actual speaker is a speaker directly used in the car, and the actual speaker is directly connected to the corresponding speaker through an audio line by building a model in an ADAU1467 development board.
In order to verify the sound field effect, the invention also provides a verification method for the effectiveness and the correctness of the in-vehicle diffused sound field environment constructed based on the virtual loudspeaker. Firstly, a sound pressure curve of the constructed virtual loudspeaker at the position of the human ear is compared with a sound pressure curve of the human ear measured by actually using the corresponding real loudspeaker, the construction effect of the virtual loudspeaker is checked according to the comparison result, then the sound pressure of the human ear measured by actually using the loudspeaker is calculated, and a binaural auditory cross-correlation coefficient (IACC) is used as an index for measuring the sound spatial sensation. IACC is defined as binaural time-domain sound pressure pL(t)、pR(t) maximum of normalized cross-correlation function:
wherein p isL(t) denotes the measured sound pressure at the left ear of the human, pR(t) represents the measured sound pressure at the right ear of the person,representing the cross-correlation function of the left and right human ears,the left-ear autocorrelation function is represented as,the left ear autocorrelation function is represented, IACC is more than or equal to 0 and less than or equal to 1, when the IACC is closer to 1, the auditory sense can generate clear sound images, the localization of the sound images is more clear, and the lower the IACC is, the more fuzzy the sound images are, so that the sound images cannot be localized.
The method and the device for constructing the in-vehicle diffused sound field environment based on the virtual loudspeaker provided by the embodiment of the invention at least have the following technical effects:
(1) the virtual loudspeaker is constructed to solve the problem that a passenger accurately positions the loudspeaker which emits sound to the active sound production (ASD) system in the vehicle to generate a false sound impression, so that the authenticity of the designed ASD sound is reduced, and the driving experience of the passenger is influenced.
(2) The invention carries out decorrelation processing on the audios according to the priority effect (Hass effect) in the spatial auditory effect, weakens the correlation among the audios, ensures that the existence of one sound source has a masking effect on the spatial position information of another sound source, is limited to the spatial position information instead of the spatial auditory information, and not only meets the aims of fuzzy sound source positioning but also realizes the sense of enclosure.
(3) The invention balances the frequency (tone) of the measured HRTF parameters, further balances the balance among energies of different frequencies, and avoids the problems of sound image distortion of reproduced sound and tone color change caused by frequency distortion.
(4) The virtual loudspeaker is constructed by utilizing the head transfer function from the sound source to the human ear, the actually measured binaural impulse response can be convoluted with the audio to achieve the addition of the space effect, but one audio needs to be convoluted twice, five virtual loudspeakers need to be convoluted ten times, and the calculation amount is overlarge, so that the IIR filter coefficient of the HRTF is calculated by adopting a CAPZ model.
(5) The method for constructing the in-vehicle diffusion sound field environment by the virtual loudspeaker solves the problem that the in-vehicle loudspeaker occupies a large space, reduces the cost for arranging the loudspeaker, and meanwhile confuses the judgment of passengers on the ASD sound source position information by the diffusion sound field, so that the reality of the ASD is improved.
Finally, it should be noted that the above embodiments are only for illustrating the technical solutions of the present invention and not for limiting, and although the present invention has been described in detail with reference to examples, it should be understood by those skilled in the art that modifications or equivalent substitutions may be made on the technical solutions of the present invention without departing from the spirit and scope of the technical solutions of the present invention, which should be covered by the claims of the present invention.
Claims (9)
1. A method for constructing an in-vehicle diffuse sound field environment based on virtual speakers is characterized by comprising the following steps:
performing decorrelation processing on input original audio information to obtain a plurality of different first audio information;
constructing a target diffusion sound field environment, obtaining first HRTF parameter information by using artificial false head measurement, and performing frequency equalization processing on the first HRTF parameter information to obtain second HRTF parameter information;
building an HRTF filter according to the second HRTF parameter information, and filtering all the first audio information by using the HRTF filter to obtain a plurality of left audio sub-information and a plurality of right audio sub-information;
superposing all the left audio sub-information to obtain left channel information, and outputting the left channel information to a first real loudspeaker; superposing all the right audio sub-information to obtain right channel information, and outputting the right channel information to a second real loudspeaker;
the sound emitted by the first real speaker and the second real speaker is used for an active sound design ASD in the vehicle.
2. The method for constructing an in-vehicle diffuse sound field environment based on virtual speakers according to claim 1, wherein before the decorrelation process, further comprising: performing analog-to-digital conversion processing on the original audio information through an audio processing chip;
before outputting the left channel information to the first real speaker, the method further comprises: performing digital-to-analog conversion processing on the left channel information;
before outputting the right channel information to the second real speaker, the method further comprises: and D/A conversion processing is carried out on the right channel information.
3. The method for constructing the in-vehicle diffuse sound field environment based on the virtual speaker according to claim 1, wherein the decorrelation processing is performed by a time-delay method, and the original audio information is respectively subjected to time-delay processing according to different time-delay times to obtain a plurality of different first audio information, wherein the audio of the left channel and the audio of the right channel in each first audio information are the same and are represented as follows:
e0L(t)=e0R(t)=e0(t)
e1L(t)=e1R(t)=e0(t-τ1)
eyL(t)=eyR(t)=e0(t-τy)
in the formula, e0Representing the original audio information, τyRepresenting the delay interval corresponding to the y-th first audio information, eyL(t) denotes the audio of the left channel in the y-th first audio information, eyR(t) represents the audio of the right channel in the y-th first audio information.
4. The method for constructing the in-vehicle diffuse sound field environment based on the virtual speaker as claimed in claim 3, wherein the specific implementation manner of obtaining the first HRTF parameter information by using the artificial dummy head measurement to construct the target diffuse sound field environment is as follows:
constructing a target diffusion sound field environment by arranging the first real loudspeaker and the second real loudspeaker, and recording sound information of each acquisition loudspeaker by using an artificial false head to obtain first HRTF parameter information, wherein the first HRTF parameter information is expressed as:
in the formula, sL(t)、sR(t) represents a time domain signal received by the artificial false head; e.g. of the typeL(t) an audio time domain signal representing the original left channel, corresponding to e0L(t)、……、eyL(t);eR(t) an audio time domain signal representing the original right channel, corresponding to e0R(t)、……、eyR(t);hL(t)、hR(t) denotes a binaural impulse response; sL(f)、SR(f)、EL(f)、ER(f)、All obtained by Fourier transform;is a virtual loudspeaker head transfer function, i.e. first HRTF parameter information.
5. The method for constructing an in-vehicle diffuse sound field environment based on virtual speakers according to claim 4, wherein the frequency equalization processing on the first HRTF parameter information adopts the following formula:
in the formula, HL、HRRespectively represent first HRTF parameter information, H'L(θ,f)、H′R(θ, f) respectively represent the second HRTF parameter information, and W is an amplitude normalization constant.
6. The method for constructing the in-vehicle diffuse sound field environment based on the virtual speaker as claimed in claim 1, wherein the specific implementation manner of constructing the HRTF filter is as follows: and calculating the IIR filter coefficient of the corresponding HRTF by using the CAPZ model, and constructing an IIR filter model in Sigma Studio by using the obtained IIR filter coefficient.
7. The method for constructing the in-vehicle diffuse sound field environment based on the virtual speaker as claimed in claim 6, wherein the IIR filter coefficient is calculated by using the following formula:
for HRTF parameters of a space direction theta, a filter system function designed by a CAPZ model is adoptedComprises the following steps:
A(z)=1+a1z-1+…+aPz-P
B(z)=b0(θ)+b1(θ)z-1+…+bQ(θ)z-Q
wherein the P coefficients of the A (z) part are independent of the sound source direction and determineP common poles; q coefficients of B (z) are related to the direction of the sound source, and they determineQ zeros related to the sound source direction;representing a common transfer function independent of the direction of the sound source; b (z) represents a direction transfer function related to a direction;
in the formula (I), the compound is shown in the specification,is andthe impulse response of the corresponding filter; δ (n-q) is the unit impulse response, i.e., HRIR;
let it be assumed that the original HRIR for M spatial directions is known, denoted as h (θ)iN), i is 0, 1, …, M-1, HRIR length in each direction is N points, i.e. N is 0, 1, …, N-1; for the ith direction, the squared error of the impulse response of the filter from the original HRIR is:
the sum of the squared errors in all directions is:
in the formula, epsilonallContaining P coefficients a independent of the direction of the sound sourcepAnd M (Q +1) coefficients b related to the direction of the sound sourceq(θi) The total number of P + M (Q +1) waiting coefficients is;
define a column matrix (vector) of [ M (N + P) ]. times.1:
writing all pending coefficients as a column matrix x of [ P + M (Q +1) ]. times.1:
x=[a1,a2,…,aP,b0(θ0),b0(θ1),…,b0(θM-1),…,bQ(θ0),bQ(θ1),…,bQ(θM-1)]T
e=h1-[A]x
in the formula, h1Is [ M (N + P)]X 1 column matrix, [ A ]]Is [ M (N + P)]×[P+M(Q+1)]Matrix, from known h (θ)iN) obtaining;
the sum of squared errors can be written as:
εall=e+e=h1-[A]x+h1-[A]x
in the formula, the symbol "+" represents the transposed conjugate of the matrix;
selecting P + M (Q +1) coefficients such that ε of the above formulaallMinimum:
the solution that can be obtained for the CAPZ filter coefficients is:
x=([A]+[A]}-1[A]+h1
in the formula, x represents the calculated IIR filter coefficient.
8. An apparatus for constructing an in-vehicle diffuse sound field environment based on virtual speakers, comprising:
the audio decorrelation processing unit is used for performing decorrelation processing on the original audio information;
the equalization processing unit is used for carrying out equalization processing on the first HRTF parameter information obtained by artificial false head measurement to obtain second HRTF parameter information;
the HRTF filter unit is used for building an HRTF filter according to the second HRTF parameter information and filtering all the first audio information;
the audio gain superposition unit is used for superposing all the left audio sub-information to obtain left channel information and superposing all the right audio sub-information to obtain right channel information;
the device for constructing the in-vehicle diffusion sound field environment based on the virtual loudspeaker is used for realizing the steps in the method for constructing the in-vehicle diffusion sound field environment based on the virtual loudspeaker in any one of claims 1 to 7.
9. The apparatus for constructing an in-vehicle diffuse sound field environment based on virtual speakers according to claim 8, further comprising:
the audio input unit is used for carrying out analog-to-digital conversion processing on the original audio information before decorrelation processing;
the audio output unit is used for performing digital-to-analog conversion processing on the left channel information before outputting the left channel information to a first real loudspeaker; and the processing module is used for performing digital-to-analog conversion processing on the right channel information before outputting the right channel information to a second real loudspeaker.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202010805149.2A CN112019994B (en) | 2020-08-12 | 2020-08-12 | Method and device for constructing in-vehicle diffusion sound field environment based on virtual loudspeaker |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202010805149.2A CN112019994B (en) | 2020-08-12 | 2020-08-12 | Method and device for constructing in-vehicle diffusion sound field environment based on virtual loudspeaker |
Publications (2)
Publication Number | Publication Date |
---|---|
CN112019994A CN112019994A (en) | 2020-12-01 |
CN112019994B true CN112019994B (en) | 2022-02-08 |
Family
ID=73504229
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN202010805149.2A Active CN112019994B (en) | 2020-08-12 | 2020-08-12 | Method and device for constructing in-vehicle diffusion sound field environment based on virtual loudspeaker |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN112019994B (en) |
Families Citing this family (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN113971013A (en) * | 2021-10-25 | 2022-01-25 | 北京字节跳动网络技术有限公司 | Sound effect playing method and equipment of music |
CN114630240B (en) * | 2022-03-16 | 2024-01-16 | 北京小米移动软件有限公司 | Direction filter generation method, audio processing method, device and storage medium |
CN116367076A (en) * | 2023-03-30 | 2023-06-30 | 潍坊歌尔丹拿电子科技有限公司 | In-vehicle audio processing method, in-vehicle audio processing device and storage medium |
CN116744216B (en) * | 2023-08-16 | 2023-11-03 | 苏州灵境影音技术有限公司 | Automobile space virtual surround sound audio system based on binaural effect and design method |
CN117676418B (en) * | 2023-12-06 | 2024-05-24 | 广州番禺职业技术学院 | Sound field equalization method and system for mixed phase system |
CN117476026A (en) * | 2023-12-26 | 2024-01-30 | 芯瞳半导体技术(山东)有限公司 | Method, system, device and storage medium for mixing multipath audio data |
Citations (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN102395098A (en) * | 2005-09-13 | 2012-03-28 | 皇家飞利浦电子股份有限公司 | Method of and device for generating 3d sound |
Family Cites Families (14)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP5341919B2 (en) * | 2008-02-14 | 2013-11-13 | ドルビー ラボラトリーズ ライセンシング コーポレイション | Stereo sound widening |
CA2732079C (en) * | 2008-07-31 | 2016-09-27 | Fraunhofer-Gesellschaft Zur Foerderung Der Angewandten Forschung E.V. | Signal generation for binaural signals |
WO2012016722A2 (en) * | 2010-08-04 | 2012-02-09 | Fraunhofer-Gesellschaft zur Förderung der angewandten Forschung e.V. | Apparatus for generating a drive dependent sound and engine driven vehicle |
KR101805110B1 (en) * | 2013-12-13 | 2017-12-05 | 앰비디오 인코포레이티드 | Apparatus and method for sound stage enhancement |
DE102014214052A1 (en) * | 2014-07-18 | 2016-01-21 | Bayerische Motoren Werke Aktiengesellschaft | Virtual masking methods |
BR112017017332B1 (en) * | 2015-02-18 | 2022-11-16 | Huawei Technologies Co., Ltd | AUDIO SIGNAL PROCESSING APPARATUS AND METHOD FOR FILTERING AN AUDIO SIGNAL |
CN109155895B (en) * | 2016-04-20 | 2021-03-16 | 珍尼雷克公司 | Active listening headset and method for regularizing inversion thereof |
GB201609089D0 (en) * | 2016-05-24 | 2016-07-06 | Smyth Stephen M F | Improving the sound quality of virtualisation |
CN206171449U (en) * | 2016-10-14 | 2017-05-17 | 武汉理工大学 | Vehicle velocity early warning system based on vehicle status road surface condition |
WO2018086701A1 (en) * | 2016-11-11 | 2018-05-17 | Huawei Technologies Co., Ltd. | Apparatus and method for weighting stereo audio signals |
US20180190306A1 (en) * | 2017-01-04 | 2018-07-05 | 2236008 Ontario Inc. | Voice interface and vocal entertainment system |
US10979844B2 (en) * | 2017-03-08 | 2021-04-13 | Dts, Inc. | Distributed audio virtualization systems |
KR20200075144A (en) * | 2018-12-13 | 2020-06-26 | 현대자동차주식회사 | A control system for making car sound index based engine sound in use with deep-learing and the method of it |
JP7470695B2 (en) * | 2019-01-08 | 2024-04-18 | テレフオンアクチーボラゲット エルエム エリクソン(パブル) | Efficient spatially heterogeneous audio elements for virtual reality |
-
2020
- 2020-08-12 CN CN202010805149.2A patent/CN112019994B/en active Active
Patent Citations (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN102395098A (en) * | 2005-09-13 | 2012-03-28 | 皇家飞利浦电子股份有限公司 | Method of and device for generating 3d sound |
Non-Patent Citations (1)
Title |
---|
环绕声重放中通路信号相关性与听觉空间印象;石蓓等;《声学学报(中文版)》;20090715(第04期);全文 * |
Also Published As
Publication number | Publication date |
---|---|
CN112019994A (en) | 2020-12-01 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN112019994B (en) | Method and device for constructing in-vehicle diffusion sound field environment based on virtual loudspeaker | |
JP4364326B2 (en) | 3D sound reproducing apparatus and method for a plurality of listeners | |
JP3913775B2 (en) | Recording and playback system | |
JP4588945B2 (en) | Method and signal processing apparatus for converting left and right channel input signals in two-channel stereo format into left and right channel output signals | |
JP4584416B2 (en) | Multi-channel audio playback apparatus for speaker playback using virtual sound image capable of position adjustment and method thereof | |
JP4343845B2 (en) | Audio data processing method and sound collector for realizing the method | |
JP4780119B2 (en) | Head-related transfer function measurement method, head-related transfer function convolution method, and head-related transfer function convolution device | |
WO2009046223A2 (en) | Spatial audio analysis and synthesis for binaural reproduction and format conversion | |
KR100647338B1 (en) | Method of and apparatus for enlarging listening sweet spot | |
WO2004103023A1 (en) | Method for preparing transfer function table for localizing virtual sound image, recording medium on which the table is recorded, and acoustic signal editing method using the medium | |
JP2009508158A (en) | Method and apparatus for generating and processing parameters representing head related transfer functions | |
EP2243136B1 (en) | Mediaplayer with 3D audio rendering based on individualised HRTF measured in real time using earpiece microphones. | |
CN113170271A (en) | Method and apparatus for processing stereo signals | |
Jot et al. | Binaural simulation of complex acoustic scenes for interactive audio | |
JP2008502200A (en) | Wide stereo playback method and apparatus | |
CN115226022A (en) | Content-based spatial remixing | |
US10321252B2 (en) | Transaural synthesis method for sound spatialization | |
CN107743713A (en) | Handle for the stereophonic signal reproduced in the car to realize the apparatus and method of single three dimensional sound by front loudspeakers | |
CN103546838A (en) | Method for establishing an optimized loudspeaker sound field | |
Ifergan et al. | On the selection of the number of beamformers in beamforming-based binaural reproduction | |
Krebber et al. | Auditory virtual environments: basics and applications for interactive simulations | |
KR100849030B1 (en) | 3D sound Reproduction Apparatus using Virtual Speaker Technique under Plural Channel Speaker Environments | |
JPH09191500A (en) | Method for generating transfer function localizing virtual sound image, recording medium recording transfer function table and acoustic signal edit method using it | |
JP2000333297A (en) | Stereophonic sound generator, method for generating stereophonic sound, and medium storing stereophonic sound | |
CN109923877A (en) | The device and method that stereo audio signal is weighted |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
GR01 | Patent grant | ||
GR01 | Patent grant |