CN107105384B - The synthetic method of near field virtual sound image on a kind of middle vertical plane - Google Patents
The synthetic method of near field virtual sound image on a kind of middle vertical plane Download PDFInfo
- Publication number
- CN107105384B CN107105384B CN201710347867.8A CN201710347867A CN107105384B CN 107105384 B CN107105384 B CN 107105384B CN 201710347867 A CN201710347867 A CN 201710347867A CN 107105384 B CN107105384 B CN 107105384B
- Authority
- CN
- China
- Prior art keywords
- hrtf
- field
- vertical plane
- middle vertical
- far
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Expired - Fee Related
Links
Classifications
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04S—STEREOPHONIC SYSTEMS
- H04S7/00—Indicating arrangements; Control arrangements, e.g. balance control
- H04S7/30—Control circuits for electronic adaptation of the sound field
- H04S7/302—Electronic adaptation of stereophonic sound system to listener position or orientation
- H04S7/303—Tracking of listener position or orientation
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04S—STEREOPHONIC SYSTEMS
- H04S7/00—Indicating arrangements; Control arrangements, e.g. balance control
- H04S7/30—Control circuits for electronic adaptation of the sound field
- H04S7/302—Electronic adaptation of stereophonic sound system to listener position or orientation
- H04S7/303—Tracking of listener position or orientation
- H04S7/304—For headphones
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04S—STEREOPHONIC SYSTEMS
- H04S2420/00—Techniques used stereophonic systems covered by H04S but not provided for in its groups
- H04S2420/01—Enhancing the perception of the sound image or of the spatial distribution using head related transfer functions [HRTF's] or equivalents thereof, e.g. interaural time difference [ITD] or interaural level difference [ILD]
Landscapes
- Physics & Mathematics (AREA)
- Engineering & Computer Science (AREA)
- Acoustics & Sound (AREA)
- Signal Processing (AREA)
- Stereophonic System (AREA)
Abstract
The present invention relates to a kind of synthetic methods of near field virtual sound image on middle vertical plane, include the following steps:Known far field head-position difficult labor HRTFfar, the wherein distance D of sound source and head center;Determine the distance d of near-field target acoustic image and head center on number of people radius a and middle vertical plane;According to formula, the angle correction θ of auris dextra R is calculated;Extract far field HRTFfarThe auris dextra HRTF of (θ, R) as near-field target acoustic image;According to the bilateral symmetry of middle vertical plane, far field HRTF is extractedfarThe left ear HRTF of (- θ, L) as near-field target acoustic image;By the ears HRTF of near-field target acoustic image and input acoustical signal progress convolution, it can be achieved that the 3D virtual playbacks of the middle vertical plane near field acoustic image based on earphone.Method provided by the invention deduces near-field HRTF using far field HRTF, can easily realize the synthesis of middle vertical plane near field virtual sound image, saves the cumbersome measurement process of near-field HRTF.
Description
Technical field
The present invention relates to 3D virtual sound technologies fields, and in particular to a kind of near field middle vertical plane Virtual Sound based on far-field information
Synthetic method.
Background technology
Virtual Sound resets the virtual sound image of the method simulation specific dimensional orientation using Underwater Acoustic channels, to make hearer produce
Raw feeling of immersion and telepresenc.Since reproducing device (be typically earphone) is simple, portable, Virtual Sound is reset in multimedia computer, a
The fields such as people's audio mobile terminal have a wide range of applications.Head-position difficult labor (head-related transfer
Function, HRTF) it is the key that virtual sound signal processing.HRTF be in the case of free field sound source to the acoustic transmission letter of ears
Number is related to sound bearing, distance, frequency of sound wave, hearer's physiological structure etc..When the distance of sound source to listener head is more than 1.0
Rice, HRTF change, referred to as far field HRTF little with distance;And the distance for working as sound source to listener head is less than 1.0 meters, HRTF
Significant change, referred to as near-field HRTF occurs with distance.Near-field HRTF contains the information of sound source Distance positioning, therefore it is in ears
It is of great significance in the research of Distance positioning mechanism;Meanwhile with universal (such as the mobile phone) of hand-held playback equipment, it is also desirable to
The signal processing of near field Virtual Sound playback is carried out using near-field HRTF.
Currently, near-field HRTF data are mainly obtained by measuring.In actual near-field HRTF measurement, it is primarily present following
The difficulty of three aspects:(1) closer due to measuring sound source and subject, multiple scattering and diffraction will inevitably occur for sound wave
Problem, it is therefore desirable to specific small size sound source;(2) due to near-field HRTF and distance dependent, it is therefore desirable to measure respectively different
The HRTF of distance, workload are very big.(3) related (such as head size, auricle are subtle for the physiological structure of HRTF and subject
Structure etc.).In principle, in order to obtain best sound replaying effect, the near field for measuring each hearer (i.e. subject) is needed
HRTF.Currently, having multiple seminars establishes far field HRTF databases, and disclose on the internet, such as MIT and CIPIC
HRTF databases.If the near-field HRTF that known far field HRTF derives different distance may be used, that can be to avoid upper
The difficulty for stating near-field HRTF measurement realizes the convenient acquisition of near-field HRTF.
The space orientation of acoustic image is broadly divided into horizontal direction positioning and vertical direction positioning.Level orientation positioning is mainly depended on
In the difference of left and right ear HRTF, and vertical direction positioning (Sound image localization especially on middle vertical plane) relies primarily on HRTF's
Spectrum signature.Therefore, the accuracy of the spectrum signature of HRTF determines the accuracy of middle vertical plane Sound image localization;In middle vertical plane far field
During HRTF derives near-field HRTF, the accurate reproduction for paying special attention to near-field HRTF spectrum signature is needed.
In general, the correlative study (including existing database) of HRTF uses head centre coordinate system, the wherein angle of HRTF
It is defined as angle of the sound source relative to head center.However, HRTF itself is defined as sound source to the transfer function of ears, HRTF data
And picked up at ears.Therefore, during far field HRTF derives near-field HRTF, in order to ensure near-field HRTF frequency
The accuracy of spectrum signature needs to consider angle of the sound source relative to ear (rather than head center).Calculation shows that far field sound source is opposite
It is roughly equal in head center and angle relative to ear;And angle of the near-field sound source relative to head center and relative to ear exist compared with
Large deviation, it might even be possible to reach 30 °.This means that in common head centre coordinate system, far field HRTF cannot be directly used
The near-field HRTF with angle is derived, needs to consider near-field sound source at the angle of two kinds of ears coordinate system and head centre coordinate system
Spend deviation.
Invention content
The angular deviation problem during near-field HRTF is derived for far field HRTF, the present invention provides one kind being based on angle
The acquisition methods of modified middle vertical plane near-field HRTF.The present invention using remote, close sound source to ear angle having the same as starting point, push away
The angle (i.e. angle correction) of middle vertical plane near-field HRTF far field HRTF corresponding in head centre coordinate system is exported, and then from
Some in the far field HRTF databases of head centre coordinate system using corresponding ears HRTF data are extracted, for subsequently synthesizing middle vertical plane
Near field virtual sound image.This method can accurately, easily synthesize the virtual sound image at the arbitrary sound source distance of middle vertical plane.
The purpose of the present invention is achieved through the following technical solutions.
The synthetic method of near field virtual sound image on a kind of middle vertical plane, including step:
Step 1, input given data, including far field HRTF baseline databases, three distance parameters, described three distances
Parameter include the distance D of far field sound source and head center, number of people radius a, on middle vertical plane near-field target acoustic image and head center distance
d;
Step 2 calculates the angle correction θ of auris dextra R according to formula:
Wherein, change the distance d of near-field target acoustic image and head center on middle vertical plane, so that it may to obtain the different distances of near field
Angle correction;
Step 3, extraction far field HRTFfarThe auris dextra HRTF of (θ, R) as near-field target acoustic image, i.e. HRTFnear(R);Extraction
Far field HRTFfarThe left ear HRTF, i.e. HRTF of (- θ, L) as near-field target acoustic imagenear(L);
Step 4, the ears HRTF by near-field target acoustic imagenearConvolution is carried out with input acoustical signal, and feeds and earphone or raises
Sound device carries out low voice speaking put.
Further, the far field is the region other than 1 meter of head center, the near field be apart from 1 meter of head center with
Interior region.
Further, the feature of the HRTF in the far field does not change with distance, and the feature of the HRTF near field is with distance
And change.
Further, the multiple far field HRTF databases being disclosed before the feelings the pulse with the finger-tip of HRTF baseline databases described in step 1,
Including Massachusetts Institute Technology's MIT HRTF databases, California, USA university Davis branch school CIPIC HRTF data
Library.
Further, the number of people radius a described in step 1 is averaged as 8.75cm.
Further, the number of people radius a described in step 1 is by directly measuring acquisition.
Further, the binaural signals synthesized in step 4 are directly fed to Headphone reproducing;When using speaker playback,
It then needs to be inserted into string sound elimination algorithm.
The principle of the present invention is:The sound wave that any sound source in space is sent out, by with hearer's physiological structure (such as head, auricle,
Trunk etc.) interaction after, reach ears.HRTF reflects the change of sound wave during the above-mentioned sound transmission.No matter far field or
Near field, as long as sound source is identical relative to the azimuth of ear, corresponding sound transmission process be it is identical, therefore same angle (relative to
For ear) remote near-field HRTF have the similar spectrum structure of height.Here it is the reasons that near-field HRTF is deduced using far field HRTF
By basis.However, the research of far field HRTF and database often use head centre coordinate system.The study found that due near field sound
Relative to head center and relative to ear, there are apparent angular deviations in source.This means that the same angle distance field for ear
HRTF is different angle in head centre coordinate system.If the far field HRTF using existing head centre coordinate system derives near field
HRTF must carry out angle modification, to ensure that the far field HRTF and near-field target HRTF of extraction are same angle for ear
's.In view of the accuracy of HRTF frequency spectrums is vital, emphasis of the present invention for the accurate reproduction of middle vertical plane virtual sound image
The angle modification formula on middle vertical plane is derived.Similar method can also be generalized to other than middle vertical plane.
Compared with prior art, advantage of the invention is that:
(1) near-field HRTF data are derived using existing far field HRTF data, can thoroughly avoids complicated near-field HRTF
It measures;
(2) angular deviation of the near-field HRTF relative to ear and relative to head center is considered, it is public to use corresponding amendment
Formula, to improve the accuracy that far field HRTF derives near-field HRTF;
(3) present invention may be used the software that algorithmic language is worked out and be realized on multimedia computer, can also use letter
Number processing chip circuit or dedicated integrated circuit are realized, are set for multimedia computer, the broadcasting of various handheld portable audios
The short distance Virtual Sound of standby (such as mobile phone) etc. is reset.
Description of the drawings
Fig. 1 is the realization principle figure of the present invention.
Fig. 2 is the schematic diagram of angle modification of the present invention.
Fig. 3 is the signal flow graph that the multimedia computer of the embodiment of the present invention is realized.
Specific implementation mode
Following further describes the present invention with reference to the drawings, but the claimed scope of the invention is not limited to implement
The range that example indicates.
Fig. 1 is the realization principle figure of the synthetic method of near field virtual sound image on a kind of middle vertical plane of the present invention.It is considered closely
Sound source is deduced corresponding angle modification formula relative to head center and relative to the angular deviation of ear in the case of.According to
Angle modification formula can accurately extract near-field target HRTF data from HRTF baseline databases known to far field, so that it is guaranteed that
The accuracy that near field virtual sound image synthesizes on middle vertical plane.The synthetic method of near field virtual sound image can be avoided multiple on this middle vertical plane
The measurement of miscellaneous near-field HRTF can easily obtain the virtual sound image of the arbitrary distance of near field.
The synthetic method of near field virtual sound image on a kind of middle vertical plane, including step:
Step 1, input given data, including far field HRTF baseline databases, three distance parameters, described three distances
Parameter include the distance D of far field sound source and head center, number of people radius a, on middle vertical plane near-field target acoustic image and head center distance
d;
Step 2, according to formula, calculate the angle correction θ of auris dextra R:
Wherein, change the distance d of near-field target acoustic image and head center on middle vertical plane, so that it may to obtain the different distances of near field
Angle correction;
Step 3, extraction far field HRTFfarThe auris dextra HRTF of (θ, R) as near-field target acoustic image, i.e. HRTFnear(R);Extraction
Far field HRTFfarThe left ear HRTF, i.e. HRTF of (- θ, L) as near-field target acoustic imagenear(L);
Step 4, the ears HRTF by near-field target acoustic imagenearConvolution is carried out with input acoustical signal, and feeds and earphone or raises
Sound device carries out low voice speaking put.
Specifically, the far field is the region other than 1 meter of head center, the near field be apart from 1 meter of head center with
Interior region.
Specifically, the feature of the HRTF in the far field does not change with distance, and the feature of the HRTF near field is with distance
And change.
Specifically, the multiple far field HRTF databases being disclosed before the feelings the pulse with the finger-tip of HRTF baseline databases described in step 1,
Including Massachusetts Institute Technology's MIT HRTF databases, California, USA university Davis branch school CIPIC HRTF data
Library.Wherein, disclosed HRTF baseline databases can all indicate the measurement distance of its data, the i.e. distance of far field sound source and head center
D。
Specifically, the number of people radius a described in step 1 is averaged as 8.75cm, can also be obtained by directly measuring
?.For example, the three-dimensional dimension of somatosensory device (kinect that such as Microsoft produces) capture listener head may be used, and then really
The head radius of fixed specific hearer.
Specifically, the binaural signals synthesized in step 4 can be directly fed to Headphone reproducing;If thought highly of using raising one's voice
It puts, then needs to be inserted into string sound elimination algorithm.Wherein, the specific mathematical form of string sound elimination algorithm and the quantity of loud speaker, put
Mode is all related.
Fig. 2 is the schematic diagram of angle modification.Since the visual angle of figure is overlooked downwards from the crown, so middle vertical plane (is defined as
By the section that human body or so is divided equally) it is rendered as a dotted line for dividing head equally in left and right.Although the number of people is not an ideal in reality
Spherical shape, but it is existing research confirm spherical shape be the number of people a kind of good approximation;Simultaneously as the sense of hearing of the mankind has centainly
Resolution ratio, this approximation will not cause the sense of hearing to distort.In figure, O points represent head center, and A points represent auris dextra, and B points, which represent, it is expected to close
At near-field target acoustic image, ∠ BOC=θ represent correction angle, and C points represent acoustic image known to the far field determined according to correction angle, OA=
A, OB=d, OC=D.It can be seen from the figure that AB and AC is on same straight line, near field acoustic image B and far field acoustic image C relative to
The angle of auris dextra is identical, and corresponding far field and near-field HRTF have similar spectrum structure.Therefore, the HRTF of far field C points is extracted
It as the HRTF of near field B points, i.e., derives near field B point HRTF from far field C points HRTF, there is reasonability.Fig. 3 is that computer is realized
Flow chart, may be used matlab or C Plus Plus programming implement.
The above embodiment of the present invention be only to clearly illustrate example of the present invention, and not be to the present invention
Embodiment restriction.For those of ordinary skill in the art, it can also make on the basis of the above description
Other various forms of variations or variation.There is no necessity and possibility to exhaust all the enbodiments.It is all the present invention
All any modification, equivalent and improvement etc., should be included in the protection of the claims in the present invention made by within spirit and principle
Within the scope of.
Claims (7)
1. the synthetic method of near field virtual sound image on a kind of middle vertical plane, which is characterized in that including step:
Step 1, input given data, including far field HRTF baseline databases, three distance parameters, described three distance parameters
The distance d of near-field target acoustic image and head center on distance D, number of people radius a, middle vertical plane including far field sound source and head center;
Step 2, according to formula, calculate the angle correction θ of auris dextra R:
Step 3, extraction far field HRTFfarThe auris dextra HRTF of (θ, R) as near-field target acoustic image, i.e. HRTFnear(R);Extract far field
HRTFfarThe left ear HRTF, i.e. HRTF of (- θ, L) as near-field target acoustic imagenear(L);
Step 4, the ears HRTF by near-field target acoustic imagenearCarry out convolution with input acoustical signal, and feed earphone or loud speaker into
Row is low voice speaking to be put.
2. the synthetic method of near field virtual sound image on a kind of middle vertical plane as described in claim 1, it is characterised in that:The far field
For the region other than 1 meter of head center, the near field is the region within 1 meter of head center.
3. the synthetic method of near field virtual sound image on a kind of middle vertical plane as claimed in claim 2, it is characterised in that:The far field
HRTF feature not with distance and change, and the feature of the HRTF near field with distance and change.
4. the synthetic method of near field virtual sound image on a kind of middle vertical plane as described in claim 1, it is characterised in that:In step 1
The HRTF baseline databases are described to be disclosed using any of the multiple far field HRTF databases being disclosed at present
Multiple far field HRTF databases include Massachusetts Institute Technology's MIT HRTF databases, California, USA university Davis
Branch school CIPIC HRTF databases.
5. the synthetic method of near field virtual sound image on a kind of middle vertical plane as described in claim 1, it is characterised in that:In step 1
The number of people radius a is averaged as 8.75cm.
6. the synthetic method of near field virtual sound image on a kind of middle vertical plane as described in claim 1, it is characterised in that:In step 1
The number of people radius a is by directly measuring acquisition.
7. the synthetic method of near field virtual sound image on a kind of middle vertical plane as described in claim 1, it is characterised in that:In step 4
The binaural signals of synthesis are directly fed to Headphone reproducing;When using speaker playback, then need to be inserted into string sound elimination algorithm.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201710347867.8A CN107105384B (en) | 2017-05-17 | 2017-05-17 | The synthetic method of near field virtual sound image on a kind of middle vertical plane |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201710347867.8A CN107105384B (en) | 2017-05-17 | 2017-05-17 | The synthetic method of near field virtual sound image on a kind of middle vertical plane |
Publications (2)
Publication Number | Publication Date |
---|---|
CN107105384A CN107105384A (en) | 2017-08-29 |
CN107105384B true CN107105384B (en) | 2018-11-02 |
Family
ID=59669348
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN201710347867.8A Expired - Fee Related CN107105384B (en) | 2017-05-17 | 2017-05-17 | The synthetic method of near field virtual sound image on a kind of middle vertical plane |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN107105384B (en) |
Families Citing this family (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN110856094A (en) * | 2018-08-20 | 2020-02-28 | 华为技术有限公司 | Audio processing method and device |
CN110856095B (en) | 2018-08-20 | 2021-11-19 | 华为技术有限公司 | Audio processing method and device |
CN109286888B (en) * | 2018-10-29 | 2021-01-29 | 中国传媒大学 | Audio and video online detection and virtual sound image generation method and device |
CN109379694B (en) * | 2018-11-01 | 2020-08-18 | 华南理工大学 | Virtual replay method of multi-channel three-dimensional space surround sound |
CN109618274B (en) * | 2018-11-23 | 2021-02-19 | 华南理工大学 | Virtual sound playback method based on angle mapping table, electronic device and medium |
CN111246345B (en) * | 2020-01-08 | 2021-09-21 | 华南理工大学 | Method and device for real-time virtual reproduction of remote sound field |
Family Cites Families (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN102572676B (en) * | 2012-01-16 | 2016-04-13 | 华南理工大学 | A kind of real-time rendering method for virtual auditory environment |
US9426589B2 (en) * | 2013-07-04 | 2016-08-23 | Gn Resound A/S | Determination of individual HRTFs |
-
2017
- 2017-05-17 CN CN201710347867.8A patent/CN107105384B/en not_active Expired - Fee Related
Also Published As
Publication number | Publication date |
---|---|
CN107105384A (en) | 2017-08-29 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN107105384B (en) | The synthetic method of near field virtual sound image on a kind of middle vertical plane | |
US10397722B2 (en) | Distributed audio capture and mixing | |
Brinkmann et al. | A high resolution and full-spherical head-related transfer function database for different head-above-torso orientations | |
Jin et al. | Creating the Sydney York morphological and acoustic recordings of ears database | |
CN108616789B (en) | Personalized virtual audio playback method based on double-ear real-time measurement | |
US9131305B2 (en) | Configurable three-dimensional sound system | |
ES2526785T3 (en) | Apparatus and procedure to derive directional information and systems | |
JP7082126B2 (en) | Analysis of spatial metadata from multiple microphones in an asymmetric array in the device | |
US9578439B2 (en) | Method, system and article of manufacture for processing spatial audio | |
US10397728B2 (en) | Differential headtracking apparatus | |
US20180199137A1 (en) | Distributed Audio Microphone Array and Locator Configuration | |
US20180206039A1 (en) | Capturing Sound | |
TW201909657A (en) | Use multi-point sound field description to generate an enhanced sound field description or modified sound field description concept | |
EP3446309A1 (en) | Merging audio signals with spatial metadata | |
CN109804559A (en) | Gain control in spatial audio systems | |
CN111050271B (en) | Method and apparatus for processing audio signal | |
TW201426738A (en) | Apparatus and method for generating a plurality of parametric audio streams and apparatus and method for generating a plurality of loudspeaker signals | |
US20200107149A1 (en) | Binaural Sound Source Localization | |
CN109314832A (en) | Acoustic signal processing method and equipment | |
GB2563670A (en) | Sound source distance estimation | |
Ahrens et al. | A head-mounted microphone array for binaural rendering | |
KR102656969B1 (en) | Discord Audio Visual Capture System | |
CN108038291B (en) | Personalized head-related transfer function generation system and method based on human body parameter adaptation algorithm | |
US10708679B2 (en) | Distributed audio capture and mixing | |
CN110583030B (en) | Incoherent idempotent ambisonics rendering |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
GR01 | Patent grant | ||
GR01 | Patent grant | ||
CF01 | Termination of patent right due to non-payment of annual fee |
Granted publication date: 20181102 |
|
CF01 | Termination of patent right due to non-payment of annual fee |