CN108574925A - The method and apparatus that audio signal output is controlled in virtual auditory environment - Google Patents
The method and apparatus that audio signal output is controlled in virtual auditory environment Download PDFInfo
- Publication number
- CN108574925A CN108574925A CN201710149313.7A CN201710149313A CN108574925A CN 108574925 A CN108574925 A CN 108574925A CN 201710149313 A CN201710149313 A CN 201710149313A CN 108574925 A CN108574925 A CN 108574925A
- Authority
- CN
- China
- Prior art keywords
- azimuth information
- sound
- variation
- orientation
- bearing data
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Pending
Links
Classifications
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04S—STEREOPHONIC SYSTEMS
- H04S7/00—Indicating arrangements; Control arrangements, e.g. balance control
- H04S7/30—Control circuits for electronic adaptation of the sound field
- H04S7/302—Electronic adaptation of stereophonic sound system to listener position or orientation
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/16—Sound input; Sound output
Landscapes
- Engineering & Computer Science (AREA)
- Physics & Mathematics (AREA)
- Theoretical Computer Science (AREA)
- Health & Medical Sciences (AREA)
- Audiology, Speech & Language Pathology (AREA)
- General Health & Medical Sciences (AREA)
- Human Computer Interaction (AREA)
- General Engineering & Computer Science (AREA)
- General Physics & Mathematics (AREA)
- Acoustics & Sound (AREA)
- Signal Processing (AREA)
- Stereophonic System (AREA)
Abstract
The present invention provides a kind of method and apparatus controlling audio signal output in virtual auditory environment.The device includes:Audio input unit is configured as receiving one or more audio signals corresponding with one or more of virtual auditory environment sound source;Azimuth information acquiring unit is configured as obtaining the azimuth information of the object in the virtual auditory environment;Determination unit, it is configured as determining that one or more sound bearings data acquisition system corresponding with the azimuth information, one or more of sound bearing data acquisition systems characterize one or more of virtual auditory environment sound source transmission sound wave to the transmission characteristic in orientation corresponding with the azimuth information respectively for the object;Processing unit is configured as being based on one or more of sound bearing data acquisition systems, handle respectively one or more of audio signals, to obtain binaural signals;Audio output unit is configured as exporting the binaural signals.
Description
Technical field
The present invention relates to virtual auditory technologies, more particularly to the method for audio signal output is controlled in virtual auditory environment
And device.
Background technology
The sense of hearing of the mankind is that another important channel of external information is perceived in addition to vision.In real acoustic environments, sound
The sound wave of source radiation is transferred to attentive listener through through and Ambient approach, and the head through attentive listener, auricle, trunk etc. are raw
Ears are transferred to after reason structural scattering or reflection, thus the sound pressure signal received by ears includes the acoustics sky of sound source and environment
Between information.Auditory system (including high-rise nervous system) generates corresponding spatial hearing using these information, and to sound source
Positioning and to the subjective sensation of ambient acoustic environment.
Virtual auditory environment is so that attentive listener is generated like placing oneself in the midst of nature acoustics by artificially generated or control acoustic enviroment
The feeling of environment.Since binaural signals include the main information of sound, thus manual simulation's binaural signals can be passed through
The method that earphone or speaker playback is used in combination makes attentive listener generate the subjective sensation like placing oneself in the midst of certain acoustic environment.
In the prior art, some are disclosed, corresponding spatial hearing is fictionalized to attentive listener by earphone or loud speaker
Scheme.For example, in patent document CN 1402593A, a kind of signal processing of the earphone repeat of 5.1 access surround sound is provided
Method orientation effect in the head of earphone repeat can be eliminated, and can be without simulating listening room by using this method
Room reflections sound under the premise of, retransmit out more natural 5.1 access surround sound effect.
In these current schemes, it is only capable of realization and provides natural spatial hearing by earphone or loud speaker for attentive listener
Effect.And when attentive listener itself is moved or is rotated, the sound field experienced without changing therewith, to obtain
Immersive effects more true to nature.
Invention content
The present invention provides a kind of method and apparatus of control audio signal output.The method and device can be built virtually
Acoustic environments, make the difference for being in attentive listener therein according to own physical orientation, can hear and be consistent with its physical orientation
Thus the sound effect of conjunction makes attentive listener obtain true virtual auditory experience.Further, which can be with
The physical orientation variation for attentive listener is adjusted, and so that attentive listener is experienced the sound being consistent with the variation of its physical orientation and is imitated
Fruit, to enhance the virtual auditory experience of attentive listener.
The method according to an embodiment of the invention that audio signal output is controlled in virtual auditory environment, including:
First obtaining step obtains the azimuth information of the object in the virtual auditory environment;First determines step, is that the object is true
Fixed one or more sound bearings data acquisition system corresponding with the azimuth information, one or more of sound bearing data
It is extremely corresponding with the azimuth information that set characterizes one or more of virtual auditory environment sound source transmission sound wave respectively
Orientation transmission characteristic;And first processing step, be based on one or more of sound bearing data acquisition systems, respectively pair with
The corresponding one or more audio signals of one or more of sound sources are handled, to obtain the binaural sound letter for output
Number.
In a further embodiment, the above method further includes:Second obtaining step obtains in the virtual auditory environment
The object variation after azimuth information;Second determines step, determines for the object and believes with the orientation after the variation
Manner of breathing corresponding another or another multi-acoustical bearing data set, another or the another multi-acoustical bearing data set point
It is extremely opposite with the azimuth information after the variation that one or more of virtual auditory environment sound source transmission sound wave is not characterized
The transmission characteristic in the orientation answered;And second processing step, based on another or the another multi-acoustical bearing data set, point
Other pair of one or more audio signals corresponding with one or more of sound sources are handled, to obtain for the double of output
Otoacoustic signal.
The device according to an embodiment of the invention that audio signal output is controlled in virtual auditory environment, including:
Audio input unit, be configured as receiving and one or more of virtual auditory environment sound source it is corresponding one or
Multiple audio signals;Azimuth information acquiring unit is configured as obtaining the orientation letter of the object in the virtual auditory environment
Breath;Determination unit is configured as determining one or more sound bearings corresponding with the azimuth information for the object
Data acquisition system, one or more of sound bearing data acquisition systems characterize one or more of described virtual auditory environment respectively
Sound source transmits sound wave to the transmission characteristic in orientation corresponding with the azimuth information;Processing unit is configured as being based on institute
One or more sound bearing data acquisition systems are stated, one or more of audio signals are handled respectively, to obtain ears
Acoustical signal;Audio output unit is configured as exporting the binaural signals.
In a further embodiment, the azimuth information acquiring unit is configured to:It obtains and described virtually listens
Feel the azimuth information after the variation of the object in environment;The determination unit is configured to:It is true for the object
Fixed corresponding with the azimuth information after the variation another or another multi-acoustical bearing data set, it is described another or separately
Multi-acoustical bearing data set characterize respectively one or more of the virtual auditory environment sound source transmission sound wave to institute
State the transmission characteristic in the corresponding orientation of azimuth information after variation;The processing unit is configured to:Based on described
Another or another multi-acoustical bearing data set, are respectively handled one or more of audio signals, double to obtain
Otoacoustic signal.
Description of the drawings
Fig. 1 shows the control audio signal output according to an embodiment of the invention in single object virtual auditory environment
The structure diagram of device;
Fig. 2 shows the schematic diagrames of single object virtual reality system according to an embodiment of the invention;
Fig. 3 shows the control audio signal output according to an embodiment of the invention in single object virtual auditory environment
The flow chart of method;
Fig. 4 shows that the audio signal according to an embodiment of the invention that controlled in the multipair environment as virtual auditory exports
The structure diagram of device;
Fig. 5 shows the multipair schematic diagram as virtual reality system according to an embodiment of the invention;And
Fig. 6 shows that the audio signal according to an embodiment of the invention that controlled in more single object virtual auditory environment exports
Method flow chart.
Specific implementation mode
The transmission characteristic in sound source transmission sound wave to a certain specific orientation can be expressed as function data set, this characterization
The function data set of sound transmission characteristics can be used for handling audio signal, and audio signal is made to embody the sound source transmission sound wave
To the transmission characteristic in the orientation.When this treated audio signal is converted into sound via playback equipment, which is
It shows the sound source and transmits sound wave to the transmission characteristic in the orientation, attentive listener is enable to experience virtual sound source space side
Position.If characterizing sound source respectively with multiple function data set transmits sound wave to the transmission characteristic in multiple specific orientation, this is used in combination
Multiple function data set handle audio signal respectively, so that it may so that audio signal shows the sound source transmission sound wave extremely respectively
The transmission characteristic in multiple specific orientation.Virtual acoustic environments can be built by this scheme, on this basis, if will incline
The real physical orientation projection of hearer is the specific orientation in virtual auditory environment, attentive listener the real physical orientations of difference and
Correspondence is established between the specific orientation of difference in virtual auditory environment, so that it may so that attentive listener is according to own physical orientation
Difference hears the sound effect being consistent with own physical orientation.This technical solution proposed by the invention can make to listen attentively to
The set sense of space direction of the no longer unalterable embodiment of the sound that person experiences, and the physical orientation with attentive listener can be embodied
The sense of space direction being consistent.
In the inventive solutions, the function data set for characterizing sound transmission characteristics is referred to as sound source
Bearing data set.It can be such as head-related transfer function data to be suitable for the invention sound bearing data acquisition system
(Head Related Transfer Function, HRTF) gathers, interaural difference data (Interaural Time
Difference, ITD) set, any sound wave appropriate that can be characterized sound source and send out such as interaural intensity difference data (IID) set
It is transmitted to the data acquisition system of the transmission characteristic in a certain orientation.
ITD refers to the distance difference from left and right ear due to sound source so that acoustical signal reaches time difference when ears.IID is
Refer to the distance difference due to sound source from left and right ear so that acoustical signal reaches intensity difference when ears.ITD and IID is sound source position
Set the function with frequency of sound wave.When sound source localization data collection is combined into ITD and IID data acquisition systems, user can differentiate sound source position
In its left side or right side.
HRTF is the acoustic transmission function from sound source to ears in the case of free field, is used for describing in free found field
The sound wave that sound source is sent out is incident on the changing features occurred when certain point in duct at a certain angle.HRTF is sound source position, sound
The function of wave frequency rate and human body surface shape and property.From sound source to anthropological measuring, the unit impulse response of point is known as head phase
Close impulse response (Head Related Impulse Response, HRIR).HRTF is the Fourier transform of HRIR.Work as sound source
When location data collection is combined into HRTF data acquisition systems, user can differentiate sound source and be located at its front, rear, top, lower section, left side also
It is right side.
For a better understanding of the present invention, hereinafter will by taking HRTF data acquisition systems as an example to sound bearing data acquisition system this
Application in inventive technique scheme illustrates.Before each embodiment of the detailed narration present invention, HRTF is obtained first
Mode is taken to briefly introduce.
HRTF can be acquired by testing to measure.It measures and is usually carried out in free field (for example, anechoic room).HRTF's
Measuring principle is similar with the common measurement of acoustic transmission function.Sound source (for example, miniature loudspeaker system) generates pumping signal,
Sound pressure signal at subject's ears is picked up using a pair of of microphone, then finds out transfer function with the method for signal processing.It is logical
The relative position changed between sound source and subject is crossed, then can obtain the HRTF of different sound source positions.As for acoustic pressure at ears
Microphone can be placed on the end of duct simulator, be equivalent to eardrum by the pickup of signal to dummy head model subject
Position measures;And to true man subject, micro microphone can be placed on to ear canal mouth and measured.
HRTF can also be acquired by way of calculating.For example, head can be reduced to rigid ball, by ears letter
Opposite on spherical surface 2 points are turned to, HRTF is then calculated to the Raylei scattering formulas of plane incidence wave using rigid ball.Example again
Such as, in the case where considering that trunk influences, head and trunk can be reduced to the ball of 2 different radiis, then use Green
The method of function and multiple scattering calculates HRTF.Further, it is also possible to using the methods of laser or CT scan, by the outer of research object
Shape is converted into computer picture, then calculates HRTF with approximation method.
It the above is only and enumerated for made by several known modes for obtaining HRTF, it should be appreciated that by any
The HRTF that the known or following mode that may be developed obtains can be applied to the basic conception of the present invention, fall therefrom
Enter within protection scope of the present invention.
It is directed to single object virtual auditory environment and multipair as virtual auditory environment is to the present invention respectively below with reference to attached drawing
Each embodiment illustrate.The virtual auditory environment can be applied to several scenes, for example, fictitious air space flight is instructed
Experienced, virtual military training, virtual entertainment system, Virtual meeting etc..Under normal conditions, various application scenarios listed above
It works in coordination including virtual auditory environment and virtual vision environment, the two simultaneously and constructs relatively complete virtual reality ring
Border.In the reality environment, user is projected as experiencer as the object in reality environment, the behavior phase of user
It is projected the activity in reality environment for object with answering, for example, user can pass through the fortune certainly in actual environment
Move the movement to realize object in virtual reality scenario.It should be noted that referred to below for being easy to understand and describing letter
Just, the object that user is projected in user with the real world and reality environment is not distinguished strictly.For example, if necessary
Azimuth information of the object in reality environment is obtained, practical operation step is the orientation detection unit inspection worn by user
Survey physical orientation of the user in reality, then the correspondence between the object in user and reality environment based on real world
Relationship obtains azimuth information of the object in reality environment.It, can directly will be such for the ease of understanding and describing simplicity
Operation is known as detecting the azimuth information of object.In addition it should also be noted that, virtual auditory environment of the present invention can not
The individualism dependent on virtual vision environment.
Single object virtual auditory environment
Fig. 1 shows the control audio signal output according to an embodiment of the invention in single object virtual auditory environment
Device 10 (hereinafter referred to as " device 10 ").In single object virtual auditory environment, the number of object is only one.As shown in Figure 1,
Device 10 includes audio input unit 110, azimuth information acquiring unit 120, determination unit 130, processing unit 140 and audio
Output unit 150.
Referring to Fig.1, audio input unit 110 is for receiving and conveying audio signal to processing unit 140.Audio signal can
To be the audio signal for example acquired by microphone, or by the audio signal of audio player output.Audio signal can be
Such as monophonic, two-channel or stereo audio signal.
In order to determine that orientation of the object in virtual auditory environment, azimuth information acquiring unit 120 obtain the orientation of object
Information.The orientation information includes orientation of the object in the single object virtual auditory environment.
The azimuth information of object can provide unit (not shown) by azimuth information and azimuth information is supplied to obtain list
Member 120.The function that azimuth information provides unit can be accomplished in several ways, these modes can be used alone, can also
It is applied in combination.For example, user can detect user in virtual auditory equipped with orientation detection unit by orientation detection unit
Then the azimuth information detected is supplied to azimuth information acquiring unit 120 by the orientation of the object projected in environment.Or
Person, user can also receive user to it in virtual auditory environment equipped with user input interface by user input interface
The input in the orientation of middle projected object, is then supplied to azimuth information acquiring unit 120 by azimuth information.Orientation detection list
Member can be, such as any suitable element or member that can be used in detecting object aspect such as gyroscope, compass, accelerometer
Part combines.The input of user input interface may include for example word, symbol, voice or to the click of the icon on interface,
The various ways such as dragging.
After azimuth information acquiring unit 120 obtains the azimuth information of object, determination unit 130 is transmitted it to.By determining
130 object-based azimuth information of unit determines one or more HRTF data corresponding with its azimuth information for the object
Set, and will determine that information is transmitted to processing unit 140.Determine that information can be in any number of forms, for example, determining that information can be with
Including one or more HRTF data acquisition systems itself, or only include the rope of identified one or more HRTF data acquisition systems
Draw.For the latter, need to be stored with HRTF data acquisition systems corresponding with each index in processing unit 140.
The above-mentioned one or more HRTF data acquisition systems determined for the object are for characterizing in single object virtual auditory environment
One or more sound sources transmission sound wave to the object residing for orientation transmission characteristic.When opposite in single object virtual auditory environment
When the sound source of the object is single sound source, determination unit 130 determines one corresponding with the single sound source for the object
HRTF data acquisition systems.When in single object virtual auditory environment relative to the sound source of the object be multi-acoustical when, determination unit 130
Multiple HRTF data acquisition systems corresponding with multiple sound source are determined for the object.
HRTF numbers for characterizing the transmission characteristic in orientation residing for the transmission sound wave to object of the sound source in virtual auditory environment
It can be acquired by testing measurement method or calculation according to set.In one example, can by test measure or
Calculation obtains one-to-one azimuth information between each other and HRTF data acquisition systems in advance.Acquired correspondence can be with
It is pre-stored in the internal storage unit (not shown) of device 10, or the external memory except device 10 can also be pre-stored in
In.In this example embodiment, determination unit 130 can be based on the azimuth information of the object, and the azimuth information that prestores and HRTF numbers
According to the correspondence of set, the one or more HRTF data acquisition systems to match for the Object Selection and its azimuth information.
It should be understood that in the practical measurement of HRTF, it is typically only capable to accomplish the measurement to M confined space orientation,
And can not measurement of the limit to all dimensional orientations, therefore, if azimuth information acquiring unit 120 obtain and be transmitted to determine it is single
The azimuth information of member 130 is not included within the M confined space orientation, and determination unit 130 can also use Spatial Interpolation Method,
Such as linear interpolation method, calculate HRTF data acquisition systems corresponding with the azimuth information by M known HRTF data acquisition systems.
In another example, it can not also obtain and store corresponding between azimuth information and HRTF data acquisition systems in advance
Relationship, but by determination unit 130 according to the azimuth information of the object, and what is prestored calculate other correlations required when HRTF
Information calculates one or more HRTF data acquisition systems corresponding with its azimuth information in real time for the object.
Processing unit 140 is based on one corresponding with its azimuth information or more that determination unit 130 is that the object determines
A HRTF data acquisition systems, respectively to from one or more of audio input unit 110 and single object virtual auditory environment
The corresponding one or more audio signals of sound source are handled, and obtain the binaural signals for output, and transmit it to
Audio output unit 150.Wherein, it is wrapped in each HRTF data acquisition systems corresponding with its azimuth information determined for the object
Include two groups of HRTF data for left and right ear.
Specifically, for each sound source in virtual auditory environment, processing unit 140 is pair corresponding with the sound source first
Audio signal carry out Fourier transform, obtain the frequency domain representation of the audio signal.Then, processing unit 140 will correspond to a left side
Two groups of HRTF data of auris dextra carry out inverse fourier transform after being multiplied respectively with the frequency domain representation of audio signal, to obtain for defeated
The binaural signals gone out.
It is understood that it is above it is corresponding be signal processing mode in a frequency domain, and corresponding in time domain, be then by
Convolution is carried out corresponding to the two groups of HRIR data and audio signal of left and right ear, to obtain the binaural signals for output.
Binaural signals are supplied to user by audio output unit 150.
In view of orientation of the object in single object virtual auditory environment may change, therefore, further
In embodiment, device 10 is also based on the azimuth information after the variation of object, for its determination and the azimuth information phase after variation
Corresponding another or another multiple HRTF data acquisition systems, are that the Object Selection matches with its present orientation in real time to realize
HRTF data acquisition systems effect.It is described in detail below for the further embodiment.
When orientation of the object in virtual auditory environment changes, unit can be provided (in figure not by azimuth information
Show) it obtains the Orientation differences of the object and Orientation differences information is supplied to azimuth information acquiring unit 120.Azimuth information carries
It can be realized in several ways for the function of unit.It is virtually being listened for example, user can be detected by orientation detection unit
Feel the Orientation differences of the object projected in environment, and/or user is received to it in virtual auditory ring by user input interface
Then Orientation differences information is supplied to azimuth information acquiring unit 120 by the input of the Orientation differences of the object projected in border.
For example, orientation detection unit can in real time or periodically be detected the Orientation differences of the object, it can also be in response to
Triggering is detected the Orientation differences of the object.On the other hand, can also the object be inputted by user input interface at any time
Orientation differences.Orientation differences of the object in single object virtual auditory environment may include rotation in the horizontal direction,
Translation, rotation in the vertical direction, translation or its arbitrary combination.
Azimuth information acquiring unit 120 obtains the Orientation differences information that unit (not shown) is provided from azimuth information
Afterwards, the Orientation differences information of object-based former azimuth information and this acquisition, calculates the azimuth information after the variation of object,
And the azimuth information after variation is supplied to determination unit 130.Azimuth information after 130 object-based variation of determination unit,
For the object determine it is corresponding with its azimuth information after changing another or separately multiple HRTF data acquisition systems.
Alternatively, azimuth information, which provides unit, can also directly acquire the orientation after the variation of the object and by the side after variation
Position information is supplied to azimuth information acquiring unit 120.In this case, azimuth information acquiring unit 120 is without executing above-mentioned meter
Calculate, and the azimuth information after the variation of acquisition need to be only transmitted to determination unit 130, by determination unit 130 according to variation after
Azimuth information, for the object determine it is corresponding with its azimuth information after changing another or separately multiple HRTF data acquisition systems.
Determination unit 130 be object determine with change after azimuth information it is corresponding another or separately multiple HRTF data
The mode of set is that object determines one or more HRTF corresponding with azimuth information with the determination unit 130 being outlined above
The mode of data acquisition system is similar, and for simplicity, therefore not to repeat here.
Determination unit 130 determine with change after azimuth information it is corresponding another or separately multiple HRTF data acquisition systems
Afterwards, processing unit 140 is based on one or more HRTF data acquisition systems, to from audio input unit 110 and virtual auditory
The corresponding one or more audio signals of one or more of environment sound source are handled, and the binaural sound for output is obtained
Signal.Binaural signals are then supplied to user by audio output unit 150.
Processing unit 140 is based on another or another multiple HRTF data acquisition systems handle one or more audio signals
Mode and the processing unit 140 that is outlined above based on one or more HRTF data acquisition systems to one or more audio signals
The mode handled is similar, and for simplicity, therefore not to repeat here.
Hereinafter, by taking the single object virtual auditory environment for simulating seabeach scene as an example, in conjunction with Fig. 2 to the control audio of Fig. 1
The device of signal output in concrete application scene with being further described.
Fig. 2 shows exemplary single object virtual reality systems 20 according to an embodiment of the invention.As shown in Fig. 2, single
The device 10 of control audio signal output of the object virtual reality system 20 including audio/video player 210, Fig. 1 is (hereinafter referred to as
" device 10 "), azimuth information unit 220, bi-telephone 230 and virtual reality glasses 240 are provided.
With reference to Fig. 2, audio/video player 210 is for providing audio signal and vision signal.Audio/video player 210 can be with
It is any setting with audio and video playing function appropriate such as PC, smart mobile phone, personal digital assistant, MP3 player
It is standby.In the seabeach scene of the present embodiment, when bi-telephone 230 in user's wearing and virtual reality glasses 240, audio and video playing
Device 210 is to 240 transmission video signal of virtual reality glasses, the virtual vision in blue sea and white sandy beach to be presented to the user
Environment, and it is transferred through the audio signal that device 10 is handled to bi-telephone 230, it is listened with providing a user the virtual of sound of sea wave
Feel environment.In addition, in the seabeach scene of the present embodiment, it is assumed that the number of sound source is only one namely sea.
When the object that user is projected enters virtual reality scenario, azimuth information provides unit 220 and is existed by obtaining user
Residing physical orientation obtains the orientation of object that user is projected in virtual reality scenario in actual environment, and by object
Azimuth information be supplied to device 10.By taking azimuth information offer unit 220 is orientation detection unit as an example, in the sea of the present embodiment
In the scene of beach, orientation detection unit obtains object in the virtual seabeach scene by detecting the physical orientation residing for user
The variation in direction, position and the two, and azimuth information and Orientation differences information are supplied to device 10.Orientation detection unit
It can be arranged in any position appropriate for being able to detect that user's physical orientation, for example, being independently worn on the head of user
Position, or be integrated in bi-telephone 230 or virtual reality glasses 240.
By the left ear of object towards for sound source (namely sea) direction, it is assumed that object surface towards orientation when Sounnd source direction be 0 degree,
And angle increases when object rotates clockwise, then the orientation that orientation detection unit detects at this time is 90 degree.In device 10
Azimuth information acquiring unit (not shown) obtains 90 degree of azimuth informations from orientation detection unit and transmits it to dress
Set the determination unit (not shown) in 10.Then, it is determined for object by determination unit corresponding with 90 degree of azimuth informations
HRTF data acquisition systems, and by determine information be transmitted to the processing unit (not shown) in device 10.Processing unit is based on
Identified HRTF data acquisition systems corresponding with 90 degree of azimuth informations, to by the audio input unit in device 10 (in figure not
Showing) audio signal of sound of sea wave that is received from audio/video player 210 handled, obtain the binaural sound for output
Signal.Binaural signals are supplied to user by the audio output unit (not shown) in device 10 by bi-telephone 230.
Since the audio signal to the sound of sea wave from audio/video player 210 uses HRTF numbers corresponding with 90 degree of azimuth informations
It is handled according to set, therefore, the sound of sea wave that user is experienced will embody orientative feature corresponding with 90 degree of azimuth informations,
That is, the sound of sea wave that user is experienced is located at its left ear side.
Further, the physical orientation as user in actual environment changes, such as rotates clockwise 90 degree, then
The object that user is projected also rotates clockwise 90 degree in the scene of seabeach, the Orientation differences that orientation detection unit detects at this time
To rotate clockwise 90 degree, and the azimuth information acquiring unit that the orientation change information is supplied in device 10 (is not shown in figure
Go out).After azimuth information acquiring unit obtains the Orientation differences information for rotating clockwise 90 degree, object-based original azimuth information
(namely 90 degree of azimuth informations) and this Orientation differences information (namely rotate clockwise 90 degree of Orientation differences information) obtained,
The azimuth information (namely 180 degree azimuth information) after variation is calculated, and the azimuth information after variation is transmitted in device 10
Determination unit (not shown).Determination unit is determined and the side after the variation based on the azimuth information after variation for object
The position corresponding HRTF data acquisition systems of information, and will determine that information is supplied to the processing unit (not shown) in device 10.
Processing unit is based on identified HRTF data acquisition systems, to being regarded from sound by the audio input unit (not shown) in device 10
The audio signal of the sound of sea wave received at frequency player 210 is handled, and the binaural signals for output are obtained.Device 10
In audio output unit (not shown) binaural signals are supplied to user by bi-telephone 230.Due to coming from
The audio signal of the sound of sea wave of audio/video player 210 uses opposite with the azimuth information (180 degree azimuth information) after variation
The HRTF data acquisition systems answered are handled, and therefore, the sound of sea wave that user is experienced will embody corresponding to 180 degree azimuth information
Orientative feature, that is, the sound of sea wave that user is experienced is located at the dead astern of user.
It should be noted that although above-described embodiment is horizontal turn clockwise of the direction with object in virtual reality scenario
For 90 degree dynamic, the change in location of object it is not related to, it will be appreciated by those skilled in the art that technical scheme of the present invention is actually
It is applicable to more complicated object aspect variation.For example, the Orientation differences of the object can also include translation, or rotation with
The combination of translation.In addition, the translation and rotation are not limited to two dimensional surface, can be the arbitrary translation in three dimensions
And rotation.
It is further to note that virtual auditory environment is combined with virtual vision environment in above-mentioned seabeach application scenarios
It is described merely to technical scheme of the present invention is better described, technical scheme of the present invention can be only applied to virtually listen
Feel the application scenarios of environment.
Fig. 3 shows the control audio signal output according to an embodiment of the invention in single object virtual auditory environment
Method.This method includes the first obtaining step S310, the first determination step S320 and the first processing step S330.
With reference to Fig. 3, in the first obtaining step S310, the orientation letter of the object in single object virtual auditory environment is obtained
Breath.The orientation information includes orientation of the object in single object virtual auditory environment.
In first determines step S320, one or more sound source sides corresponding with its azimuth information are determined for the object
Position data acquisition system.The one or more sound bearing data acquisition system characterizes one or more in single object virtual auditory environment respectively
The transmission characteristic in the orientation residing for a sound source transmission sound wave to the object.
Determine that one or more sound bearings data acquisition system corresponding with its azimuth information can be by a variety of for object
Mode is realized.For example, can be based on the azimuth information of the object, and according to the azimuth information that prestores and sound bearing data set
The correspondence of conjunction, the one or more sound bearing data acquisition systems to match for the Object Selection and its azimuth information.Example again
It such as, can be based on the azimuth information of the object, for the calculation and object one or more sound source sides corresponding with its azimuth information
Position data acquisition system.
Sound bearing data acquisition system can be for example head related transfer function data acquisition system, interaural difference data acquisition system,
Interaural intensity difference data acquisition system etc. is any appropriate can to characterize sonic transmissions that sound source is sent out to the transmission characteristic in a certain orientation
Data acquisition system.
In the first processing step S330, based on identified one or more sound bearings data acquisition systems, respectively pair with
The corresponding one or more audio signals of one or more of single object virtual auditory environment sound source are handled, to obtain
Binaural signals for output.
In view of orientation of the object in single object virtual auditory environment may change, therefore, further
Embodiment in, be also based on the azimuth information after the variation of the object, for its determination with change after azimuth information it is opposite
Another or the another multiple HRTF data acquisition systems answered, are that the Object Selection matches with its present orientation in real time to realize
The effect of HRTF data acquisition systems.
In the further embodiment, in addition to above-mentioned first obtaining step, first determine step and the first processing step
Further include the second obtaining step, the second determination step and second processing step except rapid.
In the second obtaining step, the azimuth information after the variation of the object in single object virtual auditory environment is obtained.
Azimuth information after the variation includes the orientation after variation of the object in single object virtual auditory environment.
Second determine step in, for the object determine it is corresponding with its azimuth information after changing another or separately it is more
A sound bearing data acquisition system.This another or another multi-acoustical bearing data set characterize single object virtual auditory environment respectively
One or more of sound source transmission sound wave to the object residing for variation after orientation transmission characteristic.
Corresponding with its azimuth information after changing another or another multi-acoustical bearing data collection are determined for the object
Conjunction can also be realized in several ways.For example, can be based on the azimuth information after the variation of the object, and according to prestoring
The correspondence of azimuth information and sound bearing data acquisition system matches for the Object Selection with the azimuth information after it changes
Another or another multi-acoustical bearing data set.In another example can be right for this based on the azimuth information after the variation of the object
As calculating corresponding with its azimuth information after changing another or another multi-acoustical bearing data set.
In the second processing step, another or another multi-acoustical bearing data set based on determined by, respectively pair with
The corresponding one or more audio signals of one or more of single object virtual auditory environment sound source are handled, to obtain
Binaural signals for output.
It is multipair as virtual auditory environment
Fig. 4 shows that the audio signal according to an embodiment of the invention that controlled in the multipair environment as virtual auditory exports
Device 40 (hereinafter referred to as " device 40 ").In object virtual auditory environment, the number of object is multiple.As shown in figure 4, device
40 include audio input unit 410, azimuth information acquiring unit 420, determination unit 430, processing unit 440 and audio output
Unit 450.
With reference to Fig. 4, audio input unit 410 is for receiving and conveying audio signal to processing unit 440.For it is multipair as
Virtual auditory environment, audio signal can be the audio signals for example acquired by microphone, or by audio player output
Audio signal.The audio signal typically acquired by microphone can be the audio signal from each object, typically by audio
The audio signal of player output can be the audio signal of the non-object sound source in virtual auditory environment.Described herein comes
From the audio signal of each object, actually it is meant that referring to each user is acquired by microphone and be transmitted to audio input unit 410
Audio signal, herein be referred to as the audio signal from each object;Similar, described is non-in virtual auditory environment
The audio signal of object sound source is actually meant that finger by audio player plays audio file and is transmitted to audio input unit
410 audio signal is referred to as the audio signal of the non-object sound source in virtual auditory environment at this.Audio signal can be
Such as monophonic, two-channel or stereo audio signal.
In order to determine that orientation of the object in virtual auditory environment, azimuth information acquiring unit 420 obtain multipair as virtual
The azimuth information of each object in acoustic environments.
In one example, each object only receives the audio signal of the non-object sound source in virtual auditory environment.
Under this situation, the azimuth information of each object can only include orientation of the object in virtual auditory environment.
In another example, each object receives the audio signal of the non-object sound source in virtual auditory environment, with
And the audio signal from other objects;Or only receive the audio signal from other objects.It is described under this two classes situation
The azimuth information of each object includes that orientation and the object of the object in virtual auditory environment are opposite with other objects
Orientation.
The azimuth information of object can provide unit (not shown) by azimuth information and azimuth information is supplied to obtain list
Member 420.The function that azimuth information provides unit can be accomplished in several ways, these modes can be used alone, can also
It is applied in combination.For example, each user can respectively be equipped with orientation detection unit, user is detected in void by orientation detection unit
The orientation of the object projected in quasi- acoustic environments, is then supplied to azimuth information acquiring unit by the azimuth information detected
420.Orientation detection unit can be such as gyroscope, compass, accelerometer it is any it is appropriate can be used in detect object side
Element or the element combination of position.Alternatively, each user can also be equipped with independent user input interface, by User Defined its
The relative bearing in the orientation of the object projected in virtual auditory environment and the object and other objects.User inputs boundary
The input in face may include such as word, symbol, voice or to various ways the click of the icon on interface, dragging.It is logical
The azimuth information for crossing user input interface input is thereupon provided to azimuth information acquiring unit 420.Above-described detection side
Formula and User Defined mode both can be used alone, and can also be combined with each other use.
After azimuth information acquiring unit 420 obtains the azimuth information of each object, determination unit 430 is transmitted it to, by
Azimuth information of the determination unit 430 based on each object, respectively each object determine one corresponding with its azimuth information
Or multiple HRTF data acquisition systems, and will determine that information is transmitted to processing unit 440.Determine that information can be in any number of forms, example
Such as, it determines that information may include one or more HRTF data acquisition systems itself, or only includes identified one or more
The index of HRTF data acquisition systems.For the latter, need to be stored with HRTF data sets corresponding with each index in processing unit 440
It closes.
The above-mentioned one or more HRTF data acquisition systems determined for each object are multipair as virtual auditory environment for characterizing
One or more of sound source transmission sound wave to the object residing for orientation transmission characteristic.When phase in the multipair environment as virtual auditory
When sound source for certain an object is single sound source, determination unit 430 determines corresponding with the single sound source for the object
One HRTF data acquisition system.When in the multipair environment as virtual auditory relative to the sound source of certain an object be multi-acoustical when, determine
Unit 430 is determined for the object and multiple HRTF data acquisition systems corresponding to a sound source, corresponding to multiple sound source.
Determination unit 430 is that each object determines that one or more HRTF data acquisition systems corresponding with its azimuth information can
To realize in several ways.For example, determination unit 430 can be based on the azimuth information of each object, and the side that prestores
The correspondence of position information and HRTF data acquisition systems determines the one or more to match with its azimuth information for each object
HRTF data acquisition systems.Azimuth information and the acquisition modes of the correspondence of HRTF data acquisition systems have hereinbefore been explained, are
For the sake of simplicity, therefore not to repeat here.In another example determination unit 430 can be according to each object azimuth information, and prestore
Other relevant informations required when HRTF are calculated, one corresponding with its azimuth information or more is calculated in real time for each object
A HRTF data acquisition systems.
Processing unit 440 based on determination unit 430 be each object determine one corresponding with its azimuth information or
Multiple HRTF data acquisition systems, respectively to from audio input unit 410 with one or more in the multipair environment as virtual auditory
The corresponding one or more audio signals of a sound source are handled, and the binaural signals for output are obtained.Then, audio is defeated
Go out unit 450 and binaural signals are respectively supplied to corresponding user.
In view of orientation of each object in the multipair environment as virtual auditory and/or can with the relative bearing of other objects
It can change, therefore, in a further embodiment, device 40 is also based on the letter of the orientation after the variation of each object
Breath determines corresponding with its azimuth information after changing another or another multiple HRTF data acquisition systems for each object, thus
Realize be in real time the HRTF data acquisition systems that each Object Selection matches with the orientation after it changes effect.Below for
The further embodiment is described in detail.
When orientation of certain an object in virtual auditory environment in multiple objects and/or the relative bearing with other objects
When changing, the Orientation differences that unit (not shown) obtains the object can be provided by the azimuth information that the object is equipped with
And/or changes with the relative bearing of other objects and Orientation differences information is supplied to azimuth information acquiring unit 420.
In one example, each object only receives the audio signal of the non-object sound source in virtual auditory environment.
Under this situation, the Orientation differences information of each object can only include orientation of the object in virtual auditory environment and become
Change.
In another example, each object receives the audio signal of the non-object sound source in virtual auditory environment, with
And the audio signal from other objects;Or only receive the audio signal from other objects.It is described under this two classes situation
The azimuth information of each object includes Orientation differences and the object and other objects of the object in virtual auditory environment
Relative bearing changes.
The function that azimuth information provides unit can be accomplished in several ways, these modes can be used alone, and also may be used
To be applied in combination.For example, each user can respectively be equipped with orientation detection unit, by orientation detection unit come in real time, week
Phase property or the Orientation differences of object that are projected in virtual auditory environment in response to detection trigger user, then will inspection
The Orientation differences information measured is supplied to azimuth information acquiring unit 420.Side of the object in the multipair environment as virtual auditory
Position variation may include rotation or translation in the horizontal direction, and/or rotation or flat in the vertical direction
Move or a combination thereof.Alternatively, each user can also be equipped with independent user input interface, by User Defined, it is virtually being listened
The relative bearing of the Orientation differences and the object and other objects of feeling the object projected in environment changes.User inputs boundary
The input in face may include such as word, symbol, voice or to various ways the click of the icon on interface, dragging.It is logical
The Orientation differences information for crossing user input interface input is thereupon provided to azimuth information acquiring unit 420.Above-described inspection
Survey mode and User Defined mode both can be used alone, and can also be combined with each other use.
Above-described detection mode and User Defined mode can be combined with each other, jointly to azimuth information acquiring unit
420 provide Orientation differences information of the object in the multipair environment as virtual auditory.For example, when entering virtual auditory environment,
User can be by self-defined its object that is projected of input of user input interface in virtual auditory environment orientation, and/or
The relative bearing of the object and other objects, and it is supplied to azimuth information acquiring unit using azimuth information as initial orientation information
420.Then by the orientation detection unit provisioned in each user in real time, periodically or in response to detection trigger user
The Orientation differences of institute's projective object in virtual auditory environment, and it is supplied to azimuth information to obtain the Orientation differences information detected
Take unit 420.In the process, user can also be at any time by the self-defined regulating object of user input interface in virtual auditory ring
The relative bearing in orientation and/or the object and other objects in border.
After azimuth information acquiring unit 420 obtains Orientation differences information of each object in virtual auditory environment, it is based on
The Orientation differences information of the former azimuth information and this acquisition of each object, calculates the azimuth information after variation, and will be each
Azimuth information after the variation of object is supplied to determination unit 430.Orientation after variation of the determination unit 430 based on each object
Information, for each object determine it is corresponding with its azimuth information after changing another or separately multiple HRTF data acquisition systems.
Alternatively, the azimuth information that object is equipped with provide unit can also directly acquire the orientation after the variation of the object and/
Or with the relative bearing after the variation of other objects and the azimuth information after variation is supplied to azimuth information acquiring unit 120.
In this case, azimuth information acquiring unit 420 is without executing above-mentioned calculating, and only need to will be after the variation of each object of acquisition
Azimuth information be transmitted to determination unit 430, by the azimuth information after the variation of determination unit 430 according to each object, respectively
For each object determine it is corresponding with its azimuth information after changing another or separately multiple HRTF data acquisition systems.
Determination unit 430 be each object determine it is corresponding with its azimuth information after changing another or separately it is multiple
After HRTF data acquisition systems, processing unit 440 is based on one or more HRTF data acquisition systems, to coming from audio input unit 410
One or more audio signals corresponding with one or more of virtual auditory environment sound source handled, be used for
The binaural signals of output.Binaural signals are then respectively supplied to corresponding user by audio output unit 450.
The device according to an embodiment of the invention that audio signal output is controlled in the multipair environment as virtual auditory can be with
Positioned at distal end, for example, being integrated in remote server;It can also be located locally end, for example, being integrated in local audio and video playing
In equipment or Multi-Party Conference equipment.
It should be noted that although above-described embodiment is defeated with the control audio signal in the multipair environment as virtual auditory
The device gone out obtains the azimuth information of each object in multiple objects and is determined for each object corresponding with its azimuth information
One or more HRTF data acquisition systems for be described, should it will be appreciated by those skilled in the art that in other embodiments
The device of control audio signal output can also only obtain the azimuth information of the partial objects in multiple objects and be respectively the portion
Object is divided to determine one or more HRTF data acquisition systems corresponding with its azimuth information.For example, when the control audio signal is defeated
When the device gone out is located locally end, it can only obtain the azimuth information of native object and be determined for native object and be believed with its orientation
The corresponding one or more HRTF data acquisition systems of manner of breathing, the azimuth information of native object may include native object multipair as void
Orientation in quasi- acoustic environments and/or the relative bearing with other objects.The number of the native object can be one or more
It is a.
Hereinafter, by taking the multipair environment as virtual auditory for simulating Multi-Party Conference scene as an example, the control in conjunction with Fig. 5 to Fig. 4
The device of audio signal output in concrete application scene with being further described.
Fig. 5 shows according to an embodiment of the invention exemplary multipair as virtual reality system 50.As shown in figure 5, more
Object virtual reality system 50 include Multi-Party Conference equipment 510, the control audio signal of Fig. 4 output device 40 (hereinafter referred to as
" device 40 "), user input interface 520, bi-telephone 530 and virtual reality glasses 540.In this application scene, with dress
Set 40 be located locally hold and only receive native object azimuth information and variation azimuth information for illustrate.
With reference to Fig. 5, Multi-Party Conference equipment 510 provides vision signal and the audio signal from participant, the video
Signal is used to simulate the virtual vision environment of Multi-Party Conference scene, and the audio signal is for simulating the virtual of Multi-Party Conference scene
Acoustic environments.When bi-telephone 530 in user's wearing and virtual reality glasses 540, Multi-Party Conference equipment 510 is to virtual reality eye
540 transmission video signal of mirror, virtual Multi-Party Conference scene to be presented to the user.In the virtual Multi-Party Conference scene
In, the user and other participants can be appeared to be in the object of different direction respectively.On this basis, Multi-Party Conference
Equipment 510 is also transferred through the audio signal from other participants that device 40 is handled to bi-telephone 530, is in simulation
The sound of the object of different direction.
Assuming that in the Multi-Party Conference scene of the present embodiment, three participants are shared, they are to be located at 40 institute of device respectively
Local user A on ground and strange land participant B and C located elsewhere.Local user A wears 530 He of above-mentioned bi-telephone
Virtual reality glasses 540.After Multi-Party Conference equipment 510 is activated, local user A is projected as the sheet in virtual reality scenario
Ground object a, strange land participant B and C are projected respectively as other object b and c in virtual reality scenario.Local user A at this time
The object b and c in virtual reality scenario can be seen by virtual reality glasses 540, it can also be seen that at least partly seeing
The object a being projeced into itself in virtual reality scenario.
In the present embodiment, local user A can input the sheet itself projected by the way that user input interface 520 is self-defined
The relative bearing in orientation and object a and other object bs and c of the ground object a in Multi-Party Conference scene, and by this azimuth information
The azimuth information acquiring unit (not shown) being supplied to as initial orientation information in device 40.
Azimuth information acquiring unit (not shown) in device 40 obtains the azimuth information and transmits it to device
Determination unit (not shown) in 40 is that native object a determines two corresponding with the azimuth information by determination unit
HRTF data acquisition systems, and will determine that information is transmitted to the processing unit (not shown) in device 40.One of HRTF numbers
To the transmission feature of object a, another HRTF data acquisition system characterizes the sonic transmissions sent out as sound source according to set characterization object b
The sonic transmissions that object c is sent out as sound source to object a transmission feature.Then, the processing unit in device 40 (does not show in figure
Go out) based on identified two HRTF data acquisition systems, respectively to by the audio input unit (not shown) in device 40 from
The audio signal from object b and c received at Multi-Party Conference equipment 510 is handled, and the binaural sound for output is obtained
Signal.Binaural signals are posed to local use by the audio output unit (not shown) in device 40 by bi-telephone 530
Family A.Since the audio signal from object b and c is respectively by the processing unit in device 40 based on corresponding with azimuth information
Different HRTF data acquisition systems are handled, and the sound from object b and c that local user A is experienced will embody and orientation
The corresponding orientative feature of information.According to such scheme, user can be according only to sound in the Multi-Party Conference of virtual reality scenario
Just tell the orientation of different participants.For example it is assumed that the azimuth information of the self-defined inputs of local user A is that native object a exists
Middle, object b is in its left front, and object c is in its right front, then local user A is practical pleasant to the ear in the virtual Multi-Party Conference
The sound of the participant B arrived just in its left front, the sound of participant C also just in its right front and user's A settings, or
User A is the same by the virtual vision scene that virtual reality glasses 540 can see, to acoustically enhance user
Virtual reality experience.
Further, local user A can also adjust native object a by user input interface 520 at any time and in many ways can
Discuss the relative bearing in the orientation and object a and other object b and c in scene.
Azimuth information acquiring unit (not shown) in device 40 obtains the azimuth information after variation and is transmitted
It is that native object a is determined and the azimuth information after variation by determination unit to the determination unit (not shown) in device 40
Corresponding two HRTF data acquisition systems.Wherein, the sonic transmissions that a HRTF data acquisition systems characterization object b is sent out as sound source
To the transmission feature of object a, the sonic transmissions that another HRTF data acquisition system characterizes object c as sound source and sends out are to object a's
Transmission feature.Then, the processing unit (not shown) in device 40 is based on identified two HRTF data acquisition systems, to by
The sound from object b and c that audio input unit (not shown) in device 40 receives from Multi-Party Conference equipment 510
Frequency signal is handled, and the binaural signals for output are obtained.Audio output unit (not shown) in device 40 will
Binaural signals are supplied to local user by bi-telephone 530.Since the audio signal from object b and c is respectively by device
Processing unit in 40 is handled based on HRTF data acquisition systems corresponding with the azimuth information after variation, local user A institutes
The sound of the object b and c that experience will embody variation corresponding with azimuth information variation.According to such scheme, user is in void
The Orientation differences of different participants can be just told in the Multi-Party Conference of quasi- reality scene according only to sound.For example, in user A
On the basis of the azimuth information of preceding primary self-defined input, it is assumed that the azimuth information after the variation of the self-defined inputs of user A is, local
Object a is constant in middle, and object b is moved to its front, and object c is moved to its front-right, then user A is practical virtual at this
Meeting in the sound of participant B heard move to its front, the sound of participant C also moves to its front-right,
It is this time set with user A, or the same by the virtual vision scene that virtual reality glasses 540 can see, to
In the virtual reality experience for acoustically further enhancing user.
It is understood that for two strange lands the participant B and C, can also be equipped in its respective local side
There is device 40, to provide it similar virtual auditory environment.
Being combined to be described with virtual vision environment by virtual auditory environment in above-mentioned Multi-Party Conference application scenarios is only
Technical solution in order to better illustrate the present invention, it is to be understood that technical scheme of the present invention can be only applied to virtually
The application scenarios of acoustic environments.
Fig. 6 shows that the audio signal according to an embodiment of the invention that controlled in the multipair environment as virtual auditory exports
Method.This method includes the first obtaining step S610, the first determination step S620 and the first processing step S630.
With reference to Fig. 6, in the first obtaining step S610, obtain in multiple objects in the multipair environment as virtual auditory extremely
The azimuth information of a few object.The azimuth information of at least one object includes the object in the multipair environment as virtual auditory
Orientation and/or relative bearing with other objects.
In first determines step S620, one or more corresponding with its azimuth information is determined at least one object
A sound bearing data acquisition system.The one or more sound bearing data acquisition system is characterized respectively in the multipair environment as virtual auditory
The transmission characteristic in the orientation residing for one or more sound source transmission sound waves at least one object.
Determine that one or more sound bearings data acquisition system corresponding with its azimuth information can be at least one object
It realizes in several ways.For example, can be based on the azimuth information of at least one object, and according to the azimuth information to prestore
With the correspondence of sound bearing data acquisition system, one or more to match at least one Object Selection and its azimuth information
A sound bearing data acquisition system.In another example can be based on the azimuth information of at least one object, at least one object meter
Calculate one or more sound bearings data acquisition system corresponding with its azimuth information.
Sound bearing data acquisition system can be for example head related transfer function data acquisition system, interaural difference data acquisition system,
Interaural intensity difference data acquisition system etc. is any appropriate can to characterize sonic transmissions that sound source is sent out to the transmission characteristic in a certain orientation
Data acquisition system.
In the first processing step S630, based on identified one or more sound bearings data acquisition systems, respectively pair with
It is multipair as the corresponding one or more audio signals of one or more of virtual auditory environment sound source are handled, to obtain
Binaural signals for output.
In one example, for each object at least one object, other objects are multipair as virtual as this
One or more of acoustic environments sound source.
In view of orientation of at least one object in the multipair environment as virtual auditory may change, therefore,
In a further embodiment, the azimuth information being also based on after the variation of at least one object is at least one right for this
As determining corresponding with its azimuth information after changing another or another multiple HRTF data acquisition systems, to which realization is in real time
The effect for the HRTF data acquisition systems that at least one Object Selection matches with its present orientation.
In the further embodiment, in addition to above-mentioned first obtaining step, first determine step and the first processing step
Further include the second obtaining step, the second determination step and second processing step except rapid.
In the second obtaining step, the azimuth information after the variation of at least one object is obtained.At least one object
Variation after azimuth information include orientation after variation of the object in the virtual auditory environment and/or with other objects
Variation after relative bearing.
In second determines step, determined at least one object corresponding another with its azimuth information after changing
A or another multi-acoustical bearing data set.This another or another multi-acoustical bearing data set characterize respectively it is multipair as virtual
The transmission that one or more of acoustic environments sound source transmits the orientation after the variation residing for sound wave at least one object is special
Property.
Corresponding with its azimuth information after changing another or another multi-acoustical side are determined at least one object
Position data acquisition system can also be realized in several ways.For example, orientation that can be after the variation based at least one object
Information, and according to the correspondence of the azimuth information and sound bearing data acquisition system that prestore, at least one Object Selection with
Another or the another multi-acoustical bearing data set that its azimuth information after changing matches.In another example this can be based on extremely
Azimuth information after the variation of a few object, it is corresponding with its azimuth information after changing at least one calculation and object
Another or another multi-acoustical bearing data set.
In the second processing step, another or another multi-acoustical bearing data set based on determined by, respectively pair with
It is multipair as the corresponding one or more audio signals of one or more of virtual auditory environment sound source are handled, to obtain
Binaural signals for output.
It can be completely or partially according to each unit in the device of the control audio signal of various embodiments of the present invention output
It is realized by software, hardware, firmware or its arbitrary combination on ground.The example of hardware cell include but not limited to system on chip (SOC),
Integrated circuit (IC), application-specific integrated circuit (ASIC), field programmable gate array (FPGA), etc..
The device exported according to the control audio signal of various embodiments of the present invention may be implemented as stand-alone device, that is,
All units are all realized in a physical machine.Alternatively, according to the output of the control audio signal of various embodiments of the present invention
Device can also be a distributed apparatus, that is, certain units and other units be deployed in respectively different physical machine or
Position.
According between the device and external device (ED) of the output of the control audio signal of various embodiments of the present invention or according to this hair
Can be currently known by means of any between the inside each unit of the device of the control audio signal output of bright each embodiment or
The communication technology of exploitation communicates with one another in the future, and such technology includes but not limited to:Communication between process/thread, bluetooth, it is infrared,
Computer network wirelessly or non-wirelessly, communication network, etc..
The above description of the present invention is for enabling any those of ordinary skill of this field to realize or using invention.For
For those of ordinary skill in the art, various modifications of the invention will be apparent from, and defined herein general former
Reason can also be applied to other deformations in the case where not departing from the spirit and scope of the disclosure.Therefore, the present invention is not
It is limited to the example and design of this paper, but it is consistent with the widest scope of principle disclosed herein and novel features.
Claims (27)
1. a kind of method controlling audio signal output in virtual auditory environment, including:
First obtaining step obtains the azimuth information of the object in the virtual auditory environment;
First determines step, and one or more sound bearings data set corresponding with the azimuth information is determined for the object
It closes, one or more of sound bearing data acquisition systems characterize one or more of virtual auditory environment sound source biography respectively
The transmission characteristic in defeated sound wave extremely orientation corresponding with the azimuth information;And
First processing step is based on one or more of sound bearing data acquisition systems, respectively pair with one or more of sound
The corresponding one or more audio signals in source are handled, to obtain the binaural signals for output.
2. according to the method described in claim 1, wherein, further including:
Second obtaining step obtains the azimuth information after the variation of the object in the virtual auditory environment;
Second determine step, for the object determine it is corresponding with the azimuth information after the variation another or separately multiple sound
Source bearing data set, another or the another multi-acoustical bearing data set are characterized respectively in the virtual auditory environment
The transmission characteristic in one or more sound source transmission sound waves extremely orientation corresponding with the azimuth information after the variation;And
Second processing step, based on another or the another multi-acoustical bearing data set, respectively pair with it is one or more
The corresponding one or more audio signals of a sound source are handled, to obtain the binaural signals for output.
3. according to the method described in claim 1, wherein,
Described first determines that step includes:Based on the azimuth information, and according to the azimuth information to prestore and sound bearing data
The correspondence of set, the one or more of sound bearing data to match with the azimuth information for the Object Selection
Set.
4. according to the method described in claim 1, wherein,
Described first determines that step includes:It is that the calculation and object is corresponding with the azimuth information based on the azimuth information
One or more of sound bearing data acquisition systems.
5. according to the method described in claim 2, wherein,
Described second determines that step includes:Based on the azimuth information after the variation, and according to the azimuth information and sound source to prestore
The correspondence of bearing data set, for the azimuth information after the Object Selection and the variation match it is described another
Or another multi-acoustical bearing data set.
6. according to the method described in claim 2, wherein,
Described second determines that step includes:Based on the azimuth information after the variation, after the calculation and object and the variation
Azimuth information corresponding another or the another multi-acoustical bearing data set.
7. according to the method described in claim 1, wherein, the object is an object,
First obtaining step includes:The azimuth information of one object is obtained, the azimuth information includes one
Orientation of the object in the virtual auditory environment;
Described first determines that step includes:One or more sound sources corresponding with its azimuth information are determined for one object
Bearing data set, one or more of sound bearing data acquisition systems characterize respectively one in the virtual auditory environment or
Multi-acoustical transmits sound wave to the transmission characteristic in the orientation of one object.
8. according to the method described in claim 2, wherein, the object is an object,
Second obtaining step includes:Obtain the azimuth information after the variation of one object, the orientation after the variation
Information includes the orientation after variation of one object in the virtual auditory environment;
Described second determines that step includes:For one object determine it is corresponding with its azimuth information after changing another
Or another multi-acoustical bearing data set, another or the another multi-acoustical bearing data set characterize respectively described virtually listens
Feel the transmission characteristic in the orientation after one or more of environment sound source transmission sound wave to the variation of one object.
9. according to the method described in claim 1, wherein, the object is multiple objects,
First obtaining step includes:Obtain the azimuth information of at least one of the multiple object object, it is described at least
The azimuth information of one object includes orientation of the object in the virtual auditory environment and/or the contra with other objects
Position;
Described first determines that step includes:One or more corresponding with its azimuth information is determined at least one object
Sound bearing data acquisition system, one or more of sound bearing data acquisition systems characterize one in the virtual auditory environment respectively
The transmission characteristic in the orientation of a or multi-acoustical transmission sound wave at least one object.
10. according to the method described in claim 2, wherein, the object is multiple objects,
Second obtaining step includes:The azimuth information after the variation of at least one of the multiple object object is obtained,
Azimuth information after the variation of at least one object includes the side after variation of the object in the virtual auditory environment
Position and/or with the relative bearing after the variation of other objects;
Described second determines that step includes:It is determined at least one object corresponding with its azimuth information after changing another
One or another multi-acoustical bearing data set, another or the another multi-acoustical bearing data set characterize the void respectively
The biography in the orientation after one or more of quasi- acoustic environments sound source transmission sound wave to the variation of at least one object
Defeated characteristic.
11. method according to claim 9 or 10, wherein other for each object at least one object
Object is one or more of sound sources in the virtual auditory environment.
12. according to the method described in any one of claim 1-10, wherein the sound bearing data acquisition system includes following
One:
Head related transfer function data acquisition system;
Interaural difference data acquisition system;
Interaural intensity difference data acquisition system.
13. a kind of device controlling audio signal output in virtual auditory environment, including:
Audio input unit is configured as receiving and one or more of virtual auditory environment sound source corresponding one
A or multiple audio signals;
Azimuth information acquiring unit is configured as obtaining the azimuth information of the object in the virtual auditory environment;
Determination unit is configured as determining one or more sound bearings corresponding with the azimuth information for the object
Data acquisition system, one or more of sound bearing data acquisition systems characterize one or more of described virtual auditory environment respectively
Sound source transmits sound wave to the transmission characteristic in orientation corresponding with the azimuth information;
Processing unit is configured as being based on one or more of sound bearing data acquisition systems, respectively to one or more
A audio signal is handled, to obtain binaural signals;
Audio output unit is configured as exporting the binaural signals.
14. device according to claim 13, wherein
The azimuth information acquiring unit is configured to:Obtain the variation of the object in the virtual auditory environment
Azimuth information afterwards;
The determination unit is configured to:It is determined for the object corresponding with the azimuth information after the variation another
One or another multi-acoustical bearing data set, another or the another multi-acoustical bearing data set characterize the void respectively
The biography in one or more of quasi- acoustic environments sound source transmission sound wave extremely orientation corresponding with the azimuth information after the variation
Defeated characteristic;
The processing unit is configured to:It is right respectively based on another or the another multi-acoustical bearing data set
One or more of audio signals are handled, to obtain binaural signals.
15. device according to claim 14, wherein
The azimuth information acquiring unit is configured as:Receive the Orientation differences letter of the object in the virtual auditory environment
Breath, and it is based on the Orientation differences information and the azimuth information, determine the azimuth information after the variation of the object.
16. device according to claim 13, wherein
The determination unit is configured as:Based on the azimuth information, and according to the azimuth information to prestore and sound bearing data
The correspondence of set, the one or more of sound bearing data to match with the azimuth information for the Object Selection
Set.
17. device according to claim 13, wherein
The determination unit is configured as:It is that the calculation and object is corresponding with the azimuth information based on the azimuth information
One or more of sound bearing data acquisition systems.
18. device according to claim 14, wherein
The determination unit is configured as:Based on the azimuth information after the variation, and according to the azimuth information and sound source to prestore
The correspondence of bearing data set, for the azimuth information after the Object Selection and the variation match it is described another
Or another multi-acoustical bearing data set.
19. device according to claim 14, wherein
The determination unit is configured as:Based on the azimuth information after the variation, after the calculation and object and the variation
Azimuth information corresponding another or the another multi-acoustical bearing data set.
20. device according to claim 13, wherein the object is an object,
The azimuth information acquiring unit is configured as:The azimuth information of one object is obtained, the azimuth information includes
Orientation of one object in the virtual auditory environment;
The determination unit is configured as:One or more sound sources corresponding with its azimuth information are determined for one object
Bearing data set, one or more of sound bearing data acquisition systems characterize respectively one in the virtual auditory environment or
Multi-acoustical transmits sound wave to the transmission characteristic in the orientation of one object.
21. device according to claim 14, wherein the object is an object,
The azimuth information acquiring unit is configured as:Obtain the azimuth information after the variation of one object, the variation
Azimuth information afterwards includes the orientation after variation of one object in the virtual auditory environment;
The determination unit is configured as:For one object determine it is corresponding with its azimuth information after changing another
Or another multi-acoustical bearing data set, another or the another multi-acoustical bearing data set characterize respectively described virtually listens
Feel the transmission characteristic in the orientation after one or more of environment sound source transmission sound wave to the variation of one object.
22. device according to claim 13, wherein the object is multiple objects,
The azimuth information acquiring unit is configured as:The azimuth information of at least one of the multiple object object is obtained,
The azimuth information of at least one object include orientation of the object in the virtual auditory environment and/or with other objects
Relative bearing;
The determination unit is configured as:One or more corresponding with its azimuth information is determined at least one object
Sound bearing data acquisition system, one or more of sound bearing data acquisition systems characterize one in the virtual auditory environment respectively
The transmission characteristic in the orientation of a or multi-acoustical transmission sound wave at least one object.
23. device according to claim 14, wherein the object is multiple objects,
The azimuth information acquiring unit is configured as:Obtain the side after the variation of at least one of the multiple object object
Position information, the azimuth information after the variation of at least one object includes variation of the object in the virtual auditory environment
Rear orientation and/or with the relative bearing after the variation of other objects;
The determination unit is configured as:It is determined at least one object corresponding with its azimuth information after changing another
One or another multi-acoustical bearing data set, another or the another multi-acoustical bearing data set characterize the void respectively
The biography in the orientation after one or more of quasi- acoustic environments sound source transmission sound wave to the variation of at least one object
Defeated characteristic.
24. the device according to claim 22 or 23, wherein for each object at least one object,
Its object is one or more of sound sources in the virtual auditory environment.
25. device according to claim 13, wherein further include:
Azimuth information provides unit, and the orientation for being configured as providing the object to the azimuth information acquiring unit is believed
Breath.
26. device according to claim 14, wherein further include:
Azimuth information provides unit, is configured as after the variation that the azimuth information acquiring unit provides the object
Azimuth information or the object Orientation differences information.
27. according to the device described in any one of claim 13-23, wherein the sound bearing data acquisition system includes following
Any one:
Head related transfer function data acquisition system;
Interaural difference data acquisition system;
Interaural intensity difference data acquisition system.
Priority Applications (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201710149313.7A CN108574925A (en) | 2017-03-14 | 2017-03-14 | The method and apparatus that audio signal output is controlled in virtual auditory environment |
JP2017113712A JP2018152834A (en) | 2017-03-14 | 2017-06-08 | Method and apparatus for controlling audio signal output in virtual auditory environment |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201710149313.7A CN108574925A (en) | 2017-03-14 | 2017-03-14 | The method and apparatus that audio signal output is controlled in virtual auditory environment |
Publications (1)
Publication Number | Publication Date |
---|---|
CN108574925A true CN108574925A (en) | 2018-09-25 |
Family
ID=63578455
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN201710149313.7A Pending CN108574925A (en) | 2017-03-14 | 2017-03-14 | The method and apparatus that audio signal output is controlled in virtual auditory environment |
Country Status (2)
Country | Link |
---|---|
JP (1) | JP2018152834A (en) |
CN (1) | CN108574925A (en) |
Cited By (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN109582273A (en) * | 2018-11-26 | 2019-04-05 | 联想(北京)有限公司 | Audio-frequency inputting method, electronic equipment and audio output device |
CN114025287A (en) * | 2021-10-29 | 2022-02-08 | 歌尔光学科技有限公司 | Audio output control method, system and related components |
Citations (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN102064781A (en) * | 2010-10-29 | 2011-05-18 | 华为终端有限公司 | Method and device for adjusting audio frequency of terminal and terminal |
CN105163242A (en) * | 2015-09-01 | 2015-12-16 | 深圳东方酷音信息技术有限公司 | Multi-angle 3D sound playback method and device |
CN105872928A (en) * | 2016-04-26 | 2016-08-17 | 惠州Tcl移动通信有限公司 | Method and system for generating virtual surround sound based on mobile terminal |
CN106412751A (en) * | 2016-11-14 | 2017-02-15 | 惠州Tcl移动通信有限公司 | Earphone capable of discriminating direction and realization method thereof |
CN106484099A (en) * | 2016-08-30 | 2017-03-08 | 王杰 | Content reproduction apparatus, the processing system with the replay device and method |
Family Cites Families (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JPS5944197A (en) * | 1982-09-06 | 1984-03-12 | Matsushita Electric Ind Co Ltd | Headphone device |
JP3111455B2 (en) * | 1990-02-14 | 2000-11-20 | ソニー株式会社 | Sound signal reproduction device |
JPH0879899A (en) * | 1994-09-01 | 1996-03-22 | Sanyo Electric Co Ltd | Audio signal reproducing device |
JPH10230899A (en) * | 1997-02-24 | 1998-09-02 | Motoya Takeyama | Man-machine interface of aerospace aircraft |
-
2017
- 2017-03-14 CN CN201710149313.7A patent/CN108574925A/en active Pending
- 2017-06-08 JP JP2017113712A patent/JP2018152834A/en active Pending
Patent Citations (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN102064781A (en) * | 2010-10-29 | 2011-05-18 | 华为终端有限公司 | Method and device for adjusting audio frequency of terminal and terminal |
CN105163242A (en) * | 2015-09-01 | 2015-12-16 | 深圳东方酷音信息技术有限公司 | Multi-angle 3D sound playback method and device |
CN105872928A (en) * | 2016-04-26 | 2016-08-17 | 惠州Tcl移动通信有限公司 | Method and system for generating virtual surround sound based on mobile terminal |
CN106484099A (en) * | 2016-08-30 | 2017-03-08 | 王杰 | Content reproduction apparatus, the processing system with the replay device and method |
CN106412751A (en) * | 2016-11-14 | 2017-02-15 | 惠州Tcl移动通信有限公司 | Earphone capable of discriminating direction and realization method thereof |
Cited By (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN109582273A (en) * | 2018-11-26 | 2019-04-05 | 联想(北京)有限公司 | Audio-frequency inputting method, electronic equipment and audio output device |
US11202163B2 (en) | 2018-11-26 | 2021-12-14 | Lenovo (Beijing) Co., Ltd. | Audio output method, electronic device, and audio output apparatus |
CN114025287A (en) * | 2021-10-29 | 2022-02-08 | 歌尔光学科技有限公司 | Audio output control method, system and related components |
CN114025287B (en) * | 2021-10-29 | 2023-02-17 | 歌尔科技有限公司 | Audio output control method, system and related components |
Also Published As
Publication number | Publication date |
---|---|
JP2018152834A (en) | 2018-09-27 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US10911882B2 (en) | Methods and systems for generating spatialized audio | |
CN108616789B (en) | Personalized virtual audio playback method based on double-ear real-time measurement | |
US9906885B2 (en) | Methods and systems for inserting virtual sounds into an environment | |
JP5894634B2 (en) | Determination of HRTF for each individual | |
TWI687106B (en) | Wearable electronic device, virtual reality system and control method | |
US10257630B2 (en) | Computer program and method of determining a personalized head-related transfer function and interaural time difference function | |
US20150326963A1 (en) | Real-time Control Of An Acoustic Environment | |
JP2019536395A (en) | System and method for capturing an image of the pinna and using the pinna image to characterize human auditory anatomy | |
EP3687190B1 (en) | Mapping virtual sound sources to physical speakers in extended reality applications | |
JPH10500809A (en) | Binaural signal synthesis, head related transfer function and its use | |
JPWO2017134973A1 (en) | Sound output device, sound output method, program, and sound system | |
CN110267166B (en) | Virtual sound field real-time interaction system based on binaural effect | |
US11758347B1 (en) | Dynamic speech directivity reproduction | |
Roginska | Binaural audio through headphones | |
WO2016167007A1 (en) | Head-related transfer function selection device, head-related transfer function selection method, head-related transfer function selection program, and sound reproduction device | |
JP5697079B2 (en) | Sound reproduction system, sound reproduction device, and sound reproduction method | |
CN111492342A (en) | Audio scene processing | |
EP2822301B1 (en) | Determination of individual HRTFs | |
CN108574925A (en) | The method and apparatus that audio signal output is controlled in virtual auditory environment | |
US20240031759A1 (en) | Information processing device, information processing method, and information processing system | |
Geronazzo et al. | Acoustic selfies for extraction of external ear features in mobile audio augmented reality | |
US20230401968A1 (en) | Training system and method for sound directional discrimination ability | |
US11589181B1 (en) | System and method for realistic rotation of stereo or binaural audio | |
WO2019174442A1 (en) | Adapterization equipment, voice output method, device, storage medium and electronic device | |
WO2021194487A1 (en) | Head-related transfer functions with antropometric measurements |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
TA01 | Transfer of patent application right | ||
TA01 | Transfer of patent application right |
Effective date of registration: 20181227 Address after: Tokyo, Japan Applicant after: Toshiba terminal Solutions Ltd Address before: Tokyo, Japan, Japan Applicant before: Toshiba Corp Applicant before: Toshiba terminal Solutions Ltd |
|
WD01 | Invention patent application deemed withdrawn after publication | ||
WD01 | Invention patent application deemed withdrawn after publication |
Application publication date: 20180925 |