CN104837106B - A kind of acoustic signal processing method and device for spatialized sound - Google Patents

A kind of acoustic signal processing method and device for spatialized sound Download PDF

Info

Publication number
CN104837106B
CN104837106B CN201510271560.5A CN201510271560A CN104837106B CN 104837106 B CN104837106 B CN 104837106B CN 201510271560 A CN201510271560 A CN 201510271560A CN 104837106 B CN104837106 B CN 104837106B
Authority
CN
China
Prior art keywords
sound
control
audio signal
difference
intensity
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN201510271560.5A
Other languages
Chinese (zh)
Other versions
CN104837106A (en
Inventor
房大磊
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
SHANGHAI CONSERVATORY OF MUSIC
Original Assignee
SHANGHAI CONSERVATORY OF MUSIC
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by SHANGHAI CONSERVATORY OF MUSIC filed Critical SHANGHAI CONSERVATORY OF MUSIC
Priority to CN201510271560.5A priority Critical patent/CN104837106B/en
Publication of CN104837106A publication Critical patent/CN104837106A/en
Application granted granted Critical
Publication of CN104837106B publication Critical patent/CN104837106B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Landscapes

  • Stereophonic System (AREA)

Abstract

The present invention relates to a kind of acoustic signal processing method and device for spatialized sound, wherein, processing method is specially:Pending audio signal is received, position control parameter, loudness difference control intensity and the time difference control intensity of sound in space is obtained, spatialization processing is carried out to pending audio signal using independent control pattern or overall-in-one control schema pattern.The independent control pattern is specially:Pending audio signal is controlled respectively according to position control parameter, loudness difference control intensity and the time difference control intensity of the sound of acquisition in space, obtains the audio signal after spatialization.The overall-in-one control schema pattern is specially:Overall-in-one control schema is carried out to pending audio signal in 2 dimension control interfaces according to position control parameter, loudness difference control intensity and the time difference control intensity of the sound of acquisition in space, obtains the audio signal after spatialization.Compared with prior art, the present invention has the advantages that spatialization effect is good, algorithm is easy and effective.

Description

A kind of acoustic signal processing method and device for spatialized sound
Technical field
The present invention relates to Audio Signal Processing field, more particularly, to a kind of Audio Signal Processing for spatialized sound Method and device.
Background technology
So-called its key of sound spatialization audio system is the space attribute for embodying sound source and acoustic irradiation.Sound has There are many attributes, its most basic attribute includes and the wavelength of time correlation, frequency, and the loudness related to loudness, acoustic pressure The parameters such as level.Why human auditory can judge this sequence of acoustic characteristic that most basic is a little that the mankind gather around There are a pair of ears.Two ears are located at the left side and right side on head respectively.Human brain by the loudness between two ears, frequency and The difference of time carrys out location sound so that people can perceive out the position of sound source.
Positioned for the left and right of sound, there are 2 points of most obvious difference:
(1) in high-frequency range, because sound source to ears is apart from difference, loudness is made to create a difference, i.e. ILD (Interaural Level Differences, ears level error) or IID (Interaural Intensity Differences, ears loudness are poor).
(2) in all frequency ranges, due to ears in itself apart from difference, the time is created a difference, i.e. ITD (Interaural Time Differences, ears time difference).
Under normal circumstances, loudness difference and the collective effect of time difference cause people to can interpolate that sound position.Deeper level is ground Study carefully discovery over different frequencies, both difference play the role of different.Lord Rayleigh were in the Duplex proposed in 1907 Theory points out that low frequency positioning relies on the time difference, and high frequency positioning is poor by loudness.
Spatialized sound can be picked up by using 2 microphones, then be played back by 2 loudspeakers.Such as OTRF pickup standards, i.e., using 2 single directional microphones, vibrating diaphragm is mutually 110 degree of angle at a distance of 17 centimetres.17 centimetres of difference Away from make it that the time creates a difference;110 degree of angular compliance list directional microphone causes loudness to create a difference.During playback, 2 are used Loudspeaker independently plays back the sound of 2 microphone records, you can by the audio reproduction of spatialization.Exactly because with 2 microphones pair Spatialized sound is picked up, while have recorded loudness differences and time difference so that current live performance recording has non- Normal real spatial impression.
Spatialized sound can also be simulated by artificially producing loudness difference and time difference.Such case is picked up using monophonic It is very common in sound or music making.For example, same sound, deliver to 2 loudspeakers, but the loudness of two loudspeakers if Difference, then it is artificial to generate volume difference.Similarly, if the loudness of 2 loudspeakers is identical, but existence time is poor, then people For generate the time difference.
The method for producing loudness difference at present is easily achieved.Such as on sound console or Digital Audio Workstation software " acoustic image control " and " balance control " can be realized and the loudness ratio of two sound channels in two-channel is controlled.Acoustic image control is generally used In monophonic passage.Controlled by acoustic image, monophonic sounds can be assigned to two buses by a certain percentage.Thereby produce sound Degree is poor;Balance control is generally used for stereo channel.It can change the loudness ratio of stereo left and right passage by balancing control, Also it is poor that loudness is generated to a certain extent.
But the method for current generation time difference is more complicated, Major Difficulties have it is following some:
(1) method of generation time difference is to use carryover effects device.But current most of carryover effects devices are still with effect Processor form is present, and for reflection of the simulated sound in space, the least unit of time delay is ms, with Steinberg Exemplified by ModDelay, its time delay shows that least unit is 1ms, and the least unit that can be inputted is 0.1ms.Relative to for producing The exemplary value such as 0.26ms of raw time difference, precision have seemed very low.
(2) common delay disposal device conveniently, intuitively can not show and adjust the time delay between the passage of left and right.With Exemplified by Avid Pro Tools Mod Delay II delay disposal devices, the time delay of its left passage and right passage is only by two Vertical control is completed, and rather than the acoustic image control for loudness difference, is only completed by a control.
Due to above reason, when carrying out spatialization to sound during common music making, due to only applying Loudness is poor, causes final result to be unable to reach telepresenc and the sense of reality possessed by the spatialized sound of 2 microphone pickups.
The content of the invention
It is an object of the present invention to overcome the above-mentioned drawbacks of the prior art and provide a kind of spatialization effect is good, The simple and effective acoustic signal processing method and device for spatialized sound of algorithm.
The purpose of the present invention can be achieved through the following technical solutions:
A kind of acoustic signal processing method for spatialized sound, this method are specially:
Receive pending audio signal, obtain sound position control parameter in space, loudness difference control intensity and when Between difference control intensity, spatialization processing is carried out to pending audio signal using independent control pattern or overall-in-one control schema pattern.
The independent control pattern is specially:
According to position control parameter, loudness difference control intensity and the time difference control intensity point of the sound of acquisition in space It is other that pending audio signal is controlled, obtain the audio signal after spatialization.
The overall-in-one control schema pattern is specially:
Existed according to position control parameter, loudness difference control intensity and the time difference control intensity of the sound of acquisition in space Overall-in-one control schema is carried out to pending audio signal in 2 dimension control interfaces, obtains the audio signal after spatialization.
In 2 dimension control interfaces, X-axis controls the position of sound spatialization,
Following state be present in Y-axis:
A) lower position only loudness difference in Y-axis pole influences sound, b) Y-axis pole top position only time difference influence sound, c) Y-axis Centre position, loudness difference and time difference joint effect sound, d) Y-axis other positions, then the transition between tri- kinds of states of a, b, c.
The method for producing loudness difference is realized by changing the loudness ratio of two sound channels in two-channel;
The method of generation time difference adds delay disposal device to realize by some sound channel in two-channel.
A kind of audio signal processor for spatialized sound, including:
Pending audio signal input module, for receiving pending audio signal;
Position control parameter input module, for obtaining the position control parameter of sound in space;
Loudness difference control intensity input module, for obtaining loudness difference control intensity;
Time difference controls intensity input module, for obtaining time difference control intensity;
Control mode switch module, for according to input instruction switching control pattern;
Individual control module, for position control parameter, the loudness difference control intensity according to the sound of acquisition in space Pending audio signal is controlled respectively with time difference control intensity;
Overall-in-one control schema module is strong for being controlled according to position control parameter, the loudness difference of the sound of acquisition in space Degree and time difference control intensity carry out overall-in-one control schema in 2 dimension control interfaces to pending audio signal.
The pending audio signal input module includes left channel audio signal input block and right channel audio signal Input block.
Existing " acoustic image control " and " balance control " on sound console or Digital Audio Workstation software is only capable of passing through sound Degree difference to sound carry out spatialization, compare with real world and lacked the time difference so that the sound of spatialization lack telepresenc with Spatial impression, the result of sound spatialization is only embodied between 2 loudspeakers.Although it can be increased using reverberation and other effects device Spatial impression, but the spatial impression as caused by the time difference in real world can not be simulated.The present invention compared with prior art, has Following beneficial effect:
(1) this method uses loudness difference and time difference simultaneously, can simulate the result of sound localization in real world, and Control mode is still based on traditional " acoustic image control " so that this method both can be with completely compatible existing sound localization Method, localization of sound can also be strengthened by increasing the time difference.
(2) additionally, due to can with independent control loudness difference and the time difference influence sound intensity, can also by using only Time difference obtains some special effects.
(3) result of real world can be more nearly, is greatly enriched using the sound after this method processing, spatialization The expressive ability of works.
Brief description of the drawings
Fig. 1 is the schematic flow sheet of the present invention;
Fig. 2 realizes schematic diagram for the present invention in Max platforms;
Fig. 3 is Max platforms comb~algorithm filter principle schematic;
Fig. 4 is the schematic diagram for applying the inventive method in Max platforms under independent control pattern;
Fig. 5 is the schematic diagram for applying the inventive method in Max platforms under overall-in-one control schema pattern.
Embodiment
The present invention is described in detail with specific embodiment below in conjunction with the accompanying drawings.The present embodiment is with technical solution of the present invention Premised on implemented, give detailed embodiment and specific operating process, but protection scope of the present invention is not limited to Following embodiments.
The embodiment of the present invention provides a kind of acoustic signal processing method for spatialized sound, and this method includes simultaneously To loudness difference and the simulation of time difference, and by certain algorithm, make user with the concept of traditional " acoustic image control ", it is auxiliary With intuitively state modulator so that user to loudness difference and the time difference easily controlled, allow sound spatialization control More directly, and make its result truer.
The present invention solves the difficult point of the prior art mentioned in background technology using following methods:
(1) using as the small as possible audio processing algorithms of minimum delay unit, now for audio frequency process might not It is conventional delay effect device.Such as employed in the present embodiment and be originally used for making under the Max platforms of the companies of Cycling ' 74 The effect device comb of comb-filter effects~.
(2) using certain algorithm be transformed into common control device the time delay between the passage of left and right, and be aided with Extra parameter.Traditional " acoustic image control " is remained in this example, and is aided with loudness difference control intensity and time difference control by force Spend the two additional parameters.
As shown in figure 1, the inventive method is specially:Pending audio signal is received, obtains the position of sound in space Control parameter, loudness difference control intensity and time difference control intensity, are treated using independent control pattern or overall-in-one control schema pattern Handle audio signal and carry out spatialization processing.
Wherein, independent control pattern is specially:Controlled according to position control parameter, the loudness difference of the sound of acquisition in space Intensity processed and time difference control intensity are controlled to pending audio signal respectively, obtain the audio signal after spatialization.
Overall-in-one control schema pattern is specially:Controlled according to position control parameter, the loudness difference of the sound of acquisition in space Intensity and time difference control intensity carry out overall-in-one control schema in 2 dimension control interfaces to pending audio signal, obtain spatialization Audio signal afterwards.In 2 dimension control interfaces, X-axis controls the position of sound spatialization, and Y-axis has following state:
A) lower position only loudness difference in Y-axis pole influences sound, b) Y-axis pole top position only time difference influence sound, c) Y-axis Centre position, loudness difference and time difference joint effect sound, d) Y-axis other positions, then the transition between tri- kinds of states of a, b, c.
The above method is verified and implemented under the Max software platforms of the companies of Cycling ' 74.Max is a kind of figure Change programming language, it is widely used for such performance data by composer, concert performer, artist, software designer and research staff Processing, audio frequency process and various audio units.Due to its flexibility and opening, present Max has been depicted as Development of Interactive The all-purpose language of music performance software.
It is illustrated in figure 2 using Max to realize the flow of the above method, its middle port 1,2 is used to receive pending audio Signal, port 3 are used to obtain the position control parameter of sound in space, and port 4 is used to obtain loudness difference control intensity, port 5 in acquisition time difference control intensity.Because Max is generally controlled with MIDI relevant devices, therefore the data area of port 3,4,5 7bit, i.e., 0~127 are arranged to, complies fully with midi standard.
In the present embodiment, the method for producing loudness difference can be by changing the loudness ratio of two sound channels in two-channel come real It is existing.Benjamin Bauer draw prediction loudness difference and sound orientation angle according to two-way speaker model and spherical headform The formula of degree:
Wherein φ is the sound orientation angle that people feels, θ is the angle between two-way speaker;L and R is each sound The loudness of road loudspeaker.
Calculated according to above formula, when the two-way speaker angle that hearer uses is 60 degree, if the sound of left and right acoustic channels For degree difference in 30dB, the sensation that people position to sound is in ultra-Left or ultra-Right loudspeaker.Continuing increase loudness difference will not be more Obvious positioning change.But understand in view of conventional " acoustic image control " and voice signal is fully allocated to a passage, in this method Still the conventional method is used.Method of the volume using cosine wave table (cycle~) control loudness (*~) is controlled, L channel rings Degree control uses 0 degree to 90 degree phase, and the control of R channel loudness uses 270 degree to 360 degree phases, and interchannel phase difference is 270 degree.The position data of 3 inputs is converted into the acquiescence of 0 degree to 90 degree change by expr $ f1/127*0.25-0.125 first Centre position, i.e., 45 degree;4 inputs control the intensity of its change, are converted into by expr $ f1/127 by the 0 to 127 of midi standard 0 to 1, and it is multiplied by the position phase that an expr is drawn, you can realize the intensity that loudness difference influences.The actual control numerical value drawn By delivered to after $ 120 and line~smoothing processing cycle~, output 0 to 1 data be used for control loudness.
In the present embodiment, the method for generation time difference can add delay disposal device in the sound channel of some in two-channel and come Realize.Specific time delay can be drawn by Woodworth formula:
Wherein r is a radius, and c is speed of sound, and θ is the sound orientation angle that people feels.
Calculated according to above formula, as a diameter of 8.75cm of hearer, two-way speaker angle is 60 degree, if left The time difference of R channel, the sensation that people position to sound was in ultra-Left or ultra-Right loudspeaker in about 260us.It is poor with loudness Unlike, if continuing to increase the time difference, people can feel some other effects:
When the time difference reaching 10ms or so, sound positioning can be felt feel not in undelayed loudspeaker, hearer To the sound of delay loudspeaker;
When the time difference is in 10~30ms, sound positioning is felt still in undelayed loudspeaker, but can increase certain " presence " and volume.Wherein 15~25ms is generally used for making stereophonic effect for monophonic sounds;
When the time difference is in 30~50ms, people may feel that slight carryover effects;
When the time difference is in more than 50ms, people may feel that obvious delay effect.
The maximum delay numerical value used in this method is 10ms, therefore can obtain most strong positioning and feel.In order to carry The time precision of high latency, in this example using comb~, the algorithm filter principle is as shown in Figure 3.
It is 10ms, initial delay time 0ms that comb~parameter 100010 in this method, which specifies maximum delay time, Primary signal a=0, postpones signal intensity b=1, postponed signal intensity c=0.Due to comb~carried out to time of delay Smoothing processing, thus the time delay inputted be not necessarily accurate to the time of sampled point, therefore obtained the higher delay of precision Time.5 inputs are maximum delay time, i.e., the intensity that the time difference influences, and the numerical value first will by expr f1/127*10 The 0 to 127 of midi standard is converted into 0 to 10, and corresponding 0 to 10ms time delay, then the final time of L channel delay is first First pass through expr abs ($ f1/127-0.5) * 2* ($ f1>64), judging whether acoustic image positions are more than that 64 (acoustic image positions are opened Begin to the right), and 0 to 1 will be converted into from 3 acoustic image positions, then be multiplied with maximum delay time obtained above, that is, obtain a left side The final delay time of sound channel.The final time of R channel delay as L channel, simply judges whether acoustic image positions are less than 64, therefore use expr abs ($ f1/127-0.5) * 2* ($ f1<64).
Fig. 4 is the example that algorithm above is applied to sound spatialization in Max, uses independent control pattern.Wherein 3rd, acoustic image positions, loudness difference intensity and time difference intensity are corresponded to 4,5 respectively, its excursion meet the 0 of midi standard to 127。
Fig. 5 is the example that algorithm above is applied to sound spatialization in Max, uses overall-in-one control schema pattern.2 dimensions In control interface, X-axis control sound spatialization position, Y-axis below pole only loudness difference influence sound, Y-axis above pole only Time difference influences sound.When Y-axis is in centre, loudness difference and time difference joint effect sound.In other positions, then this 3 Transition between kind state.

Claims (4)

  1. A kind of 1. acoustic signal processing method for spatialized sound, it is characterised in that this method simultaneously using loudness difference and Time difference, the result of sound localization in real world is simulated, be specially:
    Pending audio signal is received, obtains position control parameter, loudness difference control intensity and the time difference of sound in space Intensity is controlled, spatialization processing is carried out to pending audio signal using independent control pattern or overall-in-one control schema pattern;
    The method for producing loudness difference is realized by changing the loudness ratio of two sound channels in two-channel;
    The method of generation time difference adds delay disposal device to realize by some sound channel in two-channel;
    The overall-in-one control schema pattern is specially:Controlled according to position control parameter, the loudness difference of the sound of acquisition in space Intensity and time difference control intensity carry out overall-in-one control schema in 2 dimension control interfaces to pending audio signal, obtain spatialization Audio signal afterwards;
    In the 2 dimension control interface, X-axis controls the position of sound spatialization,
    Following state be present in Y-axis:
    A) lower position only loudness difference in Y-axis pole influences sound, b) the Y-axis pole top position only time difference influences sound, c) among Y-axis Position, loudness difference and time difference joint effect sound, d) Y-axis other positions, then the transition between tri- kinds of states of a, b, c.
  2. 2. the acoustic signal processing method according to claim 1 for spatialized sound, it is characterised in that the independence Control model is specially:
    Control intensity right respectively according to position control parameter, loudness difference control intensity and the time difference of the sound of acquisition in space Pending audio signal is controlled, and obtains the audio signal after spatialization.
  3. A kind of 3. audio signal processor for spatialized sound, it is characterised in that including:
    Pending audio signal input module, for receiving pending audio signal;
    Position control parameter input module, for obtaining the position control parameter of sound in space;
    Loudness difference control intensity input module, for obtaining loudness difference control intensity;
    Time difference controls intensity input module, for obtaining time difference control intensity;
    Control mode switch module, for according to input instruction switching control pattern;
    Individual control module, for according to the sound of acquisition position control parameter in space, loudness difference control intensity and when Between difference control intensity pending audio signal is controlled respectively;
    Overall-in-one control schema module, for according to the sound of acquisition position control parameter in space, loudness difference control intensity and Time difference controls intensity to carry out overall-in-one control schema to pending audio signal in 2 dimension control interfaces;
    In the 2 dimension control interface, X-axis controls the position of sound spatialization,
    Following state be present in Y-axis:
    A) lower position only loudness difference in Y-axis pole influences sound, b) the Y-axis pole top position only time difference influences sound, c) among Y-axis Position, loudness difference and time difference joint effect sound, d) Y-axis other positions, then the transition between tri- kinds of states of a, b, c.
  4. 4. the audio signal processor according to claim 3 for spatialized sound, it is characterised in that described to wait to locate Reason audio signal input module includes left channel audio signal input block and right channel audio signal input block.
CN201510271560.5A 2015-05-25 2015-05-25 A kind of acoustic signal processing method and device for spatialized sound Active CN104837106B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201510271560.5A CN104837106B (en) 2015-05-25 2015-05-25 A kind of acoustic signal processing method and device for spatialized sound

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201510271560.5A CN104837106B (en) 2015-05-25 2015-05-25 A kind of acoustic signal processing method and device for spatialized sound

Publications (2)

Publication Number Publication Date
CN104837106A CN104837106A (en) 2015-08-12
CN104837106B true CN104837106B (en) 2018-01-26

Family

ID=53814689

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201510271560.5A Active CN104837106B (en) 2015-05-25 2015-05-25 A kind of acoustic signal processing method and device for spatialized sound

Country Status (1)

Country Link
CN (1) CN104837106B (en)

Families Citing this family (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN106412792B (en) * 2016-09-05 2018-10-30 上海艺瓣文化传播有限公司 The system and method that spatialization is handled and synthesized is re-started to former stereo file

Family Cites Families (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2522529B2 (en) * 1988-10-31 1996-08-07 株式会社東芝 Sound effect device
JP2005236502A (en) * 2004-02-18 2005-09-02 Yamaha Corp Sound system
EP1600791B1 (en) * 2004-05-26 2009-04-01 Honda Research Institute Europe GmbH Sound source localization based on binaural signals
CN101116374B (en) * 2004-12-24 2010-08-18 松下电器产业株式会社 Acoustic image locating device
CN101960866B (en) * 2007-03-01 2013-09-25 杰里·马哈布比 Audio spatialization and environment simulation
WO2013051085A1 (en) * 2011-10-03 2013-04-11 パイオニア株式会社 Audio signal processing device, audio signal processing method and audio signal processing program
EP3038385B1 (en) * 2013-08-19 2018-11-14 Yamaha Corporation Speaker device and audio signal processing method

Also Published As

Publication number Publication date
CN104837106A (en) 2015-08-12

Similar Documents

Publication Publication Date Title
CN105684467B (en) The ears of the earphone handled using metadata are presented
CN105874820B (en) Binaural audio is produced by using at least one feedback delay network in response to multi-channel audio
US10021507B2 (en) Arrangement and method for reproducing audio data of an acoustic scene
Jot Real-time spatial processing of sounds for music, multimedia and interactive human-computer interfaces
US9769589B2 (en) Method of improving externalization of virtual surround sound
JP4633870B2 (en) Audio signal processing method
US9794715B2 (en) System and methods for processing stereo audio content
US11516616B2 (en) System for and method of generating an audio image
US9967693B1 (en) Advanced binaural sound imaging
CN106797524A (en) Method and apparatus and computer readable recording medium storing program for performing for rendering acoustic signal
CN113170271A (en) Method and apparatus for processing stereo signals
Jot et al. Binaural simulation of complex acoustic scenes for interactive audio
CN106465027A (en) Apparatus and method for edge fading amplitude panning
KR102119240B1 (en) Method for up-mixing stereo audio to binaural audio and apparatus using the same
CN106954139A (en) A kind of sound field rendering method and system for combining earphone and loudspeaker
US11589184B1 (en) Differential spatial rendering of audio sources
EP2946573B1 (en) Audio signal processing apparatus
CN104837106B (en) A kind of acoustic signal processing method and device for spatialized sound
JP2004023486A (en) Method for localizing sound image at outside of head in listening to reproduced sound with headphone, and apparatus therefor
CN114363793B (en) System and method for converting double-channel audio into virtual surrounding 5.1-channel audio
DK180449B1 (en) A method and system for real-time implementation of head-related transfer functions
KR20060004528A (en) Apparatus and method for creating 3d sound having sound localization function
Aarts et al. NAG
Carty Artificial Simulation of Audio Spatialisation: Developing a Binaural System
De Sena et al. Introduction to Sound Field Recording and Reproduction

Legal Events

Date Code Title Description
C06 Publication
PB01 Publication
EXSB Decision made by sipo to initiate substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant