CN103826194B - Method and device for rebuilding sound source direction and distance in multichannel system - Google Patents

Method and device for rebuilding sound source direction and distance in multichannel system Download PDF

Info

Publication number
CN103826194B
CN103826194B CN201410071545.1A CN201410071545A CN103826194B CN 103826194 B CN103826194 B CN 103826194B CN 201410071545 A CN201410071545 A CN 201410071545A CN 103826194 B CN103826194 B CN 103826194B
Authority
CN
China
Prior art keywords
distance
sound
articulation
point
sound source
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Expired - Fee Related
Application number
CN201410071545.1A
Other languages
Chinese (zh)
Other versions
CN103826194A (en
Inventor
胡瑞敏
张茂胜
姚雪春
涂卫平
王晓晨
姜林
杨乘
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Wuhan University WHU
Original Assignee
Wuhan University WHU
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Wuhan University WHU filed Critical Wuhan University WHU
Priority to CN201410071545.1A priority Critical patent/CN103826194B/en
Publication of CN103826194A publication Critical patent/CN103826194A/en
Application granted granted Critical
Publication of CN103826194B publication Critical patent/CN103826194B/en
Expired - Fee Related legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Abstract

The invention discloses a method and device for rebuilding a sound source direction and distance in a multichannel system. The particle speed and sound pressure of a sound source received by a sound listening point are calculated according to known sound source signals, the distance between the sound listening point and the sound source is (i)r(/i), four loudspeakers in the direction are selected according to the particle speed, the sound source signals are multiplied by four weight factors respectively and then are distributed to the four selected loudspeakers, then particle speed and sound pressure of a sound image inducted at the sound listening point are calculated at a playback end after the loudspeakers in a rebuilt sound field send signals, finally, direction and distance equivalent models are built according to the particle speed and sound pressure of the original sound source and the particle speed and sound pressure of the sound image in the rebuilt sound field, and the signals of the loudspeakers are distributed through the weight factors acquired according to a solution model. Compared with the prior art, the direction and distance information of the sound source in the original sound source space can be accurately restored, the operation is simple, calculation efficiency is high, and stability is good.

Description

The method and apparatus that in a kind of multi-channel system, Sounnd source direction and distance are rebuild
Technical field
The invention belongs to field of multimedia signal processing, to particularly relate in Audio Signal Processing direction the directions such as sound source reconstruction, sound field rebuilding, sound source dimensional orientation and range recovery, be specifically related to the method and apparatus that in a kind of multi-channel system, Sounnd source direction and distance are rebuild.
Background technology
Although the tonequality of stereophony and sound field effect are better than monophony greatly, it still has obvious limitation.Dual-channel stereo system can only reproduce the sound field in the sector region of one, people front, and the sound reset can not be allowed to give Sensurround.So multichannel technology also starts to grow up.Multi-sound channel digital audio system, by the expansion of number of channels, can recover multiple ambient sound, brings better feeling of immersion and higher-quality audio frequency to enjoy, surmounted the effect of traditional monophony and stereophonic sound system far away to audience.Along with the development of the storage mediums such as DVD, SACD, multichannel audio has progressively been entered multiple fields of people's life by initial exclusive cinema.
Typical multi-channel system has Dolby Digital AC-3 multichannel surround sound sound system, THX Surround EX system and 22.2 multichannel stereo systems.
Dolby digital audio compression standard (AC-3) is the international standard proposed, and it solves digital sound and simulated sound on a film and the problem of depositing.Dolby Digital AC-3 multichannel surround sound sound system by six independently sound channel form.Because the Hz-KHz of the first five separate channels in these six separate channels is all audio frequency Whole frequency band and 20Hz ~ 20kHz, and subwoofer sound channel Hz-KHz only has 15Hz ~ 150Hz, only account for 1/10th of whole frequency spectrum, therefore Dolby Digital AC-3 multichannel surround sound sound system is also called 5.l sound channel ambiophonic system.
THX Surround EX system has strictly been worked out cinema and to be correlated with the standard of audio-visual equipment and environment, as long as meet THX standard and through certification, just can have suitable level.As long as consumer selects the movie theatre with THX certification like this, just have excellent audio-visual enjoyment.THX was transplanted to home theater afterwards, for the seehears of certification high-quality, and had unique requirement for the difference of home environment.But THX is not Dolby Digital and DTS is a kind of audio format like that, but a kind of audio post-processing pattern, object obtains best seeing and hearing enjoyment.When the Dolby Digital EX of 6.1 sound channels and DTS ES out after, it is evolved into THX Surround EX system by THX further.In order to the former Bidirectional sounding of compatibility side sound channel and reinforcing ring is around sound effect Sensurround once again, so sound channel turn increase two on the basis of former side sound channel after, this just constitutes 7.1 sound channels.
22.2 multichannel stereo systems are developed as " ultra high-definition " image sound equipment by NHK.By the loud speaker of upper strata 9 sound channel, middle level 10 sound channel, lower floor 3 sound channel three layers configuration, and dual track low frequency audio (LFE) loud speaker, the sound that left and right and above-below direction are forwards, backwards propagated can be reproduced comparatively truly.The sound technique of 22.2 sound channels needs, with the position of audience's ear level, lay 10,9 and 3 loudspeaker respectively above and below the position of audience's ear, to be furnished with 2 subwoofer be made up of 36 little bass units in addition.Whole ambiophonic system forms by three layers, and the sound field of the bottom is made up of three front channels and two LFE sound channels.Middle level sound field is made up of surround channel after four front channels, two side surround channels and three.Upper strata sound field is by three front channels, two side surround channels, surround channel is added a top sound channel and formed after four.In addition, in main sound channel, in order to strengthen effect, the dynamics sense that two trumpet arrays be made up of 36 toy trumpets export to guarantee main sound channel can also be selected.
5.1 ambiophonic systems, 22.2 multi-channel systems can bring good Sensurround to audience, to the reconstruction of sound all based on the process of signal itself, gratifying effect is achieved in tonequality fidelity, but the recovery lacked the orientation of sound source, cannot rebuild the positional information of holding and recovering sound source.
Summary of the invention
The deficiency that Sounnd source direction and range information exist is being rebuild in order to overcome existing multi-channel system, the invention provides the method and apparatus that in a kind of multi-channel system, Sounnd source direction and distance are rebuild, can accurately recover Sounnd source direction and distance in multi-channel system.
The technical scheme that method of the present invention adopts is: the method and apparatus that in a kind of multi-channel system, Sounnd source direction and distance are rebuild, is characterized in that, comprise the following steps:
Step 1: according to time-domain signal s (t) of known sound source, to listen the point of articulation for after initial point sets up Descartes's rectangular coordinate system in original sound field, the particle rapidity pv of the sound source of listening point of articulation place to receive that to calculate with sound source distance be r 0with acoustic pressure p 0;
Step 2: according to the particle rapidity pv of the sound source calculated in step 1 0, select to comprise particle rapidity pv 04 loud speaker L in direction 1, L 2, L 3, L 4;
Step 3: by the loud speaker L in sound-source signal s (t) and step 2 1, L 2, L 3, L 4, s (t) is multiplied by respectively 4 weight w 1, w 2, w 3, w 4after be assigned on four selected loud speakers;
Step 4: at playback end, calculates and rebuilds loud speaker L in sound field 1, L 2, L 3, L 4particle rapidity pv and the acoustic pressure p of the acoustic image of point of articulation place perception is listened after sending signal;
Step 5: according to the particle rapidity pv of the original sound source that step 1 calculates 0with acoustic pressure p 0, the particle rapidity pv of acoustic image and acoustic pressure p in the reconstruction sound field that calculates in step 4, set up direction, distance equivalence model;
Step 6: in the direction of step 5, apart from equivalence model, with the weight w arranged in step 3 1, w 2, w 3, w 4for unknown quantity, solve the value of direction, distance equivalence model acquisition weight;
Step 7: utilize the weight w solved in step 6 1, w 2, w 3, w 4, the signal carrying out loud speaker distributes, and makes the audio direction of reconstruction and distance with the direction of acoustic source with apart from consistent.
As preferably, the particle rapidity pv of the sound source of listening point of articulation place to receive that the calculating described in step 1 and sound source distance are r 0with acoustic pressure p 0,its specific implementation comprises following sub-step:
Step 1.1: utilize Fourier transform that time-domain signal s (t) of sound source is transformed into frequency domain by time domain, obtains frequency-region signal s (ω)
s ( ω ) = ∫ - ∞ + ∞ s ( t ) e - iwt dt
Step 1.2: by frequency-region signal s (ω) and the sound source position vector ε in sound field=(ε x, ε y, ε z), listen point of articulation position coordinates r=r (r x, r y, r z), according to the definition of sound physical properties attribute particle rapidity, calculate the particle rapidity pv listening point of articulation place to receive 0:
pv 0 ( r , ω ) = G e - ik | r - ϵ | | r - ϵ | ( 1 + 1 ik | r - ϵ | ) × 1 | r - ϵ | r x - ϵ x r y - ϵ y r z - ϵ z s ( ω ) ;
Calculate the acoustic pressure p listening point of articulation place to receive 0:
p 0 ( T 0 , ω ) = G e - ik | r - ϵ | | r - ϵ | s ( ω ) ;
Wherein k represents wave number, relevant with the velocity of sound with the frequency of voice signal, and e is constant, and i is imaginary unit, and ω is the angular frequency of signal s (t), | r-ε | represent sound source and listen the distance between the point of articulation, G (ω) represents source strength.
As preferably, the loud speaker L of 4 described in step 2 1, L 2, L 3, L 4, meet following condition: point of articulation r (x, y, z) and L will be listened respectively 1, L 2, L 3, L 4connect, in the polyhedron formed, the particle rapidity pv of sound source 0unit vector to be positioned at this polyhedron inner.
As preferably, loud speaker L in sound field is rebuild in the calculating described in step 4 1, L 2, L 3, L 4listen particle rapidity pv and the acoustic pressure p of the acoustic image of point of articulation place perception after sending signal, its specific implementation comprises following sub-step:
Step 4.1: at playback end, with the center of the ambiophonic system of multichannel composition for initial point, sets up cartesian coordinate system, loud speaker L icoordinate be designated as L i(L ix, L iy, L iz), listen the coordinate of the point of articulation to be r (r x, r y, r z);
Step 4.2: by loud speaker L 1, L 2, L 3, L 4the signal q sent 1(t), q 2(t), q 3(t), q 4t () is converted to frequency domain:
q ( ω ) = q 1 ( ω ) q 2 ( ω ) q 3 ( ω ) q 4 ( ω ) , q i ( ω ) = ∫ - ∞ + ∞ q i ( t ) e - iwt dt
Step 4.3: by loud speaker L 1, L 2, L 3, L 4send signal q 1(t), q 2(t), q 3(t), q 4t () calculates the frequency domain representation pv and the acoustic pressure p that listen the particle rapidity at point of articulation place:
p v ( r , ω ) = Σ i = 1 3 G e - ik | r - L i | | r - L i | ( 1 + 1 ik | r - L i | ) × 1 | r - L i | r x - L ix r y - L iy r z - L iz q ( ω ) = Σ i = 1 3 G e - ik | r - L i | | r - L i | 2 ( 1 + 1 ik | r - L i | ) × r x - L ix r y - L iy r z - L iz q ( ω )
p = G Σ j = 1 n e - ik | r - L i | | r - L i | q ( ω )
Wherein k represents wave number, relevant with the velocity of sound with the frequency of voice signal, and e is constant, and i is imaginary unit, and ω is the angular frequency of signal s (t), | r-L i| represent loud speaker and listen the distance between the point of articulation, G (ω) represents source strength.
As preferably, the particle rapidity pv of the original sound source calculated according to step 1 described in step 5 0with acoustic pressure p 0, the particle rapidity pv of acoustic image and acoustic pressure p in the reconstruction sound field that calculates in step 4, set up direction, distance equivalence model, its specific implementation comprises following sub-step:
Step 5.1: according to the particle rapidity p of the original sound source that step 1 calculates v0particle rapidity pv with acoustic image in the reconstruction sound field calculated in step 4, sets up direction equivalence relation as follows:
pv=pv 0
Namely
Σ i = 1 3 G e - ik | r - L i | | r - L i | 2 ( 1 + 1 ik | r - L i | ) × r ix - L ix r iy - L iy r iz - L iz q ( ω ) = G e - ik | r - ϵ | | r - ϵ | ( 1 + 1 ik | r - ϵ | ) × 1 | r - ϵ | r x - ϵ x r y - ϵ x r z - ϵ z s ( ω )
Wherein k represents wave number, relevant with the velocity of sound with the frequency of voice signal, and e is constant, and i is imaginary unit, and ω is the angular frequency of signal s (t), | r-L i| represent loud speaker and listen the distance between the point of articulation, | r-ε | represent sound source and listen the distance between the point of articulation, G (ω) represents source strength;
Step 5.2: according to the acoustic pressure p of the original sound source that step 1 calculates 0acoustic pressure p with acoustic image in the reconstruction sound field calculated in step 4, sets up distance equivalence relation as follows:
p=p 0
Namely
G Σ j = 1 n e - ik | r - L i | | r - L i | q ( ω ) = G e - ik | r - ϵ | | r - ϵ | s ( ω )
Wherein k represents wave number, relevant with the velocity of sound with the frequency of voice signal, and e is constant, and i is imaginary unit, and ω is the angular frequency of signal s (t), | r-L i| represent loud speaker and listen the distance between the point of articulation, | r-ε | represent sound source and listen the distance between the point of articulation, G (ω) represents source strength;
Step 5.3: according to step 5.1 and 5.2 direction equivalence relation and distance equivalence relation, set up direction, distance equivalence model:
pv = pv 0 p = p 0
The technical scheme that device of the present invention adopts is: a kind of method that in multi-channel system utilized described in claim 1, Sounnd source direction and distance are rebuild carries out the device that in multi-channel system, Sounnd source direction and distance are rebuild, it is characterized in that, comprise: point of articulation directional information computing module (1) is listened in sound source space, point of articulation range information computing module (2) is listened in sound source space, module (3) selected by loud speaker, signal forward allocator module (4), rebuild sound field Sounnd source direction information computational module (5), rebuild sound field sound source range information computing module (6), model building module (7), model solution module (8), signal distribution module (9),
The direction of the sound source that described sound source space listens point of articulation directional information computing module (1) to perceive for the pleasant to the ear point of articulation in sound source space calculating acoustic source place, signal s (t) according to sound source calculates the particle rapidity pv listening the point of articulation 0, and by pv 0export model building module (7) to;
The distance of the sound source that described sound source space listens point of articulation range information computing module (2) to perceive for the pleasant to the ear point of articulation in sound source space calculating acoustic source place, signal s (t) according to sound source calculates the acoustic pressure p listening the point of articulation 0, and by p 0export model building module (7) to;
Described loud speaker select module (3) for set up reconstruction multi-channel system in the selection principle of loud speaker, determine the loud speaker L selected in multi-channel system 1, L 2, L 3, L 4, and export selected loud speaker to signal forward allocator module (4);
Described signal forward allocator module (4) is for being multiplied by weight w by original audio signal s (t) 1, w 2, w 3, w 4after distribute to the loud speaker L that loud speaker selects to select in module (3) 1, L 2, L 3, L 4, then the signal after distribution is exported to and rebuilds sound field Sounnd source direction information computational module (5) and rebuild sound field sound source range information computing module (6);
Described reconstruction sound field acoustic image directional information computing module (5) receives the direction of acoustic image for calculating the pleasant to the ear point of articulation place of multi-channel system, according in signal forward allocator module (4) to the preallocated signal of loud speaker, the pleasant to the ear point of articulation place of multi-channel system utilizing acoustic theory to calculate reconstruction receives the particle rapidity pv of acoustic image, and pv is exported to model building module (7);
Described reconstruction sound field audio-visual distance information computing module (6) receives the distance of acoustic image for calculating the pleasant to the ear point of articulation place of multi-channel system, according in signal forward allocator module (4) to the preallocated signal of loud speaker, the pleasant to the ear point of articulation place of multi-channel system utilizing acoustic theory to calculate reconstruction receives the acoustic pressure p of acoustic image, and p is exported to model building module (7);
Described model building module (7) for set up listen point of articulation prescription to perceived distance consistency model, the sound source particle rapidity p listening point of articulation directional information computing module (1) to export by sound source space v0direction equivalence relation is set up, the sound source acoustic pressure p listening point of articulation range information computing module (2) to export by sound source space with the acoustic image particle rapidity pv rebuilding output in sound field Sounnd source direction information computational module (4) 0distance equivalence relation is set up with the acoustic image acoustic pressure p rebuilding output in sound field Sounnd source direction information computational module (6), set up direction, distance equivalence model according to direction equivalence relation and distance equivalence relation, and export this model to model solution module (8);
Described model solution module (8) is set up the model of foundation in module (7) for solving model and then is obtained four weight w 1, w 2, w 3, w 4value, finally weights are exported to signal distribution module (9);
The weight w that described signal distribution module (9) solves for utilizing model solution module (8) 1, w 2, w 3, w 4, the signal carrying out loud speaker distributes, and makes the audio direction of reconstruction and distance with the direction of acoustic source with apart from consistent.
The present invention, relative to prior art, can recover direction and the range information of the sound source in acoustic source space accurately, and simple to operate, and computational efficiency is high, good stability.
Accompanying drawing explanation
Fig. 1: the device workflow diagram of the embodiment of the present invention.
Embodiment
By reference to the accompanying drawings technical scheme of the present invention and system are described further with specific embodiment below.
The technical scheme that method of the present invention adopts is: Sounnd source direction and distance reconstructing device and method in a kind of multi-channel system, comprise the following steps:
Step 1: according to time-domain signal s (t) of known sound source, to listen the point of articulation for after initial point sets up Descartes's rectangular coordinate system in original sound field, the particle rapidity pv of the sound source of listening point of articulation place to receive that to calculate with sound source distance be r 0with acoustic pressure p 0;
Step 1.1: utilize Fourier transform that time-domain signal s (t) of sound source is transformed into frequency domain by time domain, obtain frequency-region signal s (ω), in this example, time-domain signal s (t) adopts sinusoidal signal, and the sample rate of signal is 48000Hz.Computational process is as follows:
s ( ω ) = ∫ - ∞ + ∞ s ( t ) e - iwt dt
Step 1.2: this example, to listen the point of articulation for initial point, namely listens the coordinate r (x, y, z) of the point of articulation to be (0,0,0), by frequency-region signal s (ω) and the sound source position vector ε in sound field=(ε x, ε y, ε z), listen point of articulation position coordinates (0,0,0), according to the definition of sound physical properties attribute particle rapidity, calculate listen point of articulation place to receive particle rapidity pv 0, computational methods are as follows:
pv 0 ( r , ω ) = G e - ik | ϵ | | ϵ | ( 1 + 1 ik | ϵ | ) × 1 | ϵ | - ϵ x - ϵ y - ϵ z s ( ω )
Calculate the acoustic pressure p listening point of articulation place to receive simultaneously 0:
p 0 ( T 0 , ω ) = G e - ik | ϵ | | ϵ | s ( ω )
Wherein k represents wave number, relevant with the velocity of sound with the frequency of voice signal, and e is constant, and i is imaginary unit, and ω is the angular frequency of signal s (t), | ε | represent sound source and listen the distance between the point of articulation, G (ω) represents source strength.
Step 2: according to the particle rapidity pv of the sound source calculated in step 1 0, select to comprise particle rapidity pv 04 loud speaker L in direction 1, L 2, L 3, L 4, wherein L 1, L 2, L 3, L 4meet following condition: point of articulation r (x, y, z) and L will be listened respectively 1, L 2, L 3, L 4connect, in the polyhedron formed, the particle rapidity pv of sound source 0unit vector to be positioned at this polyhedron inner.
Step 3: by the loud speaker L in sound-source signal s (t) and step 2 1, L 2, L 3, L 4, s (t) is multiplied by respectively 4 weight w 1, w 2, w 3, w 4after be assigned on four selected loud speakers, obtain four preallocated signal q of loud speaker 1(t), q 2(t), q 3(t), q 4(t);
q ( t ) = q 1 ( t ) q 2 ( t ) q 3 ( t ) q 4 ( t ) = W · s ( t ) = w 1 w 2 w 3 w 4 s ( t )
Wherein W represents weight vector, and its value is:
W = w 1 w 2 w 3 w 4
Q (t)=(q 1(t), q 2(t), q 3(t), q 4(t)) T represents loud speaker L 1, L 2, L 3, L 4the signal phasor distributed, T represents transpose operation.
Step 4: calculate and rebuild in sound field by loud speaker L 1, L 2, L 3, L 4listen particle rapidity pv and the acoustic pressure p of the acoustic image of point of articulation place perception after sending signal, detailed process comprises following 3 sub-steps:
Step 4.1: at playback end, with the center of the ambiophonic system of multichannel composition for initial point, sets up cartesian coordinate system.The coordinate of loud speaker Li is designated as L i(L ix, L iy, L iz), in this example, listen the position of the point of articulation and the position consistency of central point, namely listen the coordinate of the point of articulation to be r (r x, r y, r z)=(0,0,0);
Step 4.2: by loud speaker L 1, L 2, L 3, L 4the signal q sent 1(t), q 2(t), q 3(t), q 4t () is converted to frequency domain:
q ( ω ) = q 1 ( ω ) q 2 ( ω ) q 3 ( ω ) q 4 ( ω ) , q i ( ω ) = ∫ - ∞ + ∞ q i ( t ) e - iwt dt
Step 4.3: by loud speaker L 1, L 2, L 3, L 4send signal q 1(t), q 2(t), q 3(t), q 4t () calculates the frequency domain representation pv and the acoustic pressure p that listen the particle rapidity at point of articulation place:
p v ( r , w ) = Σ i = 1 3 G e - ik | L i | | L i | ( 1 + 1 ik | L i | ) × 1 | L i | - L ix - L iy - L iz q ( ω ) = Σ i = 1 3 G e - ik | L i | | L i | 2 ( 1 + 1 ik | L i | ) × - L ix - L iy - L iz q ( ω )
p = G Σ j = 1 n e - ik | L i | | L i | q ( ω )
Wherein k represents wave number, relevant with the velocity of sound with the frequency of voice signal, and e is constant, and i is imaginary unit, and ω is the angular frequency of signal s (t), | L i| represent loud speaker and listen the distance between the point of articulation, G (ω) represents source strength;
Step 5: according to the particle rapidity pv of the original sound source that step 1 calculates 0with acoustic pressure p 0, the particle rapidity pv of acoustic image and acoustic pressure p in the reconstruction sound field that calculates in step 4, set up direction, distance equivalence model, comprise following 3 sub-steps:
Step 5.1: according to the particle rapidity pv of the original sound source that step 1 calculates 0particle rapidity pv with acoustic image in the reconstruction sound field calculated in step 4, sets up direction equivalence relation as follows:
pv=pv 0
Namely
Σ i = 1 3 G e - ik | L i | | L i | 2 ( 1 + 1 ik | L i | ) × - L ix - L iy - L iz q ( ω ) = G e - ik | ϵ | ϵ | ϵ | ( 1 + 1 ik | ϵ | ) × 1 | ϵ | - ϵ x - ϵ x - ϵ z s ( ω )
Wherein k represents wave number, relevant with the velocity of sound with the frequency of voice signal, and e is constant, and i is imaginary unit, and ω is the angular frequency of signal s (t), | L i| represent loud speaker and listen the distance between the point of articulation, | ε | represent sound source and listen the distance between the point of articulation, G (ω) represents source strength.
Step 5.2: according to the acoustic pressure p of the original sound source that step 1 calculates 0acoustic pressure p with acoustic image in the reconstruction sound field calculated in step 4, sets up distance equivalence relation as follows:
p=p 0
Namely
G Σ j = 1 n e - ik | L i | | L i | q ( ω ) = G e - ik | ϵ | | ϵ | s ( ω )
Wherein k represents wave number, relevant with the velocity of sound with the frequency of voice signal, and e is constant, and i is imaginary unit, and ω is the angular frequency of signal s (t), | L i| represent loud speaker and listen the distance between the point of articulation, | ε | represent sound source and listen the distance between the point of articulation, G (ω) represents source strength.
Step 5.3: according to step 5.1 and 5.2 direction equivalence relation and distance equivalence relation, set up direction, distance equivalence model:
pv = pv 0 p = p 0
Step 6: in the direction of step 5, distance equivalence model, with the weight w arranged in step 3 1, w 2, w 3, w 4for unknown quantity, solve the value of direction, distance equivalence model acquisition weight; Model is launched to obtain system of linear equations, separates this system of linear equations, can utilize but be not limited to the tool software such as Matlab, Mathematical, or other instrument, and manually calculate, obtain this solution of equations (w 1, w 2, w 3, w 4).
Step 7: utilize the weight w solved in step 6 1, w 2, w 3, w 4, the signal carrying out loud speaker distributes, and makes the audio direction of reconstruction and distance with the direction of acoustic source with apart from consistent, in this example, all loud speakers are positioned on same sphere, namely each loud speaker is equal with listening the distance between the point of articulation, therefore in signal assigning process without the need to postponing, loud speaker L 1, L 2, L 3, L 4distribute signal as shown in the formula:
q 1 ( t ) = w 1 s ( t ) q 2 ( t ) = w 2 s ( t ) q 3 ( t ) = w 3 s ( t ) q 4 ( t ) = w 4 s ( t )
Ask for an interview Fig. 1, the technical scheme that device of the present invention adopts is: the device recovering Sounnd source direction and range information in a kind of multi-channel system, it is characterized in that, comprising: point of articulation directional information computing module 1 is listened in sound source space, sound source space listens point of articulation range information computing module 2, loud speaker to select module 3, signal forward allocator module 4, rebuild sound field Sounnd source direction information computational module 5, rebuild sound field sound source range information computing module 6, model building module 7, model solution module 8, signal distribution module 9;
The direction of the sound source that described sound source space listens point of articulation directional information computing module 1 to perceive for the pleasant to the ear point of articulation in sound source space calculating acoustic source place, signal s (t) according to sound source calculates the particle rapidity pv listening the point of articulation 0, and by pv 0export model building module 7 to;
The distance of the sound source that described sound source space listens point of articulation range information computing module 2 to perceive for the pleasant to the ear point of articulation in sound source space calculating acoustic source place, signal s (t) according to sound source calculates the acoustic pressure p listening the point of articulation 0, and by p 0export model building module 7 to;
Described loud speaker select module 3 for set up reconstruction multi-channel system in the selection principle of loud speaker, determine the loud speaker L selected in multi-channel system 1, L 2, L 3, L 4, and export selected loud speaker to signal forward allocator module 4;
Described signal forward allocator module 4 is for being multiplied by weight w by original audio signal s (t) 1, w 2, w 3, w 4after distribute to the loud speaker L that loud speaker selects to select in module 3 1, L 2, L 3, L 4, then the signal after distribution is exported to and rebuilds sound field Sounnd source direction information computational module 5 and rebuild sound field sound source range information computing module 6;
Described reconstruction sound field acoustic image directional information computing module 5 receives the direction of acoustic image for calculating the pleasant to the ear point of articulation place of multi-channel system, according in signal forward allocator module 4 to the preallocated signal of loud speaker, the pleasant to the ear point of articulation place of multi-channel system utilizing acoustic theory to calculate reconstruction receives the particle rapidity pv of acoustic image, and exports pv to model building module 7;
Described reconstruction sound field audio-visual distance information computing module 6 receives the distance of acoustic image for calculating the pleasant to the ear point of articulation place of multi-channel system, according in signal forward allocator module 4 to the preallocated signal of loud speaker, the pleasant to the ear point of articulation place of multi-channel system utilizing acoustic theory to calculate reconstruction receives the acoustic pressure p of acoustic image, and exports p to model building module 7;
Described model building module 7 for set up listen point of articulation prescription to perceived distance consistency model, the sound source particle rapidity pv listening point of articulation directional information computing module 1 to export by sound source space 0direction equivalence relation is set up, the sound source acoustic pressure p listening point of articulation range information computing module 2 to export by sound source space with the acoustic image particle rapidity pv rebuilding output in sound field Sounnd source direction information computational module 4 0set up distance equivalence relation with the acoustic image acoustic pressure p rebuilding output in sound field Sounnd source direction information computational module 6, set up direction, distance equivalence model according to direction equivalence relation and distance equivalence relation, and export this model to model solution module 8;
Described model solution module 8 is set up the model of foundation in module 7 for solving model and then is obtained four weight w 1, w 2, w 3, w 4value, finally export weights to signal distribution module 9;
The weight w that described signal distribution module 9 solves for utilizing model solution module 8 1, w 2, w 3, w 4, the signal carrying out loud speaker distributes, and makes the audio direction of reconstruction and distance with the direction of acoustic source with apart from consistent.
These are only preferred embodiment of the present invention, be not intended to limit protection scope of the present invention, therefore, all any amendments done within the spirit and principles in the present invention, equivalent replacement, improvement etc., all should be included within protection scope of the present invention.

Claims (5)

1. the method that in multi-channel system, Sounnd source direction and distance are rebuild, is characterized in that, comprise the following steps:
Step 1: according to time-domain signal s (t) of known sound source, to listen the point of articulation for after initial point sets up Descartes's rectangular coordinate system in original sound field, the particle rapidity pv of the sound source of listening point of articulation place to receive that to calculate with sound source distance be r 0with acoustic pressure p 0;
Step 2: according to the particle rapidity pv of the sound source calculated in step 1 0, select to comprise particle rapidity pv 04 loud speaker L in direction 1, L 2, L 3, L 4;
Step 3: by the loud speaker L in sound-source signal s (t) and step 2 1, L 2, L 3, L 4, s (t) is multiplied by respectively 4 weight w 1, w 2, w 3, w 4after be assigned on four selected loud speakers;
Step 4: at playback end, calculates and rebuilds loud speaker L in sound field 1, L 2, L 3, L 4particle rapidity pv and the acoustic pressure p of the acoustic image of point of articulation place perception is listened after sending signal;
Step 5: according to the particle rapidity pv of the original sound source that step 1 calculates 0with acoustic pressure p 0, the particle rapidity pv of acoustic image and acoustic pressure p in the reconstruction sound field that calculates in step 4, set up direction, distance equivalence model; Its specific implementation comprises following sub-step:
Step 5.1: according to the particle rapidity p of the original sound source that step 1 calculates v0particle rapidity pv with acoustic image in the reconstruction sound field calculated in step 4, sets up direction equivalence relation as follows:
pv=pv 0
Namely
Σ i = 1 3 G e - ik | r - L i | | r - L i | 2 ( 1 + 1 ik | r - L i | ) × r ix - L ix r iy - L iy r iz - L iz q ( ω ) = G e - ik | r - ϵ | | r - ϵ | ( 1 + 1 ik | r - ϵ | ) × 1 | r - ϵ | r x - ϵ x r y - ϵ x r z - ϵ x s ( ω )
Wherein k represents wave number, relevant with the velocity of sound with the frequency of voice signal, and e is constant, and i is imaginary unit, and ω is the angular frequency of signal s (t), | r-L i| represent loud speaker and listen the distance between the point of articulation, | r-ε | represent sound source and listen the distance between the point of articulation, G (ω) represents source strength;
Step 5.2: according to the acoustic pressure p of the original sound source that step 1 calculates 0acoustic pressure p with acoustic image in the reconstruction sound field calculated in step 4, sets up distance equivalence relation as follows:
p=p 0
Namely
G Σ j = 1 n e - ik | r - L i | | r - L i | q ( ω ) = G e - ik | r - ϵ | | r - ϵ | s ( ω )
Wherein k represents wave number, relevant with the velocity of sound with the frequency of voice signal, and e is constant, and i is imaginary unit, and ω is the angular frequency of signal s (t), | r-L i| represent loud speaker and listen the distance between the point of articulation, | r-ε | represent sound source and listen the distance between the point of articulation, G (ω) represents source strength;
Step 5.3: according to step 5.1 and 5.2 direction equivalence relation and distance equivalence relation, set up direction, distance equivalence model:
pv = pv 0 p = p 0 ;
Step 6: in the direction of step 5, apart from equivalence model, with the weight w arranged in step 3 1, w 2, w 3, w 4for unknown quantity, solve the value of direction, distance equivalence model acquisition weight;
Step 7: utilize the weight w solved in step 6 1, w 2, w 3, w 4, the signal carrying out loud speaker distributes, and makes the audio direction of reconstruction and distance with the direction of acoustic source with apart from consistent.
2. the method that in multi-channel system according to claim 1, Sounnd source direction and distance are rebuild, is characterized in that: the calculating described in step 1 and sound source distance are the particle rapidity pv of the sound source of listening point of articulation place to receive of r 0with acoustic pressure p 0, its specific implementation comprises following sub-step:
Step 1.1: utilize Fourier transform that time-domain signal s (t) of sound source is transformed into frequency domain by time domain, obtains frequency-region signal s (ω)
s ( ω ) = ∫ - ∞ + ∞ s ( t ) e - iwt dt
Step 1.2: by frequency-region signal s (ω) and the sound source position vector ε in sound field=(ε x, ε y, ε z), listen point of articulation position coordinates r=r (r x, r y, r z), according to the definition of sound physical properties attribute particle rapidity, calculate the particle rapidity pv listening point of articulation place to receive 0:
pv 0 ( r , ω ) = G e - ik | r - ϵ | | r - ϵ | ( 1 + 1 ik | r - ϵ | ) × 1 | r - ϵ | r x - ϵ x r y - ϵ y r z - ϵ z s ( ω ) ;
Calculate the acoustic pressure p listening point of articulation place to receive 0:
p 0 ( T 0 , ω ) = G e - ik | r - ϵ | | r - ϵ | s ( ω ) ;
Wherein k represents wave number, relevant with the velocity of sound with the frequency of voice signal, and e is constant, and i is imaginary unit, and ω is the angular frequency of signal s (t), | r-ε | represent sound source and listen the distance between the point of articulation, G (ω) represents source strength.
3. the method that in multi-channel system according to claim 1, Sounnd source direction and distance are rebuild, is characterized in that: the loud speaker L of 4 described in step 2 1, L 2, L 3, L 4, meet following condition: point of articulation r (x, y, z) and L will be listened respectively 1, L 2, L 3, L 4connect, in the polyhedron formed, the particle rapidity pv of sound source 0unit vector to be positioned at this polyhedron inner.
4. the method that in multi-channel system according to claim 1, Sounnd source direction and distance are rebuild, is characterized in that: loud speaker L in sound field is rebuild in the calculating described in step 4 1, L 2, L 3, L 4listen particle rapidity pv and the acoustic pressure p of the acoustic image of point of articulation place perception after sending signal, its specific implementation comprises following sub-step:
Step 4.1: at playback end, with the center of the ambiophonic system of multichannel composition for initial point, sets up cartesian coordinate system, loud speaker L icoordinate be designated as L i(L ix, L iy, L iz), listen the coordinate of the point of articulation to be r (r x, r y, r z);
Step 4.2: by loud speaker L 1, L 2, L 3, L 4the signal q sent 1(t), q 2(t), q 3(t), q 4t () is converted to frequency domain:
q ( ω ) = q 1 ( ω ) q 2 ( ω ) q 3 ( ω ) q 4 ( ω ) , q i ( ω ) = ∫ - ∞ + ∞ q i ( t ) e - iwt dt
Step 4.3: by loud speaker L 1, L 2, L 3, L 4send signal q 1(t), q 2(t), q 3(t), q 4t () calculates the frequency domain representation pv and the acoustic pressure p that listen the particle rapidity at point of articulation place:
p v ( r , ω ) = Σ i = 1 3 G e - ik | r - L i | | r - L i | ( 1 + 1 ik | r - L i | ) × 1 | r - L i | r x - L ix r y - L iy r z - L iz q ( ω ) = Σ i = 1 3 G e - ik | r - L i | | r - L i | 2 ( 1 + 1 ik | r - L i | ) × r x - L ix r y - L iy r z - L iz q ( ω )
p = G Σ j = 1 n e - ik | r - L i | | r - L i | q ( ω )
Wherein k represents wave number, relevant with the velocity of sound with the frequency of voice signal, and e is constant, and i is imaginary unit, and ω is the angular frequency of signal s (t), | r-L i| represent loud speaker and listen the distance between the point of articulation, G (ω) represents source strength.
5. one kind utilizes the method that in the multi-channel system described in claim 1, Sounnd source direction and distance are rebuild to carry out the device that in multi-channel system, Sounnd source direction and distance are rebuild, it is characterized in that, comprise: point of articulation directional information computing module (1) is listened in sound source space, point of articulation range information computing module (2) is listened in sound source space, module (3) selected by loud speaker, signal forward allocator module (4), rebuild sound field Sounnd source direction information computational module (5), rebuild sound field sound source range information computing module (6), model building module (7), model solution module (8), signal distribution module (9),
The direction of the sound source that described sound source space listens point of articulation directional information computing module (1) to perceive for the pleasant to the ear point of articulation in sound source space calculating acoustic source place, signal s (t) according to sound source calculates the particle rapidity pv listening the point of articulation 0, and by pv 0export model building module (7) to;
The distance of the sound source that described sound source space listens point of articulation range information computing module (2) to perceive for the pleasant to the ear point of articulation in sound source space calculating acoustic source place, signal s (t) according to sound source calculates the acoustic pressure p listening the point of articulation 0, and by p 0export model building module (7) to;
Described loud speaker select module (3) for set up reconstruction multi-channel system in the selection principle of loud speaker, determine the loud speaker L selected in multi-channel system 1, L 2, L 3, L 4, and export selected loud speaker to signal forward allocator module (4);
Described signal forward allocator module (4) is for being multiplied by weight w by original audio signal s (t) 1, w 2, w 3, w 4after distribute to the loud speaker L that loud speaker selects to select in module (3) 1, L 2, L 3, L 4, then the signal after distribution is exported to and rebuilds sound field Sounnd source direction information computational module (5) and rebuild sound field sound source range information computing module (6);
Described reconstruction sound field acoustic image directional information computing module (5) receives the direction of acoustic image for calculating the pleasant to the ear point of articulation place of multi-channel system, according in signal forward allocator module (4) to the preallocated signal of loud speaker, the pleasant to the ear point of articulation place of multi-channel system utilizing acoustic theory to calculate reconstruction receives the particle rapidity pv of acoustic image, and pv is exported to model building module (7);
Described reconstruction sound field audio-visual distance information computing module (6) receives the distance of acoustic image for calculating the pleasant to the ear point of articulation place of multi-channel system, according in signal forward allocator module (4) to the preallocated signal of loud speaker, the pleasant to the ear point of articulation place of multi-channel system utilizing acoustic theory to calculate reconstruction receives the acoustic pressure p of acoustic image, and p is exported to model building module (7);
Described model building module (7) for set up listen point of articulation prescription to perceived distance consistency model, the sound source particle rapidity p listening point of articulation directional information computing module (1) to export by sound source space v0direction equivalence relation is set up, the sound source acoustic pressure p listening point of articulation range information computing module (2) to export by sound source space with the acoustic image particle rapidity pv rebuilding output in sound field Sounnd source direction information computational module (4) 0distance equivalence relation is set up with the acoustic image acoustic pressure p rebuilding output in sound field Sounnd source direction information computational module (6), set up direction, distance equivalence model according to direction equivalence relation and distance equivalence relation, and export this model to model solution module (8);
Set up direction, distance equivalence model according to direction equivalence relation and distance equivalence relation, its specific implementation comprises following sub-step: step 5.1: according to the particle rapidity p of the original sound source that step 1 calculates v0particle rapidity pv with acoustic image in the reconstruction sound field calculated in step 4, sets up direction equivalence relation as follows:
pv=pv 0
Namely
Σ i = 1 3 G e - ik | r - L i | | r - L i | 2 ( 1 + 1 ik | r - L i | ) × r ix - L ix r iy - L iy r iz - L iz q ( ω ) = G e - ik | r - ϵ | | r - ϵ | ( 1 + 1 ik | r - ϵ | ) × 1 | r - ϵ | r x - ϵ x r y - ϵ x r z - ϵ x s ( ω )
Wherein k represents wave number, relevant with the velocity of sound with the frequency of voice signal, and e is constant, and i is imaginary unit, and ω is the angular frequency of signal s (t), | r-L i| represent loud speaker and listen the distance between the point of articulation, | r-ε | represent sound source and listen the distance between the point of articulation, G (ω) represents source strength;
Step 5.2: according to the acoustic pressure p of the original sound source that step 1 calculates 0acoustic pressure p with acoustic image in the reconstruction sound field calculated in step 4, sets up distance equivalence relation as follows:
p=p 0
Namely
G Σ j = 1 n e - ik | r - L i | | r - L i | q ( ω ) = G e - ik | r - ϵ | | r - ϵ | s ( ω )
Wherein k represents wave number, relevant with the velocity of sound with the frequency of voice signal, and e is constant, and i is imaginary unit, and ω is the angular frequency of signal s (t), | r-L i| represent loud speaker and listen the distance between the point of articulation, | r-ε | represent sound source and listen the distance between the point of articulation, G (ω) represents source strength;
Step 5.3: according to step 5.1 and 5.2 direction equivalence relation and distance equivalence relation, set up direction, distance equivalence model:
pv = p v 0 p = p 0 ;
Described model solution module (8) is set up the model of foundation in module (7) for solving model and then is obtained four weight w 1, w 2, w 3, w 4value, finally weights are exported to signal distribution module (9);
The weight w that described signal distribution module (9) solves for utilizing model solution module (8) 1, w 2, w 3, w 4, the signal carrying out loud speaker distributes, and makes the audio direction of reconstruction and distance with the direction of acoustic source with apart from consistent.
CN201410071545.1A 2014-02-28 2014-02-28 Method and device for rebuilding sound source direction and distance in multichannel system Expired - Fee Related CN103826194B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201410071545.1A CN103826194B (en) 2014-02-28 2014-02-28 Method and device for rebuilding sound source direction and distance in multichannel system

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201410071545.1A CN103826194B (en) 2014-02-28 2014-02-28 Method and device for rebuilding sound source direction and distance in multichannel system

Publications (2)

Publication Number Publication Date
CN103826194A CN103826194A (en) 2014-05-28
CN103826194B true CN103826194B (en) 2015-06-03

Family

ID=50760977

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201410071545.1A Expired - Fee Related CN103826194B (en) 2014-02-28 2014-02-28 Method and device for rebuilding sound source direction and distance in multichannel system

Country Status (1)

Country Link
CN (1) CN103826194B (en)

Families Citing this family (11)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN104363555A (en) * 2014-09-30 2015-02-18 武汉大学深圳研究院 Method and device for reconstructing directions of 5.1 multi-channel sound sources
CN106205573B (en) * 2016-06-28 2019-09-17 青岛海信移动通信技术股份有限公司 A kind of audio data processing method and device
CN106017837B (en) * 2016-06-30 2018-12-21 北京空间飞行器总体设计部 A kind of analogy method of equivalent sound simulation source
CN106454685B (en) * 2016-11-25 2018-03-27 武汉大学 A kind of sound field rebuilding method and system
CN109302660B (en) * 2017-07-24 2020-04-14 华为技术有限公司 Audio signal compensation method, device and system
CN109068262B (en) * 2018-08-03 2019-11-08 武汉大学 A kind of acoustic image personalization replay method and device based on loudspeaker
CN109618276B (en) * 2018-11-23 2020-08-07 武汉轻工大学 Sound field reconstruction method, device, storage medium and device based on non-central point
CN110366091B (en) * 2019-08-07 2021-11-02 武汉轻工大学 Sound field reconstruction method and device based on sound pressure, storage medium and device
CN111464932A (en) * 2020-04-07 2020-07-28 武汉轻工大学 Sound field reconstruction method, device and equipment based on multiple listening points and storage medium
CN112073804B (en) * 2020-09-10 2022-05-20 深圳创维-Rgb电子有限公司 Television sound adjusting method, television and storage medium
CN113286252B (en) * 2021-07-23 2021-11-16 科大讯飞(苏州)科技有限公司 Sound field reconstruction method, device, equipment and storage medium

Family Cites Families (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP3900208B2 (en) * 1997-02-06 2007-04-04 ソニー株式会社 Sound reproduction system and audio signal processing apparatus
JP5314129B2 (en) * 2009-03-31 2013-10-16 パナソニック株式会社 Sound reproducing apparatus and sound reproducing method
CN103021414B (en) * 2012-12-04 2014-12-17 武汉大学 Method for distance modulation of three-dimensional audio system
CN103037301B (en) * 2012-12-19 2014-11-05 武汉大学 Convenient adjustment method for restoring range information of acoustic images
CN103347245B (en) * 2013-07-01 2015-03-25 武汉大学 Method and device for restoring sound source azimuth information in stereophonic sound system

Also Published As

Publication number Publication date
CN103826194A (en) 2014-05-28

Similar Documents

Publication Publication Date Title
CN103826194B (en) Method and device for rebuilding sound source direction and distance in multichannel system
CN106454685B (en) A kind of sound field rebuilding method and system
CN104363555A (en) Method and device for reconstructing directions of 5.1 multi-channel sound sources
CN105264914B (en) Audio playback device and method therefor
CN104604255A (en) Virtual rendering of object-based audio
CN103888889B (en) A kind of multichannel conversion method based on spheric harmonic expansion
CN105120418B (en) Double-sound-channel 3D audio generation device and method
CN102932730B (en) Method and system for enhancing sound field effect of loudspeaker group in regular tetrahedron structure
CN103021414B (en) Method for distance modulation of three-dimensional audio system
CN101889307A (en) Phase-amplitude 3-D stereo encoder and demoder
CN105308988A (en) Audio decoder configured to convert audio input channels for headphone listening
CN105392102A (en) Three-dimensional audio signal generation method and system for non-spherical speaker array
JP2009077379A (en) Stereoscopic sound reproduction equipment, stereophonic sound reproduction method, and computer program
CN106303843B (en) A kind of 2.5D playback methods of multizone different phonetic sound source
US20190394596A1 (en) Transaural synthesis method for sound spatialization
CN102883246A (en) Simplifying and laying method for loudspeaker groups of three-dimensional multi-channel audio system
CN106465027A (en) Apparatus and method for edge fading amplitude panning
US9066173B2 (en) Method for producing optimum sound field of loudspeaker
CN103347245B (en) Method and device for restoring sound source azimuth information in stereophonic sound system
EP3530006B1 (en) Apparatus and method for weighting stereo audio signals
CN103402158B (en) Dimensional sound extension method for handheld playing device
CN103052018A (en) Audio-visual distance information recovery method
CN109036456B (en) Method for extracting source component environment component for stereo
Wilkinson AD-3D: HRTF based 3D Audio Designer
CN105120406B (en) Three-dimensional audio compressing method and system

Legal Events

Date Code Title Description
C06 Publication
PB01 Publication
C10 Entry into substantive examination
SE01 Entry into force of request for substantive examination
C14 Grant of patent or utility model
GR01 Patent grant
CF01 Termination of patent right due to non-payment of annual fee
CF01 Termination of patent right due to non-payment of annual fee

Granted publication date: 20150603

Termination date: 20200228