CN105206281B - Sound enhancement method based on distributed microphone array network - Google Patents
Sound enhancement method based on distributed microphone array network Download PDFInfo
- Publication number
- CN105206281B CN105206281B CN201510582363.5A CN201510582363A CN105206281B CN 105206281 B CN105206281 B CN 105206281B CN 201510582363 A CN201510582363 A CN 201510582363A CN 105206281 B CN105206281 B CN 105206281B
- Authority
- CN
- China
- Prior art keywords
- node
- network
- microphone array
- multichannel
- signal
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Active
Links
Abstract
The invention discloses a kind of sound enhancement methods based on distributed microphone array network, including the following steps: establishes the distributed microphone array network based on Ad-hoc network;It is synchronous that sample rate is carried out to network node;The signal of each node is subjected to framing;Speech enhan-cement is carried out using multichannel Wiener filter in each node;By the every other node of transmitting voice signal after enhancing to network;In each node, simultaneously according to voice signal after the enhancing of the single channel of the multichannel microphone array observation signal of present node and every other node, speech enhan-cement is carried out using multichannel Wiener filter again, obtains voice signal after the updated single channel enhancing of present node.By isolated microphone array, network interconnects the present invention by wireless communication, forms microphone array network, is conducive to the speech enhan-cement effect for improving individual node.
Description
Technical field
The present invention relates to sound enhancement methods, and in particular to a kind of speech enhan-cement based on distributed microphone array network
Method.
Background technique
Various noises are frequently accompanied by environment locating for us, for example, television set sound and fan sound, automobile in room
Interior engine sound, the running car sound on road, Babble noise in coffee shop etc..Noise is to multiple voice processing system
It has a negative impact.For example, noise can interfere the sound for even covering other side in voice communication, speech quality is reduced;In language
In sound identifying system, noise can make phonetic recognization rate decline, or even keep identifying system entirely ineffective.Therefore, according to observing
Noisy Speech Signal, estimation clean speech have a very important significance, we are referred to as speech enhan-cement.
Traditional voice enhancing algorithm is handled using the observation signal of a microphone, including single channel Wiener filtering
Device, spectrum-subtraction, the maximum likelihood based on statistical model and sound enhancement method of maximum a posteriori probability etc..Although such method can
To eliminate noise to a certain extent, but deposit problem both ways.It will cause lacking for phonetic element while firstly, eliminating noise
It loses, i.e. generation speech distortion.Secondly, noise eliminate after frequency spectrum on be commonly present the extreme point of Random Discrete, experience hearer
" music noise ".The factor of these two aspects all makes the enhanced intelligibility of speech be difficult to reach expected, and makes speech recognition
Performance cannot effectively improve.
To solve the above-mentioned problems, people start with two or more microphones, form " microphone array ", to visit
Seek more preferably multi-channel speech enhancement method.Microphone in microphone array is in space different location, but clock and sampling
Rate is consistent.Therefore, multiple microphones provide the temporal redundancy and Spatial Difference of voice and noise, and more information make
The raising of speech enhan-cement performance is provided with possibility.In order to enhance voice, people can design the space filter of referred to as " Beam-former "
Wave device extracts the signal in target sound source direction, inhibits the noise in other directions.Simplest Beam-former is " delay adduction "
Beam-former, and MVDR and LCMV Beam-former theoretically can avoid speech distortion while reducing noise.In addition to
Except simple Beam-former, generalized sidelobe is eliminated (GSC) framework and is also widely used.Although can theoretically prove
The equivalence of GSC and LCMV Beam-former, but the realization of GSC is more simple, computation complexity is relatively low.Above-mentioned wave beam shape
Grow up to be a useful person be required to voice directions (even noise orientation) it is known that but sound bearing often and is not fixed under actual conditions, and make an uproar
Sound bearing under sound and reverberation is difficult to estimate.In order to avoid auditory localization, single channel Wiener filter is generalized to multichannel,
So that optimal multichannel Wiener filter can be designed according only to noise space time statistical properties, and noise space time statistical properties can
In conjunction with voice existing probability or voice activity detection algorithms estimation and update.It is compared with single-channel algorithm, even twin-channel
Sound enhancement method can obtain being obviously improved for performance.
Speech enhan-cement is carried out using microphone array and is increasingly becoming mainstream.Once microphone array hardware completes,
Microphone spacing, the parameters such as included number of microphone are difficult to change.Due to the limitation in the spaces such as handheld device, microphone array
More microphone and biggish spacing cannot be used.When microphone array is only in a lesser spatial dimension, it is difficult to
Accurate comprehensive acquisition is carried out to ambient noise and reverberation.And theoretically more microphones and bigger microphone spacing can be with
Effectively improve the performance of multicenter voice enhancing algorithm.Therefore, it is traditional based on the voice enhancement algorithm of microphone array by
The limitation of microphone array itself scalability and space.
Summary of the invention
In view of the deficiencies of the prior art, the invention discloses a kind of speech enhan-cements based on distributed microphone array network
Method.
Technical scheme is as follows:
A kind of sound enhancement method based on distributed microphone array network, including the following steps:
Step a, the distributed microphone array network based on Ad-hoc network being made of multiple microphone arrays is established;
It can be in communication with each other between any two network node;
Step b, by distributed microphone array netinit, i.e., it is synchronous sample rate to be carried out to network node;
Step c, the signal of each node is subjected to framing, the multinode multichannel microphone array after obtaining framing observes letter
Number;
Step d, in each node, for the multichannel microphone array observation signal of each frame, according to the more of present node
Road microphone array observation signal carries out speech enhan-cement using multichannel Wiener filter, and voice is believed after obtaining single channel enhancing
Number;
Step e, in each node, transmitting voice signal after which is enhanced by the single channel that the step d is obtained
To the every other node of network;
Step f, in each node, while according to the multichannel microphone array observation signal and every other section of present node
Voice signal after the single channel enhancing of point, carries out speech enhan-cement using multichannel Wiener filter again, obtains present node more
Voice signal after single channel enhancing after new;
Step g, iteration step e~step f, when voice signal is restrained after the single channel enhancing that certain node obtains,
Voice signal no longer updates after the single channel enhancing of present node;When all nodes single channel enhancing after voice signal no longer
When update, processing terminate for present frame;The finally voice signal after each node obtains present node enhancing.
Its further technical solution are as follows: the microphone array includes audio collection module and communication module.
Its further technical solution are as follows: the structure of the Ad-hoc network in the step a is planar structure or classification
Structure;Ad-hoc network using priori formula, reaction equation or hybrid-type Routing Protocol realize in network two node devices it
Between be in communication with each other.
Its further technical solution are as follows: the step b further includes carrying out time synchronization to network node;
The distribution microphone array includes network equipment clock;When the time synchronization is by the network equipment
Clock is synchronized based on ntp network time protocol.
Its further technical solution is, the step b specifically includes the following steps:
Step b1, network samples rate is initialized, makes K=1, i.e. network samples rate f0Equal to the equipment sample rate of node 1
f1;
Step b2, the equipment sample rate of node K is fK;By the equipment sample rate f of node KKIt is transferred to node K+1;
If step b3, the equipment sample rate f of node K+1K+1> fK, then f0=fK, otherwise f0=fK+1;
Step b4, K=K+1;
Step b5, step b2~step b4 is repeated, until all nodes are traversed, thus network samples rate f0For whole network institute
There is the equipment sample rate minimum value of node;
Step b6, pass through finish node for current network sample rate f0Other each nodes are transferred to, so that all nodes
Equipment sample rate is f0。
Its further technical solution are as follows: the signal framing in the step c inhibits frequency using hamming window or Hanning window
Spectrum leakage;The step c uses the framing strategy of time aliasing.
Its further technical solution are as follows: the step d uses the multi-pass of time domain multichannel Wiener filter or frequency domain
Road Wiener filter is filtered multichannel microphone array observation signal, to achieve the effect that speech enhan-cement:
In node K, the expression formula of the time domain multichannel Wiener filter are as follows:
hW, K(t)=[RXx, K(t)+λRNn, K(t)]-1RXx, K(t)u;
In above formula, RXx, K(t)=RYy, K(t)-RNn, K(t);
It is the clean speech vector x of present nodeK(t)=[x1,K(t),
x2,K(t),…,xM,K(t)]TTime domain autocorrelation matrix;
It is the noise vector n of present nodeK(t)=[n1,K(t),n2,K(t),...,
nM,K(t)]TTime domain autocorrelation matrix;
It is the multichannel microphone array observation signal vector y of present nodeK(t)=
[y1,K(t),y2,K(t),...,yM,K(t)]TTime domain autocorrelation matrix;
U=[1,0 ..., 0]T, the length is M;
M is the number of microphone of present node;
λ is the degree for controlling noise elimination and speech distortion, and λ > 0, λ is bigger, and the repressed effect of noise is more obvious, together
When bring more speech distortions;
The time-domain filtering of node K exports are as follows:
In node K, the expression formula of the frequency-domain multi-channel Wiener filter are as follows:
HW, K(ω)=[RXX, K(ω)+λRNN, K(ω)]-1RXX, K(ω)u;
In above formula, RXX, K(ω)=RYY, K(ω)-RNN, K(ω);
It is the clean speech vector X of present nodeK(ω)=[X1, K
(ω), X2, K(ω) ..., XM, K(ω)]HFrequency domain autocorrelation matrix;
It is the noise vector N of present nodeK(ω)=[N1,K(ω),N2,K
(ω),...,NM,K(ω)]HFrequency domain autocorrelation matrix;
It is the multichannel microphone array observation signal vector Y of present nodeK(ω)
=[Y1,K(ω),Y2,K(ω),...,YM,K(ω)]HFrequency domain autocorrelation matrix;
U=[1,0 ..., 0]T, the length is M;
M is the number of microphone of present node;
λ is the degree for controlling noise elimination and speech distortion, and λ > 0, λ is bigger, and the repressed effect of noise is more obvious, together
When bring more speech distortions;
The frequency domain filtering of node K exports are as follows:
Its further technical solution are as follows: the step e includes that transmitting node sequence is added in the data packet of signal transmission
Number, the information of receiving node serial number and multichannel Wiener filter number of processes.
Its further technical solution are as follows: the step f includes the multichannel Wiener filter pair using time domain or frequency domain
Signal is filtered after the enhancing of present node multichannel observation signal and other nodes;
In the multichannel Wiener filter of the time domain,
The joint that signal is constituted after the enhancing of present node K multichannel microphone array observation signal and every other node
Vector are as follows:
In above formula,For except
Vector composed by the enhanced time domain single-channel voice of other outer nodes of node K;
NiFor the number of iterations of step g;
It isIn clean speech ingredient;
It isIn noise contribution;
For clean speech ingredient in present nodeTime domain from phase
Close matrix;
For noise vector in present nodeTime domain autocorrelation matrix;
To combine vector in present nodeTime domain autocorrelation matrix;
U=[1,0 ..., 0]T, it is the node total number in network the length is M+P-1, P;
Then node K NiThe time domain multichannel Wiener filter of+1 iteration are as follows:
In the multichannel Wiener filter of the frequency domain,
The joint vector that signal is constituted after the enhancing of present node K multichannel observation signal and every other node are as follows:
In above formula,
For vector composed by the enhanced frequency domain single-channel voice of other nodes in addition to node K;
NiFor the number of iterations of step g;
ForIn clean speech ingredient;
ForIn noise contribution;
For remove node K except other nodes it is pure
The frequency domain autocorrelation matrix of speech vector;
To remove other node background noise vectors except node K
Frequency domain autocorrelation matrix;
To remove the frequency domain of other node observation vectors except node K certainly
Correlation matrix;
U=[1,0 ..., 0]T;The length is M+P-1, P is the node total number in network, then node K Ni+ 1 iteration
Frequency-domain multi-channel Wiener filter are as follows:
Its further technical solution are as follows: the step g include according to filtering front and back signal vector difference norm and
Signal energy judges the step of whether voice signal restrains after the single channel enhancing that node obtains, and method is as follows:
It is previous to filter obtained single channel time domain signal vector in node K are as follows:
This filters obtained single channel time domain signal vector are as follows:
WhenWhen, present filter output convergence;
In above formula, | | | |pP norm is represented, η is threshold value.
The method have the benefit that:
First, the invention proposes a kind of completely new frames that speech enhan-cement is carried out based on microphone array.With tradition side
Method is different, and by isolated microphone array, network interconnects the present invention by wireless communication, forms microphone array network.
Second, each of microphone network node can directly or indirectly utilize whole Mikes in network
Wind breaches the space limitation pole of each equipment, greatly extends the spatial observation range of individual node, be conducive to improve single
The speech enhan-cement effect of node.For single-channel, after it is linked into microphone array network, it can reach more
The speech enhan-cement effect in channel.
Third, microphone array network is to the microphone number in the relative position of network node quantity, each node, node
Amount and spatial position are not done any it is assumed that having great scalability and freedom degree.
4th, by way of Ad-hoc networking, so that network needs not rely on central node, it can be completed distributed
It calculates, improves the tolerance of network.
5th, each node of microphone array network obtains local optimal filter output simultaneously, this is each in network
Node provides the user experience of differentiation.
Detailed description of the invention
Fig. 1 is flow chart of the invention.
Fig. 2 is the distributed microphone array network diagram based on Ad-hoc network.
Fig. 3 is the synchronous flow chart of distributed microphone array network samples rate.
Fig. 4 is the flow chart of the single node speech enhan-cement based on multichannel Wiener filter.
Fig. 5 is the flow chart of the iterative speech enhan-cement of multinode based on multichannel Wiener filter.
Specific embodiment
Fig. 1 is flow chart of the invention.
Core content in the present invention mainly includes three parts: (1), the foundation of Ad-hoc network shown in step a and
The initialization of audio collection module shown in step b;(2), the single node voice shown in step d based on multichannel Wiener filter increases
By force;(3) the iterative speech enhan-cement of multinode shown in step f based on multichannel Wiener filter.
As shown in Figure 1, the present invention specifically includes the following steps:
(1), the foundation and initialization of Ad-hoc network
Step a, multiple microphone arrays are set, point based on Ad-hoc network being made of multiple microphone arrays is established
Cloth microphone array network;It can be in communication with each other between any two network node.
Ad-hoc network is otherwise known as interim self-organizing network.Since the network is not necessarily to additional infrastructure network, base
In construction and extension, therefore by the present invention for constructing distributed microphone array network.
Fig. 2 is the distributed microphone array network diagram based on Ad-hoc network.In microphone array network, net
Network node is each microphone array.The microphone array equipment of each node includes at least one microphone.Each node
Microphone array equipment further includes audio collection module, communication module and computing module.It is connected with each other between modules.Its
In, audio collection module is responsible for acquiring the sound in present node local environment, and communication module is responsible for and other nodes communicate mould
Data transmission between block, the speech enhan-cement that computing module is responsible for this node calculate.
Hierarchical structure or plane formula network structure can be used in the structure of Ad-hoc network.In hierarchical structure, multiple nets
Network node is divided into different " cluster ", and each node in cluster selects cluster head by certain election algorithm, cluster head safeguard this cluster and
Routing iinformation between cluster head is realized any in network jointly by the communication between communication, cluster head and the cluster interior nodes between cluster head
Communication between two nodes.In plane formula network structure, each node status equity, respective independent maintenance owns to other
The routing iinformation of node.In general, hierarchical structure is used when network node is more, and when network node is less, it uses
Plane formula network structure.
As shown in Fig. 2, the present embodiment only includes three network nodes, therefore use plane formula network structure.
The present embodiment uses standardized Ad-hoc network communication mode, and each node of Ad-hoc network passes through IEEE
802.11 agreement is communicated.When networking, start node will be set as when a certain node by software by user, and send request
The wireless signal of networking.Network node to be added searches for the signal and start node completes that the network is added after confirming.Work as institute
After having node that network is added, start node turn-off request networking signal, to complete the establishment process of network.Each node is pressed
The sequencing for being shining into net is assigned node serial number.
Step b, by distributed microphone array netinit, i.e., it is synchronous sample rate to be carried out to network node.
Fig. 3 is the synchronous flow chart of distributed microphone array network samples rate.
Specifically includes the following steps:
Step b1, network samples rate is initialized, makes K=1, i.e. network samples rate f0Equal to the equipment sample rate of node 1
f1;
Step b2, the equipment sample rate of node K is fK;By the equipment sample rate f of node KKIt is transferred to node K+1;
If step b3, the equipment sample rate f K+1 > f of node K+1K, then f0=fK, otherwise f0=fK+1;
Step b4, K=K+1;
Step b5, step b2~step b4 is repeated, until all nodes are traversed, thus network samples rate f0For whole network institute
There is the equipment sample rate minimum value of node;
Step b6, pass through finish node, i.e., the last one node in ergodic process, by current network sample rate f0Transmission
To other each nodes, so that all node device sample rates are f0。
Network samples rate in step b is the software sampling rate of whole network.Node device sample rate is that node passes through firmly
The sample rate of part equipment acquisition voice signal.
Step b further includes that clock is synchronous.
Microphone array further includes network equipment clock, is set on communication module.Time synchronization is set by network
Standby clock is synchronized based on ntp network time protocol.Number of the sequence of network as present node is added in node, initial to compile
Number be 1.Communication module between each node uses high-accuracy network time protocol NTP and number to keep for 1 start node
Clock is synchronous.Audio collection module in node reads the network equipment clock in communication module, at the beginning of audio collection
Between and communication module specific time node TsAlignment.TsValue specified by user, and the whole network is sent to by start node
Network.
(2), based on the single node speech enhan-cement of multichannel Wiener filter.
Step c, the signal of each node is subjected to framing, the multinode multichannel microphone array after obtaining framing observes letter
Number.Signal framing in step c inhibits spectral leakage using hamming window or Hanning window.Step c uses the framing of time aliasing
Strategy.
Step d: in each node, for the multichannel microphone array observation signal of each frame, according to the more of present node
Road microphone array observation signal carries out speech enhan-cement using multichannel Wiener filter, and voice is believed after obtaining single channel enhancing
Number.
It is eliminated relative to Beam-former and generalized sidelobe, a clear superiority of multichannel Wiener filter is without estimating
Meter sound bearing can effectively realize speech enhan-cement.Since target language sound bearing may often change under actual conditions, and make an uproar
The sound bearing that variation is tracked under acoustic environment is particularly difficult, therefore the present invention carries out voice increasing using multichannel Wiener filter
By force.
Wiener filter can be calculated in time domain or frequency domain, theoretically, Time-Domain algorithm and frequency domain algorithm equivalent,
But in practice due to the difference of time domain and Frequency domain noise estimated result, so that algorithm output is not fully consistent.In addition, two changes
Changing domain algorithm, there is also the differences of computation complexity.
Fig. 4 is the flow chart of the single node speech enhan-cement based on multichannel Wiener filter.It is as shown in Figure 4: to first have to pair
Speech activity detect or estimate voice existing probability, secondly estimates noise autocorrelation matrix, again
Noise autocorrelation matrix is calculated, the calculating of multichannel Wiener filter is finally carried out.
The method being filtered using the multichannel Wiener filter of time domain or frequency domain to original multiple signals is as follows:
In node K, the expression formula of the time domain multichannel Wiener filter are as follows:
hW, K(t)=[RXx, K(t)+λRNn, K(t)]-1RXx, K(t)u;
In above formula, RXx, K(t)=RYy, K(t)-RNn, K(t);
It is the clean speech vector x of present nodeK(t)=[x1,K(t),x2,K
(t),...,xM,K(t)]TTime domain autocorrelation matrix;
It is the noise vector n of present nodeK(t)=[n1,K(t),n2,K(t),...,
nM,K(t)]TTime domain autocorrelation matrix;
It is the multichannel microphone array observation signal vector y of present nodeK(t)=
[y1,K(t),y2,K(t),...,yM,K(t)]TTime domain autocorrelation matrix;
U=[1,0 ..., 0]T, the length is M;
M is the number of microphone of present node;
λ is the degree for controlling noise elimination and speech distortion, and λ > 0, when λ is bigger, the repressed effect of noise is more obvious,
More speech distortions are brought simultaneously;
The time-domain filtering of node K exports
In node K, the expression formula of the frequency-domain multi-channel Wiener filter are as follows:
HW, K(ω)=[RXX, K(ω)+λRNN, K(ω)]-1RXX, K(ω)u;
Above formula, RXX, K(ω)=RYY, K(ω)-RNN, K(ω);
It is the clean speech vector X of present nodeK(ω)=[X1, K
(ω), X2, K(ω) ..., XM, K(ω)]HFrequency domain autocorrelation matrix;
It is the noise vector N of present nodeK(ω)=[N1,K(ω),N2,K
(ω),...,NM,K(ω)]HFrequency domain autocorrelation matrix;
It is the multichannel microphone array observation signal vector Y of present nodeK(ω)
=[Y1,K(ω),Y2,K(ω),...,YM,K(ω)]HFrequency domain autocorrelation matrix;
U=[1,0 ..., 0]T, the length is M;
M is the number of microphone of present node;
λ is the degree for controlling noise elimination and speech distortion, and λ > 0, λ is bigger, and the repressed effect of noise is more obvious, together
When bring more speech distortions;
The frequency domain filtering of node K exports
When node only includes a microphone, voice signal is original observation letter after the single channel enhancing of node output
Number.
The critical issue of multichannel Wiener filter is the estimation of noise autocorrelation matrix.In the time domain, which can tie
Voice activity detection is closed to be estimated.Assuming that present frame is judged as noise, then
In above formula, 0 < α > 1 is updating factor.Otherwise, the matrix is kept not update.Similarly, combinable on frequency domain
Voice existing probability is estimated.Assuming that the voice existing probability of present frame frequency band omega is p (ω), then RNN, K(ω) updates are as follows:
RNN, K(ω)←αpRNN, K(ω)+(1-αp)X(ω)XH(ω);
Wherein, αp=α+p (ω) (1- α), similarly 0 < α < 1 is updating factor.The noise autocorrelation of time domain or frequency domain
Matrix initialisation is the average value of matrix in initial several frames.
When node only includes a microphone, in order to avoid the enhanced voice signal distortion of this node, the node is defeated
Voice signal is original multichannel microphone array observation signal after single channel enhancing out.
(3), the iterative speech enhan-cement of multinode based on multichannel Wiener filter.
Step e, in each node, transmitting voice signal after which is enhanced by the single channel that the step d is obtained
To the every other node of network.In step e, transmitting node, receiving node sequence can also be added in the data packet of signal transmission
Number and the information such as multichannel Wiener filter number of processes, to be mutually distinguishable with other data packets.
Step f, in each node, while according to the multichannel observation signal of the microphone array of present node and every other
Voice signal after the single channel enhancing of node, carries out speech enhan-cement using multichannel Wiener filter again, obtains present node
Voice signal after updated single channel enhancing;
Each node can obtain the enhanced voice letter of single channel according to the microphone array observation signal of node itself
Number.On the one hand the different enhanced voice signals of node inhibit the noise around the node, on the other hand provide pure language
The redundancy of sound, therefore can be used to further increase the effect of speech enhan-cement by other nodes.Go out from the angle of network communication
Hair, transmits enhanced single-channel voice signal, and the multi-channel original signal that not a node is observed, has been greatly saved band
Width, and ensure that the consistency of internodal data transformat.
This link is by the multi-channel GPS observations signal of single-channel voice signal after the enhancing of remaining node and this node together structure
The observation vector of Cheng Xin.Single-channel voice signal can be considered the new observation channel of local node, class after the enhancing of remaining node
As, multichannel Wiener filter can be used, according to above-mentioned new observation vector, obtains list after updated node enhancing
Channel speech signal.
Step g, iteration step e~step f, when voice signal is restrained after the single channel enhancing that certain node obtains,
Voice signal no longer updates after the single channel enhancing of present node;When all nodes single channel enhancing after voice signal no longer
When update, processing terminate for present frame;The finally voice signal after each node obtains present node enhancing.
Fig. 5 is the flow chart of the iterative speech enhan-cement of multinode based on multichannel Wiener filter.Construction is current first
The joint vector that signal is constituted after the enhancing of the multichannel observation signal and every other node of node, secondly to speech activity into
Voice existing probability is estimated in row detection, updates noise autocorrelation matrix again, calculates signal with noise later
Autocorrelation matrix finally carries out the calculating of multichannel Wiener filter.
Circular is as follows:
In node K, the multichannel Wiener filter of the time domain,
The joint vector that signal is constituted after the enhancing of present node K multichannel observation signal and every other node is
In above formula,For except section
Vector composed by the enhanced time domain single-channel voice of other outer nodes of point K:
NiFor the number of iterations of step g;
It isIn clean speech ingredient;
It isIn noise contribution;
For clean speech ingredient in present nodeTime domain from phase
Close matrix;
For noise contribution in present nodeTime domain autocorrelation matrix;
To combine vector in present nodeTime domain autocorrelation matrix;
U=[1,0 ..., 0]T, it is the node total number in network the length is M+P-1, P;
Then node K NiThe time domain multichannel Wiener filter of+1 iteration are as follows:
In node K, the multichannel Wiener filter of the frequency domain,
The joint vector that signal is constituted after the enhancing of present node K multichannel observation signal and every other node is
In above formula,
For vector composed by the enhanced frequency domain single-channel voice of other nodes in addition to node K;
NiFor the number of iterations of step g;
To combine vectorIn clean speech ingredient;
To combine vectorIn noise contribution;
For remove node K except other nodes it is pure
The frequency domain autocorrelation matrix of speech vector;
To remove other node background noise vectors except node K
Frequency domain autocorrelation matrix;
To remove the frequency domain of other node observation vectors except node K certainly
Correlation matrix;
U=[1,0 ..., 0]T;The length is M+P-1, P is the node total number in network, then node K Ni+ 1 iteration
Frequency-domain multi-channel Wiener filter are as follows:
It, equally can be by this updated signal after all nodes obtain updated single-channel voice enhancing signal
Other nodes are transmitted to, in order to which other nodes update single-channel voice enhancing signal again.It therefore, can be in distributed Mike
Wind array network repeats the above steps, and when voice signal is restrained after the single channel enhancing that certain node obtains, voice is believed after enhancing
It number no longer updates.When all node single-channel voice signals no longer update, processing terminate for present frame, finally in each node
Obtain present node enhancing after voice signal.
Step g may also include the step of whether voice signal restrains judged.We can be according to filtering front and back signal vector
Difference norm and signal energy comprehensive descision node obtain single channel enhancing after voice signal whether restrain: step g
Language after the single channel enhancing obtained according to the norm of the difference of filtering front and back signal vector and signal energy comprehensive descision node
Whether sound signal restrains, and method is as follows:
In node K, the previous obtained single channel time domain signal vector that filters is
This filters obtained single channel time domain signal vector
WhenWhen, it is believed that present filter output convergence;
||·||pP norm is represented, η is a threshold value.
What has been described above is only a preferred embodiment of the present invention, and present invention is not limited to the above embodiments.It is appreciated that this
The other improvements and change that field technical staff directly exports or associates without departing from the spirit and concept in the present invention
Change, is considered as being included within protection scope of the present invention.
Claims (9)
1. a kind of sound enhancement method based on distributed microphone array network, characterized in that it comprises the following steps:
Step a, the distributed microphone array network based on Ad-hoc network being made of multiple microphone arrays is established;Arbitrarily
It can be in communication with each other between two network nodes;
Step b, by distributed microphone array netinit, i.e., it is synchronous sample rate to be carried out to network node;
Step c, the signal of each node is subjected to framing, the multinode multichannel microphone array observation signal after obtaining framing;
Step d, in each node, for the multichannel microphone array observation signal of each frame, according to the multichannel wheat of present node
Gram wind array observation signal carries out speech enhan-cement using multichannel Wiener filter, obtains voice signal after single channel enhancing;
Step e, in each node, transmitting voice signal is to net after the single channel enhancing which is obtained by the step d
The every other node of network;
Step f, in each node, while according to the multichannel microphone array observation signal of present node and every other node
Voice signal after single channel enhancing, carries out speech enhan-cement using multichannel Wiener filter again, after obtaining present node update
Single channel enhancing after voice signal;
Step g, iteration step e~step f, when voice signal is restrained after the single channel enhancing that certain node obtains, currently
Voice signal no longer updates after the single channel enhancing of node;Voice signal no longer updates after the single channel enhancing of all nodes
When, processing terminate for present frame;The finally voice signal after each node obtains present node enhancing;
The step b specifically includes the following steps:
Step b1, network samples rate is initialized, makes K=1, i.e. network samples rate f0Equal to the equipment sample rate f of node 11;
Step b2, the equipment sample rate of node K is fK;By the equipment sample rate f of node KKIt is transferred to node K+1;
If step b3, the equipment sample rate f of node K+1K+1> fK, then f0=fK, otherwise f0=fK+1;
Step b4, K=K+1;
Step b5, step b2~step b4 is repeated, until all nodes are traversed, thus network samples rate f0For all sections of whole network
The equipment sample rate minimum value of point;
Step b6, pass through finish node for current network sample rate f0Other each nodes are transferred to, so that all node devices are adopted
Sample rate is f0。
2. the sound enhancement method as described in claim 1 based on distributed microphone array network, it is characterised in that: described
Microphone array includes audio collection module and communication module.
3. the sound enhancement method as described in claim 1 based on distributed microphone array network, it is characterised in that: described
The structure of Ad-hoc network in step a is planar structure or hierarchical structure;Ad-hoc network using priori formula, reaction equation or
The hybrid-type Routing Protocol of person realizes being in communication with each other between two node devices in network.
4. the sound enhancement method as described in claim 1 based on distributed microphone array network, it is characterised in that: described
Step b further includes carrying out time synchronization to network node;The distribution microphone array includes network equipment clock;When described
Between to synchronize be to be synchronized by the network equipment clock based on ntp network time protocol.
5. the sound enhancement method as described in claim 1 based on distributed microphone array network, it is characterised in that: described
Signal framing in step c inhibits spectral leakage using hamming window or Hanning window;The step c uses the framing of time aliasing
Strategy.
6. the sound enhancement method as described in claim 1 based on distributed microphone array network, it is characterised in that: described
Step d, which observes multichannel microphone array using the multichannel Wiener filter of time domain multichannel Wiener filter or frequency domain, to be believed
It number is filtered, to achieve the effect that speech enhan-cement:
In node K, the expression formula of the time domain multichannel Wiener filter are as follows:
hW, K(t)=[RXx, K(t)+λRNn, K(t)]-1RXx, K(t)u;
In above formula, RXx, K(t)=RYy, K(t)-RNn, K(t);
It is the clean speech vector x of present nodeK(t)=[x1,K(t),x2,K
(t),…,xM,K(t)]TTime domain autocorrelation matrix;
It is the noise vector n of present nodeK(t)=[n1,K(t),n2,K(t),...,nM,K
(t)]TTime domain autocorrelation matrix;
It is the multichannel microphone array observation signal vector y of present nodeK(t)=[y1,K(t),
y2,K(t),...,yM,K(t)]TTime domain autocorrelation matrix;
U=[1,0 ..., 0]T, the length is M;
M is the number of microphone of present node;
λ is the degree for controlling noise elimination and speech distortion, and λ > 0, λ is bigger, and the repressed effect of noise is more obvious, while band
Carry out more speech distortions;
The time-domain filtering of node K exports are as follows:
In node K, the expression formula of the frequency-domain multi-channel Wiener filter are as follows:
HW, K(ω)=[RXX, K(ω)+λRNN, K(ω)]-1RXX, K(ω)u;
In above formula, RXX, K(ω)=RYY, K(ω)-RNN, K(ω);
It is the clean speech vector X of present nodeK(ω)=[X1, K(ω),
X2, K(ω) ..., XM, K(ω)]HFrequency domain autocorrelation matrix;
It is the noise vector N of present nodeK(ω)=[N1,K(ω),N2,K
(ω),...,NM,K(ω)]HFrequency domain autocorrelation matrix;
It is the multichannel microphone array observation signal vector Y of present nodeK(ω)=
[Y1,K(ω),Y2,K(ω),...,YM,K(ω)]HFrequency domain autocorrelation matrix;
U=[1,0 ..., 0]T, the length is M;
M is the number of microphone of present node;
λ is the degree for controlling noise elimination and speech distortion, and λ > 0, λ is bigger, and the repressed effect of noise is more obvious, while band
Carry out more speech distortions;
The frequency domain filtering of node K exports are as follows:
7. the sound enhancement method as described in claim 1 based on distributed microphone array network, it is characterised in that: described
Step e includes that transmitting node serial number, receiving node serial number and multichannel Wiener filtering are added in the data packet of signal transmission
The information of device number of processes.
8. the sound enhancement method as described in claim 1 based on distributed microphone array network, it is characterised in that: described
Step f includes the increasing using the multichannel Wiener filter of time domain or frequency domain to present node multichannel observation signal and other nodes
Signal is filtered after strong;
In the multichannel Wiener filter of the time domain,
The joint vector that signal is constituted after the enhancing of present node K multichannel microphone array observation signal and every other node
Are as follows:
In above formula,For except node K
Vector composed by the enhanced time domain single-channel voice of other outer nodes;
NiFor the number of iterations of step g;
It isIn clean speech ingredient;
It isIn noise contribution;
For clean speech ingredient in present nodeTime domain auto-correlation square
Battle array;
For noise vector in present nodeTime domain autocorrelation matrix;
To combine vector in present nodeTime domain autocorrelation matrix;
U=[1,0 ..., 0]T, it is the node total number in network the length is M+P-1, P;
Then node K NiThe time domain multichannel Wiener filter of+1 iteration are as follows:
In the multichannel Wiener filter of the frequency domain,
The joint vector that signal is constituted after the enhancing of present node K multichannel observation signal and every other node are as follows:
In above formula,For
The vector composed by the enhanced frequency domain single-channel voice of other nodes in addition to node K;
NiFor the number of iterations of step g;
ForIn clean speech ingredient;
ForIn noise contribution;
For remove node K except other nodes clean speech to
The frequency domain autocorrelation matrix of amount;
For the frequency domain for removing other node background noise vectors except node K
Autocorrelation matrix;
For the frequency domain auto-correlation for removing other node observation vectors except node K
Matrix;
U=[1,0 ..., 0]T;The length is M+P-1, P is the node total number in network, then node K NiThe frequency of+1 iteration
Domain multichannel Wiener filter are as follows:
9. the sound enhancement method as described in claim 1 based on distributed microphone array network, it is characterised in that: described
Step g includes judging what node obtained according to the norm of difference and the norm of filtered signal vector of filtering front and back signal vector
The step of whether voice signal restrains after single channel enhancing, method is as follows:
It is previous to filter obtained single channel time domain signal vector in node K are as follows:
This filters obtained single channel time domain signal vector are as follows:
WhenWhen, present filter output convergence;
In above formula, | | | |pP norm is represented, η is threshold value.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201510582363.5A CN105206281B (en) | 2015-09-14 | 2015-09-14 | Sound enhancement method based on distributed microphone array network |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201510582363.5A CN105206281B (en) | 2015-09-14 | 2015-09-14 | Sound enhancement method based on distributed microphone array network |
Publications (2)
Publication Number | Publication Date |
---|---|
CN105206281A CN105206281A (en) | 2015-12-30 |
CN105206281B true CN105206281B (en) | 2019-02-15 |
Family
ID=54953910
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN201510582363.5A Active CN105206281B (en) | 2015-09-14 | 2015-09-14 | Sound enhancement method based on distributed microphone array network |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN105206281B (en) |
Families Citing this family (74)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US10509626B2 (en) | 2016-02-22 | 2019-12-17 | Sonos, Inc | Handling of loss of pairing between networked devices |
US10743101B2 (en) | 2016-02-22 | 2020-08-11 | Sonos, Inc. | Content mixing |
US9965247B2 (en) | 2016-02-22 | 2018-05-08 | Sonos, Inc. | Voice controlled media playback system based on user profile |
US10095470B2 (en) | 2016-02-22 | 2018-10-09 | Sonos, Inc. | Audio response playback |
US9947316B2 (en) | 2016-02-22 | 2018-04-17 | Sonos, Inc. | Voice control of a media playback system |
US10264030B2 (en) | 2016-02-22 | 2019-04-16 | Sonos, Inc. | Networked microphone device control |
CN105957536B (en) * | 2016-04-25 | 2019-11-12 | 深圳永顺智信息科技有限公司 | Based on channel degree of polymerization frequency domain echo cancel method |
US9978390B2 (en) | 2016-06-09 | 2018-05-22 | Sonos, Inc. | Dynamic player selection for audio signal processing |
CN106028227B (en) * | 2016-07-08 | 2019-05-24 | 乐鑫信息科技(上海)股份有限公司 | Distributed microphone array and its applicable sonic location system |
US10152969B2 (en) | 2016-07-15 | 2018-12-11 | Sonos, Inc. | Voice detection by multiple devices |
US10115400B2 (en) | 2016-08-05 | 2018-10-30 | Sonos, Inc. | Multiple voice services |
GB201615538D0 (en) * | 2016-09-13 | 2016-10-26 | Nokia Technologies Oy | A method , apparatus and computer program for processing audio signals |
US9942678B1 (en) | 2016-09-27 | 2018-04-10 | Sonos, Inc. | Audio playback settings for voice interaction |
US9743204B1 (en) | 2016-09-30 | 2017-08-22 | Sonos, Inc. | Multi-orientation playback device microphones |
US10181323B2 (en) | 2016-10-19 | 2019-01-15 | Sonos, Inc. | Arbitration-based voice recognition |
CN106782590B (en) * | 2016-12-14 | 2020-10-09 | 南京信息工程大学 | Microphone array beam forming method based on reverberation environment |
CN106846803B (en) * | 2017-02-08 | 2023-06-23 | 广西交通科学研究院有限公司 | Traffic event detection device and method based on audio frequency |
CN110169082B (en) | 2017-03-08 | 2021-03-23 | 惠普发展公司,有限责任合伙企业 | Method and apparatus for combining audio signal outputs, and computer readable medium |
CN106992010B (en) * | 2017-06-02 | 2020-02-21 | 厦门大学 | Microphone array speech enhancement device under condition of no direct sound |
US10475449B2 (en) | 2017-08-07 | 2019-11-12 | Sonos, Inc. | Wake-word detection suppression |
US10048930B1 (en) | 2017-09-08 | 2018-08-14 | Sonos, Inc. | Dynamic computation of system response volume |
US10446165B2 (en) | 2017-09-27 | 2019-10-15 | Sonos, Inc. | Robust short-time fourier transform acoustic echo cancellation during audio playback |
US10482868B2 (en) | 2017-09-28 | 2019-11-19 | Sonos, Inc. | Multi-channel acoustic echo cancellation |
US10621981B2 (en) | 2017-09-28 | 2020-04-14 | Sonos, Inc. | Tone interference cancellation |
US10466962B2 (en) | 2017-09-29 | 2019-11-05 | Sonos, Inc. | Media playback system with voice assistance |
CN107993670B (en) * | 2017-11-23 | 2021-01-19 | 华南理工大学 | Microphone array speech enhancement method based on statistical model |
US10880650B2 (en) | 2017-12-10 | 2020-12-29 | Sonos, Inc. | Network microphone devices with automatic do not disturb actuation capabilities |
US10818290B2 (en) | 2017-12-11 | 2020-10-27 | Sonos, Inc. | Home graph |
US11343614B2 (en) | 2018-01-31 | 2022-05-24 | Sonos, Inc. | Device designation of playback and network microphone device arrangements |
US11175880B2 (en) | 2018-05-10 | 2021-11-16 | Sonos, Inc. | Systems and methods for voice-assisted media content selection |
US10847178B2 (en) | 2018-05-18 | 2020-11-24 | Sonos, Inc. | Linear filtering for noise-suppressed speech detection |
US10959029B2 (en) | 2018-05-25 | 2021-03-23 | Sonos, Inc. | Determining and adapting to changes in microphone performance of playback devices |
US10681460B2 (en) | 2018-06-28 | 2020-06-09 | Sonos, Inc. | Systems and methods for associating playback devices with voice assistant services |
CN110364166B (en) * | 2018-06-28 | 2022-10-28 | 腾讯科技(深圳)有限公司 | Electronic equipment for realizing speech signal recognition |
CN109192196A (en) * | 2018-08-22 | 2019-01-11 | 昆明理工大学 | A kind of audio frequency characteristics selection method of the SVM classifier of anti-noise |
US10461710B1 (en) | 2018-08-28 | 2019-10-29 | Sonos, Inc. | Media playback system with maximum volume setting |
US11076035B2 (en) | 2018-08-28 | 2021-07-27 | Sonos, Inc. | Do not disturb feature for audio notifications |
US10587430B1 (en) | 2018-09-14 | 2020-03-10 | Sonos, Inc. | Networked devices, systems, and methods for associating playback devices based on sound codes |
US11024331B2 (en) | 2018-09-21 | 2021-06-01 | Sonos, Inc. | Voice detection optimization using sound metadata |
US10811015B2 (en) | 2018-09-25 | 2020-10-20 | Sonos, Inc. | Voice detection optimization based on selected voice assistant service |
US11100923B2 (en) | 2018-09-28 | 2021-08-24 | Sonos, Inc. | Systems and methods for selective wake word detection using neural network models |
US10692518B2 (en) | 2018-09-29 | 2020-06-23 | Sonos, Inc. | Linear filtering for noise-suppressed speech detection via multiple network microphone devices |
US11899519B2 (en) | 2018-10-23 | 2024-02-13 | Sonos, Inc. | Multiple stage network microphone device with reduced power consumption and processing load |
EP3654249A1 (en) | 2018-11-15 | 2020-05-20 | Snips | Dilated convolutions and gating for efficient keyword spotting |
US11183183B2 (en) | 2018-12-07 | 2021-11-23 | Sonos, Inc. | Systems and methods of operating media playback systems having multiple voice assistant services |
US11132989B2 (en) | 2018-12-13 | 2021-09-28 | Sonos, Inc. | Networked microphone devices, systems, and methods of localized arbitration |
US10602268B1 (en) | 2018-12-20 | 2020-03-24 | Sonos, Inc. | Optimization of network microphone devices using noise classification |
US11315556B2 (en) | 2019-02-08 | 2022-04-26 | Sonos, Inc. | Devices, systems, and methods for distributed voice processing by transmitting sound data associated with a wake word to an appropriate device for identification |
US10867604B2 (en) | 2019-02-08 | 2020-12-15 | Sonos, Inc. | Devices, systems, and methods for distributed voice processing |
US11120794B2 (en) | 2019-05-03 | 2021-09-14 | Sonos, Inc. | Voice assistant persistence across multiple network microphone devices |
US11361756B2 (en) | 2019-06-12 | 2022-06-14 | Sonos, Inc. | Conditional wake word eventing based on environment |
US10586540B1 (en) | 2019-06-12 | 2020-03-10 | Sonos, Inc. | Network microphone device with command keyword conditioning |
US11200894B2 (en) | 2019-06-12 | 2021-12-14 | Sonos, Inc. | Network microphone device with command keyword eventing |
US11138969B2 (en) | 2019-07-31 | 2021-10-05 | Sonos, Inc. | Locally distributed keyword detection |
US11138975B2 (en) | 2019-07-31 | 2021-10-05 | Sonos, Inc. | Locally distributed keyword detection |
US10871943B1 (en) | 2019-07-31 | 2020-12-22 | Sonos, Inc. | Noise classification for event detection |
US11189286B2 (en) | 2019-10-22 | 2021-11-30 | Sonos, Inc. | VAS toggle based on device orientation |
US11200900B2 (en) | 2019-12-20 | 2021-12-14 | Sonos, Inc. | Offline voice control |
US11562740B2 (en) | 2020-01-07 | 2023-01-24 | Sonos, Inc. | Voice verification for media playback |
US11556307B2 (en) | 2020-01-31 | 2023-01-17 | Sonos, Inc. | Local voice data processing |
US11308958B2 (en) | 2020-02-07 | 2022-04-19 | Sonos, Inc. | Localized wakeword verification |
CN111048106B (en) * | 2020-03-12 | 2020-06-16 | 深圳市友杰智新科技有限公司 | Pickup method and apparatus based on double microphones and computer device |
US11727919B2 (en) | 2020-05-20 | 2023-08-15 | Sonos, Inc. | Memory allocation for keyword spotting engines |
US11482224B2 (en) | 2020-05-20 | 2022-10-25 | Sonos, Inc. | Command keywords with input detection windowing |
US11308962B2 (en) | 2020-05-20 | 2022-04-19 | Sonos, Inc. | Input detection windowing |
US11698771B2 (en) | 2020-08-25 | 2023-07-11 | Sonos, Inc. | Vocal guidance engines for playback devices |
CN112735462A (en) * | 2020-12-30 | 2021-04-30 | 科大讯飞股份有限公司 | Noise reduction method and voice interaction method of distributed microphone array |
CN112652310A (en) * | 2020-12-31 | 2021-04-13 | 乐鑫信息科技(上海)股份有限公司 | Distributed speech processing system and method |
CN112820287A (en) * | 2020-12-31 | 2021-05-18 | 乐鑫信息科技(上海)股份有限公司 | Distributed speech processing system and method |
CN112954122B (en) * | 2021-01-22 | 2022-10-11 | 成都天奥信息科技有限公司 | Voice selecting method for very high frequency voice communication system |
US11551700B2 (en) | 2021-01-25 | 2023-01-10 | Sonos, Inc. | Systems and methods for power-efficient keyword detection |
CN113257270B (en) * | 2021-05-10 | 2022-07-15 | 中国科学技术大学 | Multi-channel voice enhancement method based on reference microphone optimization |
CN113744751A (en) * | 2021-08-16 | 2021-12-03 | 清华大学苏州汽车研究院(相城) | Multi-channel speech signal enhancement method applied to microphone array |
CN114283832A (en) * | 2021-09-09 | 2022-04-05 | 腾讯科技(深圳)有限公司 | Processing method and device for multi-channel audio signal |
Citations (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN101292547A (en) * | 2005-10-13 | 2008-10-22 | 摩托罗拉公司 | Method and apparatus for synchronizing a node within an ad-hoc communication system |
CN101587712A (en) * | 2008-05-21 | 2009-11-25 | 中国科学院声学研究所 | A kind of directional speech enhancement method based on minitype microphone array |
CN101772983A (en) * | 2007-07-31 | 2010-07-07 | 摩托罗拉公司 | System and method of resource allocation within a communication system |
Family Cites Families (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
AU2012236649A1 (en) * | 2011-03-28 | 2013-10-31 | Ambientz | Methods and systems for searching utilizing acoustical context |
-
2015
- 2015-09-14 CN CN201510582363.5A patent/CN105206281B/en active Active
Patent Citations (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN101292547A (en) * | 2005-10-13 | 2008-10-22 | 摩托罗拉公司 | Method and apparatus for synchronizing a node within an ad-hoc communication system |
CN101772983A (en) * | 2007-07-31 | 2010-07-07 | 摩托罗拉公司 | System and method of resource allocation within a communication system |
CN101587712A (en) * | 2008-05-21 | 2009-11-25 | 中国科学院声学研究所 | A kind of directional speech enhancement method based on minitype microphone array |
Non-Patent Citations (3)
Title |
---|
"Distributed Adaptive Node-Specific Signal Estimation in Fully Connected Sensor Networks—Part I: Sequential Node Updating";A. Bertrand , ect.;《IEEE Transactions on Signal Processing》;20101031;第58卷(第10期);第5页第1栏第2段-第6页第1栏第4段,图3 |
"Distributed GSC Beamforming Using the Relative Transfer Function";M. G. Shmulik, ect.;《20th European Signal Processing Conference》;20120831;第1页摘要,第1部分第4段-第4页第5部分 |
"on Optimal Frequency-Domain Multichannel Linear Filtering for Noise Reduction";S. Mehrez, ect.;《IEEE Transactions on Audio, speech, and language processing》;20100228;第2卷(第18期);第3页第3段-第5页,第7页第IV部分 |
Also Published As
Publication number | Publication date |
---|---|
CN105206281A (en) | 2015-12-30 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN105206281B (en) | Sound enhancement method based on distributed microphone array network | |
CN106251877B (en) | Voice Sounnd source direction estimation method and device | |
Zeng et al. | Distributed delay and sum beamformer for speech enhancement via randomized gossip | |
CN105388459B (en) | The robust sound source space-location method of distributed microphone array network | |
CN102204281B (en) | A system and method for producing a directional output signal | |
CN107211225A (en) | Hearing assistant system | |
CN106710601A (en) | Voice signal de-noising and pickup processing method and apparatus, and refrigerator | |
CN110931031A (en) | Deep learning voice extraction and noise reduction method fusing bone vibration sensor and microphone signals | |
CN105741849A (en) | Voice enhancement method for fusing phase estimation and human ear hearing characteristics in digital hearing aid | |
CN105872275B (en) | A kind of speech signal time delay estimation method and system for echo cancellor | |
CN107621625B (en) | Sound source positioning method based on double micro microphones | |
CN107316648A (en) | A kind of sound enhancement method based on coloured noise | |
JP5337072B2 (en) | Model estimation apparatus, sound source separation apparatus, method and program thereof | |
CN104995679A (en) | Signal source separation | |
CN106057210B (en) | Quick speech blind source separation method based on frequency point selection under binaural distance | |
CN103026738A (en) | Method of signal processing in a hearing aid system and a hearing aid system | |
CN108109617A (en) | A kind of remote pickup method | |
CN108389586A (en) | A kind of long-range audio collecting device, monitoring device and long-range collection sound method | |
CN106297817B (en) | A kind of sound enhancement method based on binaural information | |
CN108986832A (en) | Ears speech dereverberation method and device based on voice probability of occurrence and consistency | |
Liu | Sound source seperation with distributed microphone arrays in the presence of clocks synchronization errors | |
CN109874096A (en) | A kind of ears microphone hearing aid noise reduction algorithm based on intelligent terminal selection output | |
Cherkassky et al. | Blind synchronization in wireless sensor networks with application to speech enhancement | |
CN112201276B (en) | TC-ResNet network-based microphone array voice separation method | |
CN110739004B (en) | Distributed voice noise elimination system for WASN |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
C06 | Publication | ||
PB01 | Publication | ||
C10 | Entry into substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
GR01 | Patent grant | ||
GR01 | Patent grant |