CN103854660B - A kind of four Mike's sound enhancement methods based on independent component analysis - Google Patents

A kind of four Mike's sound enhancement methods based on independent component analysis Download PDF

Info

Publication number
CN103854660B
CN103854660B CN201410061180.4A CN201410061180A CN103854660B CN 103854660 B CN103854660 B CN 103854660B CN 201410061180 A CN201410061180 A CN 201410061180A CN 103854660 B CN103854660 B CN 103854660B
Authority
CN
China
Prior art keywords
signal
frequency
voice
source
mike
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN201410061180.4A
Other languages
Chinese (zh)
Other versions
CN103854660A (en
Inventor
张彦芳
王芳
周海瑞
王犇
朱冰
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
CETC 28 Research Institute
Original Assignee
CETC 28 Research Institute
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by CETC 28 Research Institute filed Critical CETC 28 Research Institute
Priority to CN201410061180.4A priority Critical patent/CN103854660B/en
Publication of CN103854660A publication Critical patent/CN103854660A/en
Application granted granted Critical
Publication of CN103854660B publication Critical patent/CN103854660B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Abstract

The invention discloses a kind of four Mike's sound enhancement methods based on independent component analysis, comprise the steps of four Mikes of use and form array, gather four road Noisy Speech Signals and be converted into frequency-region signal;The two paths of signals randomly choosed in four microphone arrays on same one side is as first group of input signal, using other two paths of signals as second group of input signal, estimates a separation matrix W respectivelyf;Step 3: by the frequency-region signal obtained in step 1, utilize multiple signal classification algorithm to process, obtains source speech signal direction θ;Step 4: using source speech signal direction θ calculated in step 3 as reference information, extracts voice signal during often group separates signal from step 2;Step 5: carry out frequency sequence, the amplitude that carries out gained voice signal smooths, and utilizes a window function that the signal of side frequency carries out amplitude and smooths;Step 6: obtain the time-domain signal of voice signal through windowing and inverse Fourier transform.

Description

A kind of four Mike's sound enhancement methods based on independent component analysis
Technical field
A kind of method that the present invention relates to speech enhan-cement, a kind of four Mike's voices based on independent component analysis increase Strong method.
Background technology
In the communication system of reality, it is often necessary to a Mike gathers sound, during gathering, target voice is frequent Can be polluted by background noise.Therefore, the signals with noise that Mike collects needs, through speech-enhancement system, to remove noise Impact after play again, transmit or preserve.The research of voice enhancement algorithm has had the history of more than 40 year, be the earliest by A kind of algorithm that Schroeder in Bell laboratory proposes.A lot of people is had to propose other algorithms, wherein the most successively Spectrum-subtraction is the most most widely used algorithm, but spectrum-subtraction exists obvious defect, meeting while suppression noise Damage voice, and artificial introducing noise, e.g., " music " (musical) noise.
Voice enhancement algorithm based on single Mike can not improve voice quality and the intelligibility of speech, carrying of voice quality simultaneously Height is usually associated with the reduction of the intelligibility of speech.In recent years, voice enhancement algorithm based on microphone array is increasingly subject to Paying attention to, microphone array can utilize the spatial information of source location, therefore voice enhancement algorithm based on microphone array Noise can be suppressed in the case of not damaging voice.Voice enhancement algorithm based on microphone array has become as a kind of new Research tendency.More ripe voice enhancement algorithm based on microphone array has: adaptive beamforming algorithm, nothing Distortion minimization variance response beam forming, generalized sidelobe cancellation and generalized singular value decomposition scheduling algorithm.Existing based on In the voice enhancement algorithm of microphone array, microphone array is larger, and the space complexity of algorithm and time complexity are the most relatively High, it is impossible in flexible Application and portable communication device, such as mobile phone and transmitter receiver.
Summary of the invention
In order to overcome above-mentioned existing voice to strengthen the problem that algorithm exists, the invention provides one and divide based on independent element Four Mike's sound enhancement methods of analysis, can reduce the damage to voice during speech enhan-cement, reach good noise reduction Effect.
The technical scheme used of a kind of four Mike's sound enhancement methods based on independent component analysis of the present invention is:
Step 1: utilize the four road Noisy Speech Signals that microphone array gathers, signal is carried out pretreatment, through Fourier It is converted into frequency-region signal;
Step 2: using in four Mike's square formations with on two paths of signals as first group of input signal, additionally two paths of signals As second group of input signal, to often organizing pretreated frequency-region signal, utilize frequency domain independent composition analysis algorithm, pin Each frequency f in frequency domain is estimated a separation matrix W respectivelyf, often group signal correspondence obtains including a road voice letter Number and the separation signal of a road noise signal;
Step 3: by the frequency-region signal obtained in step 1, utilize multiple signal classification algorithm to process, obtain language Tone signal source direction θ;
Step 4: will estimate in step 3 that the source speech signal direction θ obtained, as reference information, is used for from often organizing separation Signal extracts voice signal;
Step 5: the amplitude that carries out frequency domain speech signal smooths;
Step 6: obtain the time-domain signal of voice signal through windowing and inverse Fourier transform.
In step 4, utilize the source speech signal direction θ estimating to obtain in step 3 as reference information, from step 2 Extract voice signal during often group separates signal to comprise the steps:
1) the separation matrix W of frequency f is utilizedfCalculating source speech signal direction, each frequency is calculated two letters Number direction, source(f) andF (), frequency number is the half of Fourier transformation length;
2) the above-mentioned calculated signal source direction of all frequencies is carried out K mean cluster, be polymerized to two classes, utilize Source speech signal direction θ, the cluster conduct within selecting source speech signal direction θ angle to differ 10 degree The incident direction of voice signal;
3) frequency with a low credibility is rejected: for same frequency, if the incident direction of one of them signal and language Within tone signal source direction θ differs 10 °, and the incident direction of its another signal clusters apart from another Center less than 20 °, then this frequency is with a high credibility, the language that chosen distance source speech signal direction θ is near The incident direction of tone signal is as final source speech signal direction;Otherwise, it is determined that this frequency was for losing efficacy Frequency;
4) for inefficacy frequency, the dependency of adjacent frequency spectrum envelope is utilized again to extract:
First, the frequency spectrum of inefficacy frequency is calculatedHarmonic structure with effective frequencyDependency it With cor (f), computing formula is:
cor ( f ) = Σ g s ^ i ( f ) s ^ i ( g ) | s ^ i ( f ) | | s ^ i ( g ) | ,
Wherein g=..., 1/3f, 1/2f, 2f, 3f ...,
Wherein g is effective frequency,
If dependency sum cor (f) is more than 0.9, then extracting this frequency is effective frequency;
Above-mentioned iterative step is repeated. until iterations is more than 100 times, for final undrawn frequency, The frequency extracted being adjacent by described frequency calculates dependency, extracts.
In step 5, the voice signal extracted being carried out windowing to smooth, use Hanning window, windowed function is as follows:
w ~ kl ( f ) = 1 4 [ w kl ( f - Δf ) + 2 w kl ( f ) + w kl ( f + Δf ) ] ,
Wherein wklF () is separation matrix WfIn a row k l row element, Δ f is between the frequency of Fourier transformation Every, equal to signal sampling rate divided by Fourier transformation length.
Speech enhan-cement, as the pretreatment module of a lot of speech signal processing systems, has important meaning to the performance improving system Justice.Voice while suppression noise, can be produced by traditional voice enhancement algorithm based on single Mike, these algorithms Raw bigger damage, and the introducing noise that meeting is artificial.For present in voice enhancement algorithm based on single Mike this A little problems, the present invention proposes a kind of voice enhancement algorithm based on four Mikes, utilizes multiple signal classification algorithm The direction in (Multiple signal classification, MUSIC) estimated speech signal source, then utilizes frequency domain independently to become Analyze (Independent component analysis, ICA) algorithm and be individually separated two groups of signals, utilize and estimate gained language The direction in tone signal source, reconfigures separating signal, judges to obtain voice signal from two-way separation signal. Compared with original voice enhancement algorithm based on single Mike, four Mike speech enhan-cement sides based on independent component analysis Method, can under various noise circumstances preferable isolated voice signal, suppression noise while the most not to voice Signal causes damage.The algorithm that the present invention mentions can be applied with the communication apparatus such as mobile phone, transmitter receiver, reaches preferable Speech enhan-cement effect.
Accompanying drawing explanation
Fig. 1 is the placement location of microphone array in the present invention.
Fig. 2 is the present invention four Mike's sound enhancement method flow charts based on independent component analysis.
Fig. 3 is to the sort algorithm flow chart between independent component analysis isolated signal different frequency in the present invention.
Fig. 4 is the graph of a relation that in pretreatment, frame length moves with frame.
Fig. 5 is the data acquisition environment schematic diagram in embodiment.
Detailed description of the invention
As in figure 2 it is shown, the invention discloses a kind of four Mike's sound enhancement methods based on independent component analysis, including Following steps.
Step 1: as it is shown in figure 1, the four road Noisy Speech Signals utilizing microphone array to gather, signal is carried out pretreatment, It is converted into frequency-region signal through Fourier transform;
The signal s that i-th mike collectsi' (n), n=1,2 ..., i value 1~4, carry out framing such as Fig. 4, frame length is L, it is N that frame moves, and the overlap between frame is L-N.
After framing, signal being carried out windowing process, use a length of L Hanning window function w (n), after windowing, signal is:
si(n, m)=si' (n, m) * w (n) (1)
Wherein n is frequency point number n=0,1,2 ..., L-1 is frame number.
To signal si(n, m) carries out discrete Fourier transform (DFT), obtain correspondence frequency-region signal:
s i ( f , m ) = Σ n = 0 L - 1 s i ( n , m ) e - j 2 π L fn - - - ( 2 )
Wherein f is frequency, and m is frame number, and e is the nature truth of a matter, and j is imaginary unit.In order to statement below is convenient, Omit frame number, the result of step 1 is expressed as frequency-region signal si(f)
Step 2: by four microphone arrays with on two mikes, such as, No. 1 in Fig. 1 and No. 2 Mikes The signal that wind collects, after pretreatment, forms one group of input signal and carries out independent component analysis, for each frequency Point is calculated a separation matrix Wf, utilize WfIsolated includes a road voice signal and a road noise signal Separate signal s1(f)s2(f)。
y 1 ( f ) y 2 ( f ) = A f s 1 ( f ) s 2 ( f ) - - - ( 3 )
Hybrid matrix AfIt is unknown, it is considered that AfIt it is full rank.The Fast Fixed-point independent element utilizing complex field divides Analysis algorithm (Fast fix-pointed ICA) algorithm estimates WfSo that:
s 1 ( f ) s 2 ( f ) = W f y 1 ( f ) y 2 ( f ) - - - ( 4 )
Solve Wf, first structure cost function is as follows:
JG(w)=E{G (| wH y|2)}
(5)
E{wHy2}=1
Wherein G:Cr→ R is a smooth function, w ∈ CnWeight vector matrix, is WfIn row, Y=[y1(f)y2(f)].Need observation signal y is carried out whitening pretreatment so that it is meet during calculating E{yyH}=I.
Assume between each signal source it is independent, meet E{ssT}=O, s=[s1(f)s2(f)].According to Ku En- Plutarch (Kuhn-Tucker) condition, optimization above formula, wherein restrictive condition E{ | wHy|2}=| | w | |2=1 can be converted into down Formula
▿ E { G ( | w H y | 2 ) } - β ▿ E { | w H y | 2 } = 0 - - - ( 6 )
Wherein β ∈ R.
▽E{G(|wHy|2) Jacobian matrix be calculated as follows:
▿ 2 E { G ( | w H y | 2 ) } = 2 E { ( ▿ 2 | w H y | 2 ) g ( | w H y | 2 ) + 2 ( ▿ | w H y | 2 ) ( ▿ ≈ 2 E { g ( | w H y | 2 ) + | w H y | 2 g ′ ( | w H y | 2 ) } I - - - ( 7 )
β▽E{|wHy|2Jacobian matrix be β2E{|wHy|2}=2 β I, the therefore whole Jacobian matrix of formula (7) For:
J=2 (E{g (| wHy|2)+|wHy|2g′(|wHy|2)-β) I (8)
Utilizing newton (Newton) iterative method, the iterative formula that can obtain w is as follows:
w ′ = w - E { y ( w H y ) g ( | w H y | 2 ) } - βw E { g ( | w H y | 1 ) + | w H y | 2 g ′ ( | w H y | 5 ) } - β - - - ( 9 )
w new = w ′ | | w | |
Above formula is utilized to estimate the vectorial w in W matrix respectivelyi, i=1 ... n.In order to avoid finding same local Big value, needs to go phase with Glan-Schmidt-similar (Gram-Schmidt-like) in iterative process each time Close algorithm pairCarry out decorrelation.
Step 3: four road voice signals of step 1 are carried out multiple signal classification, utilizes MUSIC algorithm to estimate voice Signal source direction includes: four Mikes of step 1 are divided the signal collected after pretreatment, as multiple signal The input of sorting algorithm.The basic thought of multiple signal classification algorithm is to be carried out by the covariance matrix of array input signal Feature decomposition, obtains a signal subspace and an orthogonal to that noise subspace.Utilize voice signal and noise The character of Orthogonal Subspaces carrys out the incident direction of estimated speech signal.
The covariance matrix R of estimation array input signal:
RightCarry out Eigenvalues Decomposition, obtain noise subspace feature matrixThe guiding of signal subspace is vowed Amount aH(θ) withOrthogonal, but due to the existence of noise, both are the most orthogonal, can be made by searchMinimum θ is exactly corresponding source speech signal direction θ.
Step 4: as it is shown on figure 3, having an important problem in frequency domain independent composition analysis algorithm is exactly from separating letter Voice signal is extracted in number.In the present invention a length of the 1024 of Fourier transformation, therefore have 512 frequencies;Frequency domain Each frequency is implemented by independent composition analysis algorithm respectively, when separate two paths of signals n=2, for different frequent points i, J, i ≠ j, separates through frequency domain independent component analysis (ICA) algorithm, can obtain:
S ^ 1 ( i ) S ^ 2 ( i ) = W i y ( i ) S ^ 1 ( j ) S ^ 2 ( j ) = W j y ( j ) - - - ( 12 )
Because to the signal not priori of isolated, we judge where signal is voice signal, it is therefore desirable to FromIn determine which component is voice signal.Utilize the separation matrix W of each frequencyi, estimate two points Signal source direction from signal
θ ^ 1 ( i ) = arccos arg ( [ W i - 1 ] 11 / [ W - 1 ] 21 ) 2 π fc - 1 d - - - ( 13 )
θ ^ 2 ( i ) = arccos arg ( [ W i - 1 ] 21 / [ W - 1 ] 22 ) 2 π fc - 1 d - - - ( 14 )
Different frequent points is estimated that the direction obtained clusters, with reference to θ, selects the class in voice signal direction.If(i) be The direction of voice signal is thenI () is voice signal, combined by the voice signal of all frequencies and obtain voice signal. Concrete algorithm steps is as follows:
1) the separation matrix W of frequency f is utilizedfCalculating source speech signal direction, each frequency is calculated two letters Number direction, source(f) and(f), in an embodiment, a length of the 1024 of Fourier transformation, therefore altogether There are 512 frequencies;
2) above-mentioned all frequencies are estimated that the signal source direction obtained carries out K mean cluster, be polymerized to two classes, utilize Source speech signal direction θ, selects cluster the entering as voice signal near the θ of source speech signal direction Penetrate direction;
3) frequency with a low credibility is rejected: for same frequency, if one of them signal source direction differs with θ Within 10 °, and another signal source direction apart from the center of another class less than 20 °, then this frequency With a high credibility, the near signal source direction of chosen distance signal source direction θ is as source speech signal direction; Otherwise, it is determined that this frequency is inefficacy frequency;
4) for inefficacy frequency, the dependency of adjacent frequency spectrum envelope is utilized again to extract:
First, the frequency spectrum of inefficacy frequency is calculatedHarmonic structure with effective frequencyDependency it With cor (f), computing formula is:
cor ( f ) = Σ g s ^ i ( f ) s ^ i ( g ) | s ^ i ( f ) | | s ^ i ( g ) | ,
Wherein g=..., 1/3f, 1/2f, 2f, 3f ...,
Wherein g is effective frequency, if dependency sum cor (f) is more than 0.9, then extracts this frequency for effectively frequency Point;
5) above-mentioned iterative step is repeated, until iterations is more than 100 times, for final undrawn frequency, The frequency extracted being adjacent by described frequency calculates dependency, extracts.All frequencies The signal source direction that voice signal is corresponding isF () then separation matrix is constant, if the voice of this frequency The signal source direction that signal is corresponding isF () is then by separation matrix WfFirst row and secondary series exchange.
Step 5: amplitude smooths.To processing the new separation matrix obtained through step 4, between adjacent frequency, use one Individual window function carries out amplitude and smooths;Discrete frequency cyclicity refers to when frequency domain processes signal, L frequency of frequency domain, Sample frequency is fs, frequency interval is fs/ L, the signal at time-domain representation is to be L/f in the cyclesPeriodic signal, only When choosing a cycle, sharp thorn can be produced at cycle intersection.Applying wider a kind of algorithm is smooth spectrum.Pass through To the element w of separation matrixklF () is multiplied a window function, such as, Hanning window.
w ~ kl ( f ) = Σ φ = 0 f s - Δf g ( φ ) w kl ( f - φ ) - - - ( 15 )
Wherein, wklF () is separation matrix WfRow k l column element, g (f) is window function, Δ f=fs/L.If G (f) is Hanning window, then above formula can be expressed as follows:
w ~ kl ( f ) = 1 4 [ w kl ( f - Δf ) + 2 w kl ( f ) + w kl ( f + Δf ) ] .
WfIn element wklF () obtains new separation matrix through smooth combination afterwardsVoice can be calculated according to formula (4) Signal:
s 1 ( f ) s 2 ( f ) = W ~ f y 1 ( f ) y 2 ( f ) - - - ( 4 )
WhereinF) it is the voice signal estimating in step 4 to obtain,F () is noise signal.
Step 6: to estimating the voice signal that obtains in step 5Carry out inverse Fourier transform and windowing obtains voice The time-domain signal of signal.
To estimating that the voice signal that obtains carries out inverse Fourier transform:
s ^ i ( n ) = 1 L Σ f = 0 L - 1 s ^ 1 ( f ) * e j 2 π L fn - - - ( 16 )
To estimating that the voice signal obtained carries out windowing process, window function is identical with the window function in preprocessing process:
s ^ 1 ′ ( n ) = s ^ 1 ( n ) * w ( n ) - - - ( 17 )
For the enhancing voice finally given.
Embodiment
As it is shown in figure 5, at a long 10m, wide 5m, in the room of high 3m, as figure places microphone array and letter Number source, is at level height 1.1m, plan-position coordinate as it can be seen, the centre coordinate of microphone array is (5,2.5), Mike's spacing is 3cm, and source speech signal is positioned at (5,2.4), and other three signal sources are noise signal source: (2.5,4.5) believe Making an uproar the ratio white noise for 0dB, (2.4,0.6) signal to noise ratio is the white noise of 0dB, and (8,3.4) signal to noise ratio is to make an uproar in the pink of 5dB Sound.In that noise circumstance, gather signals with noise, utilize voice enhancement algorithm in this paper to strengthen and process, Wherein a length of the 1024 of Fourier transform, it is 512 that frame moves, i.e. L=1024, N=512 in Fig. 4.Utilize this algorithm After the enhancing obtained, the signal to noise ratio of voice is 13.5dB, utilizes traditional minimum variance undistorted response (MVDR) wave beam After the enhancing that molding algorithm obtains, voice signal to noise ratio is 9.6dB, it is seen that the noise suppression effect of algorithm is fine.

Claims (3)

1. four Mike's sound enhancement methods based on independent component analysis, it is characterised in that comprise the steps of
Step 1: use four Mikes to form rectangular array, gather four road Noisy Speech Signals, signal is carried out pretreatment, It is converted into frequency-region signal through Fourier transformation;
Step 2: randomly choose in four microphone arrays with on two paths of signals as first group of input signal, will be separately Outer two paths of signals is as second group of input signal, to often organizing pretreated frequency-region signal, utilizes frequency domain independent element to divide Analysis algorithm, estimates a separation matrix W respectively for each frequency f in frequency domainf, often group signal correspondence is included One road voice signal and the separation signal of a road noise signal;
Step 3: by the frequency-region signal obtained in step 1, utilize multiple signal classification algorithm to process, obtain language Tone signal source direction θ;
Step 4: using source speech signal direction θ calculated in step 3 as reference information, every group from step 2 Separate and signal extracts voice signal;
Step 5: carry out frequency sequence, the amplitude that carries out gained voice signal smooths, and utilizes a window function to adjacent frequency The signal of rate carries out amplitude and smooths;
Step 6: obtain the time-domain signal of voice signal through windowing and inverse Fourier transform.
2. according to four Mike's sound enhancement methods based on independent component analysis a kind of described in claim 1, its feature Be, in step 4, utilize the source speech signal direction θ estimating to obtain in step 3 as reference information, from step 2 Extract voice signal during often group separates signal to comprise the steps:
1) the separation matrix W of frequency f is utilizedfCalculating source speech signal direction, each frequency is calculated two letters Number direction, sourceWithFrequency number is the half of Fourier transformation length;
2) the above-mentioned calculated signal source direction of all frequencies is carried out K mean cluster, be polymerized to two classes, utilize Source speech signal direction θ, the cluster conduct within selecting source speech signal direction θ angle to differ 10 degree The incident direction of voice signal;
3) frequency with a low credibility is rejected: for same frequency, if the incident direction of one of them signal and language Within tone signal source direction θ differs 10 °, and the incident direction of its another signal clusters apart from another Center less than 20 °, then this frequency is with a high credibility, the language that chosen distance source speech signal direction θ is near The incident direction of tone signal is as final source speech signal direction;Otherwise, it is determined that this frequency was for losing efficacy Frequency;
4) for inefficacy frequency, the dependency of adjacent frequency spectrum envelope is utilized again to extract:
First, the frequency spectrum of inefficacy frequency is calculatedHarmonic structure with effective frequencyDependency it With cor (f), computing formula is:
c o r ( f ) = Σ g s ^ i ( f ) s ^ i ( g ) | s ^ i ( f ) | | s ^ i ( g ) | ,
Wherein g=..., 1/3f, 1/2f, 2f, 3f ...,
Wherein g is effective frequency,
If dependency sum cor (f) is more than 0.9, then extracting this frequency is effective frequency;
Above-mentioned iterative step is repeated, until iterations is more than 100 times, for final undrawn frequency Point, the frequency extracted being adjacent by described frequency calculates dependency, extracts.
3., according to a kind of based on independent component analysis the four Mike's sound enhancement methods described in right 1, its feature exists In, step 5 carrying out windowing to the voice signal extracted and smooths, use Hanning window, windowed function is as follows:
w ~ k l ( f ) = 1 4 [ w k l ( f - Δ f ) + 2 w k l ( f ) + w k l ( f + Δ f ) ] ,
Wherein wklF () is separation matrix WfIn a row k l row element, Δ f is between the frequency of Fourier transformation Every, equal to signal sampling rate divided by Fourier transformation length.
CN201410061180.4A 2014-02-24 2014-02-24 A kind of four Mike's sound enhancement methods based on independent component analysis Active CN103854660B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201410061180.4A CN103854660B (en) 2014-02-24 2014-02-24 A kind of four Mike's sound enhancement methods based on independent component analysis

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201410061180.4A CN103854660B (en) 2014-02-24 2014-02-24 A kind of four Mike's sound enhancement methods based on independent component analysis

Publications (2)

Publication Number Publication Date
CN103854660A CN103854660A (en) 2014-06-11
CN103854660B true CN103854660B (en) 2016-08-17

Family

ID=50862229

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201410061180.4A Active CN103854660B (en) 2014-02-24 2014-02-24 A kind of four Mike's sound enhancement methods based on independent component analysis

Country Status (1)

Country Link
CN (1) CN103854660B (en)

Families Citing this family (9)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN104200813B (en) * 2014-07-01 2017-05-10 东北大学 Dynamic blind signal separation method based on real-time prediction and tracking on sound source direction
CN106504763A (en) * 2015-12-22 2017-03-15 电子科技大学 Based on blind source separating and the microphone array multiple target sound enhancement method of spectrum-subtraction
CN105869627A (en) * 2016-04-28 2016-08-17 成都之达科技有限公司 Vehicle-networking-based speech processing method
US9741360B1 (en) * 2016-10-09 2017-08-22 Spectimbre Inc. Speech enhancement for target speakers
CN108074580B (en) * 2016-11-17 2021-04-02 杭州海康威视数字技术股份有限公司 Noise elimination method and device
CN109285557B (en) * 2017-07-19 2022-11-01 杭州海康威视数字技术股份有限公司 Directional pickup method and device and electronic equipment
CN111344778A (en) * 2017-11-23 2020-06-26 哈曼国际工业有限公司 Method and system for speech enhancement
CN109194422B (en) * 2018-09-04 2021-06-22 南京航空航天大学 SNR estimation method based on subspace
CN110786022A (en) * 2018-11-14 2020-02-11 深圳市大疆创新科技有限公司 Wind noise processing method, device and system based on multiple microphones and storage medium

Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN1826019A (en) * 2005-02-24 2006-08-30 索尼株式会社 Microphone apparatus
CN101517941A (en) * 2006-09-14 2009-08-26 美商富迪科技股份有限公司 Small array microphone apparatus and noise suppression methods thereof
CN102707262A (en) * 2012-06-20 2012-10-03 太仓博天网络科技有限公司 Sound localization system based on microphone array
JP5299233B2 (en) * 2009-11-20 2013-09-25 ソニー株式会社 Signal processing apparatus, signal processing method, and program
CN104459625A (en) * 2014-12-14 2015-03-25 南京理工大学 Sound source positioning device and method based on track moving double microphone arrays

Family Cites Families (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20080228470A1 (en) * 2007-02-21 2008-09-18 Atsuo Hiroe Signal separating device, signal separating method, and computer program

Patent Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN1826019A (en) * 2005-02-24 2006-08-30 索尼株式会社 Microphone apparatus
CN101517941A (en) * 2006-09-14 2009-08-26 美商富迪科技股份有限公司 Small array microphone apparatus and noise suppression methods thereof
JP5299233B2 (en) * 2009-11-20 2013-09-25 ソニー株式会社 Signal processing apparatus, signal processing method, and program
CN102707262A (en) * 2012-06-20 2012-10-03 太仓博天网络科技有限公司 Sound localization system based on microphone array
CN104459625A (en) * 2014-12-14 2015-03-25 南京理工大学 Sound source positioning device and method based on track moving double microphone arrays

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
基于四麦克风阵列的三维声源定位;王文龙 张艳萍;《南京信息工程大学学报》;20100531;全文 *

Also Published As

Publication number Publication date
CN103854660A (en) 2014-06-11

Similar Documents

Publication Publication Date Title
CN103854660B (en) A kind of four Mike's sound enhancement methods based on independent component analysis
CN109830245A (en) A kind of more speaker's speech separating methods and system based on beam forming
CN103426434B (en) Separated by the source of independent component analysis in conjunction with source directional information
CN106251877B (en) Voice Sounnd source direction estimation method and device
CN101667425A (en) Method for carrying out blind source separation on convolutionary aliasing voice signals
Kwan et al. An automated acoustic system to monitor and classify birds
CN106023996B (en) Sound recognition methods based on cross acoustic array broad-band EDFA
CN106782590A (en) Based on microphone array Beamforming Method under reverberant ambiance
CN108986838A (en) A kind of adaptive voice separation method based on auditory localization
CN102799892B (en) Mel frequency cepstrum coefficient (MFCC) underwater target feature extraction and recognition method
CN106057210B (en) Quick speech blind source separation method based on frequency point selection under binaural distance
CN102222508A (en) Matrix-transformation-based method for underdetermined blind source separation
CN105204001A (en) Sound source positioning method and system
CN106772331A (en) Target identification method and Target Identification Unit
CN109977724B (en) Underwater target classification method
CN109884591B (en) Microphone array-based multi-rotor unmanned aerial vehicle acoustic signal enhancement method
CN105580074B (en) Signal processing system and method
CN109637554A (en) MCLP speech dereverberation method based on CDR
Sainath et al. Reducing the Computational Complexity of Multimicrophone Acoustic Models with Integrated Feature Extraction.
Asaei Model-based sparse component analysis for multiparty distant speech recognition
CN110111806A (en) A kind of blind separating method of moving source signal aliasing
CN116559778B (en) Vehicle whistle positioning method and system based on deep learning
CN109741759A (en) A kind of acoustics automatic testing method towards specific birds species
CN114613384B (en) Deep learning-based multi-input voice signal beam forming information complementation method
Guo et al. Underwater target detection and localization with feature map and CNN-based classification

Legal Events

Date Code Title Description
C06 Publication
PB01 Publication
C10 Entry into substantive examination
SE01 Entry into force of request for substantive examination
C14 Grant of patent or utility model
GR01 Patent grant