CN109313909A - Assess method, unit and the system of microphone array consistency - Google Patents
Assess method, unit and the system of microphone array consistency Download PDFInfo
- Publication number
- CN109313909A CN109313909A CN201880001199.6A CN201880001199A CN109313909A CN 109313909 A CN109313909 A CN 109313909A CN 201880001199 A CN201880001199 A CN 201880001199A CN 109313909 A CN109313909 A CN 109313909A
- Authority
- CN
- China
- Prior art keywords
- microphone
- signal
- audio signal
- frequency
- reference microphone
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Granted
Links
- 238000000034 method Methods 0.000 title claims abstract description 72
- 238000001228 spectrum Methods 0.000 claims abstract description 196
- 230000005236 sound signal Effects 0.000 claims abstract description 159
- 238000011156 evaluation Methods 0.000 claims abstract description 14
- 238000012545 processing Methods 0.000 claims description 49
- 238000012360 testing method Methods 0.000 claims description 32
- 241000209140 Triticum Species 0.000 claims description 19
- 235000021307 Triticum Nutrition 0.000 claims description 19
- 239000011159 matrix material Substances 0.000 claims description 19
- 230000017105 transposition Effects 0.000 claims description 19
- 238000009432 framing Methods 0.000 claims description 15
- 230000008569 process Effects 0.000 claims description 14
- 230000007613 environmental effect Effects 0.000 claims description 8
- 230000009467 reduction Effects 0.000 claims description 6
- 230000001960 triggered effect Effects 0.000 claims description 3
- 238000004422 calculation algorithm Methods 0.000 abstract description 28
- 230000002708 enhancing effect Effects 0.000 abstract description 21
- 238000013461 design Methods 0.000 description 11
- 238000010586 diagram Methods 0.000 description 9
- 238000005516 engineering process Methods 0.000 description 7
- 230000006870 function Effects 0.000 description 7
- 238000004891 communication Methods 0.000 description 6
- 241000208340 Araliaceae Species 0.000 description 5
- 235000005035 Panax pseudoginseng ssp. pseudoginseng Nutrition 0.000 description 5
- 235000003140 Panax quinquefolius Nutrition 0.000 description 5
- 235000008434 ginseng Nutrition 0.000 description 5
- 230000003595 spectral effect Effects 0.000 description 4
- 238000010168 coupling process Methods 0.000 description 3
- 238000005859 coupling reaction Methods 0.000 description 3
- 238000001514 detection method Methods 0.000 description 3
- 230000000694 effects Effects 0.000 description 3
- 239000004615 ingredient Substances 0.000 description 3
- 241001413866 Diaphone Species 0.000 description 2
- 238000004364 calculation method Methods 0.000 description 2
- 238000010276 construction Methods 0.000 description 2
- 230000008878 coupling Effects 0.000 description 2
- 230000002452 interceptive effect Effects 0.000 description 2
- 230000009466 transformation Effects 0.000 description 2
- 230000005526 G1 to G0 transition Effects 0.000 description 1
- 230000008859 change Effects 0.000 description 1
- 238000004590 computer program Methods 0.000 description 1
- 230000010485 coping Effects 0.000 description 1
- 230000003247 decreasing effect Effects 0.000 description 1
- 230000003993 interaction Effects 0.000 description 1
- 238000005192 partition Methods 0.000 description 1
- 230000008447 perception Effects 0.000 description 1
- 238000004445 quantitative analysis Methods 0.000 description 1
- 230000004044 response Effects 0.000 description 1
- 238000000926 separation method Methods 0.000 description 1
Classifications
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04R—LOUDSPEAKERS, MICROPHONES, GRAMOPHONE PICK-UPS OR LIKE ACOUSTIC ELECTROMECHANICAL TRANSDUCERS; DEAF-AID SETS; PUBLIC ADDRESS SYSTEMS
- H04R29/00—Monitoring arrangements; Testing arrangements
- H04R29/001—Monitoring arrangements; Testing arrangements for loudspeakers
Landscapes
- Health & Medical Sciences (AREA)
- General Health & Medical Sciences (AREA)
- Otolaryngology (AREA)
- Physics & Mathematics (AREA)
- Engineering & Computer Science (AREA)
- Acoustics & Sound (AREA)
- Signal Processing (AREA)
- Circuit For Audible Band Transducer (AREA)
- Measurement Of Velocity Or Position Using Acoustic Or Ultrasonic Waves (AREA)
Abstract
The embodiment of the present application provides a kind of method, unit and system for assessing microphone array consistency, the consistency in microphone array between different microphones can be assessed, to instruct the calibration of microphone array and the robustness of assessment multichannel enhancing algorithm according to compliance evaluation result, user experience is promoted.This method comprises: obtaining N number of audio signal that N number of microphone acquires respectively, which constitutes microphone array, N >=2;According to N number of audio signal, each microphone in N number of microphone in addition to reference microphone and phase spectrum difference and/or power spectrum difference between the reference microphone are determined, which is any one microphone in N number of microphone;According to each microphone in N number of microphone in addition to the reference microphone and phase spectrum difference and/or power spectrum difference between the reference microphone, compliance evaluation is carried out to N number of microphone.
Description
Technical field
This application involves speech communications and speech-sound intelligent interaction field, and more particularly, to assessment microphone array
Method, unit and the system of consistency.
Background technique
In speech communication application, speech enhancement technique can be improved the auditory perception of people, improve understanding for speech communication
Degree, in speech-sound intelligent interactive application, speech enhancement technique can be improved the accuracy rate of speech recognition, promote user experience, because
This speech enhancement technique is either crucial in traditional speech communication or interactive voice.Speech enhancement technique point
Enhancing technology and multicenter voice for single-channel voice enhances technology, wherein single-channel voice enhancing technology can eliminate stable state
Noise cannot eliminate nonstationary noise, and signal is using speech damage as cost than improving, and signal-to-noise ratio improves more, voice damage
Hurt bigger;Multicenter voice enhances technology and acquires multiple signals using microphone array, utilizes the phase between multi-microphone signal
Position information and coherence messages eliminate noise, can eliminate nonstationary noise, and smaller to speech damage.
In multicenter voice enhancing technology, the consistency in microphone array between different microphones directly affects algorithm
Performance, existing scheme propose the innovatory algorithm of multichannel enhancing technology, increase the robustness of algorithm, while between microphone
Coherence request reduce, however, the consistency between microphone still will affect algorithm performance when very low, to affect use
Family experience.
Summary of the invention
The application provides a kind of method, unit and system for assessing microphone array consistency, can assess Mike
Consistency in wind array between different microphones, to instruct the calibration of microphone array according to compliance evaluation result and comment
The robustness for estimating multichannel enhancing algorithm, promotes user experience.
In a first aspect, providing a kind of method for assessing microphone array consistency, comprising:
N number of audio signal that N number of microphone acquires respectively is obtained, which constitutes microphone array, N >=2;
According to N number of audio signal, each microphone and the ginseng in N number of microphone in addition to reference microphone are determined
The phase spectrum difference and/or power spectrum difference between microphone are examined, which is any one in N number of microphone
Microphone;
According to each microphone in N number of microphone in addition to the reference microphone and the phase between the reference microphone
Position spectral difference value and/or power spectrum difference carry out compliance evaluation to N number of microphone.
It should be noted that carrying out compliance evaluation to N number of microphone, it can be used for instructing the wheat in microphone array
The distribution of gram wind, or guidance redesign the microphone distribution in microphone array, or guidance redesigns microphone array
Column, or the robustness of assessment multichannel enhancing algorithm.
For example, when assessment result shows that the consistency of microphone 1 and microphone 2 is poor, can instruct adjustment microphone 1 or
Distribution of the person's microphone 2 in microphone array can either be instructed to redesign microphone 1 or microphone 2.
In another example adjustment wheat can be instructed when assessment result shows that the consistency of microphone 1 and multiple microphones is all poor
Gram distribution of the wind 1 in microphone array can perhaps be instructed to redesign microphone 1 or can instruct to redesign wheat
Gram wind array.
In the embodiment of the present application, the N number of audio signal acquired respectively according to N number of microphone, determine each microphone with
Phase spectrum difference and/or power spectrum difference between reference microphone are eliminated to carry out compliance evaluation to N number of microphone
Influence of the consistency to multicenter voice enhancing algorithm between microphone, promotes user experience.
In some possible implementations, each wheat according in N number of microphone in addition to reference microphone
Phase spectrum difference gram between wind and the reference microphone carries out compliance evaluation to N number of microphone, comprising:
According to each microphone in N number of microphone in addition to the reference microphone and the phase between the reference microphone
Position spectral difference value, assesses the phase equalization between corresponding microphone and the reference microphone.
It should be noted that the phase spectrum difference between two microphones is smaller, the phase between the two microphones is indicated
Bit integrity is better.
For example, the phase spectrum difference between microphone 1 and reference microphone is A, A is smaller, indicates microphone 1 and refers to wheat
Phase equalization between gram wind is better.
It is alternatively possible to a threshold value be arranged, if the phase spectrum difference between two microphones is less than this threshold value, table
Show that the phase equalization between the two microphones meets design requirement, the consistency between the two microphones is to multichannel language
Sound enhancing algorithm influence can ignore or the two microphones between consistency to multicenter voice enhancing algorithm do not have
It influences.
It should be noted that above-mentioned threshold value can enhance algorithm flexible configuration according to different multicenter voices.
In some possible implementations, this method further include:
The each microphone and the reference microphone to sound in N number of microphone in addition to the reference microphone are measured respectively
The range difference in source;
According to measured range difference, each Mike in N number of microphone in addition to the reference microphone is calculated separately
Fixed skew between wind and the reference microphone;
According to each microphone in N number of microphone in addition to the reference microphone and consolidating between the reference microphone
Phase bit is poor, calibrates its corresponding phase spectrum difference respectively.
For example, the fixed skew between microphone 1 and reference microphone is A, between microphone 1 and reference microphone
Phase spectrum difference is B, and after calibration, the phase spectrum difference between microphone 1 and reference microphone is C, at this point, C=B-A.
In some possible implementations, the distance according to measured by is calculated separately to remove in N number of microphone and is somebody's turn to do
The fixed skew between each microphone and the reference microphone except reference microphone, comprising:
According to formulaIt calculates separately in N number of microphone in addition to the reference microphone
Fixed skew between each microphone and the reference microphone,
Wherein, Yi(ω) indicates the frequency spectrum of i-th of microphone, Y1(ω) indicates that the frequency spectrum of reference microphone, ω indicate frequency
Rate, diIndicate the range difference of i-th of microphone and reference microphone to sound source, the c expression velocity of sound, 2 π ω di/ c indicates i-th of Mike
Fixed skew between wind and reference microphone.
In some possible implementations, each wheat according in N number of microphone in addition to reference microphone
Phase spectrum difference gram between wind and the reference microphone carries out compliance evaluation to N number of microphone, comprising:
According to each microphone in N number of microphone in addition to the reference microphone and the function between the reference microphone
Rate spectral difference value assesses the amplitude coincidence between corresponding microphone and the reference microphone.
It should be noted that the power spectrum difference between two microphones is smaller, the width between the two microphones is indicated
It is better to spend consistency.
For example, the power spectrum difference between microphone 1 and reference microphone is A, A is smaller, indicates microphone 1 and refers to wheat
Amplitude coincidence between gram wind is better.
It is alternatively possible to a threshold value be arranged, if the power spectrum difference between two microphones is less than this threshold value, table
Show that the amplitude coincidence between the two microphones meets design requirement, the consistency between the two microphones is to multichannel language
Sound enhancing algorithm influence can ignore or the two microphones between consistency to multicenter voice enhancing algorithm do not have
It influences.
It should be noted that above-mentioned threshold value can enhance algorithm flexible configuration according to different multicenter voices.
In some possible implementations, when carrying out phase equalization assessment, which swept in broadcasting
The signal acquired in the environment of frequency signal data.
In some possible implementations, when carrying out amplitude coincidence assessment, which is to play height
The signal acquired in the environment of this white noise data or swept-frequency signal data.
In some possible implementations, which is Linear chirp, logarithm swept-frequency signal, linear stepping
Any one in swept-frequency signal, logarithm stepping swept-frequency signal.
It is described according to N number of audio signal in some possible implementations, it determines in N number of microphone except reference
The phase spectrum difference and/or power spectrum difference between each microphone and the reference microphone except microphone, comprising:
Each audio signal in N number of audio signal is subjected to framing, obtains K signal frame of equal length, K >=2;
Windowing process is done to each signal frame in the K signal frame, obtains K windowing signal frame;
Fast Fourier transform (Fast Fourier is done to each windowing signal frame in the K windowing signal frame
Transformation, FFT) transformation, obtain K echo signal frame;
According to the corresponding K echo signal frame of each audio signal, determine in N number of microphone except this is with reference to Mike
The phase spectrum difference and/or power spectrum difference between each microphone and the reference microphone except wind.
Optionally, K indicates that each microphone collects the totalframes of signal.
It should be noted that bring truncation effect when windowing process is used to eliminate framing.It is alternatively possible to be to the K
Each signal frame in a signal frame is done plus Hamming window processing.
In some possible implementations, any two adjacent signals frame is overlapped R%, R > 0 in the K signal frame.Example
Such as, which is 25 or 50.
Optionally, signal amplitude remains unchanged after being overlapped adding window.
It should be understood that each frame signal after overlapping has the ingredient of previous frame, prevent discontinuous between two frames.
In some possible implementations, i-th of audio signal is subjected to framing, obtains K signal of equal length
Frame is write as following vector form:
xi(t)=[xi,1(t),xi,2(t),…,xi,K(t)]T
Wherein, xi(t) i-th of audio signal is indicated, K indicates that each microphone collects the totalframes of signal, []TIt indicates
The transposition of vector or matrix.
It is described according to the corresponding K echo signal frame of each audio signal in some possible implementations, really
Each microphone in addition to the reference microphone and the phase spectrum difference between the reference microphone in fixed N number of microphone,
Include:
According to formulaIt determines in N number of microphone in addition to the reference microphone
Each microphone and the reference microphone between phase spectrum difference,
Wherein, imag () expression takes imaginary part, and ln () expression takes natural logrithm,Indicate i-th of microphone and ginseng
The phase spectrum difference between microphone is examined,Indicate j-th of echo signal frame of reference microphone,Indicate i-th
J-th of echo signal frame of a microphone,Indicate basic frequency.
It is described according to the corresponding K echo signal frame of each audio signal in some possible implementations, really
Each microphone in addition to the reference microphone and the power spectrum difference between the reference microphone in fixed N number of microphone,
Include:
According to the corresponding K echo signal frame of each audio signal, the power spectrum of each audio signal is determined;
According to the power spectrum of each audio signal, determine in N number of microphone each of in addition to the reference microphone
Power spectrum difference between microphone and the reference microphone.
It is described according to the corresponding K echo signal frame of each audio signal in some possible implementations, really
The power spectrum of fixed each audio signal, comprising:
According to formulaThe power spectrum of each audio signal is calculated,
Wherein, Pi(ω) indicates the power spectrum of i-th of audio signal, Yi,j(ω) indicates the jth in i-th of audio signal
A echo signal frame, K indicate that each microphone receives the totalframes of signal, and ω indicates frequency.
In some possible implementations, the power spectrum according to each audio signal determines N number of microphone
In each microphone in addition to the reference microphone and the power spectrum difference between the reference microphone, comprising:
According to formula PDi(ω)=P1(ω)-Pi(ω) is calculated in N number of microphone each of in addition to reference microphone
Power spectrum difference between microphone and the reference microphone,
Wherein, PDi(ω) indicates the power spectrum difference between i-th of microphone and reference microphone, P1(ω) indicates reference
The power spectrum of microphone, Pi(ω) indicates the power spectrum of i-th of microphone.
In some possible implementations, the N number of audio signal for obtaining N number of microphone and acquiring respectively, comprising:
Determine sample frequency F of the N number of microphone when carrying out audio signal samplesWith FFT points Nfft, use loudspeaking
Device plays white Gaussian noise data or swept-frequency signal data, N number of microphone acquire N number of audio signal, wherein if this is raised
The data that sound device is played are swept-frequency signal data, and the swept-frequency signal data are equal by M+1 segment length and the signal of frequency not etc.
It constitutes,
It should be noted that FFT points NfftFor even number, generally 32,64,128 ..., 1024 etc., it counts more, fortune
The saving of calculation amount is bigger.
In some possible implementations, according to formulaCalculate the frequency of every segment signal in the M+1 segment signal
Rate, and
According to formula Si(t)=sin (2 π fiT) every segment signal in the M+1 segment signal is calculated,
Wherein, fiIndicate the frequency of the i-th segment signal, FsIndicate sample frequency, NfftIndicate FFT points, Si(t) i-th is indicated
Segment signal, and S1(t) length is the integral multiple of cycle T, T=1/f1。
In some possible implementations, the swept-frequency signal data that loudspeaker is played can be write as following vector shape
Formula:
S (t)=[S0(t),S1(t),…,SM(t)]T
Wherein, S (t) indicates the swept-frequency signal data that loudspeaker is played, Si(t) the i-th segment signal is indicated,
[ ]TIndicate the transposition of vector or matrix.
In some possible implementations, which collects N number of audio signal respectively, wherein i-th of Mike
The collected audio signal of wind is expressed as xiAnd x (t),i(t) it can be write as following vector form:
xi(t)=[xi,1(t),xi,2(t),…,xi,K(t)]T
Wherein, xi(t) the collected audio signal of i-th of microphone is indicated, K indicates that each microphone collects signal
Totalframes, []TIndicate the transposition of vector or matrix.
In some possible implementations, the N number of audio signal for obtaining N number of microphone and acquiring respectively, comprising:
N number of microphone is placed in test room, loudspeaker, N number of microphone position are configured in the test room
In the front of the loudspeaker;
It controls the loudspeaker and plays white Gaussian noise data or swept-frequency signal data, and control N number of microphone point
N number of audio signal is not acquired.
In some possible implementations, there is noise reduction room environmental in the test room, which is audio-frequency test
Special artificial mouth, and the artificial mouth is calibrated with standard microphone before the use.
In some possible implementations, white Gaussian noise data or swept-frequency signal number are played controlling the loudspeaker
According to before, this method further include:
Under quiet environment, N number of microphone is obtained in the first duration T1First audio data X of interior acquisition1(n);
In the environment of playing white Gaussian noise data or swept-frequency signal data, N number of microphone is obtained at second
The second audio data X acquired in long T22(n);
According to formulaSignal to Noise Ratio (SNR) is calculated, and ensures that the SNR is greater than first threshold.
Second aspect provides a kind of equipment for assessing microphone array consistency, comprising:
Acquiring unit, the N number of audio signal acquired respectively for obtaining N number of microphone, N number of microphone constitute Mike
Wind array, N >=2;
Processing unit, for determining in N number of microphone in addition to reference microphone according to N number of audio signal
Each microphone and the reference microphone between phase spectrum difference and/or power spectrum difference, the reference microphone is
Any one microphone in N number of microphone;
The processing unit is also used to according to each Mike in N number of microphone in addition to the reference microphone
Phase spectrum difference and/or power spectrum difference between wind and the reference microphone carry out consistency to N number of microphone and comment
Estimate.
In some possible implementations, the processing unit is specifically used for:
According in N number of microphone in addition to the reference microphone each microphone and the reference microphone it
Between phase spectrum difference, assess the phase equalization between corresponding microphone and the reference microphone.
In some possible implementations, the processing unit is also used to:
Each microphone in N number of microphone in addition to the reference microphone is measured respectively with described with reference to Mike
Range difference of the wind to sound source;
According to measured range difference, calculate separately in N number of microphone each of in addition to the reference microphone
Fixed skew between microphone and the reference microphone;
According in N number of microphone in addition to the reference microphone each microphone and the reference microphone it
Between fixed skew, calibrate its corresponding phase spectrum difference respectively.
In some possible implementations, the processing unit is specifically used for:
According to formulaCalculate separately in N number of microphone except the reference microphone it
Fixed skew between outer each microphone and the reference microphone,
Wherein, Yi(ω) indicates the frequency spectrum of i-th of microphone, Y1(ω) indicates that the frequency spectrum of reference microphone, ω indicate frequency
Rate, diIndicate the range difference of i-th of microphone and reference microphone to sound source, the c expression velocity of sound, 2 π ω di/ c indicates i-th of Mike
Fixed skew between wind and reference microphone.
In some possible implementations, the processing unit is specifically used for:
According in N number of microphone in addition to the reference microphone each microphone and the reference microphone it
Between power spectrum difference, assess the amplitude coincidence between corresponding microphone and the reference microphone.
In some possible implementations, N number of audio signal is adopted in the environment of playing swept-frequency signal data
The signal of collection.
In some possible implementations, N number of audio signal is to play white Gaussian noise data or frequency sweep
The signal acquired in the environment of signal data.
In some possible implementations, the swept-frequency signal is Linear chirp, logarithm swept-frequency signal, linear step
Into any one in swept-frequency signal, logarithm stepping swept-frequency signal.
In some possible implementations, the processing unit is specifically used for:
Each audio signal in N number of audio signal is subjected to framing, obtains K signal frame of equal length, K >=
2;
Windowing process is done to each signal frame in the K signal frame, obtains K windowing signal frame;
FFT transform is done to each windowing signal frame in the K windowing signal frame, obtains K echo signal frame;
According to the corresponding K echo signal frame of each audio signal, determine in N number of microphone except described
The phase spectrum difference and/or power spectrum difference between each microphone and the reference microphone except reference microphone.
In some possible implementations, any two adjacent signals frame is overlapped R%, R > 0 in the K signal frame.
In some possible implementations, the R is 25 or 50.
In some possible implementations, i-th of audio signal is subjected to framing, obtains K signal of equal length
Frame is write as following vector form:
xi(t)=[xi,1(t),xi,2(t),…,xi,K(t)]T
Wherein, xi(t) i-th of audio signal is indicated, K indicates that each microphone collects the totalframes of signal, []TIt indicates
The transposition of vector or matrix.
In some possible implementations, the processing unit is specifically used for:
According to formulaIt determines in N number of microphone except the reference microphone
Except each microphone and the reference microphone between phase spectrum difference,
Wherein, imag () expression takes imaginary part, and ln () expression takes natural logrithm,Indicate i-th of microphone and ginseng
The phase spectrum difference between microphone is examined,Indicate j-th of echo signal frame of reference microphone,Indicate i-th
J-th of echo signal frame of a microphone,Indicate basic frequency.
In some possible implementations, the processing unit is specifically used for:
According to the corresponding K echo signal frame of each audio signal, the function of each audio signal is determined
Rate spectrum;
According to the power spectrum of each audio signal, determine in N number of microphone in addition to the reference microphone
Each microphone and the reference microphone between power spectrum difference.
In some possible implementations, the processing unit is specifically used for:
According to formulaThe power spectrum of each audio signal is calculated,
Wherein, Pi(ω) indicates the power spectrum of i-th of audio signal, Yi,j(ω) indicates the jth in i-th of audio signal
A echo signal frame, K indicate that each microphone collects the totalframes of signal, and ω indicates frequency.
In some possible implementations, the processing unit is specifically used for:
According to formula PDi(ω)=P1(ω)-Pi(ω) calculates every in addition to reference microphone in N number of microphone
Power spectrum difference between a microphone and the reference microphone,
Wherein, PDi(ω) indicates the power spectrum difference between i-th of microphone and reference microphone, P1(ω) indicates reference
The power spectrum of microphone, Pi(ω) indicates the power spectrum of i-th of microphone.
In some possible implementations, the processing unit is specifically used for:
Determine sample frequency F of the N number of microphone when carrying out audio signal samplesWith FFT points Nfft, using raising
Sound device plays white Gaussian noise data or swept-frequency signal data, controls N number of microphone and acquires N number of audio signal,
Wherein, if the data that are played of the loudspeaker are swept-frequency signal data, the swept-frequency signal data it is equal by M+1 segment length and
Equal signal is not constituted frequency,
In some possible implementations, the processing unit is also used to:
According to formulaThe frequency of every segment signal in the M+1 segment signal is calculated, and
According to formula Si(t)=sin (2 π fiT) every segment signal in the M+1 segment signal is calculated,
Wherein, fiIndicate the frequency of the i-th segment signal, FsIndicate sample frequency, NfftIndicate FFT points, Si(t) i-th is indicated
Segment signal, and S1(t) length is the integral multiple of cycle T, T=1/f1。
In some possible implementations, the swept-frequency signal data that the loudspeaker is played are write as following vector shape
Formula:
S (t)=[S0(t),S1(t),…,SM(t)]T
Wherein, S (t) indicates the swept-frequency signal data that loudspeaker is played, Si(t) the i-th segment signal is indicated,
[ ]TIndicate the transposition of vector or matrix.
In some possible implementations, N number of microphone collects N number of audio signal respectively, wherein i-th of wheat
Gram collected audio signal of wind is expressed as xiAnd x (t),i(t) it can be write as following vector form:
xi(t)=[xi,1(t),xi,2(t),…,xi,K(t)]T
Wherein, xi(t) the collected audio signal of i-th of microphone is indicated, K indicates that each microphone collects signal
Totalframes, []TIndicate the transposition of vector or matrix.
In some possible implementations, the acquiring unit is specifically used for:
N number of microphone is placed in test room, loudspeaker, N number of wheat are configured in the test room
Gram wind is located at the front of the loudspeaker;
It controls the loudspeaker and plays white Gaussian noise data or swept-frequency signal data, and control N number of Mike
Wind acquires N number of audio signal respectively.
In some possible implementations, there is noise reduction room environmental in the test room, the loudspeaker is audio
Special artificial mouth is tested, and the artificial mouth is calibrated with standard microphone before the use.
In some possible implementations, the loudspeaker is controlled in the processing unit and plays white Gaussian noise data
Or before swept-frequency signal data, the acquiring unit is also used to:
Under quiet environment, N number of microphone is obtained in the first duration T1First audio data X of interior acquisition1(n);
In the environment of playing white Gaussian noise data or swept-frequency signal data, N number of microphone is obtained second
Duration T2The second audio data X of interior acquisition2(n);
The processing unit is triggered according to formulaSignal to Noise Ratio (SNR) is calculated, and is ensured described
SNR is greater than first threshold.
The third aspect provides a kind of device for assessing microphone array consistency, comprising:
Memory, for storing program and data;And
Processor, for calling and running the program and data that store in the memory;
The device is configured as executing the method in above-mentioned first aspect or its any possible implementation.
Fourth aspect provides the system of assessment microphone array consistency, comprising:
Constitute N number of microphone of microphone array, N >=2;
At least one audio-source;
Device stores in the memory including the memory for storing program and data and for calling and running
The processor of program and data, the device are configured as the method in above-mentioned first aspect or its any possible implementation.
5th aspect, provides a kind of computer storage medium, is stored with program code in the computer storage medium, should
Program code can serve to indicate that the method executed in above-mentioned first aspect or its any possible implementation.
6th aspect, provides a kind of computer program product comprising instruction, when running on computers, makes to succeed in one's scheme
Calculation machine executes the method in above-mentioned first aspect or its any possible implementation.
Detailed description of the invention
Fig. 1 is the schematic flow chart of the method for the assessment microphone array consistency of the embodiment of the present application.
Fig. 2 is the test environment schematic according to the embodiment of the present application.
Fig. 3 is the schematic diagram according to the calculating phase spectrum difference of the embodiment of the present application.
Fig. 4 is the schematic diagram according to the calculating power spectrum difference of the embodiment of the present application.
Fig. 5 is the schematic diagram of the phase spectrum difference between two microphones according to the embodiment of the present application.
Fig. 6 is the schematic diagram of the phase spectrum difference after calibrating between two microphones according to the embodiment of the present application.
Fig. 7 a is the schematic diagram according to the power spectrum of two microphones of the embodiment of the present application.
Fig. 7 b is the schematic diagram of the power spectrum difference between two microphones according to the embodiment of the present application.
Fig. 8 is the schematic diagram according to a kind of equipment of assessment microphone array consistency of the embodiment of the present application.
Fig. 9 is the schematic diagram according to a kind of device of assessment microphone array consistency of the embodiment of the present application.
Figure 10 is the schematic diagram according to a kind of system of assessment microphone array consistency of the embodiment of the present application.
Specific embodiment
Below in conjunction with the attached drawing in the embodiment of the present application, technical solutions in the embodiments of the present application is clearly retouched
It states.
Microphone array (Microphone Array) refers to be made of the microphone (acoustic sensor) of certain amount, is used
Come the system that the spatial character of sound field is sampled and handled.The difference between the phase of sound wave is received using two microphones
It is different that sound wave is filtered, environmental background sound can be disposed to greatest extent, only the remaining sound wave needed.
Multicenter voice enhancing technique algorithm assumed condition be multiple microphones in microphone array target voice at
Divide high correlation, target voice is uncorrelated to non-targeted interference, therefore the consistency in microphone array between different microphones
Directly affect algorithm performance.
The qualitative assessment of microphone consistency can be used for instructing the design of microphone and the design of microphone array, Mike
Circuit, electronic component, the acoustic construction of wind array can all influence the consistency of microphone, can be by when designing microphone array
Influence of the item test various factors to consistency, so that the design of microphone consistency be made to reach system requirements.
The qualitative assessment of microphone consistency can be used for comparing the robustness of algorithms of different, reach same voice enhancing
The premise of performance, lower to coincident indicator requirement, algorithm robustness is better.
In the embodiment of the present application, consistency is measured in terms of amplitude spectrum difference and phase spectrum difference two, it is objective to have
Property and accuracy, and quantitative method for assessing consistency can objectively instruct the design of microphone array, it also can be objective
Comparison multicenter voice enhancing algorithm robustness.
Hereinafter, the method for the assessment microphone array consistency of the embodiment of the present application is discussed in detail in conjunction with Fig. 1 to Fig. 7.
Fig. 1 is the schematic flow chart of the method for the assessment microphone array consistency of the application one embodiment.Ying Li
The step of solution, Fig. 1 shows this method or operation, but these steps or operation are only examples, and the embodiment of the present application can also be held
The deformation of other operations of row or each operation in Fig. 1.This method can be held by the device of assessment microphone array consistency
Row, wherein the device of the assessment microphone array consistency can be mobile phone, tablet computer, portable computer, individual digital and help
Manage (Personal Digital Assistant, PDA) etc..
S110 obtains N number of audio signal that N number of microphone acquires respectively, N number of microphone composition microphone array, and N >=
2。
When carrying out compliance evaluation to N number of microphone, need to limit environment locating for N number of microphone, i.e. N number of audio
Signal is acquired under special test environment.
Specifically, as shown in Fig. 2, the microphone array 201 being made of N number of microphone is placed in test room 202
It is interior, and loudspeaker 203 is configured in the test room 202, which is being particularly located at the loudspeaker 203 just
Front, the microphone array 201 connect the control equipment 204 of such as computer with the loudspeaker 203.The control equipment 204 can
Specific audio data is played to control the loudspeaker 203, for example, white Gaussian noise data or swept-frequency signal data are played,
Meanwhile the control equipment 204 can obtain N number of audio letter of N number of microphone distribution collection from the microphone array 201
Number.
It should be noted that microphone compliance evaluation requires the signal-to-noise ratio of the audio signal of acquisition sufficiently high, background is made an uproar
Sound is weak enough, therefore tests environmental requirement under quiet environment.Particularly, noise reduction room environmental is required in test room 202.
Loudspeaker 203 requires noise relatively high, and frequency response curve is flat, and particularly, loudspeaker uses audio-frequency test Special artificial
Mouth, and calibrated before use with standard microphone.Microphone array 201 is placed on the front of loudspeaker 203, particularly,
Ask the position for being placed on standard microphone calibration.
Optionally, before carrying out formal audio signal sample, it is also necessary to carry out signal-to-noise ratio to above-mentioned test environment
(signal-to-noise ratio, SNR) detection.
Specifically, under test environment as shown in Figure 2, firstly, (i.e. loudspeaker 203, which is in, closes under quiet environment
Closed state), N number of microphone is obtained in the first duration T1First audio data X of interior acquisition1(n);Then, white Gaussian is being played
(i.e. the control equipment 204 controls the loudspeaker 203 and plays white Gaussian noise in the environment of noise data or swept-frequency signal data
Data or swept-frequency signal data), N number of microphone is obtained in the second duration T2The second audio data X of interior acquisition2(n);It connects
, 1 calculate SNR according to the following formula;Finally, then detection passes through when SNR is greater than given threshold, otherwise detects and do not pass through.
Wherein, T1Indicate the first duration, T2Indicate the second duration, X1(n) the first audio data, X are indicated2(n) second is indicated
Audio data.
It should be noted that if detection does not pass through, need that above-mentioned test environment is adjusted or is calibrated, eliminate
It may be to the factor that signal-to-noise ratio impacts, until being greater than given threshold according to the SNR calculated of above-mentioned formula 1.
It optionally, in the embodiment of the present application, specifically can be with using above-mentioned test environment shown in Fig. 2 acquisition audio signal
Include:
Determine sample frequency F of the N number of microphone when carrying out audio signal samplesWith FFT points Nfft, use loudspeaking
Device plays white Gaussian noise data or swept-frequency signal data, N number of microphone acquire N number of audio signal.
Optionally, FFT points NfftFor even number, generally 32,64,128 ..., 1024 etc., it counts more, operand
It saves bigger.
It should be noted that if the data that the loudspeaker is played are swept-frequency signal data, the swept-frequency signal data are by M+1
Segment length is equal and the signal of frequency not etc. is constituted,
It is alternatively possible to 2 frequency for calculating every segment signal in the M+1 segment signal according to the following formula, and according to as follows
Formula 3 calculates every segment signal in the M+1 segment signal.
Wherein, fiIt is the frequency of the i-th segment signal, FsIt is sample frequency, NfftIndicate FFT points.
Si(t)=sin (2 π fiT) formula 3
Wherein, Si(t) the i-th segment signal, f are indicatediIt is the frequency of the i-th segment signal.
It should be noted that the first segment signal S1(t) length is the integral multiple of cycle T, T=1/f1。
Optionally, the swept-frequency signal data that loudspeaker is played can be write as following vector form:
S (t)=[S0(t),S1(t),…,SM(t)]T
Wherein, S (t) indicates the swept-frequency signal data that loudspeaker is played, Si(t) the i-th segment signal is indicated,
[ ]TIndicate the transposition of vector or matrix.
Optionally, N number of microphone collects N number of audio signal respectively, wherein the collected audio signal of i-th of microphone
It is expressed as xiAnd x (t),i(t) it can be write as following vector form:
xi(t)=[xi,1(t),xi,2(t),…,xi,K(t)]T
Wherein, xi(t) the collected audio signal of i-th of microphone is indicated, K indicates that each microphone collects signal
Totalframes, []TIndicate the transposition of vector or matrix.
S120 determines each microphone in N number of microphone in addition to reference microphone according to N number of audio signal
Phase spectrum difference and/or power spectrum difference between the reference microphone, the reference microphone are appointing in N number of microphone
It anticipates a microphone.
Optionally, in the embodiment of the present application, after N number of audio signal sample arrives, audio signal point can be passed through
Frame does FFT transform to every frame windowing signal, seeks the phase spectrum difference between different microphones to every frame audio signal adding window.
Specifically, as shown in Figure 3, it is assumed that N number of audio signal is x1(t),x2(t),…,xN(t), which is believed
Each audio signal in number carries out framing, obtains K signal frame of equal length, K >=2, for example, by i-th of audio signal
Framing is carried out, K signal frame for obtaining equal length is write as following vector form:
xi(t)=[xi,1(t),xi,2(t),…,xi,K(t)]T
Wherein, xi(t) i-th of audio signal is indicated, K indicates that each microphone collects the totalframes of signal, []TIt indicates
The transposition of vector or matrix;
Windowing process is done to each signal frame in the K signal frame, K windowing signal frame is obtained, for example, to i-th
J-th of frame x of audio signali,jAdding window obtains j-th of windowing signal frame y of i-th of audio signali,j=xi,j×Win;
FFT transform is done to each windowing signal frame in the K windowing signal frame, obtains K echo signal frame, for example,
To j-th of windowing signal frame y of i-th of audio signali,j(t) FFT transform is done, j-th of target of i-th of audio signal is obtained
Signal frame Yi,j(ω);
According to the corresponding K echo signal frame of each audio signal, determine in N number of microphone except this is with reference to Mike
The phase spectrum difference between each microphone and the reference microphone except wind, for example, it is assumed that the master of j-th of echo signal frame
Frequency isThen can be in basic frequency with according to the following formula 4 i-th of microphone of calculating and reference microphoneThe phase spectrum at place
Difference.
Wherein, imag () expression takes imaginary part, and ln () expression takes natural logrithm,Indicate i-th of microphone with
Phase spectrum difference between reference microphone,Indicate j-th of echo signal frame of reference microphone,Indicate the
J-th of echo signal frame of i microphone,Indicate basic frequency.
It should be noted that in above-mentioned Fig. 3 being calculated separately and removing using first microphone as reference microphone
The phase spectrum difference between each microphone and first microphone except first microphone, and the first microphone diaphone
Frequency signal x1(t), second microphone corresponds to audio signal x2(t) ..., N microphone corresponds to audio signal xN(t)。
Optionally, K indicates that each microphone receives the totalframes of signal.
It should be noted that bring truncation effect when windowing process is used to eliminate framing.It is alternatively possible to be to the K
Each signal frame in a signal frame is done plus Hamming window processing.
In some possible implementations, any two adjacent signals frame is overlapped R%, R > 0 in the K signal frame.Example
Such as, which is 25 or 50.In other words, any two adjacent signals frame overlapping 25% or 50% in the K signal frame.
Optionally, signal amplitude remains unchanged after being overlapped adding window.
It should be understood that each frame signal after overlapping has the ingredient of previous frame, prevent discontinuous between two frames.
Optionally, in the embodiment of the present application, when carrying out phase equalization assessment, which is to play
The signal acquired in the environment of swept-frequency signal data.In other words, when computationally stating phase spectrum difference, N number of audio signal
It is the signal acquired in the environment of playing swept-frequency signal data.
It is consequently possible to calculate the phase difference of any frequencies omega is out to get between i-th of microphone and reference microphone
Phase spectrum difference PDiffi(ω), i.e., it is above-mentioned
Optionally, in the embodiment of the present application, after N number of audio signal sample arrives, audio signal point can be passed through
Frame does FFT transform to every frame windowing signal, seeks the power of every frame signal after FFT transform to every frame audio signal adding window
Spectrum, seeks the power spectrum difference between different microphones.
Specifically, it is assumed that N number of audio signal is x1(t),x2(t),…,xN(t), which is believed
Each audio signal in number carries out framing, obtains K signal frame of equal length, K >=2, for example, by i-th of audio signal
Framing is carried out, K signal frame for obtaining equal length is write as following vector form:
xi(t)=[xi,1(t),xi,2(t),…,xi,K(t)]T
Wherein, xi(t) i-th of audio signal is indicated, K indicates that each microphone receives the totalframes of signal, []TIt indicates
The transposition of vector or matrix;
Windowing process is done to each signal frame in the K signal frame, K windowing signal frame is obtained, for example, to i-th
J-th of frame x of audio signali,jAdding window obtains j-th of windowing signal frame y of i-th of audio signali,j=xi,j×Win;
FFT transform is done to each windowing signal frame in the K windowing signal frame, obtains K echo signal frame, for example,
To j-th of windowing signal frame y of i-th of audio signali,j(t) FFT transform is done, j-th of target of i-th of audio signal is obtained
Signal frame Yi,j(ω);
According to the corresponding K echo signal frame of each audio signal, the power spectrum of each audio signal, example are determined
Such as, according to the following formula 5 calculate i-th of audio signal power spectrum;
According to the power spectrum of each audio signal, determine in N number of microphone each of in addition to the reference microphone
Power spectrum difference between microphone and the reference microphone, for example, 6 calculating i-th of microphone and the ginseng according to the following formula
Examine the power spectrum difference between microphone.
Wherein, Pi(ω) indicates the power spectrum of i-th of audio signal, Yi,j(ω) indicates the jth in i-th of audio signal
A echo signal frame, ω indicate frequency, and K indicates that each microphone collects the totalframes of signal.
PDi(ω)=P1(ω)-Pi(ω) formula 6
Wherein, PDi(ω) indicates the power spectrum difference between i-th of microphone and reference microphone, P1(ω) indicates reference
The power spectrum of microphone, Pi(ω) indicates the power spectrum of i-th of microphone.
It should be noted that in above-mentioned Fig. 4 being calculated separately and removing using first microphone as reference microphone
The power spectrum difference between each microphone and first microphone except first microphone, and the first microphone diaphone
Frequency signal x1(t), second microphone corresponds to audio signal x2(t) ..., N microphone corresponds to audio signal xN(t)。
It should be noted that bring truncation effect when windowing process is used to eliminate framing.It is alternatively possible to be to the K
Each signal frame in a signal frame is done plus Hamming window processing.
In some possible implementations, any two adjacent signals frame is overlapped R%, R > 0 in the K signal frame.Example
Such as, which is 25 or 50.In other words, any two adjacent signals frame overlapping 25% or 50% in the K signal frame.
Optionally, signal amplitude remains unchanged after being overlapped adding window.
It should be understood that each frame signal after overlapping has the ingredient of previous frame, prevent discontinuous between two frames.
Optionally, in the embodiment of the present application, when carrying out amplitude coincidence assessment, which is to play
The signal acquired in the environment of white Gaussian noise data or swept-frequency signal data.In other words, power spectral difference is computationally stated
When value, which is the signal acquired in the environment of playing white Gaussian noise data or swept-frequency signal data.
S130, according in N number of microphone in addition to the reference microphone each microphone and the reference microphone it
Between phase spectrum difference and/or power spectrum difference, compliance evaluation is carried out to the N number of microphone.
Specifically, phase spectrum difference is for carrying out phase equalization assessment and power spectrum difference for carrying out amplitude one
The assessment of cause property.
Optionally, in the embodiment of the present application, according in N number of microphone each of in addition to the reference microphone
Phase spectrum difference between microphone and the reference microphone assesses the phase between corresponding microphone and the reference microphone
Bit integrity.
It should be noted that the phase spectrum difference between two microphones is smaller, the phase between the two microphones is indicated
Bit integrity is better.
For example, the phase spectrum difference between microphone 1 and reference microphone is A, A is smaller, indicates microphone 1 and refers to wheat
Phase equalization between gram wind is better.
It is alternatively possible to a threshold value be arranged, if the phase spectrum difference between two microphones is less than this threshold value, table
Show that the phase equalization between the two microphones meets design requirement, the consistency between the two microphones is to multichannel language
Sound enhancing algorithm influence can ignore or the two microphones between consistency to multicenter voice enhancing algorithm do not have
It influences.
It should be noted that above-mentioned threshold value can enhance algorithm flexible configuration according to different multicenter voices.
It should be noted that because when acquiring data, the distance of different microphones to sound source be difficult to it is completely the same, so not
With between microphone, there are a fixed skews.
Optionally, in the embodiment of the present application, above-mentioned phase spectrum difference can be calibrated by fixed skew.
Specifically, each microphone in N number of microphone in addition to the reference microphone is measured respectively and this refers to wheat
Gram wind to sound source range difference, for example, diIndicate the range difference of i-th of microphone with reference microphone to sound source;
According to measured range difference, each Mike in N number of microphone in addition to the reference microphone is calculated separately
Fixed skew between wind and the reference microphone, for example, i-th of microphone and reference can be calculated with according to the following formula 7
Fixed skew between microphone;
According to each microphone in N number of microphone in addition to the reference microphone and consolidating between the reference microphone
Phase bit is poor, calibrates its corresponding phase spectrum difference respectively.
Wherein, Yi(ω) indicates the frequency spectrum of i-th of microphone, Y1(ω) indicates that the frequency spectrum of reference microphone, ω indicate frequency
Rate, diIndicate the range difference of i-th of microphone and reference microphone to sound source, the c expression velocity of sound, 2 π ω di/ c indicates i-th of Mike
Fixed skew between wind and reference microphone.
It should be noted that fixed skew and signal frequency meet linear relationship, therefore, it is possible to use linear fit
Mode determines fixed skew.
For example, the fixed skew between microphone 1 and reference microphone is A, between microphone 1 and reference microphone
Phase spectrum difference is B, as shown in figure 5, straight line portion indicates the stationary phase between the obtained microphone 1 of fitting and reference microphone
Potential difference, curved portion indicate the phase spectrum difference between microphone 1 and reference microphone, and overall performance goes out, with frequency from
0Hz increases to 8000Hz, and the phase spectrum difference between microphone 1 and reference microphone is decreased to -2 radians from 0 radian.Calibrate it
Afterwards, the phase spectrum difference between microphone 1 and reference microphone is C, as shown in curve in Fig. 6, at this point, C=B-A, entirety
Show, as frequency increases to 8000Hz from 0Hz, the phase spectrum difference between microphone 1 and reference microphone 0 radian with
It is fluctuated between ± 0.5 radian.
By Fig. 5 and Fig. 6 comparison it is found that fixed skew the phase spectrum difference between two microphones can be caused it is biggish
It influences, therefore, when carrying out amplitude coincidence assessment to two microphones, needs to eliminate the fixed skew institute between two microphones
Caused by influence.
Optionally, in the embodiment of the present application, according to each Mike in N number of microphone in addition to the reference microphone
It is consistent with the amplitude between the reference microphone to assess corresponding microphone for power spectrum difference between wind and the reference microphone
Property.
It should be noted that the power spectrum difference between two microphones is smaller, the width between the two microphones is indicated
It is better to spend consistency.
For example, as shown in fig. 7, specifically, Fig. 7 a shows the power spectrum of microphone 1 and the power spectrum of reference microphone,
Fig. 7 b shows the power spectrum difference between microphone 1 and reference microphone, the power spectrum between microphone 1 and reference microphone
It is not much different, and maximum value < ± 1 decibel (dB) of its power spectrum difference.
It is alternatively possible to a threshold value be arranged, if the power spectrum difference between two microphones is less than this threshold value, table
Show that the amplitude coincidence between the two microphones meets design requirement, the consistency between the two microphones is to multichannel language
Sound enhancing algorithm influence can ignore or the two microphones between consistency to multicenter voice enhancing algorithm do not have
It influences.
It should be noted that above-mentioned threshold value can enhance algorithm flexible configuration according to different multicenter voices.
Optionally, in the embodiment of the present application, can test item by item the circuit of such as microphone array, electronic component,
Influence of the factors such as acoustic construction to microphone consistency specifically, can be guidance to instruct the calibration of microphone array
The design of microphone and the design of microphone array, the robustness of assessment multichannel enhancing algorithm.
Therefore, in the embodiment of the present application, the N number of audio signal that can be acquired respectively according to N number of microphone, determines each
Phase spectrum difference and/or power spectrum difference between microphone and reference microphone are commented to carry out consistency to N number of microphone
Estimate, eliminates influence of the consistency to multicenter voice enhancing algorithm between microphone, promote user experience.
Optionally, as shown in figure 8, the embodiment of the present application provides a kind of equipment 800 for assessing microphone array consistency,
Include:
Acquiring unit 810, the N number of audio signal acquired respectively for obtaining N number of microphone, N number of microphone are constituted
Microphone array, N >=2;
Processing unit 820, for according to N number of audio signal, determine in N number of microphone except reference microphone it
Phase spectrum difference and/or power spectrum difference between outer each microphone and the reference microphone, the reference microphone
For any one microphone in N number of microphone;
The processing unit 820 is also used to according to each wheat in N number of microphone in addition to the reference microphone
Phase spectrum difference and/or power spectrum difference gram between wind and the reference microphone carry out consistency to N number of microphone
Assessment.
Optionally, the processing unit 820 is specifically used for:
According in N number of microphone in addition to the reference microphone each microphone and the reference microphone it
Between phase spectrum difference, assess the phase equalization between corresponding microphone and the reference microphone.
Optionally, the processing unit 820 is also used to:
Each microphone in N number of microphone in addition to the reference microphone is measured respectively with described with reference to Mike
Range difference of the wind to sound source;
According to measured range difference, calculate separately in N number of microphone each of in addition to the reference microphone
Fixed skew between microphone and the reference microphone;
According in N number of microphone in addition to the reference microphone each microphone and the reference microphone it
Between fixed skew, calibrate its corresponding phase spectrum difference respectively.
Optionally, the processing unit 820 is specifically used for:
According to formulaCalculate separately in N number of microphone except the reference microphone it
Fixed skew between outer each microphone and the reference microphone,
Wherein, Yi(ω) indicates the frequency spectrum of i-th of microphone, Y1(ω) indicates that the frequency spectrum of reference microphone, ω indicate frequency
Rate, diIndicate the range difference of i-th of microphone and reference microphone to sound source, the c expression velocity of sound, 2 π ω di/ c indicates i-th of Mike
Fixed skew between wind and reference microphone.
Optionally, the processing unit 820 is specifically used for:
According in N number of microphone in addition to the reference microphone each microphone and the reference microphone it
Between power spectrum difference, assess the amplitude coincidence between corresponding microphone and the reference microphone.
Optionally, N number of audio signal is the signal acquired in the environment of playing swept-frequency signal data.
Optionally, N number of audio signal is in the environment of playing white Gaussian noise data or swept-frequency signal data
The signal of acquisition.
Optionally, the swept-frequency signal is Linear chirp, logarithm swept-frequency signal, linear stepping swept-frequency signal, logarithm
Any one in stepping swept-frequency signal.
Optionally, the processing unit 820 is specifically used for:
Each audio signal in N number of audio signal is subjected to framing, obtains K signal frame of equal length, K >=
2;
Windowing process is done to each signal frame in the K signal frame, obtains K windowing signal frame;
FFT transform is done to each windowing signal frame in the K windowing signal frame, obtains K echo signal frame;
According to the corresponding K echo signal frame of each audio signal, determine in N number of microphone except described
The phase spectrum difference and/or power spectrum difference between each microphone and the reference microphone except reference microphone.
Optionally, any two adjacent signals frame is overlapped R%, R > 0 in the K signal frame.
Optionally, the R is 25 or 50.
Optionally, i-th of audio signal is subjected to framing, K signal frame for obtaining equal length is write as following vector shape
Formula:
xi(t)=[xi,1(t),xi,2(t),…,xi,K(t)]T
Wherein, xi(t) i-th of audio signal is indicated, K indicates that each microphone collects the totalframes of signal, []TIt indicates
The transposition of vector or matrix.
Optionally, the processing unit 820 is specifically used for:
According to formulaIt determines in N number of microphone except the reference microphone
Except each microphone and the reference microphone between phase spectrum difference,
Wherein, imag () expression takes imaginary part, and ln () expression takes natural logrithm,Indicate i-th of microphone and ginseng
The phase spectrum difference between microphone is examined,Indicate j-th of echo signal frame of reference microphone,Indicate i-th
J-th of echo signal frame of a microphone,Indicate basic frequency.
Optionally, the processing unit 820 is specifically used for:
According to the corresponding K echo signal frame of each audio signal, the function of each audio signal is determined
Rate spectrum;
According to the power spectrum of each audio signal, determine in N number of microphone in addition to the reference microphone
Each microphone and the reference microphone between power spectrum difference.
Optionally, the processing unit 820 is specifically used for:
According to formulaThe power spectrum of each audio signal is calculated,
Wherein, Pi(ω) indicates the power spectrum of i-th of audio signal, Yi,j(ω) indicates the jth in i-th of audio signal
A echo signal frame, K indicate that each microphone collects the totalframes of signal, and ω indicates frequency.
Optionally, the processing unit 820 is specifically used for:
According to formula PDi(ω)=P1(ω)-Pi(ω) calculates every in addition to reference microphone in N number of microphone
Power spectrum difference between a microphone and the reference microphone,
Wherein, PDi(ω) indicates the power spectrum difference between i-th of microphone and reference microphone, P1(ω) indicates reference
The power spectrum of microphone, Pi(ω) indicates the power spectrum of i-th of microphone.
Optionally, the processing unit 820 is specifically used for:
Determine sample frequency F of the N number of microphone when carrying out audio signal samplesWith FFT points Nfft, using raising
Sound device plays white Gaussian noise data or swept-frequency signal data, controls N number of microphone and acquires N number of audio signal,
Wherein, if the data that are played of the loudspeaker are swept-frequency signal data, the swept-frequency signal data it is equal by M+1 segment length and
Equal signal is not constituted frequency,
Optionally, the processing unit 820 is also used to:
According to formulaThe frequency of every segment signal in the M+1 segment signal is calculated, and
According to formula Si(t)=sin (2 π fiT) every segment signal in the M+1 segment signal is calculated,
Wherein, fiIndicate the frequency of the i-th segment signal, FsIndicate sample frequency, NfftIndicate FFT points, Si(t) i-th is indicated
Segment signal, and S1(t) length is the integral multiple of cycle T, T=1/f1。
Optionally, the swept-frequency signal data that the loudspeaker is played are write as following vector form:
S (t)=[S0(t),S1(t),…,SM(t)]T
Wherein, S (t) indicates the swept-frequency signal data that loudspeaker is played, Si(t) the i-th segment signal is indicated,
[ ]TIndicate the transposition of vector or matrix.
Optionally, N number of microphone collects N number of audio signal respectively, wherein the collected audio of i-th of microphone
Signal is expressed as xiAnd x (t),i(t) it can be write as following vector form:
xi(t)=[xi,1(t),xi,2(t),…,xi,K(t)]T
Wherein, xi(t) the collected audio signal of i-th of microphone is indicated, K indicates that each microphone collects signal
Totalframes, []TIndicate the transposition of vector or matrix.
Optionally, the acquiring unit 810 is specifically used for:
N number of microphone is placed in test room, loudspeaker, N number of wheat are configured in the test room
Gram wind is located at the front of the loudspeaker;
It controls the loudspeaker and plays white Gaussian noise data or swept-frequency signal data, and control N number of Mike
Wind acquires N number of audio signal respectively.
Optionally, there is noise reduction room environmental in the test room, the loudspeaker is audio-frequency test Special artificial mouth, and
The artificial mouth is calibrated with standard microphone before the use.
Optionally, the loudspeaker is controlled in the processing unit 820 play white Gaussian noise data or swept-frequency signal
Before data, the acquiring unit 810 is also used to:
Under quiet environment, N number of microphone is obtained in the first duration T1First audio data X of interior acquisition1(n);
In the environment of playing white Gaussian noise data or swept-frequency signal data, N number of microphone is obtained second
Duration T2The second audio data X of interior acquisition2(n);
The processing unit 820 is triggered according to formulaSignal to Noise Ratio (SNR) is calculated, and ensures institute
SNR is stated greater than first threshold.
Optionally, as shown in figure 9, the embodiment of the present application provides a kind of device 900 for assessing microphone array consistency,
Include:
Memory 910, for storing program and data;And
Processor 920, for calling and running the program and data that store in the memory;
The device 900 is configured as executing method shown in above-mentioned Fig. 1 to 7.
Optionally, as shown in Figure 10, the embodiment of the present application provides a kind of system for assessing microphone array consistency
1000, comprising:
Constitute N number of microphone of microphone array 1010, N >=2;
At least one audio-source 1020;
Device 1030, including the memory 1031 for storing program and data and for calling and running the memory
The program of middle storage and the processor 1032 of data, the device 1030 are configured as method shown in above-mentioned Fig. 1 to 7.
It should be understood that magnitude of the sequence numbers of the above procedures are not meant to execute suitable in the various embodiments of the application
Sequence it is successive, the execution of each process sequence should be determined by its function and internal logic, the implementation without coping with the embodiment of the present application
Process constitutes any restriction.
Those of ordinary skill in the art may be aware that list described in conjunction with the examples disclosed in the embodiments of the present disclosure
Member and algorithm steps can be realized with the combination of electronic hardware or computer software and electronic hardware.These functions are actually
It is implemented in hardware or software, the specific application and design constraint depending on technical solution.Professional technician
Each specific application can be used different methods to achieve the described function, but this realization is it is not considered that exceed
Scope of the present application.
It is apparent to those skilled in the art that for convenience and simplicity of description, the system of foregoing description,
The specific work process of device and unit, can refer to corresponding processes in the foregoing method embodiment, and details are not described herein.
In several embodiments provided herein, it should be understood that disclosed systems, devices and methods, it can be with
It realizes by another way.For example, the apparatus embodiments described above are merely exemplary, for example, the unit is drawn
Point, only a kind of logical function partition, there may be another division manner in actual implementation, such as multiple units or components can
To combine or be desirably integrated into another system, or some features can be ignored or not executed.Another point, it is shown or beg for
The mutual coupling, direct-coupling or communication connection of opinion can be through some interfaces, the INDIRECT COUPLING of device or unit
Or communication connection, it can be electrical property, mechanical or other forms.
The unit as illustrated by the separation member may or may not be physically separated, aobvious as unit
The component shown may or may not be physical unit, it can and it is in one place, or may be distributed over multiple
In network unit.It can select some or all of unit therein according to the actual needs to realize the mesh of this embodiment scheme
's.
It, can also be in addition, each functional unit in each embodiment of the application can integrate in one processing unit
It is that each unit physically exists alone, can also be integrated in one unit with two or more units.Above-mentioned integrated list
Member both can take the form of hardware realization, can also realize in the form of software functional units.
If the integrated unit is realized in the form of SFU software functional unit and sells or use as independent product
When, it can store in a computer readable storage medium.Based on this understanding, the technical solution of the application is substantially
The part of the part that contributes to existing technology or the technical solution can embody in the form of software products in other words
Come, which is stored in a storage medium, including some instructions are used so that a computer equipment (can
To be personal computer, server or the network equipment etc.) execute each embodiment the method for the application all or part
Step.And storage medium above-mentioned include: USB flash disk, it is mobile hard disk, read-only memory (ROM, Read-Only Memory), random
Access various Jie that can store program code such as memory (RAM, Random Access Memory), magnetic or disk
Matter.
The above, the only specific embodiment of the application, but the protection scope of the application is not limited thereto, it is any
Those familiar with the art within the technical scope of the present application, can easily think of the change or the replacement, and should all contain
Lid is within the scope of protection of this application.Therefore, the protection scope of the application should be subject to the scope of protection of the claims.
Claims (48)
1. a kind of method for assessing microphone array consistency characterized by comprising
N number of audio signal that N number of microphone acquires respectively is obtained, N number of microphone constitutes microphone array, N >=2;
According to N number of audio signal, determine each microphone in N number of microphone in addition to reference microphone with it is described
Phase spectrum difference and/or power spectrum difference between reference microphone, the reference microphone are appointing in N number of microphone
It anticipates a microphone;
According to each microphone in N number of microphone in addition to the reference microphone and between the reference microphone
Phase spectrum difference and/or power spectrum difference carry out compliance evaluation to N number of microphone.
2. the method according to claim 1, wherein it is described according in N number of microphone remove reference microphone
Except each microphone and the reference microphone between phase spectrum difference, consistency is carried out to the N number of microphone and is commented
Estimate, comprising:
According to each microphone in N number of microphone in addition to the reference microphone and between the reference microphone
Phase spectrum difference assesses the phase equalization between corresponding microphone and the reference microphone.
3. according to the method described in claim 2, it is characterized in that, the method also includes:
Each microphone in N number of microphone in addition to the reference microphone is measured respectively to arrive with the reference microphone
The range difference of sound source;
According to measured range difference, each Mike in N number of microphone in addition to the reference microphone is calculated separately
Fixed skew between wind and the reference microphone;
According to each microphone in N number of microphone in addition to the reference microphone and between the reference microphone
Fixed skew calibrates its corresponding phase spectrum difference respectively.
4. according to the method described in claim 3, it is characterized in that, the distance according to measured by, calculates separately described N number of
The fixed skew between each microphone and the reference microphone in microphone in addition to the reference microphone, packet
It includes:
According to formulaIt calculates separately in N number of microphone in addition to the reference microphone
Fixed skew between each microphone and the reference microphone,
Wherein, Yi(ω) indicates the frequency spectrum of i-th of microphone, Y1(ω) indicates that the frequency spectrum of reference microphone, ω indicate frequency, di
Indicate the range difference of i-th of microphone and reference microphone to sound source, the c expression velocity of sound, 2 π ω di/ c indicate i-th microphone with
Fixed skew between reference microphone.
5. method according to claim 1 to 4, which is characterized in that described to be removed according in N number of microphone
The phase spectrum difference between each microphone and the reference microphone except reference microphone, to N number of microphone into
Row compliance evaluation, comprising:
According to each microphone in N number of microphone in addition to the reference microphone and between the reference microphone
Power spectrum difference assesses the amplitude coincidence between corresponding microphone and the reference microphone.
6. method according to any one of claim 2 to 4, which is characterized in that N number of audio signal is swept in broadcasting
The signal acquired in the environment of frequency signal data.
7. according to the method described in claim 5, it is characterized in that, N number of audio signal is to play white Gaussian noise number
According to or swept-frequency signal data in the environment of the signal that acquires.
8. method according to claim 6 or 7, which is characterized in that the swept-frequency signal is Linear chirp, logarithm is swept
Frequency signal, linear stepping swept-frequency signal, any one in logarithm stepping swept-frequency signal.
9. method according to any one of claim 1 to 8, which is characterized in that it is described according to N number of audio signal,
Determine each microphone in N number of microphone in addition to reference microphone and the Phase spectrum difference between the reference microphone
Value and/or power spectrum difference, comprising:
Each audio signal in N number of audio signal is subjected to framing, obtains K signal frame of equal length, K >=2;
Windowing process is done to each signal frame in the K signal frame, obtains K windowing signal frame;
FFT transform is done to each windowing signal frame in the K windowing signal frame, obtains K echo signal frame;
According to the corresponding K echo signal frame of each audio signal, determine in N number of microphone except the reference
The phase spectrum difference and/or power spectrum difference between each microphone and the reference microphone except microphone.
10. according to the method described in claim 9, it is characterized in that, any two adjacent signals frame weight in the K signal frame
Folded R%, R > 0.
11. according to the method described in claim 10, it is characterized in that, the R is 25 or 50.
12. the method according to any one of claim 9 to 11, which is characterized in that divided i-th of audio signal
Frame, K signal frame for obtaining equal length are write as following vector form:
xi(t)=[xi,1(t),xi,2(t),…,xi,K(t)]T
Wherein, xi(t) i-th of audio signal is indicated, K indicates that each microphone collects the totalframes of signal, []TIndicate vector
Or the transposition of matrix.
13. the method according to any one of claim 9 to 12, which is characterized in that described to be believed according to each audio
Number corresponding K echo signal frame, determines each microphone in N number of microphone in addition to the reference microphone
Phase spectrum difference between the reference microphone, comprising:
According to formulaIt determines in N number of microphone in addition to the reference microphone
Each microphone and the reference microphone between phase spectrum difference,
Wherein, imag () expression takes imaginary part, and ln () expression takes natural logrithm,It indicates i-th of microphone and refers to wheat
Phase spectrum difference between gram wind,Indicate j-th of echo signal frame of reference microphone,Indicate i-th of wheat
J-th of echo signal frame of gram wind,Indicate basic frequency.
14. the method according to any one of claim 9 to 13, which is characterized in that described to be believed according to each audio
Number corresponding K echo signal frame, determines each microphone in N number of microphone in addition to the reference microphone
Power spectrum difference between the reference microphone, comprising:
According to the corresponding K echo signal frame of each audio signal, the power spectrum of each audio signal is determined;
According to the power spectrum of each audio signal, determine every in addition to the reference microphone in N number of microphone
Power spectrum difference between a microphone and the reference microphone.
15. according to the method for claim 14, which is characterized in that described corresponding described according to each audio signal
K echo signal frame determines the power spectrum of each audio signal, comprising:
According to formulaThe power spectrum of each audio signal is calculated,
Wherein, Pi(ω) indicates the power spectrum of i-th of audio signal, Yi,j(ω) indicates j-th of target in i-th of audio signal
Signal frame, K indicate that each microphone collects the totalframes of signal, and ω indicates frequency.
16. method according to claim 14 or 15, which is characterized in that the power according to each audio signal
Spectrum, determines each microphone in N number of microphone in addition to the reference microphone and between the reference microphone
Power spectrum difference, comprising:
According to formula PDi(ω)=P1(ω)-Pi(ω) calculates each wheat in N number of microphone in addition to reference microphone
Power spectrum difference gram between wind and the reference microphone,
Wherein, PDi(ω) indicates the power spectrum difference between i-th of microphone and reference microphone, P1(ω) indicates to refer to Mike
The power spectrum of wind, Pi(ω) indicates the power spectrum of i-th of microphone.
17. according to claim 1 to method described in any one of 16, which is characterized in that the N number of microphone of acquisition is adopted respectively
N number of audio signal of collection, comprising:
Determine sample frequency F of the N number of microphone when carrying out audio signal samplesWith FFT points Nfft, broadcast using loudspeaker
White Gaussian noise data or swept-frequency signal data are put, N number of microphone acquires N number of audio signal, wherein if described
The data that loudspeaker is played are swept-frequency signal data, the swept-frequency signal data are equal by M+1 segment length and frequency not etc.
Signal is constituted,
18. according to the method for claim 17, which is characterized in that
According to formulaThe frequency of every segment signal in the M+1 segment signal is calculated, and
According to formula Si(t)=sin (2 π fiT) every segment signal in the M+1 segment signal is calculated,
Wherein, fiIndicate the frequency of the i-th segment signal, FsIndicate sample frequency, NfftIndicate FFT points, Si(t) i-th section of letter is indicated
Number, and S1(t) length is the integral multiple of cycle T, T=1/f1。
19. according to the method for claim 18, which is characterized in that the swept-frequency signal data that the loudspeaker is played are write as
Following vector form:
S (t)=[S0(t),S1(t),…,SM(t)]T
Wherein, S (t) indicates the swept-frequency signal data that loudspeaker is played, Si(t) the i-th segment signal is indicated,[]TTable
Show the transposition of vector or matrix.
20. according to claim 1 to method described in any one of 19, which is characterized in that N number of microphone collects respectively
N number of audio signal, wherein the collected audio signal of i-th of microphone is expressed as xiAnd x (t),i(t) it can be write as following vector
Form:
xi(t)=[xi,1(t),xi,2(t),…,xi,K(t)]T
Wherein, xi(t) the collected audio signal of i-th of microphone is indicated, K indicates that each microphone collects total frame of signal
Number, []TIndicate the transposition of vector or matrix.
21. according to claim 1 to method described in any one of 20, which is characterized in that the N number of microphone of acquisition is adopted respectively
N number of audio signal of collection, comprising:
N number of microphone is placed in test room, loudspeaker, N number of microphone are configured in the test room
Positioned at the front of the loudspeaker;
It controls the loudspeaker and plays white Gaussian noise data or swept-frequency signal data, and control N number of microphone point
N number of audio signal is not acquired.
22. according to the method for claim 21, which is characterized in that there is noise reduction room environmental in the test room, it is described
Loudspeaker is audio-frequency test Special artificial mouth, and the artificial mouth is calibrated with standard microphone before the use.
23. the method according to claim 21 or 22, which is characterized in that play white Gaussian noise controlling the loudspeaker
Before data or swept-frequency signal data, the method also includes:
Under quiet environment, N number of microphone is obtained in the first duration T1First audio data X of interior acquisition1(n);
In the environment of playing white Gaussian noise data or swept-frequency signal data, N number of microphone is obtained in the second duration T2
The second audio data X of interior acquisition2(n);
According to formulaSignal to Noise Ratio (SNR) is calculated, and ensures that the SNR is greater than first threshold.
24. a kind of equipment for assessing microphone array consistency characterized by comprising
Acquiring unit, the N number of audio signal acquired respectively for obtaining N number of microphone, N number of microphone constitute microphone array
Column, N >=2;
Processing unit, for determining every in addition to reference microphone in N number of microphone according to N number of audio signal
Phase spectrum difference and/or power spectrum difference between a microphone and the reference microphone, the reference microphone are the N
Any one microphone in a microphone;
The processing unit, be also used to according in N number of microphone in addition to the reference microphone each microphone with
Phase spectrum difference and/or power spectrum difference between the reference microphone carry out compliance evaluation to N number of microphone.
25. equipment according to claim 24, which is characterized in that the processing unit is specifically used for:
According to each microphone in N number of microphone in addition to the reference microphone and between the reference microphone
Phase spectrum difference assesses the phase equalization between corresponding microphone and the reference microphone.
26. equipment according to claim 25, which is characterized in that the processing unit is also used to:
Each microphone in N number of microphone in addition to the reference microphone is measured respectively to arrive with the reference microphone
The range difference of sound source;
According to measured range difference, each Mike in N number of microphone in addition to the reference microphone is calculated separately
Fixed skew between wind and the reference microphone;
According to each microphone in N number of microphone in addition to the reference microphone and between the reference microphone
Fixed skew calibrates its corresponding phase spectrum difference respectively.
27. equipment according to claim 26, which is characterized in that the processing unit is specifically used for:
According to formulaIt calculates separately in N number of microphone in addition to the reference microphone
Fixed skew between each microphone and the reference microphone,
Wherein, Yi(ω) indicates the frequency spectrum of i-th of microphone, Y1(ω) indicates that the frequency spectrum of reference microphone, ω indicate frequency, di
Indicate the range difference of i-th of microphone and reference microphone to sound source, the c expression velocity of sound, 2 π ω di/ c indicate i-th microphone with
Fixed skew between reference microphone.
28. the equipment according to any one of claim 24 to 27, which is characterized in that the processing unit is specifically used for:
According to each microphone in N number of microphone in addition to the reference microphone and between the reference microphone
Power spectrum difference assesses the amplitude coincidence between corresponding microphone and the reference microphone.
29. the equipment according to any one of claim 25 to 27, which is characterized in that N number of audio signal is to broadcast
Put the signal acquired in the environment of swept-frequency signal data.
30. equipment according to claim 28, which is characterized in that N number of audio signal is to play white Gaussian noise
The signal acquired in the environment of data or swept-frequency signal data.
31. the equipment according to claim 29 or 30, which is characterized in that the swept-frequency signal is Linear chirp, right
Number swept-frequency signals, linear stepping swept-frequency signal, any one in logarithm stepping swept-frequency signal.
32. the equipment according to any one of claim 24 to 31, which is characterized in that the processing unit is specifically used for:
Each audio signal in N number of audio signal is subjected to framing, obtains K signal frame of equal length, K >=2;
Windowing process is done to each signal frame in the K signal frame, obtains K windowing signal frame;
FFT transform is done to each windowing signal frame in the K windowing signal frame, obtains K echo signal frame;
According to the corresponding K echo signal frame of each audio signal, determine in N number of microphone except the reference
The phase spectrum difference and/or power spectrum difference between each microphone and the reference microphone except microphone.
33. equipment according to claim 32, which is characterized in that any two adjacent signals frame in the K signal frame
It is overlapped R%, R > 0.
34. equipment according to claim 33, which is characterized in that the R is 25 or 50.
35. the equipment according to any one of claim 32 to 34, which is characterized in that divided i-th of audio signal
Frame, K signal frame for obtaining equal length are write as following vector form:
xi(t)=[xi,1(t),xi,2(t),…,xi,K(t)]T
Wherein, xi(t) i-th of audio signal is indicated, K indicates that each microphone collects the totalframes of signal, []TIndicate vector
Or the transposition of matrix.
36. the equipment according to any one of claim 32 to 35, which is characterized in that the processing unit is specifically used for:
According to formulaIt determines in N number of microphone in addition to the reference microphone
Each microphone and the reference microphone between phase spectrum difference,
Wherein, imag () expression takes imaginary part, and ln () expression takes natural logrithm,It indicates i-th of microphone and refers to wheat
Phase spectrum difference between gram wind,Indicate j-th of echo signal frame of reference microphone,Indicate i-th of wheat
J-th of echo signal frame of gram wind,Indicate basic frequency.
37. the equipment according to any one of claim 32 to 36, which is characterized in that the processing unit is specifically used for:
According to the corresponding K echo signal frame of each audio signal, the power spectrum of each audio signal is determined;
According to the power spectrum of each audio signal, determine every in addition to the reference microphone in N number of microphone
Power spectrum difference between a microphone and the reference microphone.
38. the equipment according to claim 37, which is characterized in that the processing unit is specifically used for:
According to formulaThe power spectrum of each audio signal is calculated,
Wherein, Pi(ω) indicates the power spectrum of i-th of audio signal, Yi,j(ω) indicates j-th of target in i-th of audio signal
Signal frame, K indicate that each microphone collects the totalframes of signal, and ω indicates frequency.
39. the equipment according to claim 37 or 38, which is characterized in that the processing unit is specifically used for:
According to formula PDi(ω)=P1(ω)-Pi(ω) calculates each wheat in N number of microphone in addition to reference microphone
Power spectrum difference gram between wind and the reference microphone,
Wherein, PDi(ω) indicates the power spectrum difference between i-th of microphone and reference microphone, P1(ω) indicates to refer to Mike
The power spectrum of wind, Pi(ω) indicates the power spectrum of i-th of microphone.
40. the equipment according to any one of claim 24 to 39, which is characterized in that the processing unit is specifically used for:
Determine sample frequency F of the N number of microphone when carrying out audio signal samplesWith FFT points Nfft, broadcast using loudspeaker
White Gaussian noise data or swept-frequency signal data are put, N number of microphone is controlled and acquires N number of audio signal, wherein if
The data that the loudspeaker is played are swept-frequency signal data, the swept-frequency signal data are equal by M+1 segment length and frequency not
Deng signal constitute,
41. equipment according to claim 40, which is characterized in that the processing unit is also used to:
According to formulaThe frequency of every segment signal in the M+1 segment signal is calculated, and
According to formula Si(t)=sin (2 π fiT) every segment signal in the M+1 segment signal is calculated,
Wherein, fiIndicate the frequency of the i-th segment signal, FsIndicate sample frequency, NfftIndicate FFT points, Si(t) i-th section of letter is indicated
Number, and S1(t) length is the integral multiple of cycle T, T=1/f1。
42. equipment according to claim 41, which is characterized in that the swept-frequency signal data that the loudspeaker is played are write as
Following vector form:
S (t)=[S0(t),S1(t),…,SM(t)]T
Wherein, S (t) indicates the swept-frequency signal data that loudspeaker is played, Si(t) the i-th segment signal is indicated,[]TTable
Show the transposition of vector or matrix.
43. the equipment according to any one of claim 24 to 42, which is characterized in that N number of microphone acquires respectively
To N number of audio signal, wherein the collected audio signal of i-th of microphone is expressed as xiAnd x (t),i(t) can be write as it is following to
Amount form:
xi(t)=[xi,1(t),xi,2(t),…,xi,K(t)]T
Wherein, xi(t) the collected audio signal of i-th of microphone is indicated, K indicates that each microphone collects total frame of signal
Number, []TIndicate the transposition of vector or matrix.
44. the equipment according to any one of claim 24 to 43, which is characterized in that the acquiring unit is specifically used for:
N number of microphone is placed in test room, loudspeaker, N number of microphone are configured in the test room
Positioned at the front of the loudspeaker;
It controls the loudspeaker and plays white Gaussian noise data or swept-frequency signal data, and control N number of microphone point
N number of audio signal is not acquired.
45. equipment according to claim 44, which is characterized in that there is noise reduction room environmental in the test room, it is described
Loudspeaker is audio-frequency test Special artificial mouth, and the artificial mouth is calibrated with standard microphone before the use.
46. the equipment according to claim 44 or 45, which is characterized in that control the loudspeaker in the processing unit and broadcast
Before putting white Gaussian noise data or swept-frequency signal data, the acquiring unit is also used to:
Under quiet environment, N number of microphone is obtained in the first duration T1First audio data X of interior acquisition1(n);
In the environment of playing white Gaussian noise data or swept-frequency signal data, N number of microphone is obtained in the second duration T2
The second audio data X of interior acquisition2(n);
The processing unit is triggered according to formulaSignal to Noise Ratio (SNR) is calculated, and ensures that the SNR is big
In first threshold.
47. a kind of device for assessing microphone array consistency characterized by comprising
Memory, for storing program and data;And
Processor, for calling and running the program and data that store in the memory;
Described device is configured as: executing the method as described in any one of claim 1 to 23.
48. a kind of system for assessing microphone array consistency characterized by comprising
Constitute N number of microphone of microphone array, N >=2;
At least one audio-source;
Device, including the memory for storing program and data and for calling and running the program stored in the memory
With the processor of data, described device is configured as:
Execute the method as described in any one of claim 1 to 23.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202310466643.4A CN116437280A (en) | 2018-08-22 | 2018-08-22 | Method, device, apparatus and system for evaluating consistency of microphone array |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
PCT/CN2018/101766 WO2020037555A1 (en) | 2018-08-22 | 2018-08-22 | Method, device, apparatus, and system for evaluating microphone array consistency |
Related Child Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN202310466643.4A Division CN116437280A (en) | 2018-08-22 | 2018-08-22 | Method, device, apparatus and system for evaluating consistency of microphone array |
Publications (2)
Publication Number | Publication Date |
---|---|
CN109313909A true CN109313909A (en) | 2019-02-05 |
CN109313909B CN109313909B (en) | 2023-05-12 |
Family
ID=65221692
Family Applications (2)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN201880001199.6A Active CN109313909B (en) | 2018-08-22 | 2018-08-22 | Method, device, apparatus and system for evaluating consistency of microphone array |
CN202310466643.4A Pending CN116437280A (en) | 2018-08-22 | 2018-08-22 | Method, device, apparatus and system for evaluating consistency of microphone array |
Family Applications After (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN202310466643.4A Pending CN116437280A (en) | 2018-08-22 | 2018-08-22 | Method, device, apparatus and system for evaluating consistency of microphone array |
Country Status (2)
Country | Link |
---|---|
CN (2) | CN109313909B (en) |
WO (1) | WO2020037555A1 (en) |
Cited By (10)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN110111807A (en) * | 2019-04-27 | 2019-08-09 | 南京理工大学 | A kind of indoor sound source based on microphone array follows and Enhancement Method |
CN110636432A (en) * | 2019-09-29 | 2019-12-31 | 深圳市火乐科技发展有限公司 | Microphone testing method and related equipment |
CN111065036A (en) * | 2019-12-26 | 2020-04-24 | 北京声智科技有限公司 | Frequency response testing method and device of microphone array |
CN112672265A (en) * | 2020-10-13 | 2021-04-16 | 珠海市杰理科技股份有限公司 | Method and system for detecting microphone consistency and computer readable storage medium |
CN112889299A (en) * | 2021-01-12 | 2021-06-01 | 华为技术有限公司 | Method and apparatus for evaluating microphone array consistency |
CN113259830A (en) * | 2021-04-26 | 2021-08-13 | 歌尔股份有限公司 | Multi-microphone consistency test system and method |
CN114222234A (en) * | 2021-12-31 | 2022-03-22 | 思必驰科技股份有限公司 | Microphone array consistency detection method, electronic device and storage medium |
CN114390421A (en) * | 2021-12-03 | 2022-04-22 | 伟创力电子技术(苏州)有限公司 | Automatic testing method for microphone matrix and loudspeaker |
CN114449434A (en) * | 2022-04-07 | 2022-05-06 | 荣耀终端有限公司 | Microphone calibration method and electronic equipment |
CN115776626A (en) * | 2023-02-10 | 2023-03-10 | 杭州兆华电子股份有限公司 | Frequency response calibration method and system of microphone array |
Citations (14)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP2006033734A (en) * | 2004-07-21 | 2006-02-02 | Sanyo Electric Co Ltd | Sound inspection method and device of electric product |
CN1756444A (en) * | 2004-09-30 | 2006-04-05 | 富迪科技股份有限公司 | Self detection and correction method for electroacoustic system |
US20100135501A1 (en) * | 2008-12-02 | 2010-06-03 | Tim Corbett | Calibrating at least one system microphone |
WO2011057346A1 (en) * | 2009-11-12 | 2011-05-19 | Robert Henry Frater | Speakerphone and/or microphone arrays and methods and systems of using the same |
CN102075848A (en) * | 2011-02-17 | 2011-05-25 | 深圳市豪恩声学股份有限公司 | Method and system for testing array microphone and rotating device |
CN102111697A (en) * | 2009-12-28 | 2011-06-29 | 歌尔声学股份有限公司 | Method and device for controlling noise reduction of microphone array |
CN102461203A (en) * | 2009-06-09 | 2012-05-16 | 高通股份有限公司 | Systems, methods, apparatus, and computer-readable media for phase-based processing of multichannel signal |
CN103247298A (en) * | 2013-04-28 | 2013-08-14 | 华为技术有限公司 | Sensitivity calibration method and audio frequency apparatus |
CN103559330A (en) * | 2013-10-10 | 2014-02-05 | 上海华为技术有限公司 | Method and system for detecting data consistency |
US20140119552A1 (en) * | 2012-10-26 | 2014-05-01 | Broadcom Corporation | Loudspeaker localization with a microphone array |
CN103917886A (en) * | 2011-08-31 | 2014-07-09 | 弗兰霍菲尔运输应用研究公司 | Direction of arrival estimation using watermarked audio signals and microphone arrays |
CN105554674A (en) * | 2015-12-28 | 2016-05-04 | 努比亚技术有限公司 | Microphone calibration method, device and mobile terminal |
CN106161751A (en) * | 2015-04-14 | 2016-11-23 | 电信科学技术研究院 | A kind of noise suppressing method and device |
WO2016209098A1 (en) * | 2015-06-26 | 2016-12-29 | Intel Corporation | Phase response mismatch correction for multiple microphones |
Family Cites Families (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN103871420B (en) * | 2012-12-13 | 2016-12-21 | 华为技术有限公司 | The signal processing method of microphone array and device |
EP2942982A1 (en) * | 2014-05-05 | 2015-11-11 | Fraunhofer-Gesellschaft zur Förderung der angewandten Forschung e.V. | System, apparatus and method for consistent acoustic scene reproduction based on informed spatial filtering |
CN107864444B (en) * | 2017-11-01 | 2019-10-29 | 大连理工大学 | A kind of microphone array frequency response calibration method |
-
2018
- 2018-08-22 WO PCT/CN2018/101766 patent/WO2020037555A1/en active Application Filing
- 2018-08-22 CN CN201880001199.6A patent/CN109313909B/en active Active
- 2018-08-22 CN CN202310466643.4A patent/CN116437280A/en active Pending
Patent Citations (14)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP2006033734A (en) * | 2004-07-21 | 2006-02-02 | Sanyo Electric Co Ltd | Sound inspection method and device of electric product |
CN1756444A (en) * | 2004-09-30 | 2006-04-05 | 富迪科技股份有限公司 | Self detection and correction method for electroacoustic system |
US20100135501A1 (en) * | 2008-12-02 | 2010-06-03 | Tim Corbett | Calibrating at least one system microphone |
CN102461203A (en) * | 2009-06-09 | 2012-05-16 | 高通股份有限公司 | Systems, methods, apparatus, and computer-readable media for phase-based processing of multichannel signal |
WO2011057346A1 (en) * | 2009-11-12 | 2011-05-19 | Robert Henry Frater | Speakerphone and/or microphone arrays and methods and systems of using the same |
CN102111697A (en) * | 2009-12-28 | 2011-06-29 | 歌尔声学股份有限公司 | Method and device for controlling noise reduction of microphone array |
CN102075848A (en) * | 2011-02-17 | 2011-05-25 | 深圳市豪恩声学股份有限公司 | Method and system for testing array microphone and rotating device |
CN103917886A (en) * | 2011-08-31 | 2014-07-09 | 弗兰霍菲尔运输应用研究公司 | Direction of arrival estimation using watermarked audio signals and microphone arrays |
US20140119552A1 (en) * | 2012-10-26 | 2014-05-01 | Broadcom Corporation | Loudspeaker localization with a microphone array |
CN103247298A (en) * | 2013-04-28 | 2013-08-14 | 华为技术有限公司 | Sensitivity calibration method and audio frequency apparatus |
CN103559330A (en) * | 2013-10-10 | 2014-02-05 | 上海华为技术有限公司 | Method and system for detecting data consistency |
CN106161751A (en) * | 2015-04-14 | 2016-11-23 | 电信科学技术研究院 | A kind of noise suppressing method and device |
WO2016209098A1 (en) * | 2015-06-26 | 2016-12-29 | Intel Corporation | Phase response mismatch correction for multiple microphones |
CN105554674A (en) * | 2015-12-28 | 2016-05-04 | 努比亚技术有限公司 | Microphone calibration method, device and mobile terminal |
Non-Patent Citations (1)
Title |
---|
叶永等: "用于麦克风阵列的阵元筛选方法研究", 《传感技术学报》 * |
Cited By (13)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN110111807A (en) * | 2019-04-27 | 2019-08-09 | 南京理工大学 | A kind of indoor sound source based on microphone array follows and Enhancement Method |
CN110111807B (en) * | 2019-04-27 | 2022-01-11 | 南京理工大学 | Microphone array-based indoor sound source following and enhancing method |
CN110636432A (en) * | 2019-09-29 | 2019-12-31 | 深圳市火乐科技发展有限公司 | Microphone testing method and related equipment |
CN111065036B (en) * | 2019-12-26 | 2021-08-31 | 北京声智科技有限公司 | Frequency response testing method and device of microphone array |
CN111065036A (en) * | 2019-12-26 | 2020-04-24 | 北京声智科技有限公司 | Frequency response testing method and device of microphone array |
CN112672265A (en) * | 2020-10-13 | 2021-04-16 | 珠海市杰理科技股份有限公司 | Method and system for detecting microphone consistency and computer readable storage medium |
CN112889299A (en) * | 2021-01-12 | 2021-06-01 | 华为技术有限公司 | Method and apparatus for evaluating microphone array consistency |
WO2022150950A1 (en) * | 2021-01-12 | 2022-07-21 | 华为技术有限公司 | Method and apparatus for evaluating consistency of microphone array |
CN113259830A (en) * | 2021-04-26 | 2021-08-13 | 歌尔股份有限公司 | Multi-microphone consistency test system and method |
CN114390421A (en) * | 2021-12-03 | 2022-04-22 | 伟创力电子技术(苏州)有限公司 | Automatic testing method for microphone matrix and loudspeaker |
CN114222234A (en) * | 2021-12-31 | 2022-03-22 | 思必驰科技股份有限公司 | Microphone array consistency detection method, electronic device and storage medium |
CN114449434A (en) * | 2022-04-07 | 2022-05-06 | 荣耀终端有限公司 | Microphone calibration method and electronic equipment |
CN115776626A (en) * | 2023-02-10 | 2023-03-10 | 杭州兆华电子股份有限公司 | Frequency response calibration method and system of microphone array |
Also Published As
Publication number | Publication date |
---|---|
CN109313909B (en) | 2023-05-12 |
CN116437280A (en) | 2023-07-14 |
WO2020037555A1 (en) | 2020-02-27 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN109313909A (en) | Assess method, unit and the system of microphone array consistency | |
CN110491403B (en) | Audio signal processing method, device, medium and audio interaction equipment | |
CN106486131B (en) | A kind of method and device of speech de-noising | |
CN105469785B (en) | Voice activity detection method and device in communication terminal dual microphone noise-canceling system | |
CN108766454A (en) | A kind of voice noise suppressing method and device | |
WO2020181824A1 (en) | Voiceprint recognition method, apparatus and device, and computer-readable storage medium | |
CN107910014A (en) | Test method, device and the test equipment of echo cancellor | |
EP3826012B1 (en) | Method and device for evaluating performance of speech enhancement algorithm, electronic device and computer-readable storage medium | |
WO2021093380A1 (en) | Noise processing method and apparatus, and system | |
CN109102819A (en) | One kind is uttered long and high-pitched sounds detection method and device | |
CN103903634B (en) | The detection of activation sound and the method and apparatus for activating sound detection | |
JP2012155339A (en) | Improvement in multisensor sound quality using sound state model | |
CN106161751A (en) | A kind of noise suppressing method and device | |
CN110491373A (en) | Model training method, device, storage medium and electronic equipment | |
WO2022174727A1 (en) | Howling suppression method and apparatus, hearing aid, and storage medium | |
CN108200526A (en) | A kind of sound equipment adjustment method and device based on confidence level curve | |
WO2020112577A1 (en) | Similarity measure assisted adaptation control of an echo canceller | |
US11915718B2 (en) | Position detection method, apparatus, electronic device and computer readable storage medium | |
WO2022105571A1 (en) | Speech enhancement method and apparatus, and device and computer-readable storage medium | |
JP7352740B2 (en) | Method and apparatus for wind noise attenuation | |
EP3549355A1 (en) | Combined audio signal output | |
BR112014009647B1 (en) | NOISE Attenuation APPLIANCE AND NOISE Attenuation METHOD | |
JP2009276365A (en) | Processor, voice recognition device, voice recognition system and voice recognition method | |
CN111192569B (en) | Double-microphone voice feature extraction method and device, computer equipment and storage medium | |
CN106710602A (en) | Acoustic reverberation time estimation method and device |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
GR01 | Patent grant | ||
GR01 | Patent grant |