WO2011002823A1 - Calibrating a dual omnidirectional microphone array (doma) - Google Patents

Calibrating a dual omnidirectional microphone array (doma) Download PDF

Info

Publication number
WO2011002823A1
WO2011002823A1 PCT/US2010/040501 US2010040501W WO2011002823A1 WO 2011002823 A1 WO2011002823 A1 WO 2011002823A1 US 2010040501 W US2010040501 W US 2010040501W WO 2011002823 A1 WO2011002823 A1 WO 2011002823A1
Authority
WO
WIPO (PCT)
Prior art keywords
microphone
filter
response
signal
model
Prior art date
Application number
PCT/US2010/040501
Other languages
English (en)
French (fr)
Inventor
Gregory C. Burnett
Original Assignee
Aliph, Inc.
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Aliph, Inc. filed Critical Aliph, Inc.
Priority to CN201090001122.8U priority Critical patent/CN203086710U/zh
Publication of WO2011002823A1 publication Critical patent/WO2011002823A1/en

Links

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04RLOUDSPEAKERS, MICROPHONES, GRAMOPHONE PICK-UPS OR LIKE ACOUSTIC ELECTROMECHANICAL TRANSDUCERS; DEAF-AID SETS; PUBLIC ADDRESS SYSTEMS
    • H04R29/00Monitoring arrangements; Testing arrangements
    • H04R29/004Monitoring arrangements; Testing arrangements for microphones
    • H04R29/005Microphone arrays
    • H04R29/006Microphone matching

Definitions

  • the disclosure herein relates generally to noise suppression systems.
  • this disclosure relates to calibration of noise suppression systems, devices, and methods for use in acoustic applications.
  • the noise relationship between the microphones is then determined using an adaptive filter (such as Least-Mean-Squares as described in Haykin & Widrow, ISBN# 0471215708, Wiley, 2002, but any adaptive or stationary system identification algorithm may be used) and that relationship used to filter the noise from the desired signal.
  • an adaptive filter such as Least-Mean-Squares as described in Haykin & Widrow, ISBN# 0471215708, Wiley, 2002, but any adaptive or stationary system identification algorithm may be used
  • Multi- microphone systems have not been very successful for a variety of reasons, the most compelling being poor noise cancellation performance and/or significant speech distortion.
  • conventional multi-microphone systems attempt to increase the SNR of the user's speech by "steering" the nulls of the system to the strongest noise sources. This approach is limited in the number of noise sources removed by the number of available nulls.
  • the Jawbone earpiece (referred to as the "Jawbone), introduced in December 2006 by AliphCom of San Francisco, California, was the first known commercial product to use a pair of physical directional microphones (instead of omnidirectional microphones) to reduce environmental acoustic noise.
  • the technology supporting the Jawbone is currently described under one or more of US Patent Number 7,246,058 by Burnett and/or US Patent Application Numbers 10/400,282, 10/667,207, and/or 10/769,302.
  • multi-microphone techniques make use of an acoustic-based Voice Activity Detector (VAD) to determine the background noise characteristics, where "voice” is generally understood to include human voiced speech, unvoiced speech, or a combination of voiced and unvoiced speech.
  • VAD Voice Activity Detector
  • the Jawbone improved on this by using a microphone-based sensor to construct a VAD signal using directly detected speech vibrations in the user's cheek. This allowed the Jawbone to aggressively remove noise when the user was not producing speech.
  • a Jawbone implementation also uses a pair of omnidirectional microphones to construct two virtual microphones that are used to remove noise from speech. This construction requires that the omnidirectional microphones be calibrated, that is, that they both respond as similarly as possible when exposed to the same acoustic field.
  • the omnidirectional microphones incorporate a mechanical highpass filter, with a 3-dB frequency that varies between about 100 and about 400 Hz.
  • Figure 1 shows a continuous-time RC filter response and discrete- time model for a worst-case 3-dB frequency of 350 Hz, under an
  • Figure 2 shows a magnitude response of the calibration filter alpha for three headsets used to test this technique, under an embodiment.
  • Figure 3 shows a phase response of the calibration filter alpha for three headsets used to test this technique, under an embodiment. The peak locations and magnitudes are shown in Figure 16.
  • Figure 4 shows the magnitude response of the calibration filters from Figure 2 (solid) with the RC filter difference model results (dashed), under an embodiment.
  • the RC filter responses have been offset with constant gains (+1.75, +0.25, and -3.25 dB for 6AB5, 6C93, and 90B9 respectively) and match very well with the observed responses.
  • FIG. 5 shows the phase response of the calibration filters from
  • Figure 3 (solid) with the RC filter difference model results (dashed), under an embodiment.
  • the RC filter phase responses are very similar, within a few degrees below 1000 Hz.
  • headset 6C83 which had very little magnitude response difference above 1 kHz, has a very large phase difference.
  • Headsets 6AB5 and 90B9 has phase responses that trend toward zero degrees, as expected, but 90B9 does not, for unknown reasons.
  • Figure 6 shows the calibration flow using a standard gain target for each branch, under an embodiment.
  • the delay "d” is the linear phase delay in samples of the alpha filter.
  • the alpha filter can be either linear phase or minimum phase.
  • Figure 7 shows original Oi, O 2 , and compensated modeled responses for headset 90B9, under an embodiment.
  • the loss is 3.3 dB at 100 Hz, 1.1 dB at 200 Hz, and 0.4 dB at 300 Hz.
  • Figure 8 shows original Oi, O 2 , and compensated modeled responses for headset 6AB5, under an embodiment.
  • the loss is 6.4 dB at 100 Hz, 2.7 dB at 200 Hz, and 1.3 dB at 300 Hz.
  • Figure 9 shows original Oi, O 2 , and compensated modeled responses for headset 6C83, under an embodiment.
  • the loss is 9.4 dB at 100 Hz, 4.7 dB at 200 Hz, and 2.6 dB at 300 Hz.
  • Figure 10 shows compensated O 1 and O 2 responses for three different headsets, under an embodiment. There is a 7.0 dB difference between headset 90B9 and 6C83 at 100 Hz.
  • Figure 11 shows the magnitude response of the calibration filter for the three headsets with factory calibrations before (solid) and after (dashed) compensation, under an embodiment. There is little change except near DC, where the responses are reduced, as intended.
  • Figure 12 shows a calibration phase response for the three headsets using factory calibrations (solid) and compensated Aliph calibrations
  • Figure 14 is a flow diagram of the calibration algorithm, under an embodiment.
  • the top flow is executed on the first three-second excitation and produces the model for each microphone HP filter.
  • the middle flow calculates the LP filter needed to correct the amplitude response of the combination of Oi HA ⁇ and O 2 H AT -
  • the final flow calculates the alpha filter.
  • Figure 15 is a flow diagram of the calibration filters during normal operation, under an embodiment .
  • Figure 16 is a table that shows the locations and size of the maximum phase difference, under an embodiment. Estimated values are calculated as described herein given the peak magnitude and location of the calibration filter.
  • Figure 17 is a table that shows the boost needed to regain original Oi sensitivity for the three responses shown in Figures 6-8, under an embodiment.
  • the amount of boost needed is highly dependent on the original 3-dB frequencies.
  • Figure 18 is a table that shows magnitude responses of several simple RC filters and their combination at 125 and 375 Hz, under an embodiment.
  • Figure 19 is a table that shows a simplified version of the table of Figure 18 with ⁇ f and needed boost for each frequency band, under an embodiment.
  • Figure 20 shows a magnitude response of six test headsets using v4 (solid lines) and v5 (dashed), under an embodiment.
  • the "flares" at DC have been eliminated, reducing the 1 kHz normalized difference in responses from more than 8 dB to less than 2 dB.
  • Figure 21 shows a phase response of six test headsets using v4 (solid lines) and v5 (dashed), under an embodiment.
  • the large peaks below 500 Hz have been eliminated, reducing phase differences from 34 degrees to less than 7 degrees.
  • Figure 22 is a table that shows approximate denoising, devoicing, and SNR increase in dB using headset 931B-v5 as the standard, under an embodiment.
  • Pathfinder-only denoising and devoicing changes were used to compile the table.
  • SNR differences of up to 11 dB were compensated to within 0 to -3 dB of the standard headset.
  • Denoising differences between calibration versions were up to 21 dB before and 2 dB after.
  • Devoicing differences were up to 12 dB before and 2 dB after.
  • Figure 23 shows phase responses of 99 headsets using v4 calibration, under an embodiment.
  • the spread in max phase runs from -21 to + 17 degrees, which results in significant performance differences.
  • Figure 24 shows phase responses of 99 headsets using v5 calibration, under an embodiment.
  • the outlier yellow plot was likely due to operator error.
  • the spread in max phase has changed from -21 to + 17 degrees to +- 5 degrees below 500 Hz.
  • the magnitude variations near DC were similarly eliminated. These headsets should be indistinguishable in performance.
  • Figure 25 shows mean, +-1 ⁇ , and +-2 ⁇ of the magnitude (top) and phase (bottom) responses of 99 headsets using v4 calibration, under an embodiment.
  • the 2 ⁇ spread in magnitude at DC is almost 13 dB, and for phase is 31 degrees. If +5 and -10 degrees are taken to be the cutoff for good performance, then about 40% of these headsets will have significantly poorer performance than the others.
  • Figure 26 shows mean, +-1 ⁇ , and +-2 ⁇ of the magnitude (top) and phase (bottom) responses of 99 headsets using v5 calibration, under an embodiment.
  • the 2 ⁇ spread in magnitude at DC is now only 6 dB (within spec) with less ripple, and for phase is less than 7 degrees with significantly less ripple. These headsets should be indistinguishable in performance.
  • Figure 27 shows magnitude response for the combination of Olhat, 02hat, and H AC , under an embodiment. This will be modulated by Oi's native response to arrive at the final input response to the system.
  • the annotated line shows what the current system does when no phase correction is needed; this has been changed to a unity filter for now and will be updated to a 150 Hz HP for v6. All of the compensated responses are within +-1 dB and their 3dB points within +-25 Hz.
  • Figure 28 is a table that shows initial and final maximum phases for initial maximum near the upper limit, under an embodiment.
  • initial maximum phases For headsets with initial maximum phases above 5 degrees, there was always a reduction in maximum phase. Between 3-5 degrees, there was some reduction in phase and some small increases. Below 3 degrees there was little change or a small increase. Thus 3 degrees is a good upper limit in determining whether or not to compensate for phase differences.
  • Figure 29 is a flow chart of the v6 algorithm where headsets without significant phase difference also get normalized to the standard response, under an embodiment.
  • Figure 31 shows a flow of the v4.1 calibration algorithm, under an embodiment. Since no new information is possible, the benefits are limited to O 1HAT , O 2HAT , and H AC (z) for units that have sufficient alpha phase.
  • Figure 32 shows the use of the filters of an embodiment prior to the
  • Figure 33 is a two-microphone adaptive noise suppression system, under an embodiment.
  • Figure 34 is an array and speech source (S) configuration, under an embodiment.
  • the microphones are separated by a distance approximately equal to 2d 0 , and the speech source is located a distance d s away from the midpoint of the array at an angle ⁇ .
  • the system is axially symmetric so only d s and ⁇ need be specified.
  • Figure 35 is a block diagram for a first order gradient microphone using two omnidirectional elements O 1 and O 2 , under an embodiment.
  • Figure 36 is a block diagram for a DOMA including two physical microphones configured to form two virtual microphones Vi and V 2 , under an embodiment.
  • Figure 37 is a block diagram for a DOMA including two physical microphones configured to form N virtual microphones Vi through V N , where N is any number greater than one, under an embodiment.
  • Figure 38 is an example of a headset or head-worn device that includes the DOMA, as described herein, under an embodiment.
  • Figure 39 is a flow diagram for denoising acoustic signals using the DOMA, under an embodiment.
  • Figure 40 is a flow diagram for forming the DOMA, under an embodiment.
  • Figure 41 is a plot of linear response of virtual microphone V 2 to a 1 kHz speech source at a distance of 0.1 m, under an embodiment.
  • the null is at O degrees, where the speech is normally located.
  • Figure 42 is a plot of linear response of virtual microphone V 2 to a 1 kHz noise source at a distance of 1.0 m, under an embodiment. There is no null and all noise sources are detected.
  • Figure 43 is a plot of linear response of virtual microphone Vi to a 1 kHz speech source at a distance of 0.1 m, under an embodiment. There is no null and the response for speech is greater than that shown in Figure 9.
  • Figure 44 is a plot of linear response of virtual microphone Vi to a 1 kHz noise source at a distance of 1.0 m, under an embodiment. There is no null and the response is very similar to V 2 shown in Figure 10.
  • Figure 45 is a plot of linear response of virtual microphone V 1 to a speech source at a distance of 0.1 m for frequencies of 100, 500, 1000, 2000, 3000, and 4000 Hz, under an embodiment.
  • Figure 46 is a plot showing comparison of frequency responses for speech for the array of an embodiment and for a conventional cardioid microphone.
  • Figure 47 is a plot showing speech response for Vi (top, dashed)
  • the spatial null in V 2 is relatively broad.
  • Figure 48 is a plot showing a ratio of V 1 ZV 2 speech responses shown in Figure 10 versus B, under an embodiment. The ratio is above 10 dB for all 0.8 ⁇ B ⁇ 1.1. This means that the physical ⁇ of the system need not be exactly modeled for good performance.
  • the resulting phase difference clearly affects high frequencies more than low.
  • Non-unity B affects the entire frequency range.
  • the cancellation remains below -10 dB for frequencies below 6 kHz.
  • the cancellation is below -10 dB only for frequencies below about 2.8 kHz and a reduction in performance is expected.
  • the noise has been reduced by about 25 dB and the speech hardly affected, with no noticeable distortion.
  • bleedthrough means the undesired presence of noise during speech.
  • decoding means removing unwanted noise from the signal of interest, and also refers to the amount of reduction of noise energy in a signal in decibels (dB).
  • devoicing means removing and/or distorting the desired speech from the signal of interest.
  • DOMA refers to the Aliph Dual Omnidirectional Microphone
  • Array used in an embodiment of the invention.
  • the technique described herein is not limited to use with DOMA; any array technique that will benefit from more accurate microphone calibrations can be used.
  • omnidirectional microphone means a physical microphone that is equally responsive to acoustic waves originating from any direction.
  • the term “01" or “O 1 " refers to the first omnidirectional microphone of the array, normally closer to the user than the second omnidirectional microphone. It may also, according to context, refer to the time-sampled output of the first omnidirectional microphone or the frequency response of the first omnidirectional microphone.
  • the term "02" or “O 2 " refers to the second omnidirectional microphone of the array, normally farther from the user than the first omnidirectional microphone. It may also, according to context, refer to the time-sampled output of the second omnidirectional microphone or the frequency response of the second omnidirectional microphone.
  • the term refers to the RC filter model of the response Of O 1 .
  • the term refers to the RC filter model of the response
  • noise means unwanted environmental acoustic noise.
  • nucle means a zero or minima in the spatial response of a physical or virtual directional microphone.
  • speech means desired speech of the user.
  • SSM Skin Surface Microphone
  • Vi means the virtual directional "speech" microphone of DOMA.
  • V 2 means the virtual directional "noise" microphone of DOMA, which has a null for the user's speech.
  • VAD Voice Activity Detection
  • VM virtual microphones
  • microphones means a microphone constructed using two or more
  • Calibration methods for two omnidirectional microphones with mechanical highpass filters are described below. More than two microphones may be calibrated using this technique by selecting one omnidirectional microphone to use as a standard and calibrating all other microphones to the chosen standard microphone. Any application that requires accurately calibrated omnidirectional microphones with mechanical highpass filters can benefit from this technique.
  • the embodiment below uses the DOMA microphone array, but the technique is not so limited. Compared to conventional arrays and algorithms, which seek to reduce noise by nulling out noise sources, the array of an embodiment is used to form two distinct virtual directional microphones which are configured to have very similar noise responses and very dissimilar speech responses. The only null formed by the DOMA is one used to remove the speech of the user from V 2 .
  • the omnidirectional microphones can be combined to form two or more virtual microphones which may then be paired with an adaptive filter algorithm and/or VAD algorithm to significantly reduce the noise without distorting the speech, significantly improving the SNR of the desired speech over conventional noise suppression systems.
  • embodiments described herein are stable in operation, flexible with respect to virtual microphone pattern choice, and have proven to be robust with respect to speech source-to-array distance and orientation as well as temperature and calibration techniques, as shown herein.
  • the noise suppression system (DOMA) of an embodiment uses two combinations of the output of two omnidirectional microphones to form two virtual microphones.
  • the omnidirectional microphones In order to construct these virtual microphones, the omnidirectional microphones have to be accurately calibrated in both amplitude and phase so that they respond in both amplitude and phase as similarly as possible to the same acoustic input.
  • Many omnidirectional microphones use mechanical highpass (HP) filters (usually implemented using one or more holes in the diaphragm of the microphone) to reduce wind noise response. These mechanical filters commonly have responses similar to electronic RC filters, but small differences in the hole size and shape can lead to 3-dB frequencies that range from below 100 Hz more than 400 Hz. This difference can cause the relative phase response between the
  • An RC filter has the real-time response
  • f N is the 3-dB frequency for the Nth microphone and f s is the sampling frequency. This is now adjusted so that the magnitude matches better at low frequencies:
  • the 3-dB frequency of the microphone with white noise can be difficult to accurately determine the 3-dB frequency of the microphone with white noise because the power spectrum is only flat on average, and normally a long (15+ seconds) burst is needed to ensure acceptable spectral flatness.
  • the 3-dB frequency can be deduced by subtracting the recorded spectrum from the stored one.
  • the speaker and air transfer functions are unity, which is doubtful for low frequencies. It is possible to measure the speaker and air transfer functions for each box using a reference microphone, but if there is variance between calibration boxes then this could not be used as a general algorithm.
  • the initial calibration filter of an embodiment is determined using the unfiltered Oi and O 2 responses and an adaptive filter, as shown in Figure 14, but is not so limited.
  • the initial calibration filter relates one microphone (in this case, O 2 , but it can be any number of microphones) back to the reference microphone (in this case, O 1 ).
  • O 2 the output of O 2
  • O 1 the reference microphone
  • the output of O 2 is filtered using the initial calibration filter, the response should be the same as O 1 if the calibration process and filter are accurate.
  • the assumption is made that the peak in the calibration filter phase response below 500 Hz is due to the different 3-dB frequencies and roll-offs of the mechanical HP filters in the microphones.
  • the mechanical filter can be modeled with an RC filter model (or, for other mechanical filters, another mathematical model), then the peak value and location can be found mathematically and used to predict the locations of the individual microphone 3-dB frequencies. This has the advantage of not requiring a change to the calibration process but is not as accurate as other methods. A reduction in phase mismatch to less than +-5 degrees, though, will be accurate enough for most applications.
  • Equations 7 and 8 allow the calculation of f ⁇ and f 2 given f max and ⁇ max . Experimental testing has shown that these estimates are usually quite accurate, commonly within +-5 Hz. Then f t and f 2 can be used to calculate A 1 and A 2 in Equation 1 and thus the filter models in Equation 2.
  • Estimated values are calculated as above given the peak magnitude and location of the calibration filter. Using this information, the model magnitude and phase responses are shown along with the measured ones in Figures 4 and 5. The magnitude responses have been offset by a constant gain to make comparisons easier.
  • Figure 4 shows the magnitude response of the calibration filters from Figure 2 (solid) with the RC filter difference model results (dashed).
  • the RC filter responses have been offset with constant gains (+1.75, +0.25, and - 3.25 dB for headsets 6AB5, 6C93, and 90B9 respectively) and match very well with the observed responses.
  • the RC model fits the observed magnitude differences very well (within +-0.2 dB) with constant offsets.
  • Headset 6C83 had an offset of only 0.25 dB, indicating that with the exception of the 3-dB point, the microphones match very well in magnitude response.
  • their 3-dB frequencies are sufficiently different that they differ in magnitude by 4 dB at DC and -12.5 degrees at 250 Hz. For this headset, virtually all the mismatch is due to the difference in 3-dB frequency.
  • FIG. 5 shows the phase response of the calibration filters from
  • FIG. 3 (solid) with the RC filter difference model results (dashed).
  • the RC filter phase responses are very similar, within a few degrees below 1000 Hz.
  • headset 6C83 which had very little magnitude response difference above 1 kHz, has a very large phase difference.
  • Headsets 6AB5 and 90B9 has phase responses that trend toward zero degrees, as expected, but 90B9 does not, for unknown reasons.
  • this compensation method should significantly decrease the phase difference between the microphones.
  • the modeled phase outputs are very good matches at the peak (which just means the model is consistent) and within +-2 degrees below 500 Hz. This should be sufficient to bring the relative phase to within +-5 degrees.
  • This calibration method of an embodiment, referred to herein as the version 5 or v5 calibration method comprises: 1. Calculating the calibration filter a o (z) using O 1 Cz) and O 2 (z).
  • the minimum-phase filter a MP (z) may be transformed to a linear phase filter a LP (z) if desired.
  • the final application-ready calibrated outputs at this stage are thus
  • FIG. 6 is a flow diagram for calibration using a standard gain target for each branch, under an embodiment.
  • the delay "d” is the linear phase delay in samples of the alpha filter.
  • the alpha filter can be either linear phase or minimum phase.
  • the final filtering flow (pre-DOMA) is shown in Figure 6, where
  • the mechanical filter be constructed in such a way so that its response can be modeled using the RC model above.
  • Figure 7 shows original O 1 , O 2 , and compensated modeled responses for headset 90B9, under an embodiment.
  • the loss is 3.3 dB at 100 Hz, 1.1 dB at 200 Hz, and 0.4 dB at 300 Hz.
  • Figure 8 shows original Oi, O 2 , and compensated modeled responses for headset 6AB5, under an embodiment.
  • the loss is 6.4 dB at 100 Hz, 2.7 dB at 200 Hz, and 1.3 dB at 300 Hz.
  • Figure 9 shows original Oi, O 2 , and compensated modeled responses for headset 6C83, under an embodiment.
  • the loss is 9.4 dB at 100 Hz, 4.7 dB at 200 Hz, and 2.6 dB at 300 Hz.
  • Figure 10 shows the compensated Oi and O 2 responses for the three different headsets. There is a significant 7.0 dB difference between headset 90B9 (204) and 6C83 (206) at 100 Hz. This variation will depend on the initial Oi and O 2 responses as well as the 3-dB frequencies. If calibration is performed not to the Oi response but to a nominal value, this variation can be reduced, but some variation will always be present. In DOMA, though, some amplitude response variation below 500 Hz is preferable to large phase variations below 500 Hz, so even without normalizing the gains for the decreased response below 500 Hz the phase compensation is still worthwhile. Phase compensation test
  • Figure 11 shows the magnitude response of the calibration filter for the three headsets with factory calibrations before (solid) and after (dashed) compensation. There is little change except near DC, where the responses are reduced, as intended.
  • Figure 12 shows calibration phase response for the three headsets using factory calibrations (solid) and compensated Aliph calibrations
  • headset 90B9 the poorest performer
  • Headset 6AB5 which had very little phase below 500 Hz, has been increased and thus argues that phase responses below 5 degrees should not be adjusted.
  • the maximum in headset 6C83 has dropped from -12.5 degrees to -8- not as much as for headset 90B9, but still an improvement. To make sure the calibration or microphone drift was not to blame, the calibrations were run again on the headsets at Aliph.
  • FIG. 18 shows the responses calculated using the RC model above at 125 and 375 Hz for Oi, O 2 , and the combination of Oi and O 2 . Clearly, if one or both of the 3-dB frequencies is high, the resulting OiO 2 response is low.
  • Figure 19 shows just the response of the combination of Oi and O 2 and the boost needed to regain the response of a single-pole filter with a 3-dB frequency of 200 Hz.
  • the boost can vary between -1.1 and 12.0 dB depending on where the 3-dB frequencies of the filters in Oi and O 2 are, and the needed boost is independent of the difference in frequencies.
  • phase and frequency information is sent to the "Compensation Filter” subroutine, where fi and f 2 are calculated and the model filters O IHAT (Z) and O 2 HAT(Z) are generated.
  • the combination of O 1H AT(Z) and O 2H AT(Z) can lead to significant loss of response below 300 Hz, and the amount of loss depends on both the location of the 3-dB frequencies and their difference.
  • the next stage involves convolving Oi HA ⁇ (z) with O 2HA T(Z) and comparing it to a "Standard Response" filter (currently a 200 Hz single-pole highpass filter).
  • the linear phase FIR filter needed to correct the amplitude response of the combination of O IHAT (Z) and O 2HAT (Z) is then determined and output as H A c(z).
  • O IHAT (Z), O 2HAT (Z), and H A c(z) are used as shown in the bottom flow of Figure 14 to calculate the second calibration filter ⁇ M p(z), where "MP" denotes a minimum phase filter. That is, the filter is allowed to be nonlinear.
  • a third filter ⁇ LP (z) may also be generated by forcing the second filter ⁇ MP (z) to have linear phase with the same amplitude response, using standard techniques. It may also be truncated or zero-padded if desired. Either or both of these may be used in subsequent calculations depending on the application.
  • Figure 15 contains a flow diagram for operation of a microphone array using the calibration, under an embodiment. The minimum phase filter and its delay are used for the AVAD (acoustic voice activity detection) algorithm and the linear phase filter and its delay are used to form the virtual microphones for use in the DOMA denoising algorithm.
  • AVAD acoustic voice activity detection
  • the delays of 40 and 40.1 samples used in the top and bottom part of Figure 14 are specific to the system used for the embodiment and the algorithm is not so limited.
  • the delays used there are to time-align the signals before using them in the algorithm and should be adjusted for each embodiment to compensate for analog-to-digital channel delays and the like.
  • a (normally linear phase) "CaI chamber correction” filter as seen in Figure 14 can be used to correct for known calibration chamber issues.
  • This filter can be approximated by examining hundreds or thousands of calibration responses and looking for similarities in all responses or measured using a reference microphone or by other methods known to those skilled in the art. For optimal performance, this requires that each calibration chamber be set up in an identical manner as much as possible.
  • this correction filter is known, it is convolved with either the calibration filter ⁇ o (z) if the initial phase difference is between -5 and +3 degrees or the calibration filter ⁇ MP (z) otherwise.
  • This correction filter is optional and may be set to unity if desired.
  • the calibrated outputs of the system are where again, the minimum phase filter can be transformed to a linear phase filter of equivalent amplitude response if desired.
  • a method of reducing the phase variation of Oi and O 2 due to 3-dB frequency mismatches has been shown.
  • the method used is to estimate the 3-dB frequency of the microphones using the peak frequency and amplitude of the ⁇ o (z) peak below 500 Hz.
  • Estimates of the 3-dB frequencies for three different headsets yielded very accurate magnitude responses at all frequencies and good phase estimates below 1000 Hz.
  • Tests on three headsets showed good reduction of phase difference for headsets with significant (e.g., greater than +-6 deg) differences. This reduction in relative phase is often accompanied by a significant decrease in response below 500 Hz, but an algorithm has been presented that will restore the response to one that is desired, so that all compensated microphone combinations will end up with similar frequency responses. This is highly desirable in a consumer electronic product.
  • the version 5 (v5, ⁇ M p(z) used) calibration method or algorithm described above is a compensation subroutine that minimizes the amplitude and phase effects of mismatched mechanical filters in the microphones. These mismatched filters can cause variations of up to +-25 degrees in the phase and +-1OdB in the magnitude of the alpha filter at DC. These variations caused the noise suppression performance to vary by more than 21 dB and the devoicing performance to vary by more than 12 dB, causing significant variation in the speech and noise response of the headsets.
  • the effects that the v5 cal routine has on the amplitude and phase response mismatches are examined and the correlated denoising and devoicing performance compared to the previous conventional version 4 (v4, only ⁇ o (z) used) calibration method. These were tested first at Aliph using six headsets and then at the manufacturer using 100 headsets.
  • the v5 calibration algorithm was implemented and tested on six units.
  • Figure 20 shows magnitude response of six test headsets using v4
  • Figure 21 shows phase response of six test headsets using v4 (solid lines) and v5 (dashed). The large peaks below 500 Hz have been eliminated, reducing phase differences from 34 degrees to less than 7 degrees.
  • FIG. 22 shows a table of the approximate denoising, devoicing, and SNR increase in dB using headset 931B-v5 as the standard. Pathfinder-only denoising and devoicing changes were used to compile the table. SNR differences of up to 11 dB were compensated to within 0 to -3 dB of the standard headset. Denoising differences between calibration versions were up to 21 dB before and 2 dB after. Devoicing differences were up to 12 dB before and 2 dB after.
  • the average denoising at low frequencies varied by up to 21 dB between headsets using v4. In v5, that difference dropped to 2 dB.
  • Devoicing varied by up to 12 dB using v4; this was reduced to 2 dB in v5.
  • the large differences in denoising and devoicing manifest themselves not only in SNR differences, but the spectral tilt of the user's voice. Using v4, the spectral tilt could vary several dB at low frequencies, which means that a user could sound different on headsets with large phase and magnitude differences. With v5, a user will sound the same on any of the headsets.
  • the v5 headsets have no modulation, no swishing or musicality, significantly higher quality, intelligibility, and naturalness, and spectrally similar outputs.
  • the performance of the headsets was significantly better using v5 - even for the units that required no phase correction, due to the use of the standard response and the deletion of the phase of the anechoic/calibration chamber compensation filter.
  • phase responses for the v4 cal are shown in Figure 23.
  • This 38- degree spread (-21 to +17 degrees) is typical to what is normally observed with headsets using these microphones. These headsets would vary widely in their performance, even more than the 21 dB observed in the six headsets above.
  • the spread has been reduced to less than 10 degrees below 500 Hz, rendering these headsets practically indistinguishable in performance.
  • ripple in the phase response for v5. There was one headset that returned a spurious response (likely due to operator error) but it would have been caught by the v5 error-checking routine.
  • Figure 25 shows mean 2502, +-1 ⁇ 2504, and +-2 ⁇ 2506 of the magnitude (top) and phase (bottom) responses of 99 headsets using v4 calibration.
  • the 2 ⁇ spread in magnitude at DC is almost 13 dB, and for phase is 31 degrees. If +5 and -10 degrees are taken to be the cutoff for good performance, then about 40% of these headsets will have significantly poorer performance than the others.
  • Figure 26 shows mean 2602, +-1 ⁇ 2604, and +-2 ⁇ 2606 of the magnitude (top) and phase (bottom) responses of 99 headsets using v5 calibration.
  • the 2 ⁇ spread in magnitude at DC is now only 6 dB (within spec) with less ripple, and for phase is less than 7 degrees with significantly less ripple. These headsets should be indistinguishable in performance.
  • the mean 2502 and standard deviations (2504 for +-l ⁇ , 2506 for +- 2 ⁇ ) for the v4 cal in Figure 25 show that at DC there is a 13 dB difference in magnitude response and a 31 degree spread below 500 Hz for +-2 ⁇ . This is reduced to 6 dB in magnitude (which is the specification for the microphones, +-3 dB) and 7 degrees in phase for v5 shown in Figure 26. Also, there is significantly less ripple in both the magnitude and the phase responses. This is a phenomenal improvement in calibration accuracy and will significantly improve performance for all headsets.
  • any headset with a maximum phase more than 5 degrees is always reduced in phase difference. Between 3-5 degrees, there was some reduction in phase but some small increases (red text) as well. Below 3 degrees there was little change or a small increase. Thus 3 degrees is a good upper limit in determining whether or not to compensate for phase differences.
  • denoising artifacts such as swishing, musicality, and other irritants have been significantly reduced or eliminated.
  • the outgoing speech quality and intelligibility is significantly higher, even for units with small phase differences.
  • the spectral tilt of the microphones has been normalized, making the user sound more natural and making it easier to set the TX equalization.
  • the increase in performance and robustness that was realized with the use of the v5 calibration is significantly large.
  • the microphone outputs are normalized to a standard level so that the input to DOMA will be functionally identical for all headsets, further normalizing the user's speech so that it will sound more natural and uniform in all noise environments.
  • the v5 calibration routine described above significantly increased the performance of all headsets by a combination of eliminating phase and magnitude differences in the alpha filter caused by different mechanical HP filter 3-dB points. It also used a "Standard response" (i.e. a 200 Hz HP filter) to normalize the spectral response of Oi and O 2 for those units that were phase-corrected. However, it did not impose a standard gain (that is, the gain of Oi at 1 kHz could vary up to the spec, +-3 dB) and it also did not normalize the spectral response for units that did not require phase- correcting (units that had very small alpha filter phase peaks below 500 Hz).
  • Standard response i.e. a 200 Hz HP filter
  • the v4 calibration was a typical state-of-the-art microphone calibration system.
  • the two microphones to be calibrated were exposed to an acoustic source designed so that the acoustic input to the microphones was as similar as possible in both amplitude and phase.
  • the embodiment consisted of a 1 kHz sync tone and two 3-second white noise bursts (spectrally flat between approximately 125 Hz and 3875 Hz) separated by 1 second of silence.
  • White noise was used to equally weight the spectrums of the microphones to make the adaptive filter algorithm as accurate as possible.
  • the input to the microphones may be whitened further using a reference microphone to record and compensate for any non-ideal response from the loudspeaker used, as known to those skilled in the art.
  • Version 6 is relatively simple in that only one extra step is required from v5, and it is only required for arrays that do not require compensation - that is, phase-matched arrays whose maximum phase below 500 Hz is less than three degrees and greater than negative 5 degrees.
  • the second white noise burst instead of using the second white noise burst to calculate O iHAT , O 2HAT , and H AC , we can use it to impose the "Standard response" in Figure 14 on the phase-matched headsets.
  • a compensation filter H B c(z) is generated using the difference between the "Standard Response” and the calculated responses, and both calibrated outputs are filtered with the H BC (z) filter to recover the standard response.
  • the v6 outputs are where again, only the arrays that did not need phase compensation are used.
  • the calibrated outputs of both v5 and v6 can be normalized to the same gain at a fixed frequency - we have used 750 Hz to good effect. However, this is not required, as manufacturing
  • Figure 29 shows a flow chart of the v6 algorithm where arrays without significant phase difference also get normalized to the standard response, under an embodiment.
  • the recorded responses of Oi from the second burst of white noise are analyzed using any standard algorithm (such as the PSD) to calculate the approximate amplitude response of Oi(z).
  • the difference between the Oi amplitude response and the desired "Standard response" (in our case, a first-order highpass RC filter with a 3-dB frequency of 200 Hz) is used to generate the compensation filter H B c(z), which is then used to filter both calibrated outputs from v5.
  • v5 and v6 calibration algorithms described above are effective at normalizing the response of the microphones and reducing the effect of mismatched 3-dB frequencies on the alpha phase and amplitude near DC. But, they require the unit to be re-calibrated, and this is difficult to accomplish for previously-shipped headsets. While these shipped headsets cannot all be recalibrated, they still may gain some performance just from the reduction of the phase and magnitude differences.
  • v5 algorithm described herein reduces the amplitude and phase mismatches by determining the 3-dB frequencies fi and f 2 for Oi and O 3 . Then, RC models of the mechanical filters are constructed, as described herein, using :
  • the compensation filter ⁇ c (z) is therefore
  • Figure 31 shows a flow diagram for the v4.1 calibration algorithm, under an embodiment. Since no new information is possible, the benefits are limited to OIHAT / O 2 HAT, and H A c(z) for units that have sufficient alpha phase.
  • Figure 32 shows use of the new filters prior to the DOMA and AVAD algorithms. The implementation of O m a t; O 2hat , and H A c into the DOMA and AVAD algorithms is unchanged from v5.
  • DQMA Dual Omnidirectional Microphone Array
  • a dual omnidirectional microphone array that provides improved noise suppression is described herein.
  • Numerous systems and methods for calibrating the DOMA was described above. Compared to conventional arrays and algorithms, which seek to reduce noise by nulling out noise sources, the array of an embodiment is used to form two distinct virtual directional microphones which are configured to have very similar noise responses and very dissimilar speech responses. The only null formed by the DOMA is one used to remove the speech of the user from V 2 .
  • the two virtual microphones of an embodiment can be paired with an adaptive filter algorithm and/or VAD algorithm to significantly reduce the noise without distorting the speech, significantly improving the SNR of the desired speech over conventional noise suppression systems.
  • the embodiments described herein are stable in operation, flexible with respect to virtual microphone pattern choice, and have proven to be robust with respect to speech source- to-array distance and orientation as well as temperature and calibration techniques. Numerous systems and methods for calibrating the DOMA was described above.
  • Figure 33 is a two-microphone adaptive noise suppression system 3300, under an embodiment.
  • the two-microphone system 3300 including the combination of physical microphones MIC 1 and MIC 2 along with the processing or circuitry components to which the microphones couple
  • the dual omnidirectional microphone array (DOMA) 3310 is referred to herein as the dual omnidirectional microphone array (DOMA) 3310, but the embodiment is not so limited.
  • the total acoustic information coming into MIC 1 (3302, which can be an physical or virtual microphone) is denoted by m ⁇ n).
  • the total acoustic information coming into MIC 2 (103, which can also be an physical or virtual microphone) is similarly labeled m 2 (n).
  • m ⁇ n The total acoustic information coming into MIC 2 (103, which can also be an physical or virtual microphone
  • M 1 (Z) and M 2 (z) In the z (digital frequency) domain, these are represented as M 1 (Z) and M 2 (z).
  • Equation 1 This is the general case for all two microphone systems. Equation 1 has four unknowns and only two known relationships and therefore cannot be solved explicitly.
  • Equation 1 there is another way to solve for some of the unknowns in Equation 1.
  • the analysis starts with an examination of the case where the speech is not being generated, that is, where a signal from the VAD subsystem 3304 (optional) equals zero.
  • the function H 1 (Z) can be calculated using any of the available system identification algorithms and the microphone outputs when the system is certain that only noise is being received. The calculation can be done adaptively, so that the system can react to changes in the noise.
  • H 1 (Z) one of the unknowns in Equation 1.
  • Equation 1 After calculating H 1 (Z) and H 2 (z), they are used to remove the noise from the signal. If Equation 1 is rewritten as
  • N(z) may be substituted as shown to solve for S(z) as If the transfer functions H 1 (Z) and H 2 (z) can be described with sufficient accuracy, then the noise can be completely removed and the original signal recovered. This remains true without respect to the amplitude or spectral characteristics of the noise. If there is very little or no leakage from the speech source into M 2 , then H 2 (z) » 0 and Equation 3 reduces to
  • Equation 4 is much simpler to implement and is very stable, assuming Hi(z) is stable. However, if significant speech energy is in M 2 (z), devoicing can occur. In order to construct a well-performing system and use Equation 4, consideration is given to the following conditions:
  • H 1 (Z) cannot change substantially.
  • Condition Rl is easy to satisfy if the SNR of the desired speech to the unwanted noise is high enough. "Enough” means different things depending on the method of VAD generation. If a VAD vibration sensor is used, as in Burnett 7,256,048, accurate VAD in very low SNRs (-10 dB or less) is possible. Acoustic-only methods using information from O 1 and O 2 can also return accurate VADs, but are limited to SNRs of ⁇ 3 dB or greater for adequate performance.
  • Condition R5 is normally simple to satisfy because for most
  • the DOMA in various embodiments, can be used with the Pathfinder system as the adaptive filter system or noise removal.
  • the Pathfinder system available from AliphCom, San Francisco, CA, is described in detail in other patents and patent applications referenced herein.
  • any adaptive filter or noise removal algorithm can be used with the DOMA in one or more various alternative embodiments or configurations.
  • the Pathfinder system When the DOMA is used with the Pathfinder system, the Pathfinder system generally provides adaptive noise cancellation by combining the two microphone signals (e.g., Micl, Mic2) by filtering and summing in the time domain.
  • the adaptive filter generally uses the signal received from a first microphone of the DOMA to remove noise from the speech received from at least one other microphone of the DOMA, which relies on a slowly varying linear transfer function between the two microphones for sources of noise.
  • an output signal is generated in which the noise content is attenuated with respect to the speech content, as described in detail below.
  • Figure 34 is a generalized two-microphone array (DOMA) including an array 3401/3402 and speech source S configuration, under an embodiment.
  • Figure 35 is a system 3500 for generating or producing a first order gradient microphone V using two omnidirectional elements Oi and O 2 , under an embodiment.
  • the array of an embodiment includes two physical
  • microphones 3401 and 3402 e.g., omnidirectional microphones placed a distance 2d 0 apart and a speech source 3400 is located a distance d s away at an angle of ⁇ .
  • This array is axially symmetric (at least in free space), so no other angle is needed.
  • the output from each microphone 3401 and 3402 can be delayed (Z 1 and Z 2 ), multiplied by a gain (Ai and A 2 ), and then summed with the other as demonstrated in Figure 35.
  • the output of the array is or forms at least one virtual microphone, as described in detail below. This operation can be over any frequency range desired.
  • VMs virtual microphones
  • VMs can be realized. There are other methods known to those skilled in the art for constructing VMs but this is a common one and will be used in the enablement below.
  • Figure 36 is a block diagram for a DOMA 3600 including two physical microphones configured to form two virtual
  • the DOMA includes two first order gradient microphones Vi and V 2 formed using the outputs of two microphones or elements Oi and O 2 (3401 and 3402), under an embodiment.
  • the DOMA of an embodiment includes two physical microphones 3401 and 3402 that are omnidirectional microphones, as described above with reference to Figures 34 and 35.
  • the output from each microphone is coupled to a processing component 3602, or circuitry, and the processing component outputs signals representing or corresponding to the virtual microphones Vi and V 2 .
  • the output of physical microphone 3401 is coupled to processing component 3602 that includes a first processing path that includes application of a first delay Z 11 and a first gain A n and a second processing path that includes application of a second delay Z 12 and a second gain Ai 2 .
  • the output of physical microphone 3402 is coupled to a third processing path of the processing component 3602 that includes application of a third delay z 2i and a third gain A 2 i and a fourth processing path that includes application of a fourth delay Z 22 and a fourth gain A 22 .
  • the output of the first and third processing paths is summed to form virtual microphone Vi, and the output of the second and fourth processing paths is summed to form virtual microphone V 2 .
  • VMs virtual microphones
  • Figure 37 is a block diagram for a DOMA 3700 including two physical microphones configured to form N virtual microphones V 1 through V N , where N is any number greater than one, under an embodiment.
  • the DOMA can include a processing component 3702 having any number of processing paths as appropriate to form a number N of virtual microphones.
  • the DOMA of an embodiment can be coupled or connected to one or more remote devices.
  • the DOMA outputs signals to the remote devices.
  • the remote devices include, but are not limited to, at least one of cellular telephones, satellite telephones, portable telephones, wireline telephones, Internet telephones, wireless transceivers, wireless communication radios, personal digital assistants (PDAs), personal computers (PCs), headset devices, head-worn devices, and earpieces.
  • the DOMA of an embodiment can be a component or subsystem integrated with a host device.
  • the DOMA outputs signals to components or subsystems of the host device.
  • the host device includes, but is not limited to, at least one of cellular telephones, satellite telephones, portable telephones, wireline telephones, Internet telephones, wireless transceivers, wireless communication radios, personal digital assistants (PDAs), personal computers (PCs), headset devices, head- worn devices, and earpieces.
  • Figure 38 is an example of a headset or head-worn device 3800 that includes the DOMA, as described herein, under an embodiment.
  • the headset 3800 of an embodiment includes a housing having two areas or receptacles (not shown) that receive and hold two microphones (e.g., O 1 and O 2 ).
  • the headset 3800 is generally a device that can be worn by a speaker 3802, for example, a headset or earpiece that positions or holds the microphones in the vicinity of the speaker's mouth.
  • the headset 3800 of an embodiment places a first physical microphone (e.g., physical microphone Oi) in a vicinity of a speaker's lips.
  • a second physical microphone e.g., physical microphone O 2
  • the distance of an embodiment is in a range of a few centimeters behind the first physical microphone or as described herein (e.g., described with reference to Figures 33-37).
  • the DOMA is symmetric and is used in the same configuration or manner as a single close-talk microphone, but is not so limited.
  • FIG 39 is a flow diagram for denoising 3900 acoustic signals using the DOMA, under an embodiment.
  • the denoising 3900 begins by receiving 3902 acoustic signals at a first physical microphone and a second physical microphone. In response to the acoustic signals, a first microphone signal is output from the first physical microphone and a second microphone signal is output from the second physical microphone 3904.
  • a first virtual microphone is formed 3906 by generating a first combination of the first microphone signal and the second microphone signal.
  • a second virtual microphone is formed 3908 by generating a second combination of the first microphone signal and the second microphone signal, and the second combination is different from the first combination.
  • the first virtual microphone and the second virtual microphone are distinct virtual directional microphones with substantially similar responses to noise and substantially dissimilar responses to speech.
  • the denoising 3900 generates 3910 output signals by combining signals from the first virtual microphone and the second virtual microphone, and the output signals include less acoustic noise than the acoustic signals.
  • Figure 40 is a flow diagram for forming 4000 the DOMA, under an embodiment.
  • Formation 4000 of the DOMA includes forming 4002 a physical microphone array including a first physical microphone and a second physical microphone. The first physical microphone outputs a first microphone signal and the second physical microphone outputs a second microphone signal.
  • a virtual microphone array is formed 4004 comprising a first virtual microphone
  • the first virtual microphone comprises a first combination of the first microphone signal and the second microphone signal.
  • the second virtual microphone comprises a second combination of the first microphone signal and the second microphone signal, and the second combination is different from the first combination.
  • the virtual microphone array including a single null oriented in a direction toward a source of speech of a human speaker.
  • VMs for the adaptive noise suppression system of an embodiment includes substantially similar noise response in Vi and V 2 .
  • substantially similar noise response as used herein means that Hi(z) is simple to model and will not change much during speech, satisfying conditions R2 and R4 described above and allowing strong denoising and minimized bleedthrough.
  • the construction of VMs for the adaptive noise suppression system of an embodiment includes relatively small speech response for V 2 .
  • the relatively small speech response for V 2 means that H 2 (z) « o, which will satisfy conditions R3 and R5 described above.
  • VMs for the adaptive noise suppression system of an embodiment further includes sufficient speech response for Vi so that the cleaned speech will have significantly higher SNR than the original speech captured by Oi.
  • V 2 (z) V 2 (z)
  • the distances di and d 2 are the distance from Oi and O 2 to the speech source (see Figure 34), respectively, and ⁇ is their difference divided by c, the speed of sound, and multiplied by the sampling frequency f s .
  • is in samples, but need not be an integer.
  • fractional-delay filters (well known to those versed in the art) may be used.
  • the ⁇ above is not the conventional ⁇ used to denote the mixing of VMs in adaptive beamforming; it is a physical variable of the system that depends on the intra-microphone distance d 0 (which is fixed) and the distance d s and angle ⁇ , which can vary. As shown below, for properly calibrated microphones, it is not necessary for the system to be programmed with the exact ⁇ of the array. Errors of approximately 10- 15% in the actual ⁇ (i.e. the ⁇ used by the algorithm is not the ⁇ of the physical array) have been used with very little degradation in quality.
  • the algorithmic value of ⁇ may be calculated and set for a particular user or may be calculated adaptively during speech production when little or no noise is present. However, adaptation during use is not required for nominal performance.
  • the linear response of V 2 to noise is devoid of or includes no null, meaning all noise sources are detected.
  • V 2 (z) has a null at the speech location and will therefore exhibit minimal response to the speech. This is shown in
  • Vi(z) can be formulated using the general form for Vi(z):
  • Vi and V 2 above mean that for noise H 1 (Z) is: which, if the amplitude noise responses are about the same, has the form of an allpass filter. This has the advantage of being easily and accurately modeled, especially in magnitude response, satisfying R2.
  • is the ratio of the distances from O 1 and O 2 to the speech source, it is affected by the size of the array and the distance from the array to the speech source.
  • the linear response of virtual microphone V 1 to speech is devoid of or includes no null and the response for speech is greater than that shown in Figure 4.
  • the linear response of virtual microphone Vi to noise is devoid of or includes no null and the response is very similar to V 2 shown in Figure 5.
  • Figure 46 is a plot showing comparison of frequency responses for speech for the array of an embodiment and for a conventional cardioid microphone.
  • the response of Vi to speech is shown in Figure 43, and the response to noise in Figure 44. Note the difference in speech response compared to V 2 shown in Figure 9 and the similarity of noise response shown in Figure 42. Also note that the orientation of the speech response for V 1 shown in Figure 43 is completely opposite the orientation of conventional systems, where the main lobe of response is normally oriented toward the speech source.
  • the orientation of an embodiment, in which the main lobe of the speech response of Vi is oriented away from the speech source, means that the speech sensitivity Of V 1 is lower than a normal directional microphone but is flat for all frequencies within approximately +-30 degrees of the axis of the array, as shown in Figure 45. This flatness of response for speech means that no shaping postfilter is needed to restore omnidirectional frequency response.
  • the speech response Of V 1 is approximately 0 to ⁇ 13 dB less than a normal directional microphone between approximately 500 and 7500 Hz and approximately 0 to 10+ dB greater than a directional microphone below approximately 500 Hz and above 7500 Hz for a sampling frequency of approximately 16000 Hz.
  • the superior noise suppression made possible using this system more than compensates for the initially poorer SNR.
  • the noise distance is not required to be 1 m or more, but the denoising is the best for those distances. For distances less than approximately 1 m, denoising will not be as effective due to the greater dissimilarity in the noise responses of Vi and V 2 . This has not proven to be an impediment in practical use - in fact, it can be seen as a feature. Any "noise" source that is ⁇ 10 cm away from the earpiece is likely to be desired to be captured and transmitted.
  • the speech null of V 2 means that the VAD signal is no longer a critical component.
  • the VAD's purpose was to ensure that the system would not train on speech and then subsequently remove it, resulting in speech distortion. If, however, V 2 contains no speech, the adaptive system cannot train on the speech and cannot remove it. As a result, the system can denoise all the time without fear of devoicing, and the resulting clean audio can then be used to generate a VAD signal for use in subsequent single- channel noise suppression algorithms such as spectral subtraction.
  • constraints on the absolute value of Hi(z) i.e. restricting it to absolute values less than two) can keep the system from fully training on speech even if it is detected. In reality, though, speech can be present due to a mis-located V 2 null and/or echoes or other phenomena, and a VAD sensor or other acoustic-only VAD is recommended to minimize speech distortion.
  • ⁇ and ⁇ may be fixed in the noise suppression algorithm or they can be estimated when the algorithm indicates that speech production is taking place in the presence of little or no noise. In either case, there may be an error in the estimate of the actual ⁇ and ⁇ of the system. The following description examines these errors and their effect on the performance of the system. As above, "good performance" of the system indicates that there is sufficient denoising and minimal devoicing.
  • FIG. 47 is a plot showing speech response for Vi (top, dashed) and V 2 (bottom, solid) versus B with d s assumed to be 0.1 m, under an embodiment. This plot shows the spatial null in V 2 to be relatively broad.
  • Figure 48 is a plot showing a ratio of Vi/V 2 speech responses shown in Figure 42 versus B, under an
  • the B factor can be non-unity for a variety of reasons. Either the distance to the speech source or the relative orientation of the array axis and the speech source or both can be different than expected. If both distance and angle mismatches are included for B, then where again the T subscripts indicate the theorized values and R the actual values.
  • Figure 50 shows what happens if the speech source is located at a distance of approximately 10 cm but not on the axis of the array.
  • the angle can vary up to approximately +-55 degrees and still result in a B less than 1.1, assuring good performance. This is a significant amount of allowable angular deviation. If there is both angular and distance errors, the equation above may be used to determine if the deviations will result in adequate performance. Of course, if the value for ⁇ ⁇ is allowed to update during speech, essentially tracking the speech source, then B can be kept near unity for almost all configurations.
  • N(z) Bz "YD - 1
  • N(s) Be- Ds - l .
  • is the time difference between arrival of speech at V 1 compared to V 2 , it can be errors in estimation of the angular location of the speech source with respect to the axis of the array and/or by temperature changes.
  • the speed of sound varies with temperature as where T is degrees Celsius. As the temperature decreases, the speed of sound also decreases.
  • Setting 20 C as a design temperature and a maximum expected temperature range to -40 C to +60 C (-40 F to 140 F).
  • the design speed of sound at 20 C is 343 m/s and the slowest speed of sound will be 307 m/s at -40 C with the fastest speed of sound 362 m/s at 60 C.
  • Set the array length (2d 0 ) to be 21 mm.
  • the resulting phase difference clearly affects high frequencies more than low.
  • the amplitude response is less than approximately -10 dB for all frequencies less than 7 kHz and is only about -9 dB at 8 kHz.
  • Non-unity B affects the entire frequency range.
  • N(s) is below approximately -10 dB only for frequencies less than approximately 5 kHz and the response at low
  • a temperature sensor may be integrated into the system to allow the algorithm to adjust ⁇ ⁇ as the temperature varies.
  • D can be non-zero
  • the speech source is not where it is believed to be - specifically, the angle from the axis of the array to the speech source is incorrect.
  • the distance to the source may be incorrect as well, but that introduces an error in B, not D.
  • the cancellation is still below -10 dB for frequencies below 6 kHz.
  • the cancellation is still below approximately -10 dB for frequencies below approximately 6 kHz, so an error of this type will not significantly affect the performance of the system.
  • ⁇ 2 is increased to approximately 45 degrees, as shown in Figure 54, the cancellation is below approximately -10 dB only for frequencies below approximately 2.8 kHz.
  • the cancellation is below -10 dB only for frequencies below about 2.8 kHz and a reduction in performance is expected.
  • the poor V 2 speech cancellation above approximately 4 kHz may result in significant devoicing for those frequencies.
  • the ⁇ of the system should be fixed and as close to the real value as possible. In practice, the system is not sensitive to changes in ⁇ and errors of approximately +-5% are easily tolerated. During times when the user is producing speech but there is little or no noise, the system can train ⁇ (z) to remove as much speech as possible. This is accomplished by: 1. Construct an adaptive system as shown in Figure 33 with ⁇ Oi S (z)z - ⁇ in the "MICl" position, O 2S (z) in the "MIC2" position, and ⁇ (z) in the Hi(z) position.
  • a simple adaptive filter can be used for ⁇ (z) so that only the
  • the system of an embodiment trains only when speech is being produced by the user.
  • a sensor like the SSM is invaluable in determining when speech is being produced in the absence of noise. If the speech source is fixed in position and will not vary significantly during use (such as when the array is on an earpiece), the adaptation should be infrequent and slow to update in order to minimize any errors introduced by noise present during training.
  • This formulation also allows the virtual microphone responses to be varied but retains the all-pass characteristic of H 1 (Z).
  • the system is flexible enough to operate well at a variety of Bl values, but B2 values should be close to unity to limit devoicing for best performance.
  • Embodiments described herein include a method executing on a processor, the method comprising inputting a signal into a first microphone and a second microphone.
  • the method of an embodiment comprises determining a first response of the first microphone to the signal.
  • the method of an embodiment comprises determining a second response of the second microphone to the signal.
  • the method of an embodiment comprises generating a first filter model of the first microphone and a second filter model of the second microphone from the first response and the second response.
  • the method of an embodiment comprises forming a calibrated microphone array by applying the second filter model to the first response of the first microphone and applying the first filter model to the second response of the second microphone.
  • Embodiments described herein include a method executing on a processor, the method comprising: inputting a signal into a first microphone and a second microphone; determining a first response of the first
  • the microphone to the signal; determining a second response of the second microphone to the signal; generating a first filter model of the first microphone and a second filter model of the second microphone from the first response and the second response; and forming a calibrated microphone array by applying the second filter model to the first response of the first microphone and applying the first filter model to the second response of the second microphone.
  • the method of an embodiment comprises generating a third filter model that normalizes the first response and the second response.
  • the generating of the third filter model of an embodiment comprises convolving the first filter model with the second filter model.
  • the method of an embodiment comprises comparing a result of the convolving with a standard response filter.
  • the standard response filter of an embodiment comprises a highpass filter having a pole at a frequency of approximately 200 Hertz.
  • the third filter model of an embodiment corrects an amplitude response of the result of the convolving.
  • the third filter model of an embodiment is a linear phase finite impulse response (FIR) filter.
  • the method of an embodiment comprises applying the third filter model to a signal resulting from the applying of the second filter model to the first response of the first microphone.
  • the method of an embodiment comprises applying the third filter model to a signal resulting from the applying of the first filter model to the second response of the second microphone.
  • the method of an embodiment comprises inputting a second signal into the system.
  • the method of an embodiment comprises determining a third response of the first microphone by applying the second filter model and the third filter model to an output of the first microphone resulting from the second signal.
  • the method of an embodiment comprises determining a fourth response of the second microphone by applying the first filter model and the third filter model to an output of the second microphone resulting from the second signal.
  • the method of an embodiment comprises generating a fourth filter model from a combination of the third response and the fourth response.
  • the generating of the fourth filter model of an embodiment comprises applying an adaptive filter to the third response and the fourth response.
  • the fourth filter model of an embodiment is a minimum phase filter model.
  • the method of an embodiment comprises generating a fifth filter model from the fourth filter model.
  • the fifth filter model of an embodiment is a linear phase filter model.
  • Forming the calibrated microphone array of an embodiment comprises applying the third filter model to at least one of an output of the first filter model and an output of the second filter model.
  • Forming the calibrated microphone array of an embodiment comprises applying the third filter model to the output of the first filter model and the output of the second filter model.
  • the method of an embodiment comprises applying the second filter model and the third filter model to a signal output of the first microphone.
  • the method of an embodiment comprises applying the first filter model, the third filter model and the fifth filter model to a signal output of the second microphone.
  • the calibrated microphone array of an embodiment comprises amplitude response calibration and phase response calibration.
  • the method of an embodiment comprises generating a first delayed first microphone signal by applying a first delay filter to the first microphone signal.
  • the method of an embodiment comprises inputting the first delayed first microphone signal to a processing component, wherein the processing component generates a virtual microphone array comprising a first virtual microphone and a second virtual microphone.
  • the method of an embodiment comprises generating a second microphone signal by applying the first filter model, the third filter model and the fifth filter model to a signal output of the second microphone.
  • the method of an embodiment comprises inputting the second microphone signal to the processing component.
  • the method of an embodiment comprises generating a second delayed first microphone signal by applying a second delay filter to the first microphone signal.
  • the method of an embodiment comprises inputting the second delayed first microphone signal to an acoustic voice activity detector.
  • the microphone signal by applying the first filter model, the third filter model and the fourth filter model to a signal output of the second microphone.
  • the method of an embodiment comprises inputting the third microphone signal to the acoustic voice activity detector.
  • the microphone signal by applying the second filter model and the third filter model to a signal output of the first microphone.
  • the method of an embodiment comprises generating a second microphone signal by applying the first filter model, the third filter model and the fifth filter model to a signal output of the second microphone.
  • the method of an embodiment comprises forming a first virtual microphone by generating a first combination of the first microphone signal and the second microphone signal.
  • the method of an embodiment comprises forming a second virtual microphone by generating a second combination of the first microphone signal and the second microphone signal, wherein the second combination is different from the first combination, wherein the first virtual microphone and the second virtual microphone are distinct virtual directional microphones with substantially similar responses to noise and substantially dissimilar responses to speech.
  • Forming the first virtual microphone of an embodiment includes forming the first virtual microphone to have a first linear response to speech that is devoid of a null, wherein the speech is human speech.
  • Forming the second virtual microphone of an embodiment includes forming the second virtual microphone to have a second linear response to speech that includes a single null oriented in a direction toward a source of the speech.
  • the single null of an embodiment is a region of the second linear response having a measured response level that is lower than the measured response level of any other region of the second linear response.
  • the second linear response of an embodiment includes a primary lobe oriented in a direction away from the source of the speech.
  • the primary lobe of an embodiment is a region of the second linear response having a measured response level that is greater than the measured response level of any other region of the second linear response.
  • the second signal of an embodiment is a white noise signal.
  • the generating of the first filter model and the second filter model of an embodiment comprises: calculating a calibration filter by applying an adaptive filter to the first response and the second response;
  • the largest peak is a largest peak located below a frequency of approximately 500 Hertz.
  • the generating of the first filter model and the second filter model comprises using unity filters for each of the first filter model, the second filter model and the third filter model.
  • the method of an embodiment comprises, when a largest phase variation of the calibration filter is greater than three degrees, calculating a first frequency corresponding to the first microphone and a second frequency corresponding to the second microphone.
  • the first frequency and the second frequency of an embodiment is a 3- decible frequency.
  • the generating of the first filter model and the second filter model of an embodiment comprises using the first frequency and the second frequency to generate the first filter model and the second filter model.
  • the second filter model of an embodiment is an infinite impulse response (HR) model.
  • the signal of an embodiment is a white noise signal.
  • Embodiments described herein include a system comprising a microphone array comprising a first microphone and a second microphone.
  • the system of an embodiment comprises a first filter coupled to an output of the second microphone.
  • the first filter models a response of the first microphone to a noise signal.
  • the system of an embodiment comprises a second filter coupled to an output of the first microphone.
  • the second filter models a response of the second microphone to the noise signal.
  • the system of an embodiment comprises a processor coupled to the first filter and the second filter.
  • Embodiments described herein include a system comprising: a microphone array comprising a first microphone and a second microphone; a first filter coupled to an output of the second microphone, wherein the first filter models a response of the first microphone to a noise signal; a second filter coupled to an output of the first microphone, wherein the second filter models a response of the second microphone to the noise signal; and a processor coupled to the first filter and the second filter.
  • the system of an embodiment comprises a third filter coupled to an output of at least one of the first filter and the second filter.
  • the third filter of an embodiment normalizes the first response and the second response.
  • the third filter of an embodiment is generated by convolving a response of the first filter with a response of the second filter and comparing a result of the convolving with a standard response filter.
  • the third filter of an embodiment corrects an amplitude response of the result of the convolving.
  • the third filter of an embodiment is a linear phase finite impulse response (FIR) filter.
  • the system of an embodiment comprises coupling the third filter to an output of the second filter.
  • the system of an embodiment comprises coupling the third filter to an output of the first filter.
  • the system of an embodiment comprises a fourth filter coupled to an output of the third filter that is coupled to the second microphone.
  • the fourth filter of an embodiment is a minimum phase filter.
  • the fourth filter of an embodiment is generated by: determining a third response of the first microphone by applying a response of the second filter and a response of the third filter to an output of the first microphone resulting from a second signal; determining a fourth response of the second microphone by applying a response of the first filter and a response of the third filter to an output of the second microphone resulting from the second signal; and generating the fourth filter from a combination of the third response and the fourth response.
  • the generating of the fourth filter of an embodiment comprises applying an adaptive filter to the third response and the fourth response.
  • the system of an embodiment comprises a fifth filter that is a linear phase filter.
  • the fifth filter of an embodiment is generated from the fourth filter.
  • the system of an embodiment comprises at least one of the fourth filter and the fifth filter coupled to an output of the third filter that is coupled to the first filter and the second microphone.
  • the system of an embodiment comprises outputting a first microphone signal from a signal path including the first microphone coupled to the second filter and the third filter.
  • the system of an embodiment comprises
  • the system of an embodiment comprises inputting the first delayed first microphone signal to the processor, wherein the processor generates a virtual microphone array comprising a first virtual microphone and a second virtual microphone.
  • the system of an embodiment comprises outputting a second microphone signal from a signal path including the second microphone coupled to the first filter, the third filter and the fifth filter.
  • the system of an embodiment comprises inputting the second microphone signal to the processor.
  • the system of an embodiment comprises generating a second delayed first microphone signal by applying a second delay filter to the first
  • the system of an embodiment comprises inputting the second delayed first microphone signal to an acoustic voice activity detector (AVAD).
  • AVAD acoustic voice activity detector
  • the system of an embodiment comprises inputting the third microphone signal to the acoustic voice activity detector.
  • the system of an embodiment comprises outputting a first microphone signal from a signal path including the first microphone coupled to the second filter and the third filter.
  • the system of an embodiment comprises outputting a second microphone signal from a signal path including the second microphone coupled to the first filter, the third filter and the fifth filter.
  • the system of an embodiment comprises a first virtual microphone, wherein the first virtual microphone is formed by generating a first
  • the system of an embodiment comprises a second virtual microphone, wherein the second virtual microphone is formed by generating a second combination of the first microphone signal and the second microphone signal, wherein the second combination is different from the first combination, wherein the first virtual microphone and the second virtual microphone are distinct virtual directional microphones with substantially similar responses to noise and substantially dissimilar responses to speech.
  • Forming the first virtual microphone of an embodiment includes forming the first virtual microphone to have a first linear response to speech that is devoid of a null, wherein the speech is human speech.
  • Forming the second virtual microphone of an embodiment includes forming the second virtual microphone to have a second linear response to speech that includes a single null oriented in a direction toward a source of the speech.
  • the single null of an embodiment is a region of the second linear response having a measured response level that is lower than the measured response level of any other region of the second linear response.
  • the second linear response of an embodiment includes a primary lobe oriented in a direction away from the source of the speech.
  • the primary lobe of an embodiment is a region of the second linear response having a measured response level that is greater than the measured response level of any other region of the second linear response.
  • Generating the first filter and the second filter of an embodiment comprises: calculating a calibration filter by applying an adaptive filter to the first response and the second response; and determining a peak magnitude and a peak location of a largest peak of the calibration filter, wherein the largest peak is a largest peak located below a frequency of approximately 500 Hertz.
  • the generating of the first filter and the second filter comprises using unity filters for each of the first filter, the second filter and the third filter.
  • the system of an embodiment comprises, when a largest phase variation of the calibration filter is greater than positive three (3) degrees, calculating a first frequency corresponding to the first microphone and a second frequency corresponding to the second microphone.
  • embodiment is a three-decible frequency.
  • invention comprises using the first frequency and the second frequency to generate the first filter and the second filter.
  • the first filter of an embodiment is an infinite impulse response (HR) filter.
  • the second filter of an embodiment is an infinite impulse response (IIR) filter.
  • the signal of an embodiment is a white noise signal.
  • the microphone array of an embodiment comprises amplitude response calibration and phase response calibration.
  • Embodiments described herein include a system comprising a microphone array comprising a first microphone and a second microphone.
  • the system of an embodiment comprises a first filter coupled to an output of the second microphone.
  • the first filter models a response of the first microphone to a noise signal and outputs a second microphone signal.
  • the system of an embodiment comprises a second filter coupled to an output of the first microphone.
  • the second filter models a response of the second microphone to the noise signal and outputs a first microphone signal.
  • the first microphone signal is calibrated with the second microphone signal.
  • the system of an embodiment comprises a processor coupled to the microphone array and generating from the first microphone signal and the second microphone signal a virtual microphone array comprising a first virtual microphone and a second virtual microphone.
  • Embodiments described herein include a system comprising: a microphone array comprising a first microphone and a second microphone; a first filter coupled to an output of the second microphone, wherein the first filter models a response of the first microphone to a noise signal and outputs a second microphone signal; a second filter coupled to an output of the first microphone, wherein the second filter models a response of the second microphone to the noise signal and outputs a first microphone signal, wherein the first microphone signal is calibrated with the second microphone signal; and a processor coupled to the microphone array and generating from the first microphone signal and the second microphone signal a virtual microphone array comprising a first virtual microphone and a second virtual microphone.
  • the system of an embodiment comprises a third filter coupled to an output of at least one of the first filter and the second filter.
  • the third filter of an embodiment normalizes the first response and the second response.
  • the third filter of an embodiment is a linear phase finite impulse response (FIR) filter.
  • the third filter of an embodiment is coupled to an output of the second filter.
  • the third filter of an embodiment is coupled to an output of the first filter.
  • the system of an embodiment comprises a fourth filter coupled to an output of a signal path including the third filter and the second microphone.
  • the fourth filter of an embodiment is a minimum phase filter.
  • the fifth filter of an embodiment is a linear phase filter.
  • the fifth filter of an embodiment is derived from the fourth filter.
  • the system of an embodiment comprises at least one of the fourth filter and the fifth filter coupled to an output of a signal path including the third filter, the first filter and the second microphone.
  • the system of an embodiment comprises outputting a first microphone signal from a signal path including the first microphone coupled to the second filter and the third filter.
  • the system of an embodiment comprises
  • the system of an embodiment comprises inputting the first delayed first microphone signal to the processor, wherein the processor generates a virtual microphone array comprising a first virtual microphone and a second virtual microphone.
  • the system of an embodiment comprises outputting a second microphone signal from a signal path including the second microphone coupled to the first filter, the third filter and the fifth filter.
  • the system of an embodiment comprises inputting the second microphone signal to the processor.
  • the system of an embodiment comprises inputting the second delayed first microphone signal to a voice activity detector (VAD).
  • VAD voice activity detector
  • the system of an embodiment comprises inputting the third microphone signal to the voice activity detector (VAD).
  • VAD voice activity detector
  • the system of an embodiment comprises outputting the second microphone signal from a signal path including the second microphone coupled to the first filter, the third filter and the fifth filter.
  • the first filter and the second filter of an embodiment are generated by: calculating a calibration filter by applying an adaptive filter to the first response and the second response; and determining a peak magnitude and a peak location of a largest peak of the calibration filter, wherein the largest peak is a largest peak located below a frequency of approximately 500 Hertz.
  • the generating of the first filter and the second filter comprises using unity filters for each of the first filter, the second filter and the third filter.
  • the system of an embodiment comprises, when a largest phase variation of the calibration filter is greater than positive three (3) degrees, calculating a first frequency corresponding to the first microphone and a second frequency corresponding to the second microphone.
  • the first frequency and the second frequency of an embodiment is a three-decible frequency.
  • the first frequency and the second frequency of an embodiment are used to generate the first filter and the second filter.
  • the first filter of an embodiment is an infinite impulse response (HR) filter.
  • the second filter of an embodiment is an infinite impulse response (IIR) filter.
  • the signal of an embodiment is a white noise signal.
  • the microphone array of an embodiment comprises amplitude response calibration and phase response calibration.
  • the system of an embodiment comprises an adaptive noise removal application running on the processor and generating denoised output signals by forming a plurality of combinations of signals output from the first virtual microphone and the second virtual microphone, wherein the denoised output signals include less acoustic noise than acoustic signals received at the microphone array.
  • the first virtual microphone of an embodiment has a first linear response to speech that is devoid of a null, wherein the speech is human speech.
  • the second virtual microphone of an embodiment has a second linear response to speech that includes a single null oriented in a direction toward a source of the speech.
  • the single null of an embodiment is a region of the second linear response having a measured response level that is lower than the measured response level of any other region of the second linear response.
  • the second linear response of an embodiment includes a primary lobe oriented in a direction away from the source of the speech.
  • the primary lobe of an embodiment is a region of the second linear response having a measured response level that is greater than the measured response level of any other region of the second linear response.
  • the first microphone and the second microphone of an embodiment are positioned along an axis and separated by a first distance.
  • a midpoint of the axis of an embodiment is a second distance from a speech source that generates the speech, wherein the speech source is located in a direction defined by an angle relative to the midpoint.
  • the first virtual microphone of an embodiment comprises the second microphone signal subtracted from the first microphone signal.
  • the first microphone signal of an embodiment is delayed.
  • the delay of an embodiment is raised to a power that is proportional to a time difference between arrival of the speech at the first virtual microphone and arrival of the speech at the second virtual microphone.
  • the delay of an embodiment is raised to a power that is proportional to a sampling frequency multiplied by a quantity equal to a third distance subtracted from a fourth distance, the third distance being between the first microphone and the speech source and the fourth distance being between the second microphone and the speech source.
  • the second microphone signal of an embodiment is multiplied by a ratio, wherein the ratio is a ratio of a third distance to a fourth distance, the third distance being between the first microphone and the speech source and the fourth distance being between the second microphone and the speech source.
  • the second virtual microphone of an embodiment comprises the first microphone signal subtracted from the second microphone signal.
  • the first microphone signal of an embodiment is delayed.
  • the delay of an embodiment is raised to a power that is proportional to a time difference between arrival of the speech at the first virtual microphone and arrival of the speech at the second virtual microphone.
  • the power of an embodiment is proportional to a sampling frequency multiplied by a quantity equal to a third distance subtracted from a fourth distance, the third distance being between the first microphone and the speech source and the fourth distance being between the second microphone and the speech source.
  • the first microphone signal of an embodiment is multiplied by a ratio, wherein the ratio is a ratio of the third distance to the fourth distance.
  • the first virtual microphone of an embodiment comprises the second microphone signal subtracted from a delayed version of the first microphone signal.
  • the second virtual microphone of an embodiment comprises a delayed version of the first microphone signal subtracted from the second microphone signal.
  • the system of an embodiment comprises a voice activity detector (VAD) coupled to the processor, the VAD generating voice activity signals.
  • VAD voice activity detector
  • the system of an embodiment comprises a communication channel coupled to the processor, the communication channel comprising at least one of a wireless channel, a wired channel, and a hybrid wireless/wired channel.
  • the system of an embodiment comprises a communication device coupled to the processor via the communication channel, the communication device comprising one or more of cellular telephones, satellite telephones, portable telephones, wireline telephones, Internet telephones, wireless transceivers, wireless communication radios, personal digital assistants (PDAs), and personal computers (PCs).
  • the communication device comprising one or more of cellular telephones, satellite telephones, portable telephones, wireline telephones, Internet telephones, wireless transceivers, wireless communication radios, personal digital assistants (PDAs), and personal computers (PCs).
  • PDAs personal digital assistants
  • PCs personal computers
  • Embodiments described herein include a method executing on a processor, the method comprising receiving signals at a microphone array comprising a first microphone and a second microphone.
  • the method of an embodiment comprises filtering an output of the second microphone with a first filter.
  • the first filter comprises a first filter model that models a response of the first microphone to a noise signal and outputs a second microphone signal.
  • the method of an embodiment comprises filtering an output of the first microphone with a second filter.
  • the second filter comprises a second filter model that models a response of the second microphone to the noise signal and outputs a first microphone signal.
  • the first microphone signal is calibrated with the second microphone signal.
  • the method of an embodiment comprises generating from the first microphone signal and the second microphone signal a virtual microphone array comprising a first virtual microphone and a second virtual microphone.
  • Embodiments described herein include a method executing on a processor, the method comprising: receiving signals at a microphone array comprising a first microphone and a second microphone; filtering an output of the second microphone with a first filter, wherein the first filter comprises a first filter model that models a response of the first microphone to a noise signal and outputs a second microphone signal; filtering an output of the first microphone with a second filter, wherein the second filter comprises a second filter model that models a response of the second microphone to the noise signal and outputs a first microphone signal, wherein the first microphone signal is calibrated with the second microphone signal; and generating from the first microphone signal and the second microphone signal a virtual microphone array comprising a first virtual microphone and a second virtual microphone.
  • the method of an embodiment comprises generating a third filter model that normalizes the first response and the second response.
  • the generating of the third filter model of an embodiment comprises convolving the first filter model with the second filter model and comparing a result of the convolving with a standard response filter, wherein the third filter model corrects an amplitude response of the result of the convolving.
  • the third filter model of an embodiment is a linear phase finite impulse response (FIR) filter.
  • the method of an embodiment comprises applying the third filter model to a signal resulting from the applying of the second filter model to the first response of the first microphone.
  • the method of an embodiment comprises applying the third filter model to a signal resulting from the applying of the first filter model to the second response of the second microphone.
  • the method of an embodiment comprises determining a third response of the first microphone by applying the second filter model and the third filter model to an output of the first microphone resulting from a second signal.
  • the method of an embodiment comprises determining a fourth response of the second microphone by applying the first filter model and the third filter model to an output of the second microphone resulting from the second signal.
  • the method of an embodiment comprises generating a fourth filter model from a combination of the third response and the fourth response, wherein the generating of the fourth filter model comprises applying an adaptive filter to the third response and the fourth response.
  • the fourth filter model of an embodiment is a minimum phase filter model.
  • the method of an embodiment comprises generating a fifth filter model from the fourth filter model.
  • the fifth filter model of an embodiment is a linear phase filter model.
  • Forming the microphone array of an embodiment comprises applying the third filter model to at least one of an output of the first filter model and an output of the second filter model.
  • Forming the microphone array of an embodiment comprises applying the third filter model to the output of the first filter model and the output of the second filter model.
  • the method of an embodiment comprises applying the second filter model and the third filter model to a signal output of the first microphone.
  • the method of an embodiment comprises applying the first filter model, the third filter model and the fifth filter model to a signal output of the second microphone.
  • the microphone array of an embodiment comprises amplitude response calibration and phase response calibration.
  • the method of an embodiment comprises generating de ⁇ oised output signals by forming a plurality of combinations of signals output from the first virtual microphone and the second virtual microphone, wherein the denoised output signals include less acoustic noise than acoustic signals received at the microphone array.
  • the method of an embodiment comprises generating the first microphone signal by applying the second filter model and the third filter model to a signal output of the first microphone.
  • the method of an embodiment comprises generating a first delayed first microphone signal by applying a first delay filter to the first microphone signal.
  • the method of an embodiment comprises inputting the first delayed first microphone signal to the processor.
  • the method of an embodiment comprises generating a second microphone signal by applying the first filter model, the third filter model and the fifth filter model to a signal output of the second microphone.
  • the method of an embodiment comprises inputting the second microphone signal to the processor.
  • the method of an embodiment comprises generating a second delayed first microphone signal by applying a second delay filter to the first microphone signal.
  • the method of an embodiment comprises inputting the second delayed first microphone signal to an acoustic voice activity detector.
  • the microphone signal by applying the first filter model, the third filter model and the fourth filter model to a signal output of the second microphone.
  • the method of an embodiment comprises inputting the third microphone signal to the acoustic voice activity detector.
  • the method of an embodiment comprises generating the first microphone signal by applying the second filter model and the third filter model to a signal output of the first microphone, and generating the second microphone signal by applying the first filter model, the third filter model and the fifth filter model to a signal output of the second microphone.
  • At least one of the first filter model and the second filter model of an embodiment is an infinite impulse response (HR) model.
  • the method of an embodiment comprises forming the first virtual microphone by generating a first combination of the first microphone signal and the second microphone signal.
  • the method of an embodiment comprises forming the second virtual microphone by generating a second combination of the first microphone signal and the second microphone signal, wherein the second combination is different from the first combination, wherein the first virtual microphone and the second virtual microphone are distinct virtual directional microphones with substantially similar responses to noise and substantially dissimilar responses to speech.
  • Forming the first virtual microphone of an embodiment includes forming the first virtual microphone to have a first linear response to speech that is devoid of a null, wherein the speech is human speech.
  • Forming the second virtual microphone of an embodiment includes forming the second virtual microphone to have a second linear response to speech that includes a single null oriented in a direction toward a source of the speech.
  • the single null of an embodiment is a region of the second linear response having a measured response level that is lower than the measured response level of any other region of the second linear response.
  • the second linear response of an embodiment includes a primary lobe oriented in a direction away from the source of the speech.
  • the primary lobe of an embodiment is a region of the second linear response having a measured response level that is greater than the measured response level of any other region of the second linear response.
  • the method of an embodiment comprises positioning the first physical microphone and the second physical microphone along an axis and
  • a midpoint of the axis of an embodiment is a second distance from a speech source that generates the speech, wherein the speech source is located in a direction defined by an angle relative to the midpoint.
  • the method of an embodiment comprises raising the delay to a power that is proportional to a time difference between arrival of the speech at the first virtual microphone and arrival of the speech at the second virtual microphone.
  • the method of an embodiment comprises raising the delay to a power that is proportional to a sampling frequency multiplied by a quantity equal to a third distance subtracted from a fourth distance, the third distance being between the first physical microphone and the speech source and the fourth distance being between the second physical microphone and the speech source.
  • the method of an embodiment comprises multiplying the second microphone signal by a ratio, wherein the ratio is a ratio of a third distance to a fourth distance, the third distance being between the first physical microphone and the speech source and the fourth distance being between the second physical microphone and the speech source.
  • Forming the second virtual microphone of an embodiment comprises subtracting the first microphone signal from the second microphone signal.
  • the method of an embodiment comprises raising the delay to a power that is proportional to a time difference between arrival of the speech at the first virtual microphone and arrival of the speech at the second virtual microphone.
  • the method of an embodiment comprises raising the delay to a power that is proportional to a sampling frequency multiplied by a quantity equal to a third distance subtracted from a fourth distance, the third distance being between the first physical microphone and the speech source and the fourth distance being between the second physical microphone and the speech source.
  • the method of an embodiment comprises multiplying the first microphone signal by a ratio, wherein the ratio is a ratio of the third distance to the fourth distance.
  • Forming the first virtual microphone of an embodiment comprises subtracting the second microphone signal from a delayed version of the first microphone signal.
  • Forming the second virtual microphone of an embodiment comprises: forming a quantity by delaying the first microphone signal; and subtracting the quantity from the second microphone signal.
  • the DOMA and corresponding calibration methods can be a component of a single system, multiple systems, and/or
  • the DOMA and corresponding calibration methods can also be a subcomponent or subsystem of a single system, multiple systems, and/or geographically separate systems.
  • the DOMA and corresponding calibration methods can be coupled to one or more other components (not shown) of a host system or a system coupled to the host system.
  • One or more components of the DOMA and corresponding calibration methods (v4, v4.1, v5, v6) and/or a corresponding system or application to which the DOMA and corresponding calibration methods (v4, v4.1, v5, v6) is coupled or connected includes and/or runs under and/or in association with a processing system.
  • the processing system includes any collection of processor- based devices or computing devices operating together, or components of processing systems or devices, as is known in the art.
  • the processing system can include one or more of a portable computer, portable communication device operating in a communication network, and/or a network server.
  • the portable computer can be any of a number and/or combination of devices selected from among personal computers, cellular telephones, personal digital assistants, portable computing devices, and portable communication devices, but is not so limited.
  • the processing system can include components within a larger computer system.
  • the processing system of an embodiment includes at least one processor and at least one memory device or subsystem.
  • the processing system can also include or be coupled to at least one database.
  • the term "processor” as generally used herein refers to any logic processing unit, such as one or more central processing units (CPUs), digital signal processors (DSPs), application-specific integrated circuits (ASIC), etc.
  • the processor and memory can be monolithically integrated onto a single chip, distributed among a number of chips or components, and/or provided by some combination of algorithms. The methods described herein can be
  • Communication paths couple the components and include any medium for communicating or transferring files among the components.
  • the communication paths include wireless connections, wired connections, and hybrid wireless/wired connections.
  • the communication paths also include couplings or connections to networks including local area networks (LANs), metropolitan area networks (MANs), wide area networks (WANs), proprietary networks, interoffice or backend networks, and the Internet.
  • LANs local area networks
  • MANs metropolitan area networks
  • WANs wide area networks
  • proprietary networks interoffice or backend networks
  • the Internet and the Internet.
  • the communication paths include removable fixed mediums like floppy disks, hard disk drives, and CD-ROM disks, as well as flash RAM, Universal Serial Bus (USB) connections, RS-232 connections, telephone lines, buses, and electronic mail messages.
  • USB Universal Serial Bus
  • aspects of the DOMA and corresponding calibration methods (v4, v4.1, v5, v6) and corresponding systems and methods described herein may be implemented as functionality programmed into any of a variety of circuitry, including programmable logic devices (PLDs), such as field programmable gate arrays (FPGAs), programmable array logic (PAL) devices, electrically programmable logic and memory devices and standard cell-based devices, as well as application specific integrated circuits (ASICs).
  • PLDs programmable logic devices
  • FPGAs field programmable gate arrays
  • PAL programmable array logic
  • ASICs application specific integrated circuits
  • Some other possibilities for implementing aspects of the DOMA and corresponding calibration methods (v4, v4.1, v5, v6) and corresponding systems and methods include: microcontrollers with memory (such as electronically erasable programmable read only memory (EEPROM)), embedded
  • microprocessors having software- based circuit emulation, discrete logic (sequential and combinatorial), custom devices, fuzzy (neural) logic, quantum devices, and hybrids of any of the above device types.
  • discrete logic discrete logic (sequential and combinatorial)
  • custom devices custom devices
  • fuzzy (neural) logic fuzzy logic
  • quantum devices and hybrids of any of the above device types.
  • the underlying device technologies may be provided in a variety of
  • MOSFET metal-oxide semiconductor field-effect transistor
  • CMOS complementary metal-oxide semiconductor
  • ECL emitter-coupled logic
  • polymer technologies e.g., silicon-conjugated polymer and metal-conjugated polymer-metal structures
  • mixed analog and digital etc.
  • any system, method, and/or other components disclosed herein may be described using computer aided design tools and expressed (or represented), as data and/or instructions embodied in various computer-readable media, in terms of their behavioral, register transfer, logic component, transistor, layout geometries, and/or other characteristics.
  • Computer-readable media in which such formatted data and/or instructions may be embodied include, but are not limited to, non-volatile storage media in various forms (e.g., optical, magnetic or semiconductor storage media) and carrier waves that may be used to transfer such formatted data and/or instructions through wireless, optical, or wired signaling media or any combination thereof. Examples of transfers of such formatted data and/or instructions by carrier waves include, but are not limited to, transfers
  • data transfer protocols e.g., HTTP, FTP, SMTP, etc.
  • a processing entity e.g., one or more processors
  • DOMA and corresponding calibration methods v4, v4.1, v5, v6 and corresponding systems and methods to the specific embodiments disclosed in the specification and the claims, but should be construed to include all systems that operate under the claims. Accordingly, the DOMA and corresponding calibration methods (v4, v4.1, v5, v6) and corresponding systems and methods is not limited by the disclosure, but instead the scope is to be determined entirely by the claims.

Landscapes

  • Health & Medical Sciences (AREA)
  • General Health & Medical Sciences (AREA)
  • Otolaryngology (AREA)
  • Physics & Mathematics (AREA)
  • Engineering & Computer Science (AREA)
  • Acoustics & Sound (AREA)
  • Signal Processing (AREA)
  • Circuit For Audible Band Transducer (AREA)
PCT/US2010/040501 2009-06-29 2010-06-29 Calibrating a dual omnidirectional microphone array (doma) WO2011002823A1 (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201090001122.8U CN203086710U (zh) 2009-06-29 2010-06-29 双重全向传声器阵列校准系统

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
US22141909P 2009-06-29 2009-06-29
US61/221,419 2009-06-29

Publications (1)

Publication Number Publication Date
WO2011002823A1 true WO2011002823A1 (en) 2011-01-06

Family

ID=43411415

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/US2010/040501 WO2011002823A1 (en) 2009-06-29 2010-06-29 Calibrating a dual omnidirectional microphone array (doma)

Country Status (2)

Country Link
CN (1) CN203086710U (zh)
WO (1) WO2011002823A1 (zh)

Cited By (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US9066186B2 (en) 2003-01-30 2015-06-23 Aliphcom Light-based detection for acoustic applications
US9099094B2 (en) 2003-03-27 2015-08-04 Aliphcom Microphone array with rear venting
DE102014208445A1 (de) 2014-05-06 2015-11-12 Volkswagen Ag Bipolarplatte, Brennstoffzelle und Verfahren zur Herstellung der Bipolarplatte
US9196261B2 (en) 2000-07-19 2015-11-24 Aliphcom Voice activity detector (VAD)—based multiple-microphone acoustic noise suppression
US10070220B2 (en) 2015-10-30 2018-09-04 Dialog Semiconductor (Uk) Limited Method for equalization of microphone sensitivities

Families Citing this family (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
DE102016210008A1 (de) * 2016-06-07 2017-12-07 Robert Bosch Gmbh Sensor- und/oder Wandlervorrichtung und Verfahren zum Betreiben einer Sensor- und/oder Wandlervorrichtung mit zumindest einer mindestens eine piezoelektrische Schicht umfassenden Biegestruktur
CN110333478B (zh) * 2018-03-30 2022-05-17 华为技术有限公司 一种到达角度、出发角度确定方法及通信装置
CN109246570B (zh) * 2018-08-29 2020-12-11 北京声智科技有限公司 麦克风质检的装置及方法

Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US5402669A (en) * 1994-05-16 1995-04-04 General Electric Company Sensor matching through source modeling and output compensation
US20060147032A1 (en) * 2004-12-30 2006-07-06 Mccree Alan V Acoustic echo devices and methods
US20070183610A1 (en) * 2004-10-19 2007-08-09 Widex A/S System and method for adaptive microphone matching in a hearing aid
US20080201138A1 (en) * 2004-07-22 2008-08-21 Softmax, Inc. Headset for Separation of Speech Signals in a Noisy Environment
US20090003623A1 (en) * 2007-06-13 2009-01-01 Burnett Gregory C Dual Omnidirectional Microphone Array (DOMA)
US20090003640A1 (en) * 2003-03-27 2009-01-01 Burnett Gregory C Microphone Array With Rear Venting

Patent Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US5402669A (en) * 1994-05-16 1995-04-04 General Electric Company Sensor matching through source modeling and output compensation
US20090003640A1 (en) * 2003-03-27 2009-01-01 Burnett Gregory C Microphone Array With Rear Venting
US20080201138A1 (en) * 2004-07-22 2008-08-21 Softmax, Inc. Headset for Separation of Speech Signals in a Noisy Environment
US20070183610A1 (en) * 2004-10-19 2007-08-09 Widex A/S System and method for adaptive microphone matching in a hearing aid
US20060147032A1 (en) * 2004-12-30 2006-07-06 Mccree Alan V Acoustic echo devices and methods
US20090003623A1 (en) * 2007-06-13 2009-01-01 Burnett Gregory C Dual Omnidirectional Microphone Array (DOMA)

Cited By (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US9196261B2 (en) 2000-07-19 2015-11-24 Aliphcom Voice activity detector (VAD)—based multiple-microphone acoustic noise suppression
US9066186B2 (en) 2003-01-30 2015-06-23 Aliphcom Light-based detection for acoustic applications
US9099094B2 (en) 2003-03-27 2015-08-04 Aliphcom Microphone array with rear venting
DE102014208445A1 (de) 2014-05-06 2015-11-12 Volkswagen Ag Bipolarplatte, Brennstoffzelle und Verfahren zur Herstellung der Bipolarplatte
US10070220B2 (en) 2015-10-30 2018-09-04 Dialog Semiconductor (Uk) Limited Method for equalization of microphone sensitivities

Also Published As

Publication number Publication date
CN203086710U (zh) 2013-07-24

Similar Documents

Publication Publication Date Title
US11818534B2 (en) Forming virtual microphone arrays using dual omnidirectional microphone array (DOMA)
US8699721B2 (en) Calibrating a dual omnidirectional microphone array (DOMA)
US8731211B2 (en) Calibrated dual omnidirectional microphone array (DOMA)
WO2011002823A1 (en) Calibrating a dual omnidirectional microphone array (doma)
US9099094B2 (en) Microphone array with rear venting
US10225649B2 (en) Microphone array with rear venting
US8477961B2 (en) Microphone array with rear venting
US20090003640A1 (en) Microphone Array With Rear Venting
CA2798512A1 (en) Vibration sensor and acoustic voice activity detection system (vads) for use with electronic systems
CA2798282A1 (en) Wind suppression/replacement component for use with electronic systems
CA2741652A1 (en) Acoustic voice activity detection (avad) for electronic systems
US20140286519A1 (en) Microphone array with rear venting
US10070220B2 (en) Method for equalization of microphone sensitivities
WO2009003180A1 (en) Microphone array with rear venting
US20220417652A1 (en) Microphone array with rear venting

Legal Events

Date Code Title Description
WWE Wipo information: entry into national phase

Ref document number: 201090001122.8

Country of ref document: CN

121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 10794671

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: DE

122 Ep: pct application non-entry in european phase

Ref document number: 10794671

Country of ref document: EP

Kind code of ref document: A1