WO2016209098A1 - Correction de réponse en phase inadaptée pour de multiples microphones - Google Patents

Correction de réponse en phase inadaptée pour de multiples microphones Download PDF

Info

Publication number
WO2016209098A1
WO2016209098A1 PCT/PL2015/000105 PL2015000105W WO2016209098A1 WO 2016209098 A1 WO2016209098 A1 WO 2016209098A1 PL 2015000105 W PL2015000105 W PL 2015000105W WO 2016209098 A1 WO2016209098 A1 WO 2016209098A1
Authority
WO
WIPO (PCT)
Prior art keywords
audio
microphone
signal
phase
frequency
Prior art date
Application number
PCT/PL2015/000105
Other languages
English (en)
Inventor
Przemysław MAZIEWSKI
Łukasz KURYLO
Piotr CHLEBEK
Original Assignee
Intel Corporation
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Intel Corporation filed Critical Intel Corporation
Priority to PCT/PL2015/000105 priority Critical patent/WO2016209098A1/fr
Priority to US15/573,581 priority patent/US10694304B2/en
Publication of WO2016209098A1 publication Critical patent/WO2016209098A1/fr

Links

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04RLOUDSPEAKERS, MICROPHONES, GRAMOPHONE PICK-UPS OR LIKE ACOUSTIC ELECTROMECHANICAL TRANSDUCERS; DEAF-AID SETS; PUBLIC ADDRESS SYSTEMS
    • H04R29/00Monitoring arrangements; Testing arrangements
    • H04R29/004Monitoring arrangements; Testing arrangements for microphones
    • H04R29/005Microphone arrays
    • H04R29/006Microphone matching
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04RLOUDSPEAKERS, MICROPHONES, GRAMOPHONE PICK-UPS OR LIKE ACOUSTIC ELECTROMECHANICAL TRANSDUCERS; DEAF-AID SETS; PUBLIC ADDRESS SYSTEMS
    • H04R1/00Details of transducers, loudspeakers or microphones
    • H04R1/20Arrangements for obtaining desired frequency or directional characteristics
    • H04R1/32Arrangements for obtaining desired frequency or directional characteristics for obtaining desired directional characteristic only
    • H04R1/40Arrangements for obtaining desired frequency or directional characteristics for obtaining desired directional characteristic only by combining a number of identical transducers
    • H04R1/406Arrangements for obtaining desired frequency or directional characteristics for obtaining desired directional characteristic only by combining a number of identical transducers microphones
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04RLOUDSPEAKERS, MICROPHONES, GRAMOPHONE PICK-UPS OR LIKE ACOUSTIC ELECTROMECHANICAL TRANSDUCERS; DEAF-AID SETS; PUBLIC ADDRESS SYSTEMS
    • H04R29/00Monitoring arrangements; Testing arrangements
    • H04R29/004Monitoring arrangements; Testing arrangements for microphones
    • H04R29/005Microphone arrays
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04RLOUDSPEAKERS, MICROPHONES, GRAMOPHONE PICK-UPS OR LIKE ACOUSTIC ELECTROMECHANICAL TRANSDUCERS; DEAF-AID SETS; PUBLIC ADDRESS SYSTEMS
    • H04R3/00Circuits for transducers, loudspeakers or microphones
    • H04R3/005Circuits for transducers, loudspeakers or microphones for combining the signals of two or more microphones
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04MTELEPHONIC COMMUNICATION
    • H04M2203/00Aspects of automatic or semi-automatic exchanges
    • H04M2203/50Aspects of automatic or semi-automatic exchanges related to audio conference
    • H04M2203/509Microphone arrays
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04MTELEPHONIC COMMUNICATION
    • H04M3/00Automatic or semi-automatic exchanges
    • H04M3/42Systems providing special services or facilities to subscribers
    • H04M3/56Arrangements for connecting several subscribers to a common circuit, i.e. affording conference facilities
    • H04M3/568Arrangements for connecting several subscribers to a common circuit, i.e. affording conference facilities audio processing specific to telephonic conferencing, e.g. spatial distribution, mixing of participants
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04RLOUDSPEAKERS, MICROPHONES, GRAMOPHONE PICK-UPS OR LIKE ACOUSTIC ELECTROMECHANICAL TRANSDUCERS; DEAF-AID SETS; PUBLIC ADDRESS SYSTEMS
    • H04R25/00Deaf-aid sets, i.e. electro-acoustic or electro-mechanical hearing aids; Electric tinnitus maskers providing an auditory perception
    • H04R25/40Arrangements for obtaining a desired directivity characteristic
    • H04R25/407Circuits for combining signals of a plurality of transducers

Definitions

  • the present description relates to the field of computing devices with multiple microphones and, in particular, correcting for phase response differences between audio received from multiple microphones.
  • Computing and communications devices are increasingly being fitted with multiple microphones.
  • the microphones may be used to add additional functions to a computing or communication device. These new functions include improved noise reduction, improved speech recognition, and identification of different speakers or noise sources.
  • the microphones may also be used to add depth and space to the received sound. By analyzing differences between the sounds received at each microphone these functions and many more may be provided.
  • Any of these devices may be equipped to receive speech and other audio, to record the audio, to use the received speech for conversations with other people in other locations and to use the speech for voice command.
  • Figure 1 is a generalized process flow diagram of detecting and correcting phase response mismatch in a multiple microphone system according to an embodiment.
  • Figure 2 is a side plan view of a test bench configuration for microphone phase response mismatch with a computing device.
  • Figure 3 is a top plan view of the test configuration of Figure 2 according to an embodiment.
  • Figure 4 is a top plan view of an alternative test configuration using a different loudspeaker arrangement according to an embodiment.
  • Figure 5 is a graph of delay versus frequency for two microphones as measured according to an embodiment.
  • Figure 6 is a graph of delay versus frequency for two different microphones as measured according to an embodiment.
  • Figure 7 is a graph of a portion of an example excitation signal to measure phase response mismatch at a particular frequency range according to an embodiment.
  • Figure 8 is graph of delay in sample for an instantaneous frequency band versus cross- correlation values according to an embodiment.
  • Figure 9 is a graph of a phase mismatch correction characteristic for different frequencies according to an embodiment.
  • Figure 10 is a block diagram of a phase-based beam former using phase correction according to an embodiment.
  • Figure 11 is a process flow diagram of determining phase correction values for a phase- based beam former or other device according to an embodiment.
  • Figure 12 is a block diagram of a communications device incorporating microphone phase detection and correction according to an embodiment
  • a computer, tablet, slate, telephone, or wearable device may have two or more microphones on each side of the device to allow sounds to be received clearly in different directions.
  • a fixed system may have microphones distributed across a display or around a room. The difference in arrival time and phase may be determined after the audio from each microphone is received, amplified, converted to a digital signal, and optionally converted to a frequency domain from a time domain.
  • the microphones do not have a linear or a matched-phase response.
  • the phase response may not be matched in that the spectral components of audio from one microphone always arrives at the audio processor or codec before or after spectral components of the audio from the other microphone.
  • a phase mismatch may cause an audio source to seem always to be farther from the delayed microphone than it actually is.
  • the phase response may not be linear in that it changes with frequency, lower frequency sounds will have a different delay from higher frequency sounds. This can cause errors and instability in later processing when more than one frequency is used.
  • the frequency dependency may also be different with different microphones. Differences in the non-linearity for example, will cause an audio source to seem to move as the audio changes pitch. A lower frequency sound will seem to be coming from a different position than a higher frequency sound.
  • the differences may be caused by differences in the hardware signal processing chain, differences in the microphones, differences in the microphone housings, etc. If synchronized sounds arrive out of phase, then the time or arrival or the phasing cannot be relied upon for beam forming, spatial, and depth detection systems.
  • the signals from each microphone may be compared and corrected. Phase differences are detected, estimated and then the phase response of the microphones are corrected.
  • the techniques described herein may be applied to fixed and mobile devices such as tablets, telephones, notebooks, all-in-one computers, televisions, television accessories, wearables, and in vehicle audio and video systems.
  • a delay estimation is performed between the microphones. This may be done by selecting a reference or base microphone and then comparing the phase of at each microphone to the reference microphone. Linear-or matched-phase microphones should have zero delay for all audio frequency components when compared against each other. Any delay between any pair of microphones indicates differences between the microphones in the phase response. The delay in phase response may be used for a quick phase mismatch detection as well as for a more precise estimation of the phase differences between the microphones. The more precise estimation may be used to correct the phase difference.
  • Figure 1 is a generalized process flow diagram for detecting, estimating, and correcting phase mismatch between two microphones.
  • techniques described herein may be simplified into three steps, phase mismatch detection, phase mismatch estimation and phase mismatch correction.
  • the present disclosure will describe a process for two microphone audio processing pipelines. For a system with three microphones, there will be two pair-wise comparisons, with one microphone selected to be the reference. For a system with four microphones there will be three pair-wise comparisons.
  • the techniques described herein may be extended to many more microphones by detecting and correcting phase mismatch against a single one of the microphones. If there are multiple microphones on different surfaces of the device, then the microphones may be grouped so that microphones are only corrected against microphones that are facing the same direction. The selection of microphones and groupings may be adapted to suit any particular device configuration and planned use.
  • the techniques described herein are initialized at 102 by determining which microphones are to be grouped together, for example, all microphones on a single surface of the device, all front facing microphones, or all microphones that receive audio in a particular device configuration.
  • one microphone is selected at 104 as a reference microphone.
  • any microphone may be used as the reference.
  • some devices may have a primary microphone, such as a center microphone or a higher resolution microphone that serves as a reference for other purposes.
  • a second microphone is then selected at 106 for comparison against the reference.
  • Each other microphone may be paired with the reference microphone in turn.
  • the corrections for each pairing may be made independently of each other pairing. As a result, the techniques described herein for one pair may simply be repeated for each other pair.
  • the first operation at 108 is to detect whether there is any significant phase response mismatch between the two microphones of the pair.
  • a phase response mismatch may be detected by analyzing the microphones' responses to a chirp signal or another excitation that is supplied to the microphones. If the phase response is not matched at 110 then the phase response delay may be estimated between microphone pairs at 1 12. At 1 14, this phase response delay is used to determine a phase response delay correction. At 1 16, the correction is applied to the audio as it is received by the microphones.
  • the phase corrections may be configured to adjust the other microphones to the reference microphone. However, since even the reference microphone may not have a linear response across its frequency range, the reference microphone may also have a phase correction.
  • the reference microphone's phase correction may be used to compensate for some frequencies arriving earlier than other frequencies.
  • the other microphones may be corrected with respect to the actual reference microphone response or with respect to the corrected reference microphone. By correcting each of the other microphones with respect to the same reference microphone, they will also be corrected with respect to each other.
  • FIG. 2 is a side plan view that shows an example of a test bench configuration that may be used to detect and to estimate a phase mismatch.
  • a notebook computer 202 is placed on a test bench 204 in an open configuration as if in use.
  • An audio source such as a loudspeaker 206 is placed in front of the notebook computer.
  • the loudspeaker is placed in a position that may simulate the position of a user.
  • the acoustical axis 208 of the loudspeaker is horizontal and parallel with the test bench 204.
  • the acoustical axis is raised above the test bench by an amount to simulate a user's head. While the acoustical axis is not incident on the computer, the loudspeaker has a vertical spread that will provide acoustical signals to the computer.
  • Figure 3 is a top plan view of the configuration of Figure 2.
  • the notebook computer has two microphones 212, 214 in the top horizontal part of the display bezel, one on either side of a centerline of the screen.
  • the microphones may be closer together or farther apart than shown.
  • the computer may be tested in the position as shown. This may be done, for example, when the back microphone are for noise cancellation or in any situation in which the phase response of the back microphones is to be unified.
  • the computer may be rotated one half circle so that the back of the display is facing the loudspeaker.
  • the loudspeaker is positioned so that the acoustical axis 208 is directly aligned with the centerline of the display. This places the axis at a position that is evenly spaced between the two microphones. As a result, the loudspeaker provides about the same acoustic energy to both microphones. It also provides that the distance from the loudspeaker to each microphone is the same.
  • the notebook computer may be an ultrabook computer, a laptop computer, or a portable workstation computer. While a notebook computer is shown, the same test configuration may be used for any other type of computer, computing device, or multiple microphone device, such as a standalone external microphone array.
  • An all-in-one computer, whether fixed or portable, may be placed in a similar position.
  • a desktop monitor that includes microphones may also be placed in a similar position.
  • a tablet computer may be placed on a stand or kickstand in a similar position to simulate a desktop configuration or to simulate being held by a user.
  • the test configuration may be adapted.
  • the loudspeaker may be lowered with respect to the system and moved farther away. Similarly for a vehicle system, the position of the loudspeaker may be adjusted. In any case, the test configuration is adjusted to simulate the position of the user or other audio source with respect to the computing or communication device and to provide an acoustic signal that reaches the two microphones at the same time. In some systems, one microphone is positioned farther away from an intended position of a speaker so that there is an intended delay between the two microphones. For these systems the loudspeaker may be position evenly between the two microphones or in a position that simulates the intended position of a speaker.
  • the loudspeaker generates acoustic energy within a sound field that is maximized along its acoustical axis and is dispersed horizontally and vertically from that axis.
  • the loudspeaker may also be selected to have an approximately even dispersion of acoustical energy horizontally and vertically across the dispersion angles that are used in the field test. It is also selected for even distribution of high and low frequencies, between the woofer and tweeter, across the sound field. While a two-way loudspeaker system with dynamic voice coil drivers is shown, any other type of loudspeaker may be used with more or fewer drivers of different types. With more than one driver, the drivers are arranged so that the distance from each driver to any one single microphone is the same.
  • the loudspeaker is also selected to have an approximately linear frequency response so that lower frequency sounds are produced with about the same acoustic energy as higher frequency sounds across the useful range.
  • useful tests may be performed with a frequency range of about 100Hz to 8000Hz so that many suitable loudspeakers are available. A wider or narrower frequency range may be used depending on the intended application for the microphone system.
  • a standard active studio monitor loudspeaker for near field listening provides a suitable acoustical source. However, other types of acoustical sources may be used instead.
  • Figure 4 is a top plan view that shows an alternative test configuration in which there are two loudspeakers 216, 218 each coupled to a respective microphone 212, 214.
  • the two loudspeakers are placed directly in front of the corresponding microphone so that most of the acoustical energy is coupled to only one of the microphones.
  • the microphones may be a suitable studio monitor or with a much shorter distance from loudspeaker to microphone, a smaller driver may be used.
  • the loudspeakers 216, 218 may each be an earphone, earbud, earpiece, or headphone component.
  • the left ear piece may be acoustically coupled to one of the microphones of a pair and the right ear piece may be acoustically coupled to the other microphone of the pair.
  • the drivers may each be held in a suitable test stand or attached directly to the computer housing.
  • Many wearable speaker systems provide a wideband signal with sufficient energy to replicate a user in the position as shown by the loudspeaker of Figure 2.
  • the loudspeaker is coupled to a test signal driver.
  • the active studio monitor loudspeaker 206 is coupled to the headphone output of the computer 202 with a suitable cable 220.
  • the in-ear headphones 216, 218 are attached to the computer 202 with a standard stereo cable 222.
  • the two drivers 216, 218 will produce the same sound at the same time.
  • the test hardware is simplified.
  • a discrete test system, signal generator or other sound source may be used instead of the computer.
  • the sound generation system may be independent from the phase detection and measurement system.
  • a test is performed to determine if there is any phase mismatch between any one pair of microphones in the computing or communications device.
  • a test sound is generated via the one or more loudspeakers in a test setup as shown, for example, in Figures 1-3.
  • the sound may be a single tone at a single audio frequency or a more complex chirp.
  • a single frequency may be used for significant phase mismatches, however, in some cases the phase mismatch varies with audio frequency. As a result, the mismatch will only show at certain frequencies.
  • a single frequency tone may or may not show the phase mismatch, depending on the response of the microphones at that frequency.
  • a more comprehensive detection signal is a linear or logarithmic swept-frequency cosine signal with an instantaneous frequency that covers different frequencies through the sweep.
  • a variety of different frequency bands may be selected. For voice systems, a frequency band from
  • 100-200 Hz to 6000-8000Hz may be used.
  • the frequency band ranges from 100Hz to 8000 Hz.
  • a wider or narrower band may be used depending on the intended use and accuracy of the system.
  • the responses to the chirp signal from the two microphones are divided into frames covering a selected number of instantaneous frequency periods.
  • the audio signal is received by the microphones and then captured for comparison. As described herein, typically microphones will be compared in pairs, however, after a reference microphone is selected, all of the microphones may be compared to the reference at the same time.
  • the audio may be captured in an internal memory or mass storage of the device or in an audio processor of the device.
  • the audio may be captured in the time domain as digital samples generated by the microphone or after further processing. In some embodiments, the digital samples are provided to an FFT engine and the converted frequency domain samples are saved for determining delay.
  • the audio signals may be captured at any point in the audio processing pipeline before the audio from multiple microphones is combined. By capturing the audio at a later stage in the audio processing pipeline more of the possible sources of delay may be detected and compensated.
  • Figure 5 is a graph of delay versus frequency to provide example results for a phase detection measurement. Such a result may be obtained, for example using three periods of a frequency sweep chirp signal applied to a notebook computer with two spaced apart
  • the microphones as show in Figure 2.
  • the audio is sampled with a standard frame rate as the frequency changes.
  • the frame length can be adjusted to the wavelength of the currently analyzed instantaneous frequency of the chirp signal.
  • the delay between two microphones is estimated using a cross-correlation. As the chirp signal sweeps across frequency in time, each time frame corresponds to a different audio frequency band. A non-zero delay indicates a phase mismatch between the two microphones for the frequency range covered by the analysis time frame.
  • the results of Figure 5 show measured delay from about 100Hz to 8000Hz using a sweep signal.
  • the device measured in Figure 5 has almost no phase mismatch.
  • the positive delay indicates that the sound is received from the reference microphone about one sample time before the sound is received from the second microphone.
  • the negative delay indicates that the sound is received at the reference microphone later than at the first microphone.
  • a one sample delay is about 20 microseconds.
  • Such a phase difference is small enough that it will not significantly impact audio beam forming and other features for the normal voice system frequency band and a microphone distance of several centimeters.
  • Figure 6 is an example of the same measurement applied to a different notebook computer. This device shows rapid variations of almost 40 samples for some frequencies and no mismatch for other frequencies. Audio frequencies between about 1000Hz and 1500Hz are essentially unusable for phase comparison while audio frequencies from about 3000Hz to 7300Hz are very well matched between the two microphones.
  • the difference between the device of Figure 5 and the device of Figure 6 may be a matter of production variations, changes in internal components that have the same design, or different models that are built to different designs or with different parts.
  • the device of Figure 6 is not able to accurately perform audio beam steering without some compensation or correction for the phase mismatch.
  • phase mismatch characteristic is precisely estimated.
  • the phase mismatch estimate and correction may be avoided if no or very little phase mismatch is detected at block 108.
  • Phase mismatch estimation may be done by measuring the microphone responses to an excitation signal on a test bench as shown in Figure 2. While the excitation signal for detection may be a simple frequency sweep or chirp. The excitation signal for a precise estimation may be more complex and may isolate frequency bands more carefully.
  • the test signal may be pre-recorded or it may be generated in a signal synthesizer.
  • the excitation signal for detection may be short and simple. A frequency sweep chirp is not only short for a faster test but delayed responses to any one particular instantaneous frequency are partially masked by the following frequencies as the sound sweeps. For phase estimation a longer signal may be used.
  • the excitation signal starts and ends with a linear chirp. The start and end chirp are used to synchronize the starting and ending point of the signal.
  • the linear chirp brings the driver, the microphones, and the audio processing chain to an operating state. Ideally, a longer chirp brings the system to a steady state, so that transient effects are eliminated from the test.
  • a series of intermediate cosine signals with fixed instantaneous frequencies are produced.
  • the cosine signals are separated by silent periods or gaps.
  • the instantaneous frequencies of the intermediate cosine signals change from a lowest frequency limit to a highest frequency limit with a fixed step.
  • the lower and upper limits define the boundaries of the phase mismatch estimate.
  • the step size defines the resolution.
  • Each cosine signal in the series is time windowed.
  • the windowing limits the loudspeaker inertia which is used to generate the excitation.
  • the windowing allows the loudspeaker to rest so that it is not overheated.
  • the frequency steps may be linear from lowest to highest or the frequency may hop in a particular pattern to distribute the load on the system more evenly across the frequency range.
  • Figure 7 is a graph of an example excitation signal with an initial chirp 702 and four intermediate cosine signals 704, 706, 708, 710.
  • the signals are all aligned along a nominal pilot signal 712 at the baseline frequency. This signal extends out as far as the test cycle for each frequency.
  • the intermediate cosine signals may then be followed by more cosine signals with linear chirps in the signal spaced at intervals and between the cosine signals.
  • a variety of different components may be added to the excitation signal, depending on the loudspeaker and the microphones.
  • a linear sweep chirp may be used to exercise the voice coil and for start point and end synchronization.
  • a unique signal may also be used to represent the end point.
  • the sequence may be repeated exactly or approximately for all of the frequencies that are to be tested.
  • the phase mismatch may vary depending on audio frequency.
  • the excitation signal uses narrowband signals to test the phase mismatch at different frequencies. These signals may be the cosine signals of Figure 7 or they may have a different shape.
  • a cosine or sine wave signal allows the phase to be measured at a particular frequency.
  • the phase may be detected at using either cosine or sine wave signals.
  • the frequency may be varied within a selected narrowband.
  • a smoother waveform reduces any effect of loudspeaker or amplifier signal inertia.
  • the narrowband signals may be centered at different bands to test as many different frequencies as desired.
  • the bands may be spaced apart as much as desired to provide the desired precision to the phase mismatch.
  • the first set of narrowband pulses may be centered at about 100Hz.
  • the second set may be centered at 1 10Hz.
  • the sets may progress by 10Hz intervals until the maximum audio frequency is reached such as 5000, 6000, 8000 or
  • the frequency may be increased exponentially or logarithmically. This allows for the entire range to be tested in less time.
  • the entire test frequency range may be divided into e.g. 100 sections so that there are 100 different test sets to cover the entire frequency range.
  • the test sets may be distributed in different ways and the number of test sets may be modified to suit different applications.
  • the Table shows an example waveform structure that may be used as an excitation signal in some embodiments.
  • Each of the signals in the overall excitation sequence may be modified to suit a variety of different purposes.
  • the start and end chirps may use a linear or non-linear signal sweep and may start at a low, high or central frequency.
  • the frequency range may be different from the frequency range for the phase mismatch testing. In one example, the range is selected based on the bandwidth limitations of the microphones and the connected audio processing pipelines.
  • the silence periods allow for the system to cool and reset and allow for signals to be distinguished from each other.
  • the sine wave sequence may have any number of different waveforms which may be the same or different.
  • the results from each waveform may be averaged or combined in another way to obtain a final phase mismatch value for each particular test frequency.
  • the same instantaneous frequency is used with a smooth amplitude modulation.
  • the smooth modulation reduces impulse impacts from irregularities in the any part of the system.
  • a phase shift may be used to further reduce irregularities in the system.
  • the first and third waveforms may start in phase (0 radians shift) while the second and fourth waveforms start with inverse phase (Pi radians shift).
  • the number and order of waveforms may be modified to suit different test scenarios and different devices.
  • the responses to the excitation signal are recorded.
  • the recording may be done at different places in the audio processing chain, depending on the nature of the device as described in more detail below.
  • the recordings are used to find the starting and ending chirp positions for both microphone recordings using a cross-correlation technique.
  • the chirp positions may be used to discover any constant and larger delay between both microphone recordings and then to time-align both recordings for further analysis. A constant delay may be taken into account when creating a final mismatch characteristic.
  • both recordings are divided into time frames covering only one intermediate cosine signal from each microphone recording.
  • the delay between cosine signals from both microphones is calculated using a cross-correlation.
  • the delay estimation precision may be further improved using interpolation, such as a spline interpolation, between the maxima of the cross-correlation function or other improvement techniques.
  • interpolation such as a spline interpolation, between the maxima of the cross-correlation function or other improvement techniques.
  • the maximum of the interpolation function may be used as the delay estimation.
  • Figure 8 is a graph of delay in samples for a single narrow frequency band or instantaneous frequency plotted on the horizontal axis versus the cross-correlation values.
  • the estimated time delays expressed in Radians for each instantaneous frequency of the analyzed intermediate cosine signals create the final phase mismatch characteristic.
  • the delay varies from about -2.5 samples to +3.2 samples.
  • the cross-correlation for these data points reaches a maximum at about zero delay.
  • the linear interpolation agrees with this value but the spline interpolation function has a maximum at about +0.5 samples delay.
  • the graph of Figure 7 shows that different values may be obtained using different processes. In this case any of the values may be used, however, better results may be obtained using the maximum spline interpolation.
  • the value of +0.5 may then be used as the delay at this particular instantaneous frequency.
  • the process of cross-correlation and interpolation may be repeated for each excitation pulse at a particular frequency and then combined to obtain a final value for that frequency. This may then be done for each tested frequency.
  • the results may be plotted on a graph such as that of Figure 9.
  • Figure 9 shows the phase correction in radians on the vertical axis for different frequencies. While the graph of Figure 6 shows the detection of the phase mismatch, the graph of Figure 9 shows a phase correction characteristic. Radian units provide a value that may be more directly applied to a phase correction system.
  • FIG 10 is a block diagram of a phase-based beam former in which phase correction may be used.
  • a beam former may be used in any of the devices mentioned above and for other types of systems.
  • the phase corrections described herein may also be used for other types of systems that do or do not include beam forming.
  • the beam forming system is provided as an example but embodiments are not limited to beam forming systems.
  • the system has one pair of two microphones 902, 905 although there may be more. Each microphone has an audio input processing stage.
  • the beam formed by the beam former is a set of phases applied to the microphone audio so that sound from one direction is reinforced and sound from other directions is attenuated.
  • the direction may correspond to a region that includes a user or main speaker, i.e., the source of the target speech.
  • Another beam may correspond to a direction that includes a significant noise source.
  • the noise may be isolated and attenuated.
  • the noise may be assumed to be any sound from outside of the target source beam.
  • the beam width may be fixed for simplicity, or adapted to suit different audio sources.
  • the first microphone 902 receives the audio pressure wave from the ambient environment and converts it to an analog voltage signal.
  • the voltage signal is sampled in the microphone and converted to a sequence of digital samples.
  • the audio samples are 16 bits and the sampling rate is 48kHz. However other sampling rates and bit widths may be used.
  • the microphone signal is provided to an FFT 903 to convert the audio samples from the time domain to the frequency domain.
  • the FFT samples are then applied to a phase detection circuit 904 and also to a downmixer 920.
  • the second microphone 905 receives analog audio from the ambient and produces digital samples for an FFT 906.
  • the microphones are located at different positions so each microphone receives signals corresponding to the same source at a slightly different time.
  • the time difference corresponds to the difference in distance between the source and each microphone.
  • the time difference of arrival may be characterized by phase in radians, by time in milliseconds, or by a number of digital audio samples.
  • the FFTs 903, 906 produce a discrete complex frequency spectrum which may have real and imaginary components at discrete frequencies, or frequency bins.
  • the FFT outputs from each microphone are each applied to a respective phase detector 904, 906 and to a downmixer. For both microphones at each frequency bin, the phase is estimated at blocks 904, 906 from the complex spectrum.
  • phase difference is provided to a comparator 910 to determine the phase difference.
  • the phase difference may then be corrected with a phase correction block 914.
  • the phase correction values 912 are determined for each FFT bin of the phase-based beam-former. These may be stored in a table as part of the platform tuning parameters. During processing the phase corrections are subtracted from the observed phase differences between the captured microphone signals in an adjustment block 914
  • phase difference error 916 for each frequency bin, an allowable phase difference, may be computed up front using beam width, microphone placement, e.g., distance, sampling frequency and a processing block length, e.g., discrete Fourier transform length.
  • the phase difference errors may be used to estimate noise and to calculate gain or attenuation factors to be applied to the downmixed audio signals.
  • the phase errors are applied to gain computation block 918.
  • the gain computations may include noise and attenuation factors and may use different values for different FFT frequency bins.
  • the outputs of the FFTs 903, 906 for each microphone signal are supplied to a downmixer
  • the gain computation block receives the downmixed audio signals 920 and applies the computed gains to an amplifier 922.
  • the gains are applied 922 to the downmixed signal which is then converted back to the time domain by IFFT synthesis.
  • the result is a clean audio signal 926 in which the audio from the two microphones is accurately combined.
  • the beam former also produces accurate corrected phase error 916 values that may be used to localize sound sources, to distinguish different speakers and for many other purposes. While only two microphones are shown, the phase corrections 912 may be applied to many more simultaneous microphones in the same way.
  • phase corrections 912 result in more accurate time difference of arrival estimations for each frequency component. This improves the accuracy of the beam former processing by mitigating the mismatched phase response of the elements in the microphone array. Without the phase corrections applied for the mismatched frequencies the processing would attenuate the desired signal and preserve particular noise arriving from the direction determined by the phase mismatch.
  • Phase correction may also be used for sound source localization and dynamic beam steering. When the corrected phase values are used to combine multiple microphone signals a much cleaner audio signal is received. This may be used to improve the performance of Automatic Speech Recognition (ASR) systems by decreasing Word Error Rates (WER).
  • ASR Automatic Speech Recognition
  • phase mismatch characteristic such as that shown in Figure 8
  • the measured phase mismatch response is adapted to the frequency resolution of the beam former's FFT (Fast Fourier Transform).
  • FFT Fast Fourier Transform
  • the phase response will have a higher resolution than the FFT.
  • the high resolution phase mismatch is adapted by converting the measured phase difference values to complex numbers. Equation 1 provides an example of one approach for expressing the phase difference.
  • x(f) is a complex number value
  • p(f) is the measured phase difference for a frequency/.
  • the frequency is based on the center frequency of each test signal. These frequencies may also be selected to coincide with the frequency bins of the FFT. If they do not align, then the test frequencies may be mapped to the frequency bins so that there is phase correction value for each frequency bin.
  • Equation 2 The sum c( i) of the complex numbers for each FFT bin may be determined as shown in Equation 2, where (i) indicates a frequency bin of the FFT.
  • the argument of the sum is the phase correction that is applied during processing in the look-up table 912 or other device. If the magnitude is not used it does not need to be normalized.
  • ⁇ 0 arg( ⁇ 3 ⁇ 4 w(0 x(O) Eq. 2 where freq_low(i) and freqjki(i) are the boundaries of the FFT bin with index i.
  • a simpler averaging approach may alternatively be used. While this approach is more complex, it produces better results when the phase mismatch for different test frequencies within the same FFT bin are very different. As an example, if some of the phase mismatch values for a particular FFT bin are close to negative ⁇ (-180 degrees) and other phase mismatch values are close to positive ⁇ (180 degrees), the simple phase average would be zero or close to zero. There would be little or no phase correction. On the other hand, using the complex number form of Eq. 1 for the averaging of Eq. 2 gives a more relevant averaged correction of either + ⁇ or - ⁇ . These values both provide a more accurate correction than zero because they both represent a signal that is completely out of phase by the same amount.
  • Figure 11 is a process flow diagram of determining phase correction values for the beam former described above.
  • the same phase correction values may be applied to many other types of systems which receive audio from one source through two or more microphones or audio processing pipelines.
  • the process flow has been described briefly in the context of Figure 1.
  • Figure 11 describes some of the operations of Figure 1 in more detail.
  • a single audio signal from a single source is received at a first microphone and at a second microphone.
  • the microphones are coupled to a single device for combining the received audio.
  • this received audio as it comes from the two or more microphones is recorded.
  • the signals from the microphones are not combined so they may be compared in phase.
  • the signals are analyzed to determine whether a phase response mismatch can be detected in the recording.
  • the phase response mismatch is between audio received at the second microphone and the audio received at the first microphone. This analysis may be performed in a variety of different ways.
  • the responses to the excitation signal are separately recorded for both microphones and then at 322 the starting and ending chirps' positions in both responses are searched using a cross-correlation technique. At 324 the positions are used to discover any constant delay between both microphone recordings.
  • a phase delay between the second microphone and the first microphone is estimated. This may also be done in a variety of different ways.
  • the two recording are aligned in time using the chirp signals. After time-aligning, both recordings are divided into time frames at 328. The time frames cover only one intermediate cosine signal from each microphone recording. These small frames are for ease of use, but other time units may be used instead.
  • the delay between the cosine signals from both microphones is calculated using a cross-correlation.
  • the delay estimation precision can then be further improved using a spline interpolation between the local maxima of the cross- correlation function. From this interpolation, the maximum of the interpolation function is used as the delay estimation.
  • the audio excitation signal may have signals at many different frequencies. These are typically provided in sequence with a start and end chirp at each end. Since the delay varies with frequency, the delay determination may then be repeated at 334 for each frequency in the signal. The estimated time delays for each instantaneous frequency of the analyzed intermediate cosine signals create the final phase mismatch characteristic at 336. The constant delay is taken into account when creating the final mismatch characteristic.
  • the microphone audio is first converted to the frequency domain for processing before it is sent or recorded for other purposes.
  • the phase response mismatch characteristic is matched to the frequency bins of the FFT process at 338. This allows the phase correction to be easily applied after the FFT conversion. Knowing the phase mismatch characteristic allows for corrections to be applied in a phase-based beam-former. In order to do so the measured phase mismatch response that is of high resolution needs to be adapted to the frequency resolution of the beam-former's FFT. This may be done by converting the measured phase difference values to complex numbers. The sum of the complex numbers for each FFT bin can be calculated and the argument of the sum may be used the phase correction that is applied during processing. The magnitude of the sum is not needed for a phase correction.
  • the final estimated phase delay is stored for use in correcting the phase delay in received audio before combining the received audio from the two or more microphones.
  • the phase corrections are subtracted from the observed phase differences between the captured microphone signals resulting in better time difference of arrival estimations for each frequency component.
  • FIG. 12 illustrates a communications device 100 in accordance with one
  • the device 100 may include a number of components, including but not limited to a processor and at least one communication package 6.
  • the communication package is coupled to one or more antennas 16.
  • the processor in this example is housed with a SoC (System on a Chip) 4 which is packaged.
  • SoC System on a Chip
  • the package is physically and electrically coupled to a system board.
  • the SoC may include other components that may or may not be on the same chip or in the same package. These other components include, but are not limited to, volatile memory (e.g., DRAM), non-volatile memory (e.g., ROM), flash memory, a graphics processor, a digital signal processor, a crypto processor, and a chipset.
  • volatile memory e.g., DRAM
  • non-volatile memory e.g., ROM
  • flash memory e.g., a graphics processor, a digital signal processor, a crypto processor, and a chipset.
  • the SoC is coupled to many other components that may be on the same or a different attached system board.
  • a display 18 such as a touchscreen display with a touchscreen controller, a battery 22 and associated power management system 24, an audio codec 20, a video codec (not shown), a power amplifier (not shown), a global positioning system (GPS) device 26, a sensor suite 28, which may include a compass, an accelerometer, a gyroscope, a proximity sensor, a pressure sensor, a battery fuel gauge etc.
  • the SoC may also be connected to a speaker
  • a microphone array 32 a camera array 34, and a mass storage device (such as flash cards, hard disk drive, etc.) 10, an NFC (Near Field Communication) module 36, any of a variety of other peripheral devices, including players for optical disks and other external media (not shown).
  • a mass storage device such as flash cards, hard disk drive, etc.
  • an NFC (Near Field Communication) module 36 any of a variety of other peripheral devices, including players for optical disks and other external media (not shown).
  • the communication package enables wireless and/or wired communications for the transfer of data to and from the communications device 100.
  • Such systems currently may include a cellular telephony modem 6, a WiFi module 8, and any of a variety of other components.
  • wireless and its derivatives may be used to describe circuits, devices, systems, methods, techniques, communications channels, etc., that may communicate data through the use of modulated electromagnetic radiation through a non-solid medium. The term does not imply that the associated devices do not contain any wires, although in some embodiments they might not.
  • the communication package may implement any of a number of wireless or wired standards or protocols, including but not limited to Wi-Fi (IEEE 802.11 family), WiMAX (IEEE 802.16 family), IEEE 802.20, long term evolution (LTE), Ev-DO, HSPA+, HSDPA+, HSUPA+, EDGE, GSM, GPRS, CDMA, TDMA, DECT, Bluetooth, Ethernet derivatives thereof, as well as any other wireless and wired protocols that are designated as 3G, 4G, 5G, and beyond.
  • the audio player device 100 may include a plurality of
  • a first communication package may be dedicated to shorter range wireless communications such as Wi-Fi and Bluetooth and a second
  • the communication package may be dedicated to longer range wireless communications such as GPS, EDGE, GPRS, CDMA, WiMAX, LTE, Ev-DO, and others.
  • the wireless communications package may also include components for receiving broadcast signal from terrestrial or satellite transmitters, including AM and FM radio, DAB (Digital Audio Broadcasting) and satellite radio.
  • the communications device 100 may be a laptop, a netbook, a notebook, an ultrabook, a smartphone, a wearable device, a tablet, a personal digital assistant (PDA), an ultra mobile PC, a mobile phone, a desktop computer, a server, a printer, a scanner, a monitor, a set-top box, an entertainment control unit, a digital camera, a portable music player, or a digital video recorder.
  • the communications device may be fixed, portable, or wearable.
  • the audio player device 100 may be any other electronic device that provides analog audio through wires.
  • the device 100 receives audio from the microphone array 32 from one or more intended sources, such as a user or other speakers. Additional audio may be received remotely from microphones of other devices, such as wearable microphone, and other smart devices through any of the communications interfaces 6, 8.
  • the SoC 4 feeds the additional audio to the audio codec 20 which contains the amplifiers and noise cancellation circuitry as described above.
  • the audio codec may store the audio, return it to the SOC for transmission or analysis or even convert the received audio to an analog form suitable for amplification to the speaker 30.
  • Embodiments may be implemented using one or more memory chips, controllers, CPUs (Central Processing Unit), microchips or integrated circuits interconnected using a motherboard, an application specific integrated circuit (ASIC), and/or a field programmable gate array
  • CPUs Central Processing Unit
  • ASIC application specific integrated circuit
  • references to “one embodiment”, “an embodiment”, “example embodiment”, “various embodiments”, etc., indicate that the embodiment(s) of the invention so described may include particular features, structures, or characteristics, but not every embodiment necessarily includes the particular features, structures, or characteristics. Further, some embodiments may have some, all, or none of the features described for other embodiments.
  • Coupled is used to indicate that two or more elements co-operate or interact with each other, but they may or may not have intervening physical or electrical components between them.
  • Some embodiments pertain to a method that includes receiving audio from a first microphone and from a second microphone, the microphones being coupled to a single device for combining the received audio, recording the received audio from the first microphone and the second microphone before combining the received audio, detecting a phase response mismatch in the recording at the device between the audio received at the second microphone and the audio received at the first microphone, if a phase response mismatch is detected, then estimating a phase delay between the second microphone and the first microphone, and storing the estimated phase delay for use in correcting the phase delay in received audio before combining the received audio.
  • Further embodiments include generating an audio test signal at the device and playing the audio test signal through a loudspeaker to generate the received audio.
  • the audio test signal includes an audio frequency sweep across an audio frequency range.
  • the audio test signal includes a plurality of instantaneous frequency signals at a plurality of different instantaneous frequencies, each separated by a silence.
  • the instantaneous frequency signals include a repetition of each different instantaneous frequency with a different phase.
  • the received audio has a start signal and an end signal and wherein estimating a phase delay comprises finding the start signal and the end signal for the first and the second microphone in the recorded audio and aligning the recorded audio using the found start signal and end signal.
  • estimating a phase delay comprises selecting an instantaneous frequency signal in the recorded audio for the first microphone and for the second microphone and determining the delay between the two microphones for the selected signal, wherein the selected signal has a single frequency and is between the start signal and the end signal.
  • determining the delay comprises cross correlating the selected signal from the first microphone with the selected signal from the second microphone.
  • determining the delay comprises applying a spline interpolation to maxima of the cross correlation. Further embodiments include converting the audio for each microphone in a respective
  • each FFT engine providing audio in a plurality of different frequency bins
  • estimating a phase delay comprises repeating selecting an instantaneous frequency signal and determining the delay for a plurality of signals within an audio frequency range that corresponds to a frequency bin of the FFT engines and combining the determined delays to determine a delay for the corresponding frequency bin.
  • combining the determined delays comprises converting each delay to a complex number representation of the delay and summing the complex number representations.
  • Further embodiments include taking the argument of the sum of complex numbers as the estimated phase delay for the frequency bin.
  • a machine-readable medium has instructions thereon that when operated on by the machine cause the machine to perform operations that include receiving audio from a first microphone and from a second microphone, the microphones being coupled to a single device for combining the received audio, recording the received audio from the first microphone and the second microphone before combining the received audio, detecting a phase response mismatch in the recording at the device between the audio received at the second microphone and the audio received at the first microphone, if a phase response mismatch is detected, then estimating a phase delay between the second microphone and the first
  • Further embodiments include generating an audio test signal at the device and playing the audio test signal through a loudspeaker to generate the received audio.
  • the audio test signal includes an audio frequency sweep across an audio frequency range.
  • the audio test signal includes a plurality of instantaneous frequency signals at a plurality of different instantaneous frequencies, each separated by a silence.
  • Some embodiments pertain to an apparatus that includes a first microphone to receive a test signal, a second microphone to receive the test signal, a memory to record the received audio from the first microphone and the second microphone before combining the received audio, and a processor to detect a phase response mismatch in the recording at the apparatus between the audio received at the second microphone and the audio received at the first microphone, if a phase response mismatch is detected, then to estimate a phase delay between the second microphone and the first microphone, and to store the estimated phase delay in the memory for use in correcting the phase delay in received audio before combining the received audio.
  • the received audio has a start signal and an end signal and wherein the processor estimates a phase delay by finding the start signal and the end signal for the first and the second microphone in the recorded audio and aligning the recorded audio using the found start signal and end signal.
  • estimating a phase delay comprises selecting an instantaneous frequency signal in the recorded audio for the first microphone and for the second microphone and cross correlating the selected signal from the first microphone with the selected signal from the second microphone.

Landscapes

  • Health & Medical Sciences (AREA)
  • Otolaryngology (AREA)
  • Physics & Mathematics (AREA)
  • Engineering & Computer Science (AREA)
  • Acoustics & Sound (AREA)
  • Signal Processing (AREA)
  • General Health & Medical Sciences (AREA)
  • Circuit For Audible Band Transducer (AREA)
  • Obtaining Desirable Characteristics In Audible-Bandwidth Transducers (AREA)

Abstract

Selon la présente invention, pour un système à microphones multiples, une réponse en phase inadaptée peut être corrigée. Un mode de réalisation consiste à recevoir un signal audio d'un premier microphone et d'un second microphone, les microphones étant couplés à un seul dispositif destiné à combiner les signaux audio reçus, à enregistrer le signal audio reçu du premier microphone et du second microphone avant de combiner les signaux audio reçus, à détecter une réponse en phase inadaptée dans l'enregistrement au niveau du dispositif entre les signaux audio reçus au niveau du second microphone et les signaux audio reçus au niveau du premier microphone, si une réponse en phase inadaptée a été détectée, à estimer alors un retard de phase entre le second microphone et le premier microphone, et à stocker le retard de phase estimé pour une utilisation lors de la correction du retard de phase dans les signaux audio reçus avant de combiner les signaux audio reçus.
PCT/PL2015/000105 2015-06-26 2015-06-26 Correction de réponse en phase inadaptée pour de multiples microphones WO2016209098A1 (fr)

Priority Applications (2)

Application Number Priority Date Filing Date Title
PCT/PL2015/000105 WO2016209098A1 (fr) 2015-06-26 2015-06-26 Correction de réponse en phase inadaptée pour de multiples microphones
US15/573,581 US10694304B2 (en) 2015-06-26 2015-06-26 Phase response mismatch correction for multiple microphones

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
PCT/PL2015/000105 WO2016209098A1 (fr) 2015-06-26 2015-06-26 Correction de réponse en phase inadaptée pour de multiples microphones

Publications (1)

Publication Number Publication Date
WO2016209098A1 true WO2016209098A1 (fr) 2016-12-29

Family

ID=53719896

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/PL2015/000105 WO2016209098A1 (fr) 2015-06-26 2015-06-26 Correction de réponse en phase inadaptée pour de multiples microphones

Country Status (2)

Country Link
US (1) US10694304B2 (fr)
WO (1) WO2016209098A1 (fr)

Cited By (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN109313909A (zh) * 2018-08-22 2019-02-05 深圳市汇顶科技股份有限公司 评估麦克风阵列一致性的方法、设备、装置和系统
CN110495186A (zh) * 2017-03-30 2019-11-22 奇跃公司 非阻塞双驱动器耳机
WO2019238800A1 (fr) * 2018-06-15 2019-12-19 Widex A/S Procédé de test des performances d'un microphone d'un système d'aide auditive et système d'aide auditive
US11765536B2 (en) 2018-11-13 2023-09-19 Dolby Laboratories Licensing Corporation Representing spatial audio by means of an audio signal and associated metadata

Families Citing this family (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
GB201615538D0 (en) 2016-09-13 2016-10-26 Nokia Technologies Oy A method , apparatus and computer program for processing audio signals
CN111357216B (zh) * 2017-11-17 2022-05-17 上海诺基亚贝尔股份有限公司 用于定位通信设备的方法和设备
WO2022150950A1 (fr) * 2021-01-12 2022-07-21 华为技术有限公司 Procédé et appareil d'évaluation de la cohérence d'un réseau de microphones
EP4231663A4 (fr) 2021-03-12 2024-05-08 Samsung Electronics Co., Ltd. Dispositif électronique d'entrée audio et son procédé de fonctionnement
CN112995882B (zh) * 2021-05-11 2021-08-13 杭州兆华电子有限公司 一种智能设备音频开环测试方法
CN114363786B (zh) * 2022-01-10 2023-11-24 海菲曼(天津)科技有限公司 一种麦克风测试校准装置

Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20080175407A1 (en) * 2007-01-23 2008-07-24 Fortemedia, Inc. System and method for calibrating phase and gain mismatches of an array microphone
US20090323980A1 (en) * 2008-06-26 2009-12-31 Fortemedia, Inc. Array microphone system and a method thereof
US20140363001A1 (en) * 2013-06-06 2014-12-11 Fortemedia, Inc. Method for calibrating performance of small array microphones
WO2015013698A1 (fr) * 2013-07-26 2015-01-29 Analog Devices, Inc. Étalonnage de microphone

Family Cites Families (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US5778082A (en) * 1996-06-14 1998-07-07 Picturetel Corporation Method and apparatus for localization of an acoustic source
US9333463B2 (en) 2013-07-26 2016-05-10 General Electric Company Devices and systems for elution of biomolecules

Patent Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20080175407A1 (en) * 2007-01-23 2008-07-24 Fortemedia, Inc. System and method for calibrating phase and gain mismatches of an array microphone
US20090323980A1 (en) * 2008-06-26 2009-12-31 Fortemedia, Inc. Array microphone system and a method thereof
US20140363001A1 (en) * 2013-06-06 2014-12-11 Fortemedia, Inc. Method for calibrating performance of small array microphones
WO2015013698A1 (fr) * 2013-07-26 2015-01-29 Analog Devices, Inc. Étalonnage de microphone

Cited By (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN110495186A (zh) * 2017-03-30 2019-11-22 奇跃公司 非阻塞双驱动器耳机
CN110495186B (zh) * 2017-03-30 2021-11-19 奇跃公司 声音再现系统和头戴式设备
WO2019238800A1 (fr) * 2018-06-15 2019-12-19 Widex A/S Procédé de test des performances d'un microphone d'un système d'aide auditive et système d'aide auditive
US11245992B2 (en) 2018-06-15 2022-02-08 Widex A/S Method of testing microphone performance of a hearing aid system and a hearing aid system
CN109313909A (zh) * 2018-08-22 2019-02-05 深圳市汇顶科技股份有限公司 评估麦克风阵列一致性的方法、设备、装置和系统
CN109313909B (zh) * 2018-08-22 2023-05-12 深圳市汇顶科技股份有限公司 评估麦克风阵列一致性的方法、设备、装置和系统
US11765536B2 (en) 2018-11-13 2023-09-19 Dolby Laboratories Licensing Corporation Representing spatial audio by means of an audio signal and associated metadata

Also Published As

Publication number Publication date
US20180359584A1 (en) 2018-12-13
US10694304B2 (en) 2020-06-23

Similar Documents

Publication Publication Date Title
US10694304B2 (en) Phase response mismatch correction for multiple microphones
US11729572B2 (en) Systems and methods for calibrating speakers
US10873814B2 (en) Analysis of spatial metadata from multi-microphones having asymmetric geometry in devices
US11317231B2 (en) Spatial audio signal format generation from a microphone array using adaptive capture
US10382849B2 (en) Spatial audio processing apparatus
AU2016213897B2 (en) Adaptive room equalization using a speaker and a handheld listening device
US10334357B2 (en) Machine learning based sound field analysis
US9363596B2 (en) System and method of mixing accelerometer and microphone signals to improve voice quality in a mobile device
US9426598B2 (en) Spatial calibration of surround sound systems including listener position estimation
KR102036359B1 (ko) 다중 채널 오디오를 위한 룸 특성화 및 보정
US11223924B2 (en) Audio distance estimation for spatial audio processing
JP2020500480A5 (fr)
US20070121955A1 (en) Room acoustics correction device
CN109804559A (zh) 空间音频系统中的增益控制
Gaubitch et al. On near-field beamforming with smartphone-based ad-hoc microphone arrays
US11477596B2 (en) Calibration of synchronized audio playback on microphone-equipped speakers
Erdeljan Large-scale audio eavesdropping using smartphones
CN114650480A (zh) 能够主动降噪的设备、主动降噪方法和存储介质
Tinakari Physical size of microphone arrays in ad-hoc beamforming

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 15741617

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: DE

122 Ep: pct application non-entry in european phase

Ref document number: 15741617

Country of ref document: EP

Kind code of ref document: A1