US9271077B2 - Method and system for directional enhancement of sound using small microphone arrays - Google Patents
Method and system for directional enhancement of sound using small microphone arrays Download PDFInfo
- Publication number
- US9271077B2 US9271077B2 US14/108,883 US201314108883A US9271077B2 US 9271077 B2 US9271077 B2 US 9271077B2 US 201314108883 A US201314108883 A US 201314108883A US 9271077 B2 US9271077 B2 US 9271077B2
- Authority
- US
- United States
- Prior art keywords
- phase angle
- microphone
- microphone signal
- sound source
- coherence
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Active, expires
Links
- 238000000034 method Methods 0.000 title claims abstract description 84
- 238000003491 array Methods 0.000 title description 6
- 230000001419 dependent effect Effects 0.000 claims description 22
- 230000003595 spectral effect Effects 0.000 claims description 15
- 230000035945 sensitivity Effects 0.000 claims description 12
- 238000005259 measurement Methods 0.000 claims description 8
- 230000001965 increasing effect Effects 0.000 claims description 7
- 238000001914 filtration Methods 0.000 claims description 6
- 230000033001 locomotion Effects 0.000 claims description 3
- 230000004044 response Effects 0.000 claims description 3
- 230000006641 stabilisation Effects 0.000 claims description 3
- 238000011105 stabilization Methods 0.000 claims description 3
- 238000009499 grossing Methods 0.000 claims description 2
- 238000004091 panning Methods 0.000 claims description 2
- 238000004458 analytical method Methods 0.000 abstract description 20
- 210000000613 ear canal Anatomy 0.000 description 24
- 238000004891 communication Methods 0.000 description 21
- 239000013598 vector Substances 0.000 description 18
- 238000012545 processing Methods 0.000 description 17
- 230000006870 function Effects 0.000 description 15
- 238000005516 engineering process Methods 0.000 description 12
- 230000005236 sound signal Effects 0.000 description 12
- 238000007789 sealing Methods 0.000 description 9
- 230000003044 adaptive effect Effects 0.000 description 8
- 230000008569 process Effects 0.000 description 6
- 239000011521 glass Substances 0.000 description 5
- 230000006872 improvement Effects 0.000 description 5
- 210000003454 tympanic membrane Anatomy 0.000 description 5
- 238000004422 calculation algorithm Methods 0.000 description 4
- 230000007613 environmental effect Effects 0.000 description 4
- 230000003287 optical effect Effects 0.000 description 4
- 230000001413 cellular effect Effects 0.000 description 3
- 238000001514 detection method Methods 0.000 description 3
- 230000003993 interaction Effects 0.000 description 3
- 230000004807 localization Effects 0.000 description 3
- 238000010295 mobile communication Methods 0.000 description 3
- 238000012544 monitoring process Methods 0.000 description 3
- 230000000007 visual effect Effects 0.000 description 3
- 238000013459 approach Methods 0.000 description 2
- 238000004590 computer program Methods 0.000 description 2
- 230000008878 coupling Effects 0.000 description 2
- 238000010168 coupling process Methods 0.000 description 2
- 238000005859 coupling reaction Methods 0.000 description 2
- 230000002708 enhancing effect Effects 0.000 description 2
- 230000000873 masking effect Effects 0.000 description 2
- 238000012986 modification Methods 0.000 description 2
- 230000004048 modification Effects 0.000 description 2
- 238000001228 spectrum Methods 0.000 description 2
- 239000003826 tablet Substances 0.000 description 2
- 210000000707 wrist Anatomy 0.000 description 2
- 230000018199 S phase Effects 0.000 description 1
- 241000269400 Sirenidae Species 0.000 description 1
- 230000002159 abnormal effect Effects 0.000 description 1
- 230000009471 action Effects 0.000 description 1
- 230000006978 adaptation Effects 0.000 description 1
- 230000002238 attenuated effect Effects 0.000 description 1
- 230000003190 augmentative effect Effects 0.000 description 1
- 230000004888 barrier function Effects 0.000 description 1
- 230000008901 benefit Effects 0.000 description 1
- 230000033228 biological regulation Effects 0.000 description 1
- 230000000903 blocking effect Effects 0.000 description 1
- 238000004364 calculation method Methods 0.000 description 1
- 230000008859 change Effects 0.000 description 1
- 238000007418 data mining Methods 0.000 description 1
- 238000013500 data storage Methods 0.000 description 1
- 230000003247 decreasing effect Effects 0.000 description 1
- 230000001066 destructive effect Effects 0.000 description 1
- 210000005069 ears Anatomy 0.000 description 1
- 230000000694 effects Effects 0.000 description 1
- 238000000605 extraction Methods 0.000 description 1
- 239000006260 foam Substances 0.000 description 1
- 238000007429 general method Methods 0.000 description 1
- 230000001771 impaired effect Effects 0.000 description 1
- 239000007943 implant Substances 0.000 description 1
- 230000002452 interceptive effect Effects 0.000 description 1
- 238000002955 isolation Methods 0.000 description 1
- 238000007726 management method Methods 0.000 description 1
- 239000003550 marker Substances 0.000 description 1
- 230000000116 mitigating effect Effects 0.000 description 1
- 238000012806 monitoring device Methods 0.000 description 1
- 230000002093 peripheral effect Effects 0.000 description 1
- 230000008092 positive effect Effects 0.000 description 1
- 238000003672 processing method Methods 0.000 description 1
- 230000009467 reduction Effects 0.000 description 1
- 230000002829 reductive effect Effects 0.000 description 1
- 230000008054 signal transmission Effects 0.000 description 1
- 230000011664 signaling Effects 0.000 description 1
- 230000000153 supplemental effect Effects 0.000 description 1
- 238000012360 testing method Methods 0.000 description 1
- 238000012546 transfer Methods 0.000 description 1
Images
Classifications
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04R—LOUDSPEAKERS, MICROPHONES, GRAMOPHONE PICK-UPS OR LIKE ACOUSTIC ELECTROMECHANICAL TRANSDUCERS; DEAF-AID SETS; PUBLIC ADDRESS SYSTEMS
- H04R3/00—Circuits for transducers, loudspeakers or microphones
- H04R3/005—Circuits for transducers, loudspeakers or microphones for combining the signals of two or more microphones
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04R—LOUDSPEAKERS, MICROPHONES, GRAMOPHONE PICK-UPS OR LIKE ACOUSTIC ELECTROMECHANICAL TRANSDUCERS; DEAF-AID SETS; PUBLIC ADDRESS SYSTEMS
- H04R1/00—Details of transducers, loudspeakers or microphones
- H04R1/10—Earpieces; Attachments therefor ; Earphones; Monophonic headphones
- H04R1/1083—Reduction of ambient noise
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04R—LOUDSPEAKERS, MICROPHONES, GRAMOPHONE PICK-UPS OR LIKE ACOUSTIC ELECTROMECHANICAL TRANSDUCERS; DEAF-AID SETS; PUBLIC ADDRESS SYSTEMS
- H04R1/00—Details of transducers, loudspeakers or microphones
- H04R1/20—Arrangements for obtaining desired frequency or directional characteristics
- H04R1/32—Arrangements for obtaining desired frequency or directional characteristics for obtaining desired directional characteristic only
- H04R1/40—Arrangements for obtaining desired frequency or directional characteristics for obtaining desired directional characteristic only by combining a number of identical transducers
- H04R1/406—Arrangements for obtaining desired frequency or directional characteristics for obtaining desired directional characteristic only by combining a number of identical transducers microphones
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04R—LOUDSPEAKERS, MICROPHONES, GRAMOPHONE PICK-UPS OR LIKE ACOUSTIC ELECTROMECHANICAL TRANSDUCERS; DEAF-AID SETS; PUBLIC ADDRESS SYSTEMS
- H04R2201/00—Details of transducers, loudspeakers or microphones covered by H04R1/00 but not provided for in any of its subgroups
- H04R2201/40—Details of arrangements for obtaining desired directional characteristic by combining a number of identical transducers covered by H04R1/40 but not provided for in any of its subgroups
- H04R2201/405—Non-uniform arrays of transducers or a plurality of uniform arrays with different transducer spacing
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04R—LOUDSPEAKERS, MICROPHONES, GRAMOPHONE PICK-UPS OR LIKE ACOUSTIC ELECTROMECHANICAL TRANSDUCERS; DEAF-AID SETS; PUBLIC ADDRESS SYSTEMS
- H04R2499/00—Aspects covered by H04R or H04S not otherwise provided for in their subgroups
- H04R2499/10—General applications
- H04R2499/11—Transducers incorporated or for use in hand-held devices, e.g. mobile phones, PDA's, camera's
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04R—LOUDSPEAKERS, MICROPHONES, GRAMOPHONE PICK-UPS OR LIKE ACOUSTIC ELECTROMECHANICAL TRANSDUCERS; DEAF-AID SETS; PUBLIC ADDRESS SYSTEMS
- H04R2499/00—Aspects covered by H04R or H04S not otherwise provided for in their subgroups
- H04R2499/10—General applications
- H04R2499/13—Acoustic transducers and sound field adaptation in vehicles
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04R—LOUDSPEAKERS, MICROPHONES, GRAMOPHONE PICK-UPS OR LIKE ACOUSTIC ELECTROMECHANICAL TRANSDUCERS; DEAF-AID SETS; PUBLIC ADDRESS SYSTEMS
- H04R25/00—Deaf-aid sets, i.e. electro-acoustic or electro-mechanical hearing aids; Electric tinnitus maskers providing an auditory perception
- H04R25/40—Arrangements for obtaining a desired directivity characteristic
- H04R25/407—Circuits for combining signals of a plurality of transducers
Definitions
- the present invention relates to audio enhancement in noisy environments with particular application to mobile audio devices such as augmented reality displays, mobile computing devices, headphones, hearing aids.
- SNR signal to noise ratio
- the receive gain The improvement compared with omnidirectional reception is known as the receive gain.
- the receive gain measured as an improvement in SNR, is about 3 dB for every additional microphone, i.e. 3 dB improvement for 2 microphones, 6 dB for 3 microphones etc. This improvement occurs only at sound frequencies where the wavelength is above the spacing of the microphones.
- the beamforming approaches are directed to arrays where the microphones are spaced wide with respect to one another. There is also a need for a method and device for directional enhancement of sound using small microphone arrays.
- FIG. 1A illustrates an acoustic sensor in accordance with an exemplary embodiment
- FIG. 1B illustrates a wearable system for directional enhancement of sound in accordance with an exemplary embodiment
- FIG. 1C illustrates another wearable system for directional enhancement of sound in accordance with an exemplary embodiment
- FIG. 1D illustrates a mobile device for coupling with the wearable system in accordance with an exemplary embodiment
- FIG. 1E illustrates another mobile device for coupling with the wearable system in accordance with an exemplary embodiment
- FIG. 2 is method for updating directional enhancement filter.
- FIG. 3 is a measurement setup for acquiring target inter-microphone coherence between omni-directional microphones M 1 and M 2 for sound targets at particular angles of incidence (i.e. angle theta) in accordance with an exemplary embodiment
- FIG. 4A-4F shows analysis of coherence from measurement set-up in FIG. 3 with different target directions showing imaginary, real, and (unwrapped) phase angle in accordance with an exemplary embodiment
- FIG. 5 shows a multi-microphone configuration and control interface to select desired target direction and output source location in accordance with an exemplary embodiment
- FIG. 6 depicts a method for determining source location from analysis of measured coherence angle in accordance with an exemplary embodiment
- FIG. 7 is an exemplary earpiece for use with the coherence based directional enhancement system of FIG. 1A in accordance with an exemplary embodiment
- FIG. 8 is an exemplary mobile device for use with the coherence based directional enhancement system in accordance with an exemplary embodiment.
- FIG. 9 depicts a method for social deployment of directional enhancement of acoustic signals within social media in accordance with an exemplary embodiment.
- a method and system for affecting the directional sensitivity of a microphone array system comprised of at least two microphones, for example, such as those mounted on a headset or small mobile computing device. It overcomes the limitations experienced with conventional beamforming approaches using small microphone arrays. Briefly, in order for a useful improvement in SNR, there must be many microphones (e.g. 3-6) spaced over a large volume (e.g. for SNR enhancement at 500 Hz, the inter-microphone spacing must be over half a meter).
- FIG. 1A depicts an acoustic device 170 to increase a directional sensitivity of a microphone signal.
- the components therein can be integrated and/or incorporated into the wearable devices (e.g., headset 100 , eyeglasses 120 , mobile device 140 , wrist watch 160 , earpiece 500 ).
- the acoustic device 170 includes a first microphone 171 , and a processor 171 for receiving a first microphone signal from the first microphone 171 . It also receives a second microphone signal from a second microphone 172 .
- This second microphone 172 may be part of the device housing the acoustic device 170 or a separate device, and which is communicatively coupled to the acoustic device 170 .
- the second microphone 172 can be communicatively coupled to the processor 173 and reside on a secondary device that is one of a mobile device, a phone, an earpiece, a tablet, a laptop, a camera, a web cam, a wearable accessory, smart eyewear, or smart headwear.
- the acoustic device 170 can be communicatively coupled or integrated with a dash cam for police matters, for example, wirelessly connected to microphones within officer automobiles and/or on officer glasses, headgear, mobile device and other wearable communication equipment external to the automobile.
- the acoustic device 170 can also be coupled to other devices, for example, a security camera, for instance, to pan and focus on directional or localized sounds. Additional features and elements can be included with the acoustic device 170 , for instance, communication port 175 , also shown ahead in FIG. 6 , to include communication functionality (wireless chip set, Bluetooth, Wi-Fi) to transmit the localization data and enhanced acoustic sound signals to other devices.
- communication functionality wireless chip set, Bluetooth, Wi-Fi
- other devices in proximity or communicatively coupled can receive enhanced audio and directional data, for example, on request, responsive to an acoustic event (e.g., sound signature detection), a recognized voice (e.g., speech recognition), or combination thereof, for instance GPS localization information and voice recognition.
- acoustic event e.g., sound signature detection
- a recognized voice e.g., speech recognition
- GPS localization information and voice recognition for instance GPS localization information and voice recognition.
- the method implemented by way of the processor 173 performs the steps of calculating a complex coherence between the first and second microphone signal, determining a measured frequency dependent phase angle of the complex coherence, comparing the measured frequency dependent phase angle with a reference phase angle threshold and determining if the measured frequency dependent phase angle exceeds a predetermined threshold from the reference phase angle, outputting/updating a set of frequency dependent filter coefficients 176 based on the comparing to produce an updated filter coefficient set, and filtering the first microphone signal or the second microphone signal with the updated filter coefficient set 176 to enhance a directional sensitivity and quality of the microphone signal, from either or both microphones 171 and 172 .
- the processor 173 can further communicate directional data derived from the coherence based processing method with the microphone signal to the secondary device, where the directional data includes at least a direction of a sound source, and adjusts at least one parameter of the device in view of the directional data.
- the processor can focus or pan a camera of the secondary device to the sound source as will be described ahead in specific embodiments.
- the processor can perform an image stabilization and maintain a focused centering of the camera responsive to movement of the secondary device, and, if more than one camera is present and communicatively coupled thereto, selectively switch between one or more cameras of the secondary device responsive to detecting from the directional data whether a sound source is in view of the one or more cameras.
- aspects of signal processing performed by the processor may be performed by one or more processors residing in separate devices communicatively coupled to one another. At least one of the microphones are processed with an adaptive filter, where the filter is adaptive so that sound from one direction is passed through and sounds from other directions are blocked, with the resulting signal directed to, for instance, a loudspeaker or sound analysis system such as an Automatic Speech Recognition (ASR) system.
- ASR Automatic Speech Recognition
- spectral components e.g., magnitude, phase, onsets, decay, SNR ratios
- spectral components e.g., magnitude, phase, onsets, decay, SNR ratios
- these features are segregated by the directional enhancement and can be input to sound recognition systems to determine what type of other sounds are present (e.g., sirens, wind, rain, etc.).
- feature extraction for sound recognition is performed in conjunction with directional speech enhancement to identify sounds and sound directions and apply an importance weighting based on the environment context, for example, where is the user (e.g, GPS, navigation) and in proximity to what services (e.g, businesses, restaurants, police, games etc.) and other people (e.g., ad-hoc users, wi-fi users, internet browers, etc.)
- directional speech enhancement to identify sounds and sound directions and apply an importance weighting based on the environment context, for example, where is the user (e.g, GPS, navigation) and in proximity to what services (e.g, businesses, restaurants, police, games etc.) and other people (e.g., ad-hoc users, wi-fi users, internet browers, etc.)
- the system 100 can be configured to be part of any suitable media or computing device.
- the system may be housed in the computing device or may be coupled to the computing device.
- the computing device may include, without being limited to wearable and/or body-borne (also referred to herein as bearable) computing devices.
- wearable/body-borne computing devices include head-mounted displays, earpieces, smart watches, smartphones, cochlear implants and artificial eyes.
- wearable computing devices relate to devices that may be worn on the body.
- Bearable computing devices relate to devices that may be worn on the body or in the body, such as implantable devices.
- Bearable computing devices may be configured to be temporarily or permanently installed in the body.
- Wearable devices may be worn, for example, on or in clothing, watches, glasses, shoes, as well as any other suitable accessory.
- the system 100 can also be deployed for use in non-wearable contexts, for example, within cars equipped to take photos, that with the directional sound information captured herein and with location data, can track and identify where the car is, the occupants in the car, and the acoustic sounds from conversations in the vehicle, and interpreting what they are saying or intending, and in certain cases, predicting a destination.
- photo equipped vehicles enabled with the acoustic device 170 to direct the camera to take photos at specific directions of the sound field, and secondly, to process and analyze the acoustic content for information and data mining.
- the system 100 can also be configured for individual earpieces (left or right) or include an additional pair of microphones on a second earpiece in addition to the first earpiece.
- the system 100 can be configured to be optimized for different microphone spacing's.
- the first 121 and second 122 microphones are mechanically mounted to one side of eyeglasses.
- the embodiment 120 can be configured for individual sides (left or right) or include an additional pair of microphones on a second side in addition to the first side.
- the eyeglasses 120 can include one or more optical elements, for example, cameras 123 and 124 situated at the front or other direction for taking pictures.
- a processor 140 / 160 communicatively coupled to the first microphone 121 and the second microphone 122 for analyzing phase coherence and updating the adaptive filter may be present.
- the eyeglasses 120 may be worn by a user to enhance a directional component of a captured microphone signal to enhance the voice quality.
- the eyeglasses 120 upon detecting another person speaking can perform the method steps contemplated herein for enhancing that users voice arriving from a particular direction.
- This enhanced voice signal that of the secondary talker, or the primary talker wearing the eyeglasses, can then be directed to an automatic speech recognition system (ASR).
- ASR automatic speech recognition system
- Directional data can also be supplied to the ASR for providing supplemental information needed to parse or recognize words, phrases or sentences.
- FIG. 1D depicts a first media device 140 as a mobile device (i.e., smartphone) which can be communicatively coupled to either or both of the wearable computing devices ( 100 / 120 ).
- FIG. 1E depicts a second media device 140 as a wristwatch device which also can be communicatively coupled to the one or more wearable computing devices ( 100 / 120 ).
- the processor performing the coherence analysis for updating the adaptive filter is included thereon, for example, within a digital signal processor or other software programmable device within, or coupled to, the media device 140 or 160 .
- FIG. 9B components of the media device for implementing coherence analysis functionality will be explained in further detail.
- the mobile device 140 may be handled by a user to enhance a directional component of a captured microphone signal to enhance the voice quality.
- the mobile device 140 upon detecting another person speaking can perform the method steps contemplated herein for enhancing that users voice arriving from a separate direction.
- the mobile device 140 can adjust one or more component operating parameters, for instance, focusing or panning a camera toward the detected secondary talker.
- a back camera element 142 on the mobile device 140 can visually track a secondary talker within acoustic vicinity of the mobile device 140 .
- a front camera element 141 can visually track a secondary talker that may be in vicinity of the primary talker holding the phone.
- the mobile device 140 embodying the directional enhancement methods contemplated herein can also selectively switch between cameras, for example, deciding whether the mobile device is laying on a table, by which, the camera element on that side would be temporarily disabled. Although such methods may be performed by image processing the method of directional enhancement herein is useful in dark (e.g., nighttime) conditions where a camera may not be able to localize its direction.
- the mobile device by way of the processor can track a direction of a voice identified in the sound source, and from the tracking, adjusting a display parameter of the secondary device to visually follow the sound source.
- the directional tracking can also be used on the person directly handling the device. For instance, in an application where a camera element 141 on the mobile device 140 captures images or video of the person handling the device, the acoustic device microphone array in conjunction with the processing capabilities, either on an integrated circuit within the mobile device or through an internet connection to the mobile device 140 , detects a directional component of the user's voice, effectively localizing the user with respect to the display 142 of the mobile device, and then tracks the user on the display.
- the tracked user identified as the sound souce, for example face tracking
- another device for example, a second phone in a call with the user
- the display would update and center the user on the phone based on the voice directional data.
- the application for example, a face time application on a mobile device.
- the system 100 may represent a single device or a family of devices configured, for example, in a master-slave or master-master arrangement.
- components of the system 100 may be distributed among one or more devices, such as, but not limited to, the media device illustrated in FIG. 1D and the wristwatch in FIG. 1E . That is, the components of the system 100 may be distributed among several devices (such as a smartphone, a smartwatch, an optical head-mounted display, an earpiece, etc.).
- the devices (for example, those illustrated in FIG. 1B and FIG. 1C ) may be coupled together via any suitable connection, for example, to the media device in FIG. 1D and/or the wristwatch in FIG. 1E , such as, without being limited to, a wired connection, a wireless connection or an optical connection.
- the computing devices shown in FIGS. 1D and 1E can include any device having some processing capability for performing a desired function, for instance, as shown in FIG. 9B .
- Computing devices may provide specific functions, such as heart rate monitoring or pedometer capability, to name a few.
- More advanced computing devices may provide multiple and/or more advanced functions, for instance, to continuously convey heart signals or other continuous biometric data.
- advanced “smart” functions and features similar to those provided on smartphones, smartwatches, optical head-mounted displays or helmet-mounted displays can be included therein.
- Example functions of computing devices may include, without being limited to, capturing images and/or video, displaying images and/or video, presenting audio signals, presenting text messages and/or emails, identifying voice commands from a user, browsing the web, etc.
- FIG. 2 a general method 200 for directional enhancement of audio using analysis of the inter-microphone coherence phase angle is shown.
- the method 200 may be practiced with more or less than the number of steps shown.
- the method 200 can be practiced by the components presented in the figures herein though is not limited to the components shown.
- the processing steps may be performed by, or shared with, another device, wearable or non-wearable, communicatively coupled, such as the mobile device 140 shown in FIG. 1D , or the wristwatch 160 shown in FIG. 1E . That is, the method 200 is not limited to the devices described herein, but in fact any device providing certain functionality for performing the method steps herein described, for example, by a processor implementing programs to execute one or more computer readable instructions.
- the earpiece 500 is connected to a voice communication device (e.g. mobile telephone, radio, computer device) and/or audio content delivery device (e.g. portable media player, computer device).
- a voice communication device e.g. mobile telephone, radio, computer device
- audio content delivery device e.g. portable media player, computer device
- the communication earphone/headset system comprises a sound isolating component for blocking the users ear meatus (e.g. using foam or an expandable balloon); an Ear Canal Receiver (ECR, i.e. loudspeaker) for receiving an audio signal and generating a sound field in a user ear-canal; at least one ambient sound microphone (ASM) for receiving an ambient sound signal and generating at least one ASM signal; and an optional Ear Canal Microphone (ECM) for receiving an ear-canal signal measured in the user's occluded ear-canal and generating an ECM signal.
- ECM Ear Canal Microphone
- a signal processing system receives an Audio Content (AC) signal (e.g. music or speech audio signal) from the said communication device (e.g.
- AC Audio Content
- the signal processing system mixes the at least one ASM and AC signal and transmits the resulting mixed signal to the ECR in the loudspeaker.
- the first microphone and the second microphone capture a first signal and second signal respectively at step 202 and 204 .
- the order of the capture for which signal arrives first is a function of the sound source location; it not the microphone number; either the first or second microphone may capture the first microphone signal.
- the system analyzes a coherence between the two microphone signals (M 1 and M 2 ).
- the complex coherence estimate, Cxy as determined in step 206 is a function of the power spectral densities, Pxx(f) and Pyy(f), of x and y, and the cross power spectral density, Pxy(f), of x and y,
- the window length for the power spectral densities and cross power spectral density in the preferred embodiment are approximately 3 ms ( ⁇ 2 to 5 ms).
- the time-smoothing for updating the power spectral densities and cross power spectral density in the preferred embodiment is approximately 0.5 seconds (e.g. for the power spectral density level to increase from ⁇ 60 dB to 0 dB) but may be lower to 0.2 ms.
- the magnitude squared coherence estimate is a function of frequency with values between 0 and 1 that indicates how well x corresponds to y at each frequency.
- the signals x and y correspond to the signals from a first and second microphone.
- phase angle refers to the angular component of the polar coordinate representation, it is synonymous with the term “phase”, and as shown in step 208 can be calculated by the arctangent of the ratio of the imaginary component of the coherence to the real component of the coherence, as is well known.
- the reference phase angles can be selected based on a desired angle of incidence, where the angle can be selected using a polar plot representation on a GUI. For instance, the user can select the reference phase angle to direct the microphone array sensitivity.
- the phase angle is calculated; a measured frequency dependent phase angle of the complex coherence is determined.
- the phase vector from this phase angle can be optionally unwrapped, i.e. not bounded between ⁇ pi and +pi, but in practice this step does not affect the quality of the process.
- the phase angle of the complex coherence is unwrapped to produce an unwrapped phase angle, and the measured frequency dependent phase angle can be replaced with the unwrapped phase angle.
- Step 210 is a comparison step where the measured phase angle vector is compared with a reference (or “target”) phase angle vector stored on computer readable memory 212 . More specifically, the measured frequency dependent phase angle is compared with a reference phase angle threshold and determining if the measured frequency dependent phase angle exceeds a predetermined threshold from the reference phase angle
- the comparison 214 is simply a comparison of the relative signed difference between the measured and reference phase angles.
- the update of the adaptive filter in step 216 is such that the frequency band of the filter is increased towards unity.
- the update of the adaptive filter in step 216 is such that the frequency band of the filter is decreased towards zero.
- the step of updating the set of frequency dependent filter coefficients includes reducing the coefficient values towards zero if the phase angle differs significantly from the reference phase angle, and increasing the coefficient values are increased towards unity if the phase angle substantially matches the reference phase angle.
- the reference phase angles can be determined empirically from a calibration measurement process as will be described in FIG. 3 , or the reference phase angles can be determined mathematically.
- the reference phase angle vector can be selected from a set of reference phase angles, where there is a different reference phase angle vector for a corresponding desired direction of sensitivity (angle theta, 306 , in FIG. 3 ). For instance if the desired direction of sensitivity is zero degrees relative to the 2 microphones then one reference phase angle vector may be used, but if the desired direction of sensitivity is 90 degrees relative to the 2 microphones then a second reference phase angle vector is used. An example set of reference phase angles is shown in FIG. 4 .
- FIG. 3 depicts a measurement setup for acquiring target inter-microphone coherence between omni-directional microphones M 1 and M 2 for sound targets at particular angles of incident. It illustrates a measurement configuration 300 for depicting an exemplary method from obtaining empirical reference phase angle vectors for a desired direction of sensitivity (angle theta, 306 ).
- a test audio signal 302 e.g. a white noise audio sample, is reproduced from a loudspeaker 304 at an angle of incidence 306 relative to the first and second microphones M 1 308 and M 2 310 .
- the phase angle of the inter microphone coherence is analyzed according to the method described previously using audio analysis system 312 .
- the reference phase angles can be obtained by empirical measurement of a two microphone system in response to a close target sound source at a determined relative angle of incidence to the microphones.
- This angle gradient is similar to the group delay of a signal spectrum, and can be used as a target criteria to update the filter, as previously described.
- FIG. 5 shows a multi-microphone configuration and control interface to select desired target direction and output source location.
- the system 500 as illustrated uses three microphones M 1 502 , M 2 504 , M 3 506 although more can be supported.
- the three microphones are arranged tangentially (i.e. at vertices of a right-angled triangle), with equal spacing between M 1 -M 3 and M 1 -M 2 .
- Microphones are directed to an audio processing system 508 to process microphone pairs M 1 -M 2 and M 1 -M 3 according to the method described previously. With such a system, the angle theta for the target angle of incidence would be modified by 90 degrees for the M 1 -M 3 system, and the output of the 2 systems can be combined using a summer.
- System 500 further shows an optional output 512 that can be used in a configuration whereby the angle of incidence of the target sound source in unknown. The method for determining the angle of incidence is described next.
- FIG. 6 depicts a method 600 for determining source location from analysis of measured coherence angle in accordance with an exemplary embodiment.
- the method 600 may be practiced with more or less than the number of steps shown.
- the method 600 can be practiced by the components presented in the figures herein though is not limited to the components shown.
- Method 600 describes an exemplary method of determining the angle of incidence of a sound source relative to a two-microphone array, based on an analysis of the angle of the coherence, and associating this angle with a reference angle from a set of coherence-angle vectors.
- the inter-microphone coherence Cxy and it's phase angle is calculated as previously described in method 600 , and reproduced below for continuity.
- the first microphone and the second microphone capture a first signal and second signal respectively at step 602 and 604 .
- the order of the capture for which signal arrives first is a function of the sound source location; it not the microphone number; either the first or second microphone may capture the first microphone signal.
- the system analyzes a coherence between the two microphone signals (M 1 and M 2 ).
- the complex coherence estimate, Cxy as determined in step 206 is a function.
- the phase angle is calculated; a measured frequency dependent phase angle of the complex coherence is determined.
- the measured angle is then compared with one angle vector from a set of reference angle vectors 610 , and the Mean Square Error (MSE) calculated:
- the reference angle vector that yields the lowest MSE is then used to update the filter in step 618 as previously described.
- the angle of incidence theta for the reference angle vector that yields the lowest MSE is used as an estimate for the angle of incidence of the target sound source, and this angle of incidence is used as a source direction estimate 616 .
- the source direction estimate can be used to control a device such as a camera to move its focus in the estimated direction of the sound source.
- the source direction estimate can also be used in security systems, e.g. to detect an intruder that creates a noise in a target direction.
- FIG. 7 For a detailed view and description of the components of the earpiece 700 (which may be coupled to the aforementioned devices and media device 800 of FIG. 8 ); components which may be referred to in one implementation for practicing methods 200 and 600 .
- the aforementioned devices headset 100 , eyeglasses 120 , mobile device 140 , wrist watch 160 , earpiece 500
- the processing steps of method 200 can also implement the processing steps of method 200 for practicing the novel aspects of directional enhancement of speech signals using small microphone arrays.
- SI Sound isolating
- headsets are becoming increasingly popular for music listening and voice communication.
- SI earphones enable the user to hear and experience an incoming audio content signal (be it speech from a phone call or music audio from a music player) clearly in loud ambient noise environments, by attenuating the level of ambient sound in the user ear-canal.
- the disadvantage of such SI earphones/headsets is that the user is acoustically detached from their local sound environment, and communication with people in their immediate environment is therefore impaired: i.e. the earphone has a reduced situational awareness due to the acoustic masking properties of the earphone.
- a non Sound Isolating (SI) earphone can reduce the ability of an earphone wearer to hear local sound events as the earphone wearer can be distracted by incoming voice message or reproduced music on the earphones.
- the ambient sound microphone (ASM) located on an SI or non-SI earphone can be used to increase situation awareness of the earphone wearer by passing the ASM signal to the loudspeaker in the earphone.
- ASM ambient sound microphone
- Such a “sound pass through” utility can be enhanced by processing at least one of the microphone's signals, or a combination of the microphone signals, with a “spatial filter”, i.e.
- an electronic filter whereby sound originating from one direction (i.e. angle of incidence relative to the microphones) are passed through and sounds from other directions are attenuated.
- Such a spatial filtering system can increase perceived speech intelligibility by increasing the signal-to-noise ratio (SNR).
- FIG. 7 is an illustration of an earpiece device 500 that can be connected to the system 100 of FIG. 1A for performing the inventive aspects herein disclosed.
- the earpiece 700 contains numerous electronic components, many audio related, each with separate data lines conveying audio data.
- the system 100 can include a separate earpiece 700 for both the left and right ear. In such arrangement, there may be anywhere from 8 to 12 data lines, each containing audio, and other control information (e.g., power, ground, signaling, etc.)
- the earpiece 700 comprises an electronic housing unit 701 and a sealing unit 708 .
- the earpiece depicts an electro-acoustical assembly for an in-the-ear acoustic assembly, as it would typically be placed in an ear canal 724 of a user.
- the earpiece can be an in the ear earpiece, behind the ear earpiece, receiver in the ear, partial-fit device, or any other suitable earpiece type.
- the earpiece can partially or fully occlude ear canal 724 , and is suitable for use with users having healthy or abnormal auditory functioning.
- the earpiece includes an Ambient Sound Microphone (ASM) 720 to capture ambient sound, an Ear Canal Receiver (ECR) 714 to deliver audio to an ear canal 724 , and an Ear Canal Microphone (ECM) 706 to capture and assess a sound exposure level within the ear canal 724 .
- ASM Ambient Sound Microphone
- ECR Ear Canal Receiver
- ECM Ear Canal Microphone
- the earpiece can partially or fully occlude the ear canal 724 to provide various degrees of acoustic isolation.
- assembly is designed to be inserted into the user's ear canal 724 , and to form an acoustic seal with the walls of the ear canal 724 at a location between the entrance to the ear canal 724 and the tympanic membrane (or ear drum). In general, such a seal is typically achieved by means of a soft and compliant housing of sealing unit 708 .
- Sealing unit 708 is an acoustic barrier having a first side corresponding to ear canal 724 and a second side corresponding to the ambient environment.
- sealing unit 708 includes an ear canal microphone tube 710 and an ear canal receiver tube 714 .
- Sealing unit 708 creates a closed cavity of approximately 5 cc between the first side of sealing unit 708 and the tympanic membrane in ear canal 724 .
- the ECR (speaker) 714 is able to generate a full range bass response when reproducing sounds for the user.
- This seal also serves to significantly reduce the sound pressure level at the user's eardrum resulting from the sound field at the entrance to the ear canal 724 .
- This seal is also a basis for a sound isolating performance of the electro-acoustic assembly.
- the second side of sealing unit 708 corresponds to the earpiece, electronic housing unit 700 , and ambient sound microphone 720 that is exposed to the ambient environment.
- Ambient sound microphone 720 receives ambient sound from the ambient environment around the user.
- Electronic housing unit 700 houses system components such as a microprocessor 716 , memory 704 , battery 702 , ECM 706 , ASM 720 , ECR, 714 , and user interface 722 .
- Microprocessor 916 (or processor 716 ) can be a logic circuit, a digital signal processor, controller, or the like for performing calculations and operations for the earpiece.
- Microprocessor 716 is operatively coupled to memory 704 , ECM 706 , ASM 720 , ECR 714 , and user interface 720 .
- a wire 718 provides an external connection to the earpiece.
- Battery 702 powers the circuits and transducers of the earpiece.
- Battery 702 can be a rechargeable or replaceable battery.
- electronic housing unit 700 is adjacent to sealing unit 708 . Openings in electronic housing unit 700 receive ECM tube 710 and ECR tube 712 to respectively couple to ECM 706 and ECR 714 .
- ECR tube 712 and ECM tube 710 acoustically couple signals to and from ear canal 724 .
- ECR outputs an acoustic signal through ECR tube 712 and into ear canal 724 where it is received by the tympanic membrane of the user of the earpiece.
- ECM 714 receives an acoustic signal present in ear canal 724 though ECM tube 710 . All transducers shown can receive or transmit audio signals to a processor 716 that undertakes audio signal processing and provides a transceiver for audio via the wired (wire 718 ) or a wireless communication path.
- FIG. 8 depicts various components of a multimedia device 850 suitable for use for use with, and/or practicing the aspects of the inventive elements disclosed herein, for instance method 200 and method 300 , though is not limited to only those methods or components shown.
- the device 850 comprises a wired and/or wireless transceiver 852 , a user interface (UI) display 854 , a memory 856 , a location unit 858 , and a processor 860 for managing operations thereof.
- the media device 850 can be any intelligent processing platform with Digital signal processing capabilities, application processor, data storage, display, input modality like touch-screen or keypad, microphones, speaker 866 , Bluetooth, and connection to the internet via WAN, Wi-Fi, Ethernet or USB.
- a power supply 862 provides energy for electronic components.
- the transceiver 852 can utilize common wire-line access technology to support POTS or VoIP services.
- the transceiver 852 can utilize common technologies to support singly or in combination any number of wireless access technologies including without limitation BluetoothTM Wireless Fidelity (WiFi), Worldwide Interoperability for Microwave Access (WiMAX), Ultra Wide Band (UWB), software defined radio (SDR), and cellular access technologies such as CDMA-1X, W-CDMA/HSDPA, GSM/GPRS, EDGE, TDMA/EDGE, and EVDO.
- SDR can be utilized for accessing a public or private communication spectrum according to any number of communication protocols that can be dynamically downloaded over-the-air to the communication device. It should be noted also that next generation wireless access technologies can be applied to the present disclosure.
- the location unit 858 can utilize common technology such as a GPS (Global Positioning System) receiver that can intercept satellite signals and there from determine a location fix of the portable device 850 .
- GPS Global Positioning System
- the controller processor 860 can utilize computing technologies such as a microprocessor and/or digital signal processor (DSP) with associated storage memory such a Flash, ROM, RAM, SRAM, DRAM or other like technologies for controlling operations of the aforementioned components of the communication device.
- DSP digital signal processor
- Social media refers to interaction among people in which they create, share, and/or exchange information and ideas in virtual communities and networks and allow the creation and exchange of user-generated content.
- Social media leverages mobile and web-based technologies to create highly interactive platforms through which individuals and communities share, co-create, discuss, and modify user-generated content.
- social media is considered exclusive in that it does not adequately allow others the transfer of information from one to another, and there is disparity of information available, including issues with trustworthiness and reliability of information presented, concentration, ownership of media content, and the meaning of interactions created by social media.
- social media is personalized based on acoustic interactions through user's voices and environmental sounds in their vicinity providing positive effects allowing individuals to express themselves and form friendships in a socially recognized manner.
- the method 900 can be practiced by any one, or combination of, the devices and components expressed herein.
- the system 900 also include methods that can be realized in software or hardware by any of the devices or components disclosed herein and also coupled to other devices and systems, for example, those shown in FIGS. 1A-1E , FIG. 3 , FIGS. 6-8 .
- the method 900 is not limited to the order of steps shown in FIG. 9 , and may be practiced in a different order, and include additional steps herein contemplated.
- the method 900 can start in a state where a user of a mobile device is in a social setting and surrounded by other people, of which some may also have mobile devices (e.g., smartphone, laptop, internet device, etc) and others which do not. Some of these users may have active network (wi-fi, internet, cloud, etc) connections and others may be active on data and voice networks (cellular, packet data, wireless). Others may be interconnected over short range communication protocols (e.g., IEEE, Bluetooth, wi-fi, etc.) or not. Understandably, other social contexts are possible, for example, where a sound monitoring device incorporating the acoustic sensor 170 is positioned in a building or other location where people are present, and for instance, in combination with video monitoring.
- a sound monitoring device incorporating the acoustic sensor 170 is positioned in a building or other location where people are present, and for instance, in combination with video monitoring.
- inventive subject matter may be referred to herein, individually and/or collectively, by the term “invention” merely for convenience and without intending to voluntarily limit the scope of this application to any single invention or inventive concept if more than one is in fact disclosed.
- inventive subject matter may be referred to herein, individually and/or collectively, by the term “invention” merely for convenience and without intending to voluntarily limit the scope of this application to any single invention or inventive concept if more than one is in fact disclosed.
Landscapes
- Health & Medical Sciences (AREA)
- General Health & Medical Sciences (AREA)
- Otolaryngology (AREA)
- Physics & Mathematics (AREA)
- Engineering & Computer Science (AREA)
- Acoustics & Sound (AREA)
- Signal Processing (AREA)
- Circuit For Audible Band Transducer (AREA)
Abstract
Description
Claims (20)
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US14/108,883 US9271077B2 (en) | 2013-12-17 | 2013-12-17 | Method and system for directional enhancement of sound using small microphone arrays |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US14/108,883 US9271077B2 (en) | 2013-12-17 | 2013-12-17 | Method and system for directional enhancement of sound using small microphone arrays |
Publications (2)
Publication Number | Publication Date |
---|---|
US20150172814A1 US20150172814A1 (en) | 2015-06-18 |
US9271077B2 true US9271077B2 (en) | 2016-02-23 |
Family
ID=53370115
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US14/108,883 Active 2034-05-28 US9271077B2 (en) | 2013-12-17 | 2013-12-17 | Method and system for directional enhancement of sound using small microphone arrays |
Country Status (1)
Country | Link |
---|---|
US (1) | US9271077B2 (en) |
Cited By (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN109104683A (en) * | 2018-07-13 | 2018-12-28 | 深圳市小瑞科技股份有限公司 | A kind of method and correction system of dual microphone phase measurement correction |
US10848872B2 (en) | 2014-12-27 | 2020-11-24 | Intel Corporation | Binaural recording for processing audio signals to enable alerts |
Families Citing this family (36)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP6028502B2 (en) * | 2012-10-03 | 2016-11-16 | 沖電気工業株式会社 | Audio signal processing apparatus, method and program |
US9270244B2 (en) * | 2013-03-13 | 2016-02-23 | Personics Holdings, Llc | System and method to detect close voice sources and automatically enhance situation awareness |
EP2928211A1 (en) * | 2014-04-04 | 2015-10-07 | Oticon A/s | Self-calibration of multi-microphone noise reduction system for hearing assistance devices using an auxiliary device |
CA2949929A1 (en) * | 2014-05-26 | 2015-12-03 | Vladimir Sherman | Methods circuits devices systems and associated computer executable code for acquiring acoustic signals |
US10149047B2 (en) * | 2014-06-18 | 2018-12-04 | Cirrus Logic Inc. | Multi-aural MMSE analysis techniques for clarifying audio signals |
US9992593B2 (en) * | 2014-09-09 | 2018-06-05 | Dell Products L.P. | Acoustic characterization based on sensor profiling |
WO2016152007A1 (en) * | 2015-03-25 | 2016-09-29 | パナソニックIpマネジメント株式会社 | Image processing device, monitoring system provided with same, and image processing method |
US9401158B1 (en) * | 2015-09-14 | 2016-07-26 | Knowles Electronics, Llc | Microphone signal fusion |
US9959884B2 (en) | 2015-10-09 | 2018-05-01 | Cirrus Logic, Inc. | Adaptive filter control |
US20170125010A1 (en) * | 2015-10-29 | 2017-05-04 | Yaniv Herman | Method and system for controlling voice entrance to user ears, by designated system of earphone controlled by Smartphone with reversed voice recognition control system |
US10158932B2 (en) * | 2015-12-15 | 2018-12-18 | Westone Laboratories, Inc. | Ambient sonic low-pressure equalization |
US10165352B2 (en) * | 2015-12-15 | 2018-12-25 | Westone Laboratories, Inc. | Ambient sonic low-pressure equalization |
US9830930B2 (en) | 2015-12-30 | 2017-11-28 | Knowles Electronics, Llc | Voice-enhanced awareness mode |
US9779716B2 (en) | 2015-12-30 | 2017-10-03 | Knowles Electronics, Llc | Occlusion reduction and active noise reduction based on seal quality |
US9812149B2 (en) * | 2016-01-28 | 2017-11-07 | Knowles Electronics, Llc | Methods and systems for providing consistency in noise reduction during speech and non-speech periods |
US11445305B2 (en) | 2016-02-04 | 2022-09-13 | Magic Leap, Inc. | Technique for directing audio in augmented reality system |
CN114189793B (en) * | 2016-02-04 | 2024-03-19 | 奇跃公司 | Techniques for directing audio in augmented reality systems |
KR102468148B1 (en) * | 2016-02-19 | 2022-11-21 | 삼성전자주식회사 | Electronic device and method for classifying voice and noise thereof |
US10469976B2 (en) * | 2016-05-11 | 2019-11-05 | Htc Corporation | Wearable electronic device and virtual reality system |
US10547947B2 (en) * | 2016-05-18 | 2020-01-28 | Qualcomm Incorporated | Device for generating audio output |
US10313822B2 (en) | 2016-11-13 | 2019-06-04 | EmbodyVR, Inc. | Image and audio based characterization of a human auditory system for personalized audio reproduction |
US10375498B2 (en) * | 2016-11-16 | 2019-08-06 | Dts, Inc. | Graphical user interface for calibrating a surround sound system |
US10506327B2 (en) * | 2016-12-27 | 2019-12-10 | Bragi GmbH | Ambient environmental sound field manipulation based on user defined voice and audio recognition pattern analysis system and method |
KR102308937B1 (en) | 2017-02-28 | 2021-10-05 | 매직 립, 인코포레이티드 | Virtual and real object recording on mixed reality devices |
US10433051B2 (en) * | 2017-05-29 | 2019-10-01 | Staton Techiya, Llc | Method and system to determine a sound source direction using small microphone arrays |
DK3425928T3 (en) * | 2017-07-04 | 2021-10-18 | Oticon As | SYSTEM INCLUDING HEARING AID SYSTEMS AND SYSTEM SIGNAL PROCESSING UNIT AND METHOD FOR GENERATING AN IMPROVED ELECTRICAL AUDIO SIGNAL |
US11150869B2 (en) | 2018-02-14 | 2021-10-19 | International Business Machines Corporation | Voice command filtering |
US10817252B2 (en) * | 2018-03-10 | 2020-10-27 | Staton Techiya, Llc | Earphone software and hardware |
US11238856B2 (en) | 2018-05-01 | 2022-02-01 | International Business Machines Corporation | Ignoring trigger words in streamed media content |
US11200890B2 (en) * | 2018-05-01 | 2021-12-14 | International Business Machines Corporation | Distinguishing voice commands |
US10425745B1 (en) * | 2018-05-17 | 2019-09-24 | Starkey Laboratories, Inc. | Adaptive binaural beamforming with preservation of spatial cues in hearing assistance devices |
CN109932054B (en) * | 2019-04-24 | 2024-01-26 | 北京耘科科技有限公司 | Wearable acoustic detection and identification system |
US11355108B2 (en) | 2019-08-20 | 2022-06-07 | International Business Machines Corporation | Distinguishing voice commands |
KR102304815B1 (en) * | 2020-01-06 | 2021-09-23 | 엘지전자 주식회사 | Audio apparatus and method thereof |
US11778408B2 (en) | 2021-01-26 | 2023-10-03 | EmbodyVR, Inc. | System and method to virtually mix and audition audio content for vehicles |
CN114007157A (en) * | 2021-10-28 | 2022-02-01 | 中北大学 | Intelligent noise reduction communication earphone |
Citations (21)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20030035551A1 (en) | 2001-08-20 | 2003-02-20 | Light John J. | Ambient-aware headset |
US20050175189A1 (en) | 2004-02-06 | 2005-08-11 | Yi-Bing Lee | Dual microphone communication device for teleconference |
US20060074693A1 (en) | 2003-06-30 | 2006-04-06 | Hiroaki Yamashita | Audio coding device with fast algorithm for determining quantization step sizes based on psycho-acoustic model |
US20060133621A1 (en) | 2004-12-22 | 2006-06-22 | Broadcom Corporation | Wireless telephone having multiple microphones |
US20070053522A1 (en) | 2005-09-08 | 2007-03-08 | Murray Daniel J | Method and apparatus for directional enhancement of speech elements in noisy environments |
US20070076898A1 (en) | 2003-11-24 | 2007-04-05 | Koninkiljke Phillips Electronics N.V. | Adaptive beamformer with robustness against uncorrelated noise |
US20070076896A1 (en) | 2005-09-28 | 2007-04-05 | Kabushiki Kaisha Toshiba | Active noise-reduction control apparatus and method |
US20070088544A1 (en) | 2005-10-14 | 2007-04-19 | Microsoft Corporation | Calibration based beamforming, non-linear adaptive filtering, and multi-sensor headset |
US20070230712A1 (en) | 2004-09-07 | 2007-10-04 | Koninklijke Philips Electronics, N.V. | Telephony Device with Improved Noise Suppression |
US20080147397A1 (en) | 2006-12-14 | 2008-06-19 | Lars Konig | Speech dialog control based on signal pre-processing |
US8098844B2 (en) | 2002-02-05 | 2012-01-17 | Mh Acoustics, Llc | Dual-microphone spatial noise suppression |
WO2012078670A1 (en) | 2010-12-06 | 2012-06-14 | The Board Of Regents Of The University Of Texas System | Method and system for enhancing the intelligibility of sounds relative to background noise |
US20120163606A1 (en) * | 2009-06-23 | 2012-06-28 | Nokia Corporation | Method and Apparatus for Processing Audio Signals |
US8391501B2 (en) | 2006-12-13 | 2013-03-05 | Motorola Mobility Llc | Method and apparatus for mixing priority and non-priority audio signals |
US8401206B2 (en) | 2009-01-15 | 2013-03-19 | Microsoft Corporation | Adaptive beamformer using a log domain optimization criterion |
US8467543B2 (en) | 2002-03-27 | 2013-06-18 | Aliphcom | Microphone and voice activity detection (VAD) configurations for use with communication systems |
US8503704B2 (en) | 2009-04-07 | 2013-08-06 | Cochlear Limited | Localisation in a bilateral hearing device system |
US8583428B2 (en) | 2010-06-15 | 2013-11-12 | Microsoft Corporation | Sound source separation using spatial filtering and regularization phases |
US8600454B2 (en) | 2010-09-02 | 2013-12-03 | Apple Inc. | Decisions on ambient noise suppression in a mobile communications handset device |
US8606571B1 (en) | 2010-04-19 | 2013-12-10 | Audience, Inc. | Spatial selectivity noise reduction tradeoff for multi-microphone systems |
US8837747B2 (en) * | 2010-09-28 | 2014-09-16 | Kabushiki Kaisha Toshiba | Apparatus, method, and program product for presenting moving image with sound |
-
2013
- 2013-12-17 US US14/108,883 patent/US9271077B2/en active Active
Patent Citations (22)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20030035551A1 (en) | 2001-08-20 | 2003-02-20 | Light John J. | Ambient-aware headset |
US8098844B2 (en) | 2002-02-05 | 2012-01-17 | Mh Acoustics, Llc | Dual-microphone spatial noise suppression |
US8467543B2 (en) | 2002-03-27 | 2013-06-18 | Aliphcom | Microphone and voice activity detection (VAD) configurations for use with communication systems |
US20060074693A1 (en) | 2003-06-30 | 2006-04-06 | Hiroaki Yamashita | Audio coding device with fast algorithm for determining quantization step sizes based on psycho-acoustic model |
US20070076898A1 (en) | 2003-11-24 | 2007-04-05 | Koninkiljke Phillips Electronics N.V. | Adaptive beamformer with robustness against uncorrelated noise |
US20050175189A1 (en) | 2004-02-06 | 2005-08-11 | Yi-Bing Lee | Dual microphone communication device for teleconference |
US20070230712A1 (en) | 2004-09-07 | 2007-10-04 | Koninklijke Philips Electronics, N.V. | Telephony Device with Improved Noise Suppression |
US20090209290A1 (en) | 2004-12-22 | 2009-08-20 | Broadcom Corporation | Wireless Telephone Having Multiple Microphones |
US20060133621A1 (en) | 2004-12-22 | 2006-06-22 | Broadcom Corporation | Wireless telephone having multiple microphones |
US20070053522A1 (en) | 2005-09-08 | 2007-03-08 | Murray Daniel J | Method and apparatus for directional enhancement of speech elements in noisy environments |
US20070076896A1 (en) | 2005-09-28 | 2007-04-05 | Kabushiki Kaisha Toshiba | Active noise-reduction control apparatus and method |
US20070088544A1 (en) | 2005-10-14 | 2007-04-19 | Microsoft Corporation | Calibration based beamforming, non-linear adaptive filtering, and multi-sensor headset |
US8391501B2 (en) | 2006-12-13 | 2013-03-05 | Motorola Mobility Llc | Method and apparatus for mixing priority and non-priority audio signals |
US20080147397A1 (en) | 2006-12-14 | 2008-06-19 | Lars Konig | Speech dialog control based on signal pre-processing |
US8401206B2 (en) | 2009-01-15 | 2013-03-19 | Microsoft Corporation | Adaptive beamformer using a log domain optimization criterion |
US8503704B2 (en) | 2009-04-07 | 2013-08-06 | Cochlear Limited | Localisation in a bilateral hearing device system |
US20120163606A1 (en) * | 2009-06-23 | 2012-06-28 | Nokia Corporation | Method and Apparatus for Processing Audio Signals |
US8606571B1 (en) | 2010-04-19 | 2013-12-10 | Audience, Inc. | Spatial selectivity noise reduction tradeoff for multi-microphone systems |
US8583428B2 (en) | 2010-06-15 | 2013-11-12 | Microsoft Corporation | Sound source separation using spatial filtering and regularization phases |
US8600454B2 (en) | 2010-09-02 | 2013-12-03 | Apple Inc. | Decisions on ambient noise suppression in a mobile communications handset device |
US8837747B2 (en) * | 2010-09-28 | 2014-09-16 | Kabushiki Kaisha Toshiba | Apparatus, method, and program product for presenting moving image with sound |
WO2012078670A1 (en) | 2010-12-06 | 2012-06-14 | The Board Of Regents Of The University Of Texas System | Method and system for enhancing the intelligibility of sounds relative to background noise |
Cited By (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US10848872B2 (en) | 2014-12-27 | 2020-11-24 | Intel Corporation | Binaural recording for processing audio signals to enable alerts |
US11095985B2 (en) | 2014-12-27 | 2021-08-17 | Intel Corporation | Binaural recording for processing audio signals to enable alerts |
CN109104683A (en) * | 2018-07-13 | 2018-12-28 | 深圳市小瑞科技股份有限公司 | A kind of method and correction system of dual microphone phase measurement correction |
CN109104683B (en) * | 2018-07-13 | 2021-02-02 | 深圳市小瑞科技股份有限公司 | Method and system for correcting phase measurement of double microphones |
Also Published As
Publication number | Publication date |
---|---|
US20150172814A1 (en) | 2015-06-18 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US9271077B2 (en) | Method and system for directional enhancement of sound using small microphone arrays | |
US11294619B2 (en) | Earphone software and hardware | |
US9270244B2 (en) | System and method to detect close voice sources and automatically enhance situation awareness | |
US10219083B2 (en) | Method of localizing a sound source, a hearing device, and a hearing system | |
US9992587B2 (en) | Binaural hearing system configured to localize a sound source | |
US11605395B2 (en) | Method and device for spectral expansion of an audio signal | |
US11032640B2 (en) | Method and system to determine a sound source direction using small microphone arrays | |
US11893997B2 (en) | Audio signal processing for automatic transcription using ear-wearable device | |
CN114727212B (en) | Audio processing method and electronic equipment | |
US11741985B2 (en) | Method and device for spectral expansion for an audio signal | |
CN116324969A (en) | Hearing enhancement and wearable system with positioning feedback | |
CN113228710B (en) | Sound source separation in a hearing device and related methods | |
US11163522B2 (en) | Fine grain haptic wearable device | |
Amin et al. | Impact of microphone orientation and distance on BSS quality within interaction devices |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
AS | Assignment |
Owner name: PERSONICS HOLDINGS, LLC, FLORIDA Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:PERSONICS HOLDINGS, INC.;REEL/FRAME:032189/0304 Effective date: 20131231 |
|
AS | Assignment |
Owner name: DM STATON FAMILY LIMITED PARTNERSHIP (AS ASSIGNEE OF MARIA B. STATON), FLORIDA Free format text: SECURITY INTEREST;ASSIGNOR:PERSONICS HOLDINGS, LLC;REEL/FRAME:034170/0771 Effective date: 20131231 Owner name: DM STATON FAMILY LIMITED PARTNERSHIP (AS ASSIGNEE OF MARIA B. STATON), FLORIDA Free format text: SECURITY INTEREST;ASSIGNOR:PERSONICS HOLDINGS, LLC;REEL/FRAME:034170/0933 Effective date: 20141017 Owner name: DM STATON FAMILY LIMITED PARTNERSHIP (AS ASSIGNEE Free format text: SECURITY INTEREST;ASSIGNOR:PERSONICS HOLDINGS, LLC;REEL/FRAME:034170/0933 Effective date: 20141017 Owner name: DM STATON FAMILY LIMITED PARTNERSHIP (AS ASSIGNEE Free format text: SECURITY INTEREST;ASSIGNOR:PERSONICS HOLDINGS, LLC;REEL/FRAME:034170/0771 Effective date: 20131231 |
|
AS | Assignment |
Owner name: PERSONICS HOLDINGS, INC, FLORIDA Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:USHER, JOHN;GOLDSTEIN, STEVE;REEL/FRAME:037435/0548 Effective date: 20151221 |
|
STCF | Information on status: patent grant |
Free format text: PATENTED CASE |
|
AS | Assignment |
Owner name: DM STATION FAMILY LIMITED PARTNERSHIP, ASSIGNEE OF STATON FAMILY INVESTMENTS, LTD., FLORIDA Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:PERSONICS HOLDINGS, INC.;PERSONICS HOLDINGS, LLC;REEL/FRAME:042992/0493 Effective date: 20170620 Owner name: STATON TECHIYA, LLC, FLORIDA Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:DM STATION FAMILY LIMITED PARTNERSHIP, ASSIGNEE OF STATON FAMILY INVESTMENTS, LTD.;REEL/FRAME:042992/0524 Effective date: 20170621 Owner name: DM STATION FAMILY LIMITED PARTNERSHIP, ASSIGNEE OF Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:PERSONICS HOLDINGS, INC.;PERSONICS HOLDINGS, LLC;REEL/FRAME:042992/0493 Effective date: 20170620 |
|
AS | Assignment |
Owner name: DM STATON FAMILY LIMITED PARTNERSHIP, ASSIGNEE OF STATON FAMILY INVESTMENTS, LTD., FLORIDA Free format text: CORRECTIVE ASSIGNMENT TO CORRECT THE ASSIGNEE'S NAME PREVIOUSLY RECORDED AT REEL: 042992 FRAME: 0493. ASSIGNOR(S) HEREBY CONFIRMS THE ASSIGNMENT;ASSIGNORS:PERSONICS HOLDINGS, INC.;PERSONICS HOLDINGS, LLC;REEL/FRAME:043392/0961 Effective date: 20170620 Owner name: STATON TECHIYA, LLC, FLORIDA Free format text: CORRECTIVE ASSIGNMENT TO CORRECT THE ASSIGNOR'S NAME PREVIOUSLY RECORDED ON REEL 042992 FRAME 0524. ASSIGNOR(S) HEREBY CONFIRMS THE ASSIGNMENT OF THE ENTIRE INTEREST AND GOOD WILL;ASSIGNOR:DM STATON FAMILY LIMITED PARTNERSHIP, ASSIGNEE OF STATON FAMILY INVESTMENTS, LTD.;REEL/FRAME:043393/0001 Effective date: 20170621 Owner name: DM STATON FAMILY LIMITED PARTNERSHIP, ASSIGNEE OF Free format text: CORRECTIVE ASSIGNMENT TO CORRECT THE ASSIGNEE'S NAME PREVIOUSLY RECORDED AT REEL: 042992 FRAME: 0493. ASSIGNOR(S) HEREBY CONFIRMS THE ASSIGNMENT;ASSIGNORS:PERSONICS HOLDINGS, INC.;PERSONICS HOLDINGS, LLC;REEL/FRAME:043392/0961 Effective date: 20170620 |
|
MAFP | Maintenance fee payment |
Free format text: PAYMENT OF MAINTENANCE FEE, 4TH YR, SMALL ENTITY (ORIGINAL EVENT CODE: M2551); ENTITY STATUS OF PATENT OWNER: SMALL ENTITY Year of fee payment: 4 |
|
MAFP | Maintenance fee payment |
Free format text: PAYMENT OF MAINTENANCE FEE, 8TH YR, SMALL ENTITY (ORIGINAL EVENT CODE: M2552); ENTITY STATUS OF PATENT OWNER: SMALL ENTITY Year of fee payment: 8 |
|
AS | Assignment |
Owner name: ST PORTFOLIO HOLDINGS, LLC, FLORIDA Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:STATON TECHIYA, LLC;REEL/FRAME:067806/0722 Effective date: 20240612 Owner name: ST R&DTECH, LLC, FLORIDA Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:ST PORTFOLIO HOLDINGS, LLC;REEL/FRAME:067806/0751 Effective date: 20240612 |