US8751224B2 - Combined microphone and earphone audio headset having means for denoising a near speech signal, in particular for a “hands-free” telephony system - Google Patents

Combined microphone and earphone audio headset having means for denoising a near speech signal, in particular for a “hands-free” telephony system Download PDF

Info

Publication number
US8751224B2
US8751224B2 US13/450,361 US201213450361A US8751224B2 US 8751224 B2 US8751224 B2 US 8751224B2 US 201213450361 A US201213450361 A US 201213450361A US 8751224 B2 US8751224 B2 US 8751224B2
Authority
US
United States
Prior art keywords
signal
speech
headset
physiological sensor
microphone
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Expired - Fee Related, expires
Application number
US13/450,361
Other languages
English (en)
Other versions
US20120278070A1 (en
Inventor
Michael Herve
Guillaume Vitte
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Parrot Drones SAS
Original Assignee
Parrot SA
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Parrot SA filed Critical Parrot SA
Assigned to PARROT reassignment PARROT ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: VITTE, GUILLAUME, HERVE, MICHAEL
Publication of US20120278070A1 publication Critical patent/US20120278070A1/en
Application granted granted Critical
Publication of US8751224B2 publication Critical patent/US8751224B2/en
Assigned to PARROT DRONES reassignment PARROT DRONES ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: PARROT
Expired - Fee Related legal-status Critical Current
Adjusted expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
    • G10L21/00Speech or voice signal processing techniques to produce another audible or non-audible signal, e.g. visual or tactile, in order to modify its quality or its intelligibility
    • G10L21/02Speech enhancement, e.g. noise reduction or echo cancellation
    • G10L21/0208Noise filtering
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
    • G10L21/00Speech or voice signal processing techniques to produce another audible or non-audible signal, e.g. visual or tactile, in order to modify its quality or its intelligibility
    • G10L21/02Speech enhancement, e.g. noise reduction or echo cancellation
    • G10L21/0316Speech enhancement, e.g. noise reduction or echo cancellation by changing the amplitude
    • G10L21/0364Speech enhancement, e.g. noise reduction or echo cancellation by changing the amplitude for improving intelligibility
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04RLOUDSPEAKERS, MICROPHONES, GRAMOPHONE PICK-UPS OR LIKE ACOUSTIC ELECTROMECHANICAL TRANSDUCERS; DEAF-AID SETS; PUBLIC ADDRESS SYSTEMS
    • H04R1/00Details of transducers, loudspeakers or microphones
    • H04R1/10Earpieces; Attachments therefor ; Earphones; Monophonic headphones
    • H04R1/1083Reduction of ambient noise
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04RLOUDSPEAKERS, MICROPHONES, GRAMOPHONE PICK-UPS OR LIKE ACOUSTIC ELECTROMECHANICAL TRANSDUCERS; DEAF-AID SETS; PUBLIC ADDRESS SYSTEMS
    • H04R3/00Circuits for transducers, loudspeakers or microphones
    • H04R3/005Circuits for transducers, loudspeakers or microphones for combining the signals of two or more microphones
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
    • G10L21/00Speech or voice signal processing techniques to produce another audible or non-audible signal, e.g. visual or tactile, in order to modify its quality or its intelligibility
    • G10L21/02Speech enhancement, e.g. noise reduction or echo cancellation
    • G10L21/0208Noise filtering
    • G10L2021/02085Periodic noise
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
    • G10L21/00Speech or voice signal processing techniques to produce another audible or non-audible signal, e.g. visual or tactile, in order to modify its quality or its intelligibility
    • G10L21/02Speech enhancement, e.g. noise reduction or echo cancellation
    • G10L21/0208Noise filtering
    • G10L21/0216Noise filtering characterised by the method used for estimating noise
    • G10L2021/02161Number of inputs available containing the signal or the noise to be suppressed
    • G10L2021/02165Two microphones, one receiving mainly the noise signal and the other one mainly the speech signal
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04RLOUDSPEAKERS, MICROPHONES, GRAMOPHONE PICK-UPS OR LIKE ACOUSTIC ELECTROMECHANICAL TRANSDUCERS; DEAF-AID SETS; PUBLIC ADDRESS SYSTEMS
    • H04R2460/00Details of hearing devices, i.e. of ear- or headphones covered by H04R1/10 or H04R5/033 but not provided for in any of their subgroups, or of hearing aids covered by H04R25/00 but not provided for in any of its subgroups
    • H04R2460/13Hearing devices using bone conduction transducers

Definitions

  • the invention relates to an audio headset of the combined microphone and earphone type.
  • Such a headset may be used in particular for communications functions such as “hands-free” telephony functions, in addition to listening to an audio source (e.g. music) coming from equipment to which the headset is connected.
  • communications functions such as “hands-free” telephony functions
  • an audio source e.g. music
  • one of the difficulties is to ensure sufficient intelligibly of the signal picked up by the microphone, i.e. the signal representing the speech of the near speaker (the wearer of the headset).
  • the headset may be used in an environment that is noisy (subway, busy street, train, etc.), such that the microphone picks up not only speech from the wearer of the headset, but also interfering noises from the surroundings.
  • the wearer may be protected from these noises by the headset, particularly if it is of a kind comprising closed earpieces that isolate the ears from the outside, and even more so if the headset is provided with “active noise control”.
  • the remote listener i.e. the party at the other end of the communication channel
  • the microphone will suffer from the interfering noises picked up by the microphone, which noises are superposed on and interfere with the speech signal from the near speaker (the wearer of the headset).
  • the general problem of the invention is to provide noise reduction that is effective, enabling a voice signal to be delivered to the remote speaker that is indeed representative of the speech uttered by the near speaker, which signal has had removed therefrom the interference components from external noises present in the environment of the near speaker.
  • An important aspect of this problem is the need to play back a speech signal that is natural and intelligible, i.e. that is not distorted and that has a frequency range that is not cut down by the denoising processing.
  • One of the ideas on which the invention is based consists in picking up certain voice vibrations by means of a physiological sensor applied against the cheek or the temple of the wearer of the headset, so as to access new information relating to speech content. This information is then used for denoising and also for various auxiliary functions that are explained below, in particular for calculating a cutoff frequency of a dynamic filter.
  • the vibration propagates from the vocal cords to the pharynx and to the mouth-and-nose cavity, where it is modulated, amplified, and articulated.
  • the mouth, the soft palate, the pharynx, the sinuses, and the nasal cavity form a resonance box for the voiced sound, and since their walls are elastic, they vibrate in turn, and this vibration is transmitted by internal bone conduction and is perceptible from the cheek and from the temple.
  • the invention relies on the possibility of picking up such voice vibration that is free of noise by means of a physiological sensor applied directly against the cheek or the temple.
  • the signals as picked up in this way are not properly speaking “speech”, since speech is not made up solely of voiced sounds, given that it contains components that do not stem from the vocal cords: for example, frequency content is much richer with sounds coming from the throat and issuing from the mouth.
  • vocal cords for example, frequency content is much richer with sounds coming from the throat and issuing from the mouth.
  • internal bone conduction and passage through the skin has the effect of filtering out certain voice components.
  • the signal is indeed representative of voice content that is voiced, and can be used effectively for reducing noise and/or for various other functions.
  • the signal picked up by the physiological sensor is usable only for low frequencies.
  • the noises that are generally encountered in an everyday environment are concentrated for the most part at low frequencies, so there is a considerable advantage in terms of reducing noise in having available a physiological sensor that delivers a low-frequency signal that is naturally free of the interfering components resulting from noise (where this is not possible with a conventional microphone).
  • the invention proposes performing denoising of the near speech signal by using a combined microphone and earphone headset that comprises in conventional manner earpieces connected together by a headband and each having a transducer for sound reproduction of an audio signal housed in a shell that is provided with an ear-surrounding cushion, and at least one microphone suitable for picking up the speech of the wearer of the headset.
  • this combined microphone and earphone headset includes means for denoising a near speech signal uttered by the wearer of the headset, which means comprise: a physiological sensor incorporated in the ear-surrounding cushion and placed in a region thereof that is suitable for coming into contact with the cheek or the temple of the wearer of the headset in order to be coupled thereto and pick up non-acoustic voice vibration transmitted by internal bone conduction, the physiological sensor delivering a first speech signal; a microphone set, comprising the microphone(s) suitable for picking up the acoustic voice vibration that is transmitted through the air from the mouth of the wearer of the headset, this microphone set delivering a second speech signal; means for denoising the second speech signal; and mixer means for combining the first and second speech signals, and for outputting a third speech signal representative of the speech uttered by the wearer of the headset.
  • the combined microphone and earphone headset comprises: lowpass filter means for filtering the first speech signal before it is combined by the mixer means, and/or highpass filter means for filtering the second speech signal before it is denoised and combined by the mixer means.
  • the lowpass and/or highpass filter means comprise filters of adjustable cutoff frequency; and the headset includes cutoff frequency calculation means operating as a function of the signal delivered by the physiological sensor.
  • the cutoff frequency calculation means may in particular comprise means for analyzing the spectral content of the signal delivered by the physiological sensor, and suitable for determining the cutoff frequency as a function of the relative levels of the signal-to-noise ratios as evaluated in a plurality of distinct frequency bands of the signal delivered by the physiological sensor.
  • the means for denoising the second speech signal are non-frequency noise-reduction means that make use, in one particular embodiment of the invention, of the microphone set that has two microphones, and of a combiner suitable for applying a delay to the signal delivered by one of the microphones and for subtracting the delayed signal from the signal delivered by the other microphone.
  • the two microphones may be in alignment in a linear array having a main direction directed towards the mouth of the wearer of the headset.
  • means are provided for denoising the third speech signal as delivered by the mixer means, in particular frequency noise-reduction means.
  • means receiving as input the first and third speech signals and performing intercorrelation between them, and delivering as output a signal representative of the probability of speech being present as a function of the result of the intercorrelation.
  • the means for denoising the third speech signal receive as input this signal representative of the probability that speech is present, and they are suitable selectively for:
  • post-processing means suitable for performing equalization selectively in different frequency bands in the portion of the spectrum corresponding to the signal picked up by the physiological sensor. These means determine an equalization gain for each of the frequency bands, the gain being calculated on the basis of the respective frequency coefficients of the signals delivered by the microphone(s) and the signals delivered by the physiological sensor, as considered in the frequency domain.
  • FIG. 1 is a general view of a headset of the invention, placed on the head of a user.
  • FIG. 2 is an overall block diagram explaining how the signal processing is performed that enables a denoised signal to be output that is representative of the speech uttered by the wearer of the headset.
  • FIG. 3 is an amplitude/frequency spectrum diagram showing the intercorrelation calculation used for evaluating the probability of speech being present.
  • FIG. 4 is an amplitude/frequency spectrum diagram showing the final automatic equalization processing operated after noise reduction.
  • reference 10 is an overall reference for the headset of the invention, which comprises two earpieces 12 held together by a headband.
  • Each of the earpieces is preferably constituted by a closed shell 12 housing a sound reproduction transducer and pressed around the user's ear with an isolating cushion 16 interposed to isolate the ear from the outside.
  • the headset is provided with a physiological sensor 18 for picking up the vibration produced by a voiced signal uttered by the wearer of the headset, which vibration may be picked up via the cheek or the temple.
  • the sensor 18 is preferably an accelerometer incorporated in the cushion 16 so as to press against the user's cheek or temple with the closest possible coupling.
  • the physiological sensor may be placed on the inside face of the skin covering the cushion so that, once the headset is in position, the physiological sensor is pressed against the user's cheek or temple under the effect of a small amount of pressure that results from the material of the cushion being flattened, with only the skin of the cushion being interposed between the user and the sensor.
  • the headset also includes a microphone array or antenna, e.g. two omnidirectional microphones 20 and 22 placed on the shell of the earpiece 12 .
  • These two microphones comprise a front microphone 20 and a rear microphone 22 and they are omnidirectional microphones placed relative to each other in such a manner that they are in alignment along a direction 24 that is directed approximately towards the mouth 26 of the wearer of the headset.
  • FIG. 2 is a block diagram showing the various functional blocks used in the method of the invention, and how they interact.
  • the method of the invention is implemented by software means, that can be broken down and represented diagrammatically by various blocks 30 to 64 shown in FIG. 2 .
  • the processing is implemented in the form of appropriate algorithms executed by a microcontroller or a digital signal processor. Although for clarity of description these various processes are presented in the form of distinct blocks, they implement elements in common and in practice they correspond to a plurality of functions executed overall by the same software.
  • FIG. 2 shows the physiological sensor 18 and the front and rear omnidirectional microphones 20 and 22 .
  • Reference 28 designates the sound reproduction transducer placed inside the shell of the earpiece.
  • These various elements deliver signals that are subjected to processing by the block referenced 30 , which may be coupled to an interface 32 with communications circuits (telephone circuits) from which it receives as input E the sound that is to be reproduced by the transducer 28 (speech from the distant speaker during a telephone call, music source outside periods of telephone conversation), and to which it delivers on an output S a signal that is representative of the speech from the near speaker, i.e. the wearer of the headset.
  • communications circuits telephone circuits
  • the signal for reproduction that appears on the input E is a digital signal that is converted into an analog signal by a converter 34 , and then amplified by an amplifier 36 for reproduction by the transducer 28 .
  • the signal picked up by the physiological sensor 18 is a signal that mainly comprises components in the lower region of the sound spectrum (typically in the range 0 to 1500 hertz (Hz)). As explained above, this signal is naturally not noisy.
  • the signals picked up by the microphones 20 and 22 are used mainly for the higher portion of the spectrum (above 1500 Hz), but these signals are very noisy and it is essential to perform strong denoising processing in order to eliminate the interfering noise components, which components may in certain environments be at a level such as to completely hide the speech signal picked up by the microphones 20 and 22 .
  • the first step of the processing is anti-echo processing applied to the signals from the physiological sensor and from the microphones.
  • the sound reproduced by the transducer 28 is picked up by the physiological sensor 18 and by the microphones 20 and 22 , thereby generating an echo that disturbs the operation of the system, and that must therefore be eliminated upstream.
  • This anti-echo processing is implemented by blocks 38 , 40 , and 42 , each of these blocks having a first input receiving the signal delivered by a respective one of the sensor 18 , and the microphones 20 and 22 , and a second input receiving the signal reproduced by the transducer 28 (echo-generating signal), and it outputs a signal from which the echo has been eliminated for use in subsequent processing.
  • the anti-echo processing is performed by processing with an adaptive algorithm such as that described in FR 2 792 146 A1 (Parrot S A), to which reference may be made for more details.
  • It is an automatic echo canceling technique AEC consisting in dynamically defining a compensation filter that models the acoustic coupling between the transducer 28 and the physiological sensor 18 (or the microphone 20 or the microphone 22 , respectively) by a linear transformation between the signal reproduced by the transducer 28 (i.e. the signal E applied as input to the blocks 38 , 40 , and 42 ) and the echo picked up by the physiological sensor 18 (or the microphone 20 or 22 ).
  • This transformation defines an adaptive filter that is applied to the reproduced incident signal, and the result of this filtering is subtracted from the signal picked up by the physiological sensor 18 (or the microphone 20 or 22 ), thereby having the effect of canceling the major portion of the acoustic echo.
  • This modeling relies on searching for a correlation between the signal reproduced by the transducer 28 and the signal picked up by the physiological sensor 18 (or the microphone 20 or 22 ), i.e. an estimate of the impulse response of the coupling constituted by the body of the earpiece 12 supporting these various elements.
  • the processing is performed in particular by an adaptive algorithm of the affine projection algorithm (APA) type, that ensures rapid convergence, and that is well adapted to applications of the “hands-free type” in which voice delivery is intermittent and at a level that may vary rapidly.
  • APA affine projection algorithm
  • the iterative algorithm is executed at a variable sampling rate, as described in above-mentioned FR 2 792 146 A1.
  • the sampling interval ⁇ varies continuously as a function of the energy level of the signal picked up by the microphone, before and after filtering. This interval is increased when the energy of the signal as picked up is dominated by the energy of the echo, and conversely it is decreased when the energy of the signal that is picked up is dominated by the energy of the background noise and/or of the speech of the remote speaker.
  • the signal picked up by the physiological sensor 18 is used as an input signal to a block 44 for calculating a cutoff frequency FC.
  • the following step consists in performing signal filtering with a lowpass filter 48 for the signal from the physiological sensor 18 and with respective highpass filters 50 , 52 for the signals picked up by the microphones 20 and 22 .
  • These filters 48 , 50 , 52 are preferably digital filters of the incident impulse response (IIR) type, i.e. recursive filters, that present a relatively abrupt transition between the passband and the stop band.
  • IIR incident impulse response
  • these filters are adaptive filters with a cutoff frequency that is variable and determined dynamically by the block 44 .
  • the cutoff frequency FC which is preferably the same for the lowpass filter 48 and the highpass filters 50 and 52 , is determined from the signal from the physiological sensor 18 after the anti-echo processing 38 .
  • an algorithm calculates the signal-to-noise ratio over a plurality of frequency bands situated in a range extending for example from 0 to 2500 Hz (the level of noise being given by an energy calculation in a highest frequency band, e.g.
  • the cutoff frequency that is selected corresponds to the maximum frequency at which the signal-to-noise ratio exceeds a predetermined threshold, e.g. 10 decibels (dB).
  • the following step consists in using the block 54 to perform mixing so as to reconstruct the complete spectrum with both a low frequency region of the spectrum given by the filtered signal from the physiological sensor 18 and a high frequency portion of the spectrum given by the filtered signal from the microphones 20 and 22 after passing through a combiner-and-phaseshifter 56 that enables denoising to be performed in this portion of the spectrum.
  • This reconstruction is performed by summing the two signals that are applied synchronously to the mixer block 54 so as to avoid any deformation.
  • the signal that it is desired to denoise comes from the two microphones 20 and 22 that are placed a few centimeters apart from each other on the shell 14 of one of the earpieces of the headset. As mentioned above, these two microphones are arranged relative to each other in such a manner that the direction 24 they define points approximately towards the mouth 26 of the wearer of the headset.
  • the speech signal delivered by the mouth reaches the front microphone 20 and then reaches the rear microphone 22 with a delay and thus a phase shift that is substantially constant, whereas ambient noise is picked up by both microphones 20 and 22 without phase shifts (which microphones are omnidirectional microphones), given the remoteness of the sources of interfering noise from the two microphones 20 and 22 .
  • the noise in the signals picked up by the microphones 20 and 22 is not reduced in the frequency domain (as is often the case), but rather in the time domain, by means of the combiner-and-phaseshifter 56 that comprises a phaseshifter 58 that applies a delay ⁇ to the signal from the rear microphone 22 and a combiner 60 that enables the domain signal to be subtracted from the signal coming from the front microphone 20 .
  • This signal is subjected by a block 62 to frequency noise reduction.
  • This frequency noise reduction is preferably performed differently in the presence or in the absence of speech, by evaluating the probability p that speech is absent from the signals picked up by the physiological sensor 18 .
  • this possibility that speech is absent is derived from the information given by the physiological sensor.
  • the signal delivered by this sensor presents a very good signal-to-noise ratio up to the cutoff frequency FC as determined by the block 44 .
  • the cutoff frequency its signal-to-noise ratio still remains good, and is often better than that from the microphones 20 and 22 .
  • the information from the sensor is used by a block 64 that calculates the frequency intercorrelation between the combined signal delivered by the mixer block 54 and the non-filtered signal from the physiological sensor, prior to lowpass filtering 48 .
  • InterCorrelation ⁇ ( n , f ) ⁇ intercorr * InterCorrelation ⁇ ( n - 1 , f ) + ( 1 - ⁇ Intercorr ) * Smix ⁇ ( f ) ⁇ ⁇ Saac ⁇ ( f ) ⁇
  • Smix(f) and Saac(f) are (complex) vector representations of frequency for the frame n, respectively of the combined signal delivered by the mixer block 54 and for the signal from the physiological sensor 18 .
  • the algorithm searches for the frequencies for which there is only noise (the situation that applies when speech is absent): on the spectrum diagram of the signal delivered by the mixer block 54 certain harmonics are buried in noise, whereas they stand out more in the signal from the physiological sensor.
  • the peaks P 1 , P 2 , P 3 , P 4 , . . . in the intercorrelation calculation indicate strong correlation between the combined signal delivered by the mixer block 54 and the signal from the physiological sensor 18 , such that the emergence of such correlated frequencies indicates that speech is probably present for both frequencies.
  • normalization_coefficient enables the probability distribution to be adjusted as a function of the value of the intercorrelation, so as to obtain values in the range 0 to 1.
  • the probability p that speech is absent as obtained in this way is applied to the block 62 that acts on the signal delivered by the mixer block 54 to perform frequency noise reduction in selective manner relative to a given threshold for the probability that speech is absent:
  • the above-described system enables excellent overall performance to be obtained, typically with noise reduction of the order of 30 dB to 40 dB in the speech signal from the near speaker. Because all interfering noise is eliminated, in particular the most intrusive noise (train, subway, etc.), which is concentrated at low frequencies, gives the remote listener (i.e. the party with whom the wearer of the headset is in communication) the impression that the other party (the wearer of the headset) is in a silent room.
  • the low frequency content picked up from the cheek or the temple by the physiological sensor 18 is different from the low frequency content of the sound coming from the user's mouth, as it would be picked up by a microphone situated a few centimeters from the mouth, or even as it would be picked up by the ear of a listener.
  • Equalization may be performed automatically, from the signal delivered by the microphones 20 and 22 before filtering.
  • FIG. 4 shows an example in the frequency domain (but after a Fourier transform) of the signal ACC produced by the physiological sensor 18 compared with a microphone signal MIC as would be picked up a few centimeters from the mouth.
  • the algorithm calculates the respective Fourier transforms of those two signals, giving a series of frequency coefficients (expressed in dB) NormPhysioFreq_dB(i) and NormMicFreq_dB(i), corresponding respectively to the absolute value or “norm” of the i th Fourier coefficient of the signal from the physiological sensor and to the norm of the i th Fourier coefficient of the microphone signal.
  • Gain_dB( i ) ⁇ .Gain_dB( i ) ⁇ (1 ⁇ )DifferenceFreq_dB( i )

Landscapes

  • Engineering & Computer Science (AREA)
  • Acoustics & Sound (AREA)
  • Physics & Mathematics (AREA)
  • Signal Processing (AREA)
  • Health & Medical Sciences (AREA)
  • Audiology, Speech & Language Pathology (AREA)
  • Human Computer Interaction (AREA)
  • Quality & Reliability (AREA)
  • Computational Linguistics (AREA)
  • Multimedia (AREA)
  • General Health & Medical Sciences (AREA)
  • Otolaryngology (AREA)
  • Circuit For Audible Band Transducer (AREA)
  • Soundproofing, Sound Blocking, And Sound Damping (AREA)
  • Details Of Audible-Bandwidth Transducers (AREA)
  • Telephone Set Structure (AREA)
US13/450,361 2011-04-26 2012-04-18 Combined microphone and earphone audio headset having means for denoising a near speech signal, in particular for a “hands-free” telephony system Expired - Fee Related US8751224B2 (en)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
FR1153572 2011-04-26
FR1153572A FR2974655B1 (fr) 2011-04-26 2011-04-26 Combine audio micro/casque comprenant des moyens de debruitage d'un signal de parole proche, notamment pour un systeme de telephonie "mains libres".

Publications (2)

Publication Number Publication Date
US20120278070A1 US20120278070A1 (en) 2012-11-01
US8751224B2 true US8751224B2 (en) 2014-06-10

Family

ID=45939241

Family Applications (1)

Application Number Title Priority Date Filing Date
US13/450,361 Expired - Fee Related US8751224B2 (en) 2011-04-26 2012-04-18 Combined microphone and earphone audio headset having means for denoising a near speech signal, in particular for a “hands-free” telephony system

Country Status (5)

Country Link
US (1) US8751224B2 (fr)
EP (1) EP2518724B1 (fr)
JP (1) JP6017825B2 (fr)
CN (1) CN102761643B (fr)
FR (1) FR2974655B1 (fr)

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20230326474A1 (en) * 2022-04-06 2023-10-12 Analog Devices International Unlimited Company Audio signal processing method and system for noise mitigation of a voice signal measured by a bone conduction sensor, a feedback sensor and a feedforward sensor
US11902759B2 (en) 2019-09-12 2024-02-13 Shenzhen Shokz Co., Ltd. Systems and methods for audio signal generation

Families Citing this family (59)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US9247346B2 (en) 2007-12-07 2016-01-26 Northern Illinois Research Foundation Apparatus, system and method for noise cancellation and communication for incubators and related devices
US9135915B1 (en) * 2012-07-26 2015-09-15 Google Inc. Augmenting speech segmentation and recognition using head-mounted vibration and/or motion sensors
US9704486B2 (en) * 2012-12-11 2017-07-11 Amazon Technologies, Inc. Speech recognition power management
CN103208291A (zh) * 2013-03-08 2013-07-17 华南理工大学 一种可用于强噪声环境的语音增强方法及装置
US9560444B2 (en) * 2013-03-13 2017-01-31 Cisco Technology, Inc. Kinetic event detection in microphones
JP6123503B2 (ja) * 2013-06-07 2017-05-10 富士通株式会社 音声補正装置、音声補正プログラム、および、音声補正方法
CN109327789B (zh) 2013-06-28 2021-07-13 哈曼国际工业有限公司 一种增强声音的再现的方法和系统
DE102013216133A1 (de) * 2013-08-14 2015-02-19 Sennheiser Electronic Gmbh & Co. Kg Hörer oder Headset
US9180055B2 (en) * 2013-10-25 2015-11-10 Harman International Industries, Incorporated Electronic hearing protector with quadrant sound localization
US20150118960A1 (en) * 2013-10-28 2015-04-30 Aliphcom Wearable communication device
US9036844B1 (en) 2013-11-10 2015-05-19 Avraham Suhami Hearing devices based on the plasticity of the brain
EP2882203A1 (fr) 2013-12-06 2015-06-10 Oticon A/s Dispositif d'aide auditive pour communication mains libres
FR3019422B1 (fr) * 2014-03-25 2017-07-21 Elno Appareil acoustique comprenant au moins un microphone electroacoustique, un microphone osteophonique et des moyens de calcul d'un signal corrige, et equipement de tete associe
FR3021180B1 (fr) * 2014-05-16 2016-06-03 Parrot Casque audio a controle actif de bruit anc avec prevention des effets d'une saturation du signal microphonique "feedback"
US10219067B2 (en) 2014-08-29 2019-02-26 Harman International Industries, Incorporated Auto-calibrating noise canceling headphone
US9942848B2 (en) * 2014-12-05 2018-04-10 Silicon Laboratories Inc. Bi-directional communications in a wearable monitor
CN104486286B (zh) * 2015-01-19 2018-01-05 武汉邮电科学研究院 一种连续子载波ofdma系统的上行帧同步方法
US9905216B2 (en) * 2015-03-13 2018-02-27 Bose Corporation Voice sensing using multiple microphones
US9847093B2 (en) * 2015-06-19 2017-12-19 Samsung Electronics Co., Ltd. Method and apparatus for processing speech signal
US20160379661A1 (en) * 2015-06-26 2016-12-29 Intel IP Corporation Noise reduction for electronic devices
US9633672B1 (en) * 2015-10-29 2017-04-25 Blackberry Limited Method and device for suppressing ambient noise in a speech signal generated at a microphone of the device
FR3044197A1 (fr) 2015-11-19 2017-05-26 Parrot Casque audio a controle actif de bruit, controle anti-occlusion et annulation de l'attenuation passive, en fonction de la presence ou de l'absence d'une activite vocale de l'utilisateur de casque.
GB2552178A (en) * 2016-07-12 2018-01-17 Samsung Electronics Co Ltd Noise suppressor
CN106211012B (zh) * 2016-07-15 2019-11-29 成都定为电子技术有限公司 一种耳机时频响应的测量与校正系统及其方法
JP6634354B2 (ja) * 2016-07-20 2020-01-22 ホシデン株式会社 緊急通報システム用ハンズフリー通話装置
CN110035808A (zh) * 2016-09-14 2019-07-19 声感股份有限公司 具有同步的多设备音频流传输系统
WO2018083511A1 (fr) * 2016-11-03 2018-05-11 北京金锐德路科技有限公司 Appareil et procédé de lecture audio
US10311889B2 (en) * 2017-03-20 2019-06-04 Bose Corporation Audio signal processing for noise reduction
WO2018199846A1 (fr) * 2017-04-23 2018-11-01 Audio Zoom Pte Ltd Appareil transducteur destiné à une intelligibilité élevée de la parole dans des environnements bruyants
US10341759B2 (en) * 2017-05-26 2019-07-02 Apple Inc. System and method of wind and noise reduction for a headphone
CN107180627B (zh) * 2017-06-22 2020-10-09 潍坊歌尔微电子有限公司 去除噪声的方法和装置
US10706868B2 (en) 2017-09-06 2020-07-07 Realwear, Inc. Multi-mode noise cancellation for voice detection
US10701470B2 (en) 2017-09-07 2020-06-30 Light Speed Aviation, Inc. Circumaural headset or headphones with adjustable biometric sensor
US10764668B2 (en) 2017-09-07 2020-09-01 Lightspeed Aviation, Inc. Sensor mount and circumaural headset or headphones with adjustable sensor
CN109729463A (zh) * 2017-10-27 2019-05-07 北京金锐德路科技有限公司 用于脖戴式语音交互耳机的声麦骨麦复合收音装置
JP7194912B2 (ja) * 2017-10-30 2022-12-23 パナソニックIpマネジメント株式会社 ヘッドセット
CN107886967B (zh) * 2017-11-18 2018-11-13 中国人民解放军陆军工程大学 一种深度双向门递归神经网络的骨导语音增强方法
US10438605B1 (en) * 2018-03-19 2019-10-08 Bose Corporation Echo control in binaural adaptive noise cancellation systems in headsets
CN110931027A (zh) * 2018-09-18 2020-03-27 北京三星通信技术研究有限公司 音频处理方法、装置、电子设备及计算机可读存储介质
CN109413539A (zh) * 2018-12-25 2019-03-01 珠海蓝宝石声学设备有限公司 一种耳机及其调节装置
EP3737115A1 (fr) * 2019-05-06 2020-11-11 GN Hearing A/S Appareil auditif avec capteur de conduction osseuse
CN110265056B (zh) * 2019-06-11 2021-09-17 安克创新科技股份有限公司 音源的控制方法以及扬声设备、装置
CN110121129B (zh) * 2019-06-20 2021-04-20 歌尔股份有限公司 耳机的麦克风阵列降噪方法、装置、耳机及tws耳机
JP2022505997A (ja) * 2019-10-09 2022-01-17 大象声科(深セン)科技有限公司 骨振動センサーとマイクの信号を融合するディープラーニング音声抽出及びノイズ低減方法
TWI735986B (zh) * 2019-10-24 2021-08-11 瑞昱半導體股份有限公司 收音裝置及方法
CN113038318B (zh) * 2019-12-25 2022-06-07 荣耀终端有限公司 一种语音信号处理方法及装置
TWI745845B (zh) * 2020-01-31 2021-11-11 美律實業股份有限公司 耳機及耳機組
KR20220017080A (ko) * 2020-08-04 2022-02-11 삼성전자주식회사 음성 신호를 처리하는 방법 및 이를 이용한 장치
CN111935573B (zh) * 2020-08-11 2022-06-14 Oppo广东移动通信有限公司 音频增强方法、装置、存储介质及可穿戴设备
CN114339569B (zh) * 2020-08-29 2023-05-26 深圳市韶音科技有限公司 一种获取振动传递函数的方法和系统
WO2022060891A1 (fr) * 2020-09-15 2022-03-24 Dolby Laboratories Licensing Corporation Procédé et dispositif pour le traitement d'un enregistrement binaural
US11259119B1 (en) 2020-10-06 2022-02-22 Qualcomm Incorporated Active self-voice naturalization using a bone conduction sensor
US11337000B1 (en) * 2020-10-23 2022-05-17 Knowles Electronics, Llc Wearable audio device having improved output
JP7467317B2 (ja) * 2020-11-12 2024-04-15 株式会社東芝 音響検査装置及び音響検査方法
US11943601B2 (en) 2021-08-13 2024-03-26 Meta Platforms Technologies, Llc Audio beam steering, tracking and audio effects for AR/VR applications
US20230050954A1 (en) * 2021-08-13 2023-02-16 Meta Platforms Technologies, Llc Contact and acoustic microphones for voice wake and voice processing for ar/vr applications
US20230253002A1 (en) * 2022-02-08 2023-08-10 Analog Devices International Unlimited Company Audio signal processing method and system for noise mitigation of a voice signal measured by air and bone conduction sensors
CN114724574A (zh) * 2022-02-21 2022-07-08 大连理工大学 一种期望声源方向可调的双麦克风降噪方法
CN114333883B (zh) * 2022-03-12 2022-05-31 广州思正电子股份有限公司 一种头戴式智能语音识别装置

Citations (10)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
EP0683621A2 (fr) 1994-05-18 1995-11-22 Nippon Telegraph And Telephone Corporation Emetteur-récepteur ayant un transducteur acoustique du type embout auriculaire
JPH08214391A (ja) 1995-02-03 1996-08-20 Iwatsu Electric Co Ltd 骨導気導複合型イヤーマイクロホン装置
WO2000021194A1 (fr) 1998-10-08 2000-04-13 Resound Corporation Systeme de transmission de la voix a capteurs jumeles
JP2000261534A (ja) 1999-03-10 2000-09-22 Nippon Telegr & Teleph Corp <Ntt> 送受話器
US7383181B2 (en) * 2003-07-29 2008-06-03 Microsoft Corporation Multi-sensory speech detection system
US7930178B2 (en) * 2005-12-23 2011-04-19 Microsoft Corporation Speech modeling and enhancement based on magnitude-normalized spectra
US20110096939A1 (en) * 2009-10-28 2011-04-28 Sony Corporation Reproducing device, headphone and reproducing method
US20110135106A1 (en) * 2008-05-22 2011-06-09 Uri Yehuday Method and a system for processing signals
US20120310637A1 (en) * 2011-06-01 2012-12-06 Parrot Audio equipment including means for de-noising a speech signal by fractional delay filtering, in particular for a "hands-free" telephony system
US20130051585A1 (en) * 2011-08-30 2013-02-28 Nokia Corporation Apparatus and Method for Audio Delivery With Different Sound Conduction Transducers

Family Cites Families (12)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JPS5394918A (en) * 1977-01-28 1978-08-19 Masahisa Ikegami Combtned mtcrophone
JPH08223677A (ja) * 1995-02-15 1996-08-30 Nippon Telegr & Teleph Corp <Ntt> 送話器
JPH11265199A (ja) * 1998-03-18 1999-09-28 Nippon Telegr & Teleph Corp <Ntt> 送話器
FR2792146B1 (fr) 1999-04-07 2001-05-25 Parrot Sa Procede de suppression de l'echo acoustique d'un signal audio, notamment dans le signal capte par un microphone
JP2002125298A (ja) * 2000-10-13 2002-04-26 Yamaha Corp マイク装置およびイヤホンマイク装置
JP2003264883A (ja) * 2002-03-08 2003-09-19 Denso Corp 音声処理装置および音声処理方法
JP4348706B2 (ja) * 2002-10-08 2009-10-21 日本電気株式会社 アレイ装置および携帯端末
CN1701528A (zh) * 2003-07-17 2005-11-23 松下电器产业株式会社 通话装置
US7492889B2 (en) * 2004-04-23 2009-02-17 Acoustic Technologies, Inc. Noise suppression based on bark band wiener filtering and modified doblinger noise estimate
US7813923B2 (en) * 2005-10-14 2010-10-12 Microsoft Corporation Calibration based beamforming, non-linear adaptive filtering, and multi-sensor headset
FR2898209B1 (fr) 2006-03-01 2008-12-12 Parrot Sa Procede de debruitage d'un signal audio
JP2007264132A (ja) * 2006-03-27 2007-10-11 Toshiba Corp 音声検出装置及びその方法

Patent Citations (10)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
EP0683621A2 (fr) 1994-05-18 1995-11-22 Nippon Telegraph And Telephone Corporation Emetteur-récepteur ayant un transducteur acoustique du type embout auriculaire
JPH08214391A (ja) 1995-02-03 1996-08-20 Iwatsu Electric Co Ltd 骨導気導複合型イヤーマイクロホン装置
WO2000021194A1 (fr) 1998-10-08 2000-04-13 Resound Corporation Systeme de transmission de la voix a capteurs jumeles
JP2000261534A (ja) 1999-03-10 2000-09-22 Nippon Telegr & Teleph Corp <Ntt> 送受話器
US7383181B2 (en) * 2003-07-29 2008-06-03 Microsoft Corporation Multi-sensory speech detection system
US7930178B2 (en) * 2005-12-23 2011-04-19 Microsoft Corporation Speech modeling and enhancement based on magnitude-normalized spectra
US20110135106A1 (en) * 2008-05-22 2011-06-09 Uri Yehuday Method and a system for processing signals
US20110096939A1 (en) * 2009-10-28 2011-04-28 Sony Corporation Reproducing device, headphone and reproducing method
US20120310637A1 (en) * 2011-06-01 2012-12-06 Parrot Audio equipment including means for de-noising a speech signal by fractional delay filtering, in particular for a "hands-free" telephony system
US20130051585A1 (en) * 2011-08-30 2013-02-28 Nokia Corporation Apparatus and Method for Audio Delivery With Different Sound Conduction Transducers

Cited By (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US11902759B2 (en) 2019-09-12 2024-02-13 Shenzhen Shokz Co., Ltd. Systems and methods for audio signal generation
US20230326474A1 (en) * 2022-04-06 2023-10-12 Analog Devices International Unlimited Company Audio signal processing method and system for noise mitigation of a voice signal measured by a bone conduction sensor, a feedback sensor and a feedforward sensor
US11978468B2 (en) * 2022-04-06 2024-05-07 Analog Devices International Unlimited Company Audio signal processing method and system for noise mitigation of a voice signal measured by a bone conduction sensor, a feedback sensor and a feedforward sensor

Also Published As

Publication number Publication date
FR2974655A1 (fr) 2012-11-02
CN102761643A (zh) 2012-10-31
JP6017825B2 (ja) 2016-11-02
EP2518724A1 (fr) 2012-10-31
JP2012231468A (ja) 2012-11-22
FR2974655B1 (fr) 2013-12-20
CN102761643B (zh) 2017-04-12
EP2518724B1 (fr) 2013-10-02
US20120278070A1 (en) 2012-11-01

Similar Documents

Publication Publication Date Title
US8751224B2 (en) Combined microphone and earphone audio headset having means for denoising a near speech signal, in particular for a “hands-free” telephony system
US11657793B2 (en) Voice sensing using multiple microphones
JP6069830B2 (ja) 耳孔装着型収音装置、信号処理装置、収音方法
JP6069829B2 (ja) 耳孔装着型収音装置、信号処理装置、収音方法
EP3282678B1 (fr) Processeur de signaux à réduction de bruit dans l&#39;effet local pour un casque d&#39;écoute
JP2012253771A (ja) 特に「ハンズフリー」電話システム用の、小数遅延フィルタリングにより音声信号のノイズ除去を行うための手段を含むオーディオ装置
US9654855B2 (en) Self-voice occlusion mitigation in headsets
CN111683319A (zh) 一种通话拾音降噪方法及耳机、存储介质
US9620142B2 (en) Self-voice feedback in communications headsets
JP6315046B2 (ja) 耳孔装着型収音装置、信号処理装置、収音方法
JP6197930B2 (ja) 耳孔装着型収音装置、信号処理装置、収音方法
US20230328462A1 (en) Method, device, headphones and computer program for actively suppressing the occlusion effect during the playback of audio signals
CN116962934B (zh) 一种拾音降噪方法及系统
JP5249431B2 (ja) 信号経路を分離する方法及び電気喉頭を使用して音声を改良するための使用方法
CN113038315A (zh) 一种语音信号处理方法及装置
CN109862470A (zh) 对耳病患者播音的方法、耳机及其计算机可读存储介质

Legal Events

Date Code Title Description
AS Assignment

Owner name: PARROT, FRANCE

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:HERVE, MICHAEL;VITTE, GUILLAUME;SIGNING DATES FROM 20120531 TO 20120601;REEL/FRAME:028338/0600

AS Assignment

Owner name: PARROT DRONES, FRANCE

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:PARROT;REEL/FRAME:039323/0421

Effective date: 20160329

FEPP Fee payment procedure

Free format text: MAINTENANCE FEE REMINDER MAILED (ORIGINAL EVENT CODE: REM.)

LAPS Lapse for failure to pay maintenance fees

Free format text: PATENT EXPIRED FOR FAILURE TO PAY MAINTENANCE FEES (ORIGINAL EVENT CODE: EXP.)

STCH Information on status: patent discontinuation

Free format text: PATENT EXPIRED DUE TO NONPAYMENT OF MAINTENANCE FEES UNDER 37 CFR 1.362

FP Lapsed due to failure to pay maintenance fee

Effective date: 20180610