US20120010881A1 - Monaural Noise Suppression Based on Computational Auditory Scene Analysis - Google Patents
Monaural Noise Suppression Based on Computational Auditory Scene Analysis Download PDFInfo
- Publication number
- US20120010881A1 US20120010881A1 US12/860,043 US86004310A US2012010881A1 US 20120010881 A1 US20120010881 A1 US 20120010881A1 US 86004310 A US86004310 A US 86004310A US 2012010881 A1 US2012010881 A1 US 2012010881A1
- Authority
- US
- United States
- Prior art keywords
- sub
- noise
- speech
- band
- pitch
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Granted
Links
- 230000001629 suppression Effects 0.000 title abstract description 16
- 230000009467 reduction Effects 0.000 claims abstract description 31
- 238000000034 method Methods 0.000 claims description 40
- 210000003477 cochlea Anatomy 0.000 claims description 17
- 239000003607 modifier Substances 0.000 claims description 15
- 230000005236 sound signal Effects 0.000 claims description 6
- 238000009472 formulation Methods 0.000 claims description 3
- 239000000203 mixture Substances 0.000 claims description 3
- 230000009466 transformation Effects 0.000 claims description 3
- 230000001131 transforming effect Effects 0.000 claims 2
- 238000005516 engineering process Methods 0.000 abstract description 9
- 239000011295 pitch Substances 0.000 description 107
- 230000003595 spectral effect Effects 0.000 description 24
- 238000012545 processing Methods 0.000 description 22
- 230000004048 modification Effects 0.000 description 21
- 238000012986 modification Methods 0.000 description 21
- 230000001052 transient effect Effects 0.000 description 18
- 238000001228 spectrum Methods 0.000 description 12
- 238000010586 diagram Methods 0.000 description 10
- 230000008569 process Effects 0.000 description 9
- 238000001514 detection method Methods 0.000 description 6
- 238000000605 extraction Methods 0.000 description 6
- 230000007704 transition Effects 0.000 description 6
- 230000001934 delay Effects 0.000 description 5
- 230000003111 delayed effect Effects 0.000 description 5
- 238000001914 filtration Methods 0.000 description 5
- 238000004891 communication Methods 0.000 description 4
- 230000000694 effects Effects 0.000 description 4
- 230000004044 response Effects 0.000 description 4
- 239000000872 buffer Substances 0.000 description 3
- 230000007246 mechanism Effects 0.000 description 3
- 238000013459 approach Methods 0.000 description 2
- 230000008859 change Effects 0.000 description 2
- 238000006243 chemical reaction Methods 0.000 description 2
- 238000007906 compression Methods 0.000 description 2
- 230000006835 compression Effects 0.000 description 2
- 230000003750 conditioning effect Effects 0.000 description 2
- 238000010276 construction Methods 0.000 description 2
- 230000006870 function Effects 0.000 description 2
- 238000012805 post-processing Methods 0.000 description 2
- 230000003044 adaptive effect Effects 0.000 description 1
- 230000002411 adverse Effects 0.000 description 1
- 201000007201 aphasia Diseases 0.000 description 1
- 230000006399 behavior Effects 0.000 description 1
- 230000008901 benefit Effects 0.000 description 1
- 230000015572 biosynthetic process Effects 0.000 description 1
- 230000001413 cellular effect Effects 0.000 description 1
- 230000001143 conditioned effect Effects 0.000 description 1
- 230000021615 conjugation Effects 0.000 description 1
- 238000000354 decomposition reaction Methods 0.000 description 1
- 230000001419 dependent effect Effects 0.000 description 1
- 238000012938 design process Methods 0.000 description 1
- 238000002592 echocardiography Methods 0.000 description 1
- -1 energy level Substances 0.000 description 1
- 238000007667 floating Methods 0.000 description 1
- 239000011159 matrix material Substances 0.000 description 1
- 238000005457 optimization Methods 0.000 description 1
- 230000010363 phase shift Effects 0.000 description 1
- 238000007781 pre-processing Methods 0.000 description 1
- 230000008439 repair process Effects 0.000 description 1
- 230000000717 retained effect Effects 0.000 description 1
- 238000005070 sampling Methods 0.000 description 1
- 238000003786 synthesis reaction Methods 0.000 description 1
- 238000012546 transfer Methods 0.000 description 1
Images
Classifications
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS OR SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING; SPEECH OR AUDIO CODING OR DECODING
- G10L21/00—Processing of the speech or voice signal to produce another audible or non-audible signal, e.g. visual or tactile, in order to modify its quality or its intelligibility
- G10L21/02—Speech enhancement, e.g. noise reduction or echo cancellation
- G10L21/0208—Noise filtering
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS OR SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING; SPEECH OR AUDIO CODING OR DECODING
- G10L21/00—Processing of the speech or voice signal to produce another audible or non-audible signal, e.g. visual or tactile, in order to modify its quality or its intelligibility
- G10L21/02—Speech enhancement, e.g. noise reduction or echo cancellation
- G10L21/0272—Voice signal separating
Landscapes
- Engineering & Computer Science (AREA)
- Human Computer Interaction (AREA)
- Quality & Reliability (AREA)
- Signal Processing (AREA)
- Health & Medical Sciences (AREA)
- Audiology, Speech & Language Pathology (AREA)
- Computational Linguistics (AREA)
- Physics & Mathematics (AREA)
- Acoustics & Sound (AREA)
- Multimedia (AREA)
- Circuit For Audible Band Transducer (AREA)
- Noise Elimination (AREA)
- Soundproofing, Sound Blocking, And Sound Damping (AREA)
Abstract
Description
- This application claims the priority benefit of U.S. Provisional Application Ser. No. 61/363,638, titled “Single Channel Noise Reduction,” filed Jul. 12, 2010, the disclosure of which is incorporated herein by reference.
- 1. Field of the Invention
- The present invention relates generally to audio processing, and more particularly to processing an audio signal to suppress noise.
- 2. Description of Related Art
- Currently, there are many methods for reducing background noise in an adverse audio environment. A stationary noise suppression system suppresses stationary noise, by either a fixed or varying number of dB. A fixed suppression system suppresses stationary or non-stationary noise by a fixed number of dB. The shortcoming of the stationary noise suppressor is that non-stationary noise will not be suppressed, whereas the shortcoming of the fixed suppression system is that it must suppress noise by a conservative level in order to avoid speech distortion at low SNR.
- Another form of noise suppression is dynamic noise suppression. A common type of dynamic noise suppression systems is based on signal-to-noise ratios (SNR). The SNR may be used to determine a degree of suppression. Unfortunately, SNR by itself is not a very good predictor of speech distortion due to the presence of different noise types in the audio environment. SNR is a ratio indicating how much louder speech is then noise. However, speech may be a non-stationary signal which may constantly change and contain pauses. Typically, speech energy, over a given period of time, will include a word, a pause, a word, a pause, and so forth. Additionally, stationary and dynamic noises may be present in the audio environment. As such, it can be difficult to accurately estimate the SNR. The SNR averages all of these stationary and non-stationary speech and noise components. There is no consideration in the determination of the SNR of the characteristics of the noise signal—only the overall level of noise. In addition, the value of SNR can vary based on the mechanisms used to estimate the speech and noise, such as whether it based on local or global estimates, and whether it is instantaneous or for a given period of time.
- To overcome the shortcomings of the prior art, there is a need for an improved noise suppression system for processing audio signals.
- The present technology provides a robust noise suppression system which may concurrently reduce noise and echo components in an acoustic signal while limiting the level of speech distortion. An acoustic signal may be received and transformed to cochlear-domain sub-band signals. Features such as pitch may be identified and tracked within the sub-band signals. Initial speech and noise models may be then be estimated at least in part from a probability analysis based on the tracked pitch sources. Improved speech and noise models may be resolved from the initial speech and noise models and noise reduction may be performed on the sub-band signals and an acoustic signal may be reconstructed from the noise-reduced sub-band signals.
- In an embodiment, noise reduction may be performed by executing a program stored in memory to transform an acoustic signal from the time domain to cochlea-domain sub-band signals. Multiple sources of pitch may be tracked within the sub-band signals. A speech model and one or more noise models may be generated at least in part based on the tracked pitch sources. Noise reduction may be performed on the sub-band signals based on the speech model and one or more noise models.
- A system for performing noise reduction in an audio signal may include a memory, frequency analysis module, source inference module, and a modifier module. The frequency analysis module may be stored in the memory and executed by a processor to transform a time domain acoustic to cochlea domain sub-band signals. The source inference engine may be stored in the memory and executed by a processor to track multiple sources of pitch within a sub-band signal and to generate a speech model and one or more noise models based at least in part on the tracked pitch sources. The modifier module may be stored in the memory and executed by a processor to perform noise reduction on the sub-band signals based on the speech model and one or more noise models.
-
FIG. 1 is an illustration of an environment in which embodiments of the present technology may be used. -
FIG. 2 is a block diagram of an exemplary audio device. -
FIG. 3 is a block diagram of an exemplary audio processing system. -
FIG. 4 is a block diagram of exemplary modules within an audio processing system. -
FIG. 5 is a block diagram of exemplary components within a modifier module. -
FIG. 6 is a flowchart of an exemplary method for performing noise reduction for an acoustic signal. -
FIG. 7 is a flowchart of an exemplary method for estimating speech and noise models. -
FIG. 8 is a flowchart of an exemplary method for resolving speech and noise models. - The present technology provides a robust noise suppression system which may concurrently reduce noise and echo components in an acoustic signal while limiting the level of speech distortion. An acoustic signal may be received and transformed to cochlear-domain sub-band signals. Features such as pitch may be identified and tracked within the sub-band signals. Initial speech and noise models may be then be estimated at least in part from a probability analysis based on the tracked pitch sources. Improved speech and noise models may be resolved from the initial speech and noise models and noise reduction may be performed on the sub-band signals and an acoustic signal may be reconstructed from the noise-reduced sub-band signals.
- Multiple pitch sources may be identified in a sub-band frame and tracked over multiple frames. Each tracked pitch source (“track”) is analyzed based on several features, including pitch level, salience, and how stationary the pitch source is. Each pitch source is also compared to stored speech model information. For each track, a probability of being a target speech source is generated based on the features and comparison to the speech model information.
- A track with the highest probability may be, in some cases, designated as speech and the remaining tracks are designated as noises. In some embodiments, the there may be multiple speech sources, and a “target” speech may be the desired speech with other speech sources considered noise. Tracks with a probability over a certain threshold may be designated as speech speech. In addition, there may be a “softening” of the decision in the system. Downstream of the track probability determination, a spectrum may be constructed for each pitch track, and each track's probability may be mapped to gains through which the corresponding spectrum is added into the speech and non-stationary noise models. If the probability is high, the gain for the speech model will be 1 and the gain for the noise model will be 0, and vice versa.
- The present technology may utilize any of several techniques to provide an improved noise reduction of an acoustic signal. The present technology may estimate speech and noise models based on tracked pitch sources and probabilistic analysis of the tracks. Dominant speech detection may be used to control stationary noise estimations. Models for speech, noise and transients may be resolved into speech and noise. Noise reduction may be performed by filtering sub-bands using filters based on optimal least-squares estimation or on constrained optimization. These concepts are discussed in more detail below.
-
FIG. 1 is an illustration of an environment in which embodiments of the present technology may be used. A user may act as an audio (speech)source 102 to anaudio device 104. Theexemplary audio device 104 includes aprimary microphone 106. Theprimary microphone 106 may be omni-directional microphones. Alternatively embodiments may utilize other forms of a microphone or acoustic sensors, such as a directional microphone. - While the
microphone 106 receives sound (i.e. acoustic signals) from theaudio source 102, themicrophone 106 also picks upnoise 112. Although the noise 110 is shown coming from a single location inFIG. 1 , the noise 110 may include any sounds from one or more locations that differ from the location ofaudio source 102, and may include reverberations and echoes. These may include sounds produced by thedevice 104 itself. The noise 110 may be stationary, non-stationary, and/or a combination of both stationary and non-stationary noise. - Acoustic signals received by
microphone 106 may be tracked, for example by pitch. Features of each tracked signal may be determined and processed to estimate models for speech and noise. For example, aspeech source 102 may be associated with a pitch track with a higher energy level than thenoise source 112. Processing signals received bymicrophone 106 is discussed in more detail below. -
FIG. 2 is a block diagram of anexemplary audio device 104. In the illustrated embodiment, theaudio device 104 includesreceiver 200,processor 202,primary microphone 106,audio processing system 204, and anoutput device 206. Theaudio device 104 may include further or other components necessary foraudio device 104 operations. Similarly, theaudio device 104 may include fewer components that perform similar or equivalent functions to those depicted inFIG. 2 . -
Processor 202 may execute instructions and modules stored in a memory (not illustrated inFIG. 2 ) in theaudio device 104 to perform functionality described herein, including noise reduction for an acoustic signal.Processor 202 may include hardware and software implemented as a processing unit, which may process floating point operations and other operations for theprocessor 202. - The
exemplary receiver 200 may be configured to receive a signal from a communications network, such as a cellular telephone and/or data communication network. In some embodiments, thereceiver 200 may include an antenna device. The signal may then be forwarded to theaudio processing system 204 to reduce noise using the techniques described herein, and provide an audio signal tooutput device 206. The present technology may be used in one or both of the transmit and receive paths of theaudio device 104. - The
audio processing system 204 is configured to receive the acoustic signals from an acoustic source via theprimary microphone 106 and process the acoustic signals. Processing may include performing noise reduction within an acoustic signal. Theaudio processing system 204 is discussed in more detail below. The acoustic signal received byprimary microphone 106 may be converted into one or more electrical signals, such as for example a primary electrical signal and a secondary electrical signal. The electrical signal may be converted by an analog-to-digital converter (not shown) into a digital signal for processing in accordance with some embodiments. The primary acoustic signal may be processed by theaudio processing system 204 to produce a signal with an improved signal-to-noise ratio. - The
output device 206 is any device which provides an audio output to the user. For example, theoutput device 206 may include a speaker, an earpiece of a headset or handset, or a speaker on a conference device. - In various embodiments, the primary microphone is an omni-directional microphone; in other embodiments, the primary microphone is a directional microphone.
-
FIG. 3 is a block diagram of an exemplaryaudio processing system 204 for performing noise reduction as described herein. In exemplary embodiments, theaudio processing system 204 is embodied within a memory device withinaudio device 104. Theaudio processing system 204 may include atransform module 305, afeature extraction module 310, asource inference engine 315,modification generator module 320,modifier module 330,reconstructor module 335, and postprocessor module 340.Audio processing system 204 may include more or fewer components than illustrated inFIG. 3 , and the functionality of modules may be combined or expanded into fewer or additional modules. Exemplary lines of communication are illustrated between various modules ofFIG. 3 , and in other figures herein. The lines of communication are not intended to limit which modules are communicatively coupled with others, nor are they intended to limit the number of and type of signals communicated between modules. - In operation, an acoustic signal is received from the
primary microphone 106, is converted to an electrical signal, and the electrical signal is processed throughtransform module 305. The acoustic signal may be pre-processed in the time domain before being processed bytransform module 305. Time domain pre-processing may also include applying input limiter gains, speech time stretching, and filtering using an FIR or IIR filter. - The
transform module 305 takes the acoustic signals and mimics the frequency analysis of the cochlea. Thetransform module 305 comprises a filter bank designed to simulate the frequency response of the cochlea. Thetransform module 305 separates the primary acoustic signal into two or more frequency sub-band signals. A sub-band signal is the result of a filtering operation on an input signal, where the bandwidth of the filter is narrower than the bandwidth of the signal received by thetransform module 305. The filter bank may be implemented by a series of cascaded, complex-valued, first-order IIR filters. Alternatively, other filters or transforms such as a short-time Fourier transform (STFT), sub-band filter banks, modulated complex lapped transforms, cochlear models, wavelets, etc., can be used for the frequency analysis and synthesis. The samples of the sub-band signals may be grouped sequentially into time frames (e.g. over a predetermined period of time). For example, the length of a frame may be 4 ms, 8 ms, or some other length of time. In some embodiments there may be no frame at all. The results may include sub-band signals in a fast cochlea transform (FCT) domain. - The
analysis path 325 may be provided with anFCT domain representation 302 and optionally a high-density FCT representation 301 for improved pitch estimation and speech modeling (and system performance). A high density FCT may be a frame of sub-bands that have a higher density than theFCT 302; ahigh density FCT 301 may have more sub-bands thanFCT 302 within a frequency range of the acoustic signal. Thesignal path 330 also may be provided with anFCT representation 304 after implementing adelay 303. Using thedelay 303 provides theanalysis path 325 with a “lookahead” latency that can be leveraged to improve the speech and noise models during subsequent stages of processing. If there is no delay, theFCT 304 for thesignal path 330 is not necessary; the output ofFCT 302 in the diagram can be routed to the signal path processing as well as to theanalysis path 325. In the illustrated embodiment, thelookahead delay 303 is arranged before theFCT 304. As a result, the delay is implemented in the time domain in the illustrated embodiment, thereby saving memory resources as compared with implementing the lookahead delay in the FCT-domain. In alternative embodiments, the lookahead delay may be implemented in the FCT domain, such as by delaying the output ofFCT 302 and providing the delayed output to thesignal path 330. In doing so, computational resources may be saved compared with implementing the lookahead delay in the time-domain. - The sub-band frame signals are provided from
transform module 305 to an analysis path sub-system 325 and asignal path sub-system 330. The analysis path sub-system 325 may process the signal to identify signal features, distinguish between speech components and noise components of the sub-band signals, and generate a modification. The signal path sub-system 330 is responsible for modifying sub-band signals of the primary acoustic signal by reducing noise in the sub-band signals. Noise reduction can include applying a modifier, such as a multiplicative gain mask generated in theanalysis path sub-system 320, or applying a filter to each sub-band. The noise reduction may reduce noise and preserve the desired speech components in the sub-band signals. -
Feature extraction module 310 of the analysis path sub-system 325 receives the sub-band frame signals derived from the acoustic signal and computes features for each sub-band frame, such as pitch estimates and second-order statistics. In some embodiments, a pitch estimate may be determined byfeature extractor 310 and provided to sourceinference engine 315. In some embodiments, the pitch estimate may be determined bysource inference engine 315. The second-order statistics (instantaneous and smoothed autocorrelations/energies) are computed inblock 310 for each sub-band signal. For theHD FCT 301, only the zero-lag autocorrelations are computed and used by the pitch estimation procedure. The zero-lag autocorrelation may be a time sequence of the previous signal multiplied by itself and averaged. For themiddle FCT 302, the first-order lag autocorrelations are also computed since these may be used to generate a modification. The first-order lag autocorrelations, which may be computed by multiplying the time sequence of the previous signal with a version of itself offset by one sample, may also be used to improve the pitch estimation. -
Source inference engine 315 may process the frame and sub-band second-order statistics and pitch estimates provided by feature extraction module 310 (or generated by source inference engine 315) to derive models of the noise and speech in the sub-band signals.Source inference engine 315 processes the FCT-domain energies to derive models of the pitched components of the sub-band signals, the stationary components, and the transient components. The speech, noise and optional transient models are resolved into speech and noise models. If the present technology is utilizing non-zero lookahead,source inference engine 315 is the component wherein the lookahead is leveraged. At each frame,source inference engine 315 receives a new frame of analysis path data and outputs a new frame of signal path data (which corresponds to an earlier relative time in the input signal than the analysis path data). The lookahead delay may provide time to improve discrimination of speech and noise before the sub-band signals are actually modified (in the signal path). Also,source inference engine 315 outputs a voice activity detection (VAD) signal (for each tap) that is internally fed back to the stationary noise estimator to help prevent over-estimation of the noise. - The
modification generator module 320 receives models of the speech and noise as estimated bysource inference engine 315.Module 320 may derive a multiplicative mask for each sub-band per frame.Module 320 may also derive a linear enhancement filter for each sub-band per frame. The enhancement filter includes a suppression backoff mechanism wherein the filter output is cross-faded with its input sub-band signals. The linear enhancement filter may be used in addition or in place of the multiplicative mask, or not used at all. The cross-fade gain is combined with the filter coefficients for the sake of efficiency.Modification generator module 320 may also generate a post-mask for applying equalization and multiband compression. Spectral conditioning may also be included in this post-mask. - The multiplicative mask may be defined as a Wiener gain. The gain may be derived based on the autocorrelation of the primary acoustic signal and an estimate of the autocorrelation of the speech (e.g. the speech model) or an estimate of the autocorrelation of the noise (e.g. the noise model). Applying the derived gain yields an MMSE (minimum mean-squared error) estimate of the clean speech signal given the noisy signal.
- The linear enhancement filter is defined by a first-order Wiener filter. The filter coefficients may be derived based on the 0th and 1st order lag autocorrelation of the acoustic signal and an estimate of the 0th and 1st order lag autocorrelation of the speech or an estimate of the 0th and 1st order lag autocorrelation of the noise. In one embodiment, the filter coefficients are derived based on the optimal Wiener formulation using the following equations:
-
- where rxx[0] is the 0th order lag autocorrelation of the input signal, rxx[1] is the 1st order lag autocorrelation of the input signal, rss[0] is the estimated 0th order lag autocorrelation of the speech, and rss[1] is the estimated 1st order lag autocorrelation of the speech. In the Wiener formulations, * denotes conjugation and ∥ denotes magnitude. In some embodiments, the filter coefficients may be derived in part based on a multiplicative mask derived as described above. The coefficient β0 may be assigned the value of the multiplicative mask, and β1 may be determined as the optimal value for use in conjunction with that value of β0 according to the formula:
-
- Applying the filter yields an MMSE estimate of the clean speech signal given the noisy signal.
- The values of the gain mask or filter coefficients output from
modification generator module 320 are time and sub-band signal dependent and optimize noise reduction on a per sub-band basis. The noise reduction may be subject to the constraint that the speech loss distortion complies with a tolerable threshold limit. - In embodiments, the energy level of the noise component in the sub-band signal may be reduced to no less than a residual noise level, which may be fixed or slowly time-varying. In some embodiments, the residual noise level is the same for each sub-band signal, in other embodiments it may vary across sub-bands and frames. Such a noise level may be based on a lowest detected pitch level.
-
Modifier module 330 receives the signal path cochlear-domain samples fromtransform block 305 and applies a modification, such as for example a first-order FIR filter, to each sub-band signal.Modifier module 330 may also apply a multiplicative post-mask to perform such operations as equalization and multiband compression. For Rx applications, the post-mask may also include a voice equalization feature. Spectral conditioning may be included in the post-mask.Modifier 330 may also apply speech reconstruction at the output of the filter but prior to the post-mask. -
Reconstructor module 335 may convert the modified frequency sub-band signals from the cochlea domain back into the time domain. The conversion may include applying gains and phase shifts to the modified sub-band signals and adding the resulting signals. -
Reconstructor module 335 forms the time-domain system output by adding together the FCT-domain subband signals after optimized time delays and complex gains have been applied. The gains and delays are derived in the cochlea design process. Once conversion to the time domain is completed, the synthesized acoustic signal may be post-processed or output to a user viaoutput device 206 and/or provided to a codec for encoding. - Post-processing 340 may perform time-domain operations on the output of the noise reduction system. This includes comfort noise addition, automatic gain control, and output limiting. Speech time stretching may be performed as well, for example on an Rx signal.
- Comfort noise may be generated by a comfort noise generator and added to the synthesized acoustic signal prior to providing the signal to the user. Comfort noise may be a uniform constant noise that is not usually discernable to a listener (e.g., pink noise). This comfort noise may be added to the synthesized acoustic signal to enforce a threshold of audibility and to mask low-level non-stationary output noise components. In some embodiments, the comfort noise level may be chosen to be just above a threshold of audibility and may be settable by a user. In some embodiments, the
modification generator module 320 may have access to the level of comfort noise in order to generate gain masks that will suppress the noise to a level at or below the comfort noise. - The system of
FIG. 3 may process several types of signals received by an audio device. The system may be applied to acoustic signals received via one or more microphones. The system may also process signals, such as a digital Rx signal, received through an antenna or other connection. -
FIG. 4 is a block diagram of exemplary modules within an audio processing system. The modules illustrated in the block diagram ofFIG. 4 includesource inference engine 315,modification generator 320, andmodifier 330. -
Source inference engine 315 receives second order statistics data fromfeature extraction module 310 and provides this data to polyphonic pitch and source tracker (tracker) 420,stationary noise modeler 428 andtransient modeler 436.Tracker 420 receives the second order statistics and a stationary noise model and estimates pitches within the acoustic signal received bymicrophone 106. - Estimating the pitches may include estimating the highest level pitch, removing components corresponding to the pitch from the signal statistics, and estimating the next highest level pitch, for a number of iterations per a configurable parameter. First, for each frame, peaks may be detected in the FCT-domain spectral magnitude, which may be based on the 0th order lag autocorrelation and may further be based on a mean subtraction such that the FCT-domain spectral magnitude has zero mean. In some embodiments, the peaks must meet a certain criteria, such as being larger than their four nearest neighbors, and must have a large enough level relative to the maximum input level. The detected peaks form the first set of pitch candidates. Subsequently, sub-pitches are added to the set for each candidate, i.e., f0/2 f0/3 f0/4, and so forth, where f0 denotes a pitch candidate. Cross correlation is then performed by adding the level of the interpolated FCT-domain spectral magnitude at harmonic points over a specific frequency range, thereby forming a score for each pitch candidate. Because the FCT-domain spectral magnitude is zero-mean over that range (due to the mean subtraction), pitch candidates are penalized if a harmonic does not correspond to an area of significant amplitude (because the zero-mean FCT-domain spectral magnitude will have negative values at such points). This ensures that frequencies below the true pitch are adequately penalized relative to the true pitch. For example, a 0.1 Hz candidate would be given a near-zero score (because it would be the sum of all FCT-domain spectral magnitude points, which is zero by construction).
- The cross-correlation may then provide scores for each pitch candidate. Many candidates are very close in frequency (because of the addition of the sub-pitches f0/2 f0/3 f0/4 etc to the set of candidates). The scores of candidates that are close in frequency are compared, and only the best one is retained. A dynamic programming algorithm is used to select the best candidate in the current frame, given the candidates in previous frames. The dynamic programming algorithm ensures that the candidate with the best score is generally selected as the primary pitch, and helps avoid octave errors.
- Once the primary pitch has been chosen, the harmonic amplitudes are computed simply using the level of the interpolated FCT-domain spectral magnitude at harmonic frequencies. A basic speech model is applied to the harmonics to make sure they are consistent with a normal speech signal. Once the harmonic levels are computed, the harmonics are removed from the interpolated FCT-domain spectral magnitude to form a modified FCT-domain spectral magnitude.
- The pitch detection process is repeated, using the modified FCT-domain spectral magnitude. At the end of the second iteration, the best pitch is selected, without running another dynamic programming algorithm. Its harmonics are computed, and removed from the FCT-domain spectral magnitude. The third pitch is the next best candidate, and its harmonic levels are computed on the twice-modified FCT-domain spectral magnitude. This process is continued until a configurable number of pitches has been estimated. The configurable number may be for example three or some other number. As a last stage, the pitch estimates are refined using the phase of the 1st order lag autocorrelation.
- A number of the estimated pitches are then tracked by the polyphonic pitch and
source tracker 420. The tracking may determine changes in frequency and level of the pitch over multiple frames of the acoustic signal. In some embodiments, a subset of the estimated pitches are tracked, for example the estimated pitches having the highest energy level(s). - The output of the pitch detection algorithm consists of a number of pitch candidates. The first candidate may be continuous across frames because it is selected by the dynamic programming algorithm. The remaining candidates may be output in order of salience, and therefore may not form frequency-continuous tracks across frames. For the task of assigning types to sources (talker associated with speech or distractor associated with noise) it is important to be able to deal with pitch tracks that are continuous in time, rather than collections of candidates at each frame. This is the goal of the multi-pitch tracking step, carried out on the per-frame pitch estimates determined by the pitch detection.
- Given N input candidates, the algorithm outputs N tracks, immediately reusing a track slot when the track terminates and a new one is born. At each frame the algorithm considers the N! associations of (N) existing tracks to (N) new pitch candidates. For example, if N=3, tracks 1, 2, 3 from the previous frame can be continued to
candidates 1, 2, 3 in the current frame in 6 manners: (1-1,2-2, 3-3), (1-1,2-3, 3-2), (1-2,2-3, 3-1), (1-2,2-1, 3-3), (1-3,2-2, 3-1), (1-3,3-2, 2-1). For each of these associations, a transition probability is computed to evaluate which association is the most likely. The transition probability is computed based on how close in frequency the candidate pitch is from the track pitch, the relative candidate and track levels, and the age of the track (in frames, since its beginning). The transition probabilities tend to favor continuous pitch tracks, tracks with larger levels, and tracks that are older than other ones. - Once the N! transition probabilities are computed, the largest one is selected, and the corresponding transition is used to continue the tracks into the current frame. A track dies when its transition probability to any of the current candidates is 0 in the best association (in other words, it cannot be continued into any of the candidates). Any candidate pitch that isn't connected to an existing track forms a new track with an age of 0. The algorithm outputs the tracks, their level, and their age.
- Each of the tracked pitches may be analyzed to estimate the probability of the whether the tracked source is a talker or speech source The cues estimated and mapped to probabilities are level, stationarity, speech model similarity, track continuity, and pitch range.
- The pitch track data is provided to buffer 422 and then to pitch
track processor 424.Pitch track processor 424 may smooth the pitch tracking for consistent speech target selection.Pitch track processor 424 may also track the lowest-frequency identified pitch. The output ofpitch track processor 424 is provided to pitchspectral modeler 426 and to computemodification filter 450. -
Stationary noise modeler 428 generates a model of stationary noise. The stationary noise model may be based on second order statistics as well as a voice activity detection signal received from pitchspectral modeler 426. The stationary noise model may be provided to pitchspectral modeler 426,update control 432, and polyphonic pitch andsource tracker 420.Transient modeler 436 may receive second order statistics and provide the transient noise model totransient model resolution 442 viabuffer 438. Thebuffers analysis path 315 and thesignal path 330. - Construction of the stationary noise model may involve a combined feedback and feed-forward technique based on speech dominance. For example, in one feed-forward technique, if the constructed speech and noise models indicate that the speech is dominant in a given sub-band, the stationary noise estimator is not updated for that sub-band. Rather, the stationary noise estimator is reverted to that of the previous frame. In one feedback technique, if speech (voice) is determined to be dominant in a given sub-band for a given frame, the noise estimation is rendered inactive (frozen) in that sub-band during the next frame. Hence, a decision is made in a current frame not to estimate stationary noise in a subsequent frame.
- The speech dominance may be indicated by a voice activity detector (VAD) indicator computed for the current frame and used by
update control module 432. The VAD may be stored in the system and used by thestationary noise estimator 428 in the subsequent frame. This dual-mode VAD prevents damage to low-level speech, especially high-frequency harmonics; this reduces the “voice muffling” effect frequently incurred in noise suppressors. - Pitch
spectral modeler 426 may receive pitch track data frompitch track processor 424, a stationary noise model, a transient noise model, second orders statistics, and optionally other data and may output a speech model and a nonstationary noise model. Pitchspectral modifier 426 may also provide a VAD signal indicating whether speech is dominant in a particular sub-band and frame. - The pitch tracks (each comprising pitch, salience, level, stationarity, and speech probability) are used to construct models of the speech and noise spectra by the pitch
spectral model builder 426. To construct models of the speech and noise, the pitch tracks may be reordered based on the track saliences, such that the model for the highest salience pitch track will be constructed first. An exception is that high-frequency tracks with a salience above a certain threshold are prioritized. Alternatively, the pitch tracks may be reordered based on the speech probability, such that the model for the most probable speech track will be constructed first. - In
module 426, a broadband stationary noise estimate may be subtracted from the signal energy spectrum to form a modified spectrum. Next, the present system may iteratively estimate the energy spectra of the pitch tracks according to the processing order determined in the first step. An energy spectrum may be derived by estimating an amplitude for each harmonic (by sampling the modified spectrum), computing a harmonic template corresponding to the response of the cochlea to a sinusoid at the harmonic's amplitude and frequency, and accumulating the harmonic's template into the track spectral estimate. After the harmonic contributions are aggregated, the track spectrum is subtracted to form a new modified signal spectrum for the next iteration. - To compute the harmonic templates, the module uses a pre-computed approximation of the cochlea transfer function matrix. For a given sub-band, the approximation consists of a piecewise linear fit of the sub-band's frequency response where the approximation points are optimally selected from the set of sub-band center frequencies (so that sub-band indices can be stored instead of explicit frequencies).
- After the harmonic spectra are iteratively estimated, each spectrum is allocated in part to the speech model and in part to the non-stationary noise model, where the extent of the allocation to the speech model is dictated by the speech probability of the corresponding track, and the extent of the allocation to the noise model is determined as an inverse of the extent of the allocation to the speech model.
-
Noise model combiner 434 may combine stationary noise and non-stationary noise and provide the resulting noise totransient model resolution 442.Update control 432 may determine whether or not the stationary noise estimate is to be updated in the current frame, and provide the resulting stationary noise tonoise model combiner 434 to be combined with the non-stationary noise model. -
Transient model resolution 442 receives a noise model, speech model, and transient model and resolves the models into speech and noise. The resolution involves verifying that the speech model and noise model do not overlap, and determining whether the transient model is speech or noise. The noise and non-speech transient models are deemed noise and the speech model and transient speech are determined to be speech. The transient noise models re provided to repairmodule 462, and the resolved speech and noise modules are provided toSNR estimator 444 as well as the computemodification filter module 450. The speech model and the noise model are resolved to reduce cross-model leakage. The models are resolved into a consistent decomposition of the input signal into speech and noise. -
SNR estimator 444 determines an estimate of the signal to noise ratio (SNR). The SNR estimate can be used to determine an adaptive level of suppression in thecrossfade module 464. It can also be used to control other aspects of the system behavior. For example, the SNR may be used to adaptively change what the speech/noise model resolution does. - Compute
modification filter module 450 generates a modification filter to be applied to each sub-band signal. In some embodiments, a filter such as a first-order filter is applied in each sub-band instead of a simple multiplier.Modification filter module 450 is discussed in more detail below with respect toFIG. 5 . - The modification filter is applied to the sub-band signals by
module 460. After applying the generated filter, portions of the sub-band signal may be repaired atmodule 462 and then linearly combined with the unmodified sub-band signal atcrossfade 464. The transient components may be repaired bymodule 462 and the crossfade may be performed based on the SNR provided bySNR estimator 444. The sub-bands are then reconstructed atreconstructor module 335. -
FIG. 5 is a block diagram of exemplary components within a modifier module.Modifier module 500 includesdelays multipliers modules multipliers modification filter 500. A sub-band signal for the current frame, x[k, t], is received by thefilter 500, processed by the delays, multipliers, and summing modules, and an estimate of the speech s[k,t] is provided at the output of the final summingmodule 545. In themodifier 500, noise reduction is carried out by filtering each sub-band signal, unlike previous systems which apply a scalar mask. With respect to scalar multiplication, such per-sub-band filtering allows nonuniform spectral treatment within a given sub-band; in particular this may be relevant where speech and noise components have different spectral shapes within the sub-band (as in the higher frequency sub-bands), and the spectral response within the subband can be optimized to preserve the speech and suppress the noise. - The filter coefficients β0 and β1 are computed based on speech models derived by the
source inference engine 315, combined with a sub-pitch suppression mask (for example by tracking the lowest speech pitch and suppressing the sub-bands below this min pitch by reducing the β0 and β1 values for those sub-bands), and crossfaded based on the desired noise suppression level. In another approach, the VQOS approach is used to determine the crossfade. The β0 and β1 values are then subjected to interframe rate-of-change limits and interpolated across frames before being applied to the cochlear-domain signals in the modification filter. For the implementation of the delay, one sample of cochlear-domain signals (a time slice across sub-bands) is stored in the module state. - To implement a first-order modification filter, the received sub-band signal is multiplied by β0 and also delayed by one sample. The signal at the output of the delay is multiplied by β1. The results of the two multiplications are summed and provided as the output s[k,t]. The delay, multiplications, and summation correspond to the application of a first-order linear filter. There may be N delay-multiply-sum stages, corresponding to an Nth order filter.
- When applying a first-order filter in each sub-band instead of a simple multiplier, an optimal scalar multiplier (mask) may be used in the non-delayed branch of the filter. The filter coefficient for the delayed branch may be derived to be optimal conditioned on the scalar mask. In this way, the first-order filter is able to achieve a higher-quality speech estimate than using the scalar mask alone. The system can be extended to higher orders (an N-th order filter) if desired. Also, for an N-th order filter, the autocorrelations up to lag N may be computed in feature extraction module 310 (second-order statistics). In the first-order case, the zero-th and first-order lag autocorrelations are computed. This is a distinction from prior systems which rely solely on the zero-th order lag.
-
FIG. 6 is a flowchart of an exemplary method for performing noise reduction for an acoustic signal. First, an acoustic signal may be received atstep 605. The acoustic signal may be received bymicrophone 106. The acoustic signal may be transformed to the cochlea domain atstep 610.Transform module 305 may perform a fast cochlea transform to generate cochlea domain sub-band signals. In some embodiments, the transformation may be performed after a delay is implemented in the time domain. In such a case, there can be two cochleas, one for theanalysis path 325, and one for thesignal path 330 after the time-domain delay. - Monaural features are extracted from the cochlea domain sub-band signals at
step 615. The monaural features are extracted byfeature extractor 310 and may include second order statistics. Some features may include pitch, energy level, pitch salience, and other data. - Speech and noise models may be estimated for cochlea sub-bands at
step 620. The speech and noise models may be estimated bysource inference engine 315. Generating the speech model and noise model may include estimating a number of pitch elements for each frame, tracking a selected number of the pitch elements across frames, and selecting one of the tracked pitches as a talker based on a probabilistic analysis. The speech model is generated from the tracked talker. A non-stationary noise model may be based on the other tracked pitches and a stationary noise model may be based on extracted features provided byfeature extraction module 310. Step 620 is discussed in more detail with respect to the method ofFIG. 7 . - The speech model and noise models may be resolved at
step 625. Resolving the speech model and noise model may be performed to eliminate any cross-leakage between the two models. Step 625 is discussed in more detail with respect to the method ofFIG. 8 . Noise reduction may be performed on the subband signals based on the speech model and noise models atstep 630. The noise reduction may include applying a first order (or Nth order) filter to each sub-band in the current frame. The filter may provide better noise reduction than simply applying a scalar gain for each sub-band. The filter may be generated inmodification generator 320 and applied to the sub-band signals atstep 330. - The sub-bands may be reconstructed at
step 635. Reconstruction of the sub-bands may involve applying a series of delays and complex-multiply operations to the sub-band signals byreconstructor 335. The reconstructed time-domain signal may be post-processed atstep 640. Post-processing may consist of adding comfort noise, performing automatic gain control (AGC) and applying a final output limiter. The noise-reduced time-domain signal is output atstep 645. -
FIG. 7 is a flowchart of an exemplary method for estimating speech and noise models. The method ofFIG. 7 may provide more detail forstep 620 in the method ofFIG. 6 . First, pitch sources are identified atstep 705. Polyphonic pitch and source tracking module (tracking module) 420 may identify pitches present within a frame. The identified pitches may be tracked across frames atstep 710. The pitches may be tracked over different frames by trackingmodule 420. - A speech source is identified by a probability analysis at
step 715. The probability analysis identifies a probability that each pitch track is the desired talker based on each of several features, including level, salience, similarity to speech models, stationarity, and other features. A single probability for each pitch is determined based on the feature probabilities for that pitch, for example by multiplying the feature probabilities. The speech source may be identified as the pitch track with the highest probability of being associated with the talker. - A speech model and noise model are constructed at
step 720. The speech model is constructed in part based on the pitch track with the highest probability. The noise model is constructed based in part on the pitch tracks that have a low probability of corresponding to the desired talker. Transient components identified as speech may be included in the speech model and transient components identified as non-speech transient may be included in the noise model. Both the speech model and the noise model are determined bysource inference engine 315. -
FIG. 8 is a flowchart of an exemplary method for resolving speech and noise models. A noise model estimation may be configured using feedback and feedforward control atstep 805. When a sub-band within a current frame is determined to be dominated by speech, the noise estimate from the previous frame is frozen (e.g., used in the current frame) as well as in the next frame for that sub-band. - A speech model and noise model are resolved into speech and noise at
step 810. Portions of a speech model may leak into a noise model, and vice-versa. The speech and noise models are resolved such that there is no leakage between the two. - A delayed time-domain acoustic signal may be provided to the signal path to allow additional time (look-ahead) for the analysis path to discriminate between speech and noise in
step 815. By utilizing a time-domain delay in the look-ahead mechanism, memory resources are saved as compared to implementing the lookahead delay in the cochlear domain. - The steps discussed in
FIGS. 6-8 may be performed in a different order than that discussed, and the methods ofFIGS. 4 and 5 may each include additional or fewer steps than those illustrated. - The above described modules, including those discussed with respect to
FIG. 3 , may include instructions stored in a storage media such as a machine readable medium (e.g., computer readable medium). These instructions may be retrieved and executed by theprocessor 202 to perform the functionality discussed herein. Some examples of instructions include software, program code, and firmware. Some examples of storage media include memory devices and integrated circuits. - While the present invention is disclosed by reference to the preferred embodiments and examples detailed above, it is to be understood that these examples are intended in an illustrative rather than a limiting sense. It is contemplated that modifications and combinations will readily occur to those skilled in the art, which modifications and combinations will be within the spirit of the invention and the scope of the following claims.
Claims (20)
Priority Applications (6)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US12/860,043 US8447596B2 (en) | 2010-07-12 | 2010-08-20 | Monaural noise suppression based on computational auditory scene analysis |
KR1020137000488A KR20130117750A (en) | 2010-07-12 | 2011-05-19 | Monaural noise suppression based on computational auditory scene analysis |
PCT/US2011/037250 WO2012009047A1 (en) | 2010-07-12 | 2011-05-19 | Monaural noise suppression based on computational auditory scene analysis |
JP2013519682A JP2013534651A (en) | 2010-07-12 | 2011-05-19 | Monaural noise suppression based on computational auditory scene analysis |
TW100118902A TW201214418A (en) | 2010-07-12 | 2011-05-30 | Monaural noise suppression based on computational auditory scene analysis |
US13/859,186 US9431023B2 (en) | 2010-07-12 | 2013-04-09 | Monaural noise suppression based on computational auditory scene analysis |
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US36363810P | 2010-07-12 | 2010-07-12 | |
US12/860,043 US8447596B2 (en) | 2010-07-12 | 2010-08-20 | Monaural noise suppression based on computational auditory scene analysis |
Related Child Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US13/859,186 Continuation US9431023B2 (en) | 2010-07-12 | 2013-04-09 | Monaural noise suppression based on computational auditory scene analysis |
Publications (2)
Publication Number | Publication Date |
---|---|
US20120010881A1 true US20120010881A1 (en) | 2012-01-12 |
US8447596B2 US8447596B2 (en) | 2013-05-21 |
Family
ID=45439210
Family Applications (2)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US12/860,043 Active 2031-02-15 US8447596B2 (en) | 2010-07-12 | 2010-08-20 | Monaural noise suppression based on computational auditory scene analysis |
US13/859,186 Active US9431023B2 (en) | 2010-07-12 | 2013-04-09 | Monaural noise suppression based on computational auditory scene analysis |
Family Applications After (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US13/859,186 Active US9431023B2 (en) | 2010-07-12 | 2013-04-09 | Monaural noise suppression based on computational auditory scene analysis |
Country Status (5)
Country | Link |
---|---|
US (2) | US8447596B2 (en) |
JP (1) | JP2013534651A (en) |
KR (1) | KR20130117750A (en) |
TW (1) | TW201214418A (en) |
WO (1) | WO2012009047A1 (en) |
Cited By (35)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20150162021A1 (en) * | 2013-12-06 | 2015-06-11 | Malaspina Labs (Barbados), Inc. | Spectral Comb Voice Activity Detection |
US20150348572A1 (en) * | 2014-05-30 | 2015-12-03 | Apple Inc. | Detecting a user's voice activity using dynamic probabilistic models of speech features |
US9343056B1 (en) | 2010-04-27 | 2016-05-17 | Knowles Electronics, Llc | Wind noise detection and suppression |
US9431023B2 (en) | 2010-07-12 | 2016-08-30 | Knowles Electronics, Llc | Monaural noise suppression based on computational auditory scene analysis |
US9437188B1 (en) | 2014-03-28 | 2016-09-06 | Knowles Electronics, Llc | Buffered reprocessing for multi-microphone automatic speech recognition assist |
US9438992B2 (en) | 2010-04-29 | 2016-09-06 | Knowles Electronics, Llc | Multi-microphone robust noise suppression |
US9502048B2 (en) | 2010-04-19 | 2016-11-22 | Knowles Electronics, Llc | Adaptively reducing noise to limit speech distortion |
US9508345B1 (en) | 2013-09-24 | 2016-11-29 | Knowles Electronics, Llc | Continuous voice sensing |
US9530434B1 (en) * | 2013-07-18 | 2016-12-27 | Knuedge Incorporated | Reducing octave errors during pitch determination for noisy audio signals |
US9558755B1 (en) | 2010-05-20 | 2017-01-31 | Knowles Electronics, Llc | Noise suppression assisted automatic speech recognition |
US9584899B1 (en) | 2015-11-25 | 2017-02-28 | Doppler Labs, Inc. | Sharing of custom audio processing parameters |
US9589574B1 (en) | 2015-11-13 | 2017-03-07 | Doppler Labs, Inc. | Annoyance noise suppression |
US20170110142A1 (en) * | 2015-10-18 | 2017-04-20 | Kopin Corporation | Apparatuses and methods for enhanced speech recognition in variable environments |
US9640194B1 (en) | 2012-10-04 | 2017-05-02 | Knowles Electronics, Llc | Noise suppression for speech processing based on machine-learning mask estimation |
US20170133006A1 (en) * | 2015-11-06 | 2017-05-11 | Samsung Electronics Co., Ltd. | Neural network training apparatus and method, and speech recognition apparatus and method |
US9654861B1 (en) | 2015-11-13 | 2017-05-16 | Doppler Labs, Inc. | Annoyance noise suppression |
WO2017082974A1 (en) * | 2015-11-13 | 2017-05-18 | Doppler Labs, Inc. | Annoyance noise suppression |
US9679555B2 (en) | 2013-06-26 | 2017-06-13 | Qualcomm Incorporated | Systems and methods for measuring speech signal quality |
US9678709B1 (en) | 2015-11-25 | 2017-06-13 | Doppler Labs, Inc. | Processing sound using collective feedforward |
US9699554B1 (en) | 2010-04-21 | 2017-07-04 | Knowles Electronics, Llc | Adaptive signal equalization |
US9703524B2 (en) | 2015-11-25 | 2017-07-11 | Doppler Labs, Inc. | Privacy protection in collective feedforward |
US9799330B2 (en) | 2014-08-28 | 2017-10-24 | Knowles Electronics, Llc | Multi-sourced noise suppression |
CN107331406A (en) * | 2017-07-03 | 2017-11-07 | 福建星网智慧软件有限公司 | A kind of method of dynamic adjustment Echo-delay |
US9953634B1 (en) | 2013-12-17 | 2018-04-24 | Knowles Electronics, Llc | Passive training for automatic speech recognition |
CN108806708A (en) * | 2018-06-13 | 2018-11-13 | 中国电子科技集团公司第三研究所 | Voice de-noising method based on Computational auditory scene analysis and generation confrontation network model |
US20190096432A1 (en) * | 2017-09-25 | 2019-03-28 | Fujitsu Limited | Speech processing method, speech processing apparatus, and non-transitory computer-readable storage medium for storing speech processing computer program |
WO2019067335A1 (en) * | 2017-09-29 | 2019-04-04 | Knowles Electronics, Llc | Multi-core audio processor with phase coherency |
US10306389B2 (en) | 2013-03-13 | 2019-05-28 | Kopin Corporation | Head wearable acoustic system with noise canceling microphone geometry apparatuses and methods |
US10339952B2 (en) | 2013-03-13 | 2019-07-02 | Kopin Corporation | Apparatuses and systems for acoustic channel auto-balancing during multi-channel signal extraction |
CN111883154A (en) * | 2020-07-17 | 2020-11-03 | 海尔优家智能科技(北京)有限公司 | Echo cancellation method and apparatus, computer-readable storage medium, and electronic apparatus |
US10853025B2 (en) | 2015-11-25 | 2020-12-01 | Dolby Laboratories Licensing Corporation | Sharing of custom audio processing parameters |
US10891954B2 (en) | 2019-01-03 | 2021-01-12 | International Business Machines Corporation | Methods and systems for managing voice response systems based on signals from external devices |
US10997987B2 (en) * | 2017-06-16 | 2021-05-04 | Nxp B.V. | Signal processor for speech enhancement and recognition by using two output terminals designated for noise reduction |
US11145320B2 (en) | 2015-11-25 | 2021-10-12 | Dolby Laboratories Licensing Corporation | Privacy protection in collective feedforward |
EP4198975A1 (en) * | 2021-12-16 | 2023-06-21 | GN Hearing A/S | Electronic device and method for obtaining a user's speech in a first sound signal |
Families Citing this family (22)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US8849663B2 (en) * | 2011-03-21 | 2014-09-30 | The Intellisis Corporation | Systems and methods for segmenting and/or classifying an audio signal from transformed audio information |
US9142220B2 (en) | 2011-03-25 | 2015-09-22 | The Intellisis Corporation | Systems and methods for reconstructing an audio signal from transformed audio information |
US8620646B2 (en) | 2011-08-08 | 2013-12-31 | The Intellisis Corporation | System and method for tracking sound pitch across an audio signal using harmonic envelope |
US8548803B2 (en) | 2011-08-08 | 2013-10-01 | The Intellisis Corporation | System and method of processing a sound signal including transforming the sound signal into a frequency-chirp domain |
US9183850B2 (en) | 2011-08-08 | 2015-11-10 | The Intellisis Corporation | System and method for tracking sound pitch across an audio signal |
US8892046B2 (en) * | 2012-03-29 | 2014-11-18 | Bose Corporation | Automobile communication system |
CN104064197B (en) * | 2014-06-20 | 2017-05-17 | 哈尔滨工业大学深圳研究生院 | Method for improving speech recognition robustness on basis of dynamic information among speech frames |
TWI584275B (en) * | 2014-11-25 | 2017-05-21 | 宏達國際電子股份有限公司 | Electronic device and method for analyzing and playing sound signal |
US9712915B2 (en) | 2014-11-25 | 2017-07-18 | Knowles Electronics, Llc | Reference microphone for non-linear and time variant echo cancellation |
US9842611B2 (en) | 2015-02-06 | 2017-12-12 | Knuedge Incorporated | Estimating pitch using peak-to-peak distances |
US9870785B2 (en) | 2015-02-06 | 2018-01-16 | Knuedge Incorporated | Determining features of harmonic signals |
US9922668B2 (en) | 2015-02-06 | 2018-03-20 | Knuedge Incorporated | Estimating fractional chirp rate with multiple frequency representations |
US10262677B2 (en) * | 2015-09-02 | 2019-04-16 | The University Of Rochester | Systems and methods for removing reverberation from audio signals |
US10403259B2 (en) | 2015-12-04 | 2019-09-03 | Knowles Electronics, Llc | Multi-microphone feedforward active noise cancellation |
US20170206898A1 (en) * | 2016-01-14 | 2017-07-20 | Knowles Electronics, Llc | Systems and methods for assisting automatic speech recognition |
CN105957520B (en) * | 2016-07-04 | 2019-10-11 | 北京邮电大学 | A kind of voice status detection method suitable for echo cancelling system |
US10262673B2 (en) | 2017-02-13 | 2019-04-16 | Knowles Electronics, Llc | Soft-talk audio capture for mobile devices |
WO2019133765A1 (en) | 2017-12-28 | 2019-07-04 | Knowles Electronics, Llc | Direction of arrival estimation for multiple audio content streams |
DE102019214220A1 (en) * | 2019-09-18 | 2021-03-18 | Sivantos Pte. Ltd. | Method for operating a hearing aid and hearing aid |
US11587575B2 (en) * | 2019-10-11 | 2023-02-21 | Plantronics, Inc. | Hybrid noise suppression |
CN110769111A (en) * | 2019-10-28 | 2020-02-07 | 珠海格力电器股份有限公司 | Noise reduction method, system, storage medium and terminal |
CN110739005B (en) * | 2019-10-28 | 2022-02-01 | 南京工程学院 | Real-time voice enhancement method for transient noise suppression |
Citations (9)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20050049857A1 (en) * | 2003-08-25 | 2005-03-03 | Microsoft Corporation | Method and apparatus using harmonic-model-based front end for robust speech recognition |
US7065486B1 (en) * | 2002-04-11 | 2006-06-20 | Mindspeed Technologies, Inc. | Linear prediction based noise suppression |
US7254535B2 (en) * | 2004-06-30 | 2007-08-07 | Motorola, Inc. | Method and apparatus for equalizing a speech signal generated within a pressurized air delivery system |
US20080228474A1 (en) * | 2007-03-16 | 2008-09-18 | Spreadtrum Communications Corporation | Methods and apparatus for post-processing of speech signals |
US7574352B2 (en) * | 2002-09-06 | 2009-08-11 | Massachusetts Institute Of Technology | 2-D processing of speech |
US20090228272A1 (en) * | 2007-11-12 | 2009-09-10 | Tobias Herbig | System for distinguishing desired audio signals from noise |
US20100094622A1 (en) * | 2008-10-10 | 2010-04-15 | Nexidia Inc. | Feature normalization for speech and audio processing |
US7725314B2 (en) * | 2004-02-16 | 2010-05-25 | Microsoft Corporation | Method and apparatus for constructing a speech filter using estimates of clean speech and noise |
US7925502B2 (en) * | 2007-03-01 | 2011-04-12 | Microsoft Corporation | Pitch model for noise estimation |
Family Cites Families (213)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US3581122A (en) | 1967-10-26 | 1971-05-25 | Bell Telephone Labor Inc | All-pass filter circuit having negative resistance shunting resonant circuit |
US3989897A (en) | 1974-10-25 | 1976-11-02 | Carver R W | Method and apparatus for reducing noise content in audio signals |
US4811404A (en) | 1987-10-01 | 1989-03-07 | Motorola, Inc. | Noise suppression system |
US4910779A (en) | 1987-10-15 | 1990-03-20 | Cooper Duane H | Head diffraction compensated stereo system with optimal equalization |
IL84948A0 (en) | 1987-12-25 | 1988-06-30 | D S P Group Israel Ltd | Noise reduction system |
US5027306A (en) | 1989-05-12 | 1991-06-25 | Dattorro Jon C | Decimation filter as for a sigma-delta analog-to-digital converter |
US5050217A (en) | 1990-02-16 | 1991-09-17 | Akg Acoustics, Inc. | Dynamic noise reduction and spectral restoration system |
US5103229A (en) | 1990-04-23 | 1992-04-07 | General Electric Company | Plural-order sigma-delta analog-to-digital converters using both single-bit and multiple-bit quantization |
JPH0566795A (en) | 1991-09-06 | 1993-03-19 | Gijutsu Kenkyu Kumiai Iryo Fukushi Kiki Kenkyusho | Noise suppressing device and its adjustment device |
JP3279612B2 (en) | 1991-12-06 | 2002-04-30 | ソニー株式会社 | Noise reduction device |
JP3176474B2 (en) | 1992-06-03 | 2001-06-18 | 沖電気工業株式会社 | Adaptive noise canceller device |
US5408235A (en) | 1994-03-07 | 1995-04-18 | Intel Corporation | Second order Sigma-Delta based analog to digital converter having superior analog components and having a programmable comb filter coupled to the digital signal processor |
JP3307138B2 (en) | 1995-02-27 | 2002-07-24 | ソニー株式会社 | Signal encoding method and apparatus, and signal decoding method and apparatus |
US5828997A (en) | 1995-06-07 | 1998-10-27 | Sensimetrics Corporation | Content analyzer mixing inverse-direction-probability-weighted noise to input signal |
JPH0944186A (en) * | 1995-07-31 | 1997-02-14 | Matsushita Electric Ind Co Ltd | Noise suppressing device |
US5687104A (en) | 1995-11-17 | 1997-11-11 | Motorola, Inc. | Method and apparatus for generating decoupled filter parameters and implementing a band decoupled filter |
US5774562A (en) | 1996-03-25 | 1998-06-30 | Nippon Telegraph And Telephone Corp. | Method and apparatus for dereverberation |
JP3325770B2 (en) | 1996-04-26 | 2002-09-17 | 三菱電機株式会社 | Noise reduction circuit, noise reduction device, and noise reduction method |
US5701350A (en) | 1996-06-03 | 1997-12-23 | Digisonix, Inc. | Active acoustic control in remote regions |
US5825898A (en) | 1996-06-27 | 1998-10-20 | Lamar Signal Processing Ltd. | System and method for adaptive interference cancelling |
US5806025A (en) | 1996-08-07 | 1998-09-08 | U S West, Inc. | Method and system for adaptive filtering of speech signals using signal-to-noise ratio to choose subband filter bank |
JPH10124088A (en) | 1996-10-24 | 1998-05-15 | Sony Corp | Device and method for expanding voice frequency band width |
US5963651A (en) | 1997-01-16 | 1999-10-05 | Digisonix, Inc. | Adaptive acoustic attenuation system having distributed processing and shared state nodal architecture |
JP3328532B2 (en) | 1997-01-22 | 2002-09-24 | シャープ株式会社 | Digital data encoding method |
US6104993A (en) | 1997-02-26 | 2000-08-15 | Motorola, Inc. | Apparatus and method for rate determination in a communication system |
JP4132154B2 (en) | 1997-10-23 | 2008-08-13 | ソニー株式会社 | Speech synthesis method and apparatus, and bandwidth expansion method and apparatus |
US6343267B1 (en) | 1998-04-30 | 2002-01-29 | Matsushita Electric Industrial Co., Ltd. | Dimensionality reduction for speaker normalization and speaker and environment adaptation using eigenvoice techniques |
US6160265A (en) | 1998-07-13 | 2000-12-12 | Kensington Laboratories, Inc. | SMIF box cover hold down latch and box door latch actuating mechanism |
US6240386B1 (en) | 1998-08-24 | 2001-05-29 | Conexant Systems, Inc. | Speech codec employing noise classification for noise compensation |
US6539355B1 (en) | 1998-10-15 | 2003-03-25 | Sony Corporation | Signal band expanding method and apparatus and signal synthesis method and apparatus |
US6226606B1 (en) * | 1998-11-24 | 2001-05-01 | Microsoft Corporation | Method and apparatus for pitch tracking |
US6011501A (en) | 1998-12-31 | 2000-01-04 | Cirrus Logic, Inc. | Circuits, systems and methods for processing data in a one-bit format |
US6453287B1 (en) | 1999-02-04 | 2002-09-17 | Georgia-Tech Research Corporation | Apparatus and quality enhancement algorithm for mixed excitation linear predictive (MELP) and other speech coders |
US6381570B2 (en) | 1999-02-12 | 2002-04-30 | Telogy Networks, Inc. | Adaptive two-threshold method for discriminating noise from speech in a communication signal |
US6377915B1 (en) | 1999-03-17 | 2002-04-23 | Yrp Advanced Mobile Communication Systems Research Laboratories Co., Ltd. | Speech decoding using mix ratio table |
US6490556B2 (en) | 1999-05-28 | 2002-12-03 | Intel Corporation | Audio classifier for half duplex communication |
US20010044719A1 (en) | 1999-07-02 | 2001-11-22 | Mitsubishi Electric Research Laboratories, Inc. | Method and system for recognizing, indexing, and searching acoustic signals |
US6453284B1 (en) * | 1999-07-26 | 2002-09-17 | Texas Tech University Health Sciences Center | Multiple voice tracking system and method |
US6480610B1 (en) | 1999-09-21 | 2002-11-12 | Sonic Innovations, Inc. | Subband acoustic feedback cancellation in hearing aids |
US7054809B1 (en) | 1999-09-22 | 2006-05-30 | Mindspeed Technologies, Inc. | Rate selection method for selectable mode vocoder |
US6326912B1 (en) | 1999-09-24 | 2001-12-04 | Akm Semiconductor, Inc. | Analog-to-digital conversion using a multi-bit analog delta-sigma modulator combined with a one-bit digital delta-sigma modulator |
US6594367B1 (en) | 1999-10-25 | 2003-07-15 | Andrea Electronics Corporation | Super directional beamforming design and implementation |
US6757395B1 (en) | 2000-01-12 | 2004-06-29 | Sonic Innovations, Inc. | Noise reduction apparatus and method |
US20010046304A1 (en) | 2000-04-24 | 2001-11-29 | Rast Rodger H. | System and method for selective control of acoustic isolation in headsets |
JP2001318694A (en) | 2000-05-10 | 2001-11-16 | Toshiba Corp | Device and method for signal processing and recording medium |
US7346176B1 (en) | 2000-05-11 | 2008-03-18 | Plantronics, Inc. | Auto-adjust noise canceling microphone with position sensor |
US6377637B1 (en) | 2000-07-12 | 2002-04-23 | Andrea Electronics Corporation | Sub-band exponential smoothing noise canceling system |
US6782253B1 (en) | 2000-08-10 | 2004-08-24 | Koninklijke Philips Electronics N.V. | Mobile micro portal |
WO2002015407A2 (en) | 2000-08-11 | 2002-02-21 | Koninklijke Philips Electronics N.V. | Method and arrangement for synchronizing a σδ-modulator |
JP3566197B2 (en) | 2000-08-31 | 2004-09-15 | 松下電器産業株式会社 | Noise suppression device and noise suppression method |
US7472059B2 (en) | 2000-12-08 | 2008-12-30 | Qualcomm Incorporated | Method and apparatus for robust speech classification |
US20020128839A1 (en) | 2001-01-12 | 2002-09-12 | Ulf Lindgren | Speech bandwidth extension |
US20020097884A1 (en) | 2001-01-25 | 2002-07-25 | Cairns Douglas A. | Variable noise reduction algorithm based on vehicle conditions |
DE50104998D1 (en) | 2001-05-11 | 2005-02-03 | Siemens Ag | METHOD FOR EXPANDING THE BANDWIDTH OF A NARROW-FILTERED LANGUAGE SIGNAL, ESPECIALLY A LANGUAGE SIGNAL SENT BY A TELECOMMUNICATIONS DEVICE |
US6675164B2 (en) | 2001-06-08 | 2004-01-06 | The Regents Of The University Of California | Parallel object-oriented data mining system |
DE60120504T2 (en) | 2001-06-26 | 2006-12-07 | Nokia Corp. | METHOD FOR TRANSCODING AUDIO SIGNALS, NETWORK ELEMENT, WIRELESS COMMUNICATION NETWORK AND COMMUNICATION SYSTEM |
US6876859B2 (en) | 2001-07-18 | 2005-04-05 | Trueposition, Inc. | Method for estimating TDOA and FDOA in a wireless location system |
CA2354808A1 (en) | 2001-08-07 | 2003-02-07 | King Tam | Sub-band adaptive signal processing in an oversampled filterbank |
US6988066B2 (en) | 2001-10-04 | 2006-01-17 | At&T Corp. | Method of bandwidth extension for narrow-band speech |
US6895375B2 (en) | 2001-10-04 | 2005-05-17 | At&T Corp. | System for bandwidth extension of Narrow-band speech |
JP3870193B2 (en) | 2001-11-29 | 2007-01-17 | コーディング テクノロジーズ アクチボラゲット | Encoder, decoder, method and computer program used for high frequency reconstruction |
US8098844B2 (en) | 2002-02-05 | 2012-01-17 | Mh Acoustics, Llc | Dual-microphone spatial noise suppression |
US7050783B2 (en) | 2002-02-22 | 2006-05-23 | Kyocera Wireless Corp. | Accessory detection system |
WO2003084103A1 (en) | 2002-03-22 | 2003-10-09 | Georgia Tech Research Corporation | Analog audio enhancement system using a noise suppression algorithm |
GB2387008A (en) | 2002-03-28 | 2003-10-01 | Qinetiq Ltd | Signal Processing System |
US7072834B2 (en) | 2002-04-05 | 2006-07-04 | Intel Corporation | Adapting to adverse acoustic environment in speech processing using playback training data |
EP2866474A3 (en) | 2002-04-25 | 2015-05-13 | GN Resound A/S | Fitting methodology and hearing prosthesis based on signal-to-noise ratio loss data |
US7257231B1 (en) | 2002-06-04 | 2007-08-14 | Creative Technology Ltd. | Stream segregation for stereo signals |
CA2493105A1 (en) | 2002-07-19 | 2004-01-29 | British Telecommunications Public Limited Company | Method and system for classification of semantic content of audio/video data |
US7539273B2 (en) | 2002-08-29 | 2009-05-26 | Bae Systems Information And Electronic Systems Integration Inc. | Method for separating interfering signals and computing arrival angles |
US7283956B2 (en) | 2002-09-18 | 2007-10-16 | Motorola, Inc. | Noise suppression |
US7657427B2 (en) | 2002-10-11 | 2010-02-02 | Nokia Corporation | Methods and devices for source controlled variable bit-rate wideband speech coding |
KR100477699B1 (en) | 2003-01-15 | 2005-03-18 | 삼성전자주식회사 | Quantization noise shaping method and apparatus |
US7895036B2 (en) | 2003-02-21 | 2011-02-22 | Qnx Software Systems Co. | System for suppressing wind noise |
US7024358B2 (en) | 2003-03-15 | 2006-04-04 | Mindspeed Technologies, Inc. | Recovering an erased voice frame with time warping |
GB2401744B (en) | 2003-05-14 | 2006-02-15 | Ultra Electronics Ltd | An adaptive control unit with feedback compensation |
WO2005004113A1 (en) | 2003-06-30 | 2005-01-13 | Fujitsu Limited | Audio encoding device |
US7245767B2 (en) | 2003-08-21 | 2007-07-17 | Hewlett-Packard Development Company, L.P. | Method and apparatus for object identification, classification or verification |
CA2452945C (en) | 2003-09-23 | 2016-05-10 | Mcmaster University | Binaural adaptive hearing system |
US20050075866A1 (en) | 2003-10-06 | 2005-04-07 | Bernard Widrow | Speech enhancement in the presence of background noise |
US7461003B1 (en) | 2003-10-22 | 2008-12-02 | Tellabs Operations, Inc. | Methods and apparatus for improving the quality of speech signals |
US20060116874A1 (en) | 2003-10-24 | 2006-06-01 | Jonas Samuelsson | Noise-dependent postfiltering |
US7672693B2 (en) | 2003-11-10 | 2010-03-02 | Nokia Corporation | Controlling method, secondary unit and radio terminal equipment |
CN101014997B (en) | 2004-02-18 | 2012-04-04 | 皇家飞利浦电子股份有限公司 | Method and system for generating training data for an automatic speech recogniser |
EP1580882B1 (en) | 2004-03-19 | 2007-01-10 | Harman Becker Automotive Systems GmbH | Audio enhancement system and method |
WO2005106841A1 (en) | 2004-04-28 | 2005-11-10 | Koninklijke Philips Electronics N.V. | Adaptive beamformer, sidelobe canceller, handsfree speech communication device |
US8712768B2 (en) | 2004-05-25 | 2014-04-29 | Nokia Corporation | System and method for enhanced artificial bandwidth expansion |
US20060089836A1 (en) | 2004-10-21 | 2006-04-27 | Motorola, Inc. | System and method of signal pre-conditioning with adaptive spectral tilt compensation for audio equalization |
US7469155B2 (en) | 2004-11-29 | 2008-12-23 | Cisco Technology, Inc. | Handheld communications device with automatic alert mode selection |
GB2422237A (en) | 2004-12-21 | 2006-07-19 | Fluency Voice Technology Ltd | Dynamic coefficients determined from temporally adjacent speech frames |
US8170221B2 (en) | 2005-03-21 | 2012-05-01 | Harman Becker Automotive Systems Gmbh | Audio enhancement system and method |
RU2376657C2 (en) | 2005-04-01 | 2009-12-20 | Квэлкомм Инкорпорейтед | Systems, methods and apparatus for highband time warping |
US8249861B2 (en) | 2005-04-20 | 2012-08-21 | Qnx Software Systems Limited | High frequency compression integration |
US7813931B2 (en) | 2005-04-20 | 2010-10-12 | QNX Software Systems, Co. | System for improving speech quality and intelligibility with bandwidth compression/expansion |
US8280730B2 (en) | 2005-05-25 | 2012-10-02 | Motorola Mobility Llc | Method and apparatus of increasing speech intelligibility in noisy environments |
US20070005351A1 (en) | 2005-06-30 | 2007-01-04 | Sathyendra Harsha M | Method and system for bandwidth expansion for voice communications |
WO2007018293A1 (en) | 2005-08-11 | 2007-02-15 | Asahi Kasei Kabushiki Kaisha | Sound source separating device, speech recognizing device, portable telephone, and sound source separating method, and program |
KR101116363B1 (en) | 2005-08-11 | 2012-03-09 | 삼성전자주식회사 | Method and apparatus for classifying speech signal, and method and apparatus using the same |
US20070041589A1 (en) | 2005-08-17 | 2007-02-22 | Gennum Corporation | System and method for providing environmental specific noise reduction algorithms |
US8326614B2 (en) | 2005-09-02 | 2012-12-04 | Qnx Software Systems Limited | Speech enhancement system |
EP1760696B1 (en) | 2005-09-03 | 2016-02-03 | GN ReSound A/S | Method and apparatus for improved estimation of non-stationary noise for speech enhancement |
US20070053522A1 (en) | 2005-09-08 | 2007-03-08 | Murray Daniel J | Method and apparatus for directional enhancement of speech elements in noisy environments |
WO2007028250A2 (en) | 2005-09-09 | 2007-03-15 | Mcmaster University | Method and device for binaural signal enhancement |
JP4742226B2 (en) | 2005-09-28 | 2011-08-10 | 国立大学法人九州大学 | Active silencing control apparatus and method |
EP1772855B1 (en) | 2005-10-07 | 2013-09-18 | Nuance Communications, Inc. | Method for extending the spectral bandwidth of a speech signal |
US7813923B2 (en) | 2005-10-14 | 2010-10-12 | Microsoft Corporation | Calibration based beamforming, non-linear adaptive filtering, and multi-sensor headset |
US7546237B2 (en) | 2005-12-23 | 2009-06-09 | Qnx Software Systems (Wavemakers), Inc. | Bandwidth extension of narrowband speech |
US8345890B2 (en) | 2006-01-05 | 2013-01-01 | Audience, Inc. | System and method for utilizing inter-microphone level differences for speech enhancement |
US8032369B2 (en) | 2006-01-20 | 2011-10-04 | Qualcomm Incorporated | Arbitrary average data rates for variable rate coders |
US8744844B2 (en) | 2007-07-06 | 2014-06-03 | Audience, Inc. | System and method for adaptive intelligent noise suppression |
US9185487B2 (en) | 2006-01-30 | 2015-11-10 | Audience, Inc. | System and method for providing noise suppression utilizing null processing noise subtraction |
US8194880B2 (en) | 2006-01-30 | 2012-06-05 | Audience, Inc. | System and method for utilizing omni-directional microphones for speech enhancement |
EP1993320B1 (en) | 2006-03-03 | 2015-01-07 | Nippon Telegraph And Telephone Corporation | Reverberation removal device, reverberation removal method, reverberation removal program, and recording medium |
EP1994788B1 (en) | 2006-03-10 | 2014-05-07 | MH Acoustics, LLC | Noise-reducing directional microphone array |
US8180067B2 (en) | 2006-04-28 | 2012-05-15 | Harman International Industries, Incorporated | System for selectively extracting components of an audio input signal |
US8150065B2 (en) * | 2006-05-25 | 2012-04-03 | Audience, Inc. | System and method for processing an audio signal |
US8204253B1 (en) | 2008-06-30 | 2012-06-19 | Audience, Inc. | Self calibration of audio device |
US20070299655A1 (en) | 2006-06-22 | 2007-12-27 | Nokia Corporation | Method, Apparatus and Computer Program Product for Providing Low Frequency Expansion of Speech |
ATE450987T1 (en) | 2006-06-23 | 2009-12-15 | Gn Resound As | HEARING INSTRUMENT WITH ADAPTIVE DIRECTIONAL SIGNAL PROCESSING |
JP4836720B2 (en) | 2006-09-07 | 2011-12-14 | 株式会社東芝 | Noise suppressor |
KR101061415B1 (en) | 2006-09-14 | 2011-09-01 | 엘지전자 주식회사 | Controller and user interface for dialogue enhancement techniques |
DE102006051071B4 (en) | 2006-10-30 | 2010-12-16 | Siemens Audiologische Technik Gmbh | Level-dependent noise reduction |
EP1933303B1 (en) | 2006-12-14 | 2008-08-06 | Harman/Becker Automotive Systems GmbH | Speech dialog control based on signal pre-processing |
US7986794B2 (en) | 2007-01-11 | 2011-07-26 | Fortemedia, Inc. | Small array microphone apparatus and beam forming method thereof |
JP5401760B2 (en) | 2007-02-05 | 2014-01-29 | ソニー株式会社 | Headphone device, audio reproduction system, and audio reproduction method |
JP4882773B2 (en) | 2007-02-05 | 2012-02-22 | ソニー株式会社 | Signal processing apparatus and signal processing method |
US8060363B2 (en) | 2007-02-13 | 2011-11-15 | Nokia Corporation | Audio signal encoding |
BRPI0807703B1 (en) | 2007-02-26 | 2020-09-24 | Dolby Laboratories Licensing Corporation | METHOD FOR IMPROVING SPEECH IN ENTERTAINMENT AUDIO AND COMPUTER-READABLE NON-TRANSITIONAL MEDIA |
US20080208575A1 (en) | 2007-02-27 | 2008-08-28 | Nokia Corporation | Split-band encoding and decoding of an audio signal |
KR100905585B1 (en) | 2007-03-02 | 2009-07-02 | 삼성전자주식회사 | Method and apparatus for controling bandwidth extension of vocal signal |
EP1970900A1 (en) | 2007-03-14 | 2008-09-17 | Harman Becker Automotive Systems GmbH | Method and apparatus for providing a codebook for bandwidth extension of an acoustic signal |
TWI421856B (en) | 2007-03-19 | 2014-01-01 | Dolby Lab Licensing Corp | Speech enhancement employing a perceptual model |
US8005238B2 (en) | 2007-03-22 | 2011-08-23 | Microsoft Corporation | Robust adaptive beamforming with enhanced noise suppression |
US7873114B2 (en) | 2007-03-29 | 2011-01-18 | Motorola Mobility, Inc. | Method and apparatus for quickly detecting a presence of abrupt noise and updating a noise estimate |
US8180062B2 (en) | 2007-05-30 | 2012-05-15 | Nokia Corporation | Spatial sound zooming |
JP4455614B2 (en) | 2007-06-13 | 2010-04-21 | 株式会社東芝 | Acoustic signal processing method and apparatus |
US8428275B2 (en) | 2007-06-22 | 2013-04-23 | Sanyo Electric Co., Ltd. | Wind noise reduction device |
US8140331B2 (en) | 2007-07-06 | 2012-03-20 | Xia Lou | Feature extraction for identification and classification of audio signals |
US7817808B2 (en) | 2007-07-19 | 2010-10-19 | Alon Konchitsky | Dual adaptive structure for speech enhancement |
US7856353B2 (en) | 2007-08-07 | 2010-12-21 | Nuance Communications, Inc. | Method for processing speech signal data with reverberation filtering |
US20090043577A1 (en) | 2007-08-10 | 2009-02-12 | Ditech Networks, Inc. | Signal presence detection using bi-directional communication data |
ATE448649T1 (en) | 2007-08-13 | 2009-11-15 | Harman Becker Automotive Sys | NOISE REDUCTION USING A COMBINATION OF BEAM SHAPING AND POST-FILTERING |
ATE501506T1 (en) | 2007-09-12 | 2011-03-15 | Dolby Lab Licensing Corp | VOICE EXTENSION WITH ADJUSTMENT OF NOISE LEVEL ESTIMATES |
JP5302968B2 (en) | 2007-09-12 | 2013-10-02 | ドルビー ラボラトリーズ ライセンシング コーポレイション | Speech improvement with speech clarification |
DE602007008429D1 (en) | 2007-10-01 | 2010-09-23 | Harman Becker Automotive Sys | Efficient sub-band audio signal processing, method, apparatus and associated computer program |
CN101617245B (en) | 2007-10-01 | 2012-10-10 | 松下电器产业株式会社 | Sounnd source direction detector |
US8107631B2 (en) | 2007-10-04 | 2012-01-31 | Creative Technology Ltd | Correlation-based method for ambience extraction from two-channel audio signals |
US20090095804A1 (en) | 2007-10-12 | 2009-04-16 | Sony Ericsson Mobile Communications Ab | Rfid for connected accessory identification and method |
US8046219B2 (en) | 2007-10-18 | 2011-10-25 | Motorola Mobility, Inc. | Robust two microphone noise suppression system |
US8606566B2 (en) | 2007-10-24 | 2013-12-10 | Qnx Software Systems Limited | Speech enhancement through partial speech reconstruction |
DE602007004504D1 (en) | 2007-10-29 | 2010-03-11 | Harman Becker Automotive Sys | Partial language reconstruction |
EP2058804B1 (en) | 2007-10-31 | 2016-12-14 | Nuance Communications, Inc. | Method for dereverberation of an acoustic signal and system thereof |
KR101444100B1 (en) | 2007-11-15 | 2014-09-26 | 삼성전자주식회사 | Noise cancelling method and apparatus from the mixed sound |
US20090150144A1 (en) | 2007-12-10 | 2009-06-11 | Qnx Software Systems (Wavemakers), Inc. | Robust voice detector for receive-side automatic gain control |
US8175291B2 (en) | 2007-12-19 | 2012-05-08 | Qualcomm Incorporated | Systems, methods, and apparatus for multi-microphone based speech enhancement |
EP2232704A4 (en) | 2007-12-20 | 2010-12-01 | Ericsson Telefon Ab L M | Noise suppression method and apparatus |
US8554551B2 (en) | 2008-01-28 | 2013-10-08 | Qualcomm Incorporated | Systems, methods, and apparatus for context replacement by audio level |
US8223988B2 (en) | 2008-01-29 | 2012-07-17 | Qualcomm Incorporated | Enhanced blind source separation algorithm for highly correlated mixtures |
US8194882B2 (en) | 2008-02-29 | 2012-06-05 | Audience, Inc. | System and method for providing single microphone noise suppression fallback |
US8355511B2 (en) | 2008-03-18 | 2013-01-15 | Audience, Inc. | System and method for envelope-based acoustic echo cancellation |
US8374854B2 (en) | 2008-03-28 | 2013-02-12 | Southern Methodist University | Spatio-temporal speech enhancement technique based on generalized eigenvalue decomposition |
US9197181B2 (en) | 2008-05-12 | 2015-11-24 | Broadcom Corporation | Loudness enhancement system and method |
US8831936B2 (en) | 2008-05-29 | 2014-09-09 | Qualcomm Incorporated | Systems, methods, apparatus, and computer program products for speech signal processing using spectral contrast enhancement |
US20090315708A1 (en) | 2008-06-19 | 2009-12-24 | John Walley | Method and system for limiting audio output in audio headsets |
US9253568B2 (en) | 2008-07-25 | 2016-02-02 | Broadcom Corporation | Single-microphone wind noise suppression |
EP2151822B8 (en) | 2008-08-05 | 2018-10-24 | Fraunhofer-Gesellschaft zur Förderung der angewandten Forschung e.V. | Apparatus and method for processing an audio signal for speech enhancement using a feature extraction |
EP2321978A4 (en) | 2008-08-29 | 2013-01-23 | Dev Audio Pty Ltd | A microphone array system and method for sound acquisition |
US8392181B2 (en) | 2008-09-10 | 2013-03-05 | Texas Instruments Incorporated | Subtraction of a shaped component of a noise reduction spectrum from a combined signal |
EP2164066B1 (en) | 2008-09-15 | 2016-03-09 | Oticon A/S | Noise spectrum tracking in noisy acoustical signals |
WO2010033387A2 (en) | 2008-09-19 | 2010-03-25 | Dolby Laboratories Licensing Corporation | Upstream signal processing for client devices in a small-cell wireless network |
TWI398178B (en) | 2008-09-25 | 2013-06-01 | Skyphy Networks Ltd | Multi-hop wireless systems having noise reduction and bandwidth expansion capabilities and the methods of the same |
US20100082339A1 (en) | 2008-09-30 | 2010-04-01 | Alon Konchitsky | Wind Noise Reduction |
US8724829B2 (en) | 2008-10-24 | 2014-05-13 | Qualcomm Incorporated | Systems, methods, apparatus, and computer-readable media for coherence detection |
US8218397B2 (en) | 2008-10-24 | 2012-07-10 | Qualcomm Incorporated | Audio source proximity estimation using sensor array for noise reduction |
US8111843B2 (en) | 2008-11-11 | 2012-02-07 | Motorola Solutions, Inc. | Compensation for nonuniform delayed group communications |
US8243952B2 (en) | 2008-12-22 | 2012-08-14 | Conexant Systems, Inc. | Microphone array calibration method and apparatus |
DK2211339T3 (en) | 2009-01-23 | 2017-08-28 | Oticon As | listening System |
JP4892021B2 (en) | 2009-02-26 | 2012-03-07 | 株式会社東芝 | Signal band expander |
US8359195B2 (en) | 2009-03-26 | 2013-01-22 | LI Creative Technologies, Inc. | Method and apparatus for processing audio and speech signals |
US8144890B2 (en) | 2009-04-28 | 2012-03-27 | Bose Corporation | ANR settings boot loading |
US8611553B2 (en) | 2010-03-30 | 2013-12-17 | Bose Corporation | ANR instability detection |
US8184822B2 (en) | 2009-04-28 | 2012-05-22 | Bose Corporation | ANR signal processing topology |
US8071869B2 (en) | 2009-05-06 | 2011-12-06 | Gracenote, Inc. | Apparatus and method for determining a prominent tempo of an audio work |
US8160265B2 (en) | 2009-05-18 | 2012-04-17 | Sony Computer Entertainment Inc. | Method and apparatus for enhancing the generation of three-dimensional sound in headphone devices |
US8737636B2 (en) | 2009-07-10 | 2014-05-27 | Qualcomm Incorporated | Systems, methods, apparatus, and computer-readable media for adaptive active noise cancellation |
US7769187B1 (en) | 2009-07-14 | 2010-08-03 | Apple Inc. | Communications circuits for electronic devices and accessories |
US8571231B2 (en) | 2009-10-01 | 2013-10-29 | Qualcomm Incorporated | Suppressing noise in an audio signal |
US20110099010A1 (en) | 2009-10-22 | 2011-04-28 | Broadcom Corporation | Multi-channel noise suppression system |
US8244927B2 (en) | 2009-10-27 | 2012-08-14 | Fairchild Semiconductor Corporation | Method of detecting accessories on an audio jack |
US8526628B1 (en) | 2009-12-14 | 2013-09-03 | Audience, Inc. | Low latency active noise cancellation system |
US8848935B1 (en) | 2009-12-14 | 2014-09-30 | Audience, Inc. | Low latency active noise cancellation system |
US8385559B2 (en) | 2009-12-30 | 2013-02-26 | Robert Bosch Gmbh | Adaptive digital noise canceller |
US9008329B1 (en) | 2010-01-26 | 2015-04-14 | Audience, Inc. | Noise reduction using multi-feature cluster tracker |
US8700391B1 (en) | 2010-04-01 | 2014-04-15 | Audience, Inc. | Low complexity bandwidth expansion of speech |
KR20130038857A (en) | 2010-04-09 | 2013-04-18 | 디티에스, 인코포레이티드 | Adaptive environmental noise compensation for audio playback |
US8958572B1 (en) | 2010-04-19 | 2015-02-17 | Audience, Inc. | Adaptive noise cancellation for multi-microphone systems |
US8538035B2 (en) | 2010-04-29 | 2013-09-17 | Audience, Inc. | Multi-microphone robust noise suppression |
US8473287B2 (en) | 2010-04-19 | 2013-06-25 | Audience, Inc. | Method for jointly optimizing noise reduction and voice quality in a mono or multi-microphone system |
US8606571B1 (en) | 2010-04-19 | 2013-12-10 | Audience, Inc. | Spatial selectivity noise reduction tradeoff for multi-microphone systems |
US8781137B1 (en) | 2010-04-27 | 2014-07-15 | Audience, Inc. | Wind noise detection and suppression |
US8447595B2 (en) | 2010-06-03 | 2013-05-21 | Apple Inc. | Echo-related decisions on automatic gain control of uplink speech signal in a communications device |
US8515089B2 (en) | 2010-06-04 | 2013-08-20 | Apple Inc. | Active noise cancellation decisions in a portable audio device |
US8447596B2 (en) | 2010-07-12 | 2013-05-21 | Audience, Inc. | Monaural noise suppression based on computational auditory scene analysis |
US8719475B2 (en) | 2010-07-13 | 2014-05-06 | Broadcom Corporation | Method and system for utilizing low power superspeed inter-chip (LP-SSIC) communications |
US8761410B1 (en) | 2010-08-12 | 2014-06-24 | Audience, Inc. | Systems and methods for multi-channel dereverberation |
US8611552B1 (en) | 2010-08-25 | 2013-12-17 | Audience, Inc. | Direction-aware active noise cancellation system |
US8447045B1 (en) | 2010-09-07 | 2013-05-21 | Audience, Inc. | Multi-microphone active noise cancellation system |
US9049532B2 (en) | 2010-10-19 | 2015-06-02 | Electronics And Telecommunications Research Instittute | Apparatus and method for separating sound source |
US8682006B1 (en) | 2010-10-20 | 2014-03-25 | Audience, Inc. | Noise suppression based on null coherence |
US8311817B2 (en) | 2010-11-04 | 2012-11-13 | Audience, Inc. | Systems and methods for enhancing voice quality in mobile device |
CN102486920A (en) | 2010-12-06 | 2012-06-06 | 索尼公司 | Audio event detection method and device |
US9229833B2 (en) | 2011-01-28 | 2016-01-05 | Fairchild Semiconductor Corporation | Successive approximation resistor detection |
JP5817366B2 (en) | 2011-09-12 | 2015-11-18 | 沖電気工業株式会社 | Audio signal processing apparatus, method and program |
-
2010
- 2010-08-20 US US12/860,043 patent/US8447596B2/en active Active
-
2011
- 2011-05-19 JP JP2013519682A patent/JP2013534651A/en not_active Ceased
- 2011-05-19 KR KR1020137000488A patent/KR20130117750A/en not_active IP Right Cessation
- 2011-05-19 WO PCT/US2011/037250 patent/WO2012009047A1/en active Application Filing
- 2011-05-30 TW TW100118902A patent/TW201214418A/en unknown
-
2013
- 2013-04-09 US US13/859,186 patent/US9431023B2/en active Active
Patent Citations (10)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US7065486B1 (en) * | 2002-04-11 | 2006-06-20 | Mindspeed Technologies, Inc. | Linear prediction based noise suppression |
US7574352B2 (en) * | 2002-09-06 | 2009-08-11 | Massachusetts Institute Of Technology | 2-D processing of speech |
US20050049857A1 (en) * | 2003-08-25 | 2005-03-03 | Microsoft Corporation | Method and apparatus using harmonic-model-based front end for robust speech recognition |
US7516067B2 (en) * | 2003-08-25 | 2009-04-07 | Microsoft Corporation | Method and apparatus using harmonic-model-based front end for robust speech recognition |
US7725314B2 (en) * | 2004-02-16 | 2010-05-25 | Microsoft Corporation | Method and apparatus for constructing a speech filter using estimates of clean speech and noise |
US7254535B2 (en) * | 2004-06-30 | 2007-08-07 | Motorola, Inc. | Method and apparatus for equalizing a speech signal generated within a pressurized air delivery system |
US7925502B2 (en) * | 2007-03-01 | 2011-04-12 | Microsoft Corporation | Pitch model for noise estimation |
US20080228474A1 (en) * | 2007-03-16 | 2008-09-18 | Spreadtrum Communications Corporation | Methods and apparatus for post-processing of speech signals |
US20090228272A1 (en) * | 2007-11-12 | 2009-09-10 | Tobias Herbig | System for distinguishing desired audio signals from noise |
US20100094622A1 (en) * | 2008-10-10 | 2010-04-15 | Nexidia Inc. | Feature normalization for speech and audio processing |
Cited By (50)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US9502048B2 (en) | 2010-04-19 | 2016-11-22 | Knowles Electronics, Llc | Adaptively reducing noise to limit speech distortion |
US9699554B1 (en) | 2010-04-21 | 2017-07-04 | Knowles Electronics, Llc | Adaptive signal equalization |
US9343056B1 (en) | 2010-04-27 | 2016-05-17 | Knowles Electronics, Llc | Wind noise detection and suppression |
US9438992B2 (en) | 2010-04-29 | 2016-09-06 | Knowles Electronics, Llc | Multi-microphone robust noise suppression |
US9558755B1 (en) | 2010-05-20 | 2017-01-31 | Knowles Electronics, Llc | Noise suppression assisted automatic speech recognition |
US9431023B2 (en) | 2010-07-12 | 2016-08-30 | Knowles Electronics, Llc | Monaural noise suppression based on computational auditory scene analysis |
US9640194B1 (en) | 2012-10-04 | 2017-05-02 | Knowles Electronics, Llc | Noise suppression for speech processing based on machine-learning mask estimation |
US10339952B2 (en) | 2013-03-13 | 2019-07-02 | Kopin Corporation | Apparatuses and systems for acoustic channel auto-balancing during multi-channel signal extraction |
US10306389B2 (en) | 2013-03-13 | 2019-05-28 | Kopin Corporation | Head wearable acoustic system with noise canceling microphone geometry apparatuses and methods |
US9830905B2 (en) | 2013-06-26 | 2017-11-28 | Qualcomm Incorporated | Systems and methods for feature extraction |
US9679555B2 (en) | 2013-06-26 | 2017-06-13 | Qualcomm Incorporated | Systems and methods for measuring speech signal quality |
US9530434B1 (en) * | 2013-07-18 | 2016-12-27 | Knuedge Incorporated | Reducing octave errors during pitch determination for noisy audio signals |
US9508345B1 (en) | 2013-09-24 | 2016-11-29 | Knowles Electronics, Llc | Continuous voice sensing |
US9959886B2 (en) * | 2013-12-06 | 2018-05-01 | Malaspina Labs (Barbados), Inc. | Spectral comb voice activity detection |
US20150162021A1 (en) * | 2013-12-06 | 2015-06-11 | Malaspina Labs (Barbados), Inc. | Spectral Comb Voice Activity Detection |
US9953634B1 (en) | 2013-12-17 | 2018-04-24 | Knowles Electronics, Llc | Passive training for automatic speech recognition |
US9437188B1 (en) | 2014-03-28 | 2016-09-06 | Knowles Electronics, Llc | Buffered reprocessing for multi-microphone automatic speech recognition assist |
US9378755B2 (en) * | 2014-05-30 | 2016-06-28 | Apple Inc. | Detecting a user's voice activity using dynamic probabilistic models of speech features |
US20150348572A1 (en) * | 2014-05-30 | 2015-12-03 | Apple Inc. | Detecting a user's voice activity using dynamic probabilistic models of speech features |
US9799330B2 (en) | 2014-08-28 | 2017-10-24 | Knowles Electronics, Llc | Multi-sourced noise suppression |
US20170110142A1 (en) * | 2015-10-18 | 2017-04-20 | Kopin Corporation | Apparatuses and methods for enhanced speech recognition in variable environments |
US11631421B2 (en) * | 2015-10-18 | 2023-04-18 | Solos Technology Limited | Apparatuses and methods for enhanced speech recognition in variable environments |
US10529317B2 (en) * | 2015-11-06 | 2020-01-07 | Samsung Electronics Co., Ltd. | Neural network training apparatus and method, and speech recognition apparatus and method |
US20170133006A1 (en) * | 2015-11-06 | 2017-05-11 | Samsung Electronics Co., Ltd. | Neural network training apparatus and method, and speech recognition apparatus and method |
US9654861B1 (en) | 2015-11-13 | 2017-05-16 | Doppler Labs, Inc. | Annoyance noise suppression |
US11218796B2 (en) | 2015-11-13 | 2022-01-04 | Dolby Laboratories Licensing Corporation | Annoyance noise suppression |
US10841688B2 (en) | 2015-11-13 | 2020-11-17 | Dolby Laboratories Licensing Corporation | Annoyance noise suppression |
US10595117B2 (en) | 2015-11-13 | 2020-03-17 | Dolby Laboratories Licensing Corporation | Annoyance noise suppression |
WO2017082974A1 (en) * | 2015-11-13 | 2017-05-18 | Doppler Labs, Inc. | Annoyance noise suppression |
US10045115B2 (en) | 2015-11-13 | 2018-08-07 | Dolby Laboratories Licensing Corporation | Annoyance noise suppression |
US10531178B2 (en) | 2015-11-13 | 2020-01-07 | Dolby Laboratories Licensing Corporation | Annoyance noise suppression |
US9589574B1 (en) | 2015-11-13 | 2017-03-07 | Doppler Labs, Inc. | Annoyance noise suppression |
US10853025B2 (en) | 2015-11-25 | 2020-12-01 | Dolby Laboratories Licensing Corporation | Sharing of custom audio processing parameters |
US10275210B2 (en) | 2015-11-25 | 2019-04-30 | Dolby Laboratories Licensing Corporation | Privacy protection in collective feedforward |
US9584899B1 (en) | 2015-11-25 | 2017-02-28 | Doppler Labs, Inc. | Sharing of custom audio processing parameters |
US9678709B1 (en) | 2015-11-25 | 2017-06-13 | Doppler Labs, Inc. | Processing sound using collective feedforward |
US11145320B2 (en) | 2015-11-25 | 2021-10-12 | Dolby Laboratories Licensing Corporation | Privacy protection in collective feedforward |
US9703524B2 (en) | 2015-11-25 | 2017-07-11 | Doppler Labs, Inc. | Privacy protection in collective feedforward |
US9769553B2 (en) | 2015-11-25 | 2017-09-19 | Doppler Labs, Inc. | Adaptive filtering with machine learning |
US10275209B2 (en) | 2015-11-25 | 2019-04-30 | Dolby Laboratories Licensing Corporation | Sharing of custom audio processing parameters |
US10997987B2 (en) * | 2017-06-16 | 2021-05-04 | Nxp B.V. | Signal processor for speech enhancement and recognition by using two output terminals designated for noise reduction |
CN107331406A (en) * | 2017-07-03 | 2017-11-07 | 福建星网智慧软件有限公司 | A kind of method of dynamic adjustment Echo-delay |
US11004463B2 (en) * | 2017-09-25 | 2021-05-11 | Fujitsu Limited | Speech processing method, apparatus, and non-transitory computer-readable storage medium for storing a computer program for pitch frequency detection based upon a learned value |
US20190096432A1 (en) * | 2017-09-25 | 2019-03-28 | Fujitsu Limited | Speech processing method, speech processing apparatus, and non-transitory computer-readable storage medium for storing speech processing computer program |
US11029914B2 (en) | 2017-09-29 | 2021-06-08 | Knowles Electronics, Llc | Multi-core audio processor with phase coherency |
WO2019067335A1 (en) * | 2017-09-29 | 2019-04-04 | Knowles Electronics, Llc | Multi-core audio processor with phase coherency |
CN108806708A (en) * | 2018-06-13 | 2018-11-13 | 中国电子科技集团公司第三研究所 | Voice de-noising method based on Computational auditory scene analysis and generation confrontation network model |
US10891954B2 (en) | 2019-01-03 | 2021-01-12 | International Business Machines Corporation | Methods and systems for managing voice response systems based on signals from external devices |
CN111883154A (en) * | 2020-07-17 | 2020-11-03 | 海尔优家智能科技(北京)有限公司 | Echo cancellation method and apparatus, computer-readable storage medium, and electronic apparatus |
EP4198975A1 (en) * | 2021-12-16 | 2023-06-21 | GN Hearing A/S | Electronic device and method for obtaining a user's speech in a first sound signal |
Also Published As
Publication number | Publication date |
---|---|
JP2013534651A (en) | 2013-09-05 |
KR20130117750A (en) | 2013-10-28 |
TW201214418A (en) | 2012-04-01 |
US8447596B2 (en) | 2013-05-21 |
US20130231925A1 (en) | 2013-09-05 |
WO2012009047A1 (en) | 2012-01-19 |
US9431023B2 (en) | 2016-08-30 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US9431023B2 (en) | Monaural noise suppression based on computational auditory scene analysis | |
US9438992B2 (en) | Multi-microphone robust noise suppression | |
US9502048B2 (en) | Adaptively reducing noise to limit speech distortion | |
US9558755B1 (en) | Noise suppression assisted automatic speech recognition | |
US8880396B1 (en) | Spectrum reconstruction for automatic speech recognition | |
US8521530B1 (en) | System and method for enhancing a monaural audio signal | |
US8718290B2 (en) | Adaptive noise reduction using level cues | |
US9064498B2 (en) | Apparatus and method for processing an audio signal for speech enhancement using a feature extraction | |
US8958572B1 (en) | Adaptive noise cancellation for multi-microphone systems | |
US8682006B1 (en) | Noise suppression based on null coherence | |
Jung et al. | Noise Reduction after RIR removal for Speech De-reverberation and De-noising | |
Yang et al. | Environment-Aware Reconfigurable Noise Suppression | |
Vashkevich et al. | Speech enhancement in a smartphone-based hearing aid | |
CN117219102A (en) | Low-complexity voice enhancement method based on auditory perception |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
AS | Assignment |
Owner name: AUDIENCE, INC., CALIFORNIA Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:AVENDANO, CARLOS;LAROCHE, JEAN;GOODWIN, MICHAEL M.;AND OTHERS;SIGNING DATES FROM 20100916 TO 20100920;REEL/FRAME:025244/0475 |
|
STCF | Information on status: patent grant |
Free format text: PATENTED CASE |
|
FEPP | Fee payment procedure |
Free format text: PAT HOLDER NO LONGER CLAIMS SMALL ENTITY STATUS, ENTITY STATUS SET TO UNDISCOUNTED (ORIGINAL EVENT CODE: STOL); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY |
|
AS | Assignment |
Owner name: KNOWLES ELECTRONICS, LLC, ILLINOIS Free format text: MERGER;ASSIGNOR:AUDIENCE LLC;REEL/FRAME:037927/0435 Effective date: 20151221 Owner name: AUDIENCE LLC, CALIFORNIA Free format text: CHANGE OF NAME;ASSIGNOR:AUDIENCE, INC.;REEL/FRAME:037927/0424 Effective date: 20151217 |
|
FPAY | Fee payment |
Year of fee payment: 4 |
|
MAFP | Maintenance fee payment |
Free format text: PAYMENT OF MAINTENANCE FEE, 8TH YEAR, LARGE ENTITY (ORIGINAL EVENT CODE: M1552); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY Year of fee payment: 8 |
|
AS | Assignment |
Owner name: SAMSUNG ELECTRONICS CO., LTD., KOREA, REPUBLIC OF Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:KNOWLES ELECTRONICS, LLC;REEL/FRAME:066216/0142 Effective date: 20231219 |