WO2000068936A1 - Cancellation of non-stationary interfering signals for speech recognition - Google Patents
Cancellation of non-stationary interfering signals for speech recognition Download PDFInfo
- Publication number
- WO2000068936A1 WO2000068936A1 PCT/GB2000/001715 GB0001715W WO0068936A1 WO 2000068936 A1 WO2000068936 A1 WO 2000068936A1 GB 0001715 W GB0001715 W GB 0001715W WO 0068936 A1 WO0068936 A1 WO 0068936A1
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- acoustic
- channel
- signal
- stereo
- signals
- Prior art date
Links
Classifications
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10K—SOUND-PRODUCING DEVICES; METHODS OR DEVICES FOR PROTECTING AGAINST, OR FOR DAMPING, NOISE OR OTHER ACOUSTIC WAVES IN GENERAL; ACOUSTICS NOT OTHERWISE PROVIDED FOR
- G10K11/00—Methods or devices for transmitting, conducting or directing sound in general; Methods or devices for protecting against, or for damping, noise or other acoustic waves in general
- G10K11/16—Methods or devices for protecting against, or for damping, noise or other acoustic waves in general
- G10K11/175—Methods or devices for protecting against, or for damping, noise or other acoustic waves in general using interference effects; Masking sound
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS OR SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING; SPEECH OR AUDIO CODING OR DECODING
- G10L15/00—Speech recognition
- G10L15/20—Speech recognition techniques specially adapted for robustness in adverse environments, e.g. in noise, of stress induced speech
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS OR SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING; SPEECH OR AUDIO CODING OR DECODING
- G10L21/00—Processing of the speech or voice signal to produce another audible or non-audible signal, e.g. visual or tactile, in order to modify its quality or its intelligibility
- G10L21/02—Speech enhancement, e.g. noise reduction or echo cancellation
- G10L21/0208—Noise filtering
Definitions
- This invention relates to apparatus and method for cancellation of non-stationary interfering signals.
- the invention relates to cancellation of such signals for the purpose of recovering a wanted speech signal for use by a speech recognition application.
- the invention is especially suitable for use in an automobile where in-car devices produce interfering signals during the speech recognition process.
- a problem associated with speech recognition is that of maintaining performance in the presence of interfering signals so that the speech recognition process continues to function satisfactorily even in the presence of background noise.
- Known systems have been directed towards mitigating effects of quasi-stationary noise such as telephone channel noise or car noise.
- Proposed solutions to quasi-stationary noise interference include spectral subtraction, für filtering and parallel model combination, each of which work in the spectral domain.
- ECAD Electronic in-car Acoustic Devices
- Sound generated by ECAD could be present when a user wishes to control a device using a voice command.
- a radio may be playing in a car when the user wants to use voice control of a navigation system or the radio itself.
- the original interfering signal produced by the radio is assumed to be known and accessible but has passed through an unknown acoustic path between the radio's loudspeakers and the speech recognition system's microphone.
- the acoustic path may be determined by the position of the loudspeakers and the microphone inside the car as well as other factors, such as the number of passengers and the presence of luggage inside the car.
- Known systems which attempt to overcome the problem of non-stationary interferers have been based on time domain adaptive filters.
- adaptive filtering may produce satisfactory results, this approach suffers from a number of disadvantages.
- Such disadvantages include high computational requirements and slow convergence of adaptive filtering algorithms.
- Simple forms of adaptive filtering may require order 3N computations per sample.
- Such high computational requirements can mean that complex hardware may be required in order to perform the necessary filtering, thereby increasing costs of devices incorporating such technology to the consumer.
- apparatus for cancellation of one or more non- stationary interfering signals for speech recognition comprising: means for receiving an acoustic signal; means for generating an estimated value of a magnitude spectrum of said non- stationary interfering signals; and means for subtracting said estimated value from said received acoustic signal to produce a representation of a wanted speech magnitude spectrum.
- said means for generating estimated value includes processing means configured to estimate a transfer function for an acoustic channel between each source of said non-stationary interfering signals and said means for receiving an acoustic signal .
- said processing means is configured to estimate transfer functions for non-stationary interfering signals produced by left and right stereo channel transmissions .
- said estimation of said transfer functions is achieved by said processing means executing an iterative algorithm on a frame-by- frame basis, the frames being constituted by successive time periods.
- said processing means is configured to estimate magnitudes of said left and right channel interfer- ence signals, said magnitude of left channel interference signal estimated by subtracting said right channel interference signal magnitude estimated during previous said iteration from said acoustic signal received at current said iteration; and said magnitude of right channel interference signal is estimated by subtracting said left channel interference signal magnitude estimated during previous said iteration from said acoustic signal received at current said iteration.
- said transfer function estimate for said right stereo acoustic channel is determined by dividing said right channel interference magnitude estimate by said interfering signal transmitted from said right acoustic stereo channel; and said transfer function estimate for said left stereo acoustic channel is determined by dividing said left channel interference magnitude estimate by said interfering signal transmitted from said left acoustic stereo channel.
- said right acoustic channel transfer function estimation is performed for a said iteration only if a ratio of total energy of said right acoustic stereo channel interfering signal over total energy of said left acoustic stereo interfering channel exceeds a predetermined threshold value; and said left acoustic channel transfer function estimation is performed for a said iteration only if a ratio of total energy of said left acoustic stereo channel interfering signal over total energy of said right acoustic stereo channel interfering signal exceeds a predetermined threshold value .
- said ratio and threshold comparisons are applied to individual frequency components in spectra of said signals.
- said left and right stereo acoustic channel transfer functions are multiplied by (l-
- said transfer function estimate for said right stereo acoustic channel is obtained using an expression:
- processing means further comprises means for smoothing said estimated transfer functions in time domain.
- means for smoothing in time domain comprises a first order recursive filter.
- said processing means further comprises means for smoothing said estimated transfer functions in frequency domain.
- said means for smoothing in frequency domain comprises a Finite Impulse Response filter.
- said processing means includes means for performing a Fourier Transform.
- said non-stationary interfering signals are produced by an electronic acoustic device operating in a vehicle .
- said means for receiving an acoustic signal comprises a microphone.
- a method of cancellation of one or more non-stationary interfering signals for speech recognition comprising steps of: receiving an acoustic signal; generating an estimated value for a magnitude spectrum of said non-stationary interfering signal; and subtracting said estimated value from said received acoustic signal to produce a representation of a wanted speech magnitude spectrum.
- said step of generating an estimated value comprises estimating a transfer function for an acoustic channel between each source of said non-stationary interfering signals and said means for receiving an acoustic signal.
- said transfer functions are estimated for non-stationary interfering signals produced by left and right stereo channel transmissions.
- said step of generating an estimated value is executed iteratively on a frame-by- frame basis.
- said step of estimating a transfer function includes : estimating a magnitude of said left channel interference signal by subtracting said right channel interference signal magnitude estimated during previous said iteration from said acoustic signal received at current said iterations; and estimating magnitude of said right channel interference signal by substracting said left channel interference signal magnitude estimated during previous said iteration from said acoustic signal received at current said iteration.
- the method may further comprise steps of : dividing said right channel interference magnitude estimate by said interfering signal transmitted from said right acoustic stereo channel; and dividing said left channel interference magnitude estimated by said interfering signal transmitted from said left acoustic stereo channel.
- said step of estimating right acoustic channel transfer function is performed for a said iteration only if a ratio of total energy of said right acoustic stereo channel interfering signal over total energy of said left acoustic stereo channel interfering signal exceeds a predetermined threshold value; and said step of estimating left acoustic channel transfer function estimate is performed for a said iteration only if a ratio of total energy of said left acoustic stereo channel interfering signal over total energy of said right acoustic stereo channel interfering signal exceeds a predetermined threshold value.
- said ratio and threshold comparisons are applied to individual frequency components in spectra of said signals.
- said left and right stereo acoustic channel transfer functions are multiplied by (l-
- said transfer function estimate for said right stereo acoustic channel is obtained using an expression:
- this aspect may be comprising a step of smoothing said estimated transfer functions in time domain.
- this aspect may be further comprising a step of smoothing said estimated transfer functions in frequency domain.
- a speech recognition system including apparatus according to the first aspect of the invention.
- an electronic acoustic device including apparatus according to the first aspect of the invention.
- the present invention is a frequency domain (rather than time domain as used in known systems) technique solution which is preferably based on channel identification followed by spectral subtraction.
- Embodiments of the present application's system can substantially improve performance of a speech recognition system when non-stationary interferers are present whilst having the advantage of lower computational requirement than known systems.
- Embodiments of the present application's system provide levels of non-stationary interferer cancellation sufficient to substantially improve the performance of a speech recognition system, typically about 10 decibels of cancellation is possible in the case where loud background music is being output by ECAD. Such levels of cancellation may not be satisfactory to a human listener, however. For the purposes of speech recognition applications, such levels of cancellation will substantially improve the system's performance.
- a human listener is sensitive to levels of interference 40 decibels below the level of wanted signal, whilst known speech recognition systems can operate well with a 15 decibel signal-to-noise ratio.
- the interfering signal output by an ECAD such as a radio may be a mono or stereo transmission, typically being output from two loudspeakers located at separate locations within an automobile.
- an ECAD such as a radio
- a phase of the interferer signal is not required at the speech recognition system, as recognition feature sets such as cepstra do not normally contain phase information.
- Figure 1 illustrates schematically an example of an automobile environment having an ECAD where a speech recognition system is used to control an in-car device;
- Figure 2 illustrates a flow diagram representing steps which may be used to estimate transfer functions represent - ing a model of an in-car acoustic channel
- Figure 3 illustrates schematically components which may be used to implement a refinement of the algorithm m Figure 2;
- Figure 4 illustrates a block diagram representing a specific embodiment of the present invention.
- FIGS 5 to 8 illustrate examples of microphone signals obtained during experimental use of the present invention.
- Figure 1 illustrates schematically a simple situation in which stereo ECAD signals are transmitted from separate loudspeakers.
- Left stereo signal L( ⁇ ) is transmitted from left loudspeaker 101 and right stereo signal R( ⁇ ) is transmitted from right stereo speaker 102.
- Loudspeakers 101 and 102 are typically located in panelling on driver and passenger's doors. Further loudspeakers may also be fitted in the vehicle, for example they may be located in a boot compartment at the rear of the car. It will be appreciated by those skilled in the art that the specific embodiment described herein intended for use with two loudspeakers could be modified to function with different numbers of loudspeakers, which may or may not be configured to generate signals which correlate with signals being output from other loudspeakers present in the car.
- Figure 1 also includes a microphone 103 which is preferably connected to an in-car electronic device such as the radio for the purpose of receiving acoustic signals which may be used by a speech recognition system for controlling the device.
- a microphone 103 which is preferably connected to an in-car electronic device such as the radio for the purpose of receiving acoustic signals which may be used by a speech recognition system for controlling the device.
- a user's voice command which may be processed by the speech recognition system in order to control the electronic device is represented by wanted speech signal S(j ⁇ )104.
- a spectrum of the acoustic signal received at the microphone comprises components including a combination of the wanted speech S(j ⁇ ) and the signals produced by the loudspeaker having passed through an acoustic channel defined by the in-car environment.
- H AL (j ⁇ ) and H AR (j ⁇ ) were known, it would be possible to retrieve a signal corresponding to the wanted speech command spoken by the user by subtracting the left stereo source signal L(j ⁇ ) transferred by H AL (j ⁇ ) and the right source signal R(j ⁇ ) transferred by H ⁇ j ⁇ j ⁇ ) from the signal Y(j ⁇ ) received at microphone mono 103.
- source signals L(j ⁇ ) and R(j ⁇ ) may be accessible from the radio which produced them, the acoustic transfer functions H AR (j ⁇ ) and H AL (j ⁇ ) can only be estimated.
- a simple approach to the estimation of the acoustic transfer function is to find the long term ratio of microphone signal spectrum to each of the source stereo signals. Equations herein below describe this process for the right acoustic channel. Those skilled in the art will understand that a similar set of equations can be derived for the left acoustic channel.
- a basic transfer function H AR for the right acoustic channel may be written as follows:
- a spectrum of the signal Y(j ⁇ ) received at the microphone signal may be written as:
- Y (j ⁇ ) H M (j ⁇ ) .
- Equation (3) The following conclusions may be drawn from equation ⁇ ) :
- L(j ⁇ ) and R(j ⁇ ) are partially correlated, left and right acoustic channels cannot be unambiguously estimated. However, if L(j ⁇ ) and R(j ⁇ ) occupy different spectral regions or if corresponding time domain signals 1 (t) and r(t) have periods where one has low energy whilst the other has high energy, it may be still possible to make useful estimates of left and right channels for purposes of cancellation.
- the frequency domain estimation of the right acoustic channel response given by equation (3), and a corresponding equation for the left acoustic channel transfer function, H AL (j ⁇ ) may be used to obtain an estimate of the magnitude of the wanted speech spectrum S(j ⁇ ) .
- An estimate of the wanted speech magnitude spectrum may be obtained by subtracting the estimates of the left and right acoustic channels of the ECAD signals from the acoustic signal Y(j ⁇ ) received at the microphone:
- Equation (4) An estimate of the acoustic channel power transfer function for the right acoustic channel, derived by squaring equation (3) may be as follows:
- Equation (5) A corresponding estimate of the acoustic channel power transfer function for the left acoustic channel can also be derived by those skilled in the art.
- MFCC Mel Frequency Cepstral Co-efficient
- the phase characteristic encodes a frequency dependent delay spread associated with the acoustic transfer function.
- the minimum delay is about 3ms.
- the delay spread may be compensated when making the channel estimate using equation (5) .
- this compensation may be unnecessary if the spectral evaluation is done using a fast fourier transformer with block length much greater than the channel delay.
- a practical form of the cancellation of non-stationary interferer signals such as those produced by ECAD may therefore be achieved using an algorithm 200 as illustrated by steps in Figure 2 of the accompanying drawings .
- the steps 201 to 205 are repeated once for each single frame (i.e a signal received at the microphone in a fixed period of time) , however, initialisation steps 201 and 202 may only be performed for a first frame.
- step 203 new estimates of magnitudes of the left and right interference signals at the microphone are calculated. This is achieved for the left microphone signal by subtracting the channel estimate of the magnitude of the right channel (calculated during the algorithm iteration for the immediately previous frame) from the microphone signal received at the current iteration (n) . For the right interference channel, the magnitude estimate for the left channel derived during the previous iteration (n-1) is subtracted from the microphone signal :
- step 204 rough estimates of the left and right transfer functions, H AL (j ⁇ ) and H AR (j ⁇ ), are made. This is achieved for the left channel transfer function by dividing the estimated left interference signal calculated at step 203 by the signal transmitted from the left stereo acoustic channel. For the right transfer function, the right channel interference signal estimate calculated at step 203 is divided by the signal transmitted from the right acoustic stereo channel :
- the rough estimates of the channel transfer functions obtained at step 204 may be smoothed, preferably both in the time and frequency domains.
- Time smoothing is preferably achieved with a first order recursive filter using a time constant of several hundred milliseconds.
- time smoothing for the right channel may be as follows (a similar equation may also be obtained) :
- Frequency smoothing is preferably achieved using a Finite Impulse Response filter (represented by f ( ⁇ ) in an equation herein below) with a triangular impulse response covering about 300 Hertz.
- Frequency smoothing for the right channel may be as follows (a similar expression for the left channel may also be obtained) :
- the cancellation algorithm 200 described in steps 201 to 205 herein above may be refined by means of the four ways described herein below in order to attempt to deal with problems highlighted by equation (3) concerning correlation of left and right channel signals: 1. Updating of the recursive filter providing the smoothed channel estimate can be inhibited unless energy of one channel greatly exceeds energy of the other channel . This is preferably achieved by updating the left or right channel response only when it is assumed that only left or right channel, respectively, is active.
- a new right acoustic channel transfer function would be estimated at step 204 if a ratio of the total energy of the signal transmitted from the right acoustic stereo channel by the total energy of the signal transmitted from the left stereo acoustic channel exceeds a predetermined threshold value, otherwise the estimate calculated for the transfer function during the previous frame iteration is used.
- a corresponding estimation would also be performed for the left transfer function.
- E L to represent the total energy in the n th frame of the left stereo acoustic channel
- E R represent the total energy in the n th frame of the right stereo acoustic channel.
- the channel response estimation algorithm for the left channel is :
- Y(j ⁇ ) should consist mainly of terms due to the right channel and the wanted speech signal .
- Y(j ⁇ ) should contain very little energy due to the left channel if the threshold is set at high value. The reverse normally holds when considering the left channel .
- Time and domain smoothing substantially as described at step 205 would also be used. 2. Updating of recursively smoothed channel estimate at particular frequencies can be inhibited unless energy at that frequency in one channel greatly exceeds the energy at that frequency in the other channel .
- the threshold may apply to frequencies comprising a harmonic number in the Discreet Fourier Transforms of the signals.
- the index k refers to the harmonic number in the DFTs of the signals.
- E(k) R is the energy of the kth harmonic in the DFT of the right stereo source signal.
- the coherence function provides a measure of correlation over a period of time of phases of two different signals measured at a particular frequency.
- the coherence function may be used in various ways, normally based on the idea that the update of the acoustic channel responsible will be decreased if the left and right stereo channels are phase-correlated at a particular frequency. If the coherence approaches unity, the signals are correlated, but only at the specified frequency.
- the channel response estimates for the right channel may be derived from the following algorithm (a corresponding method for the transfer function for the left channel may also be derived) :
- ⁇ (k) is the coherence of the left and right stereo source signal at frequency index k.
- the orthogonalised signals are used to make the acoustic channel response estimates.
- the following expression may be used (a corresponding expression for the left stereo channel transfer function may also be obtained) : ft t i . ) - ⁇ k) H * (J ° ⁇ (*"(*> ⁇ gg. ( *> -c(*>> , H - ⁇ ) . (L" (k) + H CL (k) . c(k) ) _ s (k) ** R" (k) " R" lk) R" (k) R" (k) (k)
- the right stereo acoustic channel function, H AR (k) may be obtained by dividing the signal received at the microphone by R' ' (k) .
- FIG. 3 of the accompanying drawings illustrates schematically an example of components which may be used to form L' ' (j ⁇ ) and R' ' (j ⁇ ) .
- the components include two adaptive filters, 303 and 304, either implemented in the frequency domain, or preferably, the time domain.
- the coefficients of each FIR adaptive filter are adjusted using LMS or similar, to minimise the total energy in r ' ' (n) and l' 1 (n) , respectively, i.e. operate filters in standard system identification mode as in echo cancelling etc.
- the right stereo ECAD signal r(n) 301 is fed into adaptive filter 303 and a combiner 305.
- the left stereo ECAD signal 1 (n) 302 is fed into adaptive filter 304 and a combiner 306.
- the output of adaptive filter 303 is also fed into combiner 306.
- the output of adaptive filter 304 is also fed into combiner 305.
- the output of combiner 305 may be fed back via an adaption control path into adaptive filter 304.
- the output of mixer 306 may be fed back into adaptive filter 303 via an adaption control path.
- the output of combiner 305 comprises the orthogonal right stereo signal r' ' (n) 307.
- the output of combiner 306 comprises the left stereo orthogonal signal 1' ' (n) 308.
- FIG. 4 of the accompanying drawings illustrates a block diagram representing a specific embodiment of the present invention.
- Processing components of Fig. 4 may be electronic processors fitted integrally to the in-car device where the speech recognition system is located or, alternatively, may be a stand alone electronic device intended to receive acoustic signals, cancel non-stationary interfering signals and output a filtered acoustic signal to be received by the speech recognition system's microphone.
- ECAD sound source 401 (such as the signals output loudspeakers 101 and 102 of Figure 1) may be received directly by a spectral analysis process 404 so that the signal as produced by the ECAD prior to transmission through the in-car acoustic channel 403 may be analysed
- the ECAD signal is also received by a spectral analysis process 405 after transmission through acoustic channel 403 so that the signal 401 is m effect simultaneously spectrally analysed before and after transmission through the acoustic channel 403.
- the spectral analysis of processes 404 and 405 is preferably carried out at a 16 ms frame rate using a 256 point Fast Fourier Transformer. If user speech 402 (corre- sponding to wanted speech signal S(j ⁇ ) 104 of Figure 1) is also present then this acoustic signal too will also be transmitted through the acoustic channel 403 and received by spectral analysis process 405.
- acoustic channel model estimation process 406 which preferably functions in accordance with algorithm 200 described herein above.
- Acoustic channel model estimation process 406 produces an acoustic channel model 407 which may be used as an input to a spectral subtraction process 408 which also receives the acoustic signal transferred through channel 403.
- the acoustic channel model 407 is frozen for duration of the speech recognition process.
- the acoustic channel model 407 is then used to recover the speech signal from the microphone signal by subtracting the estimated spectrum of the ECAD interfering signals contained in the model 407 from the acoustic signals received at the microphone.
- the spectrally subtracted signal representing the recovered wanted speech 409 is then passed to a pattern matcher process 410 (part of the speech recognition system) which may use recognition feature sets such as Hidden Markov of models 311 in order to match the recovered speech signal 409 to a command which is recognised by the system.
- the pattern matcher 409 may then pass on an output signal to trace back and decision process 412 in order that the user's speech command be carried out by the device.
- test data was constructed by recording speech and interferer signals separately in the same car environment and then adding the two signals.
- the interfering music is a stereo signal.
- FIGS. 5A to 5D of the accompanying drawings illustrate microphone traces with and without cancellation in a case where the ECAD outputs pop music at OdB signal to 27
- Fig. 5A a signal received at the microphone prior to cancellation is illustrated. In this case, peak segmental speech and interferer levels are the same. This is a highly pessimistic way of estimating signal-to-noise ratio as amplitude variability of speech signal is higher than that of the ECAD music signal output which exceeds the speech for a considerable part of the example.
- Fig. 5B illustrates a signal resulting from an inverse transformation on the signal of Fig. 5A after spectral subtraction. The interfering signal as shown in Fig. 5B has clearly been reduced.
- Fig. 5C illustrates a signal representing normalised squared cepstral distances for application of the cancellation algorithm.
- Fig. 5A a signal received at the microphone prior to cancellation is illustrated. In this case, peak segmental speech and interferer levels are the same. This is a highly pessimistic way of estimating signal-to-noise ratio as amplitude variability of speech signal is higher than that of the ECAD music signal output which exceeds the speech
- FIG. 5D illustrates a signal trace for the normalised squared cepstral distances of Fig. 5C after spectral subtraction. Comparing the traces illustrated in Fig. 5C and 5D, it can be seen that the recovered speech cepstral are less distorted than with the interferer.
- FIG. 6A a signal received at the microphone prior to cancellation is illustrated.
- Fig. 6B illustrates a signal resulting from an inverse transform- ation on the signal of 6A after spectral subtraction. The interfering signal shown in Fig. 6B has clearly been reduced.
- Fig. 6C illustrates a signal representing normalised squared cepstral distances for application of the cancellation algorithm.
- Fig. 6D illustrates a signal trace for the normalised square cepstral distances of Fig. 6C after spectral subtraction.
- FIG. 7A a signal received at the microphone prior to cancellation is illustrated.
- Fig. 7B illustrates a signal resulting from an inverse transform- ation on the signal of 7A after spectral subtraction. The interfering signal shown in Fig. 7B has clearly been reduced.
- Fig. 7C illustrates a signal representing normalised squared cepstral distances for application of the cancellation algorithm.
- Fig. 7D illustrates a signal trace for the normalised square cepstral distances of Fig. 7C after spectral subtraction.
- FIG. 8A to 8D of the accompanying drawings illustrate microphone traces with and without cancellation in a case where the ECAD outputs opera music at 10 decibel signal to interference ratio.
- Fig. 8A a signal received at the microphone prior to cancellation is illustrated.
- Fig. 8B. illustrates a signal resulting from an inverse transformation on the signal of 8A after spectral subtraction. The interfering signal shown in Fig. 8B has clearly been reduced.
- Fig. 8C. illustrates a signal representing normalised squared cepstral distances for application of the cancellation algorithm.
- Fig. 8D illustrates a signal trace for the normalised square cepstral distances of Fig. 8C after spectral subtraction.
Abstract
Description
Claims
Priority Applications (3)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
EP00927504A EP1190417A1 (en) | 1999-05-07 | 2000-05-05 | Cancellation of non-stationary interfering signals for speech recognition |
JP2000617442A JP2002544552A (en) | 1999-05-07 | 2000-05-05 | Canceling non-stationary interference signals for speech recognition |
AU45904/00A AU4590400A (en) | 1999-05-07 | 2000-05-05 | Cancellation of non-stationary interfering signals for speech recognition |
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
GB9910448.1 | 1999-05-07 | ||
GBGB9910448.1A GB9910448D0 (en) | 1999-05-07 | 1999-05-07 | Cancellation of non-stationary interfering signals for speech recognition |
Publications (1)
Publication Number | Publication Date |
---|---|
WO2000068936A1 true WO2000068936A1 (en) | 2000-11-16 |
Family
ID=10852897
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
PCT/GB2000/001715 WO2000068936A1 (en) | 1999-05-07 | 2000-05-05 | Cancellation of non-stationary interfering signals for speech recognition |
Country Status (5)
Country | Link |
---|---|
EP (1) | EP1190417A1 (en) |
JP (1) | JP2002544552A (en) |
AU (1) | AU4590400A (en) |
GB (1) | GB9910448D0 (en) |
WO (1) | WO2000068936A1 (en) |
Cited By (119)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
WO2003017719A1 (en) * | 2001-08-08 | 2003-02-27 | Apple Computer, Inc. | Integrated sound input system |
EP1356452A1 (en) * | 2000-12-29 | 2003-10-29 | Cho, Mi-Hwa | Voice-controlled televison set and operating method thereof |
EP1362342A1 (en) * | 2001-02-20 | 2003-11-19 | Sungwoo Techno Inc. | A voice command identifier for a voice recognition system |
JP2004264816A (en) * | 2002-09-06 | 2004-09-24 | Microsoft Corp | Method of iterative noise estimation in recursive framework |
FR2864860A1 (en) * | 2004-01-07 | 2005-07-08 | Denso Corp | NOISE CANCELLATION SYSTEM, VOICE RECOGNITION SYSTEM, AND AUTOMOTIVE NAVIGATION SYSTEM |
US7349849B2 (en) | 2001-08-08 | 2008-03-25 | Apple, Inc. | Spacing for microphone elements |
US8892446B2 (en) | 2010-01-18 | 2014-11-18 | Apple Inc. | Service orchestration for intelligent automated assistant |
US9262612B2 (en) | 2011-03-21 | 2016-02-16 | Apple Inc. | Device access using voice authentication |
US9300784B2 (en) | 2013-06-13 | 2016-03-29 | Apple Inc. | System and method for emergency calls initiated by voice command |
US9330720B2 (en) | 2008-01-03 | 2016-05-03 | Apple Inc. | Methods and apparatus for altering audio output signals |
US9338493B2 (en) | 2014-06-30 | 2016-05-10 | Apple Inc. | Intelligent automated assistant for TV user interactions |
US9368114B2 (en) | 2013-03-14 | 2016-06-14 | Apple Inc. | Context-sensitive handling of interruptions |
US9430463B2 (en) | 2014-05-30 | 2016-08-30 | Apple Inc. | Exemplar-based natural language processing |
US9483461B2 (en) | 2012-03-06 | 2016-11-01 | Apple Inc. | Handling speech synthesis of content for multiple languages |
US9495129B2 (en) | 2012-06-29 | 2016-11-15 | Apple Inc. | Device, method, and user interface for voice-activated navigation and browsing of a document |
US9502031B2 (en) | 2014-05-27 | 2016-11-22 | Apple Inc. | Method for supporting dynamic grammars in WFST-based ASR |
US9535906B2 (en) | 2008-07-31 | 2017-01-03 | Apple Inc. | Mobile device having human language translation capability with positional feedback |
US9576574B2 (en) | 2012-09-10 | 2017-02-21 | Apple Inc. | Context-sensitive handling of interruptions by intelligent digital assistant |
US9582608B2 (en) | 2013-06-07 | 2017-02-28 | Apple Inc. | Unified ranking with entropy-weighted information for phrase-based semantic auto-completion |
US9620104B2 (en) | 2013-06-07 | 2017-04-11 | Apple Inc. | System and method for user-specified pronunciation of words for speech synthesis and recognition |
US9620105B2 (en) | 2014-05-15 | 2017-04-11 | Apple Inc. | Analyzing audio input for efficient speech and music recognition |
US9626955B2 (en) | 2008-04-05 | 2017-04-18 | Apple Inc. | Intelligent text-to-speech conversion |
US9633660B2 (en) | 2010-02-25 | 2017-04-25 | Apple Inc. | User profiling for voice input processing |
US9633004B2 (en) | 2014-05-30 | 2017-04-25 | Apple Inc. | Better resolution when referencing to concepts |
US9633674B2 (en) | 2013-06-07 | 2017-04-25 | Apple Inc. | System and method for detecting errors in interactions with a voice-based digital assistant |
US9646609B2 (en) | 2014-09-30 | 2017-05-09 | Apple Inc. | Caching apparatus for serving phonetic pronunciations |
US9646614B2 (en) | 2000-03-16 | 2017-05-09 | Apple Inc. | Fast, language-independent method for user authentication by voice |
US9668121B2 (en) | 2014-09-30 | 2017-05-30 | Apple Inc. | Social reminders |
US9697820B2 (en) | 2015-09-24 | 2017-07-04 | Apple Inc. | Unit-selection text-to-speech synthesis using concatenation-sensitive neural networks |
US9697822B1 (en) | 2013-03-15 | 2017-07-04 | Apple Inc. | System and method for updating an adaptive speech recognition model |
US9711141B2 (en) | 2014-12-09 | 2017-07-18 | Apple Inc. | Disambiguating heteronyms in speech synthesis |
US9715875B2 (en) | 2014-05-30 | 2017-07-25 | Apple Inc. | Reducing the need for manual start/end-pointing and trigger phrases |
US9721566B2 (en) | 2015-03-08 | 2017-08-01 | Apple Inc. | Competing devices responding to voice triggers |
US9734193B2 (en) | 2014-05-30 | 2017-08-15 | Apple Inc. | Determining domain salience ranking from ambiguous words in natural speech |
US9760559B2 (en) | 2014-05-30 | 2017-09-12 | Apple Inc. | Predictive text input |
US9785630B2 (en) | 2014-05-30 | 2017-10-10 | Apple Inc. | Text prediction using combined word N-gram and unigram language models |
US9798393B2 (en) | 2011-08-29 | 2017-10-24 | Apple Inc. | Text correction processing |
US9818400B2 (en) | 2014-09-11 | 2017-11-14 | Apple Inc. | Method and apparatus for discovering trending terms in speech requests |
US9842101B2 (en) | 2014-05-30 | 2017-12-12 | Apple Inc. | Predictive conversion of language input |
US9842105B2 (en) | 2015-04-16 | 2017-12-12 | Apple Inc. | Parsimonious continuous-space phrase representations for natural language processing |
US9858925B2 (en) | 2009-06-05 | 2018-01-02 | Apple Inc. | Using context information to facilitate processing of commands in a virtual assistant |
US9865280B2 (en) | 2015-03-06 | 2018-01-09 | Apple Inc. | Structured dictation using intelligent automated assistants |
US9886432B2 (en) | 2014-09-30 | 2018-02-06 | Apple Inc. | Parsimonious handling of word inflection via categorical stem + suffix N-gram language models |
US9886953B2 (en) | 2015-03-08 | 2018-02-06 | Apple Inc. | Virtual assistant activation |
US9899019B2 (en) | 2015-03-18 | 2018-02-20 | Apple Inc. | Systems and methods for structured stem and suffix language models |
US9922642B2 (en) | 2013-03-15 | 2018-03-20 | Apple Inc. | Training an at least partial voice command system |
US9934775B2 (en) | 2016-05-26 | 2018-04-03 | Apple Inc. | Unit-selection text-to-speech synthesis based on predicted concatenation parameters |
US9953088B2 (en) | 2012-05-14 | 2018-04-24 | Apple Inc. | Crowd sourcing information to fulfill user requests |
US9959870B2 (en) | 2008-12-11 | 2018-05-01 | Apple Inc. | Speech recognition involving a mobile device |
US9966065B2 (en) | 2014-05-30 | 2018-05-08 | Apple Inc. | Multi-command single utterance input method |
US9966068B2 (en) | 2013-06-08 | 2018-05-08 | Apple Inc. | Interpreting and acting upon commands that involve sharing information with remote devices |
US9972304B2 (en) | 2016-06-03 | 2018-05-15 | Apple Inc. | Privacy preserving distributed evaluation framework for embedded personalized systems |
US9971774B2 (en) | 2012-09-19 | 2018-05-15 | Apple Inc. | Voice-based media searching |
US10043516B2 (en) | 2016-09-23 | 2018-08-07 | Apple Inc. | Intelligent automated assistant |
US10049663B2 (en) | 2016-06-08 | 2018-08-14 | Apple, Inc. | Intelligent automated assistant for media exploration |
US10049668B2 (en) | 2015-12-02 | 2018-08-14 | Apple Inc. | Applying neural network language models to weighted finite state transducers for automatic speech recognition |
US10057736B2 (en) | 2011-06-03 | 2018-08-21 | Apple Inc. | Active transport based notifications |
US10067938B2 (en) | 2016-06-10 | 2018-09-04 | Apple Inc. | Multilingual word prediction |
US10074360B2 (en) | 2014-09-30 | 2018-09-11 | Apple Inc. | Providing an indication of the suitability of speech recognition |
US10078631B2 (en) | 2014-05-30 | 2018-09-18 | Apple Inc. | Entropy-guided text prediction using combined word and character n-gram language models |
US10079014B2 (en) | 2012-06-08 | 2018-09-18 | Apple Inc. | Name recognition system |
US10083688B2 (en) | 2015-05-27 | 2018-09-25 | Apple Inc. | Device voice control for selecting a displayed affordance |
US10089072B2 (en) | 2016-06-11 | 2018-10-02 | Apple Inc. | Intelligent device arbitration and control |
US10101822B2 (en) | 2015-06-05 | 2018-10-16 | Apple Inc. | Language input correction |
US10127911B2 (en) | 2014-09-30 | 2018-11-13 | Apple Inc. | Speaker identification and unsupervised speaker adaptation techniques |
US10127220B2 (en) | 2015-06-04 | 2018-11-13 | Apple Inc. | Language identification from short strings |
US10134385B2 (en) | 2012-03-02 | 2018-11-20 | Apple Inc. | Systems and methods for name pronunciation |
US10170123B2 (en) | 2014-05-30 | 2019-01-01 | Apple Inc. | Intelligent assistant for home automation |
US10176167B2 (en) | 2013-06-09 | 2019-01-08 | Apple Inc. | System and method for inferring user intent from speech inputs |
US10185542B2 (en) | 2013-06-09 | 2019-01-22 | Apple Inc. | Device, method, and graphical user interface for enabling conversation persistence across two or more instances of a digital assistant |
US10186254B2 (en) | 2015-06-07 | 2019-01-22 | Apple Inc. | Context-based endpoint detection |
US10192552B2 (en) | 2016-06-10 | 2019-01-29 | Apple Inc. | Digital assistant providing whispered speech |
US10199051B2 (en) | 2013-02-07 | 2019-02-05 | Apple Inc. | Voice trigger for a digital assistant |
US10223066B2 (en) | 2015-12-23 | 2019-03-05 | Apple Inc. | Proactive assistance based on dialog communication between devices |
US10241752B2 (en) | 2011-09-30 | 2019-03-26 | Apple Inc. | Interface for a virtual digital assistant |
US10241644B2 (en) | 2011-06-03 | 2019-03-26 | Apple Inc. | Actionable reminder entries |
US10249300B2 (en) | 2016-06-06 | 2019-04-02 | Apple Inc. | Intelligent list reading |
US10255907B2 (en) | 2015-06-07 | 2019-04-09 | Apple Inc. | Automatic accent detection using acoustic models |
US10269345B2 (en) | 2016-06-11 | 2019-04-23 | Apple Inc. | Intelligent task discovery |
US10276170B2 (en) | 2010-01-18 | 2019-04-30 | Apple Inc. | Intelligent automated assistant |
US10283110B2 (en) | 2009-07-02 | 2019-05-07 | Apple Inc. | Methods and apparatuses for automatic speech recognition |
US10289433B2 (en) | 2014-05-30 | 2019-05-14 | Apple Inc. | Domain specific language for encoding assistant dialog |
US10297253B2 (en) | 2016-06-11 | 2019-05-21 | Apple Inc. | Application integration with a digital assistant |
US10318871B2 (en) | 2005-09-08 | 2019-06-11 | Apple Inc. | Method and apparatus for building an intelligent automated assistant |
US10354011B2 (en) | 2016-06-09 | 2019-07-16 | Apple Inc. | Intelligent automated assistant in a home environment |
US10356243B2 (en) | 2015-06-05 | 2019-07-16 | Apple Inc. | Virtual assistant aided communication with 3rd party service in a communication session |
US10366158B2 (en) | 2015-09-29 | 2019-07-30 | Apple Inc. | Efficient word encoding for recurrent neural network language models |
US10410637B2 (en) | 2017-05-12 | 2019-09-10 | Apple Inc. | User-specific acoustic models |
US10446143B2 (en) | 2016-03-14 | 2019-10-15 | Apple Inc. | Identification of voice inputs providing credentials |
US10446141B2 (en) | 2014-08-28 | 2019-10-15 | Apple Inc. | Automatic speech recognition based on user feedback |
US10482874B2 (en) | 2017-05-15 | 2019-11-19 | Apple Inc. | Hierarchical belief states for digital assistants |
US10490187B2 (en) | 2016-06-10 | 2019-11-26 | Apple Inc. | Digital assistant providing automated status report |
US10496753B2 (en) | 2010-01-18 | 2019-12-03 | Apple Inc. | Automatically adapting user interfaces for hands-free interaction |
US10509862B2 (en) | 2016-06-10 | 2019-12-17 | Apple Inc. | Dynamic phrase expansion of language input |
US10521466B2 (en) | 2016-06-11 | 2019-12-31 | Apple Inc. | Data driven natural language event detection and classification |
US10552013B2 (en) | 2014-12-02 | 2020-02-04 | Apple Inc. | Data detection |
US10553209B2 (en) | 2010-01-18 | 2020-02-04 | Apple Inc. | Systems and methods for hands-free notification summaries |
US10567477B2 (en) | 2015-03-08 | 2020-02-18 | Apple Inc. | Virtual assistant continuity |
US10568032B2 (en) | 2007-04-03 | 2020-02-18 | Apple Inc. | Method and system for operating a multi-function portable electronic device using voice-activation |
US10592095B2 (en) | 2014-05-23 | 2020-03-17 | Apple Inc. | Instantaneous speaking of content on touch devices |
US10593346B2 (en) | 2016-12-22 | 2020-03-17 | Apple Inc. | Rank-reduced token representation for automatic speech recognition |
US10607141B2 (en) | 2010-01-25 | 2020-03-31 | Newvaluexchange Ltd. | Apparatuses, methods and systems for a digital conversation management platform |
US10659851B2 (en) | 2014-06-30 | 2020-05-19 | Apple Inc. | Real-time digital assistant knowledge updates |
US10671428B2 (en) | 2015-09-08 | 2020-06-02 | Apple Inc. | Distributed personal assistant |
US10679605B2 (en) | 2010-01-18 | 2020-06-09 | Apple Inc. | Hands-free list-reading by intelligent automated assistant |
US10691473B2 (en) | 2015-11-06 | 2020-06-23 | Apple Inc. | Intelligent automated assistant in a messaging environment |
US10706373B2 (en) | 2011-06-03 | 2020-07-07 | Apple Inc. | Performing actions associated with task items that represent tasks to perform |
US10705794B2 (en) | 2010-01-18 | 2020-07-07 | Apple Inc. | Automatically adapting user interfaces for hands-free interaction |
US10733993B2 (en) | 2016-06-10 | 2020-08-04 | Apple Inc. | Intelligent digital assistant in a multi-tasking environment |
US10747498B2 (en) | 2015-09-08 | 2020-08-18 | Apple Inc. | Zero latency digital assistant |
US10755703B2 (en) | 2017-05-11 | 2020-08-25 | Apple Inc. | Offline personal assistant |
US10762293B2 (en) | 2010-12-22 | 2020-09-01 | Apple Inc. | Using parts-of-speech tagging and named entity recognition for spelling correction |
US10791176B2 (en) | 2017-05-12 | 2020-09-29 | Apple Inc. | Synchronization and task delegation of a digital assistant |
US10791216B2 (en) | 2013-08-06 | 2020-09-29 | Apple Inc. | Auto-activating smart responses based on activities from remote devices |
US10810274B2 (en) | 2017-05-15 | 2020-10-20 | Apple Inc. | Optimizing dialogue policy decisions for digital assistants using implicit feedback |
US11010550B2 (en) | 2015-09-29 | 2021-05-18 | Apple Inc. | Unified language modeling framework for word prediction, auto-completion and auto-correction |
US11025565B2 (en) | 2015-06-07 | 2021-06-01 | Apple Inc. | Personalized prediction of responses for instant messaging |
US11217255B2 (en) | 2017-05-16 | 2022-01-04 | Apple Inc. | Far-field extension for digital assistant services |
US11587559B2 (en) | 2015-09-30 | 2023-02-21 | Apple Inc. | Intelligent device identification |
Citations (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
EP0856834A2 (en) * | 1997-01-29 | 1998-08-05 | Nec Corporation | Noise canceler |
US5864804A (en) * | 1995-06-10 | 1999-01-26 | U.S. Philips Corporation | Voice recognition system |
-
1999
- 1999-05-07 GB GBGB9910448.1A patent/GB9910448D0/en not_active Ceased
-
2000
- 2000-05-05 WO PCT/GB2000/001715 patent/WO2000068936A1/en not_active Application Discontinuation
- 2000-05-05 EP EP00927504A patent/EP1190417A1/en not_active Withdrawn
- 2000-05-05 AU AU45904/00A patent/AU4590400A/en not_active Abandoned
- 2000-05-05 JP JP2000617442A patent/JP2002544552A/en active Pending
Patent Citations (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US5864804A (en) * | 1995-06-10 | 1999-01-26 | U.S. Philips Corporation | Voice recognition system |
EP0856834A2 (en) * | 1997-01-29 | 1998-08-05 | Nec Corporation | Noise canceler |
Cited By (166)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US9646614B2 (en) | 2000-03-16 | 2017-05-09 | Apple Inc. | Fast, language-independent method for user authentication by voice |
EP1356452A1 (en) * | 2000-12-29 | 2003-10-29 | Cho, Mi-Hwa | Voice-controlled televison set and operating method thereof |
EP1356452A4 (en) * | 2000-12-29 | 2004-03-31 | Cho Mi Hwa | Voice-controlled televison set and operating method thereof |
EP1362342A1 (en) * | 2001-02-20 | 2003-11-19 | Sungwoo Techno Inc. | A voice command identifier for a voice recognition system |
EP1362342A4 (en) * | 2001-02-20 | 2005-09-14 | Sungwoo Techno Inc | A voice command identifier for a voice recognition system |
US7349849B2 (en) | 2001-08-08 | 2008-03-25 | Apple, Inc. | Spacing for microphone elements |
WO2003017719A1 (en) * | 2001-08-08 | 2003-02-27 | Apple Computer, Inc. | Integrated sound input system |
JP4491210B2 (en) * | 2002-09-06 | 2010-06-30 | マイクロソフト コーポレーション | Iterative noise estimation method in recursive construction |
JP2004264816A (en) * | 2002-09-06 | 2004-09-24 | Microsoft Corp | Method of iterative noise estimation in recursive framework |
US8010354B2 (en) | 2004-01-07 | 2011-08-30 | Denso Corporation | Noise cancellation system, speech recognition system, and car navigation system |
FR2864860A1 (en) * | 2004-01-07 | 2005-07-08 | Denso Corp | NOISE CANCELLATION SYSTEM, VOICE RECOGNITION SYSTEM, AND AUTOMOTIVE NAVIGATION SYSTEM |
US10318871B2 (en) | 2005-09-08 | 2019-06-11 | Apple Inc. | Method and apparatus for building an intelligent automated assistant |
US8930191B2 (en) | 2006-09-08 | 2015-01-06 | Apple Inc. | Paraphrasing of user requests and results by automated digital assistant |
US8942986B2 (en) | 2006-09-08 | 2015-01-27 | Apple Inc. | Determining user intent based on ontologies of domains |
US9117447B2 (en) | 2006-09-08 | 2015-08-25 | Apple Inc. | Using event alert text as input to an automated assistant |
US10568032B2 (en) | 2007-04-03 | 2020-02-18 | Apple Inc. | Method and system for operating a multi-function portable electronic device using voice-activation |
US9330720B2 (en) | 2008-01-03 | 2016-05-03 | Apple Inc. | Methods and apparatus for altering audio output signals |
US10381016B2 (en) | 2008-01-03 | 2019-08-13 | Apple Inc. | Methods and apparatus for altering audio output signals |
US9626955B2 (en) | 2008-04-05 | 2017-04-18 | Apple Inc. | Intelligent text-to-speech conversion |
US9865248B2 (en) | 2008-04-05 | 2018-01-09 | Apple Inc. | Intelligent text-to-speech conversion |
US10108612B2 (en) | 2008-07-31 | 2018-10-23 | Apple Inc. | Mobile device having human language translation capability with positional feedback |
US9535906B2 (en) | 2008-07-31 | 2017-01-03 | Apple Inc. | Mobile device having human language translation capability with positional feedback |
US9959870B2 (en) | 2008-12-11 | 2018-05-01 | Apple Inc. | Speech recognition involving a mobile device |
US11080012B2 (en) | 2009-06-05 | 2021-08-03 | Apple Inc. | Interface for a virtual digital assistant |
US10795541B2 (en) | 2009-06-05 | 2020-10-06 | Apple Inc. | Intelligent organization of tasks items |
US10475446B2 (en) | 2009-06-05 | 2019-11-12 | Apple Inc. | Using context information to facilitate processing of commands in a virtual assistant |
US9858925B2 (en) | 2009-06-05 | 2018-01-02 | Apple Inc. | Using context information to facilitate processing of commands in a virtual assistant |
US10283110B2 (en) | 2009-07-02 | 2019-05-07 | Apple Inc. | Methods and apparatuses for automatic speech recognition |
US10553209B2 (en) | 2010-01-18 | 2020-02-04 | Apple Inc. | Systems and methods for hands-free notification summaries |
US10496753B2 (en) | 2010-01-18 | 2019-12-03 | Apple Inc. | Automatically adapting user interfaces for hands-free interaction |
US11423886B2 (en) | 2010-01-18 | 2022-08-23 | Apple Inc. | Task flow identification based on user intent |
US9318108B2 (en) | 2010-01-18 | 2016-04-19 | Apple Inc. | Intelligent automated assistant |
US9548050B2 (en) | 2010-01-18 | 2017-01-17 | Apple Inc. | Intelligent automated assistant |
US10679605B2 (en) | 2010-01-18 | 2020-06-09 | Apple Inc. | Hands-free list-reading by intelligent automated assistant |
US10706841B2 (en) | 2010-01-18 | 2020-07-07 | Apple Inc. | Task flow identification based on user intent |
US10276170B2 (en) | 2010-01-18 | 2019-04-30 | Apple Inc. | Intelligent automated assistant |
US8903716B2 (en) | 2010-01-18 | 2014-12-02 | Apple Inc. | Personalized vocabulary for digital assistant |
US10705794B2 (en) | 2010-01-18 | 2020-07-07 | Apple Inc. | Automatically adapting user interfaces for hands-free interaction |
US8892446B2 (en) | 2010-01-18 | 2014-11-18 | Apple Inc. | Service orchestration for intelligent automated assistant |
US11410053B2 (en) | 2010-01-25 | 2022-08-09 | Newvaluexchange Ltd. | Apparatuses, methods and systems for a digital conversation management platform |
US10607140B2 (en) | 2010-01-25 | 2020-03-31 | Newvaluexchange Ltd. | Apparatuses, methods and systems for a digital conversation management platform |
US10984326B2 (en) | 2010-01-25 | 2021-04-20 | Newvaluexchange Ltd. | Apparatuses, methods and systems for a digital conversation management platform |
US10984327B2 (en) | 2010-01-25 | 2021-04-20 | New Valuexchange Ltd. | Apparatuses, methods and systems for a digital conversation management platform |
US10607141B2 (en) | 2010-01-25 | 2020-03-31 | Newvaluexchange Ltd. | Apparatuses, methods and systems for a digital conversation management platform |
US9633660B2 (en) | 2010-02-25 | 2017-04-25 | Apple Inc. | User profiling for voice input processing |
US10049675B2 (en) | 2010-02-25 | 2018-08-14 | Apple Inc. | User profiling for voice input processing |
US10762293B2 (en) | 2010-12-22 | 2020-09-01 | Apple Inc. | Using parts-of-speech tagging and named entity recognition for spelling correction |
US10102359B2 (en) | 2011-03-21 | 2018-10-16 | Apple Inc. | Device access using voice authentication |
US9262612B2 (en) | 2011-03-21 | 2016-02-16 | Apple Inc. | Device access using voice authentication |
US10706373B2 (en) | 2011-06-03 | 2020-07-07 | Apple Inc. | Performing actions associated with task items that represent tasks to perform |
US11120372B2 (en) | 2011-06-03 | 2021-09-14 | Apple Inc. | Performing actions associated with task items that represent tasks to perform |
US10057736B2 (en) | 2011-06-03 | 2018-08-21 | Apple Inc. | Active transport based notifications |
US10241644B2 (en) | 2011-06-03 | 2019-03-26 | Apple Inc. | Actionable reminder entries |
US9798393B2 (en) | 2011-08-29 | 2017-10-24 | Apple Inc. | Text correction processing |
US10241752B2 (en) | 2011-09-30 | 2019-03-26 | Apple Inc. | Interface for a virtual digital assistant |
US10134385B2 (en) | 2012-03-02 | 2018-11-20 | Apple Inc. | Systems and methods for name pronunciation |
US9483461B2 (en) | 2012-03-06 | 2016-11-01 | Apple Inc. | Handling speech synthesis of content for multiple languages |
US9953088B2 (en) | 2012-05-14 | 2018-04-24 | Apple Inc. | Crowd sourcing information to fulfill user requests |
US10079014B2 (en) | 2012-06-08 | 2018-09-18 | Apple Inc. | Name recognition system |
US9495129B2 (en) | 2012-06-29 | 2016-11-15 | Apple Inc. | Device, method, and user interface for voice-activated navigation and browsing of a document |
US9576574B2 (en) | 2012-09-10 | 2017-02-21 | Apple Inc. | Context-sensitive handling of interruptions by intelligent digital assistant |
US9971774B2 (en) | 2012-09-19 | 2018-05-15 | Apple Inc. | Voice-based media searching |
US10199051B2 (en) | 2013-02-07 | 2019-02-05 | Apple Inc. | Voice trigger for a digital assistant |
US10978090B2 (en) | 2013-02-07 | 2021-04-13 | Apple Inc. | Voice trigger for a digital assistant |
US9368114B2 (en) | 2013-03-14 | 2016-06-14 | Apple Inc. | Context-sensitive handling of interruptions |
US9922642B2 (en) | 2013-03-15 | 2018-03-20 | Apple Inc. | Training an at least partial voice command system |
US9697822B1 (en) | 2013-03-15 | 2017-07-04 | Apple Inc. | System and method for updating an adaptive speech recognition model |
US9620104B2 (en) | 2013-06-07 | 2017-04-11 | Apple Inc. | System and method for user-specified pronunciation of words for speech synthesis and recognition |
US9966060B2 (en) | 2013-06-07 | 2018-05-08 | Apple Inc. | System and method for user-specified pronunciation of words for speech synthesis and recognition |
US9582608B2 (en) | 2013-06-07 | 2017-02-28 | Apple Inc. | Unified ranking with entropy-weighted information for phrase-based semantic auto-completion |
US9633674B2 (en) | 2013-06-07 | 2017-04-25 | Apple Inc. | System and method for detecting errors in interactions with a voice-based digital assistant |
US10657961B2 (en) | 2013-06-08 | 2020-05-19 | Apple Inc. | Interpreting and acting upon commands that involve sharing information with remote devices |
US9966068B2 (en) | 2013-06-08 | 2018-05-08 | Apple Inc. | Interpreting and acting upon commands that involve sharing information with remote devices |
US10185542B2 (en) | 2013-06-09 | 2019-01-22 | Apple Inc. | Device, method, and graphical user interface for enabling conversation persistence across two or more instances of a digital assistant |
US10176167B2 (en) | 2013-06-09 | 2019-01-08 | Apple Inc. | System and method for inferring user intent from speech inputs |
US9300784B2 (en) | 2013-06-13 | 2016-03-29 | Apple Inc. | System and method for emergency calls initiated by voice command |
US10791216B2 (en) | 2013-08-06 | 2020-09-29 | Apple Inc. | Auto-activating smart responses based on activities from remote devices |
US9620105B2 (en) | 2014-05-15 | 2017-04-11 | Apple Inc. | Analyzing audio input for efficient speech and music recognition |
US10592095B2 (en) | 2014-05-23 | 2020-03-17 | Apple Inc. | Instantaneous speaking of content on touch devices |
US9502031B2 (en) | 2014-05-27 | 2016-11-22 | Apple Inc. | Method for supporting dynamic grammars in WFST-based ASR |
US9715875B2 (en) | 2014-05-30 | 2017-07-25 | Apple Inc. | Reducing the need for manual start/end-pointing and trigger phrases |
US9734193B2 (en) | 2014-05-30 | 2017-08-15 | Apple Inc. | Determining domain salience ranking from ambiguous words in natural speech |
US9785630B2 (en) | 2014-05-30 | 2017-10-10 | Apple Inc. | Text prediction using combined word N-gram and unigram language models |
US9760559B2 (en) | 2014-05-30 | 2017-09-12 | Apple Inc. | Predictive text input |
US9966065B2 (en) | 2014-05-30 | 2018-05-08 | Apple Inc. | Multi-command single utterance input method |
US10169329B2 (en) | 2014-05-30 | 2019-01-01 | Apple Inc. | Exemplar-based natural language processing |
US10170123B2 (en) | 2014-05-30 | 2019-01-01 | Apple Inc. | Intelligent assistant for home automation |
US10497365B2 (en) | 2014-05-30 | 2019-12-03 | Apple Inc. | Multi-command single utterance input method |
US10083690B2 (en) | 2014-05-30 | 2018-09-25 | Apple Inc. | Better resolution when referencing to concepts |
US10289433B2 (en) | 2014-05-30 | 2019-05-14 | Apple Inc. | Domain specific language for encoding assistant dialog |
US9842101B2 (en) | 2014-05-30 | 2017-12-12 | Apple Inc. | Predictive conversion of language input |
US10078631B2 (en) | 2014-05-30 | 2018-09-18 | Apple Inc. | Entropy-guided text prediction using combined word and character n-gram language models |
US9633004B2 (en) | 2014-05-30 | 2017-04-25 | Apple Inc. | Better resolution when referencing to concepts |
US11133008B2 (en) | 2014-05-30 | 2021-09-28 | Apple Inc. | Reducing the need for manual start/end-pointing and trigger phrases |
US9430463B2 (en) | 2014-05-30 | 2016-08-30 | Apple Inc. | Exemplar-based natural language processing |
US11257504B2 (en) | 2014-05-30 | 2022-02-22 | Apple Inc. | Intelligent assistant for home automation |
US9668024B2 (en) | 2014-06-30 | 2017-05-30 | Apple Inc. | Intelligent automated assistant for TV user interactions |
US10904611B2 (en) | 2014-06-30 | 2021-01-26 | Apple Inc. | Intelligent automated assistant for TV user interactions |
US9338493B2 (en) | 2014-06-30 | 2016-05-10 | Apple Inc. | Intelligent automated assistant for TV user interactions |
US10659851B2 (en) | 2014-06-30 | 2020-05-19 | Apple Inc. | Real-time digital assistant knowledge updates |
US10446141B2 (en) | 2014-08-28 | 2019-10-15 | Apple Inc. | Automatic speech recognition based on user feedback |
US9818400B2 (en) | 2014-09-11 | 2017-11-14 | Apple Inc. | Method and apparatus for discovering trending terms in speech requests |
US10431204B2 (en) | 2014-09-11 | 2019-10-01 | Apple Inc. | Method and apparatus for discovering trending terms in speech requests |
US9886432B2 (en) | 2014-09-30 | 2018-02-06 | Apple Inc. | Parsimonious handling of word inflection via categorical stem + suffix N-gram language models |
US9668121B2 (en) | 2014-09-30 | 2017-05-30 | Apple Inc. | Social reminders |
US9646609B2 (en) | 2014-09-30 | 2017-05-09 | Apple Inc. | Caching apparatus for serving phonetic pronunciations |
US9986419B2 (en) | 2014-09-30 | 2018-05-29 | Apple Inc. | Social reminders |
US10127911B2 (en) | 2014-09-30 | 2018-11-13 | Apple Inc. | Speaker identification and unsupervised speaker adaptation techniques |
US10074360B2 (en) | 2014-09-30 | 2018-09-11 | Apple Inc. | Providing an indication of the suitability of speech recognition |
US10552013B2 (en) | 2014-12-02 | 2020-02-04 | Apple Inc. | Data detection |
US11556230B2 (en) | 2014-12-02 | 2023-01-17 | Apple Inc. | Data detection |
US9711141B2 (en) | 2014-12-09 | 2017-07-18 | Apple Inc. | Disambiguating heteronyms in speech synthesis |
US9865280B2 (en) | 2015-03-06 | 2018-01-09 | Apple Inc. | Structured dictation using intelligent automated assistants |
US9721566B2 (en) | 2015-03-08 | 2017-08-01 | Apple Inc. | Competing devices responding to voice triggers |
US10567477B2 (en) | 2015-03-08 | 2020-02-18 | Apple Inc. | Virtual assistant continuity |
US9886953B2 (en) | 2015-03-08 | 2018-02-06 | Apple Inc. | Virtual assistant activation |
US11087759B2 (en) | 2015-03-08 | 2021-08-10 | Apple Inc. | Virtual assistant activation |
US10311871B2 (en) | 2015-03-08 | 2019-06-04 | Apple Inc. | Competing devices responding to voice triggers |
US9899019B2 (en) | 2015-03-18 | 2018-02-20 | Apple Inc. | Systems and methods for structured stem and suffix language models |
US9842105B2 (en) | 2015-04-16 | 2017-12-12 | Apple Inc. | Parsimonious continuous-space phrase representations for natural language processing |
US10083688B2 (en) | 2015-05-27 | 2018-09-25 | Apple Inc. | Device voice control for selecting a displayed affordance |
US10127220B2 (en) | 2015-06-04 | 2018-11-13 | Apple Inc. | Language identification from short strings |
US10356243B2 (en) | 2015-06-05 | 2019-07-16 | Apple Inc. | Virtual assistant aided communication with 3rd party service in a communication session |
US10101822B2 (en) | 2015-06-05 | 2018-10-16 | Apple Inc. | Language input correction |
US11025565B2 (en) | 2015-06-07 | 2021-06-01 | Apple Inc. | Personalized prediction of responses for instant messaging |
US10255907B2 (en) | 2015-06-07 | 2019-04-09 | Apple Inc. | Automatic accent detection using acoustic models |
US10186254B2 (en) | 2015-06-07 | 2019-01-22 | Apple Inc. | Context-based endpoint detection |
US10747498B2 (en) | 2015-09-08 | 2020-08-18 | Apple Inc. | Zero latency digital assistant |
US10671428B2 (en) | 2015-09-08 | 2020-06-02 | Apple Inc. | Distributed personal assistant |
US11500672B2 (en) | 2015-09-08 | 2022-11-15 | Apple Inc. | Distributed personal assistant |
US9697820B2 (en) | 2015-09-24 | 2017-07-04 | Apple Inc. | Unit-selection text-to-speech synthesis using concatenation-sensitive neural networks |
US11010550B2 (en) | 2015-09-29 | 2021-05-18 | Apple Inc. | Unified language modeling framework for word prediction, auto-completion and auto-correction |
US10366158B2 (en) | 2015-09-29 | 2019-07-30 | Apple Inc. | Efficient word encoding for recurrent neural network language models |
US11587559B2 (en) | 2015-09-30 | 2023-02-21 | Apple Inc. | Intelligent device identification |
US10691473B2 (en) | 2015-11-06 | 2020-06-23 | Apple Inc. | Intelligent automated assistant in a messaging environment |
US11526368B2 (en) | 2015-11-06 | 2022-12-13 | Apple Inc. | Intelligent automated assistant in a messaging environment |
US10049668B2 (en) | 2015-12-02 | 2018-08-14 | Apple Inc. | Applying neural network language models to weighted finite state transducers for automatic speech recognition |
US10223066B2 (en) | 2015-12-23 | 2019-03-05 | Apple Inc. | Proactive assistance based on dialog communication between devices |
US10446143B2 (en) | 2016-03-14 | 2019-10-15 | Apple Inc. | Identification of voice inputs providing credentials |
US9934775B2 (en) | 2016-05-26 | 2018-04-03 | Apple Inc. | Unit-selection text-to-speech synthesis based on predicted concatenation parameters |
US9972304B2 (en) | 2016-06-03 | 2018-05-15 | Apple Inc. | Privacy preserving distributed evaluation framework for embedded personalized systems |
US10249300B2 (en) | 2016-06-06 | 2019-04-02 | Apple Inc. | Intelligent list reading |
US11069347B2 (en) | 2016-06-08 | 2021-07-20 | Apple Inc. | Intelligent automated assistant for media exploration |
US10049663B2 (en) | 2016-06-08 | 2018-08-14 | Apple, Inc. | Intelligent automated assistant for media exploration |
US10354011B2 (en) | 2016-06-09 | 2019-07-16 | Apple Inc. | Intelligent automated assistant in a home environment |
US10490187B2 (en) | 2016-06-10 | 2019-11-26 | Apple Inc. | Digital assistant providing automated status report |
US10509862B2 (en) | 2016-06-10 | 2019-12-17 | Apple Inc. | Dynamic phrase expansion of language input |
US10192552B2 (en) | 2016-06-10 | 2019-01-29 | Apple Inc. | Digital assistant providing whispered speech |
US10733993B2 (en) | 2016-06-10 | 2020-08-04 | Apple Inc. | Intelligent digital assistant in a multi-tasking environment |
US10067938B2 (en) | 2016-06-10 | 2018-09-04 | Apple Inc. | Multilingual word prediction |
US11037565B2 (en) | 2016-06-10 | 2021-06-15 | Apple Inc. | Intelligent digital assistant in a multi-tasking environment |
US11152002B2 (en) | 2016-06-11 | 2021-10-19 | Apple Inc. | Application integration with a digital assistant |
US10269345B2 (en) | 2016-06-11 | 2019-04-23 | Apple Inc. | Intelligent task discovery |
US10089072B2 (en) | 2016-06-11 | 2018-10-02 | Apple Inc. | Intelligent device arbitration and control |
US10297253B2 (en) | 2016-06-11 | 2019-05-21 | Apple Inc. | Application integration with a digital assistant |
US10521466B2 (en) | 2016-06-11 | 2019-12-31 | Apple Inc. | Data driven natural language event detection and classification |
US10043516B2 (en) | 2016-09-23 | 2018-08-07 | Apple Inc. | Intelligent automated assistant |
US10553215B2 (en) | 2016-09-23 | 2020-02-04 | Apple Inc. | Intelligent automated assistant |
US10593346B2 (en) | 2016-12-22 | 2020-03-17 | Apple Inc. | Rank-reduced token representation for automatic speech recognition |
US10755703B2 (en) | 2017-05-11 | 2020-08-25 | Apple Inc. | Offline personal assistant |
US10791176B2 (en) | 2017-05-12 | 2020-09-29 | Apple Inc. | Synchronization and task delegation of a digital assistant |
US11405466B2 (en) | 2017-05-12 | 2022-08-02 | Apple Inc. | Synchronization and task delegation of a digital assistant |
US10410637B2 (en) | 2017-05-12 | 2019-09-10 | Apple Inc. | User-specific acoustic models |
US10482874B2 (en) | 2017-05-15 | 2019-11-19 | Apple Inc. | Hierarchical belief states for digital assistants |
US10810274B2 (en) | 2017-05-15 | 2020-10-20 | Apple Inc. | Optimizing dialogue policy decisions for digital assistants using implicit feedback |
US11217255B2 (en) | 2017-05-16 | 2022-01-04 | Apple Inc. | Far-field extension for digital assistant services |
Also Published As
Publication number | Publication date |
---|---|
GB9910448D0 (en) | 1999-07-07 |
AU4590400A (en) | 2000-11-21 |
JP2002544552A (en) | 2002-12-24 |
EP1190417A1 (en) | 2002-03-27 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
WO2000068936A1 (en) | Cancellation of non-stationary interfering signals for speech recognition | |
Avargel et al. | System identification in the short-time Fourier transform domain with crossband filtering | |
EP1855457B1 (en) | Multi channel echo compensation using a decorrelation stage | |
US5933495A (en) | Subband acoustic noise suppression | |
JP4283212B2 (en) | Noise removal apparatus, noise removal program, and noise removal method | |
US8594320B2 (en) | Hybrid echo and noise suppression method and device in a multi-channel audio signal | |
EP0843934B1 (en) | Arrangement for suppressing an interfering component of an input signal | |
US9992572B2 (en) | Dereverberation system for use in a signal processing apparatus | |
US6157909A (en) | Process and device for blind equalization of the effects of a transmission channel on a digital speech signal | |
EP1715669A1 (en) | A method for removing echo in an audio signal | |
US20040264610A1 (en) | Interference cancelling method and system for multisensor antenna | |
CA2416128A1 (en) | Sub-band exponential smoothing noise canceling system | |
JP2003534570A (en) | How to suppress noise in adaptive beamformers | |
CN111213359B (en) | Echo canceller and method for echo canceller | |
EP0789476B1 (en) | Noise reduction arrangement | |
JP2003500936A (en) | Improving near-end audio signals in echo suppression systems | |
US8199928B2 (en) | System for processing an acoustic input signal to provide an output signal with reduced noise | |
US5241692A (en) | Interference reduction system for a speech recognition device | |
KR100470523B1 (en) | Process and Apparatus for Eliminating Loudspeaker Interference from Microphone Signals | |
JPH07283796A (en) | Signal source characterizing system | |
KR100386488B1 (en) | Arrangement for communication with a subscriber | |
JP3786038B2 (en) | Input signal processing method and input signal processing apparatus | |
JP2005514668A (en) | Speech enhancement system with a spectral power ratio dependent processor | |
KR100875264B1 (en) | Post-processing method for blind signal separation | |
KR101537653B1 (en) | Method and system for noise reduction based on spectral and temporal correlations |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
AK | Designated states |
Kind code of ref document: A1 Designated state(s): AE AG AL AM AT AU AZ BA BB BG BR BY CA CH CN CR CU CZ DE DK DM DZ EE ES FI GB GD GE GH GM HR HU ID IL IN IS JP KE KG KP KR KZ LC LK LR LS LT LU LV MA MD MG MK MN MW MX NO NZ PL PT RO RU SD SE SG SI SK SL TJ TM TR TT TZ UA UG US UZ VN YU ZA ZW |
|
AL | Designated countries for regional patents |
Kind code of ref document: A1 Designated state(s): GH GM KE LS MW SD SL SZ TZ UG ZW AM AZ BY KG KZ MD RU TJ TM AT BE CH CY DE DK ES FI FR GB GR IE IT LU MC NL PT SE BF BJ CF CG CI CM GA GN GW ML MR NE SN TD TG |
|
121 | Ep: the epo has been informed by wipo that ep was designated in this application | ||
DFPE | Request for preliminary examination filed prior to expiration of 19th month from priority date (pct application filed before 20040101) | ||
WWE | Wipo information: entry into national phase |
Ref document number: 2000927504 Country of ref document: EP |
|
ENP | Entry into the national phase |
Ref country code: JP Ref document number: 2000 617442 Kind code of ref document: A Format of ref document f/p: F |
|
REG | Reference to national code |
Ref country code: DE Ref legal event code: 8642 |
|
WWP | Wipo information: published in national office |
Ref document number: 2000927504 Country of ref document: EP |
|
WWE | Wipo information: entry into national phase |
Ref document number: 10030524 Country of ref document: US |
|
WWW | Wipo information: withdrawn in national office |
Ref document number: 2000927504 Country of ref document: EP |