CA2069675C - Flexible vocabulary recognition - Google Patents

Flexible vocabulary recognition Download PDF

Info

Publication number
CA2069675C
CA2069675C CA002069675A CA2069675A CA2069675C CA 2069675 C CA2069675 C CA 2069675C CA 002069675 A CA002069675 A CA 002069675A CA 2069675 A CA2069675 A CA 2069675A CA 2069675 C CA2069675 C CA 2069675C
Authority
CA
Canada
Prior art keywords
model
word
phoneme
sequence
network
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Expired - Lifetime
Application number
CA002069675A
Other languages
French (fr)
Other versions
CA2069675A1 (en
Inventor
Vishwa Nath Gupta
Matthew Lennig
Patrick Joseph Kenny
Christopher Kenneth Toulson
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Bell Canada Inc
Original Assignee
Bell Canada Inc
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Bell Canada Inc filed Critical Bell Canada Inc
Publication of CA2069675A1 publication Critical patent/CA2069675A1/en
Application granted granted Critical
Publication of CA2069675C publication Critical patent/CA2069675C/en
Anticipated expiration legal-status Critical
Expired - Lifetime legal-status Critical Current

Links

Classifications

    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
    • G10L15/00Speech recognition
    • G10L15/08Speech classification or search
    • G10L15/14Speech classification or search using statistical models, e.g. Hidden Markov Models [HMMs]
    • G10L15/142Hidden Markov Models [HMMs]
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
    • G10L15/00Speech recognition
    • G10L15/08Speech classification or search
    • G10L15/14Speech classification or search using statistical models, e.g. Hidden Markov Models [HMMs]
    • G10L15/142Hidden Markov Models [HMMs]
    • G10L15/144Training of HMMs
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
    • G10L25/00Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00
    • G10L25/03Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00 characterised by the type of extracted parameters
    • G10L25/24Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00 characterised by the type of extracted parameters the extracted parameters being the cepstrum

Landscapes

  • Engineering & Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • Probability & Statistics with Applications (AREA)
  • Computational Linguistics (AREA)
  • Health & Medical Sciences (AREA)
  • Audiology, Speech & Language Pathology (AREA)
  • Human Computer Interaction (AREA)
  • Acoustics & Sound (AREA)
  • Multimedia (AREA)
  • Telephonic Communication Services (AREA)
  • Machine Translation (AREA)

Abstract

A flexible vocabulary speech recognition system is provided for recognizing speech transmitted via the public switched telephone network. The flexible vocabulary recognition (FVR) system is a phoneme based system. The phonemes are modelled as hidden Markov models. The vocabulary is represented as concatenated phoneme models. The phoneme models are trained using Viterbi training enhanced by: substituting the covariance matrix of given phonemes by others, applying energy level thresholds and voiced, unvoiced, silence labelling constraints during Viterbi training. Specific vocabulary members, such as digits, are represented by allophone model. A* searching of the lexical network is facilitated by providing a reduced network which provides estimate scores used to evaluate the recognition path through the lexical network. Joint recognition and rejection of out-of-vocabulary words are provided by using both cepstrum and LSP parameter vectors

Description

FLEXIBLE VOCABULARY RECOGNITION
This invention relates to a method of speech recognition and is particularly concerned with such speech recognition for user-independent and flexible vocabulary applications.
Statistical methods of speech recognition are often based on the assumption that the sequence of acoustic observations corresponding to an unknown utterance to be recognized is accounted for by a path through a hidden Markov Modal (HMM) which serves as a speech production model. The recognition problem is usually formulated as one of finding the sequence of states in the hidden Markov chain 1 0 whose posterior probability (given the acoustic observations) is maximal.
There is a large computational burden associated with such a scheme, which in systems of medium or larger vocabulary (100 words or greater}, and particularly systems intended for speaker-independent use, may inhibit the ability of such systems to respond in real time intervals, 1 5 -the use of !-1MM is well known in the art as described in the article by S.E. Levinson, L.R Rabiner and M.M. Sondhi (1983) entitled: "An introduction to the application of the theory of probabilistic functions of a Markov process to automatic speech recognition" B.a."f.J., 62(4), 1035-1x74.
Other problems arise when applying speech recognition techniques in
2 0 systems for recognition of words spoken over a public switched telephone network.
An object of the present invention is to provide an improved method of speech recognition for medium vocabulary.
In accordance with an aspect of the present invention there is provided a method of preparing phoneme models for recognition of speech received via 2 5 telephone lines camprising the steps of: a) analyzing a training word to generate a frame sequence of acoustic parameter vectors representative thereof; b) providing a first set of mode! parameters representative of a sequence of state-transition models corresponding to the training word, the parameters including a mean vector and weighting factor for each transition and a covariance matrix for each model;
3 0 c) computing a set of observation probabilities for the training word and the first set of model parameters; d) aligning the frame sequence of acoustic parameter vectors with the sequence of state-transition models to provide a mapping therebetween representative of a path through the sequence of state-transition models having a highest likelihood; e) accumulating statistics for a plurality of utterances of said 3 5 training word using the mapping of step d); f) generating a second set of model ~~~'~~~
parameters representative of the sequence of state-transition models corresponding to the training word; g) repeating step d) for the second set of model parameters;
h) comparing the likelihood of the first and second sets of model parameters;
and i) repeating step b) through h), replacing the first set of mode! parameters by the second set of model parameters when the second set of parameters provides at least a predetermined improvement in likelihood; and j) substituting the covariance matrix of a first model with the covariance matrix of a second madei to provide a smooth covariance matrix thereby improving recognition accuracy for the first model.
In accordance with another aspect of the present invention there is 1 0 provided a method of preparing phoneme models for recognition of speech received via telephone lines comprising the steps of: a) analyzing a training word to generate a frame sequence of acoustic parameter vectors representative thereof, having a parameter indicative of energy level for each frame of the frame sequence;
b} providing a first set of model parameters representative of a sequence of state-'I 5 transition models corresponding to the training word; c) computing a set of observation probabilities by, for each frame of the frame sequence and each model of the sequence of state-transition models, comparing the energy level of the frame with a predetermined, relative to noise on the telephone lines, energy threshold for the model, and if the energy level is below the energy threshold, setting the observation 2 0 probability for the frame to zero, otherwise computing the observation probability for the frame; d) aligning the frame sequence of acoustic parameter vectors with the sequence of state-transition models to provide a mapping therebetween representative of a path through the sequence of state-transition models having a highest likelihood; e) accumulating statistics for a plurality of utterances of said 2 5 training word using the mapping of step d); f) generating a second set of model parameters representative of the sequence of state-transition models corresponding to the training word; g) repeating step d) for the second set of model parameters;
h) comparing the likelihood of the first and second sets of model parameters;
and i) repeating step b) through h), replacing the first sat of model parameters by the 3 0 second set of model parameters when the second set of parameters provides at least a predetermined improvement in likelihood.
In accordance with a further aspect of the present invention there is provided a method of preparing phoneme models fbr recognition of speech received via telephone lines comprising the steps of: a) analyzing a training word to generate a 3 5 flame sequence of acoustic parameter vectors representative thereof and to label each frame of the frame sequence as voiced, unvoiced or silence; b) providing a first set of model parameters representative of a sequence of state-transition models corresponding to the training word, including the step of IabellirDg each model as voiced, unvoiced or silence in dependence upon a phoneme represented by the model and a relative position of the model in the sequence; c) computing a set of observation probabilities by, for each frame of the frame sequence and each model of the mod~I
sequence of state-transition models, comparing a voiced-unvoiced-silence (1/LJS) label of the frame with a ~1US labe6 of the model and if the labels do not match, setting the observation probability for the frame to zero, otherwise computing the 1 0 observation probability for the frame for the training word and the first set of model parameters; d) aligning the frame sequence of acoustic parameter vectors with the sequence of state-transition models to provide a mapping therebetween representative of a path through the sequence of state-transition models having a highest likelihood; e) accumulating statistics for a plurality of utterances of said 7 5 training word using the mapping of step d); f) generating a second set of model parameters representative of the sequence of state-transition models corresponding to the training word; g) repeating step d) far the second set of model parameters;
h) comparing tha likelihood of the first and second sets of model parameters;
and i) repeating step b) through h), replacing the first set of model parameters by the 2 0 second set of model parameters when the second set of parameters provides at least a predetermined improvement in likelihood.
In accordance with a furthor aspect of the present invention there is provided a method of speech recognition for speech received via telephone lines comprising the steps of: a) analyzing an unknown utterance to generate a frame 2 5 sequence of acoustic parameter vectors representative thoreaf; b) providing a first network representing a recognition vocabulary, wherein each branch of the first network is a model representing a phoneme and each complete path through the first network is a soquence of models representing a word in the recognition vocabulary;
a) providing a second network derived from the first network, in which all sequences 3 0 of three consecutive phonemes present in the first network are present;
d) computing transitional probabilities for each node of the second network given the frame sequence of acoustic parameter vectors; e) searching the second network to determine optimal cumulative probabilities for each node of the second network for all frames of dhe frame sequence; f) storing the cumulative probabilities as estimate 3 5 scores for estimating partial paths in the first network; g) computing point scores
4 for all phonemes in dependence upon the frame sequence of acoustic parameter vectors; and h) determining a complete path through the first network by evaluating successive one phoneme extensions of partial paths using the estimate scores for the nodes of the second network to find the partial path to extend.
In accordance with a further aspect of the present invention there is provided apparatus for speech recognition, comprising: a) means for analyzing an unknown utterance to generate a frame sequence of acoustic parameter vectors representative thereof; b) means for providing a first network representing a recognition vocabulary, wherein each branch of the first network is a model 1 0 representing a phoneme and each complete path through the first network is a sequence of models representing a word in the recognition vocabulary; c) means for providing a second network derived from the first network, in which all sequences of three consecutive phonemes present in the first network are present; d) means for computing transitional probabilities for each node of the second network given the 1 5 frame sequence of acoustic parameter vectors; e) means for searching the second network to determine optimal cumulative probabilities for each node of the second network for all frames of the frame sequence; f) means for storing the cumulative probabilities as estimate scores for estimating partial paths in the first network;
g) means for computing point scores for all phonemes in dependence upon the frame 2 0 sequence of acoustic parameter vectors; and h) means for determining a complete path through the first network by evaluating successive one phoneme extensions of partial paths using the estimate scores for the nodes of the second network to find the partial path to extend.
A flexible vocabulary speech recognition system~can be defined as having 25 the ability to configure the recognizer to recognize new vocabulary members without having to retrain for the new members.

~~'~~
The present flexible vocabulary recognition (FVR) system is intended to recognize words spoken over the public switched telephone netwark. The system is designed to be flexible in vocabulary size and to recognize words spoken by any speaker. In order fio provide flexible vocabulary, the system recognizes wards as a
5 sequence of phonemes. There are approximately 40 phonemes which can generate any word in American English.
~~f de~,or' t'~~ the prawinas The present invention will be further understood from the following description with reference to the drawings in which:
1 0 Fig. 1 illustrates a Markov model;
Fig. 2 illustrates a Markov model of a phoneme;
Fig. 3 illustrates in a flow chart the general steps of a method known in the art as Viterbi training;
Fig. 4 graphically illustrates the most likely path of state transitions 1 5 thraugh the phoneme model sequence found by the well known Viterbi method;
Fig. 5 illustrates in a flow chart the method of covariance matrix substitution in accordance with an embodiment of the present invention;
Fig. 6 illustrates in a flow chart the method of including energy thresholds for the observed vector in the Viterbi training method in accordance with 2 0 an embodiment of the present invention;
Fig. 7 illustrates in a flow chart the method of including voiced unvoiced, and silence classifications in the Viterbi training method in accordance with an embodiment of the present invention;
Fig. 8 graphically ilBustrates a lexical network for a vocabulary 2 5 consisting of the words: rabid, tabby and gabby;
Fig. 9 illustrates a trigram network corresponding to the lexical network of Fig ~;
Fig. 10 graphically illustrates a trellis for staring the best cumulative probability of each state in the trigram network;
3 0 Fig. 11 illustrates a functional block diagram of the flexible vocabulary recognition system in accordance with an embodiment of the present invention;
Fig. 9 ~a illustrates in a flow chart the sequence of frame synchronous steps executed by the speech recognition system of Fig. 11;
Fig. 12b iNustrates in a flow chart the sequence of steps of the A' search ~ 5 executed by the speech recognition system of Fig. 11;
6 Fig. 13 graphically illustrates an array of point scores;
Fig. 14 graphically illustrates the A* search method;
Fig. 15 illustrates in a flow chart the method of recognizing using both cepstral and LSP parameters in accordance with an embodiment of the present invention; and Fig. 16 illustrates in a flow chart the method of rejecting out-of-vocabulary utterances in accordance with an embodiment of the present invention.
A flexible vocabulary recognition system presents several challenges. To provide flexibility in the vocabulary, that is, the ability to delete, add, and substitute l0 vocabulary words, without incurring the burden of retraining for those words, vocabulary members are described in terms of their elemental phonemes.
In accordance with the present invention, the recognition system is based on hidden Markov modelling (HMM). The recognition system will be described in two parts, the first is training of hidden Markov models for phonemes (phoneme models), and the second is recognition of the unknown utterance.
A Markov Model is a set of states with transitions describing possible paths within the states that model some phenomena. Every time a transition occurs, some characteristic vector is emitted. Referring to Fig. 1, there is illustrated a Markov model having four states, 10, 12, 14 and 16. Transitions within the model 2o are shown as directed lines. The state 10 has three possible transitions, a self transition 18, a next state transition 20, and a skip state transition 22. The state 12 also has self, next and skip transitions, 24, 26, 28 respectively. The state 14 has self and next transitions, 30 and 32, while the state 16, representing the end of the phoneme has only a self-transition 34 and a null transition 36. The null transition 36 merely provides a link to a next phoneme model. In the present invention, the assumption is that each phoneme in a given utterance is generated by a Markov Model.
Every time a transition is made in the Markov Model, a cepstrum vector, representing the utterance, is emitted.
In the case of Hidden Markov Models, it is further assumed that there is no access to the original Markov process which generated the emitted vectors, there is only the observation of the vectors themselves.
' In the present invention, the FVR system is based on first order Markov Models. That is, the probability of entering a given state is dependent only on the
7 immediately previous state (i.e. not earlier states). Given this, there is a transition probability (mean vector and covariance matrix) associated with each transition.
The probability of proceeding on a particular transition can be calculated given an observation vector. The idea, then, is to maximize the cumulative probability through the state machine given the observed vectors.
The training method estimates the parameters of phoneme models from a large number of utterances. 1~r11odels for these phonemes are trained in a vocabulary-independent and speaker-independent manner, so that new words can be added to the vocabulary without any need for retraining. Speaker-independence is achieved by i 0 training the phoneme models from a large number of speakers, while the vocabulary-independence is achieved by training the phoneme models from a large number of words not in the reaognitian vooabulary. A system having a flexible vocabulary is desirable in many applications where the vocabulary changes dvith time, for example, in applications where queries about companies listed in a stock exchange have to be answered.
For the present invention, as described above, phonemes are modelled by first order hidden Markov processes. Conceptually, traversing the states in the model may be likened to progressing through the various portions of a phoneme (onset, body, termination).
2 0 'I~he phoneme models are left-to-right (or Bakis) I-+MMs as described in the article by F. Jelinek; (196) "Continuous speech recognition by statistical methods", Proceedings of the IEEE, 64(4), pp. 53~-556, and the number of states in the phoneme models vary from 3 to i 0 depending on the phoneme. The phoneme models are transition based, with each transition output probability represented by a 2 5 Gaussian mixture distribution as described in Deng et al. (1990) "The Acoustic Recognition Component of the II~RS-Telecom 86,000-word Speech Recognizer", Proceedings of the IEEE International Conference on Acoustics, Speech and Signal Processing, pp. 741-744: This Gaussian mixture is represented by 12 moans and a covariance matrix. One pooled covariance matrix is estimated for all the transitions 3 0 in the phoneme model.
Referring to Fig. 2, there is illustrated a Markov model of a phoneme.
Transitions within the phoneme models, then, are constrained as follows:
1 ) the initial state 10 has only a self loop '18 as an entry path; and a null transition 38 from the last state of the previous 3 5 phoneme.
8 2) the second state 12 has only the previous state transition 30 or a self loop 24 as entry paths, 3 ) the last state 15 has no self loop as an entry path, and has a null transition 36 to the next phoneme.
4 ) all other states 14 have a setf loop ~~, a previous state 26, and a skip last state 22 transition as entry paths.
oPJith these constraints, the last state 16 in each phonefrre maybe considered a "colteotor" and its cumulative probability may be propagated to the first state of all connected phonemes via the null transition ~6.
1 g In order to reduce the size of the data the covariance matrix is pooled across the complete phoneme. Also, the covariance matrix is pooled because there is insufficient data available to gbnerate covariance matrices for each transition. Thus, all modes for all transitions in a given phoneme share a common covariance matrix.
The specific data then; associated with a model is as fellows:
1 5 - a 15 x 15 matrix of floating point numbers that is a covariance matrix.
a set of states, S, and associated transitions, T, which define the allowable next states. A model has, on average, 5 states with 11 transitions.
2 0 Each transition eleme~tt has a number of mixtures or modes, AA, associated with it. The mode consists of a floating point number W, that represents the weight and an array of 15 floating point numbers, that represents the moan.
The modes or mixtures can be considered to describe a general class of speech type (e.g. male; female, high pitched, etc). In the present embodiment of the 2 5 EVFf system, 12 mixtures are used. Each mixture can have its own moan, covariance and weight. The weight can be considered as the general probability of using that mode for the given tra~sitiora. The transition probability is computed from the following formula:
max 1 2 ~x w jt ~ 1 ~x hi) w.
3 0 12 ~ ~~~d/2 ~ ~ ~-1/2 where:
x is the observation vector ~;i is the mean vector for mode i Phoneme Model wi is the weight for mode i parameters ~ is the covariance matrix ~ 5 t indicates transpose d is the vector dimention
9 To estimate parameters for these phoneme models, Viterbi training is used. In Viterbi training, the sequence of most likely state transitions through the string of phoneme models corresponding to the words in the training set is found. The mapping thus obtained between the state transitions and the feature parameter vectors is used to estimate the parameters of the phoneme models.
Referring to Fig. 8, there is illustrated in a flow chart, the general steps of a method known in the art as Viterbi training. Training is initiated at 70 once samples of spoken words have been collected for a set of training words. each training word, 7~, is represented by a phoneme sequence 7~. The phoneme sequence is then 1 0 aligned with the observed feature vectors for that word using the well known Viterbi method, 76. Once the phoneme model transitions have been mapped to the observed feature vectors (one transition to one vector), statistics are accumulated using such mapping from the word spoken by various speakers, 78. The statistics are used to modify the phoneme modal parameters 80. The old model is compared to the new one 9 5 by computing the likelihood for each model given the observod feature vectors, 82. If the difference is below a predetermined amount, the training is complete, 84.
Otherwise, the process is repeated beginning with block 72.
The step represented by box 76 will be described in greater detail in connection with Figs: 6 and 7, that illustrate modifications of the standard Viterbi 2 0 training method in accordance with the present invention.
Referring to Fig. 4, there is graphically illustrated the most likely path of state transitions through the phoneme model sequence found by the well known Viterbi method. In the Viterbi method the states of the phoneme model sequence for an input word are labelled sequentially and plotted along one axis (in Fig. 4, the y-axis).
2 5 The frame numbers for the corresponding observed feature vector sequence are plotted sequentially along another axis (in Fig. 4, the x-axis). For each transition between states, for each frame, the probability is calculated and stored in a trellis.
Once all such probabilities are computed, a path is traced from beginning to end of the sequence. The path represents the highest likelihood path. The path provides a 3 0 mapping between the transitions of the phoneme model sequence and the observed feature vector sequence.
In the present invention, the training method has been modified to obtain phoneme models that improve recognition accuracy significantly. The modifications include substitution of covariance matrix following training, use of energy thresholds
10 during training phoneme models, and imposition of voiced-unvoiced-silence constraints during training of the phoneme models.
Referring to Fig. 5, there is illustrated in a flow chart the method of covariance matrix substitution in accordance with an embodiment of the present invention. A box 100 represents the well known Viterbi training method as described hereinabove in connection with Fig. 3. A box 102 represents the substitution of covariance matrices. In step 102a) the covariance matrix for left-silence or utterance-initial silence, represented by the symbol { and the right-silence or utterance-final silence, represented by the symbol } is replaced by the covariance 1 0 matrix for the phoneme f. in step 102b) the covariance matrix for the phonemes /~ ~ u/ is replaced by the covariance matrix for the phoneme a. In step 102c) the covariance matrix for the phoneme a before r is replaced by the covariance matrix for the phoneme a . The reasons for these covariance matrix substitutions is described below.
1 5 The phoneme models trained include models for utterance-initial and utterance-final silences. (Table 1 shows a complete list of the phoneme models.) Even though the utterance-initial and utterance-final silence models are trained from a large amount of training data, the covariance matrices for the silence models are sharper than other phoneme models. These covariance matrices can cause recognition 2 0 errors if the utterance-initial or utterance-final silence segments contain impulse noise, cough, breath noise, aspiration, or silence segments due to a naisy telephone line. The recognition errors arise from the silence segments, which include noise, more closely matching the phoneme models than to the utterance-initial or utterance-final silence models. The recognition errors for the siionce modols can be mitigated 2 5 by reducing the sharpness of the covariance matrices for utterance-initial and utterartee-final silences. The preferred method is to substitute the covariance matrix for the ph~neme Ifl model for the covariance matrices for the utterance-initial and utterance-final silences . The spectral characteristics for phoneme /f/ look similar to those of telephone noise for a noisy telephone channel, for breath noise, and for 3 0 aspiration. Such a substitution of covariance matrices results in dramatic improvement in recognition accuracy. (Substitution of the utterance-initial and utterance-final covariance matrices by matrices of other fricative phonemes also works, but not as well as that by phoneme /f/.)
11 phoraea~ne exurazple phonergae example ~vord word /a j/ I /aw/ out /o j/ boy /a/ far /i/ eat /z/ bit /e/ bait /s/ bet /ae/ bat /s/ before /r/ air ~n/ but /u/ put /u/ hoot /o/ coal a (unstressed)/j/ you /w/ was /1/ live 1 /r/ are inter-word silence ~

utterance-initial utterance-final silence silence /p/ pit /b/ bit /t/ t~ /d/ dog /k/ could /g/ go /t.f/ church /d~/ judge /f / fill /v/ have /~/ thing /~/ the /s/ say /z/ zoo /.f/ shoe /3/ garage /m/ me /n/ no going /h/ he 'f able 1 List of all the phoneme models used in flexible vocabulary recognition. The phonemes are explained through an example 2 5 word in which they accur.
The phonemes /~ 3 u/ do not occur frequently in the training data, as a consequence the covariance matrices for these phoneme models are sharper than the covariances for other phonemes that occur more frequently. Hence, many recognition 3 0 errors occur for utterances containing phonemes /~ 3 u/. To reduce the sharpness of the covariances for these phonemes, the covariance matrix for the phoneme /3/is substituted for the crrvariance matrices for these phonemes . The covariance matrix for /a/ has the least sharpness and provides the best improvement in recognition accu racy.
12 The phoneme lei before /rl has its own phoneme model. Fven though this phoneme occurs quite frequently, its covariance matrix is quite sharp. The reason is that the spectral characteristics within the phoneme are quite uniform, but the spectral characteristics across different samples of the phoneme vary considerably.
Reducing the sharpness of the covariance matrix for /e/ before /r/, reduces recognition errors for words which contain /s/ before /r/. This is accomplished by substituting the covariance matrix for lei by covariance matrix for /a/.
The smoothing of covariance matrices, as described above, is done after the Viterbi training method.
1 0 As described hereinabove, the Viterbi training method is used to estimate the parameters of the phoneme models During the Viterbi training method, boundary markers for each phoneme in the spoken utterance are established. These boundaries correspond to the sequence of phonemes making up the utterance. Many of these boundaries turn out to be incorrect. For example, speech segments identified as 1 5 vowels may actually contain silent segments or parts of fricative segments. The effect of errors in phoneme boundaries is to train phoneme models from speech segments not corresponding to those phonemes. Some of these errors may be eliminated by using energy thresholds. For example, to avoid vowel phonemes from mapping to noise or weak fricatives, an energy threshold is used. Only those segments of speech 2 0 with energy greater ttaan the energy threshold can map to the vowels.
Since, in the public switched telephone network, the background noise level can vary from one dialled-up line to another, the energy thresholds are set relative to the noise level an the telephone line.
Referring to Fig. 6, there is illustrated in a flow chart the method of 2 5 including energy thresholds for the observed vector in the Viterbi training method.
The box 76 of Fig. ~ is subdivided into two boxes 110 and 112. The box 110 computes observation probabilities and the box 112 aligns the phoneme mode!
sequence with the observed feature vector. The method begins with 114. A box initiates the method by examining each phoneme model, frame by frame. A
decision 3 0 box 118 compares the frame energy level with the phoneme model energy threshold.
If the vector energy is greater than or equal to the threshold, a box 120 computes the observation probabilities as in the standard Viterbi method. If the vector energy is below the threshold, a box 24 sets the probability to zero. The method ends with 122.
The energy thresholds for different phonemes are outlined as follows 3 5 (noise represents energy level for noise on the line:

1. Vowel thresholds (except ! a!):
i. for vowel in first four phonemes, energy > noise + lOdB, i i . else if utterance > 10 phonemes, vowel in last 4 phonemes, energy > noise +ldB, i i i . else energy > noise + 6dB.
2. ! a! thresholds:
i. if in first four phonemes, energy > noise +6dB, 1 0 if. else if utterance > 10 phonemes, !a! in last 4 phonemes, energy > noise + ldB, i i i . else energy > noise +3dB.
/I r j ur/ threshoi~s:
i , if utterance > 10 phonemes, phoneme in last 4 phonemes, 1 5 energy > noise + ldB.
i i . else energy > noise + 3dB.
a. if/f v' 0 a h/ in first four phonemes, energy > noise + ldB.
5. if /s z n m ~/~ energy > nois~ +ldB.
6. /J ~/ thresholds:
2 0 i . if utterance > '10 phonemes, phoneme in last 4 phonemes, enemy > noise+ ldB, i i . ~Ise energy > noise + ~dB.
P~ote that some of the energy thresholds are dependent on phonemic 2 5 context. Those energy thresholds are applied during Viterbi training.
During Viterbi training, the likelihood of every speech frame for all possible transitions in the phoneme models is computed. Far speech frames whose energies are lower than the energy threshold #or that phoneme, the likelihood of that speech frame for ail possible transitians through that phoneme is set to zero. Application off such energy thresholds 3 0 during training results in phoneme boundaries which are much more reasonable. The phoneme models thus obtained rosult in significant improvement in recognition accuracy.
The use of energy constraint has been described in connection with the Viterbi method, however, these constraints will else work with the so-called Forward-~ 5 Backward training method.

During training, another method of improving the wrapping of phoneme modals to speech segments is by preclassifying of speech into voiced, unvoiced, or silence categories. Then the phoneme models are constrained to map to speech segments of a specific category. These constraints can be imposed in the same fashion as the energy constraints during Viterbi training. Such constraints improve the mapping of phoneme models to speech segments during training, resulting in phoneme models which improve recognition accuracy significantly.
Referring to Fig. 7, there is illustrated in a flow chart the method of including voiced, unvoiced and silence classifications in the Viterbi training method.
1 0 As in Fig. S, the box 76 of Fig. 3 is subdivided into two boxes 110 and 112. The method begins vrith 140. A box 142 initiates the method by examining each phoneme model, frame by frame. A decision box 144 compares the frames (VUS) label with the phone mode! {VUS) label. If the labels are the same a box 148 computes the observation probabilities. If the labels differ, the observation probability for that 1 5 frame is set to zero. The method ends with 150. Thus, the computation of observation probabilities computed at 110 in the Viterbi method are computed in accordance with steps 140 through 150.
One issue here is how to classify speech into voiced, unvoiced or silence category. A voicedlunvoiced/silence (VUS) classifier known in the art from the 2 0 article entitled "Application of an LPC Distance Measure to the Voiced-Unvoiced Silence Detection Problem" by L.Fi. Ftabiner and L.M.R. Sarnbur (1977) IEEE
Trartsactior~s on Acoustics, Speech, and Signal Processing, ASSP-45(4), 338-343, has been shown to perform well. This classifier is used with some additional improvements. The classifier may provide erroneous classification when a strong 2 5 fricative has high aner~y. ~ften segments are classified as voiced even though the correct decision should be unvoiced. A pitch detector based on the well known average magnitude difference function (AMDF) algorithm published in the article entitled "Average magnitude difference function pitch extractor", by M.J. Floss et al.
(1974), IEEE Transactions on Acoustics, Speech and Signal Processing, ASSP-22, 353-362 3 0 applied to linear predictive coefficients (LPC) residual signal has bean used in order to correct these errors.
The question is what mapping between phoneme models and speech segments classified as either voiced, unvoiced, or silence, would result in improved mapping of phoneme models to speech segments. Use of linguistically motivated 3 5 mappings (vowels, nasals, voiced fricatives mapping to voiced segmehts, unvoiced fricatives mapping to unvoiced segments, silence mapping to silence segments, etc.) would actually result in additional errors in mappings between phoneme models and speech segments. The reason is that the VUS classifier makes many errors (approximately 5% of the speech segments are misclassified). In order to eliminate mapping errors due to the misclass'sfications by the VUS classifier, mappings are formulated that take into account errors caused by the VUS classifier. Many of the VUS classification errors are predictable (for example, phoneme /i/ and phoneme sequence / ju/ are classified as unvoiced in many instances). These problems are mitigated by including the performance of the VUS detector in the mappings.
The 1 0 mapping used during Viterbi training of the phoneme models is as follows:
1, vowels and /i r w/:
i . map to voiced or unvoiced speech segments in initial position or if in last four phonemes when the total number of phonemes in the utterance is greater than 10.
1 5 i i . map to voiced segments everywhere else, except some phonemes map to voiced or unvoiced segments in special contexts:
a. phoneme /rl when followed by a vowel and preceded by /t/ or /f/.
b. phoneme/w/ when preceded by /k/.
c. phoneme /i/ when followed by inter-word silence.
2 0 d. phoneme /t/ when preceded by /d/ or inter-word silence.
e. phoneme lEl when preceded by inter-word silence.
f. phoneme /ul when preceded by/j/.
g. phoneme /a/ when preceded by lsl, Ijl, Irl, ldl, or inter-word silence.
2 5 2. phonemes /n m x~/ are voiced or unvoiced if not in last four phonemes when the total number of phonemes in the utterance is greaser than 10, otherwise voiced, unvoiced, and silence.
3. /~l is always voiced or unvoiced.
4. The intervocalic silence is always unvoiced or silence.
3 0 Note that the initial vowel can be either voiced or unvoiced. This mapping is necessary as it has been observed that initial 20 ms of the utterance-initiai vowel is classified as unvoiced in many instances. The utterance-initial and utterance-final silence models can map to speech segments labelled as either voiced, unvoiced or silence. This is due to the fact that impulse noise, and sometimes strong 3 5 breath or aspiration noise can be classified as voiced or unvoiced. When the word ~~~~~

consists of ten or more phonemes, it has been observed that the last three or four phonemes get classified as unvoiced, even though they are really voiced. This is evident from the mappings shown above, where the last four phonemes can map to unvoiced segments, even though they are voiced. All the errors caused by the VUS
classifier are reflected in the mapping rules. The mapping rules impose constraints on the VUS classifier that result in improvements in mapping phoneme models to speech segments.
The second part of the flexible vocabulary recognition system is recognition of an unknown utterance. In the present invention unknown utterances 1 0 are converted to acoustical parameter vectors, for example cepstrum vectors. As described above, Maricov models of the phonemes are trained for use during recognition.
The vocabulary is represented by a lexical network. Within the lexical network, each branch in the network represents a phoneme. Every path through the 1 5 network must define a vocabulary member and alt vocabulary members must be represented. Path merging may occur, but only if no over-generation or member deletion results.
Iqeferring to Fig. 8, there is graphically illustrated a lexical network for a vocabulary consisting of the w~rds: rabid, tabby and gabby. Ths lexical network 2 0 includes start and end nodes 160 and 162, respectively, and paths 164, 166, and 168. Each word in the vocabulary is represented by a path from the start node to the end node 162. The path 164 (for the word rabid) consists of a sequence of branches 170, 172, 174, 176 and 178, interconnecting nodes 160 and 180, '180 and 182, 182 and 184, 184 and 186, and 186 and 162 respectively. Similarly, the 2 5 path 166 (for the word tabby) consists of a sequence of branches 190, 192, 194 and 196, interconnecting nodes 160 and 198, 198 and 200, 200 and 202, and 202 and 162, resp6ctively. The path 188 (for the word gabby) consists of a branch 204 representing g and a branch 208, a null branch which merges the path 168 with the path 166 at the node 198. Thus, the path 168 consists of a sequence of branches 3 0 204, 206, 192, 194, and 196 interconnecting nodes 160 and 208, 208 and 198, '9 98 and 200, 200 and 202, and 202 and 162, respectively.
The purpose of any speech recognition system is to convert speech into text. The recognition problem is usually formulated as one of finding the sequence of states in the hidden ~Ilarkov chain whose posterior probability (given the acoustic ~ 5 observations) is maximal.

Two possible statistical speech recognition methods can be used: Viterbi recognition method (Levinson et al., 1983), and stack decoding or A* search method (Jelinek, 1976). Both these methods recognize the unknown utterance as the most likely path through a lexical network. In the present invention, the lexical graph has phoneme branches, and any complete path through the graph corresponds to a phoneme sequence of a complete utterance. The utterance could be a word or a sequence of words. It is well known in the art to use the Viterbi recognition method to find the most likely path through the lexical network. However, this method suffers from several drawbacks. Because it performs an exhaustive search, for phoneme-1 0 based recognizers with large vocabularies, the search space can be very large and the search prohibitively expensive, both in time and CPU power. A further limitation is it generates only ~ne recognition hypothesis. Also, imposing constraints at the segment level {for example, duration constraints) causes a substantial increase in computations.
1 5 The principal alternative to the Viterbi method is the A* or stack decoding method. This method was applied to speech recognition by IBM in the 1970's> using the formula dictated by the Bayes decision rule to score recognition hypotheses (Jelinek, 1976). The principal advantage of the A* method is that multiple recognition hypotheses can be generated very easily. The principal 2 0 disadvantage is that ~t will be inefficient unless partial paths through the lexical network can be estimated with sufficient accuracy.
In accordance with the present invention a method is provided to estimate scores for partial paths while reducing computational burden. To understand the need for these partial path estimates, a brief discussion of the A* search method is 2 5 provided. The A* searoh method is an optimal search procedure to find the best path through a lexical network. The A* search method searches through a lexical network by scoring partial paths; and extending the most promising partial paths, until a complete path with the highest score is obtained. In scoring a partial path, an estimate score is used. The estimate score corresponds to an estimate of the score for 3 0 the most likely extension of the partial path to form a complete path. The closer the estimate score is to the most likely path extension (but not less than this score), the faster is the search. The article by P. I(enny et al., entitled "A* -admissible heuristics for rapid lexical access", Proceedings of 1991 International Conference on Acoustics, apeech and Bignal Processing, pp. 689-692, provides a mathematical explanation of the method of searching the Bexical network and is hereby incorporated by reference.
A most precise method would be to use the Viterbi method together with the lexical network to compute the estimate scores. However, this would result in the same computing complexity as that of the Viterbi recognition method. Deduction of the computing required can only be accomplished by basing the estimator on a much smaller network. The lexical network can be reduced by partitioning its nodes into equivalence classes. If fewer equivalence classes are used, then the computing involved in estimating the partial path scores becomes smaller, but the search ~ 0 becomes less efficient. '~Uhen more equivalence classes are used, the search is more efficient, but estimator requires more computation. fin effective compromise is an equivalence network with two phone equivalence. This results in significant reduction in computing as compared to the Viterbi recognition method.
Eaci1 branch in the network used for the estimator does not need to 1 5 uniquely define a vocabulary member. That is, the network may over-generate vocabulary members; its paths can define more vacabulary members than there are in the original vocabulary. As long as there is a mapping between the full and the reduced network, any rules may be used in generating the reduced network. in the present invention, the rule used in defining the reduced network is as follows:
2 0 all sequences of three consecutive phonemes found in the original vocabulary are present in the vocabulary generated by the reduced network.
Hence, this network is termed the trigram (or tri-phone) network. The trigram network is necessary for a "two phone look ahead" technique that results in 2 5 reduced computing requirements. One may easily determine a mapping function between the full and reduced networks by defining adjacent phoneme sequences.
It should be noted, Though, that there are some minor differences between the construction of this network and that of the full vocabulary network.
First, the trigrarrt network maps the phoneme models to the nodes rather than to branches.
3 0 Second, each node in the trigram carries a phoneme pair label. Only the model of the second member of the pair is used in the probability calculations, hence the node is similar to a branch in the full network.
Deferring to Fig. 9, there is illustrated a trigram network corresponding to the lexical network of Fig. B. The trigram network includes hipbone nodes which ~ 5 begin with a starting node 210 for the phoneme { (initial-silence). For each complete path (164, 166, and 168) in the lexical network there is also a complete path in the trigram network. The path 164 (rabid} maps to the path consisting of nodes 210, 212, 214, 216, 218, 220, 222, and 232. The path 166 (tabby) maps to the path consisting of nodes 210, 224, 226, 216, 228, 230, and 232. The path 168 (gabby} maps to the path consisting of nodes 210, 234, 236, 216, 230 and 232. Unlike the lexical graph in which each complete path corresponds to a word in the vocabulary, paths exist in the trigram network that form non words for example nodes 210, 212, 214, 216, 228, 230, and 232 form rabby. Similarly, nodes 210, 234, 236, 216, 218, 220, 222, and 232 form gabid.
1 0 The Viterbi search over the trigram network is used to generate the estimator function required for the A* search. Essentially, an estimate is required for the probabilistic distance from each node (in the full lexical network) to the beginning of the utterance ( the A* search goes backwards through the utterance) from each point in time: The trigram search, then, must generate the probabilities 1 5 for these estimates.
In fact, what the trigram search will do is generate accurate numbers of these probabilities for the over-generated vocabulary. The reason that they are only estimates for the complete network is that there is not a one-to-one mapping between the nodes in the trigram network and the branches in the lexical network.
2 0 The "two-phone look ahead technique" proceeds as follows:
a) from a currently expanded phoneme in the lexical network, for example phoneme (i) 196 of Fig. 8;
b ) look "ahead" (A* search goes backwards through lexical network) two phonemes to see (b) 194 and (a) 192 of Fig. 8 thus 2 5 an estimate is required for (a b);
c ) refarring to the trigram network find a node for a b which is the node 216 of Fig. 9;
d) estimate scores for this node are available for all frame times, thus the estirnates for a specific time may be obtained. (The 3 0 calculation and storage of the estimates is described below.).
To track the best cumulative probability of each state for each time increment, a two dimensional array is constructed, where one axis denotes time and the other enumerates the states. This array, called a Trellis, has dimensions Statestnta~ x Tlme~,~,ax~

Referring the Fig. 10, there is illustrated a trellis for scoring the best cumulative probability of each state. The trellis shows a state, Sm,t, 250, and three previous states Sm, t-1, 252, S~p.i .t-t, 254, and Sm.2,t.i , 256. The three previous states 252, 254, and 256 represent the self, next state, and skip state transitions of the phoneme Markov models.
Fig. 10 also plots phoneme boundaries 258 and 260 for biphones ab and bi respectively. starting at an input state 262, a path 264 for the phoneme b of the biphone ab may diverge resulting in several trellis output states 266 at the phoneme boundary 258. The accumulated transitional probabilities for these trellis output 1 0 states are stored, as described above, for use as estimates for the !~*
search method.
The trellis output states are also used as trellis input states for the next phoneme (i) of biphone bi. For simplicity; only a single input state is shown for the biphone ab.
The Viterbi recognition method is an iterative method for calculating the cumulative probabilities rewired at every point on the Trellis. Each state Sm at a 1 5 given time t has a number of allowable transitions leading to it from certain states at t-1. Typically, Sm,t-t , Sm.t ,t-~ and Sm.~,t-1 can lead to Sm,t. These transitions correspond to a self loop, previous state transition and skip state transition. Each state has a oert~in cumulative probability at t-i and each one of these transitions has a certain probability of occurring given the cepstral vector observed. In order to 2 0 determine the best cumulative probability for S~,,t only the probability of each possible transition needs to be evaluated and summed to the corresponding previous cumulative prababitity, and the one with the best result chosen.
Note that the time is not measured in seconds but rather in terms of frames with one observed cepstral vector per frame. That is, each observed vector is 2 5 an increment on the time axis in the Trellis array. For the 'Jiterbi search of Trigram network, tire Trellis dimensions are of the order of 5000 states by 256 vectors.
The calculation for the full Trellis starts at t=0 and fills in the array one "time increment" at a time (if the transitions chosen are recordod, a description of the path taken would be obtained). The complete ~Jiterbi recognition method for the 3 0 Trellis, then, is as follows:
1 ) Set all log probabilities in the Trellis to -~, except the state defined as the beginning of the utterance. Set this one to 0.
2 } For tune from 0 to the end of utterance For ail states in all possible nodes of the trigram network.

a) get the cumulative probability at i-1 ofi all possible previous states leading to Sm,q and the associated transition probabilities, b ) calculate all possible cumulative probabilities fior Sm,t, c ) choose the one resulting in the best snore for Sm,t, and store it in Trellis.
In an embodbment of the present invention, the FVP~ system uses cepstrum vectors, with a dimension of 't5 as input. The A* search method itself, however, is not dependent on the type ofi data input; in fiact, it rraay use any type ofi 1 0 speech vector as input. Mathematioalky, the cepstrai veotors are computed in a known manner, by dividing the speech spectrum between a and 4 kHz into 2r3 channels spaced according to the mel sale of frequencies. The energy in each bhannel is computer by gumming a triangularly weighted spectrum located at the center of the channel. The cepstral coefficients are computed by taking the cosine transform in 1 5 accordance with the following equatibn:

- ~ ~j ~s (i (j -0.5) 2 Q I~ i - 1, 2, .., 7 j-i where Ej ~ log ohannet energy in the jth filter. The first two and last two cepstral vectors are multiplied by 0.01. Further details concerning cepstral vectors may be gleaned from United States Patent No. 4,955,865, issued 11 September 1990 to M.
2 0 Lennig et al., and assignod to Northern Telecom Limited the entire disclosure of which is incorporated by reference. Conveniently, one vector is generated every 12.75 ms.
This fieature veotor efi dimension 15 is the speech input which drives the A*
search method.
As previously stated, the A* search method requires estimates of the 2 5 probability from each node to the beginning of the utterance fior all time. It does not require, however, estimates for internal state transitions within each node.
Therefore, although the trigram search must calculate the probabilities for all Trellis eiernents, it need only store (fior the A*) the exit slates 265 ofi each node fior all time (i.e. a reduced Trellosj.

Referring to Fig: 11, there is illustrated a functional block diagram of the flexible vocabulary recognition (FVR} system. An input 300 accepts cepstrum data from known speech processors, for example function blocks 10 through 20 of Fig. 1 in the above-incorporated United States Patent PJo. 4,956,865.
The cepstrum data accepted is stored in an input data buffer, 302. The input data buffer 302 is connected to other system components by a bus 304.
These system components connected to the bus 304, include processing elements (CPU) 306, a recognition data tables store 308, an intermediate result store 310.
and a result output device 312. The result output device 312 includes an output 314 to an 1 0 application (not shown in Fig. 10} coupled to the FVR system.
The operation of the FVR system is illustrated as flow charts in Fig. 12a and Fig. 12b. The sequence of frame synchronous steps executed by the FVR
system is illustrated in Fig. 12a. The frame synchronous operation begins with blocks 400 and 402. Following the production of cepstrum vectors from tPie input speech in a known 1 5 manner, the presence of the cepstrum vector is determined by a decision block 404.
On a frame-by-frame baSiS, a black 406 updates Trellis input states for the Viterbi search through the trigram network.
This step is followed by a block 408 representing the calculation of transitional probabilities for all phoneme model transitions given the input cepstrum 2 0 vector frame. The resultant probabilities are then stored in the intermediate result store 310 of Fig. 10. A step of updating the trellis is represented by a block 410.
This step is followed by saving trellis output states from this frame for all phoneme boundaries as represented by a block 412. The trellis output states are used, far the next frame, to update trellis input states at the block 406.
2 5 Once all the input data has been processed as determined by the decision block 404, a block 414 computes all point scores at a block 416, for each phoneme model given the input cepstrum vector. (Point scores are described in detail in conjunction with Fig. 13.) The corr~putation continues until completed for ail input data as determined by a decision block 4i6. The frame synchronous operation then 3 0 initiates the A* search method 418 as illustrated in Fig. 12b.
For each cepstral vector received, the following are the steps to the trigram search:
1 } CogtTrap itional Prg~abilitie (408}
Calculate the probabilities associated with each transition, 3 5 given the input vector. This consists of:

a) multiplying the input vector by the model's (rotated) covariance matrix, b ) calculating the magnitude (norm) of the result, o ) for each transition within the model, determining which mode (or mixture) results in the best transition probability. This calculation uses the results of (a) and (b) and each mode's mean and weight.
2 ) ?~~ll~ ~t?date (4'10) 1 0 This step determines the best transitions and cumulative probability within each node. For each state in the node:
a) get the cumulative probability at t-1 of all possible previous states leading to Sm,t and the associated transition probabilities (calculated in step 1), 1 5 b ) calculate all possible cumulative probabilities for Sm,t, c ) chaose the one resulting in the best score for Sm,t, and store it.
3 ) ~.~v~, Trellis ~utout States (412) 2 0 Save the scores for the exit state of all nodes. These are the estimators required for the A" search.
4 ) trellis ~~ut/State~. (~b06) This step propagates exit state snores from all nodes to the 2 5 entrance states of attached nodes if they are better than the score currently stored in that state. Fir each node:
a) determine alt branch nodes to which this node attaches;
b ) If the cumulative probability of the last state in this node, Piast- state. is greater than that of the first state in 3 0 the attached node, Pt;rst-state replace first-state with P last-state ~9ote that this is reaNy part of the full Trellis update, but the task has been divided into two separatt? steps (2 and 3).

When the utterance has bean completed and the estimators have been calculated, the A* portion of the search may be applied. For the A* search portion some commonly used data are point scores for the phonemes, estimates of the partial paths and a window of best scores is to be considered .
The A* search 418 is initialized to recognize an unknown utterance by a block 420, sefiting up a stack (or heap) to keep track of the path scores arranged in descending order of Total score. A block 422 accesses the stack to retrieve the top entry. A block 424 determines if the search has ended. A yes response leads to a block 428 returns the top stack entry as top choice for recognition. This ends the A*
1 0 search 428. If the search is not complete, a block 430 determines the five exit scores from the top enter and using the points scores for the phonerrre being expanded, determines the best scores for a range of entry times for the phoneme. Using the best scores together with estimates for the partial paths obtained from the trigram network using the two phase look ahead technique, a block 432 computes a new stack 1 5 entry. A block 434 determines if there are other one phoneme extensions in the lexical network beyond the currently expanded phoneme. A yes results in computation of new stack entries for each remaining branch. Otherwise the method is repeated beginning at a block 422. The steps represented by blocks 422, 430, and 434 are explained in detail in conjunction with Fig. l4.
2 0 An array of point scores is constructed for each phoneme which represent the probability of traversing that phoneme between any two times, T1 and T2. The array is represented as a matrix where the vertical axis is the exit time and the horizontal axis is the entrance time, as shown in Fig. 13. These probabilities are used by the A* search method in determining the exact probability (or cost) of 2 5 traversing a given branch In the lexical network botween two given tames.
Point scores depend upon the sequence of observation vectors. This may be expressed in the equation form as follows:
P ~ OTz I M ) T1 <TZ , where 01 is the utterance and M is the phoneme model As described above in connection with Figs. 8 and 9, while traversing a path in the A* search, estimates for the remainder of that path to the beginning of the utterance must be continually obtained. it is useful to think of the retrieval of these estimates being provided by an estimator function (although the implementation need not treat it as such), This function can be considered to take two major parameter sets:
1 ) ~.j~-I xt Two"~~r~nemes in thePath The next two phonemes in the path of the lexical network determine which trigram node is to be used to obtain the estimate.
Thus, the next rivo phonemes define the ' mapping between the Pexicai network and the trigram network. This process of using the 't 0 next two phonemes in the path to obtain the estimate is termed "two phone look-ahead"
(Note; though, that the ~* search is actually going backwards through the network, so it is really "look-behind").
~S
2 ) Th~~T~L~,.~c~t'ne Numberl Once the trigram node to be used is established, it is still necessary to define the timd far which the estimate is required. This 2 0 ~i~e is defined ire terms of the frame number.
Hour the time is obtained is described below in connection with Fig. ~ 4.
For simplicity, the discussion thus far has boen based on using a single point in time for determination of the estimate and truo probability. In practice, 2 5 however, there may be uncertainty that a phoneme boundary is placed optimally by the first positioning. Therefore, the A* search expands each phoneme from 5 exit times to allow for movement of the phoneme boundary around this window.
Conveniently, far the present embodiment, the window has been set to the bast frames.
~ 0 For the ~* search, the probability of a given path through the lexical network is constructed of two components; the estimate for the remainder of the path, as provided by the two phone look ahead technique, on the trigram network and the exact probability calculated up to that point in the path. The object then is to continue to calculate the exact probability far the path which has the best total probability 9 5 until the beginning of the- utterance is reached.

Referring to Fig. 14, there is graphically illustrated a single phoneme expansion of the A* search method.
The current A* search stack entry consists of n probability scores , where n = 5 for Fig. 14, (exit scores) together with the corresponding frame times, one phone sequence which includes the expanded phonemes and the two-phone of the look ahead + best total score. These five probabilities are represented as shaded blocks 500, 502, 504, 506 and 508. Their positions indicate the entry times for the previously expanded phoneme, and hence the exit times for the phoneme to be expanded (Fig. 12b, block 430). The last phoneme from the two-phone look ahead 1 0 technique is expanded in the following manner. foxes 510, 512, 514, 516 and 518 represent the point scores far the last phoneme from this pair. The point scores are given from the earliest to the latest entry times possible for the selected phoneme.
The best score (point score + exit score) for each frame time is determined across the five possible as represented by boxes 520 (this corresponds to block 430 of Fig.
1 5 12b) These scores are added to the estimates for the next biphone pair taken over the range of time frames as represented by boxes 530. The total scores are represented by boxes 540. The shaded boxes 542, 544, 546, 548 and 550 represent the five best total scores. These provide the new stack entry determined by mapping boxes 560 back to 520. The highest total is kept as the best total score. The accumulated 2 0 point scores of the boxes 540 corresponding to the five best scores 542, 544, 546, 548 end 550 are also saved. These are represented as shaded boxes 572, 574, and 578 within boxes 520. For each, the accumulated total point score and the frarne time is recorded. Storing the phoneme sequence completes the new stack entry.
The new stack entry computation indicated by bracket 580, and corresponding to block 2 5 432 of Fig. 12b, is repeated for each branch in the lexical network beyond the currently expanded phoneme represented by boxes 520. Thus, all one phonome extensions at a network branch are expanded.
The steps of the A* search method are as follows:
1 ) 1C11.tiali~atio~, (420, Fib. 12b) 3 0 a ) Set up a stack which will keep track of the selection with the best total probability, where ptotal = Fact + ~ est~ Set all Iog P~S to 0.
b ) Submit all possible biphone (two phone) word endings to the estimator function and record all 1'e$t. Select the 3 5 one with the best probability for expansion, by accessing the top entry in the stack. dote that, for the end of the utterance, there is only one exit time and ptota~ = Pest.
2 } ~~~eme Exl as r~sio_n (430, 432, Fig. 12b) Calculation of the actual probability for a given phone is termed "phoneme expansion". The idea is to calculate the exact likelihood of the first phone in the biphone pair specified to the estimator function (the one oloser to the end of the utterance).
For the first phon~ in the biphcrne pair:... (boxes 500, 502, 504, 1 0 506, 508, Fig. l4) For every exit time in the win~dow....(boxes 510, 512, 514, 516, 518; Fig. 14) for al! possibiaw entrance times....
determine tha entrance time which yields the best total actual -probability (boxes 520, Fig. 14).
Pact(T2) = Pact(T1)-~P(T1->T2) 3 } ~~
Deterrrrine tEie new total probability for the path being expanded 2 0 by:
a) determining all possible '°next biphane pairs" using the two phone look ahead technique on the trigram network and calling up the estimator function for each of these biphones, (step 580, Fig. 14}
2 5 b } computing a new stack entry for each of the biphones.
The five choices for Pact (boxes 570, 572, 574, 576, 578, Fig.
14 corresponding to the five bast total scores (542, 544, 546, 548, 550, Fig. 14) form part of the new stack entry.
3 0 4 } ~St~on repeat steps ~ and 3 until the top entry of the stack has been expanded to reach the beginning of the utterance.

~8 Each entry in the stack has a score made up of estimate scores plus actual score. As several paths are expanded, those consisting of largely estimate scores will eventually appear at the top of the stack because their total score is higher than any other. However, once these are expanded they drop back down in position, only to reach the top on a subsequent reordering. To eliminate unnecessary expansions, the following technique is used:
a) Keep track of the least number of frames used in estimate scores for any entry in the stack q for an utterance length s frames.
b ) before expanding the top entry, shack its number ~f estimate 1 0 scores r. If r is greater than the greater of (q + 75 or q + sf2) discard the entry from the stack.
For example:
For an utterance whose length s is 160 frames if the entry with the least estimate scores has q = 10 frames 1 5 then ail stack entries with i 0 + .i~.Q = 10 + 80 = 90 estimate scores or more will be discarded (as they appear at the top of the stack).
An embodiment of the present invention includes other improvements for 2 0 racocJni~ing unknown utterances. As described above, an improved heuristic function for the A* search recognition method results in significant reduction in computing.
The recognition method is sensitive to the parameter vectors being used.
For simplicity, only cepstral vectors have been discussed. However, two different parameter vectors may be used for speech recognition: cepstral parameters (for 2 5 example in United States Patent No. 4,956,865 incorporated by reforonce hereinabove) and line spectral pair (LSP) parameters (as described by Soong, F.K.
and Juang, B.-N. (1984), "Line Spectrum Pair (LSP) and Speech Data Compression", Proceedings of the 1984 IEEE Bnternational Conference on Acoustics, Speech and Signal Processing, pp. 1.10.1-1.10.4). Even though the recognition 3 0 accuracy using either the cepstrai parameters or the LSP parameters is about the same, the individual recognition results may be different. Only half the errors are common to both the parameter sets. 'this happens because the two parameters represent the speech spectrum differently: cepstral parameters represent the smoothed spectrurn accurately, while the LSP parameters represent the spectral 3 5 pocks accurately. in order to benefit from both the representations, an embodiment of the present invention uses the two parameters separately for recognition and takes a joint decision. V~Ihen the recognized words using the two parameters are different, a joint likelihood is determined for each of these words using the two parameter sets and the one giving the highest likelihood is chosen as the recognized word.
Combining the results of the two parameter sets in this manner results in significant increase in recognition accuracy.
Referring to Fig. 15, there is illustrated, as a flow chart, the method of recognizing using both cepstral and LSP parameters, given separate recognition results for each type of parameter. Blocks 700 and 702 represent a recognition result using cepstral parameters and LSP parameters, respectively. The cepstral 1 0 parameter recognition 700 results in an output word 1 having a likelihood L1 (word 1). Tha LSP parameter recognition 702 results in an output word 2 having a likelihood L2 (word 2). A decision block 704 determines if the results are the same, i.e., word 1 = word 2. A yes means recognition is complete 708.
tAlhen the block 70~ detarmines that the words are not the same, i.e.
1 5 word i~ word 2, additional likelihoods are computed by a block 710. The likelihood of word 1 is determined usirrg LSP parameters, i.e. L2 (word 1). Also the likelihood of word 2 is determined using LSP parameters, i.e. Lt (word 2). Then a joint likelihood is taken by a block 712 by determining if:
L1 (word 1) x L2(word 1) > L2(word 2) x Lfi (word 2) 2 0 A yes results in an output of word 1 by a block 714, while a no results in an output of word 2 by a block 716.
A capability for rejection of out-of-vocabulary utterances is important in many applications. In accordance with an embodiment of the present invention a rejection method is provided far applications requiring medium size vocabulary (up 2 5 to several thousand words). The recognition results from two different parameter sets are used to provide effective rejection of oui of vocabulary words. The two ' parameter sets are cepstral pararneters and LSP parameters. Referring to I'ig. 18 there is illustrated as a flow chart, the steps of the rejection method. The rejection method works as follows. The cepstral parameters recognition 800 output word 1 and 3 0 the LSP parameters recognition 802 outlaut word 2. If the two parameters give different results, as determined by decision block 80.x, then the utterance is rejected as not belonging to th~ lexicon as represented by block 806. i~Vhen the two results are the same, the utterance is rejected based on a likelihood threshold. Let the log likelihood per frame using cepstraf parameters be Li , the log likelihood per frame using LSP parameters be L2, and the recognized phoneme sequence be p phonemes long. Then the recognized utterance is rejected if a block 808 determines p < 7, and (L1+L2) < T1, or a block 810 determines 6 < p < 15, and (Li + L2) < T2, or a block 812 determines p > 14, and (L1+L2) < T3.
The thresholds T1, T2, and T3 are fixed thresholds. Typical values for T1, T2, and T3 are -3, -9, and -10, respectively. The use of likelihood thresholds for rejection with only one parameter set, yields poor rejection results. The combination of two parameters (LSP and cepstrum) for rejection, and the likelihood 1 0 thresholds for rejection makes the rejection method work effectively.
It is well known to use hidden Markov models for digits (e.g., digit models) in digit-string recognition systems. An example of such a system is taught by Stephen P. Levinson et al., in U.S. Patent No. x,587,670 issued May 6, 1986 and reissued May 28, 1991 as Re.33,597. Unfortunately, digit-models will not work 1 5 with a phoneme based recognition system. Also the use of phoneme models for digits instead of digit-models may result in an increase in the digit-string recognition error rate.
In accordance with an embodiment of the present invention the phoneme base is expanded to include 33 allophone models used for digit recognition.
2 0 The use of context-dependent allophones does improve recognition accuracy: These allophones depend both on the left and the right contexts. A
tote! of 33 allophone models and three silence models for digit-string recognition are used (see Table 2). A important aspect in digit-string recognition is the smoothing of the covariance matrix for silence models. The smoothing is performed by replacing the 2 5 covariance matrix for silence models by the covariance matrix for phoneme /f/ as explained hereinabove with respect to covariance matrix substitution.
Similarly allbphone models may be used for specific vocabulary words provided sufficient training data exists for the words. Specific words, such as yes and no, may be represented by allophone models to improve recognition accuracy.

phoneme allophones phoneme al lo phones /Z/ ZI /I/ _ _ I ZII

/r/ Iro 6ri or /o/ ro for o /w/ wn n wnn /n/ ~n an na /t/ to et j a jn /u/ to /A/ 8r /1/ ri /f/ fo fa j /a j/ fa jv na /v/ a jv ova jn /s/ sI ks s~ /k/ Iks // sev /a/ van /e/ et silence inter-digit,erance-initial,utterance-final utt Table 2 Allophone models for recognizing digit-strings.

Claims (31)

WHAT IS CLAIMED IS:
1. A method of preparing phoneme models for recognition of speech received via telephone lines comprising the steps of:
a) analyzing a training word to generate a frame sequence of acoustic parameter vectors representative thereof;
b) providing a first set of model parameters representative of a sequence of state-transition models corresponding to the training word, the parameters including a mean vector and weighting factor for each transition and a covariance matrix for each model;
c) computing a set of observation probabilities for the training word and the first set of model parameters;
d) aligning the frame sequence of acoustic parameter vectors with the sequence of state-transition models to provide a mapping therebetween representative of a path through the sequence of state-transition models having a highest likelihood;
e) accumulating statistics for a plurality of utterances of said training word using the mapping of step d);
f) generating a second set of model parameters representative of the sequence of state-transition models corresponding to the training word;
g) repeating step d) for the second set of model parameters;
h) comparing the likelihood of the first and second sets of model parameters; and i) repeating step b) through h), replacing the first set of model parameters by the second set of model parameters when the second set of parameters provides at least a predetermined improvement in likelihood; and j) substituting the covariance matrix of a first model with the covariance matrix of a second model to provide a smooth covariance matrix thereby improving recognition accuracy for the first model.
2. A method as claimed in claim 1 wherein the first model represents a left silence phoneme ({) and the second model represents a phoneme (f).
3. A method as claimed in claim 1 wherein the first model represents a right silence phoneme (}) and the second model represents a phoneme (f).
4. A method as claimed in claim 1 wherein the first model represents a phoneme selected from the group consisting of .delta. , ~ , ~, and the second model represents a phoneme (~).
5. A method as claimed in claim 1 wherein the first model represents a phoneme (.epsilon. before r) and the second model represents a phoneme (~).
6. A method of preparing phoneme models for recognition of speech received via telephone lines comprising the steps of:
a) analyzing a training word to generate a frame sequence of acoustic parameter vectors representative thereof, having a parameter indicative of energy level for each frame of the frame sequence;
b) providing a first set of model parameters representative of a sequence of state-transition models corresponding to the training word;
c) computing a set of observation probabilities by, for each frame of the frame sequence and each model of the sequence of state-transition models, comparing the energy level of the frame with a predetermined, relative to noise on the telephone lines, energy threshold for the model, and if the energy level is below the energy threshold, setting the observation probability for the frame to zero, otherwise computing the observation probability for the frame;

d) aligning the frame sequence of acoustic parameter vectors with the sequence of state-transition models to provide a mapping therebetween representative of a path through the sequence of state-transition models having a highest likelihood;
e) accumulating statistics for a plurality of utterances of said training word using the mapping of step d);
f) generating a second set of model parameters representative of the sequence of state-transition models corresponding to the training word;
g) repeating step d) for the second set of model parameters;

h) comparing the likelihood of the first and second sets of model parameters: and i) repeating step b) through h), replacing the first set of model parameters by the second set of model parameters when the second set of parameters provides at least a predetermined improvement in likelihood.
7. A method as claimed in claim 6 wherein the model represents a vowel other than (~) and if the model is one of four first models of the sequence of state-transition models, the energy threshold is 10dB above background noise;
otherwise, if the sequence of state-transition models is greater than 10 models and the model is one of four last models of the sequence, the energy threshold is 1dB above background noise;
otherwise, the energy is 6dB above background noise.
8. A method as claimed in claim 6 wherein the model represents a vowel (~) and if the model is one of four first models of the sequence of state-transition models, the energy threshold is 6dB above background noise; otherwise, if the sequence of state-transition models is greater than 10 models and the model is one of four last models of the sequence, the energy threshold is 1dB above background noise; otherwise, the energy is 3dB above background noise.
9. A method as claimed in claim 6 wherein the model represents a phoneme selected from the group consisting of 1, r, j, w, ~, and ~. and if the sequence of state-transition models is greater than 10 models and the model is one of four last models of the sequence, the energy threshold is 1dB above background noise;
otherwise, the energy is 3dB above background noise.
10. A method as claimed in claim a wherein the model represents a phoneme selected from the group consisting of f, v, .theta., .delta., and h and the model is one of four first models, the energy threshold is 1dB above background noise.
11. A method as claimed in claim 6 wherein the model represents a phoneme selected from the group consisting of s, z, n, m, and ~, the energy threshold is 1 dB above background noise.
12. A method of preparing phoneme models for recognition of speech received via telephone lines comprising the steps of:
a) analyzing a training word to generate a frame sequence of acoustic parameter vectors representative thereof and to label each frame of the frame sequence as voiced, unvoiced or silence;
b) providing a first set of model parameters representative of a sequence of state-transition models corresponding to the training word, including the step of labelling each model as voiced, unvoiced or silence in dependence upon a phoneme represented by the model and a relative position of the model in the sequence;
c) computing a set of observation probabilities by, for each frame of the frame sequence and each model of the model sequence of state-transition models, comparing a voiced-unvoiced-silence (VUS) label of the frame with a VUS label of the model and if the labels do not match, setting the observation probability for the frame to zero, otherwise computing the observation probability for the frame for the training word and the first set of model parameters;
d) aligning the frame sequence of acoustic parameter vectors with the sequence of state-transition models to provide a mapping therebetween representative of a path through the sequence of state-transition models having a highest likelihood;
e) accumulating statistics for a plurality of utterances of said training word using the mapping of step d);
f) generating a second set of model parameters representative of the sequence of state-transition models corresponding to the training word;
g) repeating step d) for the second set of model parameters;
h) comparing the likelihood of the first and second sets of model parameters; and i) repeating step b) through h), replacing the first set of model parameters by the second set of model parameters when the second set of parameters provides at least a predetermined improvement in likelihood.
13. A method as claimed in claim 12 wherein the phoneme represented is intervocalic silence and wherein the step of labelling maps both unvoiced and silence labels to the model.
14. A method as claimed in claim 12 wherein the phoneme represented is (j) and wherein the step of labelling maps both unvoiced and voiced labels to the model.
15. A method as claimed in claim 12 wherein the phoneme represented is selected from the group consisting of (n, m, and ~) and wherein if the sequence is greater than 10 models and the model is one of four last models, then the step of labelling maps voiced, unvoiced, and silence labels to the phoneme, otherwise the step of labelling maps both voiced and unvoiced labels to the model.
16. A method as claimed in claim 12 wherein the phoneme represented is selected from the group consisting of (vowel, 1, r, and w) and wherein if the model is in an initial position or is one of four last models in a sequence of greater than 10 models then the step of labelling maps both voiced and unvoiced labels to the model, otherwise maps a voiced label to the model.
17. A method as claimed in claim 16 wherein the model represents the phoneme (r) when followed by a vowel and preceded by (t) or (f) and the step of labelling maps both voiced and unvoiced labels to the model.
1 8. A method as claimed in claim 16 wherein the model represents the phoneme (w) when preceded by (k) and the step of labelling maps both voiced and unvoiced labels to the model.
19. A method as claimed in claim 16 wherein the model represents the phoneme (i) when followed by inter-word silence and the step of labelling maps both voiced and unvoiced labels to the model.
20. A method as claimed in claim 16 wherein the model represents the phoneme (I) when preceded by (d) or inter-word silence and the step of labelling maps both voiced and unvoiced labels to the model.
21. A method as claimed in claim 16 wherein the model represents the phoneme (.epsilon.) when preceded by inter-word silence and the step of labelling maps both voiced and unvoiced labels to the model.
22. A method as claimed in claim 16 wherein the model represents the phoneme (u) when preceded by (j) and the step of labelling maps both voiced and unvoiced labels to the model.
23. A method as claimed in claim 16 wherein the model represents the phoneme (~) when preceded by (s), (~), (r), (d), or inter-word silence and the step of labelling maps both voiced and unvoiced labels to the model.
24. A method of speech recognition for speech received via telephone lines comprising the steps of:
a) analyzing an unknown utterance to generate a frame sequence of acoustic parameter vectors representative thereof;
b) providing a first network representing a recognition vocabulary, wherein each branch of the first network is a model representing a phoneme and each complete path through the first network is a sequence of models representing a word in the recognition vocabulary;
c) providing a second network derived from the first network, in which all sequences of three consecutive phonemes present in the first network are present;
d) computing transitional probabilities for each node of the second network given the frame sequence of acoustic parameter vectors;
e) searching the second network to determine optimal cumulative probabilities for each node of the second network for all frames of the frame sequence;
f) storing the cumulative probabilities as estimate scores for estimating partial paths in the first network;

g) computing point scores for all phonemes in dependence upon the frame sequence of acoustic parameter vectors; and h) determining a complete path through the first network by evaluating successive one phoneme extensions of partial paths using the estimate scores for the nodes of the second network to find the partial path to extend.
25. The method of claim 24 wherein the step of determining includes the step of initiating a search through the first network by:
establishing a stack for recording all paths from the end of the first network;
looking along all branches of the network two phonemes;
obtaining estimate scores for each node of the second network corresponding to each two phoneme branch of the first network; and entering each estimate score into the stack that arranges the scores in descending order.
26. The method of claim 25 wherein the step of determining includes the steps of expanding a top entry in the stack by:
obtaining point scores for the first phoneme of the two phoneme branch closest to the end of the first network;
for every exit time in the stack entry and for all possible entrance times for the phoneme determine total actual probability by adding exit scores from the stack entry to point scores for the first phoneme; and computing a new stack entry by adding estimate scores (P est) for a next two phoneme node of the second network to the total actual probability (P act) for all possible entrance times, selecting n best total probabilities (P act + P est), where n is an integer, and storing the total actual probabilities P act and frame times for each, together with the best total probability (P act + P est and a phoneme sequence as the new stack entry.
27. A method as claimed in claim 26 wherein the step of expanding a top entry includes the steps of:

a) storing a least number, q, of frames used in the estimate score for the top entry of the stack;
b) prior to expanding any top entry of the stack, determining the number of frames, r, used in its estimate score; and c) discarding, from the stack, the top entry when r is greater than the greater of (q + 75 and q + s/2) where s is the length of the unknown utterance in frames.
28. A method as claimed in claim 24 wherein the steps are completed using both cepstral parameters and LSP parameters, the step of providing the word in the recognition vocabulary providing word 1 with a cumulative probability L1 (word 1) for the cepstral parameter and word 2 with a cumulative probability L2(word 2) for the LSP parameters and wherein a joint recognition includes the further steps of:
if word 1 and word 2 are the same, providing word 1 as the speech recognition output, otherwise;
determining a cumulative probability for word 1 using the LSP parameters (L2(word 1)) and a cumulative probability for word 2 using the cepstral parameters (L1 (word 2));
if L1(word 1) x L2(word 1) is greater than L1 (word 2) x L2(word 2), providing word 1 as the speech recognition output, otherwise;
providing word 2 as the speech recognition output.

29. A method as claimed in claim 24 wherein the steps are completed using both cepstral parameters and LSP parameters, the step of providing the word in the recognition vocabulary providing word 1 with a cumulative probability L1(word 1) for the cepstral parameter and word 2 with a cumulative probability L2(word 2) for the LSP parameters and wherein a joint recognition includes the further steps of:
if word 1 and word 2 are the same;
if the word length is less than 7 phonemes and (Li(word 1) + L2(word 1)) is less than a first threshold T1, reject the unknown utterance as out-of-vocabulary;

otherwise, if the word length is between 6 phonemes and 15 phonemes and (L1(word 1) + L2(word 1)) is less than a second threshold T2, reject the unknown utterances as out-of-vocabulary;
otherwise, if the word length is greater than 14 phonemes and (L1(word 1) + L2(word 1)) is less than a third threshold T3, reject the unknown utterance as out-of-vocabulary;
otherwise, if word 1 and word 2 are different, reject the unknown utterance as out-of-vocabulary.
29. A method as claimed in claim 1 wherein the state-transition models include allophone models for digit-string recognition and wherein the first model represents a silence model and the second model represents a phoneme (f).
30. A method as claimed in claim 1 wherein the state transition models include allophone models for a word that is both a member of a recognition vocabulary and a training word set.
31. Apparatus for speech recognition, comprising:
a) means for analyzing an unknown utterance to generate a frame sequence of acoustic parameter vectors representative thereof;
b) means for providing a first network representing a recognition vocabulary, wherein each branch of the first network is a model representing a phoneme and each complete path through the first network is a sequence of models representing a word in the recognition vocabulary;
c) means for providing a second network derived from the first network, in which all sequences of three consecutive phonemes present in the first network are present;
d) means for computing transitional probabilities for each node of the second network given the frame sequence of acoustic parameter vectors;
e) means for searching the second network to determine optimal cumulative probabilities for each node of the second network for all frames of the frame sequence;

f) means for storing the cumulative probabilities as estimate scores for estimating partial paths in the first network;
g) means for computing point scores for all phonemes in the second network in dependence upon the frame sequence of acoustic parameter vectors;
and h) means for determining a complete path through the first network by evaluating successive one phoneme extensions of partial paths using the estimate scores for the nodes of the second network to find the partial path to extend.
CA002069675A 1991-10-08 1992-05-27 Flexible vocabulary recognition Expired - Lifetime CA2069675C (en)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
US772,903 1991-10-08
US07/772,903 US5390278A (en) 1991-10-08 1991-10-08 Phoneme based speech recognition

Publications (2)

Publication Number Publication Date
CA2069675A1 CA2069675A1 (en) 1993-04-09
CA2069675C true CA2069675C (en) 2000-06-27

Family

ID=25096581

Family Applications (1)

Application Number Title Priority Date Filing Date
CA002069675A Expired - Lifetime CA2069675C (en) 1991-10-08 1992-05-27 Flexible vocabulary recognition

Country Status (2)

Country Link
US (1) US5390278A (en)
CA (1) CA2069675C (en)

Families Citing this family (120)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
DE4131387A1 (en) * 1991-09-20 1993-03-25 Siemens Ag METHOD FOR RECOGNIZING PATTERNS IN TIME VARIANTS OF MEASURING SIGNALS
US5297183A (en) * 1992-04-13 1994-03-22 Vcs Industries, Inc. Speech recognition system for electronic switches in a cellular telephone or personal communication network
JP2818362B2 (en) * 1992-09-21 1998-10-30 インターナショナル・ビジネス・マシーンズ・コーポレイション System and method for context switching of speech recognition device
AU5803394A (en) * 1992-12-17 1994-07-04 Bell Atlantic Network Services, Inc. Mechanized directory assistance
ES2141824T3 (en) * 1993-03-25 2000-04-01 British Telecomm VOICE RECOGNITION WITH PAUSE DETECTION.
CA2157496C (en) * 1993-03-31 2000-08-15 Samuel Gavin Smyth Connected speech recognition
US6230128B1 (en) 1993-03-31 2001-05-08 British Telecommunications Public Limited Company Path link passing speech recognition with vocabulary node being capable of simultaneously processing plural path links
US5515475A (en) * 1993-06-24 1996-05-07 Northern Telecom Limited Speech recognition method using a two-pass search
US5528701A (en) * 1994-09-02 1996-06-18 Panasonic Technologies, Inc. Trie based method for indexing handwritten databases
GB2290684A (en) * 1994-06-22 1996-01-03 Ibm Speech synthesis using hidden Markov model to determine speech unit durations
NL9401036A (en) * 1994-06-23 1996-02-01 Tno Anaerobic removal of sulfur compounds from wastewater.
US5794198A (en) * 1994-10-28 1998-08-11 Nippon Telegraph And Telephone Corporation Pattern recognition method
US5729656A (en) * 1994-11-30 1998-03-17 International Business Machines Corporation Reduction of search space in speech recognition using phone boundaries and phone ranking
JP3008799B2 (en) * 1995-01-26 2000-02-14 日本電気株式会社 Speech adaptation device, word speech recognition device, continuous speech recognition device, and word spotting device
US5675706A (en) * 1995-03-31 1997-10-07 Lucent Technologies Inc. Vocabulary independent discriminative utterance verification for non-keyword rejection in subword based speech recognition
EP0769184B1 (en) * 1995-05-03 2000-04-26 Koninklijke Philips Electronics N.V. Speech recognition methods and apparatus on the basis of the modelling of new words
WO1996037881A2 (en) * 1995-05-26 1996-11-28 Applied Language Technologies Method and apparatus for dynamic adaptation of a large vocabulary speech recognition system and for use of constraints from a database in a large vocabulary speech recognition system
JPH08335091A (en) * 1995-06-09 1996-12-17 Sony Corp Voice recognition device, voice synthesizer, and voice recognizing/synthesizing device
US5684925A (en) * 1995-09-08 1997-11-04 Matsushita Electric Industrial Co., Ltd. Speech representation by feature-based word prototypes comprising phoneme targets having reliable high similarity
US5822728A (en) * 1995-09-08 1998-10-13 Matsushita Electric Industrial Co., Ltd. Multistage word recognizer based on reliably detected phoneme similarity regions
US5825977A (en) * 1995-09-08 1998-10-20 Morin; Philippe R. Word hypothesizer based on reliably detected phoneme similarity regions
JPH0981183A (en) * 1995-09-14 1997-03-28 Pioneer Electron Corp Generating method for voice model and voice recognition device using the method
FR2744277B1 (en) * 1996-01-26 1998-03-06 Sextant Avionique VOICE RECOGNITION METHOD IN NOISE AMBIENCE, AND IMPLEMENTATION DEVICE
US5719921A (en) * 1996-02-29 1998-02-17 Nynex Science & Technology Methods and apparatus for activating telephone services in response to speech
DE19610848A1 (en) * 1996-03-19 1997-09-25 Siemens Ag Computer unit for speech recognition and method for computer-aided mapping of a digitized speech signal onto phonemes
US5905773A (en) * 1996-03-28 1999-05-18 Northern Telecom Limited Apparatus and method for reducing speech recognition vocabulary perplexity and dynamically selecting acoustic models
US5978756A (en) * 1996-03-28 1999-11-02 Intel Corporation Encoding audio signals using precomputed silence
US5937381A (en) * 1996-04-10 1999-08-10 Itt Defense, Inc. System for voice verification of telephone transactions
US5963903A (en) * 1996-06-28 1999-10-05 Microsoft Corporation Method and system for dynamically adjusted training for speech recognition
US6038528A (en) * 1996-07-17 2000-03-14 T-Netix, Inc. Robust speech processing with affine transform replicated data
US5987414A (en) * 1996-10-31 1999-11-16 Nortel Networks Corporation Method and apparatus for selecting a vocabulary sub-set from a speech recognition dictionary for use in real time automated directory assistance
EP0994462B1 (en) * 1996-11-07 2002-04-03 Matsushita Electric Industrial Co., Ltd Excitation vector generation
US5915001A (en) * 1996-11-14 1999-06-22 Vois Corporation System and method for providing and using universally accessible voice and speech data files
US5839107A (en) * 1996-11-29 1998-11-17 Northern Telecom Limited Method and apparatus for automatically generating a speech recognition vocabulary from a white pages listing
US5987408A (en) * 1996-12-16 1999-11-16 Nortel Networks Corporation Automated directory assistance system utilizing a heuristics model for predicting the most likely requested number
US5884259A (en) * 1997-02-12 1999-03-16 International Business Machines Corporation Method and apparatus for a time-synchronous tree-based search strategy
US6041300A (en) * 1997-03-21 2000-03-21 International Business Machines Corporation System and method of using pre-enrolled speech sub-units for efficient speech synthesis
WO1999000758A1 (en) * 1997-06-26 1999-01-07 Charles Schwab & Co., Inc. System and method for automatically providing financial services to a user using speech signals
US6018708A (en) * 1997-08-26 2000-01-25 Nortel Networks Corporation Method and apparatus for performing speech recognition utilizing a supplementary lexicon of frequently used orthographies
US6122361A (en) * 1997-09-12 2000-09-19 Nortel Networks Corporation Automated directory assistance system utilizing priori advisor for predicting the most likely requested locality
US5995929A (en) * 1997-09-12 1999-11-30 Nortel Networks Corporation Method and apparatus for generating an a priori advisor for a speech recognition dictionary
CA2216224A1 (en) * 1997-09-19 1999-03-19 Peter R. Stubley Block algorithm for pattern recognition
JP2001507482A (en) * 1997-10-08 2001-06-05 コーニンクレッカ フィリップス エレクトロニクス エヌ ヴィ Vocabulary and / or language model training
US6006186A (en) * 1997-10-16 1999-12-21 Sony Corporation Method and apparatus for a parameter sharing speech recognition system
US6253173B1 (en) 1997-10-20 2001-06-26 Nortel Networks Corporation Split-vector quantization for speech signal involving out-of-sequence regrouping of sub-vectors
US6016470A (en) * 1997-11-12 2000-01-18 Gte Internetworking Incorporated Rejection grammar using selected phonemes for speech recognition system
US6246988B1 (en) 1998-02-10 2001-06-12 Dsc Telecom L.P. Method and apparatus for accessing a data base via speaker/voice verification
US6151572A (en) * 1998-04-27 2000-11-21 Motorola, Inc. Automatic and attendant speech to text conversion in a selective call radio system and method
US6229880B1 (en) 1998-05-21 2001-05-08 Bell Atlantic Network Services, Inc. Methods and apparatus for efficiently providing a communication system with speech recognition capabilities
US6233315B1 (en) 1998-05-21 2001-05-15 Bell Atlantic Network Services, Inc. Methods and apparatus for increasing the utility and interoperability of peripheral devices in communications systems
US6775652B1 (en) 1998-06-30 2004-08-10 At&T Corp. Speech recognition over lossy transmission systems
US6173258B1 (en) * 1998-09-09 2001-01-09 Sony Corporation Method for reducing noise distortions in a speech recognition system
US6768979B1 (en) 1998-10-22 2004-07-27 Sony Corporation Apparatus and method for noise attenuation in a speech recognition system
US6744860B1 (en) 1998-12-31 2004-06-01 Bell Atlantic Network Services Methods and apparatus for initiating a voice-dialing operation
US6574596B2 (en) * 1999-02-08 2003-06-03 Qualcomm Incorporated Voice recognition rejection scheme
EP1159688A2 (en) * 1999-03-05 2001-12-05 Canon Kabushiki Kaisha Database annotation and retrieval
US7139708B1 (en) 1999-03-24 2006-11-21 Sony Corporation System and method for speech recognition using an enhanced phone set
US6487531B1 (en) 1999-07-06 2002-11-26 Carol A. Tosaya Signal injection coupling into the human vocal tract for robust audible and inaudible voice recognition
WO2001004875A1 (en) * 1999-07-08 2001-01-18 Koninklijke Philips Electronics N.V. Speech recognition device with transfer means
JP3632529B2 (en) * 1999-10-26 2005-03-23 日本電気株式会社 Voice recognition apparatus and method, and recording medium
US7310600B1 (en) 1999-10-28 2007-12-18 Canon Kabushiki Kaisha Language recognition using a similarity measure
EP1228452B1 (en) * 1999-10-28 2007-09-19 Canon Kabushiki Kaisha Pattern matching method and apparatus
US6882970B1 (en) 1999-10-28 2005-04-19 Canon Kabushiki Kaisha Language recognition using sequence frequency
KR100434538B1 (en) * 1999-11-17 2004-06-05 삼성전자주식회사 Detection apparatus and method for transitional region of speech and speech synthesis method for transitional region
JP3814459B2 (en) * 2000-03-31 2006-08-30 キヤノン株式会社 Speech recognition method and apparatus, and storage medium
US6594393B1 (en) * 2000-05-12 2003-07-15 Thomas P. Minka Dynamic programming operation with skip mode for text line image decoding
GB0011798D0 (en) * 2000-05-16 2000-07-05 Canon Kk Database annotation and retrieval
GB0015233D0 (en) 2000-06-21 2000-08-16 Canon Kk Indexing method and apparatus
DE10038517A1 (en) * 2000-08-08 2002-02-21 Philips Corp Intellectual Pty Automatic recognition of company names in linguistic statements
US7292678B2 (en) * 2000-08-31 2007-11-06 Lamson Holdings Llc Voice activated, voice responsive product locator system, including product location method utilizing product bar code and aisle-situated, aisle-identifying bar code
US6813341B1 (en) * 2000-08-31 2004-11-02 Ivoice, Inc. Voice activated/voice responsive item locator
GB0023930D0 (en) 2000-09-29 2000-11-15 Canon Kk Database annotation and retrieval
US7003455B1 (en) * 2000-10-16 2006-02-21 Microsoft Corporation Method of noise reduction using correction and scaling vectors with partitioning of the acoustic space in the domain of noisy speech
GB0027178D0 (en) 2000-11-07 2000-12-27 Canon Kk Speech processing system
GB0028277D0 (en) * 2000-11-20 2001-01-03 Canon Kk Speech processing system
GB2370401A (en) * 2000-12-19 2002-06-26 Nokia Mobile Phones Ltd Speech recognition
EP1407447A1 (en) * 2001-07-06 2004-04-14 Koninklijke Philips Electronics N.V. Fast search in speech recognition
US20020010715A1 (en) * 2001-07-26 2002-01-24 Garry Chinn System and method for browsing using a limited display device
EP1306768A1 (en) * 2001-10-26 2003-05-02 Sensoria Technology Limited Method and system for adaptive learning and pattern recognition
US6990445B2 (en) * 2001-12-17 2006-01-24 Xl8 Systems, Inc. System and method for speech recognition and transcription
US20030115169A1 (en) * 2001-12-17 2003-06-19 Hongzhuan Ye System and method for management of transcribed documents
US7266497B2 (en) * 2002-03-29 2007-09-04 At&T Corp. Automatic segmentation in speech synthesis
US7117148B2 (en) * 2002-04-05 2006-10-03 Microsoft Corporation Method of noise reduction using correction vectors based on dynamic aspects of speech and noise normalization
JP4407305B2 (en) * 2003-02-17 2010-02-03 株式会社ケンウッド Pitch waveform signal dividing device, speech signal compression device, speech synthesis device, pitch waveform signal division method, speech signal compression method, speech synthesis method, recording medium, and program
US7359860B1 (en) 2003-02-27 2008-04-15 Lumen Vox, Llc Call flow object model in a speech recognition system
US7324940B1 (en) 2003-02-28 2008-01-29 Lumen Vox, Llc Speech recognition concept confidence measurement
US20050239022A1 (en) * 2003-05-13 2005-10-27 Harless William G Method and system for master teacher knowledge transfer in a computer environment
US20050239035A1 (en) * 2003-05-13 2005-10-27 Harless William G Method and system for master teacher testing in a computer environment
US7797146B2 (en) * 2003-05-13 2010-09-14 Interactive Drama, Inc. Method and system for simulated interactive conversation
FR2856217B1 (en) * 2003-06-12 2005-08-26 Nortel Networks Ltd METHOD FOR ANALYZING THE OPERATION OF A PARQUET DATA TRANSMISSION NETWORK INTERFACE
US20050049871A1 (en) * 2003-08-26 2005-03-03 Yifan Gong Speaker-dependent recognition of voice command embedded in arbitrary utterance
JP4150645B2 (en) * 2003-08-27 2008-09-17 株式会社ケンウッド Audio labeling error detection device, audio labeling error detection method and program
KR100554442B1 (en) * 2003-10-06 2006-02-22 주식회사 팬택앤큐리텔 Mobile Communication Terminal with Voice Recognition function, Phoneme Modeling Method and Voice Recognition Method for the same
US7440895B1 (en) 2003-12-01 2008-10-21 Lumenvox, Llc. System and method for tuning and testing in a speech recognition system
US20050187772A1 (en) * 2004-02-25 2005-08-25 Fuji Xerox Co., Ltd. Systems and methods for synthesizing speech using discourse function level prosodic features
US20070033041A1 (en) * 2004-07-12 2007-02-08 Norton Jeffrey W Method of identifying a person based upon voice analysis
US20060069562A1 (en) * 2004-09-10 2006-03-30 Adams Marilyn J Word categories
DE602005012596D1 (en) * 2004-10-19 2009-03-19 France Telecom LANGUAGE RECOGNITION METHOD WITH TEMPORARY MARKET INSERTION AND CORRESPONDING SYSTEM
US8694317B2 (en) * 2005-02-05 2014-04-08 Aurix Limited Methods and apparatus relating to searching of spoken audio data
US8438027B2 (en) * 2005-05-27 2013-05-07 Panasonic Corporation Updating standard patterns of words in a voice recognition dictionary
JP2009527024A (en) * 2006-02-14 2009-07-23 インテレクチャル ベンチャーズ ファンド 21 エルエルシー Communication device with speaker-independent speech recognition
US7877255B2 (en) * 2006-03-31 2011-01-25 Voice Signal Technologies, Inc. Speech recognition using channel verification
KR100883652B1 (en) * 2006-08-03 2009-02-18 삼성전자주식회사 Method and apparatus for speech/silence interval identification using dynamic programming, and speech recognition system thereof
JP5386692B2 (en) * 2007-08-31 2014-01-15 独立行政法人情報通信研究機構 Interactive learning device
US8543393B2 (en) * 2008-05-20 2013-09-24 Calabrio, Inc. Systems and methods of improving automated speech recognition accuracy using statistical analysis of search terms
US9484019B2 (en) * 2008-11-19 2016-11-01 At&T Intellectual Property I, L.P. System and method for discriminative pronunciation modeling for voice search
JP2010282199A (en) * 2009-06-02 2010-12-16 Honda Motor Co Ltd Lexical acquisition apparatus, multi dialogue behavior system, and lexical acquisition program
US8548807B2 (en) * 2009-06-09 2013-10-01 At&T Intellectual Property I, L.P. System and method for adapting automatic speech recognition pronunciation by acoustic model restructuring
US8903847B2 (en) * 2010-03-05 2014-12-02 International Business Machines Corporation Digital media voice tags in social networks
US8688090B2 (en) 2011-03-21 2014-04-01 International Business Machines Corporation Data session preferences
US20120246238A1 (en) 2011-03-21 2012-09-27 International Business Machines Corporation Asynchronous messaging tags
US20120244842A1 (en) 2011-03-21 2012-09-27 International Business Machines Corporation Data Session Synchronization With Phone Numbers
US8781825B2 (en) * 2011-08-24 2014-07-15 Sensory, Incorporated Reducing false positives in speech recognition systems
US9390708B1 (en) * 2013-05-28 2016-07-12 Amazon Technologies, Inc. Low latency and memory efficient keywork spotting
US10074363B2 (en) * 2015-11-11 2018-09-11 Apptek, Inc. Method and apparatus for keyword speech recognition
CN108305619B (en) * 2017-03-10 2020-08-04 腾讯科技(深圳)有限公司 Voice data set training method and device
US10380997B1 (en) * 2018-07-27 2019-08-13 Deepgram, Inc. Deep learning internal state index-based search and classification
US11783818B2 (en) * 2020-05-06 2023-10-10 Cypress Semiconductor Corporation Two stage user customizable wake word detection
CN112420075B (en) * 2020-10-26 2022-08-19 四川长虹电器股份有限公司 Multitask-based phoneme detection method and device
CN112668310B (en) * 2020-12-17 2023-07-04 杭州国芯科技股份有限公司 Method for outputting phoneme probability by voice deep neural network model

Family Cites Families (10)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US4587670A (en) * 1982-10-15 1986-05-06 At&T Bell Laboratories Hidden Markov model speech recognition arrangement
USRE33597E (en) * 1982-10-15 1991-05-28 Hidden Markov model speech recognition arrangement
US4956865A (en) * 1985-01-30 1990-09-11 Northern Telecom Limited Speech recognition
US4783804A (en) * 1985-03-21 1988-11-08 American Telephone And Telegraph Company, At&T Bell Laboratories Hidden Markov model speech recognition arrangement
US4903305A (en) * 1986-05-12 1990-02-20 Dragon Systems, Inc. Method for representing word models for use in speech recognition
US4803729A (en) * 1987-04-03 1989-02-07 Dragon Systems, Inc. Speech recognition method
US4805219A (en) * 1987-04-03 1989-02-14 Dragon Systems, Inc. Method for speech recognition
US5072452A (en) * 1987-10-30 1991-12-10 International Business Machines Corporation Automatic determination of labels and Markov word models in a speech recognition system
EP0438662A2 (en) * 1990-01-23 1991-07-31 International Business Machines Corporation Apparatus and method of grouping utterances of a phoneme into context-de-pendent categories based on sound-similarity for automatic speech recognition
US5193142A (en) * 1990-11-15 1993-03-09 Matsushita Electric Industrial Co., Ltd. Training module for estimating mixture gaussian densities for speech-unit models in speech recognition systems

Also Published As

Publication number Publication date
CA2069675A1 (en) 1993-04-09
US5390278A (en) 1995-02-14

Similar Documents

Publication Publication Date Title
CA2069675C (en) Flexible vocabulary recognition
US6092045A (en) Method and apparatus for speech recognition
US6195634B1 (en) Selection of decoys for non-vocabulary utterances rejection
US5072452A (en) Automatic determination of labels and Markov word models in a speech recognition system
EP0398574B1 (en) Speech recognition employing key word modeling and non-key word modeling
US5509104A (en) Speech recognition employing key word modeling and non-key word modeling
JP3672595B2 (en) Minimum false positive rate training of combined string models
EP0880126B1 (en) Speech-silence discrimination based on unsupervised HMM adaptation
US6125345A (en) Method and apparatus for discriminative utterance verification using multiple confidence measures
US6260013B1 (en) Speech recognition system employing discriminatively trained models
Jelinek et al. 25 Continuous speech recognition: Statistical methods
Kenny et al. A*-admissible heuristics for rapid lexical access
Normandin Maximum mutual information estimation of hidden Markov models
EP0314908A2 (en) Automatic determination of labels and markov word models in a speech recognition system
US20080300875A1 (en) Efficient Speech Recognition with Cluster Methods
US20030055640A1 (en) System and method for parameter estimation for pattern recognition
EP0786761A2 (en) Method of speech recognition using decoded state sequences having constrained state likelihoods
US6182036B1 (en) Method of extracting features in a voice recognition system
US6253178B1 (en) Search and rescoring method for a speech recognition system
US6301561B1 (en) Automatic speech recognition using multi-dimensional curve-linear representations
Knill et al. Hidden Markov models in speech and language processing
Deligne et al. A robust high accuracy speech recognition system for mobile applications
US6055499A (en) Use of periodicity and jitter for automatic speech recognition
Christensen et al. A comparison of three methods of extracting resonance information from predictor-coefficient coded speech
Lee et al. Acoustic modeling of subword units for speech recognition

Legal Events

Date Code Title Description
EEER Examination request
MKEX Expiry