WO2003005345A1 - Speech recognition with dynamic grammars - Google Patents

Speech recognition with dynamic grammars Download PDF

Info

Publication number
WO2003005345A1
WO2003005345A1 PCT/US2002/021364 US0221364W WO03005345A1 WO 2003005345 A1 WO2003005345 A1 WO 2003005345A1 US 0221364 W US0221364 W US 0221364W WO 03005345 A1 WO03005345 A1 WO 03005345A1
Authority
WO
WIPO (PCT)
Prior art keywords
grammar
word
sub
runtime
context
Prior art date
Application number
PCT/US2002/021364
Other languages
French (fr)
Inventor
Johan Schalkwyk
Michael S. Phillips
Original Assignee
Speechworks International, Inc.
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Priority claimed from US09/906,390 external-priority patent/US20030009335A1/en
Application filed by Speechworks International, Inc. filed Critical Speechworks International, Inc.
Publication of WO2003005345A1 publication Critical patent/WO2003005345A1/en

Links

Classifications

    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS OR SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING; SPEECH OR AUDIO CODING OR DECODING
    • G10L15/00Speech recognition
    • G10L15/08Speech classification or search
    • G10L15/18Speech classification or search using natural language modelling
    • G10L15/183Speech classification or search using natural language modelling using context dependencies, e.g. language models
    • G10L15/187Phonemic context, e.g. pronunciation rules, phonotactical constraints or phoneme n-grams
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS OR SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING; SPEECH OR AUDIO CODING OR DECODING
    • G10L15/00Speech recognition
    • G10L15/08Speech classification or search
    • G10L15/18Speech classification or search using natural language modelling
    • G10L15/183Speech classification or search using natural language modelling using context dependencies, e.g. language models
    • G10L15/19Grammatical context, e.g. disambiguation of the recognition hypotheses based on word sequence rules
    • G10L15/193Formal grammars, e.g. finite state automata, context free grammars or word networks

Abstract

The invention includes a method for speech recognition using cross-word contexts on dynamic grammars. The invention also includes a method for constructing a speech recognizer capable of speech recognition using cross-word contexts on dynamic grammars, by expanding a word of the main grammar into a corresponding network of sub-word units. The sub-word units are selected from the plurality of sub-word units based in part on a pronunciation of the word. Each sub-word unit has a permissible context including constraints on neighboring sub-word units within the corresponding network. The corresponding network is chosen to satisfy the constraints of the permissible context of each sub-word unit within the corresponding network. When the context of sub-word units would have apply to words provided by a runtime grammar, the expansion includes every sub-word unit that satisfies the permissible context, when compared to the corresponding network.

Description

SPEECH RECOGNITION WITH DYNAMIC GRAMMARS
TECHNICAL FIELD
This invention relates to machine-based speech recognition, and more particularly to machine-based speech recognition with dynamic grammar, and machine-based speech recognition with context dependency.
BACKGROUND
A speech recognition system maps sounds to words, typically by converting audio input, representing speech, to a sequence of phoneme or phones. The phoneme sequence is mapped to words based on one or more pronunciations per word. Words and acceptable sequences of words are defined in a main grammar. The chain of these mappings, from audio input through to acceptable sentences in a grammar, allows the speech recognition process to recognize speech within the audio input and to map the speech input to output values, such as the recognized text string and a confidence measure.
Context-dependent speech recognition uses context-dependent models to improve the accuracy of speech recognition. Context-dependent models are models of how an utterance can occur in the audio input stream. Typically, a context- dependent model corresponds to a linguistic component of a word, such as a phoneme or a phone, as it might be uttered in speech - that is, in context. Because the corresponding component will usually have contexts in which it might occur, several context-dependent models can correspond to one component. One form of context- dependent speech recognition, therefore, maps audio input to context-dependent models, context-dependent models to pronunciations, and pronunciations to words.
The generation of the mappings from audio input to grammar is performed on a computer.
FINITE STATE MACHINES
Finite state machines (FSMs) can encode linguistic models on a computer. An FSM can represent a system that accepts inputs and responds predictably by changing state among a finite number of possible states. Thus, an FSM can be a recognizer, if it meets the following criteria. An initial state receives input submissions. (A submission is an instance of an FSM's operation on an input string. Even if the same input string is submitted twice, there are two submissions.) For each submission, and at any given moment, an FSM has exactly one state that is current. A final state causes an FSM to finish operating on a submission. Since it is desirable that a recognizer halt and return a result for each submission, we require that an FSM recognizer have at least one final state. A state may be both initial and final.
A recognition attempt begins with a submission, which provides an input string. The FSM allocates a session to the submission. The session will return a result indicating acceptance or rejection of the input string.
A finite state transducer (FST) differs from a finite state acceptor (FST) in that the FST arcs include output labels that are added to an output string for each submission. For an FST, each session will return an output string along with its result.
The session includes a current state and an input pointer. The current state is initialized to one of the machine's initial states. The input pointer is set to the beginning of the input string. The FSM evaluates the state transitions departing the current state as follows. A state transition has at least one input symbol and a next state, while the input string has a substring starting from a location defined by the input pointer. The input symbol has a defined pattern of characters that it will match. If the characters at the beginning of the substring qualify to match the input symbol's pattern, the transition accepts the input. Acceptance moves the current state to the transition's "next" state, and the input pointer moves to the first character beyond the portion matched by the pattern. In this manner, the transition "consumes" the matched portion. An epsilon transition has the empty string "" (also known as "epsilon" or "eps") for its input symbol. An epsilon transition accepts without consuming any input. One use of an epsilon transition is, in effect, to join a second state (pointed to by the epsilon transition) to a first state, since any path that reaches the first state can also reach the second state on identical inputs. If an epsilon transition joins a state in a first FSM to a state
If the transition has an output symbol, the output is put out during acceptance. Evaluation of the state transitions begins anew from the current state. The session becomes stuck if no transitions from the current state accept the input. This can happen if there are no transitions to match the input; or, in the absence of epsilon transitions, this can happen if the input string is entirely consumed, so that there is no input to match the transitions. The session halts (a different and more constructive result than becoming stuck) when the current state is a final state. The recognition attempt succeeds if the session halts on a final state with the input string entirely consumed. Otherwise, the recognition attempt fails.
A FSM is sometimes described as a network or graph.
SUMMARY
In general, in one aspect, the invention includes a method for speech recognition. The method includes providing a main grammar containing a placeholder for a runtime grammar. The main grammar is provided at design-time and contains zero or more words. The method includes providing a plurality of sub- word units. In one embodiment, the sub-word units are context-dependent models. The method further includes expanding a word of the main grammar into a corresponding network of sub-word units selected from the plurality of sub-word units. The expanding is based in part on a pronunciation of the word. Each such sub- word unit has a permissible context, including constraints on neighboring sub-word units within the corresponding network. The corresponding network is chosen to satisfy the constraints of the permissible context of each sub- word unit within the corresponding network. When the sub-word units provided by expansion would have permissible context applicable to words provided by a runtime grammar, the expansion of the word of the main grammar includes multiple sub-word units, one for every sub-word unit whose permissible context allows it to be a neighbor to the corresponding network.
Preferred embodiments include one or more of the following features. The method may include specifying after design-time the runtime grammar to associate with the placeholder, and expanding a word of the runtime grammar somewhat similarly to the expansion of word in the main grammar. Specifically, a word of the runtime grammar is expanded into a conesponding network of sub-word units selected from the plurality of sub-word units. Each sub-word unit has a permissible context including constraints on neighboring sub-word units within the conesponding network. The conesponding network is chosen to satisfy the constraints of the permissible context of each sub-word unit within the conesponding network. When the sub-word units conesponding to the word to be expanded would have permissible context applicable to words in the main grammar, the expansion of the word into the conesponding network includes every sub-word unit from the plurality of sub-word units subject to the constraints of the sub-word unit's permissible context as compared to the conesponding network. In a further embodiment, the method may include j oining the runtime grammar to the main grammar by creating paths from the main grammar into runtime grammar and back again. The paths are selected to satisfy the constraints of the permissible context of each sub-word unit in the expansion of the main grammar and in the expansion of the runtime grammar. The method further includes accepting audio input representing speech from a speaker, analyzing the audio input to produce a string of sub-word units, and using the expansion of the main grammar joined with the expansion of the runtime grammar to recognize the string of sub-word units.
The invention includes the following advantages. It is not always desirable to prepare every step of the speech recognizer in advance of deploying the speech recognition system. Preparing the mappings, from audio input through to acceptable sentences in a grammar, consumes computing resources. An total preparation may be an inefficient use of these resources. For instance, portions of a mapping may never be needed, so the resources used to prepare these portions may be wasted. Also, for large grammars, the mappings may require large amounts of storage. The processing time may also increase with grammar size.
There are advantages to being able to leave portions of the grammar incomplete until runtime. Not every component of the grammar may be knowable at design time. A dynamic grammar adds flexibility to the speech recognition system. The speech recognition system might adapt, for instance, to the characteristics, including needs or identities, of specific users. A dynamic grammar can also usefully constrain the range of speech that the speech recognition system must be prepared to recognize, by expanding or contracting the grammar as necessary.
The details of one or more embodiments of the invention are set forth in the accompanying drawings and the description below. Other features, objects, and advantages of the invention will be apparent from the description and drawings, and from the claims.
DESCRIPTION OF DRAWINGS
FIG. 1 A is a block diagram of a speech recognition system. FIG. IB is a block diagram of a computing platform.
FIG. 2A is a flowchart of a process including a design-time mode and a runtime mode.
FIG. 2B is a block diagram of a recognizer process.
FIG. 3 A is a block diagram of a transducer combination process. FIG. 3B is a block diagram of basic grammar structures.
FIG. 4 is a block diagram of design-time preparations.
FIG. 5 is a block diagram of a finite state machine optimization of a lexicon.
FIG. 6 is a block diagram of a context-factoring example.
FIG. 7 is a block diagram of a grammar-to-phoneme compiler. FIG. 8 A is a block diagram of a composition process.
FIG. 8B is a block diagram of an example of a finite state machine rewrite.
FIG. 9 is a flowchart of a known finite state machine composition process.
FIG. 10 is a flowchart of a finite state machine composition process.
FIG. 11 is a block diagram of a finite state machine composition process, with examples.
FIG. 12 illustrates deriving context-dependent models.
Like reference symbols in the various drawings indicate like elements.
DETAILED DESCRIPTION
One approach to context-dependent speech recognition maps audio input to context-dependent models, context-dependent models to pronunciations, and pronunciations to words, h the present embodiment, finite state machines represent words, pronunciations, variations in pronunciation, and context-dependent models. The necessary mappings between them are encoded in a single FSM recognizer by constructing the recognizer from smaller machines using FSM composition. Contexts at the boundary of a dynamic grammar, as will be explained in more detail, are not fully known in advance. The invention allows speech recognition using context-dependent models, even when contexts span boundaries between a main grammar (known at design-time) and dynamic portions (provided later). one embodiment, and with regard to FIG. 1 A, a speech recognition system 22 includes an audio input source 23, a sound-to-model converter 24, and a recognizer
40.
The audio input source 23 provides a sound signal (not shown) in digitized form to the sound-to-phoneme converter 24. The sound signal may capture speech of a live speaker whose voice is sampled by a microphone. The sampled voice is then digitized to create the sound signal. Alternatively, the sound signal may be derived from a pre-recorded source.
As shown in FIG. 2 A, in a design-time mode 61, a main grammar 30, which contains words and sentences to recognize, becomes a main transducer 43 that includes context-dependent phoneme models. Broadly speaking, the main transducer 43 can process phoneme strings (such as provided by the sound-to-phoneme converter
24) into the words and sentences of the main grammar 30.
The words to be recognized, i.e. the main grammar 30, might not always be known during the design-time mode 61. We may wish to recognize words and sentences that are not provided until runtime; this requires a "dynamic" grammar. A dynamic portion of the grammar may be provided as a runtime grammar 32. After the speech recognition system 22 transitions to a runtime mode 66, a runtime grammar 32 is converted to a runtime transducer 44. A transducer combination process 42 then integrates the runtime transducer 44 and the main transducer 43, using phoneme context models even across boundaries between words in the main grammar 30 and words in the runtime grammar 32. COMPUTING ENVIRONMENT
FIG. IB shows a speech recognition system 22 on a computing platform 63.
The speech recognition system 22 contains computer instructions and runs on an operating system 631. The operating system 631 is a software process, or set of computer instructions, resident in either main memory 634 or a non- volatile storage device 637 or both. A processor 633 can access main memory 634 and the nonvolatile storage device 637 to execute the computer instructions that comprise the operating system 631 and the speech recognition system 22.
A user interacts with the computing platform via an input device 632 and an output device 636. Possible input devices 632 include a keyboard, a microphone, a touch-sensitive screen, and a pointing device such as a mouse, while possible output devices 636 include a display screen, a speaker, and a printer.
The non- volatile storage device 637 includes a computer-writable and computer-readable medium, such as a disk drive. A bus 635 interconnects the processor and motherboard 633, the input device 632, the output device 636, the storage device 637, main memory 634, and optional network connection 638. The network connection 638 includes a device and software driver to provide network functionality, such as an Ethernet card configured to run TCP/IP, for example.
The recognizer 40 may be written in the programming language C. The C code of the recognizer 40 is compiled into lower-level code, such as machine code, for execution on a computing platform 63. Some components of the recognizer 40 may be written in other languages such as C++ and incorporated into the main body of software code via component interoperability standards, as is also known in the art. In the Microsoft Windows computing platform, for example, component interoperability standards include COM (Common Object Model) and OLE (Object Linking and Embedding).
DESIGN-TIME MODE
FIG. 2 A shows a design-time mode 61, which represents a state of the recognizer 40 before it is deployed to a runtime environment. A runtime transition 65 represents the transition to a runtime mode 66. The design-time mode 61 includes a main grammar 30, a grammar-to- phoneme compiler 50, a design-time preparations process 71, and a main transducer 43. As is shown in FIG. 2B, the main transducer 43 is included in the recognizer 40.
MAIN GRAMMAR Broadly speaking, the main grammar 30 specifies the words and sentences that the recognizer 40 will accept.
Some general properties of a grammar are illustrated in FIG. 3B. As will be explained in more detail, subgrammars can be integrated into the main grammar 30. General grammar properties are shared by the main grammar and its subgrammars. A main grammar 30 and a runtime grammar 32 (see FIG. 2B) have properties in common, some of which are shown in FIG. 3B.
An alphabet 316 is a set of symbols (not shown), which can be used to spell a word 312 or token 321.
A word 312 is an arrangement of symbols from the alphabet 316; the anangement is called the spelling (not shown) of the word 312. Spelling is known in the art. Not all symbols in the alphabet 316 need be used in words 312; some may have special purposes, including notation.
A sequence of one or more words 312 forms a sentence 313. A word 312 may appear in more than one sentence 313, as shown by sentences 313a and 313b of FIG. 3, which both contain word 312a. The spelling of a word 312 is not necessarily unique: two identical spellings may be distinguished by their meaning.
Like a word 312, a token 321 is an anangement of symbols from the alphabet 316. The collection of all words 312 and tokens 321 in a grammar is called the namespace 314. Unlike a word 312, each token 321 has a unique spelling within the namespace 314. A word 312 usually has semantic meaning in some domain (for instance, the domain of speech that the speech recognition system 22 is designed to recognize), while a token 321 is usually a placeholder for which some other entity can be substituted. DESIGN-TIME PREPARATIONS
With reference to FIG. 4, design-time preparations 71 include providing linguistic models 72, lexicon preparations 73, and context factoring 35.
The linguistic models 72 are constructed by processes that include a raw lexicon 721 , called "raw" here to distinguish its initial form from the lexicon produced by lexicon preparations 73, as well as phonological rules 722, context dependent models 723, a pronunciation dictionary 724, and a pronunciation algorithm 725.
RAW LEXICON The raw lexicon 721 contains pronunciation rules for words in the main grammar 30. The rules are encoded in an FSM transducer by using input symbols on the arcs of the FSM drawn from a phonemic alphabet. The output of the raw lexicon 721 transducer includes words in the main grammar 30 and words provided by runtime grammars 32.
CONTEXT DEPENDENT MODELS
The context dependent models 723 model the sound of phonemes spoken in real speech. FIG. 12 shows elements in a process (77) to derive the context dependent models 723. Context dependent models 723 are a form of sub-word units.
The context dependent models 723 are derived empirically from training data 771 using data-driven statistical techniques 775 such as clustering. The training data 771 includes recordings 772 of a variety of utterances selected to be representative of speech that will be presented to the speech recognition system 22. Selecting training data 771 is complex and subjective. Too little training data 771 will not provide sufficient grounds for statistical distinction between two different yet acoustically similar phonemes, or between contextual changes for a given phoneme. On the other hand, too much training data 771 can cause the system to infer undesirable statistical patterns, for example, patterns that happen to appear in the training data but are not characteristic of the general range of input.
A recording 772 has a time measure 770. Alignments 774 relate a sequence of phonemic symbols 773 to the time measure 770 within the recording 772, to indicate the portions of the recording 772 that represents an utterance of the phonemic symbols 773.
For a given phoneme, its phonological context describes permissible neighbors. The neighbors can occur both before and after in time, notated as left and right, respectively. There are several ways to model context, including tri-phonic, penta-phonic, and tree-based models. This embodiment uses tri-phonic contexts with phonemes, which weigh three phonemes at a time: a current phoneme and the phonemes to the left and right.
For a given phoneme, the data-driven statistical techniques 775 derive a phonemic decision tree 776, which categorizes all possible context models for the given phoneme according to a tree of questions. The questions are Boolean- valued (yes/no) tests that can be applied to the given phoneme and its context. An example question is "Is it a vowel?", although the questions are phrased in machine-readable code. For a given branch of the tree, traversing outward from the root, subsequent questions refine earlier questions. Thus, a subsequent question for the earlier question might be "Is it a front vowel?"
The data-driven statistical techniques 775 select a question as the most distinctive question (according to a statistical measure) and label it the root question. Subsequent questions are added as children of the root question. The recursive addition of questions can continue automatically to some predetermined threshold of statistical confidence. However, the structure of the phonemic decision tree 776 - that is, the infrastructure of the questions - may also be tuned by human designers.
The phonemic decision tree 776 is a binary tree, reflecting the Boolean values of the questions. The leaves of the tree are model collections 778, which contain zero or more models 779. Initially the model collections 778 contain models 779 detected in the training data 771 by the data-driven statistical techniques 775. The context dependent models derivation process 77 adds models 779 that do not occur in the training data 771 to the phonemic decision tree 776, only after all questions have been added by the traversing the tree for each model 779. Modelss 779 are added by evaluating the question nodes against the model 779, then following the corresponding branches recursively until reaching a model collection 778 that receives the model 779. Like the raw lexicon 721, context dependent models 723 are also encoded in an FSM transducer. The transducer maps sequences of names of context-dependent phone models to the conesponding phone sequence. The topology of this transducer is determined by the kind of context dependency used in modeling. The input symbols of a tri-phonic phonemic context FSM use the phonemic alphabet with additional characters to represent positional infonriation or other information "tags" such as end- of-word, end-of-sentence, or a homophonic variant. Input symbols are of the form "x/y_z", where x represents the cunent phoneme in the input string, y and z are left and right neighbors, respectively. In this case, the center character x is never a tag character. Positional characters include "#h" (which indicates a sentence beginning) or "h#" (sentence end). Homophonic characters include "#1", "#2", etc. A word- boundary character is ".wb".
PHONOLOGICAL RULES
Phonological rales 722 are also encoded in an FSM transducer. Phonological rules 722 introduce variant pronunciations as well as phonetic realizations of phonemes. Unlike a lexicon L, which maps phoneme sequences to words, P affects phoneme sequences that are not necessarily entire words. P's rules are contextual, and the contexts may apply across word boundaries. In practice, though, there can be benefits to expressing any phonological rules that are context-dependent in the context dependent models 723 instead of the phonological rules 722. This centralizes all contextual concerns into a single machine and also simplifies the role of the phonological transducer 57.
The input symbols of the phonological rules 722 FSM use the same extended phonemic alphabet and the same matching rales as the context dependent models 723 FSM, but the contexts of the phonological rules are not restricted to triplets, and the phonological rules 722 may rewrite their inputs with one more characters from the pure phonemic alphabet.
The pronunciation generator 726 offers a way to find a pronunciation of a word. The pronunciation generator 726 therefore allows the use of dynamic grammars that are not constrained against the vocabulary of the lexicons 721 and 52. The pronunciation generator 726 takes input in the form of a word and returns a sequence of phonemes. The sequence of phonemes is a pronunciation of the input word. The pronunciation generator 726 uses a pronunciation dictionary 724 and a pronunciation algorithm 725. The pronunciation dictionary 724 provides known phonemic spellings of words. The pronunciation algorithm 725 contains rules hand- crafted to a phoneme set known to be acceptable to the context dependent models 723. Basing the pronunciation algorithm 725 on this phoneme set insures against collisions between algorithmic guesses and impermissible contexts. The pronunciation algorithm 725 is tuned by its human designers to meet subjective parameters for acceptability; in English, for example, which is not an especially phonetic language, the parameters can be quite approximate.
The pronunciation generator 726 works as follows. The pronunciation generator 726 first consults the pronunciation dictionary 724 to see if a known pronunciation for the input words exists. If so, the pronunciation generator 726 returns the pronunciation; otherwise, the pronunciation generator 726 returns the best- guess produced by passing the input word to the pronunciation algorithm 725. More than one pronunciation may be acceptable, and thus more than one pronunciation may be returned.
LEXICON PREPARATIONS
Lexicon preparations 73 include a disambiguate homophones process 731, a denote word boundaries process 732, and an FSM optimization process 74. The disambiguate homophones process 731 introduces auxiliary symbols into the raw lexicon 721 to denote two words that sound alike. An example in English is "red" and "read", which both map to the phonemes /r eh αV. This sort of homophone ambiguity can cause infinite loops in the determinization of the raw lexicon 721. Auxiliary notation, such as /r eh d #1/ for red and IT eh d #21 for read, can remove the ambiguity. The auxiliary notation can be removed after determinization, for instance by extending the function of the right transducer Cr 55 with self-looping transitions on each such auxiliary symbol. The self-looping transitions would consume the auxiliary symbols. The denote word boundaries process 732 also adds an auxiliary symbol: ".wb" indicates a word boundary. The FSM optimization process 74 performs FSM algorithms for determinization 741, minimization 743, closure 745, and epsilon removal 747 on the raw lexicon 721 FSM. FIG. 5 illustrates the effects of these operations on an example raw lexicon 721. The output of the FSM optimization process 74 is the lexicon transducer L 52, ready for composition with the main grammar 30.
CONTEXT FACTORING
With regard to FIG. 6, the context factoring process 35 derives (step 331) the left transducer CI 54 and the right transducer Cr 55 from the FSM transducer for the context dependent models 723. The right transducer Cr 55 is extended to include self- looping transitions on each such homophone disambiguation symbol. Both the left transducer CI 54 and the right transducer Cr 55 may include a phonological symbol indicating unknown context, as for instance may exist for a neighbor of a runtime grammar 32. Following the derivation, the context factoring process 35 determinizes the transducers 54 and 55. Among other reasons, determinizing improves performance of the transducers 54 and 55 after composition.
GRAMMAR-TO-PHONEME COMPILER
Referring now to FIG. 7, the grammar-to-phoneme compiler 50 takes input in the form of an input grammar G 51 and returns a phonological and context-dependent lexical-grammar machine 59, also called "PoCoLoG" for the FSM compositions it contains. The grammar-to-phoneme compiler 50 uses linguistic models encoded as
FSMs, including: a lexicon transducer L 52; a set of context transducers 501 that includes a left transducer CI 54 and a right transducer Cr 55; and a phoneme transducer 57. As will be explained in more detail, the grammar-to-phoneme compiler 50 uses a chain of compositions, passing the output of one as input to the next. The chain includes a composition with L 53, a composition with C 56, and a composition with P 58.
COMPOSITION OF G WITH L
With regard to FIG. 8 A, the composition with L 53 produces an FSM that takes in phonemes and turns out words. More specifically, the composition with L 53 composes (step 532) an input grammar G 51 with the lexicon transducer L 52. The input grammar G 51 may include the main grammar 30, which is shown in the design- time mode of FIG. 2 A, or a runtime grammar 32 from the runtime grammar collection 33, which is shown in the runtime mode 66, also in FIG. 2A. FIG. 8B illustrates an example of the composition with L process 53 in action.
For clarity, FIG. 8B uses subsets of the example machines shown in FIG 8 A. An arc in G 512 has an input symbol 513, a departed state 516, and a next state 517. A pronunciation path 521 in L 52 contains a first arc having an output symbol 524 and an input symbols that represents a first phoneme in a pronunciation of a word represented in the output symbol 524. The pronunciation path 521 optionally contains subsequent states and arcs after the first arc, daisy-chained in the manner shown in FIG. 8B. Subsequent arcs have output symbols of "eps" if they exist. The final arc in the pronunciation path 521 points to a final state 529 in L, although the final state 529 is not included in the pronunciation path 521. The final state 529, by being final, denotes a word boundary. Thus, the sequence of arcs in the pronunciation path 521 conesponds to a word, as follows: the sequence's first arc outputs a word; no subsequent arcs output anything but "eps"; the first arc accepts a first phoneme of a word's pronunciation; and subsequent arcs contribute subsequent phonemes until the final arc, which points to a word boundary which terminates the word. The resulting FSM 539, which can be denoted LoG, is a rewrite of G 51 by L
52. The composition according to the following known composition process 591 is illustrated in FIG. 9. The known composition process 591 initializes an empty output FSM 539 and copies all states of G into the empty output FSM 539 (step592). The known composition process 591 loops first through one arc 512 in G 51 at a time (step 593). In a sub-loop for each input symbol 513 on the current arc 512 (step 594), the known composition process 591 compares each input symbol 513 to each output symbol 524 on arcs in L 52 (step 595). When this comparison 595 yields a match, the known composition process 591 copies each matching pronunciation path 521 from L 52 into LoG 539 (step 596). The pronunciation path 521 conesponds to an acceptable pronunciation of the input symbol 513.
The pronunciation path 521 begins with the arc in L whose output symbol matched the input symbol and continues until a word boundary is matched. In the example of FIG. 8B, the input symbol 513 is "Works," while the pronunciation path 521 contains arcs having input symbols /w/, /er/, Ikl, and Is/ respectively. The first arc on the pronunciation path 521 has an output symbol 524 of "Works" which matches the input symbol 513 of the arc in G 512. Any intermediate states on the path are copied into LoG 539 as well; in the example, these include states labeled "1", "3", and "5" in L, which are mapped to states labeled "1", "3a", and "5" in the output LoG 539. Additional minimization and other optimization steps may be performed on LoG 539 which may rename its states to achieve the final naming shown in FIG. 8 A, where the internal states of the pronunciation path 521 are named 6, 7, and 8, respectively. The first arc in the pronunciation path 521 when written into LoG 539 departs from the same state in LoG 539 that the original departing state 516 in G maps to. h terms of the example of FIG. 8B, the state labeled "2" of LoG 539 has a departing arc with label "w: Works" that conesponds to the first arc in path 521. Similarly, the final arc in the pronunciation path 521 points to the same state in LoG 539 that the original next state 517 arc maps to. Again put in terms of the example of FIG. 8B, the state labeled "3" of LoG 539 has an incoming arc with label "s:eps" that conesponds to the last arc in path 521. The state labeled "3" happens to be a final state in LoG 539 because that was its role in G 51 in this example, as shown in G 51 of FIG. 8 A, but in the general case the state labeled "3" could be any state in G 51. When the comparison 595 does not yield a match, the known composition process 591 can invoke a pronunciation generator 726 to find a pronunciation and convert the pronunciation to a representation as a pronunciation path 521.
The known composition process 591 continues looping on symbols (step 597) and arcs (step 598) until all arcs and symbols in G have been processed, at which time the known composition process 591 may apply FSM operations to LoG 539 such as minimization, determinization, closure, and epsilon removal to normalize the LoG 539 FSM (step 599).
The composition with L process 53 is similar to the composition process 591 but has at least two differences. Referring now to FIG 10, one difference is that before comparing the input symbol 513 with output symbols 524 of arcs in L 52 (step 595), the composition with L process 53 checks whether the input symbol 513 matches a token 321 in the runtime grammar collection 33 (step 534). A second difference is that if the input symbol 513 matches such a token 321, the composition with L process 53 writes a one-arc path into LoG 539. The sole arc has the phonemic symbol for runtime class 735 as its input symbol, which is "*", and the value of the token 321 as its output symbol. (The 5 symbol "*" is a placeholder that helps manage ambiguous context at the border of a runtime grammar 32.) The composition with L process 53 then returns to looping on input symbols (step 597).
When the known composition process 591 has processed all arcs in G, LoG 539 accepts input strings in the form that L does: phonemes. Acceptance of a o phoneme string by LoG 539 is precisely the acceptance one would see if the string were first submitted to L 52, which transduces phonemes to words, and the words were then submitted to G 51 as input. The acceptance behavior and output of the transducer LoG 539 will match the acceptance behavior and output of G 51.
COMPOSITION WITH C 5 The grammar-to-phoneme compiler 50 uses the composition with C process
56 to convert a phoneme-accepting transducer to a transducer that accepts context- dependent models. Specifically, the composition with C process 56 factors the context dependent models FSM 723 into FSMs for right and left context, then uses these FSMs to rewrite LoG 539, where LoG 539 may be based on the main grammar 0 30 or a runtime grammar 32.
The result of the composition with C process 56 is an FSM transducer that can use context-dependent models as input and has the outputs and word-acceptance behavior of the underlying grammar in LoG 539. Thus, the chain of recognition is extended from grammar down to context-dependent models. The composition with C 5 process 56 also constrains the number of phoneme combinations that must be examined when considering phonemic context across the edge of a runtime grammar 32. Constraining the number of combinations improves runtime performance of the recognizer 40.
More specifically, the composition with C process 56 accepts the LoG 0 machine 539 as input; composes the reverse of the machine 539 with the right transducer Cr 55 to form a machine Cr o rev(LoG), then reverses Cr o rev(LoG) and composes it with CI. This final context-dependent LoG machine 569 is returned as output.
Thus, the formula for the context-dependent LoG machine 569 in terms of FSM operations is: Cl o rev( Cr o [ rev(LoG) ] )
The standard FSM composition operation must be extended to handle "*", the phonemic symbol for runtime class 735. The composition with C process 56 replaces arcs in LoG 539 having phonemic input labels matching "*" with a collection of arcs, each arc in the collection conesponding to an input label given by a context model in the context dependent models FSM 723. Broadly speaking, therefore, the composition with C process 56 constrains the values of"*" to known permissible values, where "permission" entails being part of a context for which a context model exists.
The replacement includes a departing arc collection 561 and a returning arc collection 562.
FIG. 11 shows a sequence of steps in the composition with C process 56 and the effects of the steps on two samples: a portion of an example input LoG 539, and a sample runtime grammar 32, refened to in this example by its token "$try".
The composition with C process 56 copies the input machine 539 to a cunent machine FSM 565. The cunent machine FSM 565 is the work-in-progress version of the FSM that will be returned as the output FSM 569.
The composition with C process 56 sets the cunent machine FSM 565 to be the FSM reversal of the input LoG machine 569 (step 564). The composition with C process 56 then composes Cr 55 with the reversed LoG 569 (step 566). The input FSM is reversed so that it may be traversed to find right contexts without backtracking: post-reversal, the right context of the cunent arc is always in the portion of the machine already traversed.
The input label for an arc in LoG 569 is a phoneme, to be replaced with one or more context-dependent models. When rewriting a given arc with Cr 55 (step 566), the composition with C process 56 considers the arc's input label, as well as the input label of the previous arc (in the reversed LoG 569), which gives the right context for the cunent phoneme. The given arc label is then replaced with every context- dependent model 779 that matches the current phoneme and its right context. For the examples shown in FIG. 11, the input label on the arc passing from state "in" to state "iv" is rewritten from the phoneme "r" to the context-dependent models "r.4", "r.8", and "r.15". (The sequence for these is written as "r.4.8.15".) This indicates that three models were found for the phoneme "r" having right context "y". Similarly, the input label on the arc passing from state "iv" to state "v" is rewritten from the phoneme "y" to "y.1-20". All models on y from "y.l" to "y.20" matched the context because the right context is "*", which represents the border of a runtime grammar 32. Since "*" could be anything, it matches every context. Also in composing Cr 55 with the reversed LoG 569 (step 566), the composition with C process 56 removes any homophone symbols from the current machine FSM 565 that were introduced into L by the disambiguate homophones process 731.
Next, the composition with C process 56 reverses (step 567) the cunent machine FSM 565 again. This second application of FSM reversal restores the original order of paths within LoG 539.
The composition with C process 56 then composes CI 54 with the cunent machine FSM 565 (step 568). This traversal of the cunent machine FSM 565 matches a phoneme (no longer represented by a phonemic symbol, but readily apparent from the context-dependent model that has replaced it) and its left phonemic context with the context-dependent models encoded in CI 54. The matching further constrains the context-dependent models which have replaced the phoneme; and, since constraints for both right context and left context have now been applied, the constraints are the same as would be applied by the un-factored FSM of context dependent models 723.
When both the left and right phonemic contexts of an input label are known (in triphone-based context schemes), they uniquely determine a context dependent model for the input label.
After composition of the cunent machine 565 with CI to produce a new cunent machine 565 (step 568), the composition with C process 56 returns the cunent machine 565 as the context-depended LoG machine 569. COMPOSITION WITHP
The grammar-to-phoneme compiler 50 uses the composition with P process 58 to include phonemic rewrite rules in the phoneme transducer that the grammar-to- phoneme compiler 50 constructs. The phonemic rewrite rules are encoded in the phonological rales FSM 722, also known as P, and include rales for alternate pronunciations. The phonemic rewrite rales can be contextual, and their contexts can cross word (and therefore runtime grammar 32) boundaries.
The transducer P 722 maps phonemes to phones, but the machine 569 returned by the composition with C process 56 has context-dependent models for input labels. However, since a phonemic symbol is readily apparent from the context-dependent model that has replaced it, the composition with P process 58 can use known FSM composition techniques.
The composition with P process 58 returns a context-dependent lexical- grammar machine 589 (not shown) to the grammar-to-phoneme compiler 50. The grammar-to-phoneme compiler 50, in turn, returns the same machine as output: the phonological and context-dependent lexical-grammar machine 59.
TRANSDUCER COMBINATION
The transducer combination process 42 enables context-dependent recognition of input strings that cross a boundary between the main transducer 43 and a runtime transducer 44.
The transducer combination process 42 includes at least two modes: an endset transducer 45 and a subroutine transducer 60.
ENDSET TRANSDUCER
The endset transducer 45 creates paths across boundaries between the main transducer 43 and a runtime transducer 44, subject to context constraints, by linking arcs and states at the edge of each transducer 43 and 44 with epsilon transitions. The endset transducer 45 produces continuous paths from the main transducer 43 into the runtime transducer 44 and vice versa.
FIG. 3 A shows example portions of a main transducer 43 and a runtime transducer 44. The endset transducer 45 rewrites an arc 452 in the main transducer 43 that represents a rantime transducer 44. Such an arc 452 has "*" as an input label and a token 321 as an output label. The arc 452 is not removed permanently but is routed around: the endset transducer 45 adds a temporary path using two epsilon transitions. One epsilon transition 454 goes from the main transducer 43 into the runtime transducer 44. Specifically, the epsilon transition 454 departs from the same state that arc 452 departs from and points to the state in the runtime transducer 44 after its first arc. (The first arc in the rantime transducer 44 has "*" as an input label, acting as a placeholder at the border of a dynamic grammar.)
The second epsilon transition 458 returns from the rantime transducer 44 to the main transducer 43. Specifically, the second epsilon transition 458 departs the same state in the runtime transducer 44 that a last arc departs. (Each last arc in the runtime transducer 44 has "*" as an input label, acting as a placeholder at the border of a dynamic grammar.) The second epsilon transition 458 points to the same state in the main transducer 43 that the arc 452 points to. The endset transducer 45 adds epsilon transitions 454 and 458 subject to context constraints encoded in the context dependent models 723. For epsilon transition 454, and with regard to the path that it would enable from the main transducer 43 into the runtime transducer 44, there exists an arc 453 immediately prior to transition 454, as well as an arc 455 immediately after. The input labels of arc 453 provide a left context to the input labels of arc 455, just as the input labels of arc 455 provide a right context to the input labels of arc 453. The endset transducer 45 requires that the context requirements of both arcs 453 and 455 be satisfied before adding epsilon transition 454.
Similarly, an arc 457 exists prior to epsilon transition 458 on the return path from the rantime transducer 44 to the main transducer 43, and an arc 459 exists after. Arc 457 provides arc 459's left context, just as arc 459 provides arc 457's right context. The endset transducer 45 requires that the context requirements of both arcs 457 and 459 be satisfied before adding epsilon transition 458.
The main transducer 43 includes a main departing arc collection 421, a main returning arc collection 422, a main last arc 423, and a main first arc 424. The runtime transducer 44 includes a runtime departing arc collection 426, a runtime returning arc collection 427, a runtime last arc 428, and a runtime first arc 429. ALTERNATE EMBODIMENTS
A number of embodiments of the invention have been described. Nevertheless, it will be understood that various modifications may be made without departing from the spirit and scope of the invention. For example, instead of triphonic models of phonological context, penta-phonic and tree-based context models may be used. Instead of phoneme-based context-dependent models, context-dependent models based on phones may be used. Tokens 321 may be replaced with respective runtime grammars prior to composition. Also, the composition with L process 53 could switch the order in which it tests whether the input symbol 513 is a placeholder for a rantime class, i.e., it could perfonn this test after looking in L for a match. Accordingly, other embodiments are within the scope of the following claims.

Claims

WHAT IS CLAIMED IS:
1. A method for speech recognition comprising: designating a main grammar containing a placeholder for a rantime grammar; providing the runtime grammar; and performing context-dependent speech recognition based on the main grammar containing the placeholder and on the runtime grammar, the speech recognition using cross-word contexts for words adjacent in the main grammar and the rantime grammar.
2. The method of claim 1 in which performing speech recognition includes joining the runtime grammar to the main grammar at the location of the placeholder.
3. The method of claim 1 in which the runtime grammar is joined to the main grammar at the location of the placeholder, prior to performing context-dependent speech recognition.
4. A method for constructing a speech recognition process, the method comprising: designating a main grammar containing a placeholder for a rantime grammar; and processing the main grammar to construct a speech recognition process that uses context-dependent speech recognition, the speech recognition process prepared to receive a rantime grammar for later use in place of the placeholder.
5. The method of claim 4, wherein the processing preserves context.
6. The method of claim 5, wherein for a first word and a second word adjacent in the main grammar, the processing preserves a cross-word context between the first word and the second word.
7. The method of claim 6, wherein the first word or the second word or both may be placeholders for a runtime grammar.
8. A method for speech recognition comprising: designating a main grammar containing a placeholder for a runtime grammar, wherein the runtime grammar includes names; providing the runtime grammar; and performing context-dependent speech recognition based on the main grammar containing the placeholder and on the runtime grammar.
9. A method for speech recognition comprising: designating a main grammar containing a placeholder for a runtime grammar; presenting a context-dependent speech recognition process to a speaker; selecting the runtime grammar from among a plurality of runtime grammars, the selection based on a characteristic of the speaker; and using the runtime grammar in place of the placeholder in the context- dependent speech recognition process, preserving cross-word context.
10. The method of claim 9, wherein the characteristic of the speaker is the speaker's identity.
11. A method for speech recognition comprising: designating a main grammar containing a placeholder for a runtime grammar, the main grammar and the runtime grammar encoded as finite state machines; providing linguistic models encoded as finite state machines, the linguistic models including word pronunciations for words in the main grammar, context- dependent models, phonological rewrite rules, and generalized word pronunciation algorithms; producing a main speech recognition machine by composing the main grammar with the linguistic models, the composing based on finite state machine composition and preserving cross-word context, such that the main speech recognition machine accepts context-dependent models as input; processing the runtime grammar, optionally before the completion of the step of producing the main speech recognition machine; and performing context-dependent speech recognition based on the runtime grammar and the main speech recognition machine.
12. The method of claim 11, wherein processing the runtime grammar includes composing the runtime grammar with the linguistic models to produce a runtime speech recognition machine.
13. The method of claim 12, wherein processing the runtime grammar further includes creating paths from the main speech recognition machine into the runtime speech recognition machine and back again.
14. The method of claim 13, wherein the paths are appropriate to linguistic contexts encoded in the main speech recognition machine and in the runtime speech recognition machine.
15. A method for constructing a speech recognizer, comprising: providing at design-time a main grammar containing one or more placeholders for a runtime grammar, the main grammar also containing zero or more words; providing a plurality of sub-word units; and expanding a word of the main grammar into a conesponding network of sub- word units selected from the plurality of sub-word units, each sub-word unit having a permissible context including constraints on neighboring sub-word units within the conesponding network, the conesponding network being chosen to satisfy the constraints of the permissible context of each sub-word unit within the conesponding network, such that when the sub-word units conesponding to the word to be expanded would have permissible context applicable to words provided by a runtime grammar, the expansion of the word of the main grammar into the conesponding network includes every sub-word unit from the plurality of sub-word units subject to the constraints of the sub-word unit's permissible context as compared to the conesponding network.
16. The method of claim 15, further comprising: expanding a word of the runtime grammar into a conesponding network of sub-word units selected from the plurality of sub-word units, each sub-word unit having a permissible context including constraints on neighboring sub-word units within the conesponding network, the conesponding network being chosen to satisfy the constraints of the permissible context of each sub-word unit within the conesponding network, the expanding being such that, when the sub-word units conesponding to the word to be expanded would have permissible context applicable to words in the main grammar, the expansion of the word into the conesponding network includes every sub-word unit from the plurality of sub- word units subject to the constraints of the sub-word unit's permissible context as compared to the conesponding network.
17. A method for speech recognition, comprising: providing at design-time a main grammar containing a placeholder for a runtime grammar, the main grammar also containing zero or more words; providing a plurality of sub-word units; expanding a word of the main grammar into a conesponding network of sub- word units selected from the plurality of sub-word units, the expanding based in part on a pronunciation of the word, each sub-word unit having a permissible context including constraints on neighboring sub-word units within the conesponding network, the conesponding network being chosen to satisfy the constraints of the permissible context of each sub-word unit within the conesponding network, the expanding being such that, when the sub-word units conesponding to the word to be expanded would have permissible context applicable to words provided by a runtime grammar, the expansion of the word of the main grammar into the conesponding network includes every sub-word unit from the plurality of sub-word units that satisfies the constraints of the sub-word unit's permissible context, when compared to the conesponding network; specifying after design-time the runtime grammar to associate with the placeholder; expanding a word of the runtime grammar into a conesponding network of sub-word units selected from the plurality of sub-word units, the expanding based in part on a pronunciation of the word, each sub-word unit having a permissible context including constraints on neighboring sub-word units within the conesponding network, the conesponding network being chosen to satisfy the constraints of the permissible context of each sub-word unit within the conesponding network, such that when the sub-word units conesponding to the word to be expanded would have permissible context applicable to words in the main grammar, the expansion of the word into the conesponding network includes every sub-word unit from the plurality of sub-word units subject to the constraints of the sub-word unit's permissible context as compared to the conesponding network; joining the runtime grammar to the main grammar by creating paths from the main grammar into runtime grammar and back again, the paths selected to satisfy the constraints of the permissible context of each sub-word unit in the expansion of the main grammar and in the expansion of the runtime grammar; accepting audio input representing speech from a speaker; analyzing the audio input to produce a string of sub-word units; and using the expansion of the main grammar j oined with the expansion of the runtime grammar to recognize the string of sub-word units.
18. A computer program comprising code means adapted to perform all the steps of any of claims 1 to 17 when said program is executed on a data processing system.
19. The computer program of claim 18 embodied on a computer-readable medium.
PCT/US2002/021364 2001-07-05 2002-07-03 Speech recognition with dynamic grammars WO2003005345A1 (en)

Applications Claiming Priority (3)

Application Number Priority Date Filing Date Title
US60/303,049 2001-07-05
US09/906,390 US20030009335A1 (en) 2001-07-05 2001-07-16 Speech recognition with dynamic grammars
US09/906,390 2001-07-16

Publications (1)

Publication Number Publication Date
WO2003005345A1 true WO2003005345A1 (en) 2003-01-16

Family

ID=25422361

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/US2002/021364 WO2003005345A1 (en) 2001-07-05 2002-07-03 Speech recognition with dynamic grammars

Country Status (1)

Country Link
WO (1) WO2003005345A1 (en)

Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US5581655A (en) * 1991-01-31 1996-12-03 Sri International Method for recognizing speech using linguistically-motivated hidden Markov models
WO2001026092A2 (en) * 1999-10-06 2001-04-12 Lernout & Hauspie Speech Products N.V. Attribute-based word modeling
WO2001091107A1 (en) * 2000-05-23 2001-11-29 Thomson Licensing S.A. Dynamic language models for speech recognition

Patent Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US5581655A (en) * 1991-01-31 1996-12-03 Sri International Method for recognizing speech using linguistically-motivated hidden Markov models
WO2001026092A2 (en) * 1999-10-06 2001-04-12 Lernout & Hauspie Speech Products N.V. Attribute-based word modeling
WO2001091107A1 (en) * 2000-05-23 2001-11-29 Thomson Licensing S.A. Dynamic language models for speech recognition

Non-Patent Citations (5)

* Cited by examiner, † Cited by third party
Title
BOULIANNE G ET AL: "French large vocabulary recognition with cross-word phonology transducers", 2000 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH, AND SIGNAL PROCESSING. PROCEEDINGS (CAT. NO.00CH37100), vol. 3, 5 June 2000 (2000-06-05) - 9 June 2000 (2000-06-09), ISTANBUL, TURKEY, Piscataway, NJ, USA, IEEE, USA, pages 1675 - 1678 vol.3, XP002215625, ISBN: 0-7803-6293-4 *
DEMUYNCK K ET AL: "An efficient search space representation for large vocabulary continuous speech recognition", SPEECH COMMUNICATION, AMSTERDAM, NL, vol. 30, no. 1, January 2000 (2000-01-01), pages 37 - 53, XP004363634, ISSN: 0167-6393 *
MOHRI M ET AL: "Full expansion of context-dependent networks in large vocabulary speech recognition", ACOUSTICS, SPEECH AND SIGNAL PROCESSING, 1998. PROCEEDINGS OF THE 1998 IEEE INTERNATIONAL CONFERENCE ON SEATTLE, WA, USA 12-15 MAY 1998, NEW YORK, NY, USA,IEEE, US, 12 May 1998 (1998-05-12), pages 665 - 668, XP010279334, ISBN: 0-7803-4428-6 *
NUANCE: "Nuance Speech Recognition System, Version 8.0, Grammar Developer's Guide", 2001, NUANCE COMMUNICATION INC., 1005 HAMILTON AVENUE, MENLO PARK, CALIFORNIA 94025 USA, XP001117344 *
WANG YE-YI ET AL: "Unified context-free grammar and n-gram model for spoken language processing", 2000 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH, AND SIGNAL PROCESSING. PROCEEDINGS. (ICASSP)., vol. 3 OF 6, 5 June 2000 (2000-06-05) - 9 June 2000 (2000-06-09), ISTANBUL, TURKEY, NEW YORK, NY: IEEE, US, pages 1639 - 1642, XP002181416, ISBN: 0-7803-6294-2 *

Similar Documents

Publication Publication Date Title
US20030009335A1 (en) Speech recognition with dynamic grammars
US7149688B2 (en) Multi-lingual speech recognition with cross-language context modeling
Young et al. The HTK book
US7072837B2 (en) Method for processing initially recognized speech in a speech recognition session
JP3741156B2 (en) Speech recognition apparatus, speech recognition method, and speech translation apparatus
KR100486733B1 (en) Method and apparatus for speech recognition using phone connection information
Lee Voice dictation of mandarin chinese
JP4215418B2 (en) Word prediction method, speech recognition method, speech recognition apparatus and program using the method
JPH0855122A (en) Context tagger
JP6453631B2 (en) Recognition system, recognition method and program
KR100735559B1 (en) Apparatus and method for constructing language model
KR102094935B1 (en) System and method for recognizing speech
JPWO2009081895A1 (en) Speech recognition system, speech recognition method, and speech recognition program
KR101424193B1 (en) System And Method of Pronunciation Variation Modeling Based on Indirect data-driven method for Foreign Speech Recognition
Hasegawa-Johnson et al. Grapheme-to-phoneme transduction for cross-language ASR
KR100726875B1 (en) Speech recognition with a complementary language model for typical mistakes in spoken dialogue
Buchsbaum et al. Algorithmic aspects in speech recognition: An introduction
Rojc et al. Time and space-efficient architecture for a corpus-based text-to-speech synthesis system
Szarvas et al. Finite-state transducer based modeling of morphosyntax with applications to Hungarian LVCSR
JP4649207B2 (en) A method of natural language recognition based on generated phrase structure grammar
JP4733436B2 (en) Word / semantic expression group database creation method, speech understanding method, word / semantic expression group database creation device, speech understanding device, program, and storage medium
Lecorvé et al. Adaptive statistical utterance phonetization for French
Mohri Weighted grammar tools: the GRM library
WO2003005345A1 (en) Speech recognition with dynamic grammars
JP2000222406A (en) Voice recognition and translation device and its method

Legal Events

Date Code Title Description
AK Designated states

Kind code of ref document: A1

Designated state(s): AE AG AL AM AT AU AZ BA BB BG BR BY BZ CA CH CN CO CR CU CZ DE DK DM DZ EC EE ES FI GB GD GE GH GM HR HU ID IL IN IS JP KE KG KP KR KZ LC LK LR LS LT LU LV MA MD MG MK MN MW MX MZ NO NZ PL PT RO RU SD SE SG SI SK SL TJ TM TR TT TZ UA UG US UZ VN YU ZA ZW

AL Designated countries for regional patents

Kind code of ref document: A1

Designated state(s): GH GM KE LS MW MZ SD SL SZ TZ UG ZM ZW AM AZ BY KG KZ MD RU TJ TM AT BE BG CH CY CZ DE DK EE ES FI FR GB GR IE IT LU MC NL PT SE SK TR BF BJ CF CG CI CM GA GN GQ GW ML MR NE SN TD TG

121 Ep: the epo has been informed by wipo that ep was designated in this application
DFPE Request for preliminary examination filed prior to expiration of 19th month from priority date (pct application filed before 20040101)
REG Reference to national code

Ref country code: DE

Ref legal event code: 8642

122 Ep: pct application non-entry in european phase
NENP Non-entry into the national phase

Ref country code: JP

WWW Wipo information: withdrawn in national office

Country of ref document: JP