US20130158998A1 - Systems and Methods for Extracting Meaning from Multimodal Inputs Using Finite-State Devices - Google Patents
Systems and Methods for Extracting Meaning from Multimodal Inputs Using Finite-State Devices Download PDFInfo
- Publication number
- US20130158998A1 US20130158998A1 US13/690,037 US201213690037A US2013158998A1 US 20130158998 A1 US20130158998 A1 US 20130158998A1 US 201213690037 A US201213690037 A US 201213690037A US 2013158998 A1 US2013158998 A1 US 2013158998A1
- Authority
- US
- United States
- Prior art keywords
- gesture
- finite
- multimodal
- speech
- lattice
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Granted
Links
- 238000000034 method Methods 0.000 title claims abstract description 80
- 230000008569 process Effects 0.000 abstract description 14
- 238000013459 approach Methods 0.000 description 14
- 238000012545 processing Methods 0.000 description 12
- 239000000284 extract Substances 0.000 description 10
- 230000008520 organization Effects 0.000 description 9
- 230000010354 integration Effects 0.000 description 7
- 230000002123 temporal effect Effects 0.000 description 7
- 239000013598 vector Substances 0.000 description 7
- 230000033001 locomotion Effects 0.000 description 6
- 238000010586 diagram Methods 0.000 description 5
- 239000012634 fragment Substances 0.000 description 5
- 239000000203 mixture Substances 0.000 description 5
- 230000006870 function Effects 0.000 description 3
- 230000007704 transition Effects 0.000 description 3
- 238000013528 artificial neural network Methods 0.000 description 2
- 238000009826 distribution Methods 0.000 description 2
- 238000000605 extraction Methods 0.000 description 2
- 230000003993 interaction Effects 0.000 description 2
- KOOAFHGJVIVFMZ-WZPXRXMFSA-M micafungin sodium Chemical compound [Na+].C1=CC(OCCCCC)=CC=C1C1=CC(C=2C=CC(=CC=2)C(=O)N[C@@H]2C(N[C@H](C(=O)N3C[C@H](O)C[C@H]3C(=O)N[C@H](C(=O)N[C@H](C(=O)N3C[C@H](C)[C@H](O)[C@H]3C(=O)N[C@H](O)[C@H](O)C2)[C@H](O)CC(N)=O)[C@H](O)[C@@H](O)C=2C=C(OS([O-])(=O)=O)C(O)=CC=2)[C@@H](C)O)=O)=NO1 KOOAFHGJVIVFMZ-WZPXRXMFSA-M 0.000 description 2
- 238000012546 transfer Methods 0.000 description 2
- 230000009471 action Effects 0.000 description 1
- 230000004888 barrier function Effects 0.000 description 1
- 230000008901 benefit Effects 0.000 description 1
- 230000001413 cellular effect Effects 0.000 description 1
- 238000007635 classification algorithm Methods 0.000 description 1
- 238000004891 communication Methods 0.000 description 1
- 238000007796 conventional method Methods 0.000 description 1
- 230000008878 coupling Effects 0.000 description 1
- 238000010168 coupling process Methods 0.000 description 1
- 238000005859 coupling reaction Methods 0.000 description 1
- 230000001419 dependent effect Effects 0.000 description 1
- 238000013461 design Methods 0.000 description 1
- 238000011143 downstream manufacturing Methods 0.000 description 1
- 230000000694 effects Effects 0.000 description 1
- 238000011156 evaluation Methods 0.000 description 1
- 230000008921 facial expression Effects 0.000 description 1
- 238000004519 manufacturing process Methods 0.000 description 1
- 230000005012 migration Effects 0.000 description 1
- 238000013508 migration Methods 0.000 description 1
- 238000012986 modification Methods 0.000 description 1
- 230000004048 modification Effects 0.000 description 1
- 238000003058 natural language processing Methods 0.000 description 1
- 230000037361 pathway Effects 0.000 description 1
- 230000002093 peripheral effect Effects 0.000 description 1
- 238000012805 post-processing Methods 0.000 description 1
- 230000002195 synergetic effect Effects 0.000 description 1
- 230000001755 vocal effect Effects 0.000 description 1
Images
Classifications
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
- G10L15/00—Speech recognition
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
- G10L15/00—Speech recognition
- G10L15/24—Speech recognition using non-acoustical features
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/16—Sound input; Sound output
- G06F3/167—Audio in a user interface, e.g. using voice commands for navigating, audio feedback
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V40/00—Recognition of biometric, human-related or animal-related patterns in image or video data
- G06V40/20—Movements or behaviour, e.g. gesture recognition
- G06V40/28—Recognition of hand or arm movements, e.g. recognition of deaf sign language
Definitions
- This invention is directed to parsing and understanding of utterances whose content is distributed across multiple input modes.
- Multimodal interfaces allow input and/or output to be conveyed over multiple different channels, such as speech, graphics, gesture and the like. Multimodal interfaces enable more natural and effective interaction, because particular modes are best-suited for particular kinds of content. Multimodal interfaces are likely to play a critical role in the ongoing migration of interaction from desktop computing to wireless portable computing devices, such as personal digital assistants, like the Palm Pilot®, digital cellular telephones, public information kiosks that are wirelessly connected to the Internet or other distributed networks, and the like. One barrier to adopting such wireless portable computing devices is that they offer limited screen real estate, and often have limited keyboard interfaces, if any keyboard interface at all.
- multimodal interfaces need to support not just input from multiple modes. Rather, multimodal interfaces also need to support synergistic multimodal utterances that are optimally distributed over the various available modes. In order to achieve this, the content from different modes needs to be effectively integrated.
- Johnston 1 disclosed a pen-based device that allows a variety of gesture utterances to be input through a gesture mode, while a variety of speech utterances can be input through a speech mode.
- Unification operations determine the consistency of two pieces of partial information. If the two pieces of partial information are determined to be consistent, the unification operation combines the two pieces of partial information into a single result. Unification operations were used to determine whether a given piece of gestural input received over the gesture mode was compatible with a given piece of spoken input received over the speech mode. If the gestural input was determined to be compatible with the spoken input, the two inputs were combined into a single result that could be further interpreted.
- Johnston 1 typed feature structures were used as a common meaning representation for both the gestural inputs and the spoken inputs.
- the multimodal integration was modeled as a cross-product unification of feature structures assigned to the speech and gestural inputs. While the technique disclosed in Johnston 1 overcomes many of the limitations of earlier multimodal systems, this technique does not scale well to support multi-gesture utterances, complex unimodal gestures, or other modes and combinations of modes. To address these limitations, the unification-based multimodal integration technique disclosed in Johnston 1 was extended in “Unification-Based Multimodal Parsing”, M. Johnston, Proceedings of COLING - ACL 98, p. 624-630, 1998 (Johnston 2), herein incorporated by reference in its entirety.
- the multimodal integration technique disclosed in Johnston 2 uses a multi-dimensional chart parser.
- elements of the multimodal input are treated as terminal edges by the parser.
- the multimodal input elements are combined together in accordance with a unification-based multimodal grammar.
- the unification-based multimodal parsing technique disclosed in Johnston 2 was further extended in “Multimodal Language Processing”, M. Johnston, Proceedings of ICSLP 1998, 1998 (published on CD-ROM only) (Johnston 3), incorporated herein by reference in its entirety.
- Johnston 2 and 3 disclosed how techniques from natural language processing can be adapted to support parsing and interpretation of utterances distributed over multiple modes.
- speech and gesture recognition produce n-best lists of recognition results.
- the n-best recognition results are assigned typed feature structure representations by speech interpretation and gesture interpretation components.
- the n-best lists of feature structures from the spoken inputs and the gestural inputs are passed to a multi-dimensional chart parser that uses a multimodal unification-based grammar to combine the representations assigned to the input elements. Possible multimodal interpretations are then ranked. The optimal interpretation is then passed on for execution.
- the unification-based approach disclosed in Johnston 1-Johnston 3 does not allow for tight coupling of multimodal parsing with speech and gesture recognition. Compensation effects are dependent on the correct answer appearing in each of the n-best list of interpretations obtained from the recognitions obtained from the inputs of each mode. Moreover, multimodal parsing cannot directly influence the progress of either speech recognition or gesture recognition.
- the multi-dimensional parsing approach is also subject to significant concerns in terms of computational complexity. In the worst case, for the multi-dimensional parsing technique disclosed in Johnston 2, the number of parses to be considered is exponential relative to the number of input elements and the number of interpretations the input elements have. This complexity is manageable when the inputs yield only n-best results for small n. However, the complexity quickly gets out of hand if the inputs are sizable lattices with associated probabilities.
- the unification-based approach also runs into significant problems when choosing between multiple competing parses and interpretations. Probabilities associated with composing speech events and multiple gestures need to be combined. Uni-modal interpretations need to be compared to multimodal interpretations and so on. While this can all be achieved using the unification-based approach disclosed in Johnston 1-Johnston 3, significant post-processing of sets of competing multimodal interpretations generated by the multimodal parser will be involved.
- This invention provides systems and methods that allow parsing understanding and/or integration of the gestural inputs and the spoken inputs using one or more finite-state devices.
- This invention separately provides systems and methods that allow multi-dimensional parsing and understanding using weighted finite-state automata.
- This invention further provides systems and methods that allow multi-dimensional parsing and understanding using a three-tape weighted finite-state automaton.
- This invention separately provides systems and methods that use combinations of finite-state transducers to integrate the various modes of the multimodal interface.
- This invention separately provides systems and methods that use the recognition results of one mode of the multimodal input received from the multimodal interface as a language model or other model in the recognition process of other modes of the multimodal inputs received from the multimodal interface.
- This invention separately provides systems and methods that use the recognition results of one mode of the multimodal input received from the multimodal interface to constrain the recognition process of one or more of the other modes of the multimodal input received from the multimodal interface.
- This invention further provides systems and methods that integrate the recognition results from the second multimodal input, which are based on the recognition results of the first multimodal input, with the recognition results of the first multimodal input and then extract meaning from the combined recognition results.
- This invention further provides systems and methods that base the speech recognition on the results of the gesture recognition.
- the various exemplary embodiments of the systems and methods according to this invention allow spoken language and gesture input streams to be parsed and integrated by a single weighted finite-state device.
- This single weighted finite-state device provides language models for speech and gesture recognition and composes the meaning content from the speech and gesture input streams into a single semantic representation.
- the systems and methods according to this invention not only address multimodal language recognition, but also encode the semantics as well as the syntax into a single weighted finite-state device.
- the systems and methods according to this invention provide the potential for direct compensation among the various multimodal input modes.
- Various exemplary embodiments of the systems and methods according to this invention allow the gestural input to dynamically alter the language model used for speech recognition.
- Various exemplary embodiments of the systems and methods according to this invention reduce the computational complexity of multi-dimensional multimodal parsing.
- the weighted finite-state devices used in various exemplary embodiments of the systems and methods according to this invention provide a well-understood probabilistic framework for combining the probability distributions associated with the speech and gesture or other input modes and for selecting among multiple competing multimodal interpretations.
- FIG. 1 is a block diagram illustrating one exemplary embodiment of a conventional automatic speech recognition system usable with a multimodal meaning recognition system according to this invention
- FIG. 2 is a block diagram illustrating one exemplary embodiment of a multimodal user input device and one exemplary embodiment of a multimodal meaning recognition system according to this invention
- FIG. 3 is a block diagram illustrating in greater detail one exemplary embodiment of the gesture recognition system of FIG. 2 ;
- FIG. 4 is a block diagram illustrating in greater detail one exemplary embodiment of the multimodal parser and meaning recognition system of FIG. 2 ;
- FIG. 5 is a block diagram illustrating in greater detail one exemplary embodiment of the multimodal user input device of FIG. 2 ;
- FIG. 6 is one exemplary embodiment of a multimodal grammar fragment usable by the multimodal meaning recognition system according to this invention.
- FIG. 7 is one exemplary embodiment of a three-tape multimodal finite-state automaton usable to recognize the multimodal inputs received from the exemplary embodiment of the multimodal user input device shown in FIG. 5 ;
- FIG. 8 is one exemplary embodiment of a gesture finite-state machine generated by recognizing the gesture inputs shown in the exemplary embodiment of the multimodal user input device shown in FIG. 5 ;
- FIG. 9 is one exemplary embodiment of a gesture-to-speech finite-state transducer that represents the relationship between speech and gesture for the exemplary embodiment of the multimodal user input device shown in FIG. 5 ;
- FIG. 10 is one exemplary embodiment of a speech/gesture/meaning finite-state transducer that represents the relationship between the combined speech and gesture symbols and the semantic meaning of the multimodal input for the exemplary embodiment of the multimodal input device shown in FIG. 5 ;
- FIG. 11 is a flowchart outlining one exemplary embodiment of a method for extracting meaning from a plurality of multimodal inputs
- FIG. 12 is one exemplary embodiment of a gesture/language finite-state transducer illustrating the composition of the gesture finite-state machine shown in FIG. 8 with the gesture-to-speech finite-state transducer shown in FIG. 9 ;
- FIG. 13 is one exemplary embodiment of a finite-state machine generated by taking a projection on the output tape of the gesture/language finite-state transducer shown in FIG. 12 ;
- FIG. 14 is one exemplary embodiment of a lattice of possible word sequences generated by the automatic speech recognition system shown in FIG. 1 when using the finite-state machine shown in FIG. 13 as a language model in view of the speech input received from the exemplary embodiment of the multimodal user input device shown in FIG. 5 ;
- FIG. 15 illustrates one exemplary embodiment of a gesture/speech finite-state transducer generated by composing the gesture/language finite-state transducer shown in FIG. 12 with the word sequence lattice shown in FIG. 14 ;
- FIG. 16 is one exemplary embodiment of a gesture/speech finite-state machine obtained from the gesture/speech finite-state transducer shown in FIG. 14 ;
- FIG. 17 is one exemplary embodiment of a finite-state transducer, obtained from composing the gesture/speech finite-state machine shown in FIG. 16 with the speech/gesture/meaning finite-state transducer shown in FIG. 10 , which extracts the meaning from the multimodal gestural and spoken inputs received when using the exemplary embodiment of the multimodal user input device shown in FIG. 5 .
- FIG. 1 illustrates one exemplary embodiment of an automatic speech recognition system 100 usable with the multimodal recognition and/or meaning system 1000 according to this invention that is shown in FIG. 2 .
- automatic speech recognition can be viewed as a processing pipeline or cascade.
- lattice In each step of the processing cascade, one or two lattices are input and composed to produce an output lattice.
- lattice denotes a directed and labeled graph, which is possibly weighted.
- start node In each lattice, there is typically a designated start node “s” and a designated final node “t”. Each possible pathway through the lattice from the start node s to the final node t induces a hypothesis based on the arc labels between each pair of nodes in the path.
- the arc labels are words and the various paths between the start node s and the final node t form sentences.
- the weights on the arcs on each path between the start node s and the final node t are combined to represent the likelihood that that path will represent a particular portion of the utterance.
- one exemplary embodiment of a known automatic speech recognition system 100 includes a signal processing subsystem 110 , an acoustic model lattice 120 , a phonetic recognition subsystem 130 , a lexicon lattice 140 , a word recognition subsystem 150 , a grammar or language model lattice 160 , and a task recognition subsystem 170 .
- uttered speech is input via a microphone, which converts the sound waves of the uttered speech into an electronic speech signal.
- the electronic speech signal is input to the signal processing subsystem 110 on a speech signal input line 105 .
- the signal processing subsystem 110 digitizes the electronic speech signal to generate a feature vector lattice 115 .
- the feature vector lattice 115 is a lattice of acoustic feature vectors.
- the feature vector lattice 115 is input along with the acoustic model lattice 120 to the phonetic recognition subsystem 130 .
- the acoustic model lattice 120 represents a set of acoustic models and is applied to transform the feature vector lattice 115 into a phone lattice.
- Each node of the phone lattice represents a spoken sound, such as, for example, the vowel /e/ in “bed”.
- the phone lattice 135 is input along with the lexicon lattice 140 into the word recognition subsystem 150 .
- the lexicon lattice 140 describes different pronunciations of various words and transforms the phone lattice 135 into a word lattice 155 .
- the word lattice 155 is then input, along with the grammar or language model lattice 160 , into the utterance recognition subsystem 170 .
- the grammar or language model lattice 160 represents task-specific information and is used to extract the most likely sequence of uttered words from the word lattice 155 .
- the utterance recognition subsystem 170 uses the grammar or language model lattice 160 to extract the most likely sentence or other type of utterance from the word lattice 155 .
- the grammar or language model lattice 160 will be selected based on the task associated with the uttered speech.
- the most likely sequence of words, or the lattice of n most-likely sequences of words, is output as the recognized utterance 175 .
- one conventional method of implementing automatic speech recognition forms each of the acoustic model lattice 120 , the lexicon lattice 140 and the grammar or language model lattice 160 as a finite-state transducer.
- each of the phonetic recognition subsystem 130 , the word recognition subsystem 150 , and the utterance recognition 170 performs a generalized composition operation between its input finite-state transducers.
- the signal processing subsystem 110 outputs the features vector lattice 115 as a finite-state transducer.
- the grammar or language model lattice 160 is predetermined and incorporated into the automatic speech recognition system 100 based on the particular recognition task that the automatic speech recognition system 100 is to perform.
- any of the acoustic model lattice 120 , the lexicon lattice 140 and/or the grammar or language model 160 can be non-deterministic finite-state transducers.
- these non-deterministic finite-state transducers can be determinized using the various techniques disclosed in “Finite-state transducers in Language and Speech Processing”, M. Mohri, Computational Linguistics, 23:2, p. 269-312, 1997, U.S. patent application Ser. No. 09/165,423, filed Oct. 2, 1998, and/or U.S. Pat. No. 6,073,098 to Buchsbaum et al., each incorporated herein by reference in its entirety.
- the automatic speech recognition system 100 uses a grammar or language model lattice 160 that is obtained from the recognized gestural input received in parallel with the speech signal 105 . This is shown in greater detail in FIG. 2 . In this way, the output of the gesture recognition system 200 can be used to compensate for uncertainties in the automatic speech recognition system.
- the output of the automatic speech recognition system 100 and output of the gesture recognition system 200 can be combined only after each output is independently obtained. In this way, it becomes possible to extract meaning from the composition of two or more different input modes, such as the two different input modes of speech and gesture.
- the output of the gesture recognition system 200 can be used to provide compensation to the automatic speech recognition system 100 . Additionally, their combined output can be further processed to extract meaning from the combination of the two different input modes. In general, when there are two or more different input modes, any of one or more of the input modes can be used to provide compensation to one or more other ones of the input modes.
- any two or more input modes that can provide compensation between the modes, which can be combined to allow meaning to be extracted from the two or more recognized outputs, or both, can be used in place of, or in addition to, the speech and gesture input modes discussed herein.
- a multimodal user input device 400 when speech and gesture are the implemented input modes, includes a gesture input portion 410 and a speech input portion 420 .
- the gesture input portion 410 outputs a gesture signal 205 to a gesture recognition system 200 of the multimodal recognition and/or meaning system 1000 .
- the speech input portion 420 outputs the speech signal 105 to the automatic speech recognition system 100 .
- the gesture recognition system 200 generates a gesture recognition lattice 255 based on the input gesture signal 205 and outputs the gesture recognition lattice 255 to a multimodal parser and meaning recognition system 300 of the multimodal recognition and/or meaning system 1000 .
- the multimodal parser/meaning recognition system 300 generates a new grammar or language model lattice 160 for the utterance recognition subsystem 170 of the automatic speech recognition system 100 from the gesture recognition lattice 255 .
- this new grammar or language model lattice 160 generated by the multimodal parser/meaning recognition system 300 is specific to the particular sets of gestural inputs generated by a user through the gesture input portion 410 of the multimodal user input device 400 .
- this new grammar or language model lattice 160 represents all of the possible spoken strings that can successfully combine with the particular sequence of gestures input by the user through the gesture input portion 410 . That is, the recognition performed by the automatic speech recognition system 100 can be improved because the particular grammar or language model lattice 160 being used to recognize that spoken utterance is highly specific to the particular sequence of gestures made by the user.
- the automatic speech recognition system 100 then outputs the recognized possible word sequence lattice 175 back to the multimodal parser/meaning recognition system 300 .
- the recognized possible word sequences lattice 175 is then output to a downstream processing task.
- the multimodal recognition and/or meaning system 1000 then waits for the next set of inputs from the multimodal user input device 400 .
- the multimodal parser/meaning recognition system 300 extracts meaning from the combination of the gesture recognition lattice 255 and the recognized possible word sequences lattice 175 . Because the spoken utterances input by the user through the speech input portion 420 are presumably closely related to the gestures input at the same time by the user through the gesture input portion 410 , the meaning of those gestures can be tightly integrated with the meaning of the spoken input generated by the user through the speech input portion 420 .
- the multimodal parser/meaning recognition system 300 outputs a recognized possible meaning lattice 375 in addition to, or in place of, one or both of the gesture recognition lattice 255 and/or the recognized possible word sequences lattice 175 .
- the multimodal parser and meaning recognition system 300 combines the recognized lattice of possible word sequences 175 generated by the automatic speech recognition system 100 with the gesture recognition lattice 255 output by the gesture recognition system 200 to generate the lattice of possible meaning sequences 375 corresponding to the multimodal gesture and speech inputs received from the user through the multimodal user input device 400 .
- the multimodal parser/meaning recognition system 300 does not generate the new grammar or language model lattice 160 .
- the gesture recognition lattice 255 does not provide compensation to the automatic speech recognition system 100 .
- the multimodal parser/meaning recognition system 300 only combines the gesture recognition lattice 255 and the recognized possible word sequences lattice 175 to generate the recognition meaning lattice 375 .
- the gesture recognition system 200 When the gesture recognition system 200 generates only a single recognized possible sequence of gestures as the gesture recognition lattice 255 , that means there is essentially no uncertainty in the gesture recognition. In this case, the gesture recognition lattice 255 provides compensation to the automatic speech recognition system 100 for any uncertainty in the speech recognition process. However, the gesture recognition system 200 can generate a lattice of n possible recognized gesture sequences as the gesture recognition lattice 255 . This recognizes that there may also be uncertainty in the gesture recognition process.
- the gesture recognition lattice 255 and the word lattice 155 provide mutual compensation for the uncertainties in both the speech recognition process and the gesture recognition process. That is, in the face of this uncertainty, the best, i.e., most-probable, combination of one of the n-best word sequences in the word lattice 155 with one of the n-best gesture sequences in the gesture recognition lattice may not include the best recognition possible sequence from either the word lattice 155 or the gesture recognition lattice 255 .
- the most-probable sequence of gestures in the gesture recognition lattice may combine only with a rather low-probability word sequence through the word lattice, while the most-probable word sequence may combine well only with a rather low-probability gesture sequence.
- a medium-probability word sequence may match very well with a medium-probability gesture sequence.
- the net probability of this latter combination of word and gesture sequences may be higher than the probability of the combination of the best word sequence with any of the gesture sequences through the gesture recognition lattice 255 and may be higher than the probability of the combination of the best gesture sequence with any of the word sequences through the lattice of possible word sequences 155 . In this way, mutual compensation is provided between the gesture recognition system 200 and the automatic speech recognition system 100 .
- FIGS. 3-5 illustrate in greater detail various exemplary embodiments of the gesture recognition system 200 , the multimodal parser/meaning recognition system 300 , and the multimodal user input device 400 .
- one exemplary embodiment of the gesture recognition system 200 includes a gesture feature extraction subsystem 210 and a gesture recognition subsystem 230 .
- Various other exemplary embodiments may include a gesture language model lattice and a gesture meaning subsystem.
- gesture utterances are input through the gesture input portion 410 of the multimodal user input device 400 , which converts the movements of an input device, such as a mouse, a pen, a trackball, a track pad or any other known or later-developed gestural input device, into an electronic gesture signal 205 .
- the multimodal user input device 400 converts the gestural input into digital ink that can be viewed and understood by the user. This is shown in greater detail in FIG. 5 .
- the gesture feature extraction subsystem 210 converts the motions of the gesture input device represented by the gesture signal 205 into a gesture feature lattice 220 .
- the various gestures that can be made can be as simple as pointing gestures to a particular information element at a particular location within the gesture input portion 410 of the multimodal user input device 400 , or can be as complex as a specialized symbol that represents a type of military unit on a military map displayed in the gesture input portion 410 of the multimodal user input portion 400 and includes an indication of how the unit is to move, and which unit is to move and how far that unit is to move, as described in detail in Johnston 1.
- the gesture feature lattice 220 is input to the gesture recognition subsystem 230 .
- the gesture recognition subsystem 230 may be implemented as a neural network, as a Hidden-Markov Model (HMM) or as a simpler template-based gesture classification algorithm.
- the gesture recognition subsystem 230 converts the gesture feature lattice 220 into the gesture recognition lattice 255 .
- the gesture recognition lattice 255 includes the identities of graphical elements against which diectic and other simple “identification” gestures are made, possible recognition of more complex gestures that the user may have made and possibly the locations on the displayed graphics where the more complex gesture was made, such as in Johnston 1, and the like.
- the gesture recognition system 200 outputs the gesture recognition lattice 255 to the multimodal parser/meaning recognition system 300 .
- gesture feature recognition subsystem 210 and the gesture recognition subsystem 230 can each be implemented using any known or later-developed system, circuit or technique that is appropriate.
- the entire gesture recognition system 200 can be implemented using any known or later-developed system that generates a directed graph from a gesture input.
- one known system captures the time and location or locations of the gesture.
- these inputs are then normalized and/or rotated.
- the gestures are then provided to a pattern classification device that is implemented as part of the gesture feature recognition subsystem 210 .
- this pattern classification device is a template matching system, which transforms the gesture into a feature vector.
- this pattern classification device is a neural network or a Hidden Markov Model that has been trained to recognize certain patterns of one or more temporally and/or spatially related gesture components as a specific set of features.
- gesture recognition subsystem 230 When a single gesture is formed by two or more temporally and/or spatially related gesture components, those gesture components can be combined into a single gesture either during the recognition process or by the multimodal parser/meaning recognition system 300 .
- the gesture recognition subsystem 230 combines the temporally adjacent gestures into a lattice of one or more recognized possible gesture sequences that represent how the recognized gestures follow each other in time.
- the multimodal parser and meaning recognition system 300 can be implemented using a single three-tape finite-state device that inputs the output lattices from the speech recognition system 100 and the gesture recognition system 200 and directly obtains and outputs a meaning result.
- the three-tape finite-state device is a three-tape grammar model that relates the gestures and the words to a meaning of the combination of a gesture and a word.
- FIG. 7 shows a portion of such a three-tape grammar model usable in the multimodal parser and meaning recognition system 300 to generate a meaning output from gesture and speech recognition inputs.
- the multimodal parser and meaning recognition system 300 can be implemented using an n-tape finite-state device that inputs n ⁇ 1 lattices from a plurality of recognition systems usable to recognize an utterance having a plurality of different modes.
- FIG. 4 shows the multimodal parser/meaning recognition system 300 in greater detail.
- the multimodal parser/meaning recognition system 300 may include one or more of a gesture-to-speech composing subsystem 310 , a gesture-to-speech finite-state transducer 320 , a lattice projection subsystem 330 , a gesture and speech composing subsystem 340 , a speech/gesture combining subsystem 350 , a speech/gesture/meaning lattice 360 and/or a meaning recognition subsystem 370 .
- the gesture-to-speech composing subsystem 310 inputs the gesture recognition lattice 255 output by the gesture recognition system 200 and composes it with the gesture-to-speech finite-state transducer 320 to generate a gesture/language finite-state transducer 325 .
- the gesture/language finite-state transducer 325 is output to both the lattice projection subsystem 330 and the gesture and speech composing subsystem 340 .
- the lattice projection subsystem 330 generates a projection of the gesture/language finite-state transducer 325 and outputs the projection of the gesture/language finite-state transducer 325 as the grammar or language model lattice 160 to the automatic speech recognition system 100 .
- the multimodal parser/meaning recognition system 300 does not also extract meaning
- the gesture and speech composing subsystem 340 , the speech/gesture combining subsystem 350 , the speech/gesture/meaning lattice 360 and the meaning recognition subsystem 370 can be omitted.
- the multimodal parser/meaning recognition system 300 does not generate a new grammar or language model lattice 160 for the automatic speech recognition system 100 , at least the lattice projection subsystem 330 can be omitted.
- the lattice of possible word sequences 175 is input by the multimodal parser/meaning recognition system 300 .
- the gesture and speech composing subsystem 340 inputs both the lattice of possible word sequences 175 and the gesture/language finite-state transducer 325 .
- the gesture/language finite-state transducer 325 can be generated using any known or later-developed technique for relating the gesture recognition lattice 255 to the recognized possible lattice of word sequences 175 in place of the gesture-to-speech composing subsystem 310 and the gesture-to-speech finite-state transducer 320 .
- the gesture and speech composing subsystem 340 composes these lattices to generate a gesture/speech finite-state transducer 345 .
- the gesture and speech composing subsystem 340 outputs the gesture/speech finite-state transducer 345 to the speech/gesture combining subsystem 350 .
- the speech/gesture combining subsystem 350 converts the gesture/speech finite-state transducer 345 to a gesture/speech finite-state machine 355 .
- the gesture/speech finite-state machine 355 is output by the speech/gesture combining subsystem 350 to the meaning recognition subsystem 370 .
- the meaning recognition subsystem 370 composes the gesture/speech finite-state machine 355 with the speech/gesture/meaning finite-state transducer 360 to generate a meaning lattice 375 .
- the meaning lattice 375 combines the recognition of the speech utterance input through the speech input portion 420 and the recognition of the gestures input through the gesture input portion 410 of the multimodal user input device 400 . The most probable meaning is then selected from the meaning lattice 375 and output to a downstream task.
- the systems and methods disclosed herein use certain simplifying assumptions with respect to temporal constraints.
- the primary function of temporal constraints is to force an order on the gestures. For example, if a user generates the spoken utterance “move this here” and simultaneously makes two gestures, then the first gesture corresponds to the spoken utterance “this”, while the second gesture corresponds to the spoken utterance “here”.
- the multimodal grammars encode order, but do not impose explicit temporal constraints.
- there are multimodal applications in which more specific temporal constraints are relevant. For example, specific temporal constraints can be relevant in selecting among unimodal and multimodal interpretations. That is, if a gesture is temporally distant from the speech, then the unimodal interpretation should be preferred.
- the multimodal user input device 400 includes the gesture input portions 410 and speed input portion 420 as shown in FIG. 5 .
- the gesture input portion 410 displays a graphical user interface that allows the user to direct either e-mail messages or pager messages to the various persons, departments, and/or organizations represented by the objects 412 displayed in the gesture input portion 410 .
- the multimodal user input device 400 also allows the user to input spoken commands to the speech input portion, or microphone, 420 . For simple illustration, further assume that the user has generated the two gestures 414 shown in FIG.
- a multimodal context-free grammar can be defined formally as the quadruple MCFG as follows:
- a context-free grammar can be approximated by a finite-state automaton.
- the transition symbols of the finite-state automaton are the terminals of the context-free grammar.
- these terminals contain three components, W, G and M.
- FIG. 6 illustrates a fragment of such a multimodal context-free grammar that is capable of handling the gesture and spoken utterances illustrated in FIG. 5 .
- FIG. 7 illustrates a three-tape finite-state automaton corresponding to the multimodal context-free grammar fragment shown in FIG. 6 that is capable of handling the gesture and spoken utterances illustrated in FIG. 5 .
- the non-terminals in the multimodal context-free grammar shown in FIG. 6 are atomic symbols.
- the multimodal aspects of the context-free grammar become apparent in the terminals.
- Each terminal contains three components “W:G:M”, corresponding to the n+1 tapes, where:
- the ⁇ symbol is used to indicate when one of these components is empty in a given terminal.
- the symbols in the spoken language component W are words from the speech recognition lattice, i.e., the lattice of possible word sequences 175 .
- the symbols in the gesture component G include the gesture symbols discussed above, as well as the identifier variables e.
- the gestures 414 are simple deictic circling gestures.
- the gesture meaning subsystem 250 assigns semantic types to each gesture 414 based on the underlining portion of the gesture input portion 410 against which the gestures 414 are made.
- the gestures 414 are made relative to the objects 412 that can represent people, organizations or departments to which an e-mail message or a pager message can be directed. If the gesture input portion 410 were instead a map, the gestures would be referenced against particular map coordinates, where the gesture indicates the action to be taken at particular map coordinates or the location of people or things at the indicated map location.
- these semantic types constitute a set of atomic categories which make the relevant distinctions for gesture events to predict speech events and vice versa. For example, if the gesture is a deictic, i.e., pointing, gesture to an object in the gesture input portion 410 that represents a particular person, then spoken utterances like “this person”, “him”, “her”, and the like, are the preferred or predicted speech events and vice versa.
- These categories also play a role in constraining the semantic representation when the speech is underspecified with respect to the semantic type, such as, for example, spoken utterances like “this one”.
- the gesture symbols G can be organized into a type hierarchy reflecting the ontology of the entities in the application domain.
- a pointing gesture may be assigned the general semantic type “G”.
- This general semantic gesture “G” may have various subtypes, such as “Go” and “Gp”, where “Go” represents a gesture made against an organization object, while the “Gp” gesture is made against a person object.
- the “Gp” type gesture may itself have subtypes, such as, for example, “Gpm” and “Gpf” for objects that respectively represent male and female persons.
- each gesture feature structure includes a content portion that allows the specific location of the gesture on the gesture input portion of the multimodal user input device 400 to be specified.
- a unique identifier is needed for each object or location in the gesture input portion 410 that a user can gesture on.
- the finite-state automata would need to include a unique identifier for each object 412 .
- these identifiers would need to be copied from the gesture feature lattice 220 into the semantic representation represented by the gesture recognition lattice 255 generated by the gesture meaning subsystem 250 .
- the systems and methods according to this invention overcome this problem by storing the specific identifiers of the elements within the gesture input portion 410 associated with incoming gestures into a finite set of variables labeled “e 1 ”, “e 2 ”, “e 3 ”, . . . . Then, in place of the specific object identifier, the labels of the variables storing that specific object identifier are incorporated into the various finite-state automata.
- the finite-state automata include arcs labeled with the labels of the variables.
- finite-state automaton instead of having the specific values for each possible object identifier in a finite-state automaton, that finite-state automaton instead incorporates the transitions “ ⁇ :e 1 :e 1 ”, “ ⁇ :e 2 :e 2 ”, “ ⁇ :e 3 :e 3 ”, . . . in each location in the finite-state automaton where specific content needs to be transferred from the gesture tape to the meaning tape.
- transitions labeled with the variable labels are generated from the “ENTRY” productions in the multimodal context-free grammar shown in FIG. 6 .
- the gesture recognition system 200 empties the variables e 1 , e 2 , e 3 . . . after each multimodal command, so that all of the variables can be reused after each multimodal command. This allows the finite-state automaton to be built using a finite set of variables. However, this limits the number of distinct gesture events in a single utterance to no more than the available number of variables.
- the gesture recognition system 200 would output, as the gesture recognition lattice 255 , the finite-state machine shown in FIG. 8 .
- the arc labeled “Gp” corresponds to the gesture made against a person object while the arc labeled “Go” represents a gesture made against an organization object.
- the are labeled “e 1 ” stores the identifier of the person object 412 , in this case, the person object 412 labeled “Robert DeNiro”, against which the person-object gesture “Gp” 414 was made.
- the arc labeled “e 2 ” represents the variable storing the identifier of the organization object 412 , in this case “Monumental Pictures”, against which the organization gesture 414 was made.
- finite-state automata in addition to capturing the structure of language with the finite-state device, meaning is also captured. This is significant in multimodal language processing, because the central goal is to capture how the multiple modes contribute to the combined interpretation.
- symbols are written onto the third tape of the three-tape finite-state automaton, which, when concatenated together, yield the semantic representation for the multimodal utterance.
- This simple representation includes predicates “pred ( . . . )” and lists “[a,b, . . . ]”.
- predicates “pred ( . . . )”
- lists [a,b, . . . ]”.
- semantic representations could be generated, such as code in a programming or scripting language that could be executed directly.
- the recognized word “e-mail” causes the predicate “e-mail ([” to be added to the semantics tape.
- the recognized word “person” causes the predicate “person (” to be written on the semantics tape.
- the e-mail predicate and the list internal to the e-mail predicate are closed when the rule “S ⁇ V MP ⁇ : ⁇ :])”, as shown in FIG. 6 , applies.
- the multimodal recognition and/or meaning system 1000 implements this three-tape finite-state automaton approach by using a series of finite-state transducers in place of the single three-tape finite-state automaton shown in FIG. 7 , as described below.
- the three-tape finite-state automaton shown in FIG. 7 and illustrated by the grammar fragment shown in FIG. 6 can be decomposed into an input component relating the gesture symbols G and the word symbols W and an output component that relates the input component to the meaning symbols M.
- FIG. 7 shows a three-tape finite-state automaton that corresponds to the grammar fragment shown in FIG. 6 and that is usable to recognize the meaning of the various spoken and gestural inputs that can be generated using the graphical user interface displayed in the gesture input portion 410 of the multimodal user input device 400 shown in FIG. 5 .
- the three-tape finite-state automaton shown in FIG. 7 is decomposed into the gesture-to-speech finite-state transducer shown in FIG. 9 and the speech/gesture/meaning finite-state transducer shown in FIG. 10 .
- the gesture-to-speech finite-state transducer shown in FIG. 9 maps the gesture symbols G to the word symbols W that are expected to coincide with each other.
- the verbal pointers “that” and “this” are expected to be accompanied by the deictic gestures 414 made against either a department object, an organization object or a person object 412 , where the object identifier for the object 412 against which the deictic gesture 414 was made is stored in one of the variables e 1 , e 2 , or e 3 .
- FIG. 9 captures the constraints that the gestures made by the user through the gesture input portion 410 of the multimodal user input device 400 place on the speech utterance that accompanies those gestures. Accordingly, a projection of the output tape of the gesture-to-speech finite-state transducer shown in FIG. 9 can be used, in conjunction with the recognized gesture string, such as the recognized gesture string shown in FIG. 8 that represents the gestures illustrated in the exemplary embodiment of the multimodal user input device 400 shown in FIG. 5 , as a language model usable to constrain the possible sequences of words to be recognized by the utterance recognition subsystem 170 of the automatic speech recognition system 100 .
- the speech/gesture/meaning finite-state transducer shown in FIG. 10 uses the cross-product of the gesture symbols G and the word symbols Was an input component or first tape.
- the gesture-to-speech finite-state transducer shown in FIG. 9 implements the function : G ⁇ W.
- the output or second tape of the speech/gesture/meaning finite-state transducer shown in FIG. 10 contains the meaning symbols M that capture the semantic representation of the multimodal utterance, as shown in FIG. 7 and outlined above.
- the speech/gesture/meaning finite-state transducer shown in FIG. 10 implements the function I: (G ⁇ W) ⁇ M. That is, the speech/gesture/meaning finite-state transducer shown in FIG. 10 is a finite-state transducer in which gesture symbols and words are on the input tape and the meaning is on the output tape.
- the gesture-to-speech finite-state transducer and the speech/gesture/meaning finite-state transducers shown in FIGS. 9 and 10 are used with the speech recognition system 100 and the multimodal parser/meaning recognition system 300 to recognize, parse, and/or extract the meaning from the multimodal inputs received from the gesture and speech input portions 410 and 420 of the multimodal user input device 400 .
- multimodal finite-state transducers can be integrated with the automatic speech recognition system 100 , the gesture recognition system 200 and the multimodal parser/meaning recognition system 300 .
- the more appropriate approach will depend on the properties of the particular multimodal user input interface 400 through which the multimodal inputs are generated and/or received.
- the approach outlined in the following description of FIGS. 8-17 involves recognizing the gesture string first.
- the recognized gesture string is then used to modify the language model used by the automatic speech recognition system 100 .
- this will be appropriate when there is limited ambiguity in the recognized gesture string.
- there will be limited ambiguity in the recognized gesture string when the majority of gestures are unambiguous deictic pointing gestures.
- more complex gestures such as the multi-element gestures described in Johnston 1-3, other ways of combining the gesture and speech recognition systems may be more appropriate.
- the gesture recognition system 200 first processes the incoming gestures to construct a gesture finite-state machine, such as that shown in FIG. 8 , corresponding to the range of gesture interpretations.
- a gesture finite-state machine such as that shown in FIG. 8
- the gesture input is unambiguous.
- a simple linearly-connected set of states forms the gesture finite-state machine shown in FIG. 8 .
- the recognized string of gestures would be represented as a lattice indicating all of the possible gesture recognitions and interpretations for the received gesture stream.
- a weighted finite-state transducer could be used to incorporate the likelihoods of the various paths in such a lattice.
- FIG. 11 is a flowchart outlining one exemplary embodiment of a method for combining and converting the various multimodal input streams into a combined finite-state transducer representing the semantic meaning of the combined multimodal input streams.
- control continues to step 510 , where gesture and speech utterances are input through one or more input devices that together combine to form a multimodal user input device.
- step 520 a gesture lattice or finite-state machine is generated from the input gesture utterance.
- the gesture lattice is composed with the gesture-to-speech transducer to generate a gesture/language finite-state transducer.
- the gesture finite-state machine shown in FIG. 8 is composed with the gesture-to-speech finite-state transducer shown in FIG. 9 to form the gesture/language finite-state transducer shown in FIG. 12 .
- the gesture/language finite-state transducer represents the relationship between the recognized stream of gestures and all of the possible word sequences that could occur with those gestures of the recognized stream of gestures.
- a projection of the gesture/language finite-state transducer is generated.
- a projection on the output tape or speech portion of the gesture/language finite-state transducer shown in FIG. 12 is taken to yield a finite-state machine.
- a projection of the gesture/language finite-state transducer shown in FIG. 12 is illustrated in FIG. 13 .
- step 550 the speech utterance is recognized using the projection of the gesture/language finite-state transducer as the language model.
- the projection of the gesture/language finite-state transducer as the language model enables the gestural information to directly influence the recognition process performed by the automatic speech recognition system 100 .
- the automatic speech recognition system generates a word sequence lattice based on the projection of the gesture/language finite-state transducer in view of the word lattice 155 .
- using the projection of the gesture/language finite-state transducer shown in FIG. 13 as the language model for the speech recognition process results in the recognized word sequence lattice “e-mail this person and that organization”, as shown in FIG. 14 .
- the gesture/language finite-state transducer is composed with the recognized word sequences lattice to generate a gesture/speech finite-state transducer. This reintegrates the gesture information that was removed when the projection of the gesture/language finite-state transducer was generated in step 540 .
- the generated gesture/speech finite-state transducer contains the information both from the speech utterance and the gesture utterance received from the various portions of the multimodal user input device 400 . For the example outlined above, composing the gesture/language finite-state transducer shown in FIG. 12 with the word sequences lattice shown in FIG. 14 generates the gesture/speech finite-state transducer shown in FIG. 15 .
- the gesture/speech finite-state transducer is converted to a gesture/speech finite-state machine.
- the gesture/speech finite-state machine combines the input and output tapes of the gesture/speech finite-state transducer onto a single tape.
- converting the gesture/speech finite-state transducer shown in FIG. 15 results in the gesture/speech finite-state machine shown in FIG. 16 .
- the gesture/speech finite-state machine is composed with the speech/gesture/meaning finite-state transducer shown in FIG. 10 to generate the meaning finite-state transducer shown in FIG. 17 .
- the speech/gesture/meaning finite-state transducer relates the speech and gesture symbols to meaning
- composing the gesture/speech finite-state machine results in the meaning finite-state transducer which captures the combined semantic meaning or representation contained in the independent modes input using the multimodal user input device.
- the meaning of the multimodal input received from the multimodal user input device can be read from the output tape of the meaning finite-state transducer.
- the meaning finite-state transducer shown in FIG. 17 is a linear finite-state transducer that unambiguously yields the meaning “e-mail ([person (e 1 ), org (e 2 )])”.
- the meaning finite-state transducer may very well be a weighted finite-state transducer having multiple paths between the start and end nodes representing the various possible meanings for the multimodal input and the probability corresponding to each path.
- the most likely meaning would be selected from the meaning finite-state transducer based on the path through the meaning finite-state transducer having the highest probability.
- step 595 is optional and can be omitted. Then, in step 600 , the process ends.
- the various exemplary embodiments described herein allow spoken language and gesture input streams to be parsed and integrated by a single weighted finite-state device.
- This single weighted finite-state device provides language models for speech and gesture recognition and composes the meaning content from the speech and gesture input streams into a single semantic representation.
- the various systems and methods according to this invention not only address multimodal language recognition, but also encode the semantics as well as the syntax into a single weighted finite-state device.
- the systems and methods according to this invention provide the potential for mutual compensation among the various multimodal input modes.
- the systems and methods according to this invention allow the gestural input to dynamically alter the language model used for speech recognition. Additionally, the systems and methods according to this invention reduce the computational complexity of multi-dimensional multimodal parsing.
- the weighted finite-state devices used in the systems and methods according to this invention provide a well-understood probabilistic framework for combining the probability distributions associated with the speech and gesture input streams and for selecting among multiple competing multimodal interpretations.
- the multimodal recognition and/or meaning system 1000 shown in FIG. 2 and/or each of the gesture recognition system 200 , the multimodal parser/meaning recognition system 300 and/or the automatic speech recognition system 100 can each be implemented on a programmed general purpose computer. However, any or all of these systems can also be implemented on a special purpose computer, a programmed microprocessor or microcontroller and peripheral integrated circuit elements, an ASIC or other integrated circuit, a digital signal processor, a hardwired electronic or a logic circuit such as a discrete element circuit, a programmable logic device such as a PLD, a PLA, a FPGA or a PAL, or the like.
- any device capable of implementing a finite-state machine that is in turn capable of implementing the flowchart shown in FIG. 10 and/or the various finite-state machines and transducers shown in FIGS. 7-9 and 11 - 17 can be used to implement one or more of the various systems shown in FIGS. 1-4 .
- each of the various systems and subsystems shown in FIGS. 1-4 can be implemented as portions of a suitably programmed general purpose computer.
- each of the systems or subsystems shown in FIGS. 1-4 can be implemented as physically distinct hardware circuits within an ASIC, or using a FPGA, a PLD, a PLA, or a PAL, or using discrete logic elements or discrete circuit elements.
- the particular form each of the systems and/or subsystems shown in FIGS. 1-4 will take is a design choice and will be obvious and predictable to those skilled in the art.
- any known or later-developed set of two or more input streams representing different modes of information or communication such as speech, electronic-ink-based gestures or haptic modes, keyboard input, inputs generated by observing or sensing human body motions, including hand motions, gaze motions, facial expressions, or other human body motions, or any other known or later-developed method for communicating information, can be combined and used as one of the input streams in the multimodal utterance.
Landscapes
- Engineering & Computer Science (AREA)
- Health & Medical Sciences (AREA)
- Human Computer Interaction (AREA)
- Physics & Mathematics (AREA)
- Multimedia (AREA)
- Audiology, Speech & Language Pathology (AREA)
- Theoretical Computer Science (AREA)
- Computational Linguistics (AREA)
- Acoustics & Sound (AREA)
- General Health & Medical Sciences (AREA)
- General Physics & Mathematics (AREA)
- General Engineering & Computer Science (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Psychiatry (AREA)
- Social Psychology (AREA)
- User Interface Of Digital Computer (AREA)
- Machine Translation (AREA)
Abstract
Description
- 1. Field of Invention
- This invention is directed to parsing and understanding of utterances whose content is distributed across multiple input modes.
- 2. Description of Related Art
- Multimodal interfaces allow input and/or output to be conveyed over multiple different channels, such as speech, graphics, gesture and the like. Multimodal interfaces enable more natural and effective interaction, because particular modes are best-suited for particular kinds of content. Multimodal interfaces are likely to play a critical role in the ongoing migration of interaction from desktop computing to wireless portable computing devices, such as personal digital assistants, like the Palm Pilot®, digital cellular telephones, public information kiosks that are wirelessly connected to the Internet or other distributed networks, and the like. One barrier to adopting such wireless portable computing devices is that they offer limited screen real estate, and often have limited keyboard interfaces, if any keyboard interface at all.
- To realize the full potential of such wireless portable computing devices, multimodal interfaces need to support not just input from multiple modes. Rather, multimodal interfaces also need to support synergistic multimodal utterances that are optimally distributed over the various available modes. In order to achieve this, the content from different modes needs to be effectively integrated.
- One previous attempt at integrating the content from the different modes is disclosed in “Unification-Based Multimodal Integration”, M. Johnston et al., Proceedings of the 35th ACL, Madrid Spain, p. 281-288, 1997 (Johnston 1), incorporated herein by reference in its entirety. Johnston 1 disclosed a pen-based device that allows a variety of gesture utterances to be input through a gesture mode, while a variety of speech utterances can be input through a speech mode.
- In Johnston 1, a unification operation over typed feature structures was used to model the integration between the gesture mode and the speech mode. Unification operations determine the consistency of two pieces of partial information. If the two pieces of partial information are determined to be consistent, the unification operation combines the two pieces of partial information into a single result. Unification operations were used to determine whether a given piece of gestural input received over the gesture mode was compatible with a given piece of spoken input received over the speech mode. If the gestural input was determined to be compatible with the spoken input, the two inputs were combined into a single result that could be further interpreted.
- In Johnston 1, typed feature structures were used as a common meaning representation for both the gestural inputs and the spoken inputs. In Johnston 1, the multimodal integration was modeled as a cross-product unification of feature structures assigned to the speech and gestural inputs. While the technique disclosed in Johnston 1 overcomes many of the limitations of earlier multimodal systems, this technique does not scale well to support multi-gesture utterances, complex unimodal gestures, or other modes and combinations of modes. To address these limitations, the unification-based multimodal integration technique disclosed in Johnston 1 was extended in “Unification-Based Multimodal Parsing”, M. Johnston, Proceedings of COLING-ACL 98, p. 624-630, 1998 (Johnston 2), herein incorporated by reference in its entirety. The multimodal integration technique disclosed in Johnston 2 uses a multi-dimensional chart parser. In Johnston 2, elements of the multimodal input are treated as terminal edges by the parser. The multimodal input elements are combined together in accordance with a unification-based multimodal grammar. The unification-based multimodal parsing technique disclosed in Johnston 2 was further extended in “Multimodal Language Processing”, M. Johnston, Proceedings of ICSLP 1998, 1998 (published on CD-ROM only) (Johnston 3), incorporated herein by reference in its entirety.
- Johnston 2 and 3 disclosed how techniques from natural language processing can be adapted to support parsing and interpretation of utterances distributed over multiple modes. In the approach disclosed by Johnston 2 and 3, speech and gesture recognition produce n-best lists of recognition results. The n-best recognition results are assigned typed feature structure representations by speech interpretation and gesture interpretation components. The n-best lists of feature structures from the spoken inputs and the gestural inputs are passed to a multi-dimensional chart parser that uses a multimodal unification-based grammar to combine the representations assigned to the input elements. Possible multimodal interpretations are then ranked. The optimal interpretation is then passed on for execution.
- However, the unification-based approach disclosed in Johnston 1-Johnston 3 does not allow for tight coupling of multimodal parsing with speech and gesture recognition. Compensation effects are dependent on the correct answer appearing in each of the n-best list of interpretations obtained from the recognitions obtained from the inputs of each mode. Moreover, multimodal parsing cannot directly influence the progress of either speech recognition or gesture recognition. The multi-dimensional parsing approach is also subject to significant concerns in terms of computational complexity. In the worst case, for the multi-dimensional parsing technique disclosed in Johnston 2, the number of parses to be considered is exponential relative to the number of input elements and the number of interpretations the input elements have. This complexity is manageable when the inputs yield only n-best results for small n. However, the complexity quickly gets out of hand if the inputs are sizable lattices with associated probabilities.
- The unification-based approach also runs into significant problems when choosing between multiple competing parses and interpretations. Probabilities associated with composing speech events and multiple gestures need to be combined. Uni-modal interpretations need to be compared to multimodal interpretations and so on. While this can all be achieved using the unification-based approach disclosed in Johnston 1-Johnston 3, significant post-processing of sets of competing multimodal interpretations generated by the multimodal parser will be involved.
- This invention provides systems and methods that allow parsing understanding and/or integration of the gestural inputs and the spoken inputs using one or more finite-state devices.
- This invention separately provides systems and methods that allow multi-dimensional parsing and understanding using weighted finite-state automata.
- This invention further provides systems and methods that allow multi-dimensional parsing and understanding using a three-tape weighted finite-state automaton.
- This invention separately provides systems and methods that use combinations of finite-state transducers to integrate the various modes of the multimodal interface.
- This invention separately provides systems and methods that use the recognition results of one mode of the multimodal input received from the multimodal interface as a language model or other model in the recognition process of other modes of the multimodal inputs received from the multimodal interface.
- This invention separately provides systems and methods that use the recognition results of one mode of the multimodal input received from the multimodal interface to constrain the recognition process of one or more of the other modes of the multimodal input received from the multimodal interface.
- This invention further provides systems and methods that integrate the recognition results from the second multimodal input, which are based on the recognition results of the first multimodal input, with the recognition results of the first multimodal input and then extract meaning from the combined recognition results.
- This invention further provides systems and methods that base the speech recognition on the results of the gesture recognition.
- The various exemplary embodiments of the systems and methods according to this invention allow spoken language and gesture input streams to be parsed and integrated by a single weighted finite-state device. This single weighted finite-state device provides language models for speech and gesture recognition and composes the meaning content from the speech and gesture input streams into a single semantic representation. Thus, the systems and methods according to this invention not only address multimodal language recognition, but also encode the semantics as well as the syntax into a single weighted finite-state device. Compared to the previous approaches for integrating multimodal input streams, such as those described in Johnston 1-3, which compose elements from n-best lists of recognition results, the systems and methods according to this invention provide the potential for direct compensation among the various multimodal input modes.
- Various exemplary embodiments of the systems and methods according to this invention allow the gestural input to dynamically alter the language model used for speech recognition. Various exemplary embodiments of the systems and methods according to this invention reduce the computational complexity of multi-dimensional multimodal parsing. In particular, the weighted finite-state devices used in various exemplary embodiments of the systems and methods according to this invention provide a well-understood probabilistic framework for combining the probability distributions associated with the speech and gesture or other input modes and for selecting among multiple competing multimodal interpretations.
- These and other features and advantages of this invention are described in, or are apparent from, the following detailed description of various exemplary embodiments of the systems and methods according to this invention.
- Various exemplary embodiments of this invention will be described in detail, with reference to the following figures, wherein:
-
FIG. 1 is a block diagram illustrating one exemplary embodiment of a conventional automatic speech recognition system usable with a multimodal meaning recognition system according to this invention; -
FIG. 2 is a block diagram illustrating one exemplary embodiment of a multimodal user input device and one exemplary embodiment of a multimodal meaning recognition system according to this invention; -
FIG. 3 is a block diagram illustrating in greater detail one exemplary embodiment of the gesture recognition system ofFIG. 2 ; -
FIG. 4 is a block diagram illustrating in greater detail one exemplary embodiment of the multimodal parser and meaning recognition system ofFIG. 2 ; -
FIG. 5 is a block diagram illustrating in greater detail one exemplary embodiment of the multimodal user input device ofFIG. 2 ; -
FIG. 6 is one exemplary embodiment of a multimodal grammar fragment usable by the multimodal meaning recognition system according to this invention; -
FIG. 7 is one exemplary embodiment of a three-tape multimodal finite-state automaton usable to recognize the multimodal inputs received from the exemplary embodiment of the multimodal user input device shown inFIG. 5 ; -
FIG. 8 is one exemplary embodiment of a gesture finite-state machine generated by recognizing the gesture inputs shown in the exemplary embodiment of the multimodal user input device shown inFIG. 5 ; -
FIG. 9 is one exemplary embodiment of a gesture-to-speech finite-state transducer that represents the relationship between speech and gesture for the exemplary embodiment of the multimodal user input device shown inFIG. 5 ; -
FIG. 10 is one exemplary embodiment of a speech/gesture/meaning finite-state transducer that represents the relationship between the combined speech and gesture symbols and the semantic meaning of the multimodal input for the exemplary embodiment of the multimodal input device shown inFIG. 5 ; -
FIG. 11 is a flowchart outlining one exemplary embodiment of a method for extracting meaning from a plurality of multimodal inputs; -
FIG. 12 is one exemplary embodiment of a gesture/language finite-state transducer illustrating the composition of the gesture finite-state machine shown inFIG. 8 with the gesture-to-speech finite-state transducer shown inFIG. 9 ; -
FIG. 13 is one exemplary embodiment of a finite-state machine generated by taking a projection on the output tape of the gesture/language finite-state transducer shown inFIG. 12 ; -
FIG. 14 is one exemplary embodiment of a lattice of possible word sequences generated by the automatic speech recognition system shown inFIG. 1 when using the finite-state machine shown inFIG. 13 as a language model in view of the speech input received from the exemplary embodiment of the multimodal user input device shown inFIG. 5 ; -
FIG. 15 illustrates one exemplary embodiment of a gesture/speech finite-state transducer generated by composing the gesture/language finite-state transducer shown inFIG. 12 with the word sequence lattice shown inFIG. 14 ; -
FIG. 16 is one exemplary embodiment of a gesture/speech finite-state machine obtained from the gesture/speech finite-state transducer shown inFIG. 14 ; and -
FIG. 17 is one exemplary embodiment of a finite-state transducer, obtained from composing the gesture/speech finite-state machine shown inFIG. 16 with the speech/gesture/meaning finite-state transducer shown inFIG. 10 , which extracts the meaning from the multimodal gestural and spoken inputs received when using the exemplary embodiment of the multimodal user input device shown inFIG. 5 . -
FIG. 1 illustrates one exemplary embodiment of an automaticspeech recognition system 100 usable with the multimodal recognition and/ormeaning system 1000 according to this invention that is shown inFIG. 2 . As shown inFIG. 1 , automatic speech recognition can be viewed as a processing pipeline or cascade. - In each step of the processing cascade, one or two lattices are input and composed to produce an output lattice. In automatic speech recognition and in the following description of the exemplary embodiments of the systems and methods of this invention, the term “lattice” denotes a directed and labeled graph, which is possibly weighted. In each lattice, there is typically a designated start node “s” and a designated final node “t”. Each possible pathway through the lattice from the start node s to the final node t induces a hypothesis based on the arc labels between each pair of nodes in the path. For example, in a word lattice, the arc labels are words and the various paths between the start node s and the final node t form sentences. The weights on the arcs on each path between the start node s and the final node t are combined to represent the likelihood that that path will represent a particular portion of the utterance.
- As shown in
FIG. 1 , one exemplary embodiment of a known automaticspeech recognition system 100 includes asignal processing subsystem 110, anacoustic model lattice 120, aphonetic recognition subsystem 130, alexicon lattice 140, aword recognition subsystem 150, a grammar orlanguage model lattice 160, and atask recognition subsystem 170. In operation, uttered speech is input via a microphone, which converts the sound waves of the uttered speech into an electronic speech signal. The electronic speech signal is input to thesignal processing subsystem 110 on a speechsignal input line 105. Thesignal processing subsystem 110 digitizes the electronic speech signal to generate afeature vector lattice 115. Thefeature vector lattice 115 is a lattice of acoustic feature vectors. Thefeature vector lattice 115 is input along with theacoustic model lattice 120 to thephonetic recognition subsystem 130. Theacoustic model lattice 120 represents a set of acoustic models and is applied to transform thefeature vector lattice 115 into a phone lattice. Each node of the phone lattice represents a spoken sound, such as, for example, the vowel /e/ in “bed”. - The
phone lattice 135 is input along with thelexicon lattice 140 into theword recognition subsystem 150. Thelexicon lattice 140 describes different pronunciations of various words and transforms thephone lattice 135 into aword lattice 155. Theword lattice 155 is then input, along with the grammar orlanguage model lattice 160, into theutterance recognition subsystem 170. The grammar orlanguage model lattice 160 represents task-specific information and is used to extract the most likely sequence of uttered words from theword lattice 155. Thus, theutterance recognition subsystem 170 uses the grammar orlanguage model lattice 160 to extract the most likely sentence or other type of utterance from theword lattice 155. In general, the grammar orlanguage model lattice 160 will be selected based on the task associated with the uttered speech. The most likely sequence of words, or the lattice of n most-likely sequences of words, is output as the recognizedutterance 175. - In particular, one conventional method of implementing automatic speech recognition forms each of the
acoustic model lattice 120, thelexicon lattice 140 and the grammar orlanguage model lattice 160 as a finite-state transducer. Thus, each of thephonetic recognition subsystem 130, theword recognition subsystem 150, and theutterance recognition 170 performs a generalized composition operation between its input finite-state transducers. In addition, thesignal processing subsystem 110 outputs thefeatures vector lattice 115 as a finite-state transducer. - Conventionally, the grammar or
language model lattice 160 is predetermined and incorporated into the automaticspeech recognition system 100 based on the particular recognition task that the automaticspeech recognition system 100 is to perform. In various exemplary embodiments, any of theacoustic model lattice 120, thelexicon lattice 140 and/or the grammar orlanguage model 160 can be non-deterministic finite-state transducers. In this case, these non-deterministic finite-state transducers can be determinized using the various techniques disclosed in “Finite-state transducers in Language and Speech Processing”, M. Mohri, Computational Linguistics, 23:2, p. 269-312, 1997, U.S. patent application Ser. No. 09/165,423, filed Oct. 2, 1998, and/or U.S. Pat. No. 6,073,098 to Buchsbaum et al., each incorporated herein by reference in its entirety. - In contrast, in various exemplary embodiments of the systems and methods according to this invention, in the multimodal recognition or
meaning system 1000 shown inFIG. 2 , the automaticspeech recognition system 100 uses a grammar orlanguage model lattice 160 that is obtained from the recognized gestural input received in parallel with thespeech signal 105. This is shown in greater detail inFIG. 2 . In this way, the output of thegesture recognition system 200 can be used to compensate for uncertainties in the automatic speech recognition system. - Alternatively, in various exemplary embodiments of the systems and methods according this invention, the output of the automatic
speech recognition system 100 and output of thegesture recognition system 200 can be combined only after each output is independently obtained. In this way, it becomes possible to extract meaning from the composition of two or more different input modes, such as the two different input modes of speech and gesture. - Furthermore, it should be appreciated that, in various exemplary embodiments of the systems and methods according to this invention, the output of the
gesture recognition system 200 can be used to provide compensation to the automaticspeech recognition system 100. Additionally, their combined output can be further processed to extract meaning from the combination of the two different input modes. In general, when there are two or more different input modes, any of one or more of the input modes can be used to provide compensation to one or more other ones of the input modes. - Thus, it should further be appreciated that, while the following detailed description focuses on speech and gesture as the two input modes, any two or more input modes that can provide compensation between the modes, which can be combined to allow meaning to be extracted from the two or more recognized outputs, or both, can be used in place of, or in addition to, the speech and gesture input modes discussed herein.
- In particular, as shown in
FIG. 2 , when speech and gesture are the implemented input modes, a multimodaluser input device 400 includes agesture input portion 410 and aspeech input portion 420. Thegesture input portion 410 outputs agesture signal 205 to agesture recognition system 200 of the multimodal recognition and/ormeaning system 1000. At the same time, thespeech input portion 420 outputs thespeech signal 105 to the automaticspeech recognition system 100. Thegesture recognition system 200 generates agesture recognition lattice 255 based on the input gesture signal 205 and outputs thegesture recognition lattice 255 to a multimodal parser and meaningrecognition system 300 of the multimodal recognition and/ormeaning system 1000. - In those various exemplary embodiments that provide compensation between the gesture and
speech recognition systems recognition system 300 generates a new grammar orlanguage model lattice 160 for theutterance recognition subsystem 170 of the automaticspeech recognition system 100 from thegesture recognition lattice 255. In particular, this new grammar orlanguage model lattice 160 generated by the multimodal parser/meaningrecognition system 300 is specific to the particular sets of gestural inputs generated by a user through thegesture input portion 410 of the multimodaluser input device 400. Thus, this new grammar orlanguage model lattice 160 represents all of the possible spoken strings that can successfully combine with the particular sequence of gestures input by the user through thegesture input portion 410. That is, the recognition performed by the automaticspeech recognition system 100 can be improved because the particular grammar orlanguage model lattice 160 being used to recognize that spoken utterance is highly specific to the particular sequence of gestures made by the user. - The automatic
speech recognition system 100 then outputs the recognized possibleword sequence lattice 175 back to the multimodal parser/meaningrecognition system 300. In those various exemplary embodiments that do not extract meaning from the combination of the recognized gesture and the recognized speech, the recognized possibleword sequences lattice 175 is then output to a downstream processing task. The multimodal recognition and/ormeaning system 1000 then waits for the next set of inputs from the multimodaluser input device 400. - In contrast, in those exemplary embodiments that additionally extract meaning from the combination of the recognized gesture and the recognized speech, the multimodal parser/meaning
recognition system 300 extracts meaning from the combination of thegesture recognition lattice 255 and the recognized possibleword sequences lattice 175. Because the spoken utterances input by the user through thespeech input portion 420 are presumably closely related to the gestures input at the same time by the user through thegesture input portion 410, the meaning of those gestures can be tightly integrated with the meaning of the spoken input generated by the user through thespeech input portion 420. - The multimodal parser/meaning
recognition system 300 outputs a recognizedpossible meaning lattice 375 in addition to, or in place of, one or both of thegesture recognition lattice 255 and/or the recognized possibleword sequences lattice 175. In various exemplary embodiments, the multimodal parser and meaningrecognition system 300 combines the recognized lattice ofpossible word sequences 175 generated by the automaticspeech recognition system 100 with thegesture recognition lattice 255 output by thegesture recognition system 200 to generate the lattice ofpossible meaning sequences 375 corresponding to the multimodal gesture and speech inputs received from the user through the multimodaluser input device 400. - Moreover, in contrast to both of the embodiments outlined above, in those exemplary embodiments that only extract meaning from the combination of the recognized multimodal inputs, the multimodal parser/meaning
recognition system 300 does not generate the new grammar orlanguage model lattice 160. Thus, thegesture recognition lattice 255 does not provide compensation to the automaticspeech recognition system 100. Rather, the multimodal parser/meaningrecognition system 300 only combines thegesture recognition lattice 255 and the recognized possibleword sequences lattice 175 to generate therecognition meaning lattice 375. - When the
gesture recognition system 200 generates only a single recognized possible sequence of gestures as thegesture recognition lattice 255, that means there is essentially no uncertainty in the gesture recognition. In this case, thegesture recognition lattice 255 provides compensation to the automaticspeech recognition system 100 for any uncertainty in the speech recognition process. However, thegesture recognition system 200 can generate a lattice of n possible recognized gesture sequences as thegesture recognition lattice 255. This recognizes that there may also be uncertainty in the gesture recognition process. - In this case, the
gesture recognition lattice 255 and theword lattice 155 provide mutual compensation for the uncertainties in both the speech recognition process and the gesture recognition process. That is, in the face of this uncertainty, the best, i.e., most-probable, combination of one of the n-best word sequences in theword lattice 155 with one of the n-best gesture sequences in the gesture recognition lattice may not include the best recognition possible sequence from either theword lattice 155 or thegesture recognition lattice 255. For example, the most-probable sequence of gestures in the gesture recognition lattice may combine only with a rather low-probability word sequence through the word lattice, while the most-probable word sequence may combine well only with a rather low-probability gesture sequence. In contrast, a medium-probability word sequence may match very well with a medium-probability gesture sequence. Thus, the net probability of this latter combination of word and gesture sequences may be higher than the probability of the combination of the best word sequence with any of the gesture sequences through thegesture recognition lattice 255 and may be higher than the probability of the combination of the best gesture sequence with any of the word sequences through the lattice ofpossible word sequences 155. In this way, mutual compensation is provided between thegesture recognition system 200 and the automaticspeech recognition system 100. -
FIGS. 3-5 illustrate in greater detail various exemplary embodiments of thegesture recognition system 200, the multimodal parser/meaningrecognition system 300, and the multimodaluser input device 400. In particular, as shown inFIG. 3 , one exemplary embodiment of thegesture recognition system 200 includes a gesturefeature extraction subsystem 210 and agesture recognition subsystem 230. Various other exemplary embodiments may include a gesture language model lattice and a gesture meaning subsystem. In operation, gesture utterances are input through thegesture input portion 410 of the multimodaluser input device 400, which converts the movements of an input device, such as a mouse, a pen, a trackball, a track pad or any other known or later-developed gestural input device, into anelectronic gesture signal 205. At the same time, the multimodaluser input device 400 converts the gestural input into digital ink that can be viewed and understood by the user. This is shown in greater detail inFIG. 5 . - The gesture
feature extraction subsystem 210 converts the motions of the gesture input device represented by the gesture signal 205 into agesture feature lattice 220. As disclosed in Johnston 1-3, the various gestures that can be made can be as simple as pointing gestures to a particular information element at a particular location within thegesture input portion 410 of the multimodaluser input device 400, or can be as complex as a specialized symbol that represents a type of military unit on a military map displayed in thegesture input portion 410 of the multimodaluser input portion 400 and includes an indication of how the unit is to move, and which unit is to move and how far that unit is to move, as described in detail inJohnston 1. - The
gesture feature lattice 220 is input to thegesture recognition subsystem 230. Thegesture recognition subsystem 230 may be implemented as a neural network, as a Hidden-Markov Model (HMM) or as a simpler template-based gesture classification algorithm. Thegesture recognition subsystem 230 converts thegesture feature lattice 220 into thegesture recognition lattice 255. Thegesture recognition lattice 255 includes the identities of graphical elements against which diectic and other simple “identification” gestures are made, possible recognition of more complex gestures that the user may have made and possibly the locations on the displayed graphics where the more complex gesture was made, such as inJohnston 1, and the like. As shown inFIG. 2 , thegesture recognition system 200 outputs thegesture recognition lattice 255 to the multimodal parser/meaningrecognition system 300. - It should be appreciated that the gesture
feature recognition subsystem 210 and thegesture recognition subsystem 230 can each be implemented using any known or later-developed system, circuit or technique that is appropriate. In general, the entiregesture recognition system 200 can be implemented using any known or later-developed system that generates a directed graph from a gesture input. - For example, one known system captures the time and location or locations of the gesture. Optionally, these inputs are then normalized and/or rotated. The gestures are then provided to a pattern classification device that is implemented as part of the gesture
feature recognition subsystem 210. In various exemplary embodiments, this pattern classification device is a template matching system, which transforms the gesture into a feature vector. In various other exemplary embodiments, this pattern classification device is a neural network or a Hidden Markov Model that has been trained to recognize certain patterns of one or more temporally and/or spatially related gesture components as a specific set of features. - When a single gesture is formed by two or more temporally and/or spatially related gesture components, those gesture components can be combined into a single gesture either during the recognition process or by the multimodal parser/meaning
recognition system 300. Once the gesture features are extracted, thegesture recognition subsystem 230 combines the temporally adjacent gestures into a lattice of one or more recognized possible gesture sequences that represent how the recognized gestures follow each other in time. - In various exemplary embodiments, the multimodal parser and meaning
recognition system 300 can be implemented using a single three-tape finite-state device that inputs the output lattices from thespeech recognition system 100 and thegesture recognition system 200 and directly obtains and outputs a meaning result. In various exemplary embodiments, the three-tape finite-state device is a three-tape grammar model that relates the gestures and the words to a meaning of the combination of a gesture and a word.FIG. 7 shows a portion of such a three-tape grammar model usable in the multimodal parser and meaningrecognition system 300 to generate a meaning output from gesture and speech recognition inputs. In general, the multimodal parser and meaningrecognition system 300 can be implemented using an n-tape finite-state device that inputs n−1 lattices from a plurality of recognition systems usable to recognize an utterance having a plurality of different modes. -
FIG. 4 shows the multimodal parser/meaningrecognition system 300 in greater detail. As shown inFIG. 4 , the multimodal parser/meaningrecognition system 300 may include one or more of a gesture-to-speech composing subsystem 310, a gesture-to-speech finite-state transducer 320, alattice projection subsystem 330, a gesture andspeech composing subsystem 340, a speech/gesture combining subsystem 350, a speech/gesture/meaninglattice 360 and/or a meaningrecognition subsystem 370. In particular, the gesture-to-speech composing subsystem 310 inputs thegesture recognition lattice 255 output by thegesture recognition system 200 and composes it with the gesture-to-speech finite-state transducer 320 to generate a gesture/language finite-state transducer 325. The gesture/language finite-state transducer 325 is output to both thelattice projection subsystem 330 and the gesture andspeech composing subsystem 340. - The
lattice projection subsystem 330 generates a projection of the gesture/language finite-state transducer 325 and outputs the projection of the gesture/language finite-state transducer 325 as the grammar orlanguage model lattice 160 to the automaticspeech recognition system 100. Thus, if the multimodal parser/meaningrecognition system 300 does not also extract meaning, the gesture andspeech composing subsystem 340, the speech/gesture combining subsystem 350, the speech/gesture/meaninglattice 360 and the meaningrecognition subsystem 370 can be omitted. Similarly, if the multimodal parser/meaningrecognition system 300 does not generate a new grammar orlanguage model lattice 160 for the automaticspeech recognition system 100, at least thelattice projection subsystem 330 can be omitted. - In those various embodiments that combine the
gesture recognition lattice 255 and the recognized possible lattice ofword sequences 175, whether or not theautomatic speech recognition 100 has generated the lattice ofpossible word sequences 175 based on using the projection of the gesture/language finite-state transducer 325 as the grammar or language model orlattice 160, the lattice ofpossible word sequences 175 is input by the multimodal parser/meaningrecognition system 300. In particular, the gesture andspeech composing subsystem 340 inputs both the lattice ofpossible word sequences 175 and the gesture/language finite-state transducer 325. In those various exemplary embodiments that do not use the output of thegesture recognition system 200 to provide compensation between the speech andgesture recognition systems state transducer 325 can be generated using any known or later-developed technique for relating thegesture recognition lattice 255 to the recognized possible lattice ofword sequences 175 in place of the gesture-to-speech composing subsystem 310 and the gesture-to-speech finite-state transducer 320. - In those various exemplary embodiments that extract meaning from the multimodal inputs, the gesture and
speech composing subsystem 340 composes these lattices to generate a gesture/speech finite-state transducer 345. The gesture andspeech composing subsystem 340 outputs the gesture/speech finite-state transducer 345 to the speech/gesture combining subsystem 350. The speech/gesture combining subsystem 350 converts the gesture/speech finite-state transducer 345 to a gesture/speech finite-state machine 355. The gesture/speech finite-state machine 355 is output by the speech/gesture combining subsystem 350 to the meaningrecognition subsystem 370. The meaningrecognition subsystem 370 composes the gesture/speech finite-state machine 355 with the speech/gesture/meaning finite-state transducer 360 to generate a meaninglattice 375. The meaninglattice 375 combines the recognition of the speech utterance input through thespeech input portion 420 and the recognition of the gestures input through thegesture input portion 410 of the multimodaluser input device 400. The most probable meaning is then selected from the meaninglattice 375 and output to a downstream task. - It should be appreciated that the systems and methods disclosed herein use certain simplifying assumptions with respect to temporal constraints. In multi-gesture utterances, the primary function of temporal constraints is to force an order on the gestures. For example, if a user generates the spoken utterance “move this here” and simultaneously makes two gestures, then the first gesture corresponds to the spoken utterance “this”, while the second gesture corresponds to the spoken utterance “here”. In the various exemplary embodiments of the systems and methods according to this invention described herein, the multimodal grammars encode order, but do not impose explicit temporal constraints. However, it should be appreciated that there are multimodal applications in which more specific temporal constraints are relevant. For example, specific temporal constraints can be relevant in selecting among unimodal and multimodal interpretations. That is, if a gesture is temporally distant from the speech, then the unimodal interpretation should be preferred.
- To illustrate the operation of the multimodal recognition and/or
meaning system 1000, assume the multimodaluser input device 400 includes thegesture input portions 410 andspeed input portion 420 as shown inFIG. 5 . In particular, thegesture input portion 410 displays a graphical user interface that allows the user to direct either e-mail messages or pager messages to the various persons, departments, and/or organizations represented by theobjects 412 displayed in thegesture input portion 410. The multimodaluser input device 400 also allows the user to input spoken commands to the speech input portion, or microphone, 420. For simple illustration, further assume that the user has generated the twogestures 414 shown inFIG. 5 and has spoken the utterance “e-mail this person and that organization” in association with generating thegestures 414 against the graphicaluser interface object 412 labeled “Robert DeNiro” and the graphicaluser interface object 412 labeled “Monumental Pictures”, respectively. - The structure interpretation of multimodal commands of this kind can be captured declaratively in a multimodal context-free grammar. A multimodal context-free grammar can be defined formally as the quadruple MCFG as follows:
-
- MCFG=<N, T, P, S> where
- N is the set of non-terminals;
- P is the set of projections of the form:
- A→α where AεN and αε(N∪T)*;
- S is the start symbol for the grammar;
- T is the set of terminals:
- ((W∪ε)×(G∪ε)×(M∪ε)+),
- where W is the vocabulary of the speech;
- G is the vocabulary of gesture:
- G=(GestureSymbols ∪ EventSymbols);
- GestureSymbols={Gp,Go,Gpf,Gpm . . . };
- Finite collections of EventSymbols={e1,e2 . . . }; and
- M is the vocabulary that represents meaning and includes EventSymbols⊂M.
- In general, a context-free grammar can be approximated by a finite-state automaton. The transition symbols of the finite-state automaton are the terminals of the context-free grammar. In the case of the multimodal context-free grammar defined above, these terminals contain three components, W, G and M. With respect to the discussion outlined above regarding temporal constraints, more specific temporal constraints than order can be encoded in the finite-state approach by writing symbols representing the passage of time onto the gesture tape and referring to such symbols in the multimodal grammar.
-
FIG. 6 illustrates a fragment of such a multimodal context-free grammar that is capable of handling the gesture and spoken utterances illustrated inFIG. 5 .FIG. 7 illustrates a three-tape finite-state automaton corresponding to the multimodal context-free grammar fragment shown inFIG. 6 that is capable of handling the gesture and spoken utterances illustrated inFIG. 5 . The non-terminals in the multimodal context-free grammar shown inFIG. 6 are atomic symbols. The multimodal aspects of the context-free grammar become apparent in the terminals. Each terminal contains three components “W:G:M”, corresponding to the n+1 tapes, where: -
- W represents the spoken language component;
- G represents the gesture component; and
- M represents the combined meaning of the spoken language and gesture components.
- The ε symbol is used to indicate when one of these components is empty in a given terminal. The symbols in the spoken language component W are words from the speech recognition lattice, i.e., the lattice of
possible word sequences 175. The symbols in the gesture component G include the gesture symbols discussed above, as well as the identifier variables e. - In the exemplary embodiment of the
gesture input portion 410 shown inFIG. 5 , thegestures 414 are simple deictic circling gestures. The gesture meaning subsystem 250 assigns semantic types to eachgesture 414 based on the underlining portion of thegesture input portion 410 against which thegestures 414 are made. In the exemplary embodiment shown inFIG. 5 , thegestures 414 are made relative to theobjects 412 that can represent people, organizations or departments to which an e-mail message or a pager message can be directed. If thegesture input portion 410 were instead a map, the gestures would be referenced against particular map coordinates, where the gesture indicates the action to be taken at particular map coordinates or the location of people or things at the indicated map location. - Compared with a feature-based multimodal grammar, these semantic types constitute a set of atomic categories which make the relevant distinctions for gesture events to predict speech events and vice versa. For example, if the gesture is a deictic, i.e., pointing, gesture to an object in the
gesture input portion 410 that represents a particular person, then spoken utterances like “this person”, “him”, “her”, and the like, are the preferred or predicted speech events and vice versa. These categories also play a role in constraining the semantic representation when the speech is underspecified with respect to the semantic type, such as, for example, spoken utterances like “this one”. - In some exemplary embodiments, the gesture symbols G can be organized into a type hierarchy reflecting the ontology of the entities in the application domain. For example, in the exemplary embodiment of the
gesture input portion 410 shown inFIG. 5 , a pointing gesture may be assigned the general semantic type “G”. This general semantic gesture “G” may have various subtypes, such as “Go” and “Gp”, where “Go” represents a gesture made against an organization object, while the “Gp” gesture is made against a person object. Furthermore, the “Gp” type gesture may itself have subtypes, such as, for example, “Gpm” and “Gpf” for objects that respectively represent male and female persons. - In the unification-based multimodal grammar disclosed in Johnston 1-3, spoken phrases and gestures are assigned typed feature structures by the natural language and gesture interpretation systems, respectively. In particular, each gesture feature structure includes a content portion that allows the specific location of the gesture on the gesture input portion of the multimodal
user input device 400 to be specified. - In contrast, when using finite-state automata, a unique identifier is needed for each object or location in the
gesture input portion 410 that a user can gesture on. For example, in the exemplary embodiment shown inFIG. 5 , the finite-state automata would need to include a unique identifier for eachobject 412. In particular, as part of the composition process performed by thegesture recognition system 200, as well as the various composition processes described below, these identifiers would need to be copied from thegesture feature lattice 220 into the semantic representation represented by thegesture recognition lattice 255 generated by the gesture meaning subsystem 250. - In the unification-based approach to multimodal integration disclosed in Johnston 1-3, this is achieved by feature sharing. In the finite-state approach used in the systems and methods according to this invention, one possible, but ultimately unworkable solution, would be to incorporate all of the different possible identifiers for all of the different possible elements of the
gesture input device 410, against which a gesture could be made, into the finite-state automata. For example, for an object having an identifier “object identifier 345”, an arc in the lattices would need to be labeled with that identifier to transfer that piece of information from the gesture tape to the meaning tape of the finite-state automaton. Moreover, the arc for each different identifier would have to be repeated numerous times in the network wherever this transfer of information would be needed. Furthermore, the various arcs would have to be updated as the underlying objects within thegesture input portion 410 were updated or changed. - In various exemplary embodiments, the systems and methods according to this invention overcome this problem by storing the specific identifiers of the elements within the
gesture input portion 410 associated with incoming gestures into a finite set of variables labeled “e1”, “e2”, “e3”, . . . . Then, in place of the specific object identifier, the labels of the variables storing that specific object identifier are incorporated into the various finite-state automata. Thus, instead of having arcs labeled with specific values in the finite-state automata, the finite-state automata include arcs labeled with the labels of the variables. - Therefore, instead of having the specific values for each possible object identifier in a finite-state automaton, that finite-state automaton instead incorporates the transitions “ε:e1:e1”, “ε:e2:e2”, “ε:e3:e3”, . . . in each location in the finite-state automaton where specific content needs to be transferred from the gesture tape to the meaning tape. These transitions labeled with the variable labels are generated from the “ENTRY” productions in the multimodal context-free grammar shown in
FIG. 6 . - In operation, the
gesture recognition system 200 empties the variables e1, e2, e3 . . . after each multimodal command, so that all of the variables can be reused after each multimodal command. This allows the finite-state automaton to be built using a finite set of variables. However, this limits the number of distinct gesture events in a single utterance to no more than the available number of variables. - Accordingly, assuming a user using the
gesture input portion 410 shown inFIG. 5 made thegestures 414 shown inFIG. 5 , thegesture recognition system 200 would output, as thegesture recognition lattice 255, the finite-state machine shown inFIG. 8 . In this case, as shown inFIG. 8 , the arc labeled “Gp” corresponds to the gesture made against a person object while the arc labeled “Go” represents a gesture made against an organization object. The are labeled “e1” stores the identifier of theperson object 412, in this case, the person object 412 labeled “Robert DeNiro”, against which the person-object gesture “Gp” 414 was made. Similarly, the arc labeled “e2” represents the variable storing the identifier of theorganization object 412, in this case “Monumental Pictures”, against which theorganization gesture 414 was made. - In the finite-state automata approach used in the systems and methods according to this invention, in addition to capturing the structure of language with the finite-state device, meaning is also captured. This is significant in multimodal language processing, because the central goal is to capture how the multiple modes contribute to the combined interpretation. In the finite-state automata technique used in the systems and methods according to this invention, symbols are written onto the third tape of the three-tape finite-state automaton, which, when concatenated together, yield the semantic representation for the multimodal utterance.
- In the following discussion, based on the exemplary utterance outlined above with respect to
FIG. 5 , a simple logical representation can be used. This simple representation includes predicates “pred ( . . . )” and lists “[a,b, . . . ]”. However, it should be appreciated that many other kinds of semantic representations could be generated, such as code in a programming or scripting language that could be executed directly. - In the simple logical representation outlined above, referring to the exemplary multimodal utterance outlined above with respect to
FIG. 5 , the recognized word “e-mail” causes the predicate “e-mail ([” to be added to the semantics tape. Similarly, the recognized word “person” causes the predicate “person (” to be written on the semantics tape. The e-mail predicate and the list internal to the e-mail predicate are closed when the rule “S→V MP ε:ε:])”, as shown inFIG. 6 , applies. - Returning to the exemplary utterance “e-mail this person and that organization” and the associated gestures outlined above with respect to
FIG. 5 , assume that the objects against which thegestures 414 have been made have the identifiers “objid367” and “objid893”. Then, the elements on the meaning tape of the three-tape finite-state automaton are concatenated and the variable references are replaced to yield the meaning “e-mail([person(objid367), organization (objid893)])”. - As more recursive semantic phenomena, such as possessives and other complex noun phrases, are added to the grammar, the resulting finite-state automata become ever larger. The computational consequences of this can be lessened by the lazy evaluation techniques in Mohri.
- While a three-tape finite-state automaton is feasible in principle, currently available tools for finite-state language processing generally only support two-tape finite-state automata, i.e., finite-state transducers. Furthermore, speech recognizers typically do not support the use of a three-tape finite-state automaton as a language model. Accordingly, the multimodal recognition and/or
meaning system 1000 implements this three-tape finite-state automaton approach by using a series of finite-state transducers in place of the single three-tape finite-state automaton shown inFIG. 7 , as described below. In particular, the three-tape finite-state automaton shown inFIG. 7 and illustrated by the grammar fragment shown inFIG. 6 can be decomposed into an input component relating the gesture symbols G and the word symbols W and an output component that relates the input component to the meaning symbols M. - As indicated above,
FIG. 7 shows a three-tape finite-state automaton that corresponds to the grammar fragment shown inFIG. 6 and that is usable to recognize the meaning of the various spoken and gestural inputs that can be generated using the graphical user interface displayed in thegesture input portion 410 of the multimodaluser input device 400 shown inFIG. 5 . The three-tape finite-state automaton shown inFIG. 7 is decomposed into the gesture-to-speech finite-state transducer shown inFIG. 9 and the speech/gesture/meaning finite-state transducer shown inFIG. 10 . - The gesture-to-speech finite-state transducer shown in
FIG. 9 maps the gesture symbols G to the word symbols W that are expected to coincide with each other. Thus, in the exemplary embodiment of the multimodaluser input device 400 shown inFIG. 4 , the verbal pointers “that” and “this” are expected to be accompanied by thedeictic gestures 414 made against either a department object, an organization object or aperson object 412, where the object identifier for theobject 412 against which thedeictic gesture 414 was made is stored in one of the variables e1, e2, or e3. The gesture-to-speech transducer shown inFIG. 9 captures the constraints that the gestures made by the user through thegesture input portion 410 of the multimodaluser input device 400 place on the speech utterance that accompanies those gestures. Accordingly, a projection of the output tape of the gesture-to-speech finite-state transducer shown inFIG. 9 can be used, in conjunction with the recognized gesture string, such as the recognized gesture string shown inFIG. 8 that represents the gestures illustrated in the exemplary embodiment of the multimodaluser input device 400 shown inFIG. 5 , as a language model usable to constrain the possible sequences of words to be recognized by theutterance recognition subsystem 170 of the automaticspeech recognition system 100. - It should be appreciated that, in those exemplary embodiments that do not also extract meaning, the further processing outlined below with respect to
FIGS. 10-17 can be omitted. Similarly, in those exemplary embodiments that do not use one or more of the multimodal inputs to provide compensation to one or more of the other multimodal inputs, the processing outlined above with respect toFIGS. 7-9 can be omitted. - The speech/gesture/meaning finite-state transducer shown in
FIG. 10 uses the cross-product of the gesture symbols G and the word symbols Was an input component or first tape. Thus, the gesture-to-speech finite-state transducer shown inFIG. 9 implements the function : G→W. The output or second tape of the speech/gesture/meaning finite-state transducer shown inFIG. 10 contains the meaning symbols M that capture the semantic representation of the multimodal utterance, as shown inFIG. 7 and outlined above. Thus, the speech/gesture/meaning finite-state transducer shown inFIG. 10 implements the function ℑ: (G×W)→M. That is, the speech/gesture/meaning finite-state transducer shown inFIG. 10 is a finite-state transducer in which gesture symbols and words are on the input tape and the meaning is on the output tape. - Thus, the gesture-to-speech finite-state transducer and the speech/gesture/meaning finite-state transducers shown in
FIGS. 9 and 10 are used with thespeech recognition system 100 and the multimodal parser/meaningrecognition system 300 to recognize, parse, and/or extract the meaning from the multimodal inputs received from the gesture andspeech input portions user input device 400. - It should be appreciated that there are any variety of ways in which the multimodal finite-state transducers can be integrated with the automatic
speech recognition system 100, thegesture recognition system 200 and the multimodal parser/meaningrecognition system 300. Clearly, for any particular recognition task, the more appropriate approach will depend on the properties of the particular multimodaluser input interface 400 through which the multimodal inputs are generated and/or received. - The approach outlined in the following description of
FIGS. 8-17 involves recognizing the gesture string first. The recognized gesture string is then used to modify the language model used by the automaticspeech recognition system 100. In general, this will be appropriate when there is limited ambiguity in the recognized gesture string. For example, there will be limited ambiguity in the recognized gesture string when the majority of gestures are unambiguous deictic pointing gestures. Obviously, if more complex gestures are used, such as the multi-element gestures described in Johnston 1-3, other ways of combining the gesture and speech recognition systems may be more appropriate. - Accordingly, for the specific exemplary embodiment of the multimodal
user input device 400 shown inFIG. 5 , thegesture recognition system 200 first processes the incoming gestures to construct a gesture finite-state machine, such as that shown inFIG. 8 , corresponding to the range of gesture interpretations. In the exemplary embodiments described above with respect toFIGS. 5 , 6 and 7, the gesture input is unambiguous. Thus, as shown inFIG. 8 , a simple linearly-connected set of states forms the gesture finite-state machine shown inFIG. 8 . It should be appreciated that, if the received gestures involved more complex gesture recognition or were otherwise ambiguous, the recognized string of gestures would be represented as a lattice indicating all of the possible gesture recognitions and interpretations for the received gesture stream. Moreover, a weighted finite-state transducer could be used to incorporate the likelihoods of the various paths in such a lattice. -
FIG. 11 is a flowchart outlining one exemplary embodiment of a method for combining and converting the various multimodal input streams into a combined finite-state transducer representing the semantic meaning of the combined multimodal input streams. Beginning instep 500, control continues to step 510, where gesture and speech utterances are input through one or more input devices that together combine to form a multimodal user input device. Then, instep 520, a gesture lattice or finite-state machine is generated from the input gesture utterance. - Next, in
step 530, the gesture lattice is composed with the gesture-to-speech transducer to generate a gesture/language finite-state transducer. For example, in the exemplary embodiment described above, the gesture finite-state machine shown inFIG. 8 is composed with the gesture-to-speech finite-state transducer shown inFIG. 9 to form the gesture/language finite-state transducer shown inFIG. 12 . The gesture/language finite-state transducer represents the relationship between the recognized stream of gestures and all of the possible word sequences that could occur with those gestures of the recognized stream of gestures. - Then, in
step 540, in order to use this information to guide thespeech recognition system 100, a projection of the gesture/language finite-state transducer is generated. In particular, a projection on the output tape or speech portion of the gesture/language finite-state transducer shown inFIG. 12 is taken to yield a finite-state machine. In particular, in the exemplary embodiment outlined above, a projection of the gesture/language finite-state transducer shown inFIG. 12 is illustrated inFIG. 13 . - Next, in
step 550, the speech utterance is recognized using the projection of the gesture/language finite-state transducer as the language model. Using the projection of the gesture/language finite-state transducer as the language model enables the gestural information to directly influence the recognition process performed by the automaticspeech recognition system 100. In particular, as shown instep 560, the automatic speech recognition system generates a word sequence lattice based on the projection of the gesture/language finite-state transducer in view of theword lattice 155. In the exemplary embodiment outlined above, using the projection of the gesture/language finite-state transducer shown inFIG. 13 as the language model for the speech recognition process results in the recognized word sequence lattice “e-mail this person and that organization”, as shown inFIG. 14 . - Then, in
step 570, the gesture/language finite-state transducer is composed with the recognized word sequences lattice to generate a gesture/speech finite-state transducer. This reintegrates the gesture information that was removed when the projection of the gesture/language finite-state transducer was generated instep 540. The generated gesture/speech finite-state transducer contains the information both from the speech utterance and the gesture utterance received from the various portions of the multimodaluser input device 400. For the example outlined above, composing the gesture/language finite-state transducer shown inFIG. 12 with the word sequences lattice shown inFIG. 14 generates the gesture/speech finite-state transducer shown inFIG. 15 . - Then, in
step 580, the gesture/speech finite-state transducer is converted to a gesture/speech finite-state machine. In particular, the gesture/speech finite-state machine combines the input and output tapes of the gesture/speech finite-state transducer onto a single tape. In the exemplary embodiment outlined above, converting the gesture/speech finite-state transducer shown inFIG. 15 results in the gesture/speech finite-state machine shown inFIG. 16 . - Next, in
step 590, the gesture/speech finite-state machine is composed with the speech/gesture/meaning finite-state transducer shown inFIG. 10 to generate the meaning finite-state transducer shown inFIG. 17 . Because the speech/gesture/meaning finite-state transducer relates the speech and gesture symbols to meaning, composing the gesture/speech finite-state machine results in the meaning finite-state transducer which captures the combined semantic meaning or representation contained in the independent modes input using the multimodal user input device. Thus, the meaning of the multimodal input received from the multimodal user input device can be read from the output tape of the meaning finite-state transducer. In the exemplary embodiment outlined above, composing the gesture/speech finite-state machine shown inFIG. 16 with the speech/gesture/meaning finite-state transducer shown inFIG. 10 results in the meaning finite-state transducer shown inFIG. 17 . In particular, it should be appreciated that the meaning finite-state transducer shown inFIG. 17 is a linear finite-state transducer that unambiguously yields the meaning “e-mail ([person (e1), org (e2)])”. - It should be appreciated that, in embodiments that use much more complex multimodal interfaces, such as those illustrated in Johnston 1-3, the meaning finite-state transducer may very well be a weighted finite-state transducer having multiple paths between the start and end nodes representing the various possible meanings for the multimodal input and the probability corresponding to each path. In this case, in
step 595, the most likely meaning would be selected from the meaning finite-state transducer based on the path through the meaning finite-state transducer having the highest probability. However, it should be appreciated thatstep 595 is optional and can be omitted. Then, instep 600, the process ends. - As outlined above, the various exemplary embodiments described herein allow spoken language and gesture input streams to be parsed and integrated by a single weighted finite-state device. This single weighted finite-state device provides language models for speech and gesture recognition and composes the meaning content from the speech and gesture input streams into a single semantic representation. Thus, the various systems and methods according to this invention not only address multimodal language recognition, but also encode the semantics as well as the syntax into a single weighted finite-state device. Compared to the previous approaches for integrating multimodal input streams, such as those described in Johnston 1-3, which compose elements from n-best lists of recognition results, the systems and methods according to this invention provide the potential for mutual compensation among the various multimodal input modes.
- The systems and methods according to this invention allow the gestural input to dynamically alter the language model used for speech recognition. Additionally, the systems and methods according to this invention reduce the computational complexity of multi-dimensional multimodal parsing. In particular, the weighted finite-state devices used in the systems and methods according to this invention provide a well-understood probabilistic framework for combining the probability distributions associated with the speech and gesture input streams and for selecting among multiple competing multimodal interpretations.
- It should be appreciated that the multimodal recognition and/or
meaning system 1000 shown inFIG. 2 , and/or each of thegesture recognition system 200, the multimodal parser/meaningrecognition system 300 and/or the automaticspeech recognition system 100 can each be implemented on a programmed general purpose computer. However, any or all of these systems can also be implemented on a special purpose computer, a programmed microprocessor or microcontroller and peripheral integrated circuit elements, an ASIC or other integrated circuit, a digital signal processor, a hardwired electronic or a logic circuit such as a discrete element circuit, a programmable logic device such as a PLD, a PLA, a FPGA or a PAL, or the like. In general, any device capable of implementing a finite-state machine that is in turn capable of implementing the flowchart shown inFIG. 10 and/or the various finite-state machines and transducers shown inFIGS. 7-9 and 11-17 can be used to implement one or more of the various systems shown inFIGS. 1-4 . - Thus, it should be understood that each of the various systems and subsystems shown in
FIGS. 1-4 can be implemented as portions of a suitably programmed general purpose computer. Alternatively, each of the systems or subsystems shown inFIGS. 1-4 can be implemented as physically distinct hardware circuits within an ASIC, or using a FPGA, a PLD, a PLA, or a PAL, or using discrete logic elements or discrete circuit elements. The particular form each of the systems and/or subsystems shown inFIGS. 1-4 will take is a design choice and will be obvious and predictable to those skilled in the art. - It should also be appreciated that, while the above-outlined description of the various systems and methods according to this invention and the figures focus on speech and gesture as the multimodal inputs, any known or later-developed set of two or more input streams representing different modes of information or communication, such as speech, electronic-ink-based gestures or haptic modes, keyboard input, inputs generated by observing or sensing human body motions, including hand motions, gaze motions, facial expressions, or other human body motions, or any other known or later-developed method for communicating information, can be combined and used as one of the input streams in the multimodal utterance.
- Thus, while this invention has been described in conjunction with the exemplary embodiments outlined above, it is evident that many alternatives, modifications and variations will be apparent to those skilled in the art. Accordingly, the exemplary embodiments of these systems and methods according to this invention, as set forth above, are intended to be illustrative, not limiting. Various changes may be made without departing from the spirit and scope of this invention.
Claims (19)
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US13/690,037 US8626507B2 (en) | 2001-07-12 | 2012-11-30 | Systems and methods for extracting meaning from multimodal inputs using finite-state devices |
Applications Claiming Priority (6)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US09/904,252 US6868383B1 (en) | 2001-07-12 | 2001-07-12 | Systems and methods for extracting meaning from multimodal inputs using finite-state devices |
US10/970,215 US7295975B1 (en) | 2001-07-12 | 2004-10-21 | Systems and methods for extracting meaning from multimodal inputs using finite-state devices |
US11/904,085 US8103502B1 (en) | 2001-07-12 | 2007-09-26 | Systems and methods for extracting meaning from multimodal inputs using finite-state devices |
US13/291,427 US8214212B2 (en) | 2001-07-12 | 2011-11-08 | Systems and methods for extracting meaning from multimodal inputs using finite-state devices |
US13/485,574 US8355916B2 (en) | 2001-07-12 | 2012-05-31 | Systems and methods for extracting meaning from multimodal inputs using finite-state devices |
US13/690,037 US8626507B2 (en) | 2001-07-12 | 2012-11-30 | Systems and methods for extracting meaning from multimodal inputs using finite-state devices |
Related Parent Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US13/485,574 Continuation US8355916B2 (en) | 2001-07-12 | 2012-05-31 | Systems and methods for extracting meaning from multimodal inputs using finite-state devices |
Publications (2)
Publication Number | Publication Date |
---|---|
US20130158998A1 true US20130158998A1 (en) | 2013-06-20 |
US8626507B2 US8626507B2 (en) | 2014-01-07 |
Family
ID=34275033
Family Applications (6)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US09/904,252 Expired - Lifetime US6868383B1 (en) | 2001-07-12 | 2001-07-12 | Systems and methods for extracting meaning from multimodal inputs using finite-state devices |
US10/970,215 Expired - Lifetime US7295975B1 (en) | 2001-07-12 | 2004-10-21 | Systems and methods for extracting meaning from multimodal inputs using finite-state devices |
US11/904,085 Expired - Lifetime US8103502B1 (en) | 2001-07-12 | 2007-09-26 | Systems and methods for extracting meaning from multimodal inputs using finite-state devices |
US13/291,427 Expired - Fee Related US8214212B2 (en) | 2001-07-12 | 2011-11-08 | Systems and methods for extracting meaning from multimodal inputs using finite-state devices |
US13/485,574 Expired - Lifetime US8355916B2 (en) | 2001-07-12 | 2012-05-31 | Systems and methods for extracting meaning from multimodal inputs using finite-state devices |
US13/690,037 Expired - Fee Related US8626507B2 (en) | 2001-07-12 | 2012-11-30 | Systems and methods for extracting meaning from multimodal inputs using finite-state devices |
Family Applications Before (5)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US09/904,252 Expired - Lifetime US6868383B1 (en) | 2001-07-12 | 2001-07-12 | Systems and methods for extracting meaning from multimodal inputs using finite-state devices |
US10/970,215 Expired - Lifetime US7295975B1 (en) | 2001-07-12 | 2004-10-21 | Systems and methods for extracting meaning from multimodal inputs using finite-state devices |
US11/904,085 Expired - Lifetime US8103502B1 (en) | 2001-07-12 | 2007-09-26 | Systems and methods for extracting meaning from multimodal inputs using finite-state devices |
US13/291,427 Expired - Fee Related US8214212B2 (en) | 2001-07-12 | 2011-11-08 | Systems and methods for extracting meaning from multimodal inputs using finite-state devices |
US13/485,574 Expired - Lifetime US8355916B2 (en) | 2001-07-12 | 2012-05-31 | Systems and methods for extracting meaning from multimodal inputs using finite-state devices |
Country Status (1)
Country | Link |
---|---|
US (6) | US6868383B1 (en) |
Cited By (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN104134060A (en) * | 2014-08-03 | 2014-11-05 | 上海威璞电子科技有限公司 | Sign language interpreting, displaying and sound producing system based on electromyographic signals and motion sensors |
CN105919591A (en) * | 2016-04-12 | 2016-09-07 | 东北大学 | Surface myoelectrical signal based sign language recognition vocal system and method |
WO2020146134A1 (en) | 2019-01-07 | 2020-07-16 | Cerence Operating Company | Contextual utterance resolution in multimodal systems |
Families Citing this family (100)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20030055644A1 (en) * | 2001-08-17 | 2003-03-20 | At&T Corp. | Systems and methods for aggregating related inputs using finite-state devices and extracting meaning from multimodal inputs using aggregation |
US7711570B2 (en) * | 2001-10-21 | 2010-05-04 | Microsoft Corporation | Application abstraction with dialog purpose |
US8229753B2 (en) * | 2001-10-21 | 2012-07-24 | Microsoft Corporation | Web server controls for web enabled recognition and/or audible prompting |
EP1331630A3 (en) * | 2002-01-07 | 2006-12-20 | AT&T Corp. | Systems and methods for generating weighted finite-state automata representing grammars |
US7949513B2 (en) * | 2002-01-22 | 2011-05-24 | Zi Corporation Of Canada, Inc. | Language module and method for use with text processing devices |
US6990639B2 (en) * | 2002-02-07 | 2006-01-24 | Microsoft Corporation | System and process for controlling electronic components in a ubiquitous computing environment using multimodal integration |
US8601096B2 (en) * | 2002-05-14 | 2013-12-03 | Motorola Mobility Llc | Method and system for multi-modal communication |
US7587318B2 (en) * | 2002-09-12 | 2009-09-08 | Broadcom Corporation | Correlating video images of lip movements with audio signals to improve speech recognition |
US7257575B1 (en) * | 2002-10-24 | 2007-08-14 | At&T Corp. | Systems and methods for generating markup-language based expressions from multi-modal and unimodal inputs |
US8745541B2 (en) | 2003-03-25 | 2014-06-03 | Microsoft Corporation | Architecture for controlling a computer using hand gestures |
US7665041B2 (en) * | 2003-03-25 | 2010-02-16 | Microsoft Corporation | Architecture for controlling a computer using hand gestures |
JP4027269B2 (en) * | 2003-06-02 | 2007-12-26 | キヤノン株式会社 | Information processing method and apparatus |
US8160883B2 (en) * | 2004-01-10 | 2012-04-17 | Microsoft Corporation | Focus tracking in dialogs |
US20050216254A1 (en) * | 2004-03-24 | 2005-09-29 | Gupta Anurag K | System-resource-based multi-modal input fusion |
US7794713B2 (en) * | 2004-04-07 | 2010-09-14 | Lpath, Inc. | Compositions and methods for the treatment and prevention of hyperproliferative diseases |
US7430324B2 (en) * | 2004-05-25 | 2008-09-30 | Motorola, Inc. | Method and apparatus for classifying and ranking interpretations for multimodal input fusion |
DE602005012596D1 (en) * | 2004-10-19 | 2009-03-19 | France Telecom | LANGUAGE RECOGNITION METHOD WITH TEMPORARY MARKET INSERTION AND CORRESPONDING SYSTEM |
US7451125B2 (en) * | 2004-11-08 | 2008-11-11 | At&T Intellectual Property Ii, L.P. | System and method for compiling rules created by machine learning program |
US20060155546A1 (en) * | 2005-01-11 | 2006-07-13 | Gupta Anurag K | Method and system for controlling input modalities in a multimodal dialog system |
US8204995B2 (en) * | 2005-06-29 | 2012-06-19 | Nokia Corporation | Multidevice session establishment for multimodal browsing |
US8977965B1 (en) | 2005-08-19 | 2015-03-10 | At&T Intellectual Property Ii, L.P. | System and method for controlling presentations using a multimodal interface |
US9116989B1 (en) * | 2005-08-19 | 2015-08-25 | At&T Intellectual Property Ii, L.P. | System and method for using speech for data searching during presentations |
EP3133809A3 (en) * | 2005-09-02 | 2017-03-01 | GVBB Holdings S.A.R.L | Automatic metadata extraction and metadata controlled production process |
JP4047885B2 (en) * | 2005-10-27 | 2008-02-13 | 株式会社東芝 | Machine translation apparatus, machine translation method, and machine translation program |
US9026915B1 (en) | 2005-10-31 | 2015-05-05 | At&T Intellectual Property Ii, L.P. | System and method for creating a presentation using natural language |
US7716039B1 (en) * | 2006-04-14 | 2010-05-11 | At&T Intellectual Property Ii, L.P. | Learning edit machines for robust multimodal understanding |
US7756710B2 (en) * | 2006-07-13 | 2010-07-13 | Sri International | Method and apparatus for error correction in speech recognition applications |
US7627541B2 (en) * | 2006-09-15 | 2009-12-01 | Microsoft Corporation | Transformation of modular finite state transducers |
US7624075B2 (en) * | 2006-09-15 | 2009-11-24 | Microsoft Corporation | Transformation of modular finite state transducers |
US9830912B2 (en) | 2006-11-30 | 2017-11-28 | Ashwin P Rao | Speak and touch auto correction interface |
US8355915B2 (en) * | 2006-11-30 | 2013-01-15 | Rao Ashwin P | Multimodal speech recognition system |
CA2572116A1 (en) * | 2006-12-27 | 2008-06-27 | Ibm Canada Limited - Ibm Canada Limitee | System and method for processing multi-modal communication within a workgroup |
US20080168478A1 (en) | 2007-01-07 | 2008-07-10 | Andrew Platzer | Application Programming Interfaces for Scrolling |
US20080168402A1 (en) | 2007-01-07 | 2008-07-10 | Christopher Blumenberg | Application Programming Interfaces for Gesture Operations |
US7770136B2 (en) | 2007-01-24 | 2010-08-03 | Microsoft Corporation | Gesture recognition interactive feedback |
US8717305B2 (en) | 2008-03-04 | 2014-05-06 | Apple Inc. | Touch event model for web pages |
US8645827B2 (en) | 2008-03-04 | 2014-02-04 | Apple Inc. | Touch event model |
US20110115702A1 (en) * | 2008-07-08 | 2011-05-19 | David Seaberg | Process for Providing and Editing Instructions, Data, Data Structures, and Algorithms in a Computer System |
US9922640B2 (en) | 2008-10-17 | 2018-03-20 | Ashwin P Rao | System and method for multimodal utterance detection |
US9129601B2 (en) | 2008-11-26 | 2015-09-08 | At&T Intellectual Property I, L.P. | System and method for dialog modeling |
US9311917B2 (en) | 2009-01-21 | 2016-04-12 | International Business Machines Corporation | Machine, system and method for user-guided teaching of deictic references and referent objects of deictic references to a conversational command and control system |
JP5530729B2 (en) * | 2009-01-23 | 2014-06-25 | 本田技研工業株式会社 | Speech understanding device |
US9684521B2 (en) | 2010-01-26 | 2017-06-20 | Apple Inc. | Systems having discrete and continuous gesture recognizers |
US8566045B2 (en) | 2009-03-16 | 2013-10-22 | Apple Inc. | Event recognition |
US8285499B2 (en) | 2009-03-16 | 2012-10-09 | Apple Inc. | Event recognition |
US9123341B2 (en) * | 2009-03-18 | 2015-09-01 | Robert Bosch Gmbh | System and method for multi-modal input synchronization and disambiguation |
US20100281435A1 (en) * | 2009-04-30 | 2010-11-04 | At&T Intellectual Property I, L.P. | System and method for multimodal interaction using robust gesture processing |
US20110242305A1 (en) * | 2010-04-01 | 2011-10-06 | Peterson Harry W | Immersive Multimedia Terminal |
US10216408B2 (en) | 2010-06-14 | 2019-02-26 | Apple Inc. | Devices and methods for identifying user interface objects based on view hierarchy |
US8296151B2 (en) * | 2010-06-18 | 2012-10-23 | Microsoft Corporation | Compound gesture-speech commands |
US9298363B2 (en) | 2011-04-11 | 2016-03-29 | Apple Inc. | Region activation for touch sensitive surface |
US8620113B2 (en) | 2011-04-25 | 2013-12-31 | Microsoft Corporation | Laser diode modes |
US9263045B2 (en) | 2011-05-17 | 2016-02-16 | Microsoft Technology Licensing, Llc | Multi-mode text input |
US8760395B2 (en) | 2011-05-31 | 2014-06-24 | Microsoft Corporation | Gesture recognition techniques |
US8751972B2 (en) * | 2011-09-20 | 2014-06-10 | Google Inc. | Collaborative gesture-based input language |
US20130096919A1 (en) * | 2011-10-12 | 2013-04-18 | Research In Motion Limited | Apparatus and associated method for modifying media data entered pursuant to a media function |
US8954330B2 (en) | 2011-11-28 | 2015-02-10 | Microsoft Corporation | Context-aware interaction system using a semantic model |
US9152376B2 (en) * | 2011-12-01 | 2015-10-06 | At&T Intellectual Property I, L.P. | System and method for continuous multimodal speech and gesture interaction |
US8635637B2 (en) | 2011-12-02 | 2014-01-21 | Microsoft Corporation | User interface presenting an animated avatar performing a media reaction |
US9100685B2 (en) | 2011-12-09 | 2015-08-04 | Microsoft Technology Licensing, Llc | Determining audience state or interest using passive sensor data |
US8811938B2 (en) | 2011-12-16 | 2014-08-19 | Microsoft Corporation | Providing a user interface experience based on inferred vehicle state |
JP5916888B2 (en) * | 2011-12-29 | 2016-05-11 | インテル・コーポレーション | Direct grammar access |
US8898687B2 (en) | 2012-04-04 | 2014-11-25 | Microsoft Corporation | Controlling a media program based on a media reaction |
CA2775700C (en) | 2012-05-04 | 2013-07-23 | Microsoft Corporation | Determining a future portion of a currently presented media program |
US9823742B2 (en) | 2012-05-18 | 2017-11-21 | Microsoft Technology Licensing, Llc | Interaction and management of devices using gaze detection |
US20140007115A1 (en) * | 2012-06-29 | 2014-01-02 | Ning Lu | Multi-modal behavior awareness for human natural command control |
KR20140028174A (en) * | 2012-07-13 | 2014-03-10 | 삼성전자주식회사 | Method for recognizing speech and electronic device thereof |
JP5916566B2 (en) * | 2012-08-29 | 2016-05-11 | アルパイン株式会社 | Information system |
KR102150289B1 (en) * | 2012-08-30 | 2020-09-01 | 삼성전자주식회사 | User interface appratus in a user terminal and method therefor |
US8972243B1 (en) * | 2012-11-20 | 2015-03-03 | Amazon Technologies, Inc. | Parse information encoding in a finite state transducer |
US10629188B2 (en) * | 2013-03-15 | 2020-04-21 | International Business Machines Corporation | Automatic note taking within a virtual meeting |
KR20140143034A (en) * | 2013-06-05 | 2014-12-15 | 삼성전자주식회사 | Method for providing service based on a multimodal input and an electronic device thereof |
US9733716B2 (en) | 2013-06-09 | 2017-08-15 | Apple Inc. | Proxy gesture recognizer |
US9196246B2 (en) * | 2013-06-14 | 2015-11-24 | Mitsubishi Electric Research Laboratories, Inc. | Determining word sequence constraints for low cognitive speech recognition |
US9280970B1 (en) * | 2013-06-25 | 2016-03-08 | Google Inc. | Lattice semantic parsing |
US11199906B1 (en) | 2013-09-04 | 2021-12-14 | Amazon Technologies, Inc. | Global user input management |
US11138971B2 (en) * | 2013-12-05 | 2021-10-05 | Lenovo (Singapore) Pte. Ltd. | Using context to interpret natural language speech recognition commands |
EP2924543B1 (en) * | 2014-03-24 | 2019-12-04 | Tata Consultancy Services Limited | Action based activity determination system and method |
CN110675866B (en) * | 2014-04-22 | 2023-09-29 | 纳宝株式会社 | Method, apparatus and computer readable recording medium for improving at least one semantic unit set |
EP2947635B1 (en) * | 2014-05-21 | 2018-12-19 | Samsung Electronics Co., Ltd. | Display apparatus, remote control apparatus, system and controlling method thereof |
US10140981B1 (en) * | 2014-06-10 | 2018-11-27 | Amazon Technologies, Inc. | Dynamic arc weights in speech recognition models |
EP2958010A1 (en) | 2014-06-20 | 2015-12-23 | Thomson Licensing | Apparatus and method for controlling the apparatus by a user |
US9569174B2 (en) * | 2014-07-08 | 2017-02-14 | Honeywell International Inc. | Methods and systems for managing speech recognition in a multi-speech system environment |
US20160021051A1 (en) * | 2014-07-16 | 2016-01-21 | Theplatform, Llc | Providing Access To Content Via Social Media |
US10048934B2 (en) * | 2015-02-16 | 2018-08-14 | International Business Machines Corporation | Learning intended user actions |
KR102217292B1 (en) * | 2015-02-26 | 2021-02-18 | 네이버 주식회사 | Method, apparatus and computer-readable recording medium for improving a set of at least one semantic units by using phonetic sound |
US9704482B2 (en) * | 2015-03-11 | 2017-07-11 | International Business Machines Corporation | Method and system for order-free spoken term detection |
US20180012227A1 (en) * | 2016-07-05 | 2018-01-11 | NXT-ID, Inc. | Biometric, Behavioral-Metric, Knowledge-Metric, and Electronic-Metric Directed Authentication and Transaction Method and System |
US10419428B2 (en) * | 2015-07-05 | 2019-09-17 | NXT-ID, Inc. | System and method to authenticate electronics using electronic-metrics |
US9679497B2 (en) * | 2015-10-09 | 2017-06-13 | Microsoft Technology Licensing, Llc | Proxies for speech generating devices |
US10148808B2 (en) | 2015-10-09 | 2018-12-04 | Microsoft Technology Licensing, Llc | Directed personal communication for speech generating devices |
US10262555B2 (en) | 2015-10-09 | 2019-04-16 | Microsoft Technology Licensing, Llc | Facilitating awareness and conversation throughput in an augmentative and alternative communication system |
US9966066B1 (en) * | 2016-02-03 | 2018-05-08 | Nvoq Incorporated | System and methods for combining finite state transducer based speech recognizers |
US11262909B2 (en) | 2016-06-02 | 2022-03-01 | Myscript | System and method for input recognition linguistic resource management |
JP6301427B1 (en) * | 2016-10-11 | 2018-03-28 | 株式会社日本総合研究所 | Natural language processing apparatus, natural language processing method, and natural language processing program |
KR102564855B1 (en) * | 2018-01-08 | 2023-08-08 | 삼성전자주식회사 | Device and method to recognize object and face expression, and device and method to train obejct and face expression robust to facial change |
US11520561B1 (en) * | 2018-11-28 | 2022-12-06 | Amazon Technologies, Inc. | Neural network accelerator with compact instruct set |
US11570016B2 (en) | 2018-12-14 | 2023-01-31 | At&T Intellectual Property I, L.P. | Assistive control of network-connected devices |
CN110136713A (en) * | 2019-05-14 | 2019-08-16 | 苏州思必驰信息科技有限公司 | Dialogue method and system of the user in multi-modal interaction |
CN112101045B (en) * | 2020-11-02 | 2021-12-14 | 北京淇瑀信息科技有限公司 | Multi-mode semantic integrity recognition method and device and electronic equipment |
Citations (9)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US6484136B1 (en) * | 1999-10-21 | 2002-11-19 | International Business Machines Corporation | Language model adaptation via network of similar users |
US20040119754A1 (en) * | 2002-12-19 | 2004-06-24 | Srinivas Bangalore | Context-sensitive interface widgets for multi-modal dialog systems |
US6823308B2 (en) * | 2000-02-18 | 2004-11-23 | Canon Kabushiki Kaisha | Speech recognition accuracy in a multimodal input system |
US7054812B2 (en) * | 2000-05-16 | 2006-05-30 | Canon Kabushiki Kaisha | Database annotation and retrieval |
US7240004B1 (en) * | 2002-03-04 | 2007-07-03 | At&T Corp. | Systems and methods for determining the determinizability of finite-state automata and transducers |
US7321854B2 (en) * | 2002-09-19 | 2008-01-22 | The Penn State Research Foundation | Prosody based audio/visual co-analysis for co-verbal gesture recognition |
US20080052080A1 (en) * | 2005-11-30 | 2008-02-28 | University Of Southern California | Emotion Recognition System |
US20100100509A1 (en) * | 2002-10-24 | 2010-04-22 | At&T Corp. | Systems and Methods for Generating Markup-Language Based Expressions from Multi-Modal and Unimodal Inputs |
US20100281435A1 (en) * | 2009-04-30 | 2010-11-04 | At&T Intellectual Property I, L.P. | System and method for multimodal interaction using robust gesture processing |
Family Cites Families (13)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US5502774A (en) * | 1992-06-09 | 1996-03-26 | International Business Machines Corporation | Automatic recognition of a consistent message using multiple complimentary sources of information |
JP2973726B2 (en) * | 1992-08-31 | 1999-11-08 | 株式会社日立製作所 | Information processing device |
JPH06131437A (en) * | 1992-10-20 | 1994-05-13 | Hitachi Ltd | Method for instructing operation in composite form |
JP3416268B2 (en) * | 1994-06-30 | 2003-06-16 | キヤノン株式会社 | Image recognition apparatus and method |
US5748974A (en) * | 1994-12-13 | 1998-05-05 | International Business Machines Corporation | Multimodal natural language interface for cross-application tasks |
JP3363283B2 (en) * | 1995-03-23 | 2003-01-08 | 株式会社日立製作所 | Input device, input method, information processing system, and input information management method |
US5855000A (en) * | 1995-09-08 | 1998-12-29 | Carnegie Mellon University | Method and apparatus for correcting and repairing machine-transcribed input using independent or cross-modal secondary input |
US6438523B1 (en) * | 1998-05-20 | 2002-08-20 | John A. Oberteuffer | Processing handwritten and hand-drawn input and speech input |
US6735566B1 (en) | 1998-10-09 | 2004-05-11 | Mitsubishi Electric Research Laboratories, Inc. | Generating realistic facial animation from speech |
US6167376A (en) * | 1998-12-21 | 2000-12-26 | Ditzik; Richard Joseph | Computer system with integrated telephony, handwriting and speech recognition functions |
US6529863B1 (en) * | 1999-08-30 | 2003-03-04 | Lucent Technologies Inc. | Method and apparatus for providing interactive services with multiple interfaces |
US6665640B1 (en) | 1999-11-12 | 2003-12-16 | Phoenix Solutions, Inc. | Interactive speech based learning/training system formulating search queries based on natural language parsing of recognized user queries |
US7069215B1 (en) * | 2001-07-12 | 2006-06-27 | At&T Corp. | Systems and methods for extracting meaning from multimodal inputs using finite-state devices |
-
2001
- 2001-07-12 US US09/904,252 patent/US6868383B1/en not_active Expired - Lifetime
-
2004
- 2004-10-21 US US10/970,215 patent/US7295975B1/en not_active Expired - Lifetime
-
2007
- 2007-09-26 US US11/904,085 patent/US8103502B1/en not_active Expired - Lifetime
-
2011
- 2011-11-08 US US13/291,427 patent/US8214212B2/en not_active Expired - Fee Related
-
2012
- 2012-05-31 US US13/485,574 patent/US8355916B2/en not_active Expired - Lifetime
- 2012-11-30 US US13/690,037 patent/US8626507B2/en not_active Expired - Fee Related
Patent Citations (10)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US6484136B1 (en) * | 1999-10-21 | 2002-11-19 | International Business Machines Corporation | Language model adaptation via network of similar users |
US6823308B2 (en) * | 2000-02-18 | 2004-11-23 | Canon Kabushiki Kaisha | Speech recognition accuracy in a multimodal input system |
US7054812B2 (en) * | 2000-05-16 | 2006-05-30 | Canon Kabushiki Kaisha | Database annotation and retrieval |
US7240004B1 (en) * | 2002-03-04 | 2007-07-03 | At&T Corp. | Systems and methods for determining the determinizability of finite-state automata and transducers |
US7783485B2 (en) * | 2002-03-04 | 2010-08-24 | At&T Intellectual Property Ii, L.P. | Systems and methods for determining the determinizability of finite-state automata and transducers |
US7321854B2 (en) * | 2002-09-19 | 2008-01-22 | The Penn State Research Foundation | Prosody based audio/visual co-analysis for co-verbal gesture recognition |
US20100100509A1 (en) * | 2002-10-24 | 2010-04-22 | At&T Corp. | Systems and Methods for Generating Markup-Language Based Expressions from Multi-Modal and Unimodal Inputs |
US20040119754A1 (en) * | 2002-12-19 | 2004-06-24 | Srinivas Bangalore | Context-sensitive interface widgets for multi-modal dialog systems |
US20080052080A1 (en) * | 2005-11-30 | 2008-02-28 | University Of Southern California | Emotion Recognition System |
US20100281435A1 (en) * | 2009-04-30 | 2010-11-04 | At&T Intellectual Property I, L.P. | System and method for multimodal interaction using robust gesture processing |
Cited By (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN104134060A (en) * | 2014-08-03 | 2014-11-05 | 上海威璞电子科技有限公司 | Sign language interpreting, displaying and sound producing system based on electromyographic signals and motion sensors |
CN105919591A (en) * | 2016-04-12 | 2016-09-07 | 东北大学 | Surface myoelectrical signal based sign language recognition vocal system and method |
WO2020146134A1 (en) | 2019-01-07 | 2020-07-16 | Cerence Operating Company | Contextual utterance resolution in multimodal systems |
US11455982B2 (en) | 2019-01-07 | 2022-09-27 | Cerence Operating Company | Contextual utterance resolution in multimodal systems |
Also Published As
Publication number | Publication date |
---|---|
US8626507B2 (en) | 2014-01-07 |
US8103502B1 (en) | 2012-01-24 |
US6868383B1 (en) | 2005-03-15 |
US8355916B2 (en) | 2013-01-15 |
US20120116768A1 (en) | 2012-05-10 |
US8214212B2 (en) | 2012-07-03 |
US20120303370A1 (en) | 2012-11-29 |
US7295975B1 (en) | 2007-11-13 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US8626507B2 (en) | Systems and methods for extracting meaning from multimodal inputs using finite-state devices | |
US7069215B1 (en) | Systems and methods for extracting meaning from multimodal inputs using finite-state devices | |
EP1291753B1 (en) | Systems and methods for classifying and representing gestural inputs | |
US9563395B2 (en) | Systems and methods for generating markup-language based expressions from multi-modal and unimodal inputs | |
Waibel et al. | Multimodal interfaces | |
Vo et al. | Building an application framework for speech and pen input integration in multimodal learning interfaces | |
Li et al. | Sign transition modeling and a scalable solution to continuous sign language recognition for real-world applications | |
Wang et al. | An approach based on phonemes to large vocabulary Chinese sign language recognition | |
Johnston et al. | Finite-state multimodal integration and understanding | |
Marquez et al. | A machine learning approach to POS tagging | |
Plamondon et al. | Online handwriting recognition | |
JP2011154099A (en) | Dictionary for speech recognition utilized in speech recognition device and learning method of language model for speech recognition | |
Vo | A framework and toolkit for the construction of multimodal learning interfaces | |
Quek | The catchment feature model: A device for multimodal fusion and a bridge between signal and sense | |
Sagawa et al. | Methods to describe and recognize sign language based on gesture components represented by symbols and numerical values | |
Mukherjee et al. | A visual context-aware multimodal system for spoken language processing. | |
Oviatt | Multimodal Speech and Pen Systems’ | |
Quek et al. | Gesture and speech multimodal conversational interaction | |
Kaiser | Dynamic New Vocabulary Enrollment through Handwriting and Speech in a Multimodal Scheduling Application. | |
Watters et al. | Exploiting a sensed environment to improve human-agent communication | |
Vuurpijl et al. | Overview of the state of the art in fusion of speech and pen input |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
AS | Assignment |
Owner name: AT&T CORP., NEW YORK Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:BANGALORE, SRINIVAS;JOHNSTON, MICHAEL J.;REEL/FRAME:029380/0452 Effective date: 20010709 |
|
FEPP | Fee payment procedure |
Free format text: PAYOR NUMBER ASSIGNED (ORIGINAL EVENT CODE: ASPN); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY |
|
STCF | Information on status: patent grant |
Free format text: PATENTED CASE |
|
AS | Assignment |
Owner name: AT&T INTELLECTUAL PROPERTY II, L.P., GEORGIA Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:AT&T PROPERTIES, LLC;REEL/FRAME:034429/0474 Effective date: 20141006 Owner name: AT&T PROPERTIES, LLC, NEVADA Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:AT&T CORP.;REEL/FRAME:034429/0467 Effective date: 20141006 |
|
AS | Assignment |
Owner name: AT&T ALEX HOLDINGS, LLC, TEXAS Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:AT&T INTELLECTUAL PROPERTY II, L.P.;REEL/FRAME:034467/0822 Effective date: 20141210 |
|
AS | Assignment |
Owner name: INTERACTIONS LLC, MASSACHUSETTS Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:AT&T ALEX HOLDINGS, LLC;REEL/FRAME:034642/0640 Effective date: 20141210 |
|
AS | Assignment |
Owner name: ORIX VENTURES, LLC, NEW YORK Free format text: SECURITY INTEREST;ASSIGNOR:INTERACTIONS LLC;REEL/FRAME:034677/0768 Effective date: 20141218 |
|
AS | Assignment |
Owner name: ARES VENTURE FINANCE, L.P., NEW YORK Free format text: SECURITY INTEREST;ASSIGNOR:INTERACTIONS LLC;REEL/FRAME:036009/0349 Effective date: 20150616 |
|
AS | Assignment |
Owner name: SILICON VALLEY BANK, MASSACHUSETTS Free format text: FIRST AMENDMENT TO INTELLECTUAL PROPERTY SECURITY AGREEMENT;ASSIGNOR:INTERACTIONS LLC;REEL/FRAME:036100/0925 Effective date: 20150709 |
|
AS | Assignment |
Owner name: ARES VENTURE FINANCE, L.P., NEW YORK Free format text: CORRECTIVE ASSIGNMENT TO CORRECT THE CHANGE PATENT 7146987 TO 7149687 PREVIOUSLY RECORDED ON REEL 036009 FRAME 0349. ASSIGNOR(S) HEREBY CONFIRMS THE ASSIGNMENT;ASSIGNOR:INTERACTIONS LLC;REEL/FRAME:037134/0712 Effective date: 20150616 |
|
FPAY | Fee payment |
Year of fee payment: 4 |
|
AS | Assignment |
Owner name: BEARCUB ACQUISITIONS LLC, CALIFORNIA Free format text: ASSIGNMENT OF IP SECURITY AGREEMENT;ASSIGNOR:ARES VENTURE FINANCE, L.P.;REEL/FRAME:044481/0034 Effective date: 20171107 |
|
AS | Assignment |
Owner name: SILICON VALLEY BANK, MASSACHUSETTS Free format text: INTELLECTUAL PROPERTY SECURITY AGREEMENT;ASSIGNOR:INTERACTIONS LLC;REEL/FRAME:049388/0082 Effective date: 20190603 |
|
AS | Assignment |
Owner name: ARES VENTURE FINANCE, L.P., NEW YORK Free format text: RELEASE BY SECURED PARTY;ASSIGNOR:BEARCUB ACQUISITIONS LLC;REEL/FRAME:052693/0866 Effective date: 20200515 |
|
FEPP | Fee payment procedure |
Free format text: MAINTENANCE FEE REMINDER MAILED (ORIGINAL EVENT CODE: REM.); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY |
|
LAPS | Lapse for failure to pay maintenance fees |
Free format text: PATENT EXPIRED FOR FAILURE TO PAY MAINTENANCE FEES (ORIGINAL EVENT CODE: EXP.); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY |
|
STCH | Information on status: patent discontinuation |
Free format text: PATENT EXPIRED DUE TO NONPAYMENT OF MAINTENANCE FEES UNDER 37 CFR 1.362 |
|
FP | Lapsed due to failure to pay maintenance fee |
Effective date: 20220107 |
|
AS | Assignment |
Owner name: INTERACTIONS LLC, MASSACHUSETTS Free format text: TERMINATION AND RELEASE OF SECURITY INTEREST IN INTELLECTUAL PROPERTY;ASSIGNOR:ORIX GROWTH CAPITAL, LLC;REEL/FRAME:061749/0825 Effective date: 20190606 Owner name: INTERACTIONS CORPORATION, MASSACHUSETTS Free format text: TERMINATION AND RELEASE OF SECURITY INTEREST IN INTELLECTUAL PROPERTY;ASSIGNOR:ORIX GROWTH CAPITAL, LLC;REEL/FRAME:061749/0825 Effective date: 20190606 |
|
AS | Assignment |
Owner name: INTERACTIONS LLC, MASSACHUSETTS Free format text: RELEASE OF SECURITY INTEREST IN INTELLECTUAL PROPERTY RECORDED AT REEL/FRAME: 049388/0082;ASSIGNOR:SILICON VALLEY BANK;REEL/FRAME:060558/0474 Effective date: 20220624 |
|
AS | Assignment |
Owner name: INTERACTIONS LLC, MASSACHUSETTS Free format text: RELEASE OF SECURITY INTEREST IN INTELLECTUAL PROPERTY RECORDED AT REEL/FRAME: 036100/0925;ASSIGNOR:SILICON VALLEY BANK;REEL/FRAME:060559/0576 Effective date: 20220624 |