WO1999018556A2 - Vocabulary and/or language model training - Google Patents
Vocabulary and/or language model training Download PDFInfo
- Publication number
- WO1999018556A2 WO1999018556A2 PCT/IB1998/001421 IB9801421W WO9918556A2 WO 1999018556 A2 WO1999018556 A2 WO 1999018556A2 IB 9801421 W IB9801421 W IB 9801421W WO 9918556 A2 WO9918556 A2 WO 9918556A2
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- documents
- vocabulary
- search criterion
- search
- training corpus
- Prior art date
Links
Classifications
-
- G—PHYSICS
- G09—EDUCATION; CRYPTOGRAPHY; DISPLAY; ADVERTISING; SEALS
- G09B—EDUCATIONAL OR DEMONSTRATION APPLIANCES; APPLIANCES FOR TEACHING, OR COMMUNICATING WITH, THE BLIND, DEAF OR MUTE; MODELS; PLANETARIA; GLOBES; MAPS; DIAGRAMS
- G09B19/00—Teaching not covered by other main groups of this subclass
- G09B19/06—Foreign languages
-
- Y—GENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
- Y10—TECHNICAL SUBJECTS COVERED BY FORMER USPC
- Y10S—TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
- Y10S707/00—Data processing: database and file management or data structures
- Y10S707/99931—Database or file accessing
- Y10S707/99933—Query processing, i.e. searching
-
- Y—GENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
- Y10—TECHNICAL SUBJECTS COVERED BY FORMER USPC
- Y10S—TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
- Y10S707/00—Data processing: database and file management or data structures
- Y10S707/99931—Database or file accessing
- Y10S707/99933—Query processing, i.e. searching
- Y10S707/99934—Query formulation, input preparation, or translation
-
- Y—GENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
- Y10—TECHNICAL SUBJECTS COVERED BY FORMER USPC
- Y10S—TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
- Y10S707/00—Data processing: database and file management or data structures
- Y10S707/99931—Database or file accessing
- Y10S707/99933—Query processing, i.e. searching
- Y10S707/99936—Pattern matching access
Definitions
- the invention relates to a method for creating a vocabulary and/or statistical language model from a textual training corpus for subsequent use by a pattern recognition system.
- the invention further relates to a system for creating a vocabulary and/or a statistical language model for subsequent use by a pattern recognition system; the system comprising means for creating the vocabulary and/or statistical language model from a textual training corpus.
- the invention also relates to a pattern recognition system for recognising a time-sequential input pattern using a vocabulary and/or statistical language model; the pattern recognition system comprising the system for creating a vocabulary and/or statistical language model from a textual training corpus.
- Pattern recognition systems such as large vocabulary continuous speech recognition systems or handwriting recognition systems, typically use a vocabulary to recognise words and a language model to improve the basic recognition result.
- Figure 1 illustrates a typical large vocabulary continuous speech recognition system 100 [refer L.Rabiner, B-H. Juang, "Fundamentals of speech recognition", Prentice Hall 1993, pages 434 to 454] .
- the system 100 comprises a spectral analysis subsystem 110 and a unit matching subsystem 120.
- the speech input signal (SIS) is spectrally and/or temporally analysed to calculate a representative vector of features
- observation vector OV
- the speech signal is digitised (e.g. sampled at a rate of 6.67 kHz.) and pre-processed, for instance by applying pre-emphasis.
- Consecutive samples are grouped (blocked) into frames, corresponding to, for instance, 32 msec, of speech signal. Successive frames partially overlap, for instance, 16 msec.
- LPC Linear Predictive Coding
- the feature vector may, for instance, have 24, 32 or 63 components.
- the observation vectors are matched against an inventory of speech recognition units.
- a speech recognition unit is represented by a sequence of acoustic references.
- Various forms of speech recognition units may be used.
- a whole word or even a group of words may be represented by one speech recognition unit.
- a word model (WM) provides for each word of a given vocabulary a transcription in a sequence of acoustic references.
- WM word model
- Other systems in particular large vocabulary systems, may use for the speech recognition unit linguistically based sub- word units, such as phones, diphones or syllables, as well as derivative units, such as fenenes and fenones.
- a word model is given by a lexicon 134, describing the sequence of sub-word units relating to a word of the vocabulary, and the sub-word models 132, describing sequences of acoustic references of the involved speech recognition unit.
- a word model composer 136 composes the word model based on the subword model 132 and the lexicon 134.
- Figure 2A illustrates a word model 200 for a system based on whole- word speech recognition units, where the speech recognition unit of the shown word is modelled using a sequence of ten acoustic references (201 to 210).
- Figure 2B illustrates a word model 220 for a system based on sub- word units, where the shown word is modelled by a sequence of three sub-word models (250, 260 and 270), each with a sequence of four acoustic references (251 , 252, 253, 254; 261 to 264; 271 to 274).
- the word models shown in Fig. 2 are based on Hidden Markov Models, which are widely used to stochastically model speech and handwriting signals.
- each recognition unit (word model or subword model) is typically characterised by an HMM, whose parameters are estimated from a training set of data.
- HMM Hidden Markov Model
- a HMM state corresponds to an acoustic reference (for speech recognition) or an allographic reference (for handwriting recognition).
- Various techniques are known for modelling a reference, including discrete or continuous probability densities.
- a word level matching system 130 matches the observation vectors against all sequences of speech recognition units and provides the likelihoods of a match between the vector and a sequence. If sub- word units are used, constraints are placed on the matching by using the lexicon 134 to limit the possible sequence of sub-word units to sequences in the lexicon 134. This reduces the outcome to possible sequences of words.
- a sentence level matching system 140 uses a language model (LM) to place further constraints on the matching so that the paths investigated are those corresponding to word sequences which are proper sequences as specified by the language model. In this way, the outcome of the unit matching subsystem 120 is a recognised sentence (RS).
- the language model used in pattern recognition may include syntactical and/or semantical constraints 142 of the language and the recognition task.
- a language model based on syntactical constraints is usually referred to as a grammar 144. Similar systems are known for recognising handwriting.
- the language model used for a handwriting recognition system may in addition to or as an alternative to specifying word sequences specify character sequences.
- N-gram word models are widely used.
- P(w w 1 w 2 w 3 ...w J . 1 ) is approximated by P(W j
- bigrams or trigrams are used.
- P(W j ⁇ w,w 2 w 3 ...w J . 1 ) is approximated by P(W j
- the invention relates to recognition systems which use a vocabulary and/or a language model which can, preferably automatically, be build from a textual training corpus.
- a vocabulary can be simply retrieved from a document by collecting all different words in the document.
- the set of words may be reduced, for instance, to words which occur frequently in the document (in absolute terms or relative terms, like relative to other words in the document, or relative with respect to a frequency of occurrence in default documents) .
- a way of automatically building an N-gram language model is to estimate the conditional probabilities P(W j j W j ⁇ . + . -W j .,) by a simple relative frequency: F(w J.N + 1 ...w J .,w J )/F(w J . N + 1 ...w J . 1 ), in which F is the number of occurrences of the string in its argument in the given textual training corpus.
- F(W j . N+ ,...W j . ,W j ) has to be substantial in the given corpus.
- One way of achieving this is to use an extremely large training corpus, which covers most relevant word sequences.
- Another approach is to ensure that the training corpus is representative of many words and word sequences used for a specific recognition task. This can be achieved by manually collecting documents relevant for a specific category of user, such as a radiologist, a surgeon or a legal practitioner. However, such an approach is not possible for recognition systems targeted towards users whose specific interests are not known in advance. Moreover, if a user develops a new interest, a default provided vocabulary and language model will not reflect this, resulting in a degraded recognition result.
- a further object is to create a vocabulary and/or language model which allows improved or faster recognition.
- the method comprises the steps of determining at least one context identifier; deriving at least one search criterion from the context identifier; selecting documents from a set of documents based on the search criterion; and composing the training corpus from the selected documents.
- searching for documents based on a search criterion derived from a context identifier pertinent documents are collected in an effective way, ensuring that pertinent language elements are covered. This increases the quality of recognition. Moreover, also many irrelevant language elements will not be included, allowing the creation of a relatively small vocabulary or language model. This in turn can lead to a faster recognition or, alternatively, improve the recognition rate by adding more elements, such as acoustic data or allographic data, in other parts of the recognition system.
- the context identifier comprises a keyword, which acts as the search criterion. For instance, the (prospective) user of a pattern recognition system specifies one or more keywords, based on which the documents are selected.
- the context identifier indicates a sequence of words, such as a phrase or a text. From this sequence of words, one or more keywords are extracted, which act as the search criterion. For instance, the (prospective) user of a pattern recognition system specifies one or more documents representative of his interests. Keywords are extracted from the documents, and additional documents are selected based on the keywords. In this way, the user is relieved from choosing keywor- ds.
- the set of documents is formed by a document database or document file system.
- a large volume storage system such as a CD-ROM or DVD, containing a large and diverse set of documents may be supplied with the pattern recognition system, allowing the (prospective) user to select pertinent documents from this set.
- the set of documents is formed by documents in a distributed computer system. This allows for centrally storing (e.g. in a server) a larger set of documents than would normally be feasible to store or provide to a client computer on which the pattern recognition system is to be executed.
- a very large set of documents may be distributed over several servers.
- a good example of this last situation is Internet. Particularly if a system like Internet is used, many of the selected documents will reflect the language used at that moment, allowing for an up- to-date vocabulary and/or language model to be created.
- a network search engine like those commonly used on Internet, is used to identify relevant documents based on the search criteria supplied to the search engine.
- a network search agent which autonomously searches the distributed computer system based on the search criterion, is used to identify relevant documents and, optionally, for retrieving the documents.
- the training corpus is updated at a later moment selecting at least one further document from the set of documents and combining the further document with at least one previously selected document to form the training corpus. Particularly, if such updating is based on recent documents (e.g. retrieved via Internet), the language model can be kept up-to-date.
- the pattern recognition system is characterised in that the system comprises: means for determining at least one context identifier; means for deriving at least one search criterion from the context identifier; means for selecting documents from a set of documents based on the search criterion; and means for composing the training corpus from the selected documents.
- Figure 1 illustrates a speech recognition system
- Figure 2 shows Hidden Markov Models for modelling word or sub-word units
- Figure 3 illustrates a block diagram of the system according to the invention
- Figure 4 shows a further embodiment of the system
- Figure 5 shows the system operating in a distributed computer system
- Figure 3 illustrates a block diagram of a system 300 according to the invention.
- the system comprising means 310 for creating a vocabulary and/ or statistical language model from a textual training corpus.
- the created vocabulary and/or language model is stored using storing means 320.
- the system will normally be implemented on a computer, such as a PC or a workstation, and operated under control of a suitable program loaded in the processor of the computer.
- the output of the system (the vocabulary and/or the language model) is supplied to a pattern recognition system, for instance like the one illustrate in Fig. 1.
- the information in the storing means 320 may be loaded onto any removable storage medium, such as a CD-ROM, or DVD and reloaded into the pattern recognition system. It will be appreciated that the transfer may also occur via other means, such as a computer network.
- the system according to the invention is physically separate from the pattern recognition system. Such an approach may advantageously be used where the system according to the invention is operational at the site of a retailer, which creates the vocabulary and/or language model according to the wishes of a customer system and stores the output into the pattern recognition system acquired by the user.
- the system 300 is combined with the pattern recognition system 350 as shown in figure 3, where the vocabulary and/or the language model form the interface 330 between both sub-systems 300 and 350.
- the pattern recognition system (100, 350) is capable of recognising patterns representing language representative signals created by a person.
- An example of such signals are speech signals or handwriting signals.
- the pattern is a time-sequential pattern, although it will be appreciated that handwriting may also be supplied to the system as an image, wherein a detailed time sequential behaviour, which is present in an on-line handwriting recognition system, is lost.
- the input signal is analysed using the signal analysing subsystem 354 and recognised by the unit matching subsystem 356, giving as the output 358 a recognition result, for instance in the form of text or control instructions.
- the system 300 comprises means 335 for determining at least one context identifier.
- a user of the system is requested to supply the context identifier.
- the system further comprises means 332 means for deriving at least one search criterion from the context identifier.
- Means 334 are used for selecting documents from a set of documents based on the search criterion. The selection may be performed in any suitable way, for instance by successively opening each of the documents of the set and inspecting the contents of the document to determine whether the document meets the selection criterion. The selection may also be performed by checking descriptive attributes, such as keywords, of a document (if available) against the criterion.
- a document may be selected if one match with the criterion is fulfilled.
- a document is only selected if the level of matching meets a predetermined matching level.
- the matching level may be an absolute level, such as the criterion has to be matched a predetermined number of times, or a relative level, for instance related to the size of the document.
- Means 336 are used for composing the training corpus from the selected documents. The composition may simply involve combining all selected documents.
- the system 300 comes with a default vocabulary and/or language model which is updated using the selected documents. It will be appreciated that adding a new word to a vocabulary may, in itself, not be sufficient to ensure that the word can be recognised. For a speech recognition system a transcription in acoustic references is additionally required. For many languages, a reasonably accurate transcription can be achieved automatically for most words. By comparing a new word to words already in the vocabulary and having a transcription, a suitable transcription can be created.
- a phonetic transcription can be made for a word based on phonetic transcriptions of known words. Even if the transcription is of only moderate quality, the new word will be present in the vocabulary and, preferably, also in the language model. This allows recognition of the word (which otherwise would not be possible) and, with the assistance of the language model, the recognition of the word may be of an acceptable level in its context. Once the word has been recognised, the transcription can automatically be adapted to better match the actual utterance for which the word is recognised. Alternatively, the transcription can be improved with the assistance of the user of the speech recognition system in the form of an acoustical training.
- the context identifier is formed by one or more keywords, which act as the search criterion.
- the context identifier indicates a sequence of words, such as a phrase or a text.
- the user may enter via a user interface a string or indicate a text document.
- the system 300 comprises means 400 for extracting at least one keyword from the indicated sequence of words.
- the keywords act as the search criterion. Automatic methods for extracting keywords from a document are known in itself.
- the set of documents from which the documents are selected may be formed by a document file system, such as, for instance, used in computer systems. Using conventional documents, the selection can be performed by scanning the contents of the document.
- the set of documents is formed by a document database, such as a document management system.
- a document management system such as a document management system.
- attributes describing the contents of the documents can be used for the selection.
- the set of documents is formed by documents in a distributed computer system 500.
- the distributed system 500 may range from a group of local computers within one building or site of a company, connected via a local area network, to a world-wide network of computers of different companies, connected via a wide area network, such as Internet.
- the distributed system 500 comprises several document stores; shown are 510, 520 and 530. In Internet term, the stores are referred to as servers.
- the distributed computer system 500 may be able to accommodate very many systems like 300.
- the distributed computer system 500 also comprises at least one network search engine 540.
- the system 300 comprises communication means 550 for supplying the search criterion to the network search engine 540.
- the network search engine 540 searches the document stores connected to the network 505 for documents meeting the search criterion.
- Such search engines are well known, particularly for Internet.
- the network search engine 540 regularly scans the distributed system 500 to determine which documents are available and to extract attributes, such as keywords, from the documents. The outcome of the scan is stored in a database of the search engine 540. The search is then performed on the database.
- the communication means 550 is also used for receiving the result of the search back from the search engine 540.
- the composition means 336 composes the training corpus. This will usually include using the communication means 550 to acquire the document from the document store indicated by the search engine 540.
- the search engine 540 may already have supplied the document to the system 300 as a result of the search.
- the system 300 uses a network search agent to search through the stores 510, 520 and 530 of the network 505.
- the system 300 provides the search criterion to the search agent.
- the search agent autonomously searches stores in the network. Whenever a document fulfilling the search criterion is located the agent may deliver this to the system 300, for instance via regular e-mail.
- search agents are known, particularly for Internet.
- the agent may be active only in the system 300, where it in turn (or in parallel) accesses stores in the distributed system, which respond to queries of the agent.
- the agent may move through the distributed system, e.g. by hopping from one server to another, where the agent becomes active at the server it is 'visiting' at that moment.
- the system 300 is operative to update the training corpus by at a later moment selecting at least one further document from the set of documents and combining the further document with at least one previously selected document to form the training corpus.
- a search engine or a search agent is used to select further documents based on the same search criterion as before.
- the search criterion is updated based on documents recently recognised by the pattern recognition system, for instance by extracting keywords from recognised document(s).
- the updating may take place as a result of a direct instruction of a user. Alternatively, the updating may be autonomous and occur at regular moments.
Abstract
Description
Claims
Priority Applications (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
JP52133299A JP2001507482A (en) | 1997-10-08 | 1998-09-14 | Vocabulary and / or language model training |
EP98940501A EP0979497A1 (en) | 1997-10-08 | 1998-09-14 | Vocabulary and/or language model training |
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
EP97203124.9 | 1997-10-08 | ||
EP97203124 | 1997-10-08 |
Publications (2)
Publication Number | Publication Date |
---|---|
WO1999018556A2 true WO1999018556A2 (en) | 1999-04-15 |
WO1999018556A3 WO1999018556A3 (en) | 1999-06-24 |
Family
ID=8228802
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
PCT/IB1998/001421 WO1999018556A2 (en) | 1997-10-08 | 1998-09-14 | Vocabulary and/or language model training |
Country Status (4)
Country | Link |
---|---|
US (1) | US6430551B1 (en) |
EP (1) | EP0979497A1 (en) |
JP (1) | JP2001507482A (en) |
WO (1) | WO1999018556A2 (en) |
Cited By (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
EP1136982A2 (en) * | 2000-03-24 | 2001-09-26 | Philips Corporate Intellectual Property GmbH | Generation of a language model and an acoustic model for a speech recognition system |
US8335688B2 (en) | 2004-08-20 | 2012-12-18 | Multimodal Technologies, Llc | Document transcription system training |
US8412521B2 (en) | 2004-08-20 | 2013-04-02 | Multimodal Technologies, Llc | Discriminative training of document transcription system |
Families Citing this family (156)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US6904405B2 (en) * | 1999-07-17 | 2005-06-07 | Edwin A. Suominen | Message recognition using shared language model |
US7689416B1 (en) | 1999-09-29 | 2010-03-30 | Poirier Darrell A | System for transferring personalize matter from one computer to another |
US8645137B2 (en) | 2000-03-16 | 2014-02-04 | Apple Inc. | Fast, language-independent method for user authentication by voice |
US7243069B2 (en) * | 2000-07-28 | 2007-07-10 | International Business Machines Corporation | Speech recognition by automated context creation |
DE10100725C1 (en) * | 2001-01-10 | 2002-01-24 | Philips Corp Intellectual Pty | Automatic dialogue system for speech interrogation of databank entries uses speech recognition system assisted by speech model obtained before beginning of dialogue |
FR2820872B1 (en) * | 2001-02-13 | 2003-05-16 | Thomson Multimedia Sa | VOICE RECOGNITION METHOD, MODULE, DEVICE AND SERVER |
US6934675B2 (en) * | 2001-06-14 | 2005-08-23 | Stephen C. Glinski | Methods and systems for enabling speech-based internet searches |
US20030009331A1 (en) * | 2001-07-05 | 2003-01-09 | Johan Schalkwyk | Grammars for speech recognition |
JP2003036093A (en) * | 2001-07-23 | 2003-02-07 | Japan Science & Technology Corp | Speech input retrieval system |
AUPR824301A0 (en) * | 2001-10-15 | 2001-11-08 | Silverbrook Research Pty. Ltd. | Methods and systems (npw001) |
US7006972B2 (en) * | 2002-03-20 | 2006-02-28 | Microsoft Corporation | Generating a task-adapted acoustic model from one or more different corpora |
US7031918B2 (en) * | 2002-03-20 | 2006-04-18 | Microsoft Corporation | Generating a task-adapted acoustic model from one or more supervised and/or unsupervised corpora |
GB0224806D0 (en) * | 2002-10-24 | 2002-12-04 | Ibm | Method and apparatus for a interactive voice response system |
US7149688B2 (en) * | 2002-11-04 | 2006-12-12 | Speechworks International, Inc. | Multi-lingual speech recognition with cross-language context modeling |
US7584102B2 (en) * | 2002-11-15 | 2009-09-01 | Scansoft, Inc. | Language model for use in speech recognition |
JP3667332B2 (en) * | 2002-11-21 | 2005-07-06 | 松下電器産業株式会社 | Standard model creation apparatus and standard model creation method |
US20050004799A1 (en) * | 2002-12-31 | 2005-01-06 | Yevgenly Lyudovyk | System and method for a spoken language interface to a large database of changing records |
US7672845B2 (en) * | 2004-06-22 | 2010-03-02 | International Business Machines Corporation | Method and system for keyword detection using voice-recognition |
US7738705B2 (en) * | 2004-06-30 | 2010-06-15 | Stefano Casadei | Hierarchical method and system for pattern recognition and edge detection |
US7739286B2 (en) * | 2005-03-17 | 2010-06-15 | University Of Southern California | Topic specific language models built from large numbers of documents |
US7921011B2 (en) * | 2005-05-20 | 2011-04-05 | Sony Computer Entertainment Inc. | Structure for grammar and dictionary representation in voice recognition and method for simplifying link and node-generated grammars |
US7593908B2 (en) * | 2005-06-27 | 2009-09-22 | Microsoft Corporation | Training with heterogeneous data |
US8677377B2 (en) | 2005-09-08 | 2014-03-18 | Apple Inc. | Method and apparatus for building an intelligent automated assistant |
US7697827B2 (en) | 2005-10-17 | 2010-04-13 | Konicek Jeffrey C | User-friendlier interfaces for a camera |
US9318108B2 (en) | 2010-01-18 | 2016-04-19 | Apple Inc. | Intelligent automated assistant |
US7272558B1 (en) | 2006-12-01 | 2007-09-18 | Coveo Solutions Inc. | Speech recognition training method for audio and video file indexing on a search engine |
US8977255B2 (en) | 2007-04-03 | 2015-03-10 | Apple Inc. | Method and system for operating a multi-function portable electronic device using voice-activation |
US9330720B2 (en) | 2008-01-03 | 2016-05-03 | Apple Inc. | Methods and apparatus for altering audio output signals |
US8996376B2 (en) | 2008-04-05 | 2015-03-31 | Apple Inc. | Intelligent text-to-speech conversion |
US10496753B2 (en) | 2010-01-18 | 2019-12-03 | Apple Inc. | Automatically adapting user interfaces for hands-free interaction |
JP4962416B2 (en) * | 2008-06-03 | 2012-06-27 | 日本電気株式会社 | Speech recognition system |
US20100030549A1 (en) | 2008-07-31 | 2010-02-04 | Lee Michael M | Mobile device having human language translation capability with positional feedback |
US9959870B2 (en) | 2008-12-11 | 2018-05-01 | Apple Inc. | Speech recognition involving a mobile device |
US10706373B2 (en) | 2011-06-03 | 2020-07-07 | Apple Inc. | Performing actions associated with task items that represent tasks to perform |
US9858925B2 (en) | 2009-06-05 | 2018-01-02 | Apple Inc. | Using context information to facilitate processing of commands in a virtual assistant |
US10241644B2 (en) | 2011-06-03 | 2019-03-26 | Apple Inc. | Actionable reminder entries |
US10241752B2 (en) | 2011-09-30 | 2019-03-26 | Apple Inc. | Interface for a virtual digital assistant |
US9431006B2 (en) | 2009-07-02 | 2016-08-30 | Apple Inc. | Methods and apparatuses for automatic speech recognition |
US11416214B2 (en) | 2009-12-23 | 2022-08-16 | Google Llc | Multi-modal input on an electronic device |
EP3091535B1 (en) | 2009-12-23 | 2023-10-11 | Google LLC | Multi-modal input on an electronic device |
US10705794B2 (en) | 2010-01-18 | 2020-07-07 | Apple Inc. | Automatically adapting user interfaces for hands-free interaction |
US10679605B2 (en) | 2010-01-18 | 2020-06-09 | Apple Inc. | Hands-free list-reading by intelligent automated assistant |
US10553209B2 (en) | 2010-01-18 | 2020-02-04 | Apple Inc. | Systems and methods for hands-free notification summaries |
WO2011089450A2 (en) | 2010-01-25 | 2011-07-28 | Andrew Peter Nelson Jerram | Apparatuses, methods and systems for a digital conversation management platform |
US8682667B2 (en) | 2010-02-25 | 2014-03-25 | Apple Inc. | User profiling for selecting user specific voice input processing information |
US8463786B2 (en) | 2010-06-10 | 2013-06-11 | Microsoft Corporation | Extracting topically related keywords from related documents |
US10762293B2 (en) | 2010-12-22 | 2020-09-01 | Apple Inc. | Using parts-of-speech tagging and named entity recognition for spelling correction |
US9081760B2 (en) | 2011-03-08 | 2015-07-14 | At&T Intellectual Property I, L.P. | System and method for building diverse language models |
US9262612B2 (en) | 2011-03-21 | 2016-02-16 | Apple Inc. | Device access using voice authentication |
US9679561B2 (en) | 2011-03-28 | 2017-06-13 | Nuance Communications, Inc. | System and method for rapid customization of speech recognition models |
US10658074B1 (en) | 2011-04-11 | 2020-05-19 | Zeus Data Solutions, Inc. | Medical transcription with dynamic language models |
US10057736B2 (en) | 2011-06-03 | 2018-08-21 | Apple Inc. | Active transport based notifications |
US8756064B2 (en) | 2011-07-28 | 2014-06-17 | Tata Consultancy Services Limited | Method and system for creating frugal speech corpus using internet resources and conventional speech corpus |
US8994660B2 (en) | 2011-08-29 | 2015-03-31 | Apple Inc. | Text correction processing |
US10134385B2 (en) | 2012-03-02 | 2018-11-20 | Apple Inc. | Systems and methods for name pronunciation |
US9483461B2 (en) | 2012-03-06 | 2016-11-01 | Apple Inc. | Handling speech synthesis of content for multiple languages |
US9280610B2 (en) | 2012-05-14 | 2016-03-08 | Apple Inc. | Crowd sourcing information to fulfill user requests |
KR101359718B1 (en) | 2012-05-17 | 2014-02-13 | 포항공과대학교 산학협력단 | Conversation Managemnt System and Method Thereof |
US9721563B2 (en) | 2012-06-08 | 2017-08-01 | Apple Inc. | Name recognition system |
US9495129B2 (en) | 2012-06-29 | 2016-11-15 | Apple Inc. | Device, method, and user interface for voice-activated navigation and browsing of a document |
US9576574B2 (en) | 2012-09-10 | 2017-02-21 | Apple Inc. | Context-sensitive handling of interruptions by intelligent digital assistant |
US9547647B2 (en) | 2012-09-19 | 2017-01-17 | Apple Inc. | Voice-based media searching |
EP2954514B1 (en) | 2013-02-07 | 2021-03-31 | Apple Inc. | Voice trigger for a digital assistant |
US9368114B2 (en) | 2013-03-14 | 2016-06-14 | Apple Inc. | Context-sensitive handling of interruptions |
US10652394B2 (en) | 2013-03-14 | 2020-05-12 | Apple Inc. | System and method for processing voicemail |
WO2014144579A1 (en) | 2013-03-15 | 2014-09-18 | Apple Inc. | System and method for updating an adaptive speech recognition model |
AU2014233517B2 (en) | 2013-03-15 | 2017-05-25 | Apple Inc. | Training an at least partial voice command system |
US9582608B2 (en) | 2013-06-07 | 2017-02-28 | Apple Inc. | Unified ranking with entropy-weighted information for phrase-based semantic auto-completion |
WO2014197334A2 (en) | 2013-06-07 | 2014-12-11 | Apple Inc. | System and method for user-specified pronunciation of words for speech synthesis and recognition |
WO2014197336A1 (en) | 2013-06-07 | 2014-12-11 | Apple Inc. | System and method for detecting errors in interactions with a voice-based digital assistant |
WO2014197335A1 (en) | 2013-06-08 | 2014-12-11 | Apple Inc. | Interpreting and acting upon commands that involve sharing information with remote devices |
US10176167B2 (en) | 2013-06-09 | 2019-01-08 | Apple Inc. | System and method for inferring user intent from speech inputs |
DE112014002747T5 (en) | 2013-06-09 | 2016-03-03 | Apple Inc. | Apparatus, method and graphical user interface for enabling conversation persistence over two or more instances of a digital assistant |
KR101809808B1 (en) | 2013-06-13 | 2017-12-15 | 애플 인크. | System and method for emergency calls initiated by voice command |
AU2014306221B2 (en) | 2013-08-06 | 2017-04-06 | Apple Inc. | Auto-activating smart responses based on activities from remote devices |
US9620105B2 (en) | 2014-05-15 | 2017-04-11 | Apple Inc. | Analyzing audio input for efficient speech and music recognition |
US10592095B2 (en) | 2014-05-23 | 2020-03-17 | Apple Inc. | Instantaneous speaking of content on touch devices |
US9502031B2 (en) | 2014-05-27 | 2016-11-22 | Apple Inc. | Method for supporting dynamic grammars in WFST-based ASR |
US9760559B2 (en) | 2014-05-30 | 2017-09-12 | Apple Inc. | Predictive text input |
US9785630B2 (en) | 2014-05-30 | 2017-10-10 | Apple Inc. | Text prediction using combined word N-gram and unigram language models |
US9715875B2 (en) | 2014-05-30 | 2017-07-25 | Apple Inc. | Reducing the need for manual start/end-pointing and trigger phrases |
US10078631B2 (en) | 2014-05-30 | 2018-09-18 | Apple Inc. | Entropy-guided text prediction using combined word and character n-gram language models |
US10170123B2 (en) | 2014-05-30 | 2019-01-01 | Apple Inc. | Intelligent assistant for home automation |
US9842101B2 (en) | 2014-05-30 | 2017-12-12 | Apple Inc. | Predictive conversion of language input |
US10289433B2 (en) | 2014-05-30 | 2019-05-14 | Apple Inc. | Domain specific language for encoding assistant dialog |
US9734193B2 (en) | 2014-05-30 | 2017-08-15 | Apple Inc. | Determining domain salience ranking from ambiguous words in natural speech |
AU2015266863B2 (en) | 2014-05-30 | 2018-03-15 | Apple Inc. | Multi-command single utterance input method |
US9430463B2 (en) | 2014-05-30 | 2016-08-30 | Apple Inc. | Exemplar-based natural language processing |
US9633004B2 (en) | 2014-05-30 | 2017-04-25 | Apple Inc. | Better resolution when referencing to concepts |
US9396726B2 (en) * | 2014-06-26 | 2016-07-19 | Nvoq Incorporated | System and methods to create and determine when to use a minimal user specific language model |
US9338493B2 (en) | 2014-06-30 | 2016-05-10 | Apple Inc. | Intelligent automated assistant for TV user interactions |
US10659851B2 (en) | 2014-06-30 | 2020-05-19 | Apple Inc. | Real-time digital assistant knowledge updates |
US10446141B2 (en) | 2014-08-28 | 2019-10-15 | Apple Inc. | Automatic speech recognition based on user feedback |
US9818400B2 (en) | 2014-09-11 | 2017-11-14 | Apple Inc. | Method and apparatus for discovering trending terms in speech requests |
US10789041B2 (en) | 2014-09-12 | 2020-09-29 | Apple Inc. | Dynamic thresholds for always listening speech trigger |
US9886432B2 (en) | 2014-09-30 | 2018-02-06 | Apple Inc. | Parsimonious handling of word inflection via categorical stem + suffix N-gram language models |
US9646609B2 (en) | 2014-09-30 | 2017-05-09 | Apple Inc. | Caching apparatus for serving phonetic pronunciations |
US10127911B2 (en) | 2014-09-30 | 2018-11-13 | Apple Inc. | Speaker identification and unsupervised speaker adaptation techniques |
US9668121B2 (en) | 2014-09-30 | 2017-05-30 | Apple Inc. | Social reminders |
US10074360B2 (en) | 2014-09-30 | 2018-09-11 | Apple Inc. | Providing an indication of the suitability of speech recognition |
US10552013B2 (en) | 2014-12-02 | 2020-02-04 | Apple Inc. | Data detection |
US9711141B2 (en) | 2014-12-09 | 2017-07-18 | Apple Inc. | Disambiguating heteronyms in speech synthesis |
KR102325724B1 (en) * | 2015-02-28 | 2021-11-15 | 삼성전자주식회사 | Synchronization of Text Data among a plurality of Devices |
US9865280B2 (en) | 2015-03-06 | 2018-01-09 | Apple Inc. | Structured dictation using intelligent automated assistants |
US10567477B2 (en) | 2015-03-08 | 2020-02-18 | Apple Inc. | Virtual assistant continuity |
US9721566B2 (en) | 2015-03-08 | 2017-08-01 | Apple Inc. | Competing devices responding to voice triggers |
US9886953B2 (en) | 2015-03-08 | 2018-02-06 | Apple Inc. | Virtual assistant activation |
US9899019B2 (en) | 2015-03-18 | 2018-02-20 | Apple Inc. | Systems and methods for structured stem and suffix language models |
US9842105B2 (en) | 2015-04-16 | 2017-12-12 | Apple Inc. | Parsimonious continuous-space phrase representations for natural language processing |
US10083688B2 (en) | 2015-05-27 | 2018-09-25 | Apple Inc. | Device voice control for selecting a displayed affordance |
US10127220B2 (en) | 2015-06-04 | 2018-11-13 | Apple Inc. | Language identification from short strings |
US10101822B2 (en) | 2015-06-05 | 2018-10-16 | Apple Inc. | Language input correction |
US9578173B2 (en) | 2015-06-05 | 2017-02-21 | Apple Inc. | Virtual assistant aided communication with 3rd party service in a communication session |
US10186254B2 (en) | 2015-06-07 | 2019-01-22 | Apple Inc. | Context-based endpoint detection |
US11025565B2 (en) | 2015-06-07 | 2021-06-01 | Apple Inc. | Personalized prediction of responses for instant messaging |
US10255907B2 (en) | 2015-06-07 | 2019-04-09 | Apple Inc. | Automatic accent detection using acoustic models |
US10671428B2 (en) | 2015-09-08 | 2020-06-02 | Apple Inc. | Distributed personal assistant |
US10747498B2 (en) | 2015-09-08 | 2020-08-18 | Apple Inc. | Zero latency digital assistant |
US9697820B2 (en) | 2015-09-24 | 2017-07-04 | Apple Inc. | Unit-selection text-to-speech synthesis using concatenation-sensitive neural networks |
US11010550B2 (en) | 2015-09-29 | 2021-05-18 | Apple Inc. | Unified language modeling framework for word prediction, auto-completion and auto-correction |
US10366158B2 (en) | 2015-09-29 | 2019-07-30 | Apple Inc. | Efficient word encoding for recurrent neural network language models |
US11587559B2 (en) | 2015-09-30 | 2023-02-21 | Apple Inc. | Intelligent device identification |
US10691473B2 (en) | 2015-11-06 | 2020-06-23 | Apple Inc. | Intelligent automated assistant in a messaging environment |
US10049668B2 (en) | 2015-12-02 | 2018-08-14 | Apple Inc. | Applying neural network language models to weighted finite state transducers for automatic speech recognition |
US10223066B2 (en) | 2015-12-23 | 2019-03-05 | Apple Inc. | Proactive assistance based on dialog communication between devices |
US10446143B2 (en) | 2016-03-14 | 2019-10-15 | Apple Inc. | Identification of voice inputs providing credentials |
US9934775B2 (en) | 2016-05-26 | 2018-04-03 | Apple Inc. | Unit-selection text-to-speech synthesis based on predicted concatenation parameters |
US9972304B2 (en) | 2016-06-03 | 2018-05-15 | Apple Inc. | Privacy preserving distributed evaluation framework for embedded personalized systems |
US10249300B2 (en) | 2016-06-06 | 2019-04-02 | Apple Inc. | Intelligent list reading |
US10049663B2 (en) | 2016-06-08 | 2018-08-14 | Apple, Inc. | Intelligent automated assistant for media exploration |
DK179309B1 (en) | 2016-06-09 | 2018-04-23 | Apple Inc | Intelligent automated assistant in a home environment |
US10509862B2 (en) | 2016-06-10 | 2019-12-17 | Apple Inc. | Dynamic phrase expansion of language input |
US10490187B2 (en) | 2016-06-10 | 2019-11-26 | Apple Inc. | Digital assistant providing automated status report |
US10586535B2 (en) | 2016-06-10 | 2020-03-10 | Apple Inc. | Intelligent digital assistant in a multi-tasking environment |
US10067938B2 (en) | 2016-06-10 | 2018-09-04 | Apple Inc. | Multilingual word prediction |
US10192552B2 (en) | 2016-06-10 | 2019-01-29 | Apple Inc. | Digital assistant providing whispered speech |
DK179343B1 (en) | 2016-06-11 | 2018-05-14 | Apple Inc | Intelligent task discovery |
DK179415B1 (en) | 2016-06-11 | 2018-06-14 | Apple Inc | Intelligent device arbitration and control |
DK201670540A1 (en) | 2016-06-11 | 2018-01-08 | Apple Inc | Application integration with a digital assistant |
DK179049B1 (en) | 2016-06-11 | 2017-09-18 | Apple Inc | Data driven natural language event detection and classification |
US20180061408A1 (en) * | 2016-08-24 | 2018-03-01 | Semantic Machines, Inc. | Using paraphrase in accepting utterances in an automated assistant |
US10043516B2 (en) | 2016-09-23 | 2018-08-07 | Apple Inc. | Intelligent automated assistant |
US10824798B2 (en) | 2016-11-04 | 2020-11-03 | Semantic Machines, Inc. | Data collection for a new conversational dialogue system |
US10593346B2 (en) | 2016-12-22 | 2020-03-17 | Apple Inc. | Rank-reduced token representation for automatic speech recognition |
WO2018148441A1 (en) | 2017-02-08 | 2018-08-16 | Semantic Machines, Inc. | Natural language content generator |
US11069340B2 (en) | 2017-02-23 | 2021-07-20 | Microsoft Technology Licensing, Llc | Flexible and expandable dialogue system |
US10586530B2 (en) | 2017-02-23 | 2020-03-10 | Semantic Machines, Inc. | Expandable dialogue system |
US10762892B2 (en) | 2017-02-23 | 2020-09-01 | Semantic Machines, Inc. | Rapid deployment of dialogue system |
DK201770439A1 (en) | 2017-05-11 | 2018-12-13 | Apple Inc. | Offline personal assistant |
DK179496B1 (en) | 2017-05-12 | 2019-01-15 | Apple Inc. | USER-SPECIFIC Acoustic Models |
DK179745B1 (en) | 2017-05-12 | 2019-05-01 | Apple Inc. | SYNCHRONIZATION AND TASK DELEGATION OF A DIGITAL ASSISTANT |
DK201770432A1 (en) | 2017-05-15 | 2018-12-21 | Apple Inc. | Hierarchical belief states for digital assistants |
DK201770431A1 (en) | 2017-05-15 | 2018-12-20 | Apple Inc. | Optimizing dialogue policy decisions for digital assistants using implicit feedback |
DK179549B1 (en) | 2017-05-16 | 2019-02-12 | Apple Inc. | Far-field extension for digital assistant services |
US11132499B2 (en) | 2017-08-28 | 2021-09-28 | Microsoft Technology Licensing, Llc | Robust expandable dialogue system |
US11455907B2 (en) | 2018-11-27 | 2022-09-27 | International Business Machines Corporation | Adaptive vocabulary improvement |
Citations (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US4994983A (en) * | 1989-05-02 | 1991-02-19 | Itt Corporation | Automatic speech recognition system using seed templates |
US5195167A (en) * | 1990-01-23 | 1993-03-16 | International Business Machines Corporation | Apparatus and method of grouping utterances of a phoneme into context-dependent categories based on sound-similarity for automatic speech recognition |
US5390278A (en) * | 1991-10-08 | 1995-02-14 | Bell Canada | Phoneme based speech recognition |
US5621809A (en) * | 1992-06-09 | 1997-04-15 | International Business Machines Corporation | Computer program product for automatic recognition of a consistent message using multiple complimentary sources of information |
Family Cites Families (11)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US5317507A (en) * | 1990-11-07 | 1994-05-31 | Gallant Stephen I | Method for document retrieval and for word sense disambiguation using neural networks |
US5278980A (en) * | 1991-08-16 | 1994-01-11 | Xerox Corporation | Iterative technique for phrase query formation and an information retrieval system employing same |
US5742807A (en) * | 1995-05-31 | 1998-04-21 | Xerox Corporation | Indexing system using one-way hash for document service |
EP0788649B1 (en) * | 1995-08-28 | 2001-06-13 | Koninklijke Philips Electronics N.V. | Method and system for pattern recognition based on tree organised probability densities |
US5970490A (en) * | 1996-11-05 | 1999-10-19 | Xerox Corporation | Integration platform for heterogeneous databases |
US5924105A (en) * | 1997-01-27 | 1999-07-13 | Michigan State University | Method and product for determining salient features for use in information searching |
US5920856A (en) * | 1997-06-09 | 1999-07-06 | Xerox Corporation | System for selecting multimedia databases over networks |
US5933822A (en) * | 1997-07-22 | 1999-08-03 | Microsoft Corporation | Apparatus and methods for an information retrieval system that employs natural language processing of search results to improve overall precision |
US6067541A (en) * | 1997-09-17 | 2000-05-23 | Microsoft Corporation | Monitoring document changes in a file system of documents with the document change information stored in a persistent log |
CA2216224A1 (en) * | 1997-09-19 | 1999-03-19 | Peter R. Stubley | Block algorithm for pattern recognition |
WO1999021172A2 (en) * | 1997-10-20 | 1999-04-29 | Koninklijke Philips Electronics N.V. | Pattern recognition enrolment in a distributed system |
-
1998
- 1998-09-14 JP JP52133299A patent/JP2001507482A/en not_active Withdrawn
- 1998-09-14 EP EP98940501A patent/EP0979497A1/en not_active Ceased
- 1998-09-14 WO PCT/IB1998/001421 patent/WO1999018556A2/en not_active Application Discontinuation
- 1998-10-06 US US09/167,424 patent/US6430551B1/en not_active Expired - Fee Related
Patent Citations (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US4994983A (en) * | 1989-05-02 | 1991-02-19 | Itt Corporation | Automatic speech recognition system using seed templates |
US5195167A (en) * | 1990-01-23 | 1993-03-16 | International Business Machines Corporation | Apparatus and method of grouping utterances of a phoneme into context-dependent categories based on sound-similarity for automatic speech recognition |
US5390278A (en) * | 1991-10-08 | 1995-02-14 | Bell Canada | Phoneme based speech recognition |
US5621809A (en) * | 1992-06-09 | 1997-04-15 | International Business Machines Corporation | Computer program product for automatic recognition of a consistent message using multiple complimentary sources of information |
Cited By (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
EP1136982A2 (en) * | 2000-03-24 | 2001-09-26 | Philips Corporate Intellectual Property GmbH | Generation of a language model and an acoustic model for a speech recognition system |
EP1136982A3 (en) * | 2000-03-24 | 2004-03-03 | Philips Intellectual Property & Standards GmbH | Generation of a language model and an acoustic model for a speech recognition system |
US8335688B2 (en) | 2004-08-20 | 2012-12-18 | Multimodal Technologies, Llc | Document transcription system training |
US8412521B2 (en) | 2004-08-20 | 2013-04-02 | Multimodal Technologies, Llc | Discriminative training of document transcription system |
Also Published As
Publication number | Publication date |
---|---|
EP0979497A1 (en) | 2000-02-16 |
JP2001507482A (en) | 2001-06-05 |
US6430551B1 (en) | 2002-08-06 |
WO1999018556A3 (en) | 1999-06-24 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US6430551B1 (en) | Vocabulary and/or language model training | |
EP0980574B1 (en) | Pattern recognition enrolment in a distributed system | |
EP1171871B1 (en) | Recognition engines with complementary language models | |
EP1349145B1 (en) | System and method for providing information using spoken dialogue interface | |
US6073098A (en) | Method and apparatus for generating deterministic approximate weighted finite-state automata | |
EP1181684B1 (en) | Client-server speech recognition | |
Bennacef et al. | A spoken language system for information retrieval. | |
US5930746A (en) | Parsing and translating natural language sentences automatically | |
EP1429313B1 (en) | Language model for use in speech recognition | |
US7016830B2 (en) | Use of a unified language model | |
US7162423B2 (en) | Method and apparatus for generating and displaying N-Best alternatives in a speech recognition system | |
US20010041977A1 (en) | Information processing apparatus, information processing method, and storage medium | |
WO2000058946A1 (en) | Client-server speech recognition | |
JP2001101187A (en) | Device and method for translation and recording medium | |
WO2000033211A2 (en) | Automatic segmentation of a text | |
US20100153366A1 (en) | Assigning an indexing weight to a search term | |
WO2001093246A2 (en) | Creating a unified task dependent language models with information retrieval techniques | |
US6345249B1 (en) | Automatic analysis of a speech dictated document | |
US20050267755A1 (en) | Arrangement for speech recognition | |
JPH08248980A (en) | Voice recognition device | |
EP1055228A1 (en) | Speech operated automatic inquiry system | |
EP1135768B1 (en) | Spell mode in a speech recognizer | |
Waibel et al. | Recent advances in JANUS: A speech translation system | |
Yamamoto et al. | Out‐of‐vocabulary word recognition using a hierarchical language model based on multiple Markov models |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
AK | Designated states |
Kind code of ref document: A2 Designated state(s): JP |
|
AL | Designated countries for regional patents |
Kind code of ref document: A2 Designated state(s): AT BE CH CY DE DK ES FI FR GB GR IE IT LU MC NL PT SE |
|
WWE | Wipo information: entry into national phase |
Ref document number: 1998940501 Country of ref document: EP |
|
ENP | Entry into the national phase |
Ref country code: JP Ref document number: 1999 521332 Kind code of ref document: A Format of ref document f/p: F |
|
121 | Ep: the epo has been informed by wipo that ep was designated in this application | ||
AK | Designated states |
Kind code of ref document: A3 Designated state(s): JP |
|
AL | Designated countries for regional patents |
Kind code of ref document: A3 Designated state(s): AT BE CH CY DE DK ES FI FR GB GR IE IT LU MC NL PT SE |
|
WWP | Wipo information: published in national office |
Ref document number: 1998940501 Country of ref document: EP |
|
WWR | Wipo information: refused in national office |
Ref document number: 1998940501 Country of ref document: EP |
|
WWW | Wipo information: withdrawn in national office |
Ref document number: 1998940501 Country of ref document: EP |