US20030115060A1 - System and interactive form filling with fusion of data from multiple unreliable information sources - Google Patents

System and interactive form filling with fusion of data from multiple unreliable information sources Download PDF

Info

Publication number
US20030115060A1
US20030115060A1 US10245144 US24514402A US20030115060A1 US 20030115060 A1 US20030115060 A1 US 20030115060A1 US 10245144 US10245144 US 10245144 US 24514402 A US24514402 A US 24514402A US 20030115060 A1 US20030115060 A1 US 20030115060A1
Authority
US
Grant status
Application
Patent type
Prior art keywords
information
user
input
based
system
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Abandoned
Application number
US10245144
Inventor
Jean-claude Junqua
Kirill Stoimenov
Roland Kuhn
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Panasonic Corp
Original Assignee
Panasonic Corp
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date

Links

Images

Classifications

    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS OR SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING; SPEECH OR AUDIO CODING OR DECODING
    • G10L15/00Speech recognition
    • G10L15/24Speech recognition using non-acoustical features

Abstract

An automated form filling system includes an input receptive of a plurality of information inputs from a plurality of information sources. An information fuser is operable to select information from the plurality of information inputs based on a comparison of the information inputs, and based on knowledge relating to reliability of the information sources. A form filler is operable to fill an electronic form with the selected information.

Description

    CROSS-REFERENCE TO RELATED APPLICATIONS
  • This application is a continuation-in-part of U.S. patent application Ser. No. 10/022,023 filed on Dec. 13, 2001. The disclosure of the above application is incorporated herein by reference.[0001]
  • FIELD OF THE INVENTION
  • The present invention generally relates to speech recognition and particularly relates to automated form filling over a telephone system. [0002]
  • BACKGROUND OF THE INVENTION
  • Automatic electronic form filling by a user, particularly over a telephone system, is a notoriously laborious and error prone process. Use of numerical keypad entries to attempt retrieval of user information is one existing process that proves to be only as reliable as completeness and correctness of the user information database and user knowledge of the required information and successful operation of the telephone keypad. Other applications, such as confirming availability and/or delivery of a product over the phone or on a networked computer system, suffer from similar problems. [0003]
  • The key to improving the automated form filling process, by increasing reliability of information and decreasing effort on the part of a user, involves recognizing that information from multiple information sources can be fused in an intelligent manner. To be successful, the information fusion process should not trust the information sources to be reliable, and should intelligently use information from the multiple sources to constrain and supplement one another based on differences in reliability between sources and of particular information inputs. To be further successful, the fusion process should be able to incorporate a dialogue with a user to increase knowledge relating to reliability of information content, and/or gather additional information inputs. [0004]
  • Automated form filling processes do not currently succeed in intelligently fusing information from multiple information sources based on knowledge relating to differences in reliability information form different sources, thereby simultaneously increasing reliability of form contents and decreasing effort on the part of the user. The need remains, therefore, for a solution to the problems associated with automated form filling as detailed above. The present invention provides such a solution. [0005]
  • SUMMARY OF THE INVENTION
  • In accordance with the present invention, an automated form filling system includes an input receptive of a plurality of information inputs from a plurality of information sources. An information fuser is operable to select information from the plurality of information inputs based on a comparison of the information inputs, and based on knowledge relating to reliability of the information sources. A form filler is operable to fill an electronic form with the selected information. [0006]
  • The form filling system of the present invention is advantageous over previous form filling systems in that it has knowledge relating to reliability of multiple information sources and is able to compare and select information content accordingly. It is further advantageous in its further aspects, wherein a prompt formulator is capable of formulating a prompt based on a comparison of the information content of the sources, and based on a level of knowledge relating to reliability of the source content, wherein the formulated prompt is designed to elicit a response from a user operable to increase knowledge of reliability relating to the multiple information sources. A resulting dialog with a user succeeds in filling out an electronic form while simultaneously decreasing effort on the user's part and increasing reliability of the form contents. [0007]
  • The present invention is particularly suitable for use with electronic form filling over a telephone, wherein user input and accessible databases are unreliable information sources. Further areas of applicability of the present invention will become apparent from the detailed description provided hereinafter. It should be understood that the detailed description and specific examples, while indicating the preferred embodiment of the invention, are intended for purposes of illustration only and are not intended to limit the scope of the invention.[0008]
  • BRIEF DESCRIPTION OF THE DRAWINGS
  • The present invention will become more fully understood from the detailed description and the accompanying drawings, wherein: [0009]
  • FIG. 1 is an illustrated block diagram depicting a general implementation of the present invention with a telephone call center; [0010]
  • FIG. 2 is a flow chart depicting the method of the present invention; [0011]
  • FIG. 3 is a block diagram depicting a form filling system according to the present invention; and [0012]
  • FIG. 4 is a partial block and information flow diagram depicting a detailed implementation of the present invention with a telephone call center.[0013]
  • DETAILED DESCRIPTION OF THE PREFERRED EMBODIMENTS
  • The following description of the preferred embodiment(s) is merely exemplary in nature and is in no way intended to limit the invention, its application, or uses. In a preferred embodiment, the present invention is employed to perform form filling by fusing multimodal user input over a telephone with one or more user information databases. The form filling system [0014] 100 of FIG. 1 exemplifies one implementation of the preferred embodiment of the present invention to fill out a form requiring a user's name, address, and telephone number.
  • According to form filling system [0015] 100, a user, one Mr. Baker, is initially prompted to speak his name into the telephone receiver as at 102, and to spell his name via the telephone keypad as at 104. The speech input 106 is communicated to an automatic speech recognizer 108, whereas the keypad entry 110 is communicated to an information database 112 of names indexed by predefined classes defined by the telephone keypad. In turn, a constraint list 114 of candidate names is generated from the keypad entry 110 and the information database 112 of names, and the generated constraint list 114 is further communicated to the speech recognizer 108. The speech recognizer functions to recognize the speech input 106 by generating a plurality of speech recognition hypotheses, and then selects the N best generated hypotheses by comparing them to the constraint list 114. Thus, if “Bater” were one of the original speech recognition hypotheses, it would be discarded based on it's absence from the constraint list.
  • The information database [0016] 112 further serves as an information source having the names and addresses of most residents of the nation in which the user, Mr. Baker, is located, so frequency information 116 relating to the frequency with which names appear in the database can be communicated to rescoring module 118 used to further rescore the N best speech recognition hypotheses. Thus, if “Baker” and “Bakes” are both present in the N best speech recognition hypotheses, they can be rescored to increase the ranking of “Baker” with respect to the ranking of “Bakes” based on a higher frequency of “Baker” in the information database 112 compared to the frequency of “Bakes”.
  • The user is additionally prompted to enter his or her postal code, such as a zip code, and the received postal code [0017] 120, the rescored speech recognition hypotheses, and the information database 112 are communicated to an information fuser 122. The information fuser 122, in turn, selects information content of the information database 112 based on the rescored recognition hypotheses and the received zip code 120. The information fuser 122, however, does not merely trust the database 112, the rescored hypotheses, and the received zip code 120 to be accurate. Instead, the information fuser 122 recognizes and adjusts for several potential causes of unreliability.
  • The data communicated to information fuser [0018] 122 may be unreliable for several reasons. For example, the user's name may be spoken in such a way that the speech recognition system misrecognizes it. Also, the customer may mistype his or her postal code. Further, the customer database may be incomplete (address missing) or incorrect (name or address misspelled, information out of date). For these and similar reasons, the information fuser employs an information selection strategy that selects the most reliable information available.
  • The information fuser [0019] 122 selects the most reliable information available based on knowledge relating to reliability of various types of information sources, and based on a comparison of the information content from the different sources. For example, if the top-ranking speech recognition hypothesis does not match any name in the information database 112 having the received zip code 120, but the second-highest ranking speech recognition hypothesis has only a slightly lower score than the top-ranking hypothesis and does match a name in the information database 112 having the received zip code 120, then the information fuser can select the name indicated by the second-highest ranking speech recognition hypothesis and prompt the user for confirmation. Similarly, if the top-ranking speech recognition hypothesis does not match any name in the information database 112 having the received postal code 120, but has a much higher score than the second-highest ranking speech recognition hypothesis, then the information fuser 122 can select the name and address matching the highest ranked speech recognition hypothesis and/or prompt the user to reenter the postal code or confirm whether the postal code is correct. Further, the information fuser may take a different approach by distinguishing between an entirely incorrect postal code and one that is only partially incorrect, and further consider the first two digits of the postal code (in the case of a zip code) more reliable than the last three digits of the postal code.
  • Form filling system [0020] 100 preferably has an interviewer 123 for implementing a strategy to prompt the user for input increasing knowledge relating to reliability of information content. The interviewer 123 can be and/or incorporate a human agent to assist in confirming, correcting, selecting, and/or supplementing information. Alternatively, the interviewer 123 can be partially or wholly automated using a prompt formulator to request different inputs from the user in different situations. Accordingly, information fuser 122 can select, deselect, and/or reselect information content based on the increased knowledge relating to the reliability of the information content, and the user responses constitute new information sources to be used in the fusion process. Thus, the prompt formulator can preferably ask the user to supply supplemental information for fields of the form 124 for which reliable information cannot be obtained with the information sources currently available. Form filling system 100 also has a form filler for filling appropriate fields of the electronic form 124 with the selected information content.
  • The method [0021] 200 of the present invention is illustrated in FIG. 2, and begins at 202. Thence, the method 200 proceeds to step 203, wherein information inputs are sought, preferably by initially prompting a user for specific inputs where a user is an applicable information source. Thence, the method proceeds to step 204, wherein multiple information inputs are received from multiple information sources. In addition to information inputs provided by a user, these information inputs may also include data from an information database, and/or additional data such as that provided by caller ID or a biometric (measured physical characteristic: fingerprint, retina scan, voice pattern, DNA, etc.) of a user. Thence, the method 200 proceeds to step 206, wherein information content is selected from one or more of the information sources based on a comparison of source contents and knowledge relating to reliability of the information sources. In one aspect, the knowledge of reliability of a source relates to the type of source, and thus is prior knowledge. In another aspect, however, the knowledge relating to reliability of an information source and/or specific information contents stems from a comparison of the contents of the information sources. Thus, content of an information source of a reliable type may be deemed less reliable based on comparison with content of another information source when the information content conflicts. Similarly, content of an information source of an unreliable type may be deemed more reliable based on comparison with content of another information source when the information content matches. This process can be used to identify more and less reliable portions of information content within one or more information sources.
  • Once selection of information has taken place, method [0022] 200 proceeds to step 208, wherein an electronic form is filled with the selected information. If the form is deemed reliably completed as at 210, then the method ends at 214 and a filled form has been generated. On the other hand, some or all of the selected information may be deemed insufficiently reliable, and/or a sufficient amount of fields of the form may not be deemed completed. In either of these latter cases, method 200 returns to step 203.
  • In step [0023] 203, the user is prompted for additional information inputs, and the prompt is designed to elicit a response to increase knowledge of reliability relating to selectable information and/or gather supplemental information. Thus, the request for additional sources may be a request for confirmation of selected information, a request to reenter one of the user inputs, a request for a different information type, or a similar type of request. With this step, speech generation is preferred to communicate the request, especially over a telephone system. Following, the prompt for additional information inputs, the method 200 returns to step 204, wherein the response is received. The method then proceeds again to step 206, wherein new information content is selected and/or it is determined whether selected content is reliable based on the new information sources. The method 200 further proceeds to step 208 and fills the form based on the revised selection. If the newly filled form is deemed reliably completed at step 210, then the method 200 ends at 216. Otherwise, processing continues in a recursive fashion until a reliably completed form is obtained or the process is otherwise interrupted.
  • A form filling system [0024] 300 of the present invention is more generally illustrated in FIG. 3, wherein a first information input 302 from a first information source and a second information input 304 from a second information source are received by input 306 and communicated to information fuser 308. Information fuser 308 compares and selects information based on reliability of the information as determined based on the comparison and in accordance with predefined rules developed with regard to reliability of different types of information sources and/or information inputs. The selected information 310 is communicated to form filler 312, which fills an electronic form 314 with the selected information to generate a filled form 316 and communicate it to an outside system via output 316.
  • Information fuser [0025] 308 mutually communicates with prompt formulator 320, and prompt formulator 320 formulates a prompt 322 for additional information sources, if needed, based on reliability of the information sources, comparative and/or objective reliability of the available information inputs, and/or requirements for supplemental information content. Prompt formulator 320 further communicates its current state to information fuser 308 so that information fuser 308 is aware of the type of information input(s) requested and how to interpret its information content in view of the other information inputs. The formulated prompt 322 is communicated to a dialogue manager 324 that generates a prompt in a manner communicable to and understandable by a user, preferably by speech generation. The generated prompt is communicated to the user via output 328. A response from the user constitutes an additional information input communicable to information fuser 308 via input 306.
  • A detailed implementation of the present invention with a telephone call center is described with reference to FIG. 4. Therein, multi-modal information inputs [0026] 400 from an information source corresponding to a user include a user speech input 402 and a user keypad entry 404. The user speech input includes a spoken user name, and the user keypad entry includes a spelling and/or initials of the user name and a zip code for the user location. Also, textual information inputs 406 from database information sources include nickname data input 408 from a nickname database and personal data input 410 from a personal information database. Use of nickname data input 408 in the present detailed implementation represents an improvement over the more simplified implementation previously discussed with reference to FIG. 1. Further, additional data inputs 412 that also represent an improvement include telephone subscriber data 414 provided by a caller ID service. Other examples of additional data inputs 412 that can also be used include biometrics identifying a user and gathered, for example, by a handheld device the user employs to communicate with the call center. Further examples include an IP address for the user location, assuming the user communicates with the call center using a computer network. These types of information inputs all assist to varying degrees in identifying the caller, and the form filling system of the present invention is adapted to use some or all of these types of information inputs accordingly.
  • Form filling system [0027] 100 receives the multi-modal information inputs 400, textual information inputs 406, and additional data inputs 412 and uses the various information inputs to constrain and supplement one another according to their varying modalities, utilities, and levels of reliability. For example, user speech input 402 is processed by speech recognizer 108A to produce a plurality of speech recognition hypotheses 414. Also, user keypad entry 404, additional data inputs 412, nickname data input 408, and personal data input 410 are all communicated to constraint list generator 416.
  • Constraint list generator [0028] 416 is adapted in the present implementation to generate a constraint list of candidate names by using a user keypad entry 404 containing a first and last name is used to access the nickname database and generate a plurality of first names based on the input first name, such that a keypad entry generating a constraint list containing “Bob” will also contain “Robert”, “Robby”, “Bobby”, and so on. The plurality of first names thus generated, a zip code from the user keypad entry, and any matching caller ID information are then used to access the personal information database, such as Phonedisc, containing names and addresses of all telephone subscribes in the United States, to generate a constraint list 114 that is communicated to N best hypotheses generator 108B.
  • The speech recognition hypotheses [0029] 414 are processes by N best hypotheses generator 108B to generate an N best list of speech recognition hypotheses 418. One skilled in the art of speech recognition will recognize that an alternative embodiment may be realized with a word lattice. The N best list of speech recognition hypotheses 418 is communicated to rescoring mechanism 118, as is data from the personal information database pertaining to frequency of appearance of names in the personal information database. Rescoring mechanism 118 rescores the N best list of speech recognition hypotheses 418 to generate a list of rescored hypotheses 420. Caller ID information and a postal code portion of the user keypad entry 404 may alternatively and/or additionally be used during this process to rescore hypotheses based on the frequency of names in the database having the corresponding postal code and/or names matching the caller ID information.
  • The rescored hypotheses [0030] 420, caller ID information, postal code, and personal data input 410 are communicated to the information fuser 122 and the prompt formulator 320. The information fuser selects most reliable information from the personal data input 410, caller ID information, and rescored hypotheses 420, and the selected data 310 is communicated to the prompt formulator 320 and to form filler 312. Form filler fills electronic form 314 to produce a filled form 316 that is partially or wholly filled with the selected data. Meanwhile, the prompt formulator 320 determines whether more information is needed based on the selected data, the information available for selection, and requirements of the electronic form. If more information inputs are required, the prompt formulator formulates an appropriate prompt 322 and communicates a current state 422 to information fuser 122. Otherwise, the form is deemed completed.
  • The description of the invention is merely exemplary in nature and, thus, variations that do not depart from the gist of the invention are intended to be within the scope of the invention. For example, one may recognize that information fusion according to the present invention occurs at several levels and at several points in the information selection process. Thus, information fusion is used in generating a constraint list, and in altering confidence scores associated with speech recognition hypotheses. A multi-layered information fusion-based form filling system is thus within the scope of the present invention, and various embodiments may be realized with respect to various types of available inputs, various modalities of input, and various applications of form filling. Such variations are not to be regarded as a departure from the spirit and scope of the invention. [0031]

Claims (27)

    What is claimed is:
  1. 1. An automated form filling system comprising:
    an input receptive of a plurality of information inputs from a plurality of information sources, including a first information input from a first information source and a second information input from a second information source;
    an information fuser operable to select information from at least one of the first information input and the second information input based on a comparison of the first information input and the second information input, and based on knowledge relating to reliability of the first information source and the second information source; and
    a form filler operable to fill an electronic form with the selected information.
  2. 2. The system of claim 1, wherein said information fuser is operable to select information from at least one of the first information input and the second information input based on knowledge relating to reliability of the first information input and the second information input.
  3. 3. The system of claim 1 comprising a prompt formulator operable to formulate a prompt designed to elicit a response from a user operable to increase knowledge relating to reliability of selectable information.
  4. 4. The system of claim 3 comprising a dialog manager operable to receive the formulated prompt and communicate the formulated prompt to a user in a manner so as to be readily understood by the user.
  5. 5. The system of claim 4, wherein said dialog manager operable to generate speech based on the formulated prompt and to communicate the generated speech to the user.
  6. 6. The system of claim 3, wherein said prompt formulator includes a human performing a review of at least one of the selected information and the plurality of information inputs, formulating the prompt based on the review, and at least one of confirming and correcting the selected information entered in the filled form based on the elicited responses.
  7. 7. The system of claim 3, wherein said prompt formulator is operable to formulate the prompt based on a comparison of the first information input and the second information input, and based on knowledge relating to reliability of the first information source and the second information source.
  8. 8. The system of claim 1, wherein the first information source corresponds to the user, and the second information source corresponds to an information database.
  9. 9. The system of claim 8, wherein the first information input corresponds to at least one of a user speech input and a user keypad entry.
  10. 10. The system of claim 9, wherein the first information input corresponds to the user speech input, wherein said plurality of information inputs includes a third information input from the user, wherein the third information input corresponds to a user keypad entry.
  11. 11. The system of claim 1, wherein said information fuser is adapted to yield numerical scores for various portions of information of the second information input based on a comparison of the first information input and the second information input.
  12. 12. The system of claim 1, wherein said information fuser is adapted to yield numerical scores for various portions of information of the second information input based on knowledge relating to reliability of various portions of information of the second information input.
  13. 13. The system of claim 1, wherein the first information source corresponds to a user, and the first information input corresponds to a measured physical characteristic of the user.
  14. 14. The system of claim 1, wherein the first information source corresponds to a user providing the first information input via telephone at a user location, and wherein the second information source corresponds to an automated caller identification system providing at least one of a phone number at the user location and a name associated with the user location.
  15. 15. The system of claim 1, wherein the first information source corresponds to a user, and wherein the first information input corresponds to a user handwritten entry.
  16. 16. The system of claim 1, wherein the first information source corresponds to a user, the first information input corresponds to a name of a user, the second information source corresponds to a database of names and other personal data of potential users, and the information fuser is operable to make a comparison of the name of the user and the names of potential users, and operable to select at least one of a name and associated personal data from the database based on the comparison.
  17. 17. A data fusion-based method of automated form filling comprising:
    receiving a plurality of information inputs from a plurality of information sources;
    selecting information from the plurality of information inputs based on a comparison of the information inputs and knowledge relating to reliability of the information sources; and
    filling an electronic form with the selected information.
  18. 18. The method of claim 17, wherein said selecting information from the plurality of information inputs is based on knowledge relating to reliability of the information inputs.
  19. 19. The method of claim 17, wherein said receiving a plurality of information inputs from a plurality of information sources corresponds to receiving a user input from a user.
  20. 20. The method of claim 19 comprising:
    formulating a prompt designed to elicit a new information input from the user; and
    prompting the user for the new information input.
  21. 21. The method of claim 19, wherein said formulating the prompt includes formulating a prompt designed to elicit a new information input from the user that increases knowledge relating to reliability of selectable information.
  22. 22. The method of claim 19, wherein said formulating the prompt includes formulating a prompt designed to elicit a new information input from the user that supplements selectable information.
  23. 23. The method of claim 19 comprising selecting unselected information based on the new information input.
  24. 24. The method of claim 19 comprising deselecting selected information based on the new information input.
  25. 25. The method of claim 19 comprising comparing the new information input to the information inputs previously received.
  26. 26. A data fusion-based form filling system for filling an electronic form based on user input communicated over a telephone system, comprising:
    an input receptive of personal information input by a user via a telephone system using input speech and a keypad entry;
    a speech recognizer operable to recognize the personal information based on the input speech, the keypad entry, and knowledge relating to reliability of the input speech and the keypad entry;
    an electronic form filler operable to access a database of stored personal information for potential users, operable to make a comparison between the recognized personal information and the stored personal information, operable to select personal information for the user based on the comparison and based on knowledge relating to reliability of the recognized personal information and the stored personal information, and operable to fill an electronic form with the selected personal information.
  27. 27. The system of claim 26 comprising an interviewer operable to prompt the user for additional input required to assist in at least one of:
    (a) selecting the personal information;
    (b) confirming the selected personal information;
    (c) correcting the selected personal information; and
    (d) supplementing the selected personal information.
US10245144 2001-12-13 2002-09-16 System and interactive form filling with fusion of data from multiple unreliable information sources Abandoned US20030115060A1 (en)

Priority Applications (2)

Application Number Priority Date Filing Date Title
US10022023 US7124085B2 (en) 2001-12-13 2001-12-13 Constraint-based speech recognition system and method
US10245144 US20030115060A1 (en) 2001-12-13 2002-09-16 System and interactive form filling with fusion of data from multiple unreliable information sources

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
US10245144 US20030115060A1 (en) 2001-12-13 2002-09-16 System and interactive form filling with fusion of data from multiple unreliable information sources
US10392549 US7240007B2 (en) 2001-12-13 2003-03-20 Speaker authentication by fusion of voiceprint match attempt results with additional information

Related Parent Applications (1)

Application Number Title Priority Date Filing Date
US10022023 Continuation-In-Part US7124085B2 (en) 2001-12-13 2001-12-13 Constraint-based speech recognition system and method

Related Child Applications (1)

Application Number Title Priority Date Filing Date
US10392549 Continuation-In-Part US7240007B2 (en) 2001-12-13 2003-03-20 Speaker authentication by fusion of voiceprint match attempt results with additional information

Publications (1)

Publication Number Publication Date
US20030115060A1 true true US20030115060A1 (en) 2003-06-19

Family

ID=21807419

Family Applications (2)

Application Number Title Priority Date Filing Date
US10022023 Active 2024-10-30 US7124085B2 (en) 2001-12-13 2001-12-13 Constraint-based speech recognition system and method
US10245144 Abandoned US20030115060A1 (en) 2001-12-13 2002-09-16 System and interactive form filling with fusion of data from multiple unreliable information sources

Family Applications Before (1)

Application Number Title Priority Date Filing Date
US10022023 Active 2024-10-30 US7124085B2 (en) 2001-12-13 2001-12-13 Constraint-based speech recognition system and method

Country Status (4)

Country Link
US (2) US7124085B2 (en)
EP (1) EP1326234A3 (en)
JP (1) JP2003216181A (en)
CN (1) CN1238832C (en)

Cited By (71)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20050288934A1 (en) * 2004-06-29 2005-12-29 Canon Kabushiki Kaisha Multimodal input method
US20060009974A1 (en) * 2004-07-09 2006-01-12 Matsushita Electric Industrial Co., Ltd. Hands-free voice dialing for portable and remote devices
US20060074652A1 (en) * 2004-09-20 2006-04-06 International Business Machines Corporation Method and system for voice-enabled autofill
US20060136564A1 (en) * 2004-11-19 2006-06-22 W.A. Krapf, Inc. Bi-directional communication between a web client and a web server
US20060167684A1 (en) * 2005-01-24 2006-07-27 Delta Electronics, Inc. Speech recognition method and system
US20060167685A1 (en) * 2002-02-07 2006-07-27 Eric Thelen Method and device for the rapid, pattern-recognition-supported transcription of spoken and written utterances
US20060276916A1 (en) * 2004-12-22 2006-12-07 Dearing Stephen M System and method for electronically processing address information
US20070094155A1 (en) * 2005-05-17 2007-04-26 Dearing Stephen M System and method for automated management of an address database
US20080281582A1 (en) * 2007-05-11 2008-11-13 Delta Electronics, Inc. Input system for mobile search and method therefor
US7502835B1 (en) * 2004-11-17 2009-03-10 Juniper Networks, Inc. Virtual folders for tracking HTTP sessions
US20090187410A1 (en) * 2008-01-22 2009-07-23 At&T Labs, Inc. System and method of providing speech processing in user interface
US20090248415A1 (en) * 2008-03-31 2009-10-01 Yap, Inc. Use of metadata to post process speech recognition output
US20100057713A1 (en) * 2008-09-03 2010-03-04 International Business Machines Corporation Entity-driven logic for improved name-searching in mixed-entity lists
US20100070360A1 (en) * 2008-09-13 2010-03-18 At&T Intellectual Property I, L.P. System and method for creating a speech search platform for coupons
US20110169906A1 (en) * 2010-01-13 2011-07-14 Seizo Suzuki Optical scanning device and image forming apparatus
US8024347B2 (en) 2007-09-27 2011-09-20 International Business Machines Corporation Method and apparatus for automatically differentiating between types of names stored in a data collection
US20110264994A1 (en) * 2010-04-13 2011-10-27 ForenSight, Inc. Method of retrieving interactive electronic data and accessing productivity tools for real property valuation and mortgage underwriting software systems
US20110282653A1 (en) * 2008-12-26 2011-11-17 Akihiro Tamura Text processing apparatus, text processing method, and computer-readable recording medium
US20120078627A1 (en) * 2010-09-27 2012-03-29 Wagner Oliver P Electronic device with text error correction based on voice recognition data
US20140095971A1 (en) * 2012-09-28 2014-04-03 Interactive Memories, Inc. Method for Form Filling an Address on a Mobile Computing Device based on Zip Code Lookup
US8892446B2 (en) 2010-01-18 2014-11-18 Apple Inc. Service orchestration for intelligent automated assistant
US9190062B2 (en) 2010-02-25 2015-11-17 Apple Inc. User profiling for voice input processing
US20160042732A1 (en) * 2005-08-26 2016-02-11 At&T Intellectual Property Ii, L.P. System and method for robust access and entry to large structured data using voice form-filling
US9262612B2 (en) 2011-03-21 2016-02-16 Apple Inc. Device access using voice authentication
US20160062971A1 (en) * 2012-04-05 2016-03-03 Mitesh L. THAKKER Systems and methods to input or access data using remote submitting mechanism
US9300784B2 (en) 2013-06-13 2016-03-29 Apple Inc. System and method for emergency calls initiated by voice command
US9330720B2 (en) 2008-01-03 2016-05-03 Apple Inc. Methods and apparatus for altering audio output signals
US9338493B2 (en) 2014-06-30 2016-05-10 Apple Inc. Intelligent automated assistant for TV user interactions
US9368114B2 (en) 2013-03-14 2016-06-14 Apple Inc. Context-sensitive handling of interruptions
US9430463B2 (en) 2014-05-30 2016-08-30 Apple Inc. Exemplar-based natural language processing
US9483461B2 (en) 2012-03-06 2016-11-01 Apple Inc. Handling speech synthesis of content for multiple languages
US9495129B2 (en) 2012-06-29 2016-11-15 Apple Inc. Device, method, and user interface for voice-activated navigation and browsing of a document
US9502031B2 (en) 2014-05-27 2016-11-22 Apple Inc. Method for supporting dynamic grammars in WFST-based ASR
US9535906B2 (en) 2008-07-31 2017-01-03 Apple Inc. Mobile device having human language translation capability with positional feedback
US9576574B2 (en) 2012-09-10 2017-02-21 Apple Inc. Context-sensitive handling of interruptions by intelligent digital assistant
US9582608B2 (en) 2013-06-07 2017-02-28 Apple Inc. Unified ranking with entropy-weighted information for phrase-based semantic auto-completion
US9583107B2 (en) 2006-04-05 2017-02-28 Amazon Technologies, Inc. Continuous speech transcription performance indication
US9620104B2 (en) 2013-06-07 2017-04-11 Apple Inc. System and method for user-specified pronunciation of words for speech synthesis and recognition
US9620105B2 (en) 2014-05-15 2017-04-11 Apple Inc. Analyzing audio input for efficient speech and music recognition
US9626955B2 (en) 2008-04-05 2017-04-18 Apple Inc. Intelligent text-to-speech conversion
US9633674B2 (en) 2013-06-07 2017-04-25 Apple Inc. System and method for detecting errors in interactions with a voice-based digital assistant
US9633004B2 (en) 2014-05-30 2017-04-25 Apple Inc. Better resolution when referencing to concepts
US9646614B2 (en) 2000-03-16 2017-05-09 Apple Inc. Fast, language-independent method for user authentication by voice
US9646609B2 (en) 2014-09-30 2017-05-09 Apple Inc. Caching apparatus for serving phonetic pronunciations
US9668121B2 (en) 2014-09-30 2017-05-30 Apple Inc. Social reminders
US9697820B2 (en) 2015-09-24 2017-07-04 Apple Inc. Unit-selection text-to-speech synthesis using concatenation-sensitive neural networks
US9697822B1 (en) 2013-03-15 2017-07-04 Apple Inc. System and method for updating an adaptive speech recognition model
US9711141B2 (en) 2014-12-09 2017-07-18 Apple Inc. Disambiguating heteronyms in speech synthesis
US9715875B2 (en) 2014-05-30 2017-07-25 Apple Inc. Reducing the need for manual start/end-pointing and trigger phrases
US9721566B2 (en) 2015-03-08 2017-08-01 Apple Inc. Competing devices responding to voice triggers
US9734193B2 (en) 2014-05-30 2017-08-15 Apple Inc. Determining domain salience ranking from ambiguous words in natural speech
US9760559B2 (en) 2014-05-30 2017-09-12 Apple Inc. Predictive text input
US9785630B2 (en) 2014-05-30 2017-10-10 Apple Inc. Text prediction using combined word N-gram and unigram language models
US9798393B2 (en) 2011-08-29 2017-10-24 Apple Inc. Text correction processing
US9818400B2 (en) 2014-09-11 2017-11-14 Apple Inc. Method and apparatus for discovering trending terms in speech requests
US9842105B2 (en) 2015-04-16 2017-12-12 Apple Inc. Parsimonious continuous-space phrase representations for natural language processing
US9842101B2 (en) 2014-05-30 2017-12-12 Apple Inc. Predictive conversion of language input
US9858925B2 (en) 2009-06-05 2018-01-02 Apple Inc. Using context information to facilitate processing of commands in a virtual assistant
US9865280B2 (en) 2015-03-06 2018-01-09 Apple Inc. Structured dictation using intelligent automated assistants
US9886953B2 (en) 2015-03-08 2018-02-06 Apple Inc. Virtual assistant activation
US9886432B2 (en) 2014-09-30 2018-02-06 Apple Inc. Parsimonious handling of word inflection via categorical stem + suffix N-gram language models
US9899019B2 (en) 2015-03-18 2018-02-20 Apple Inc. Systems and methods for structured stem and suffix language models
US9922642B2 (en) 2013-03-15 2018-03-20 Apple Inc. Training an at least partial voice command system
US9934775B2 (en) 2016-05-26 2018-04-03 Apple Inc. Unit-selection text-to-speech synthesis based on predicted concatenation parameters
US9953088B2 (en) 2012-05-14 2018-04-24 Apple Inc. Crowd sourcing information to fulfill user requests
US9959870B2 (en) 2008-12-11 2018-05-01 Apple Inc. Speech recognition involving a mobile device
US9966065B2 (en) 2014-05-30 2018-05-08 Apple Inc. Multi-command single utterance input method
US9966068B2 (en) 2013-06-08 2018-05-08 Apple Inc. Interpreting and acting upon commands that involve sharing information with remote devices
US9973450B2 (en) 2007-09-17 2018-05-15 Amazon Technologies, Inc. Methods and systems for dynamically updating web service profile information by parsing transcribed message strings
US9971774B2 (en) 2012-09-19 2018-05-15 Apple Inc. Voice-based media searching
US9972304B2 (en) 2016-06-03 2018-05-15 Apple Inc. Privacy preserving distributed evaluation framework for embedded personalized systems

Families Citing this family (23)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US7444286B2 (en) 2001-09-05 2008-10-28 Roth Daniel L Speech recognition using re-utterance recognition
US7526431B2 (en) * 2001-09-05 2009-04-28 Voice Signal Technologies, Inc. Speech recognition using ambiguous or phone key spelling and/or filtering
US7809574B2 (en) 2001-09-05 2010-10-05 Voice Signal Technologies Inc. Word recognition using choice lists
JP3542578B2 (en) * 2001-11-22 2004-07-14 キヤノン株式会社 Speech recognition apparatus and method, program
US7222072B2 (en) * 2003-02-13 2007-05-22 Sbc Properties, L.P. Bio-phonetic multi-phrase speaker identity verification
US20050049858A1 (en) * 2003-08-25 2005-03-03 Bellsouth Intellectual Property Corporation Methods and systems for improving alphabetic speech recognition accuracy
US6983244B2 (en) * 2003-08-29 2006-01-03 Matsushita Electric Industrial Co., Ltd. Method and apparatus for improved speech recognition with supplementary information
GB0322516D0 (en) * 2003-09-25 2003-10-29 Canon Europa Nv Cellular mobile communication device
JP4012143B2 (en) * 2003-12-16 2007-11-21 キヤノン株式会社 The information processing apparatus and a data input method
US7363224B2 (en) * 2003-12-30 2008-04-22 Microsoft Corporation Method for entering text
WO2005076259A1 (en) * 2004-02-10 2005-08-18 Nec Corporation Speech input system, electronic device, speech input method, and speech input program
JP2006011641A (en) * 2004-06-23 2006-01-12 Fujitsu Ltd Information input method and device
EP1635328B1 (en) * 2004-09-14 2018-05-30 Swisscom AG Speech recognition method constrained with a grammar received from a remote system.
JP4027357B2 (en) * 2004-10-08 2007-12-26 キヤノン株式会社 String input apparatus and a control method thereof
DE112006000225T5 (en) * 2005-02-28 2007-12-13 Honda Motor Co., Ltd. Dialog System and Software dialog
CN101297351B (en) 2005-10-27 2013-01-02 纽昂斯奥地利通讯有限公司 Method and system for processing dictated information
KR20090019198A (en) * 2007-08-20 2009-02-25 삼성전자주식회사 Method and apparatus for automatically completed text input using speech recognition
KR101502003B1 (en) * 2008-07-08 2015-03-12 엘지전자 주식회사 A mobile terminal and a text input method
US20120280915A1 (en) * 2011-05-02 2012-11-08 Nokia Corporation Method and apparatus for facilitating interacting with a multimodal user interface
US9257115B2 (en) 2012-03-08 2016-02-09 Facebook, Inc. Device for extracting information from a dialog
KR20140028540A (en) * 2012-08-29 2014-03-10 엘지전자 주식회사 Display device and speech search method thereof
US9946699B1 (en) * 2012-08-29 2018-04-17 Intuit Inc. Location-based speech recognition for preparation of electronic tax return
US9361084B1 (en) 2013-11-14 2016-06-07 Google Inc. Methods and systems for installing and executing applications

Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US5530950A (en) * 1993-07-10 1996-06-25 International Business Machines Corporation Audio data processing
US5855000A (en) * 1995-09-08 1998-12-29 Carnegie Mellon University Method and apparatus for correcting and repairing machine-transcribed input using independent or cross-modal secondary input
US5960399A (en) * 1996-12-24 1999-09-28 Gte Internetworking Incorporated Client/server speech processor/recognizer

Family Cites Families (16)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US5131045A (en) * 1990-05-10 1992-07-14 Roth Richard G Audio-augmented data keying
JPH07210190A (en) 1993-12-30 1995-08-11 Internatl Business Mach Corp <Ibm> Method and system for speech recognition
USRE42868E1 (en) 1994-10-25 2011-10-25 Cisco Technology, Inc. Voice-operated services
US5963666A (en) * 1995-08-18 1999-10-05 International Business Machines Corporation Confusion matrix mediated word prediction
US5917890A (en) * 1995-12-29 1999-06-29 At&T Corp Disambiguation of alphabetic characters in an automated call processing environment
US6952799B2 (en) * 1996-06-17 2005-10-04 British Telecommunications User interface for network browser including pre-processor for links embedded in hypermedia documents
US5912949A (en) 1996-11-05 1999-06-15 Northern Telecom Limited Voice-dialing system using both spoken names and initials in recognition
US6078886A (en) * 1997-04-14 2000-06-20 At&T Corporation System and method for providing remote automatic speech recognition services via a packet network
US6044347A (en) * 1997-08-05 2000-03-28 Lucent Technologies Inc. Methods and apparatus object-oriented rule-based dialogue management
JP2002505460A (en) * 1998-02-25 2002-02-19 ルノー・アンド・オスピー・スピーチ・プロダクツ・ナームローゼ・ベンノートシャープ Comprehensive runtime engine for interfacing between an application and voice engine
FI981154A (en) 1998-05-25 1999-11-26 Nokia Mobile Phones Ltd Method and apparatus for speech recognition
US6505162B1 (en) * 1999-06-11 2003-01-07 Industrial Technology Research Institute Apparatus and method for portable dialogue management using a hierarchial task description table
JP2001188781A (en) * 1999-12-28 2001-07-10 Sony Corp Device and method for processing conversation and recording medium
US7143043B1 (en) 2000-04-26 2006-11-28 Openwave Systems Inc. Constrained keyboard disambiguation using voice recognition
WO2002005263A1 (en) 2000-07-07 2002-01-17 Siemens Aktiengesellschaft Method for voice input and voice recognition
US6728348B2 (en) 2000-11-30 2004-04-27 Comverse, Inc. System for storing voice recognizable identifiers using a limited input device such as a telephone key pad

Patent Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US5530950A (en) * 1993-07-10 1996-06-25 International Business Machines Corporation Audio data processing
US5855000A (en) * 1995-09-08 1998-12-29 Carnegie Mellon University Method and apparatus for correcting and repairing machine-transcribed input using independent or cross-modal secondary input
US5960399A (en) * 1996-12-24 1999-09-28 Gte Internetworking Incorporated Client/server speech processor/recognizer

Cited By (99)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US9646614B2 (en) 2000-03-16 2017-05-09 Apple Inc. Fast, language-independent method for user authentication by voice
US20060167685A1 (en) * 2002-02-07 2006-07-27 Eric Thelen Method and device for the rapid, pattern-recognition-supported transcription of spoken and written utterances
US20050288934A1 (en) * 2004-06-29 2005-12-29 Canon Kabushiki Kaisha Multimodal input method
US7630901B2 (en) * 2004-06-29 2009-12-08 Canon Kabushiki Kaisha Multimodal input method
US20060009974A1 (en) * 2004-07-09 2006-01-12 Matsushita Electric Industrial Co., Ltd. Hands-free voice dialing for portable and remote devices
US20060074652A1 (en) * 2004-09-20 2006-04-06 International Business Machines Corporation Method and system for voice-enabled autofill
US7953597B2 (en) * 2004-09-20 2011-05-31 Nuance Communications, Inc. Method and system for voice-enabled autofill
US8065390B2 (en) 2004-11-17 2011-11-22 Juniper Networks, Inc. Virtual folders for tracking HTTP sessions
US7502835B1 (en) * 2004-11-17 2009-03-10 Juniper Networks, Inc. Virtual folders for tracking HTTP sessions
US20110093603A1 (en) * 2004-11-17 2011-04-21 Juniper Networks, Inc. Virtual folders for tracking http sessions
US7886061B1 (en) 2004-11-17 2011-02-08 Juniper Networks, Inc. Virtual folders for tracking HTTP sessions
US7461134B2 (en) * 2004-11-19 2008-12-02 W.A. Krapf, Inc. Bi-directional communication between a web client and a web server
US20060136564A1 (en) * 2004-11-19 2006-06-22 W.A. Krapf, Inc. Bi-directional communication between a web client and a web server
US20060276916A1 (en) * 2004-12-22 2006-12-07 Dearing Stephen M System and method for electronically processing address information
US7801925B2 (en) * 2004-12-22 2010-09-21 United States Postal Service System and method for electronically processing address information
US20060167684A1 (en) * 2005-01-24 2006-07-27 Delta Electronics, Inc. Speech recognition method and system
US8165909B2 (en) 2005-05-17 2012-04-24 The United States Postal Service System and method for automated management of an address database
US20070094155A1 (en) * 2005-05-17 2007-04-26 Dearing Stephen M System and method for automated management of an address database
US20160042732A1 (en) * 2005-08-26 2016-02-11 At&T Intellectual Property Ii, L.P. System and method for robust access and entry to large structured data using voice form-filling
US9824682B2 (en) * 2005-08-26 2017-11-21 Nuance Communications, Inc. System and method for robust access and entry to large structured data using voice form-filling
US9583107B2 (en) 2006-04-05 2017-02-28 Amazon Technologies, Inc. Continuous speech transcription performance indication
US9117447B2 (en) 2006-09-08 2015-08-25 Apple Inc. Using event alert text as input to an automated assistant
US8942986B2 (en) 2006-09-08 2015-01-27 Apple Inc. Determining user intent based on ontologies of domains
US8930191B2 (en) 2006-09-08 2015-01-06 Apple Inc. Paraphrasing of user requests and results by automated digital assistant
US20080281582A1 (en) * 2007-05-11 2008-11-13 Delta Electronics, Inc. Input system for mobile search and method therefor
US9973450B2 (en) 2007-09-17 2018-05-15 Amazon Technologies, Inc. Methods and systems for dynamically updating web service profile information by parsing transcribed message strings
US8024347B2 (en) 2007-09-27 2011-09-20 International Business Machines Corporation Method and apparatus for automatically differentiating between types of names stored in a data collection
US9330720B2 (en) 2008-01-03 2016-05-03 Apple Inc. Methods and apparatus for altering audio output signals
US20090187410A1 (en) * 2008-01-22 2009-07-23 At&T Labs, Inc. System and method of providing speech processing in user interface
US9530415B2 (en) 2008-01-22 2016-12-27 At&T Intellectual Property I, L.P. System and method of providing speech processing in user interface
US9177551B2 (en) * 2008-01-22 2015-11-03 At&T Intellectual Property I, L.P. System and method of providing speech processing in user interface
US20090248415A1 (en) * 2008-03-31 2009-10-01 Yap, Inc. Use of metadata to post process speech recognition output
US8676577B2 (en) * 2008-03-31 2014-03-18 Canyon IP Holdings, LLC Use of metadata to post process speech recognition output
US9865248B2 (en) 2008-04-05 2018-01-09 Apple Inc. Intelligent text-to-speech conversion
US9626955B2 (en) 2008-04-05 2017-04-18 Apple Inc. Intelligent text-to-speech conversion
US9535906B2 (en) 2008-07-31 2017-01-03 Apple Inc. Mobile device having human language translation capability with positional feedback
US20100057713A1 (en) * 2008-09-03 2010-03-04 International Business Machines Corporation Entity-driven logic for improved name-searching in mixed-entity lists
US9411877B2 (en) 2008-09-03 2016-08-09 International Business Machines Corporation Entity-driven logic for improved name-searching in mixed-entity lists
US20100070360A1 (en) * 2008-09-13 2010-03-18 At&T Intellectual Property I, L.P. System and method for creating a speech search platform for coupons
US9959870B2 (en) 2008-12-11 2018-05-01 Apple Inc. Speech recognition involving a mobile device
US8886519B2 (en) * 2008-12-26 2014-11-11 Nec Corporation Text processing apparatus, text processing method, and computer-readable recording medium
US20110282653A1 (en) * 2008-12-26 2011-11-17 Akihiro Tamura Text processing apparatus, text processing method, and computer-readable recording medium
US9858925B2 (en) 2009-06-05 2018-01-02 Apple Inc. Using context information to facilitate processing of commands in a virtual assistant
US20110169906A1 (en) * 2010-01-13 2011-07-14 Seizo Suzuki Optical scanning device and image forming apparatus
US9548050B2 (en) 2010-01-18 2017-01-17 Apple Inc. Intelligent automated assistant
US8903716B2 (en) 2010-01-18 2014-12-02 Apple Inc. Personalized vocabulary for digital assistant
US9318108B2 (en) 2010-01-18 2016-04-19 Apple Inc. Intelligent automated assistant
US8892446B2 (en) 2010-01-18 2014-11-18 Apple Inc. Service orchestration for intelligent automated assistant
US9190062B2 (en) 2010-02-25 2015-11-17 Apple Inc. User profiling for voice input processing
US9633660B2 (en) 2010-02-25 2017-04-25 Apple Inc. User profiling for voice input processing
US20110264994A1 (en) * 2010-04-13 2011-10-27 ForenSight, Inc. Method of retrieving interactive electronic data and accessing productivity tools for real property valuation and mortgage underwriting software systems
US9075783B2 (en) * 2010-09-27 2015-07-07 Apple Inc. Electronic device with text error correction based on voice recognition data
US20120078627A1 (en) * 2010-09-27 2012-03-29 Wagner Oliver P Electronic device with text error correction based on voice recognition data
US8719014B2 (en) * 2010-09-27 2014-05-06 Apple Inc. Electronic device with text error correction based on voice recognition data
US9262612B2 (en) 2011-03-21 2016-02-16 Apple Inc. Device access using voice authentication
US9798393B2 (en) 2011-08-29 2017-10-24 Apple Inc. Text correction processing
US9483461B2 (en) 2012-03-06 2016-11-01 Apple Inc. Handling speech synthesis of content for multiple languages
US20160062971A1 (en) * 2012-04-05 2016-03-03 Mitesh L. THAKKER Systems and methods to input or access data using remote submitting mechanism
US9953088B2 (en) 2012-05-14 2018-04-24 Apple Inc. Crowd sourcing information to fulfill user requests
US9495129B2 (en) 2012-06-29 2016-11-15 Apple Inc. Device, method, and user interface for voice-activated navigation and browsing of a document
US9576574B2 (en) 2012-09-10 2017-02-21 Apple Inc. Context-sensitive handling of interruptions by intelligent digital assistant
US9971774B2 (en) 2012-09-19 2018-05-15 Apple Inc. Voice-based media searching
US20140095971A1 (en) * 2012-09-28 2014-04-03 Interactive Memories, Inc. Method for Form Filling an Address on a Mobile Computing Device based on Zip Code Lookup
US9332137B2 (en) * 2012-09-28 2016-05-03 Interactive Memories Inc. Method for form filling an address on a mobile computing device based on zip code lookup
US9368114B2 (en) 2013-03-14 2016-06-14 Apple Inc. Context-sensitive handling of interruptions
US9697822B1 (en) 2013-03-15 2017-07-04 Apple Inc. System and method for updating an adaptive speech recognition model
US9922642B2 (en) 2013-03-15 2018-03-20 Apple Inc. Training an at least partial voice command system
US9966060B2 (en) 2013-06-07 2018-05-08 Apple Inc. System and method for user-specified pronunciation of words for speech synthesis and recognition
US9633674B2 (en) 2013-06-07 2017-04-25 Apple Inc. System and method for detecting errors in interactions with a voice-based digital assistant
US9582608B2 (en) 2013-06-07 2017-02-28 Apple Inc. Unified ranking with entropy-weighted information for phrase-based semantic auto-completion
US9620104B2 (en) 2013-06-07 2017-04-11 Apple Inc. System and method for user-specified pronunciation of words for speech synthesis and recognition
US9966068B2 (en) 2013-06-08 2018-05-08 Apple Inc. Interpreting and acting upon commands that involve sharing information with remote devices
US9300784B2 (en) 2013-06-13 2016-03-29 Apple Inc. System and method for emergency calls initiated by voice command
US9620105B2 (en) 2014-05-15 2017-04-11 Apple Inc. Analyzing audio input for efficient speech and music recognition
US9502031B2 (en) 2014-05-27 2016-11-22 Apple Inc. Method for supporting dynamic grammars in WFST-based ASR
US9715875B2 (en) 2014-05-30 2017-07-25 Apple Inc. Reducing the need for manual start/end-pointing and trigger phrases
US9734193B2 (en) 2014-05-30 2017-08-15 Apple Inc. Determining domain salience ranking from ambiguous words in natural speech
US9785630B2 (en) 2014-05-30 2017-10-10 Apple Inc. Text prediction using combined word N-gram and unigram language models
US9760559B2 (en) 2014-05-30 2017-09-12 Apple Inc. Predictive text input
US9430463B2 (en) 2014-05-30 2016-08-30 Apple Inc. Exemplar-based natural language processing
US9966065B2 (en) 2014-05-30 2018-05-08 Apple Inc. Multi-command single utterance input method
US9842101B2 (en) 2014-05-30 2017-12-12 Apple Inc. Predictive conversion of language input
US9633004B2 (en) 2014-05-30 2017-04-25 Apple Inc. Better resolution when referencing to concepts
US9338493B2 (en) 2014-06-30 2016-05-10 Apple Inc. Intelligent automated assistant for TV user interactions
US9668024B2 (en) 2014-06-30 2017-05-30 Apple Inc. Intelligent automated assistant for TV user interactions
US9818400B2 (en) 2014-09-11 2017-11-14 Apple Inc. Method and apparatus for discovering trending terms in speech requests
US9668121B2 (en) 2014-09-30 2017-05-30 Apple Inc. Social reminders
US9986419B2 (en) 2014-09-30 2018-05-29 Apple Inc. Social reminders
US9646609B2 (en) 2014-09-30 2017-05-09 Apple Inc. Caching apparatus for serving phonetic pronunciations
US9886432B2 (en) 2014-09-30 2018-02-06 Apple Inc. Parsimonious handling of word inflection via categorical stem + suffix N-gram language models
US9711141B2 (en) 2014-12-09 2017-07-18 Apple Inc. Disambiguating heteronyms in speech synthesis
US9865280B2 (en) 2015-03-06 2018-01-09 Apple Inc. Structured dictation using intelligent automated assistants
US9886953B2 (en) 2015-03-08 2018-02-06 Apple Inc. Virtual assistant activation
US9721566B2 (en) 2015-03-08 2017-08-01 Apple Inc. Competing devices responding to voice triggers
US9899019B2 (en) 2015-03-18 2018-02-20 Apple Inc. Systems and methods for structured stem and suffix language models
US9842105B2 (en) 2015-04-16 2017-12-12 Apple Inc. Parsimonious continuous-space phrase representations for natural language processing
US9697820B2 (en) 2015-09-24 2017-07-04 Apple Inc. Unit-selection text-to-speech synthesis using concatenation-sensitive neural networks
US9934775B2 (en) 2016-05-26 2018-04-03 Apple Inc. Unit-selection text-to-speech synthesis based on predicted concatenation parameters
US9972304B2 (en) 2016-06-03 2018-05-15 Apple Inc. Privacy preserving distributed evaluation framework for embedded personalized systems

Also Published As

Publication number Publication date Type
US20030115057A1 (en) 2003-06-19 application
US7124085B2 (en) 2006-10-17 grant
CN1424711A (en) 2003-06-18 application
JP2003216181A (en) 2003-07-30 application
EP1326234A3 (en) 2004-12-22 application
CN1238832C (en) 2006-01-25 grant
EP1326234A2 (en) 2003-07-09 application

Similar Documents

Publication Publication Date Title
Gorin et al. On adaptive acquisition of language
US6513006B2 (en) Automatic control of household activity using speech recognition and natural language
US6510235B1 (en) Method of storing data in and searching through an electronic phone number data book in a digital mobile phone with character recognition function
US7216073B2 (en) Dynamic natural language understanding
US6460029B1 (en) System for improving search text
US6223155B1 (en) Method of independently creating and using a garbage model for improved rejection in a limited-training speaker-dependent speech recognition system
US5488652A (en) Method and apparatus for training speech recognition algorithms for directory assistance applications
US6073101A (en) Text independent speaker recognition for transparent command ambiguity resolution and continuous access control
US6751606B1 (en) System for enhancing a query interface
US6954755B2 (en) Task/domain segmentation in applying feedback to command control
US6671672B1 (en) Voice authentication system having cognitive recall mechanism for password verification
US7848926B2 (en) System, method, and program for correcting misrecognized spoken words by selecting appropriate correction word from one or more competitive words
US7190770B2 (en) Method and system for acquiring information with voice input
US20060293889A1 (en) Error correction for speech recognition systems
US20020142787A1 (en) Method to select and send text messages with a mobile
US6081782A (en) Voice command control and verification system
US20040193420A1 (en) Mobile systems and methods for responding to natural language speech utterance
US20050171775A1 (en) Automatically improving a voice recognition system
US20030115062A1 (en) Method for automated sentence planning
US7299186B2 (en) Speech input system, speech portal server, and speech input terminal
US20030191648A1 (en) Method and system for voice recognition menu navigation with error prevention and recovery
US7606714B2 (en) Natural language classification within an automated response system
US20120173464A1 (en) Method and apparatus for exploiting human feedback in an intelligent automated assistant
US20040098245A1 (en) Method for automated sentence planning in a task classification system
US5987414A (en) Method and apparatus for selecting a vocabulary sub-set from a speech recognition dictionary for use in real time automated directory assistance

Legal Events

Date Code Title Description
AS Assignment

Owner name: MATSUSHITA ELECTRIC INDUSTRIAL CO., LTD., JAPAN

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:JUNQUA, JEAN-CLAUDE;STOIMENOV, KIRILL;KUHN, ROLAND;REEL/FRAME:013300/0049

Effective date: 20020913

AS Assignment

Owner name: MATSUSHITA ELECTRIC INDUSTRIAL CO., LTD., JAPAN

Free format text: RE-RECORD TO CORRECT THE RECEIVING PARTY S ADDRESS, PREVIOUSLY RECORDED AT REEL 013300, FRAME 0049.;ASSIGNORS:JUNQUA, JEAN-CLAUDE;STOIMENOV, KIRILL;KUHN, ROLAND;REEL/FRAME:013701/0263

Effective date: 20020913