US9973450B2 - Methods and systems for dynamically updating web service profile information by parsing transcribed message strings - Google Patents
Methods and systems for dynamically updating web service profile information by parsing transcribed message strings Download PDFInfo
- Publication number
- US9973450B2 US9973450B2 US12/212,644 US21264408A US9973450B2 US 9973450 B2 US9973450 B2 US 9973450B2 US 21264408 A US21264408 A US 21264408A US 9973450 B2 US9973450 B2 US 9973450B2
- Authority
- US
- United States
- Prior art keywords
- user
- client device
- message
- profile information
- profile
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Active, expires
Links
Images
Classifications
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04L—TRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
- H04L51/00—User-to-user messaging in packet-switching networks, transmitted according to store-and-forward or real-time protocols, e.g. e-mail
- H04L51/04—Real-time or near real-time messaging, e.g. instant messaging [IM]
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS OR SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING; SPEECH OR AUDIO CODING OR DECODING
- G10L15/00—Speech recognition
- G10L15/26—Speech to text systems
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04M—TELEPHONIC COMMUNICATION
- H04M1/00—Substation equipment, e.g. for use by subscribers
- H04M1/72—Mobile telephones; Cordless telephones, i.e. devices for establishing wireless links to base stations without route selection
- H04M1/724—User interfaces specially adapted for cordless or mobile telephones
- H04M1/72403—User interfaces specially adapted for cordless or mobile telephones with means for local support of applications that increase the functionality
- H04M1/7243—User interfaces specially adapted for cordless or mobile telephones with means for local support of applications that increase the functionality with interactive means for internal management of messages
- H04M1/72436—User interfaces specially adapted for cordless or mobile telephones with means for local support of applications that increase the functionality with interactive means for internal management of messages for text messaging, e.g. SMS or e-mail
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04M—TELEPHONIC COMMUNICATION
- H04M1/00—Substation equipment, e.g. for use by subscribers
- H04M1/72—Mobile telephones; Cordless telephones, i.e. devices for establishing wireless links to base stations without route selection
- H04M1/724—User interfaces specially adapted for cordless or mobile telephones
- H04M1/72403—User interfaces specially adapted for cordless or mobile telephones with means for local support of applications that increase the functionality
- H04M1/72445—User interfaces specially adapted for cordless or mobile telephones with means for local support of applications that increase the functionality for supporting Internet browser applications
-
- H04M1/72552—
-
- H04M1/72561—
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04W—WIRELESS COMMUNICATION NETWORKS
- H04W8/00—Network data management
- H04W8/18—Processing of user or subscriber data, e.g. subscribed services, user preferences or user profiles; Transfer of user or subscriber data
- H04W8/20—Transfer of user or subscriber data
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04L—TRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
- H04L51/00—User-to-user messaging in packet-switching networks, transmitted according to store-and-forward or real-time protocols, e.g. e-mail
- H04L51/06—Message adaptation to terminal or network requirements
- H04L51/066—Format adaptation, e.g. format conversion or compression
Definitions
- provisional application 60/789,837 is contained in APPENDIX A attached hereto and, likewise, is incorporated herein in its entirety by reference and is intended to provide background and technical information with regard to the systems and environments of the inventions of the current provisional patent application.
- the disclosure of the brochure of APPENDIX B is incorporated herein in its entirety by reference.
- SMS Short Message Service
- mobile client devices such as smart phones or PDAs.
- SMSes Short Message Service
- mobile client devices such as smart phones or PDAs.
- SMSes also are now used to interact with automated systems, such as ordering products and services for mobile client devices or participating in contests using mobile client devices such as, for example, voting for contestants in American Idol competitions.
- IM instant messaging
- IM is a form of “real-time” communication between two or more people that is based on the transmission of text.
- the text is conveyed over a network such as the Internet.
- Instant messaging requires an IM client that connects to an IM service.
- the IM client commonly is installed on a computer such as a laptop or desktop.
- IM clients are now available for use on mobile client devices. Because IM is considered “real-time,” communications back and forth between users of IM clients sometimes is deemed a “conversation,” just as if the people were speaking directly to one another.
- the present invention has applicability both in text messaging as well as in instant messaging and, except where context clearly implies otherwise, aspects and features of the present invention apply in the context of both (a) SMS systems, methods, applications, and implementations as well as (b) IM systems, methods, applications, and implementations.
- speech recognition refers to the process of converting a speech (audio) signal to a sequence of words or a representation thereof (message strings), by means of an algorithm implemented as a computer program.
- Speech recognition applications that have emerged over the last few years include voice dialing (e.g., “Call home”), call routing (e.g., “I would like to make a collect call”), simple data entry (e.g., entering a credit card number), preparation of structured documents (e.g., a radiology report), and content-based spoken audio searching (e.g. finding a podcast where particular words were spoken).
- ASR systems have become commonplace in recent years.
- ASR systems have found wide application in customer service centers of companies.
- the customer service centers offer middleware and solutions for contact centers. For example, they answer and route calls to decrease costs for airlines, banks, etc.
- companies such as IBM and Nuance create assets known as IVR (Interactive Voice Response) that answer the calls, then use ASR (Automatic Speech Recognition) paired with TTS (Text-To-Speech) software to decode what the caller is saying and communicate back to them.
- IVR Interactive Voice Response
- Text messaging and instant messaging usually involves the input of a textual message by a sender who presses letters and/or numbers associated with the sender's mobile phone or other mobile device.
- a sender who presses letters and/or numbers associated with the sender's mobile phone or other mobile device.
- text messages can be advantageous to a message receiver as compared to voicemail, as the receiver actually sees the message content in a written format rather than having to rely on an auditory signal.
- users can or will be able to use mobile client devices to interface with many web services via an IM client and/or SMSes. It is believed, for example, that users can or will interact with web services using text messages and/or instant messages such as those provided by Amazon, Facebook, and MySpace. This may be accomplished, for example, using either manually-typed text messages and/or instant messages or such messages that are transcribed from speech using an ASR engine.
- inventive aspects and features of the present invention are believed to further enable and facilitate the use and acceptance of text messaging and instant messaging with mobile client devices.
- inventive aspects and features of the invention relate to parsing and/or filtering of message strings (text of instant messages or text messages) that are either manually typed, transcribed from speech, or part of a stream web services query, in order to identify keywords, phrases, or fragments based on which user preferences of user profiles are dynamically updated.
- One or more steps of inventive aspects and features of methods of the invention may be performed in client and/or server side processing.
- the present invention includes many aspects and features. Moreover, while many aspects and features relate to, and are described in, the context of providing profile information to a web service, the present invention is not limited to use only in such field, as will become apparent from the following summaries and detailed descriptions of aspects, features, and one or more embodiments of the present invention.
- one aspect of the present invention relates to a method of providing profile information, derived from an utterance, from a mobile communication device to a web service.
- An exemplary such method includes the steps of receiving, at the mobile communication device, audio data representing an utterance; transcribing the audio data to text; processing the transcribed text, including parsing the text for profile information appropriate for use at one or more web services; and communicating, to the web service, the profile information parsed from the transcribed text.
- the processing step may be performed by a profile filter; the method may further comprise providing an interface to a user for manual user editing of the transcribed text; the transcription step may be performed at the mobile communication device; the transcription step may be performed by a separate automatic speech recognition system; the audio data may be a voicemail; the method may further comprise delivering ad impressions to a user based on the processed text; and the method may further comprise communicating the transcribed text, as a text-based message, from the mobile communication device to a recipient.
- the recipient may be a cell phone; the recipient may be a smart phone; the recipient may be a PDA; the recipient may be a tablet notebook; the recipient may be a desktop computer; the recipient may be a laptop computer; the recipient may be a web service; the text-based message may be a text message, communicated using Short Message Service; and the text-based message may be an instant message, communicated via an instant message service.
- Another aspect of the invention relates to a method of providing profile information, derived from an utterance, from a mobile communication device to a web service.
- An exemplary such method includes transcribing audio data, received as an utterance at the mobile communication device, to text; providing an interface to a user for manual user editing of the transcribed text; processing the edited text, including parsing the text for profile information appropriate for use at one or more web services; and communicating, to the web service, the profile information parsed from the transcribed text.
- the processing step may be performed by a profile filter; the transcription step may be performed at the mobile communication device; the transcription step may be performed by a separate automatic speech recognition system; the audio data may be a voicemail; the method may further comprise delivering ad impressions to a user based on the processed text; and the method may further comprise communicating the transcribed text, as a text-based message, from the mobile communication device to the recipient.
- the recipient may be a cell phone; the recipient may be a smart phone; the recipient may be a PDA; the recipient may be a tablet notebook; the recipient may be a desktop computer; the recipient may be a laptop computer; the recipient may be a web service; the text-based message may be a text message, communicated using Short Message Service; and the text-based message may be an instant message, communicated via an instant message service.
- Another aspect of the invention relates to a method of providing profile information, derived from an utterance, from a mobile communication device to a web service.
- An exemplary such method includes receiving, at the mobile communication device, audio data representing an utterance that is to be sent from the mobile communication device to a recipient; transcribing the utterance to text; parsing the transcribed text to identify relevant profile information for input to a web service; and communicating the transcribed text, as a text-based message, from the mobile communication device to the recipient.
- the parsing step may be performed by a profile filter; the method may further comprise providing an interface to a user for manual user editing of the transcribed text; the transcription step may be performed at the mobile communication device; the transcription step may be performed by a separate automatic speech recognition system; the audio data may be a voicemail; the method may further comprise delivering ad impressions to a user based on the parsed text; and the method may further comprise communicating, to the web service, the profile information parsed from the transcribed text.
- the recipient may be a cell phone; the recipient may be a smart phone; the recipient may be a PDA; the recipient may be a tablet notebook; the recipient may be a desktop computer; the recipient may be a laptop computer; the recipient may be a web service; the text-based message may be a text message, communicated using Short Message Service; and the text-based message may be an instant message, communicated via an instant message service.
- Another aspect of the invention relates to a method of providing profile information, derived from an utterance, from a mobile communication device to a web service.
- An exemplary such method includes receiving, at the mobile communication device, audio data representing an utterance that is to be sent from the mobile communication device to a recipient; transcribing the utterance to text; parsing the transcribed text to identify relevant profile information for input to a web service; and storing, in a profile information index, the profile information parsed from the transcribed text.
- the parsing step may be performed by a profile filter; the method may further comprise providing an interface to a user for manual user editing of the transcribed text; the transcription step may be performed at the mobile communication device; the transcription step may be performed by a separate automatic speech recognition system; the audio data may be a voicemail; the method may further comprise delivering ad impressions to a user based on the parsed text; the method may further comprise communicating, to the web service, the profile information parsed from the transcribed text; and the method may further comprise communicating the transcribed text, as a text-based message, from the mobile communication device to the recipient.
- the recipient may be a cell phone; the recipient may be a smart phone; the recipient may be a PDA; the recipient may be a tablet notebook; the recipient may be a desktop computer; the recipient may be a laptop computer; the recipient may be a web service; the text-based message may be a text message, communicated using Short Message Service; and the text-based message may be an instant message, communicated via an instant message service.
- Another aspect of the invention relates to a method of providing profile information, derived from a message string, from a mobile communication device to a web service.
- An exemplary such method includes receiving, at the mobile communication device, input representing a text-based message that is to be sent from the mobile communication device to a recipient; producing a message string from the input; parsing the message string to identify relevant profile information for input to a web service; communicating, to the web service, the profile information parsed from the message string; and communicating the message string, as a text-based message, from the mobile communication device to the recipient.
- the input may be audio data representing an utterance and the producing step includes transcribing the utterance to text;
- the parsing step may be performed by a profile filter;
- the method may further comprise providing an interface to a user for manual user editing of the transcribed text;
- the transcription step may be performed at the mobile communication device;
- the transcription step may be performed by a separate automatic speech recognition system;
- the audio data may be a voicemail; and the method may further comprise delivering ad impressions to a user based on the parsed text.
- the recipient may be a cell phone; the recipient may be a smart phone; the recipient may be a PDA; the recipient may be a tablet notebook; the recipient may be a desktop computer; the recipient may be a laptop computer; the recipient may be a web service; the text-based message may be a text message, communicated using Short Message Service; and the text-based message may be an instant message, communicated via an instant message service.
- Another aspect of the invention relates to a method of providing profile information, derived from an instant message, from a client device to a web service.
- An exemplary such method includes receiving, at the client device, input representing an instant message that is to be sent from the client device to a recipient; producing a message string from the input; parsing the message string to identify relevant profile information for input to a web service; communicating, to the web service, the profile information parsed from the message string; and communicating the message string, as an instant message, from the client device to the recipient.
- the input may be audio data representing an utterance and the producing step includes transcribing the utterance to text;
- the parsing step may be performed by a profile filter;
- the method may further comprise providing an interface to a user for manual user editing of the transcribed text;
- the transcription step may be performed at the mobile communication device;
- the transcription step may be performed by a separate automatic speech recognition system;
- the audio data may be a voicemail; and the method may further comprise delivering ad impressions to a user based on the parsed text.
- the recipient may be a cell phone; the recipient may be a smart phone; the recipient may be a PDA; the recipient may be a tablet notebook; the recipient may be a desktop computer; the recipient may be a laptop computer; the recipient may be a web service; the text-based message may be a text message, communicated using Short Message Service; and the text-based message may be an instant message, communicated via an instant message service.
- Still another aspect of the invention relates to a method of dynamically providing profile information, derived from message strings, to a web service.
- An exemplary such method includes establishing a user account configured to interface with a user profile at a web service; thereafter, repeatedly receiving message strings at a profile filter, each message string being representative of a text-based message to be communicated to a recipient; processing each message string, including parsing each message string for profile information appropriate for use by the user profile at the web service; and communicating, to the web service, the profile information parsed from the message strings.
- Still yet another aspect of the invention relates to a system for providing profile information, derived from message strings, to a web service.
- An exemplary such system includes a mobile communication device; an automatic speech recognition engine adapted to transcribe audio data, received as an utterance at the mobile communication device, to text; a user account configured to interface with a user profile at a web service; and a profile filter adapted to parse the transcribed text, according to the configured user account, for profile information appropriate for use at the web service.
- the system may further comprise a profile information index, adapted to store profile information, for the user account.
- a system for parsing and/or filtering message strings of text messages and/or instant messages in order to identify keywords, phrases, or fragments as a function of which user preferences of user profiles are dynamically updated.
- a method is disclosed for parsing and/or filtering message strings of text messages or instant messages in order to identify keywords, phrases, or fragments as a function of which user preferences of user profiles are dynamically updated.
- software may be provided for parsing and/or filtering message strings of text messages or instant messages in order to identify keywords, phrases, or fragments as a function of which user preferences of user profiles are dynamically updated, as disclosed.
- the user profiles are associated with user accounts of web services and/or social networking sites; an automatic speech recognition system generates the message strings from audio dictated by a user using a mobile device; and/or the parsing and/or filtering is performed by client side software and/or server side software.
- users can grant, to a contact (e.g., a friend, family member, or associate), access to the user preferences of that user's profile such that the contact can query that user's profile for user profile data of known fields in the user preferences.
- a contact e.g., a friend, family member, or associate
- the known fields include favorite bands and movies; the query by the contact is performed by sending a message string including an identification of the user and a known field; and/or the query by the contact is performed by sending a text message including an identification of the user and a known field.
- ad impressions may be delivered to a user based on the parsing and/or filtering of one or more message strings of text messages and/or instant messages of the user.
- ad impressions are delivered to a user based at least in part on data of the user maintained in the user profile; ad impressions are delivered to a user based at least in part on data of the user maintained in the user profile; and/or an ad impression that is delivered is presented as a text message or an instant message.
- such an ad impression is delivered to a mobile device of an author of a message string and/or presented to an author of a message string prior to sending of the message string as an instant message or text message; and an author of a message string is provided with an option of forwarding an ad impression to a recipient of the message string prior to sending of the message string as an instant message or text message.
- an ad impression is delivered to a mobile device for presentation to a user of the mobile device as disclosed herein; a method is provided for delivering an ad impression to a mobile device for presentation to a user of the mobile device as disclosed herein; and a method is provided for granting, by a user to a contact (e.g., a friend, family member, or associate), access to user preferences of that user maintained in a user profile of that user, and querying, by the contact, that user's profile for user profile data of known fields in the user preferences.
- a contact e.g., a friend, family member, or associate
- the known fields include favorite bands and movies; the query by the contact is performed by sending a message string including an identification of the user and a known field; and/or the query by the contact is performed by sending a text message including an identification of the user and a known field.
- the user profile may be dynamically updated based on parsing and/or filtering message strings of text messages and/or instant messages authored by the user; and/or the user profile is static.
- FIG. 1 is a block diagram of a communication system in accordance with a preferred embodiment of the present invention
- FIG. 2 is a block diagram of a communication system in accordance with another preferred embodiment of the present invention.
- FIG. 3 is a block diagram of an exemplary implementation of the system of FIG. 1 ;
- FIG. 4 is a schematic diagram illustrating communications between two users via a portion of the communication system of FIGS. 1 and 3 ;
- FIG. 5 is a flowchart illustrating a method of updating profile information in a web service in accordance with one or more preferred embodiments of the present invention
- FIG. 6 is a block diagram of the system architecture of one commercial implementation
- FIG. 7 is a block diagram of a portion of FIG. 6 ;
- FIG. 8 is a typical header section of an HTTP request from the client in the commercial implementation
- FIG. 9 illustrates exemplary protocol details for a request for a location of a login server and a subsequent response
- FIG. 10 illustrates exemplary protocol details for a login request and a subsequent response
- FIG. 11 illustrates exemplary protocol details for a submit request and a subsequent response
- FIG. 12 illustrates exemplary protocol details for a results request and a subsequent response
- FIG. 13 illustrates exemplary protocol details for an XML hierarchy returned in response to a results request
- FIG. 14 illustrates exemplary protocol details for a text to speech request and a subsequent response
- FIG. 15 illustrates exemplary protocol details for a correct request
- FIG. 16 illustrates exemplary protocol details for a ping request
- FIG. 17 illustrates exemplary protocol details for a debug request.
- any sequence(s) and/or temporal order of steps of various processes or methods that are described herein are illustrative and not restrictive. Accordingly, it should be understood that, although steps of various processes or methods may be shown and described as being in a sequence or temporal order, the steps of any such processes or methods are not limited to being carried out in any particular sequence or order, absent an indication otherwise. Indeed, the steps in such processes or methods generally may be carried out in various different sequences and orders while still falling within the scope of the present invention. Accordingly, it is intended that the scope of patent protection afforded the present invention is to be defined by the appended claims rather than the description set forth herein.
- a picnic basket having an apple describes “a picnic basket having at least one apple” as well as “a picnic basket having apples.”
- a picnic basket having a single apple describes “a picnic basket having only one apple.”
- FIG. 1 is a block diagram of a communication system 10 in accordance with a preferred embodiment of the present invention.
- the communication system 10 includes at least one transmitting device 12 and at least one receiving device 14 , one or more network systems 16 for connecting the transmitting device 12 to the receiving device 14 , and an ASR system 18 , including an ASR engine.
- Transmitting and receiving devices 12 , 14 may include cell phones 21 , smart phones 22 , PDAs 23 , tablet notebooks 24 , various desktop and laptop computers 25 , 26 , 27 , and the like.
- One or more of the devices 12 , 14 such as the illustrated iMac and laptop computers 25 , 26 , may connect to the network systems 16 via wireless access point 28 .
- the various transmitting and receiving devices 12 , 14 (one or both types of which being sometimes referred to herein as “client devices”) may be of any conventional design and manufacture.
- FIG. 2 is a block diagram of a communication system 60 in accordance with another preferred embodiment of the present invention.
- This system 60 is similar to the system 10 of FIG. 1 , except that the ASR system 18 of FIG. 1 has been omitted and the ASR engine has instead been incorporated into the various transmitting devices 12 , including cell phones 61 , smart phones 62 , PDAs 63 , tablet notebooks 64 , various desktop and laptop computers 65 , 66 , 67 , and the like.
- the communication systems 10 , 60 each preferably includes, inter alia, a telecommunications network.
- the communications systems 10 , 60 each preferably includes, inter alia, the Internet.
- FIG. 3 is a block diagram of an exemplary implementation of the system 10 of FIG. 1 .
- the transmitting device 12 is a mobile phone
- the ASR system 18 is implemented in one or more backend servers 160
- the one or more network systems 16 include transceiver towers 130 , one or more mobile communication service providers 140 (operating or joint or independent control) and the Internet 150 .
- the backend server 160 is or may be placed in communication with the mobile phone 12 via the mobile communication service provider 140 and the Internet 150 .
- the mobile phone has a microphone, a speaker and a display.
- a first transceiver tower 130 A is positioned between the mobile phone 12 (or the user 32 of the mobile phone 12 ) and the mobile communication service provider 140 , for receiving an audio message (V 1 ), a text message (T 1 ) and/or a verified text message (V/T 1 ) from one of the mobile phone 12 and the mobile communication service provider 140 and transmitting it (V 2 , T 1 , V/T 1 ) to the other of the mobile phone 12 and the mobile communication service provider 140 .
- a second transceiver tower 130 B is positioned between the mobile communication service provider 140 and mobile devices 170 , generally defined as receiving devices 14 equipped to communicate wirelessly via mobile communication service provider 140 , for receiving a verified text message (V/T 1 ) from the mobile communication service provider 140 and transmitting it (V 5 and T 1 ) to the mobile devices 170 .
- the mobile devices 170 are adapted for receiving a text message converted from an audio message created in the mobile phone 12 .
- the mobile devices 170 are also capable of receiving an audio message from the mobile phone 12 .
- the mobile devices 170 include, but are not limited to, a pager, a palm PC, a mobile phone, or the like.
- the system 10 also includes software, as disclosed below in more detail, installed in the mobile phone 12 and the backend server 160 for causing the mobile phone 12 and/or the backend server 160 to perform the following functions.
- the first step is to initialize the mobile phone 12 to establish communication between the mobile phone 12 and the backend server 160 , which includes initializing a desired application from the mobile phone 12 and logging into a user account in the backend server 160 from the mobile phone 12 .
- the user 32 presses and holds one of the buttons of the mobile phone 12 and speaks an utterance, thus generating an audio message, V 1 .
- the audio message V 1 is recorded in the mobile phone 12 .
- the recorded audio message V 1 is sent to the backend server 160 through the mobile communication service provider 140 .
- the recorded audio message V 1 is first transmitted to the first transceiver tower 130 A from the mobile phone 12 .
- the first transceiver tower 130 A outputs the audio message V 1 into an audio message V 2 that is, in turn, transmitted to the mobile communication service provider 140 .
- the mobile communication service provider 140 outputs the audio message V 2 into an audio message V 3 and transmits it (V 3 ) to the Internet 150 .
- the Internet 150 outputs the audio message V 3 into an audio message V 4 and transmits it (V 4 ) to the backend server 160 .
- the content of all the audio messages V 1 -V 4 is identical.
- the backend server 160 then converts the audio message V 4 into a text message, T 1 , and/or a digital signal, D 1 , in the backend server 160 by means of a speech recognition algorithm including a grammar algorithm and/or a transcription algorithm.
- the text message T 1 and the digital signal D 1 correspond to two different formats of the audio message V 4 .
- the text message T 1 and/or the digital signal D 1 are sent back to the Internet 150 that outputs them into a text message T 1 and a digital signal D 2 , respectively.
- the digital signal D 2 is transmitted to a digital receiver 180 , generally defined as a receiving device 14 equipped to communicate with the Internet and capable of receiving the digital signal D 2 .
- the digital receiver 180 is adapted for receiving a digital signal converted from an audio message created in the mobile phone 12 . Additionally, in at least some embodiments, the digital receiver 180 is also capable of receiving an audio message from the mobile phone 12 .
- a conventional computer is one example of a digital receiver 180 .
- a digital signal D 2 may represent, for example, an email or instant message.
- the digital signal D 2 can either be transmitted directly from the backend server 160 or it can be provided back to the mobile phone 12 for review and acceptance by the user 32 before it is sent on to the digital receiver 180 .
- the text message T 1 is sent to the mobile communication service provider 140 that outputs it (T 1 ) into a text message T 1 .
- the output text message T 1 is then transmitted to the first transceiver tower 130 A.
- the first transceiver tower 130 A then transmits it (T 1 ) to the mobile phone 12 in the form of a text message T 1 .
- the substantive content of all the text messages T 1 -T 1 may be identical, which are the corresponding text form of the audio messages V 1 -V 4 .
- the user 32 Upon receiving the text message T 1 , the user 32 verifies it and sends the verified text message V/T 1 to the first transceiver tower 130 A that in turn, transmits it to the mobile communication service provider 140 in the form of a verified text V/T 1 .
- the verified text V/T 1 is transmitted to the second transceiver tower 130 B in the form of a verified text V/T 1 from the mobile communication service provider 140 . Then, the transceiver tower 130 B transmits the verified text V/T 1 to the mobile devices 170 .
- the audio message is simultaneously transmitted to the backend server 160 from the mobile phone 12 , when the user 32 speaks to the mobile phone 12 .
- it is preferred that no audio message is recorded in the mobile phone 12 although it is possible that an audio message could be both transmitted and recorded.
- Such a system may be utilized to convert an audio message into a text message.
- this may be accomplished by first initializing a transmitting device so that the transmitting device is capable of communicating with a backend server 160 .
- a user 32 speaks to or into the client device so as to create a stream of an audio message.
- the audio message can be recorded and then transmitted to the backend server 160 , or the audio message can be simultaneously transmitted to the backend server 160 through a client-server communication protocol.
- Streaming may be accomplished according to processes described elsewhere herein and, in particular, in FIG. 4 , and accompanying text, of the aforementioned U.S. Patent Application Pub. No. US 2007/0239837.
- the transmitted audio message is converted into the text message in the backend server 160 .
- the converted text message is then sent back to the client device 12 .
- the converted text message is forwarded to one or more recipients 34 and their respective receiving devices 14 , where the converted text message may be displayed on the device 14 .
- Incoming messages may be handled, for example, according to processes described elsewhere herein and, in particular, in FIG. 2 , and accompanying text, of the aforementioned U.S. Patent Application Pub. No. US 2007/0239837.
- one or both types of client device 12 , 14 may be located through a global positioning system (GPS); and listing locations, proximate to the position of the client device 12 , 14 , of a target of interest may be presented in the converted text message.
- GPS global positioning system
- FIG. 4 is a block diagram illustrating communications between two users 32 , 34 via a portion of the communication system 10 of FIGS. 1 and 3 .
- a first user 32 sometimes referred to herein as a transmitting user
- a second user 34 sometimes referred to herein as a receiving user
- the transmitting user 32 may send text messages using his transmitting device 12 , for example via SMS
- the receiving user 34 receives text messages on his receiving device 14 , in this case also via SMS.
- the transmitting user 32 may send instant messages via an IM client using his transmitting device 12 , and the receiving user 34 receives instant messages on his receiving device 14 via an IM client.
- the transmitting user 32 preferably speaks into his transmitting device 12 with his utterances being converted to text for communicating to the receiving device 14 , all as more fully described hereinbelow.
- the recorded speech audio is sent to the ASR system 18 , as described previously.
- the utterance 36 is “I really liked the movie snakes on a plane.”
- the ASR engine in the ASR system 18 attempts to recognize and transcribe the utterance 36 into text.
- Speech recognition requests received by the ASR engine may be handled, for example, according to processes described elsewhere herein and, in particular, in FIG. 3 , and accompanying text, of the aforementioned U.S. Patent Application Pub. No. US 2007/0239837. Further, speech recognition may be carried out, for example, according to processes described elsewhere herein and, in particular, in FIGS. 6A-6H , and accompanying text, of the aforementioned U.S. Patent Application Pub. No. US 2007/0239837.
- speech transcription performance indications may be provided to the receiving user 34 in accordance with the disclosure of the aforementioned U.S. patent application Ser. No. 12/197,213.
- the ASR system preferably makes use of both statistical language models (SLMs) for returning results from the audio data, and finite grammars used to post-process the text results, in accordance with the disclosure of the aforementioned U.S. patent application Ser. No. 12/198,112. This is believed to result in messages that are formatted in a way that looks more typical of how a human would have typed the message using a mobile device.
- SLMs statistical language models
- finite grammars used to post-process the text results
- FIG. 5 is a flowchart illustrating a method of updating profile information in a web service in accordance with one or more preferred embodiments of the present invention. Such a method may begin at step 505 with the creation of a profile at each of one or more web services.
- web service may include any website at which user-specific “profile” information is established and maintained by the user, and includes, by way of example, websites offered by Amazon, Facebook, and MySpace. Such information may include personal data, favorites, likes and dislikes, or the like.
- one or more accounts are established for interfacing to user profiles established at the various web services.
- Such accounts may be established at the backend server 160 /ASR system 18 , the user's client device 12 , 14 , or both.
- Accounts may be designated in any of a variety of ways. For example, a user may maintain one account for text messages and one for IMs, or may maintain a single unified account for both types of messages.
- the account or accounts are next configured at step 515 to interface with the user profile at each web service.
- such configuration may be effected by the user by selecting one or more web services from a list of available web services displayed on the client device 12 , 14
- such configuration may be effected by the user by using a browser on the client device 12 , 14 to access the web service and select an option for such configuration from the web service.
- each web service makes use of a standard protocol by which one or both of the backend server 160 /ASR system 18 and the user's device 12 , 14 may communicate with the web service to update the user profile.
- a browser on the client device 12 , 14 may be utilized to access the web service and download a protocol specific to that web service.
- the various configurations are organized and managed in one or more user accounts that correspond, for example, to the client device 12 , 14 .
- preferences may be established for the configuration. These may, for example, be established directly via a user interface at the client device 12 , 14 or indirectly at the web service via a browser on the client device 12 , 14 or via a browser on a separate device. Preferences may include types of filters or the like to be employed as part of a “profile filter” described below, groups of web service profiles to be updated, message types (e.g., text messages, IMs, other messages, or the like, or a combination thereof) or utterance types to be considered, and the like. In at least one embodiment, default preferences are provided and utilized until if and when the user chooses to update the preferences.
- the method may be used to examine message strings for relevant information as shown at step 525 .
- the backend server 160 /ASR system 18 may further include a profile filter for processing the text results thereof. Specifically, as transcribed text result is produced by the ASR system 18 , whether the result is a message for communication to one or more other users and/or to one or more web services, or is some other type of transcription, the transcribed text result is parsed in order to identify keywords, fragments, or phrases that may represent relevant personal preference information. Such identification process may include, for example, keyword or grammar lookups, natural language understanding, semantic analysis, or other techniques in order to derive interestingness for further processing.
- the filter may constitute, at least in part, one or more of those filters found in the disclosure of one or more of the patent applications incorporated by reference herein.
- the identification process also may include audio fingerprinting or audio watermarking, which may involve placing human-inaudible audio artifacts in an audio stream that can carry identification or configuration information. Audio fingerprinting or audio watermarking may help the backend server 160 /ASR system 18 select the type of noise suppression done or may help it select from a given acoustic model (for example, by providing an indication as to what accent an individual is most likely to have). This may be particularly useful for client-less applications such as voicemail, where the chipset can tag these things which are eventually picked up by the backend server 160 /ASR system 18 after it traverses the normal carrier audio factories. It may be desirable to have hidden parameters that would normally be passed if the audio data originated from a corresponding application on a client device.
- a profile filter may be implemented on the client device 12 , 14 , whether or not an ASR engine is present in the device 12 , 14 . Still further, it will be appreciated that, in the context of text messaging, a profile filter may be implemented at the mobile communication service provider 140 , and in the context of instant messaging, a profile filter may be implemented at an IM service provider (not specifically illustrated).
- a separate ASR system 18 provides a convenient platform at which the profile filter may be disposed.
- a profile filter may additionally or alternatively be disposed at a transmitting device 12 .
- the transmitting user 32 can use a keyboard, keypad or other user input device on the transmitting device 12 to manually edit the transcription results before transmission.
- the user may choose to enter the entire intended message using such user input device on the transmitting device 12 . In either case, the manually-edited or -created message may then be processed by a profile filter on the transmitting device 12 .
- a profile filter may be implemented at a receiving device 14 , such that incoming text messages, IMs and other message strings may be processed in a manner similar to that of transcribed utterances or outgoing messages strings.
- web service user profiles that are linked to the user account(s) can then be updated dynamically at step 530 as a function of the keywords, fragments, or phrases identified at step 525 .
- the message is first transcribed by the ASR engine and then processed by the profile filter. Assuming an accurate transcription is obtained, then as the message passes through the profile filter, the keyword “movie” may be identified and the phrase “snakes on a plane” may be further identified using a client, server, or web based database of current and past movies.
- Analysis of the message string further indicates that there is some likelihood that the user has an interest in the movie “Snakes on a Plane.” If the user's preferences have been set to take action when this type of information (the type of message, the existence of the user's interest, the name or type of movie, or the like) is identified (or alternatively, if the user's preferences have been set to look for this type of information to begin with), then such interest in the movie “Snakes on a Plane” is dynamically posted, as appropriate, to the user's social networking profile pages that are linked to the user's client device account(s). As noted previously, the user's account on the client device 12 , 14 may be an instant messaging account, a text messaging account, a unified account, or the like. Such dynamic updating of the user's social network profile page then would enable their contacts (having suitable permissions to access this data) to ask for that user's favorite films and get an automated response from this dynamically populated information.
- ad impressions further can be targeted to the user based on the identified keyword and phrase, such as ad impressions relating to movie rentals for “Snakes on a Plane” or movie times of a local theater for showings of “Snakes on a Plane.”
- the advertisements may be pushed either prior to message strings actually being sent to recipients or to web services, or thereafter, as applicable.
- the advertising that is pushed to a user's mobile device preferably comprises an ad impression that is displayed to the user in the form of an ad bubble.
- the ad impression elements may contain text, graphics, videos, and/or audio and may be downloaded from a server infrastructure or may already be resident within the mobile device and accessed directly there from.
- each ad impression is designed to be as unobtrusive as possible to the user and allows the user to view or hear the advertisement or take some further action regarding the advertisement, if and as desired by the user, which may include opening a separate mobile browser with additional content relevant to the advertisement.
- the ad impression may be delivered only to the author of the message string.
- the ad impression may be delivered both to the author of the message string and to the intended recipient of the message string, especially where the message string is intended to be sent to the mobile device of another user.
- the ad impression is sent to either of, but not both of, the author and intended recipient, then such person may be provided with the option of conveniently forwarding the ad impression to the other person if desired, whether by text message, instant message, email, hyperlink, or injection of the ad impression into a message string itself.
- ad impression In taking further action with regard to an ad impression that is presented to a user, if desired, such user having seen or heard the ad impression may manually click on a displayed advertisement or portion thereof resulting in, for example, the launch of a mobile browser. The mobile browser may then allow the user to either complete a purchase or find relevant information associated with the advertisement. Moreover, rather than manually clicking on the displayed advertisement, the user may speak a keyword as a “voice click,” thereby resulting in the further action being taken. Such use of “voice click” may be in accordance with the disclosure of the aforementioned U.S. patent application Ser. No. 12/198,116, which is hereby incorporated herein by reference.
- indexes for storing, in association with the particular user involved, some or all of the profile information that has been parsed from the message string.
- the index or indexes may include databases, grammars, language models, or the like.
- profile information As profile information is identified, it may be stored in the appropriate index. If no index exists for the particular user, then it may be created automatically as profile information for the user is gathered.
- the index or indexes are stored at the backend server 160 /ASR system 18 and updated directly by the profile filter or other element of the system 18 .
- corresponding indexes are maintained on the client devices 12 , 14 and synchronized at appropriate times with the system index. Synchronization may be accomplished by transmitting, from the client device 12 , 14 to the system 18 , a delta model representing the differences between the new client device indexes (as updated most recently with profile information) and the last-synchronized information in the client device indexes. Use of delta models enables time and bandwidth to be conserved in the synchronization process. Still further, in at least one other embodiment, the indexes are maintained only on the client devices 12 , 14 .
- the index or indexes may be used as a specific interface point for the web services where the user maintains profiles to be updated according to one or more of the methods and systems disclosed herein. More particularly, updated profile information may be placed in the index or indexes, and a separate process may be used to provide profile information from the index or indexes to the web services. These two separate processes may occur synchronously or asynchronously.
- index(es) may be updated to include static profile information as well as the dynamic profile information derived as described herein.
- the index or indexes may be separately queried by one or more users. Any of a variety of means may be utilized to establish which users are to be given access to some or all of a particular user's profile information in the index(es).
- any user (or corresponding user device) in a particular user's contact list, as stored in the particular user's client device 12 , 14 may be permitted to query the index(es).
- a user's contacts are allowed to query the user's preferences whereby they ask for areas of known content, such as the user's favorite bands or movies.
- the preferences/profile information could include both dynamic profile information and static profile information. This blend of static and dynamic profile data could also be utilized to target ads and/or promotions.
- the first user 32 in FIG. 4 may have established preferences, at step 520 in FIG. 5 , for movie preferences to be included in his or her profile information and maintained in an index maintained for him or her. Thereafter, a transcription of the utterance 36 of first user 32 in FIG.
- a manually-entered message string similar thereto may trigger an additional entry in the profile information index for that user 32 , where the entry indicates an interest in the movie “Snakes on a Plane.” If another user, authorized by the first user 32 to query profile information in the first user's profile information index, subsequently wishes to learn what movies the first user 32 likes (perhaps as part of researching an appropriate birthday gift or the like), the other user may query the profile information index.
- the query may be accomplished via a plain language query (for example, using manually-entered text or making use of the ASR engine), via a special user interface, or any other suitable means.
- the Yap service includes one or more web applications and a client device application.
- the Yap web application is a J2EE application built using Java 5. It is designed to be deployed on an application server like IBM WebSphere Application Server or an equivalent J2EE application server. It is designed to be platform neutral, meaning the server hardware and OS can be anything supported by the web application server (e.g. Windows, Linux, MacOS X).
- FIG. 6 is a block diagram of the system architecture of the Yap commercial implementation.
- the operating system may be implemented in Red Hat Enterprise Linux 5 (RHEL 5);
- the application servers may include the Websphere Application Server Community Edition (WAS-CE) servers, available from IBM;
- the web server may be an Apache server;
- the CTTS Servlets may include CTTS servlets from Loquendo, including US/UK/ES male and US/UK/ES female;
- the Grammar ASP may be the latest WebSphere Voice Server, available from IBM;
- suitable third party ads may be provided by Google; a suitable third party IM system is Google Talk, available from Google; and a suitable database system is the DB2 Express relational database system, available from IBM.
- FIG. 7 is a block diagram of the Yap EAR of FIG. 6 .
- the audio codec JARs may include the VoiceAge AMR JAR, available from VoiceAge of Montreal, Quebec and/or the QCELP JAR, available from Qualcomm of San Diego, Calif.
- the Yap web application includes a plurality of servlets.
- servlet refers to an object that receives a request and generates a response based on the request.
- a servlet is a small Java program that runs within a Web server.
- Servlets receive and respond to requests from Web clients, usually across HTTP and/or HTTPS, the HyperText Transfer Protocol.
- the Yap web application includes nine servlets: Correct, Debug, Install, Login, Notify, Ping, Results, Submit, and TTS. Each servlet is described below in the order typically encountered.
- the communication protocol used for all messages between the Yap client and Yap server applications is HTTP and HTTPS.
- HTTP and HTTPS The communication protocol used for all messages between the Yap client and Yap server applications.
- HTTP and HTTPS are standard web protocols.
- Using these standard web protocols allows the Yap web application to fit well in a web application container. From the application server's point of view, it cannot distinguish between the Yap client midlet and a typical web browser. This aspect of the design is intentional to convince the web application server that the Yap client midlet is actually a web browser.
- the Yap client uses the POST method and custom headers to pass values to the server.
- the body of the HTTP message in most cases is irrelevant with the exception of when the client submits audio data to the server in which case the body contains the binary audio data.
- the Server responds with an HTTP code indicating the success or failure of the request and data in the body which corresponds to the request being made.
- the server does not depend on custom header messages being delivered to the client as the carriers can, and usually do, strip out unknown header values.
- FIG. 8 is a typical header section of an HTTP request from the Yap client.
- the Yap client is operated via a user interface (UI), known as “Yap9,” which is well suited for implementing methods of converting an audio message into a text message and messaging in mobile environments.
- UI user interface
- Yap9 is a combined UI for SMS and web services (WS) that makes use of the buttons or keys of the client device by assigning a function to each button (sometimes referred to as a “Yap9” button or key). Execution of such functions is carried out by “Yaplets.” This process, and the usage of such buttons, are described elsewhere herein and, in particular, in FIGS. 9A-9D , and accompanying text, of the aforementioned U.S. Patent Application Pub. No. US 2007/0239837.
- the install fails, or the install is canceled by the user, the Notify servlet is sent a message by the phone with a short description. This can be used for tracking purposes and to help diagnose any install problems.
- the first step is to create a new session by logging into the Yap web application using the Login servlet.
- multiple login servers exist, so as a preliminary step, a request is sent to find a server to log in to. Exemplary protocol details for such a request can be seen in FIG. 9 .
- An HTTP string pointing to a selected login server will be returned in response to this request. It will be appreciated that this selection process functions as a poor man's load balancer.
- a login request is sent. Exemplary protocol details for such a request can be seen in FIG. 10 .
- a cookie holding a session ID is returned in response to this request.
- the session ID is a pointer to a session object on the server which holds the state of the session. This session data will be discarded after a period determined by server policy.
- Sessions are typically maintained using client-side cookies, however, a user cannot rely on the set-cookie header successfully returning to the Yap client because the carrier may remove that header from the HTTP response.
- the solution to this problem is to use the technique of URL rewriting. To do this, the session ID is extracted from the session API, which is returned to the client in the body of the response. This is called the “Yap Cookie” and is used in every subsequent request from the client.
- the Yap Cookie looks like this:
- audio data may be submitted.
- the user presses and holds one of the Yap-9 buttons, speaks aloud, and releases the pressed button.
- the speech is recorded, and the recorded speech is then sent in the body of a request to the Submit servlet, which returns a unique receipt that the client can use later to identify this utterance. Exemplary protocol details for such a request can be seen in FIG. 11 .
- One of the header values sent to the server during the login process is the format in which the device records. That value is stored in the session so the Submit servlet knows how to convert the audio into a format required by the ASR engine. This is done in a separate thread as the process can take some time to complete.
- the Yap9 button and Yap9 screen numbers are passed to the Submit server in the HTTP request header. These values are used to lookup a user-defined preference of what each button is assigned to. For example, the 1 button may be used to transcribe audio for an SMS message, while the 2 button is designated for a grammar based recognition to be used in a web services location based search.
- the Submit servlet determines the appropriate “Yaplet” to use. When the engine has finished transcribing the audio or matching it against a grammar, the results are stored in a hash table in the session.
- filters can be applied to the text returned from the ASR engine.
- filters may include, but are not limited to, those shown Table 3.
- Engine Used to remove speech engine words Filter Number Used to convert the spelled out numbers returned Filter from the speech engine into a digit based number (e.g., “one hundred forty seven” ⁇ > “147”)
- Obscenity Used to place asterisks in for the vowels in Filter street slang e.g., “sh*t”, “f*ck”, etc.
- Time Filter Used to format time phrases Notably, after all of the filters are applied
- the client retrieves the results of the audio by taking the receipt returned from the Submit servlet and submitting it as a request to the Results servlet.
- Exemplary protocol details for such a request can be seen in FIG. 12 . This is done in a separate thread on the device and a timeout parameter may be specified which will cause the request to return after a certain amount of time if the results are not available.
- a block of XML is preferably returned.
- Exemplary protocol details for such a return response can be seen in FIG. 13 .
- a serialized Java Results object may be returned.
- This object contains a number of getter functions for the client to extract the type of results screen to advance to (i.e., SMS or results list), the text to display, the text to be used for TTS, any advertising text to be displayed, an SMS trailer to append to the SMS message, etc.
- the client to extract the type of results screen to advance to (i.e., SMS or results list), the text to display, the text to be used for TTS, any advertising text to be displayed, an SMS trailer to append to the SMS message, etc.
- TTS Transmission Control Protocol
- the TTS string is extracted from the results and sent via an HTTP request to the TTS servlet. Exemplary protocol details for such a request can be seen in FIG. 14 .
- the request blocks until the TTS is generated and returns audio in the format supported by the phone in the body of the result. This is performed in a separate thread on the device since the transaction may take some time to complete.
- the resulting audio is then played to the user through the AudioService object on the client.
- TTS speech from the server is encrypted using Corrected Block Tiny Encryption Algorithm (XXTEA) encryption.
- XXTEA Corrected Block Tiny Encryption Algorithm
- the corrected text is submitted to the Correct servlet along with the receipt for the request.
- This information is stored on the server for later use in analyzing accuracy and compiling a database of typical SMS messages. Exemplary protocol details for such a submission can be seen in FIG. 15 .
- the Ping servlet can be used to send a quick message from the client to keep the session alive. Exemplary protocol details for such a message can be seen in FIG. 16 .
- the Debug servlet sends logging messages from the client to a debug log on the server. Exemplary protocol details can be seen in FIG. 17 .
- an HTTP logout request needs to be issued to the server.
- the Yap website has a section where the user can log in and customize their Yap client preferences. This allows them to choose from available Yaplets and assign them to Yap9 keys on their phone.
- the user preferences are stored and maintained on the server and accessible from the Yap web application. This frees the Yap client from having to know about all of the different back-end Yaplets. It just records the audio, submits it to the server along with the Yap9 key and Yap9 screen used for the recording and waits for the results.
- the server handles all of the details of what the user actually wants to have happen with the audio.
- the client needs to know what type of format to utilize when presenting the results to the user. This is accomplished through a code in the Results object.
- the majority of requests fall into one of two categories: sending an SMS message, or displaying the results of a web services query in a list format. Notably, although these two are the most common, the Yap architecture supports the addition of new formats.
Abstract
Description
- (1) U.S. provisional patent application Ser. No. 60/972,851, filed Sep. 17, 2007 and titled “SYSTEM AND METHOD FOR DELIVERING MOBILE ADVERTISING WITHIN A THREADED SMS OR IM CHAT CONVERSATION ON A MOBILE DEVICE CLIENT”;
- (2) U.S. provisional patent application Ser. No. 60/972,853, filed Sep. 17, 2007 and titled “METHOD AND SYSTEM FOR DYNAMIC PERSONALIZATION AND QUERYING OF USER PROFILES BASED ON SMS/IM CHAT MESSAGING ON A MOBILE DEVICE”;
- (3) U.S. provisional patent application Ser. No. 60/972,854, filed Sep. 17, 2007 and titled “LOCATION, TIME & SEASON AWARE MOBILE ADVERTISING DELIVERY”;
- (4) U.S. provisional patent application Ser. No. 60/972,936, filed Sep. 17, 2007 and titled “DELIVERING TARGETED ADVERTISING TO MOBILE DEVICE FOR PRESENTATION WITHIN SMSes OR IM CONVERSATIONS”;
- (5) U.S. provisional patent application Ser. No. 60/972,943, filed Sep. 17, 2007 and titled “DYNAMIC PERSONALIZATION AND QUERYING OF USER PROFILES BASED ON SMSes AND IM CONVERSATIONS”; and
- (6) U.S. provisional patent application Ser. No. 60/972,944, filed Sep. 17, 2007 and titled “LOCATION, TIME, AND SEASON AWARE ADVERTISING DELIVERY TO AND PRESENTATION ON MOBILE DEVICE WITHIN SMSes OR IM CONVERSATIONS OR USER INTERFACE THEREOF”.
-
- ;jsessionid=C240B217F2351E3C420A599B0878371A
-
- /Yap/Submit;jsessionid=C240B217F2351E3C420A599B0878371A
TABLE 3 | |
Filter Type | Function |
Ad Filter | Used to scan the text and identify keywords that |
can be used to insert targeted advertising | |
messages, and/or convert the keywords into | |
hyperlinks to ad sponsored web pages | |
Currency | Used to format currency returned from the speech |
Filter | engine into the user's preferred format. (e.g., |
“one hundred twenty dollars” −> “$120.00”) | |
Date Filter | Used to format dates returned from the speech |
engine into the user's preferred format. (e.g., | |
“march fourth two thousand seven” −> “3/4/2007”) | |
Digit Filter | User to format spelled out single digits returned |
from the speech engine into a multi-digit number | |
such as a zip code (e.g., “two eight two one | |
one” −> “28211”) | |
Engine | Used to remove speech engine words |
Filter | |
Number | Used to convert the spelled out numbers returned |
Filter | from the speech engine into a digit based number |
(e.g., “one hundred forty seven” −> “147”) | |
Obscenity | Used to place asterisks in for the vowels in |
Filter | street slang (e.g., “sh*t”, “f*ck”, etc.) |
Punctuation | Used to format punctuation |
Filter | |
SMS Filter | Used to convert regular words into a spelling |
which more closely resembles an SMS message | |
(e.g., “don't forget to smile” −> “don't 4get 2 | |
:)”, etc.) | |
Time Filter | Used to format time phrases |
Notably, after all of the filters are applied, both the filtered text and original text are returned to the client so that if text to speech is enabled for the user, the original unfiltered text can be used to generate the TTS audio.
Claims (32)
Priority Applications (5)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US12/212,644 US9973450B2 (en) | 2007-09-17 | 2008-09-17 | Methods and systems for dynamically updating web service profile information by parsing transcribed message strings |
US13/620,716 US9037473B2 (en) | 2008-01-16 | 2012-09-15 | Using a physical phenomenon detector to control operation of a speech recognition engine |
US14/081,983 US9330401B2 (en) | 2007-04-05 | 2013-11-15 | Validation of mobile advertising from derived information |
US14/341,054 US9384735B2 (en) | 2007-04-05 | 2014-07-25 | Corrective feedback loop for automated speech recognition |
US15/201,188 US9940931B2 (en) | 2007-04-05 | 2016-07-01 | Corrective feedback loop for automated speech recognition |
Applications Claiming Priority (7)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US97285407P | 2007-09-17 | 2007-09-17 | |
US97285107P | 2007-09-17 | 2007-09-17 | |
US97293607P | 2007-09-17 | 2007-09-17 | |
US97294307P | 2007-09-17 | 2007-09-17 | |
US97294407P | 2007-09-17 | 2007-09-17 | |
US97285307P | 2007-09-17 | 2007-09-17 | |
US12/212,644 US9973450B2 (en) | 2007-09-17 | 2008-09-17 | Methods and systems for dynamically updating web service profile information by parsing transcribed message strings |
Publications (2)
Publication Number | Publication Date |
---|---|
US20090083032A1 US20090083032A1 (en) | 2009-03-26 |
US9973450B2 true US9973450B2 (en) | 2018-05-15 |
Family
ID=40472647
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US12/212,644 Active 2033-01-27 US9973450B2 (en) | 2007-04-05 | 2008-09-17 | Methods and systems for dynamically updating web service profile information by parsing transcribed message strings |
Country Status (1)
Country | Link |
---|---|
US (1) | US9973450B2 (en) |
Cited By (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US10714080B2 (en) * | 2017-02-10 | 2020-07-14 | Samsung Electronics Co., Ltd. | WFST decoding system, speech recognition system including the same and method for storing WFST data |
US11803532B2 (en) | 2017-01-04 | 2023-10-31 | Palantir Technologies Inc. | Integrated data analysis |
Families Citing this family (35)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20090124272A1 (en) | 2006-04-05 | 2009-05-14 | Marc White | Filtering transcriptions of utterances |
EP2008193B1 (en) | 2006-04-05 | 2012-11-28 | Canyon IP Holdings LLC | Hosted voice recognition system for wireless devices |
US9436951B1 (en) | 2007-08-22 | 2016-09-06 | Amazon Technologies, Inc. | Facilitating presentation by mobile device of additional content for a word or phrase upon utterance thereof |
US8510109B2 (en) | 2007-08-22 | 2013-08-13 | Canyon Ip Holdings Llc | Continuous speech transcription performance indication |
US8326636B2 (en) * | 2008-01-16 | 2012-12-04 | Canyon Ip Holdings Llc | Using a physical phenomenon detector to control operation of a speech recognition engine |
US20090076917A1 (en) * | 2007-08-22 | 2009-03-19 | Victor Roditis Jablokov | Facilitating presentation of ads relating to words of a message |
US8352264B2 (en) | 2008-03-19 | 2013-01-08 | Canyon IP Holdings, LLC | Corrective feedback loop for automated speech recognition |
US8611871B2 (en) | 2007-12-25 | 2013-12-17 | Canyon Ip Holdings Llc | Validation of mobile advertising from derived information |
US8352261B2 (en) * | 2008-03-07 | 2013-01-08 | Canyon IP Holdings, LLC | Use of intermediate speech transcription results in editing final speech transcription results |
US8238528B2 (en) * | 2007-06-29 | 2012-08-07 | Verizon Patent And Licensing Inc. | Automatic analysis of voice mail content |
US7996473B2 (en) * | 2007-07-30 | 2011-08-09 | International Business Machines Corporation | Profile-based conversion and delivery of electronic messages |
US9053489B2 (en) | 2007-08-22 | 2015-06-09 | Canyon Ip Holdings Llc | Facilitating presentation of ads relating to words of a message |
US8296377B1 (en) | 2007-08-22 | 2012-10-23 | Canyon IP Holdings, LLC. | Facilitating presentation by mobile device of additional content for a word or phrase upon utterance thereof |
US8136034B2 (en) | 2007-12-18 | 2012-03-13 | Aaron Stanton | System and method for analyzing and categorizing text |
US8676577B2 (en) * | 2008-03-31 | 2014-03-18 | Canyon IP Holdings, LLC | Use of metadata to post process speech recognition output |
US20100037288A1 (en) * | 2008-08-06 | 2010-02-11 | International Business Machines Corporation | Inherited Access Authorization to a Social Network |
US8301454B2 (en) | 2008-08-22 | 2012-10-30 | Canyon Ip Holdings Llc | Methods, apparatuses, and systems for providing timely user cues pertaining to speech recognition |
US20100081461A1 (en) * | 2008-10-01 | 2010-04-01 | Microsoft Corporation | SMS Based Social Networking |
US20100094627A1 (en) * | 2008-10-15 | 2010-04-15 | Concert Technology Corporation | Automatic identification of tags for user generated content |
US20110029618A1 (en) * | 2009-08-02 | 2011-02-03 | Hanan Lavy | Methods and systems for managing virtual identities in the internet |
WO2011101848A1 (en) * | 2010-02-18 | 2011-08-25 | United Parents Online Ltd. | Methods and systems for managing virtual identities |
EP2306339A1 (en) * | 2009-09-23 | 2011-04-06 | Adobe Systems Incorporated | Algorith and implementation for fast computation of content recommendation |
US9191509B2 (en) * | 2009-11-12 | 2015-11-17 | Collider Media | Multi-source profile compilation for delivering targeted content |
US8510653B2 (en) * | 2010-11-15 | 2013-08-13 | Yahoo! Inc. | Combination creative advertisement targeting system |
US20120150592A1 (en) * | 2010-12-10 | 2012-06-14 | Endre Govrik | Systems and methods for user marketing and endorsement on social networks |
US20120197648A1 (en) * | 2011-01-27 | 2012-08-02 | David Moloney | Audio annotation |
US9009041B2 (en) * | 2011-07-26 | 2015-04-14 | Nuance Communications, Inc. | Systems and methods for improving the accuracy of a transcription using auxiliary data such as personal data |
US20130332307A1 (en) * | 2011-11-21 | 2013-12-12 | Facebook, Inc. | Method for notifying a sender of a gifting event |
US20150269672A1 (en) * | 2014-03-21 | 2015-09-24 | Hybrid Tittan Management, Llc | Trading platform currently known as alphametrics and it's accompanying api (application programming interface) for its usage; to include a voice recognition software platform designed to aggregate end of day order imbalance sentiment for nyse traded issues |
US10600085B2 (en) * | 2014-05-15 | 2020-03-24 | Alan Rodriguez | Systems and methods for communicating privacy and marketing preferences |
JP2018191145A (en) * | 2017-05-08 | 2018-11-29 | オリンパス株式会社 | Voice collection device, voice collection method, voice collection program, and dictation method |
US10839008B2 (en) * | 2017-07-06 | 2020-11-17 | Sync Floor, Inc. | System and method for natural language music search |
US11494797B1 (en) | 2018-03-08 | 2022-11-08 | Inmar Clearing, Inc. | Electronic system including electronic message based electronic shopping list generation and related methods |
US11487720B2 (en) | 2018-05-08 | 2022-11-01 | Palantir Technologies Inc. | Unified data model and interface for databases storing disparate types of data |
KR102371568B1 (en) * | 2019-10-18 | 2022-03-07 | 주식회사 카카오 | Method of displaying profile view in instant messaging service |
Citations (342)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US5036538A (en) | 1989-11-22 | 1991-07-30 | Telephonics Corporation | Multi-station voice recognition and processing system |
US5623609A (en) | 1993-06-14 | 1997-04-22 | Hal Trust, L.L.C. | Computer system and computer-implemented process for phonology-based automatic speech recognition |
US5675507A (en) | 1995-04-28 | 1997-10-07 | Bobo, Ii; Charles R. | Message storage and delivery system |
US5822730A (en) | 1996-08-22 | 1998-10-13 | Dragon Systems, Inc. | Lexical tree pre-filtering in speech recognition |
US5852801A (en) | 1995-10-04 | 1998-12-22 | Apple Computer, Inc. | Method and apparatus for automatically invoking a new word module for unrecognized user input |
US5864603A (en) | 1995-06-02 | 1999-01-26 | Nokia Mobile Phones Limited | Method and apparatus for controlling a telephone with voice commands |
US5948061A (en) | 1996-10-29 | 1999-09-07 | Double Click, Inc. | Method of delivery, targeting, and measuring advertising over networks |
US5974413A (en) | 1997-07-03 | 1999-10-26 | Activeword Systems, Inc. | Semantic user interface |
US5995928A (en) | 1996-10-02 | 1999-11-30 | Speechworks International, Inc. | Method and apparatus for continuous spelling speech recognition with early identification |
US6026368A (en) | 1995-07-17 | 2000-02-15 | 24/7 Media, Inc. | On-line interactive system and method for providing content and advertising information to a targeted set of viewers |
US6100882A (en) | 1994-01-19 | 2000-08-08 | International Business Machines Corporation | Textual recording of contributions to audio conference using speech recognition |
US6173259B1 (en) | 1997-03-27 | 2001-01-09 | Speech Machines Plc | Speech to text conversion |
US6212498B1 (en) | 1997-03-28 | 2001-04-03 | Dragon Systems, Inc. | Enrollment in speech recognition |
US6219407B1 (en) | 1998-01-16 | 2001-04-17 | International Business Machines Corporation | Apparatus and method for improved digit recognition and caller identification in telephone mail messaging |
US6219638B1 (en) | 1998-11-03 | 2001-04-17 | International Business Machines Corporation | Telephone messaging and editing system |
US6253177B1 (en) | 1999-03-08 | 2001-06-26 | International Business Machines Corp. | Method and system for automatically determining whether to update a language model based upon user amendments to dictated text |
US6298326B1 (en) | 1999-05-13 | 2001-10-02 | Alan Feller | Off-site data entry system |
US20010047294A1 (en) | 2000-01-06 | 2001-11-29 | Rothschild Anthony R. | System and method for adding an advertisement to a personal communication |
US20010056350A1 (en) | 2000-06-08 | 2001-12-27 | Theodore Calderone | System and method of voice recognition near a wireline node of a network supporting cable television and/or video delivery |
US20010056369A1 (en) | 2000-06-16 | 2001-12-27 | Kuniharu Takayama | Advertisement posting system, advertisement-cost calculating method, and record medium storing advertisement-cost calculating program |
US20020016712A1 (en) | 2000-07-20 | 2002-02-07 | Geurts Lucas Jacobus Franciscus | Feedback of recognized command confidence level |
US20020029101A1 (en) | 2000-09-05 | 2002-03-07 | Hunter Engineering Company | Method and apparatus for networked wheel alignment communications and services |
US20020035474A1 (en) | 2000-07-18 | 2002-03-21 | Ahmet Alpdemir | Voice-interactive marketplace providing time and money saving benefits and real-time promotion publishing and feedback |
US6366886B1 (en) | 1997-04-14 | 2002-04-02 | At&T Corp. | System and method for providing remote automatic speech recognition services via a packet network |
US20020052781A1 (en) | 1999-09-10 | 2002-05-02 | Avantgo, Inc. | Interactive advertisement mechanism on a mobile device |
US6401075B1 (en) | 2000-02-14 | 2002-06-04 | Global Network, Inc. | Methods of placing, purchasing and monitoring internet advertising |
US20020087330A1 (en) | 2001-01-03 | 2002-07-04 | Motorola, Inc. | Method of communicating a set of audio content |
US20020091570A1 (en) | 2000-12-01 | 2002-07-11 | Hiroaki Sakagawa | Electronic mail advertisement system, method, and program storage medium |
US6453290B1 (en) | 1999-10-04 | 2002-09-17 | Globalenglish Corporation | Method and system for network-based speech recognition |
US20020161579A1 (en) | 2001-04-26 | 2002-10-31 | Speche Communications | Systems and methods for automated audio transcription, translation, and transfer |
US20020165773A1 (en) | 2000-05-31 | 2002-11-07 | Takeshi Natsuno | Method and system for distributing advertisements over network |
US20020165719A1 (en) | 2001-05-04 | 2002-11-07 | Kuansan Wang | Servers for web enabled speech recognition |
US6490561B1 (en) | 1997-06-25 | 2002-12-03 | Dennis L. Wilson | Continuous speech voice transcription |
EP1274222A2 (en) | 2001-07-02 | 2003-01-08 | Nortel Networks Limited | Instant messaging using a wireless interface |
US20030008661A1 (en) | 2001-07-03 | 2003-01-09 | Joyce Dennis P. | Location-based content delivery |
US20030028601A1 (en) | 2001-07-31 | 2003-02-06 | Rowe Lorin Bruce | Method and apparatus for providing interactive text messages during a voice call |
US6519562B1 (en) | 1999-02-25 | 2003-02-11 | Speechworks International, Inc. | Dynamic semantic control of a speech recognition system |
US6532446B1 (en) | 1999-11-24 | 2003-03-11 | Openwave Systems Inc. | Server based speech recognition user interface for wireless devices |
US20030050778A1 (en) | 2001-09-13 | 2003-03-13 | Patrick Nguyen | Focused language models for improved speech input of structured documents |
US20030093315A1 (en) | 2000-09-26 | 2003-05-15 | Kenji Sato | System and method for using e-mail as advertisement medium |
US6571210B2 (en) | 1998-11-13 | 2003-05-27 | Microsoft Corporation | Confidence measure system using a near-miss pattern |
US20030101054A1 (en) | 2001-11-27 | 2003-05-29 | Ncc, Llc | Integrated system and method for electronic speech recognition and transcription |
US20030105630A1 (en) | 2001-11-30 | 2003-06-05 | Macginitie Andrew | Performance gauge for a distributed speech recognition system |
US20030115060A1 (en) | 2001-12-13 | 2003-06-19 | Junqua Jean-Claude | System and interactive form filling with fusion of data from multiple unreliable information sources |
US20030126216A1 (en) | 2001-09-06 | 2003-07-03 | Avila J. Albert | Method and system for remote delivery of email |
US20030125955A1 (en) | 2001-12-28 | 2003-07-03 | Arnold James F. | Method and apparatus for providing a dynamic speech-driven control and remote service access system |
US20030139922A1 (en) | 2001-12-12 | 2003-07-24 | Gerhard Hoffmann | Speech recognition system and method for operating same |
US20030144906A1 (en) | 2002-01-31 | 2003-07-31 | Nissan Motor Co., Ltd. | Advertisement distribution method, advertisement distribution apparatus and advertisement displaying vehicle |
US20030149566A1 (en) | 2002-01-02 | 2003-08-07 | Esther Levin | System and method for a spoken language interface to a large database of changing records |
US20030182113A1 (en) | 1999-11-22 | 2003-09-25 | Xuedong Huang | Distributed speech recognition for mobile communication devices |
US20030187643A1 (en) | 2002-03-27 | 2003-10-02 | Compaq Information Technologies Group, L.P. | Vocabulary independent speech decoder system and method using subword units |
US20030191639A1 (en) | 2002-04-05 | 2003-10-09 | Sam Mazza | Dynamic and adaptive selection of vocabulary and acoustic models based on a call context for speech recognition |
US20030200086A1 (en) | 2002-04-17 | 2003-10-23 | Pioneer Corporation | Speech recognition apparatus, speech recognition method, and computer-readable recording medium in which speech recognition program is recorded |
US20030200093A1 (en) | 1999-06-11 | 2003-10-23 | International Business Machines Corporation | Method and system for proofreading and correcting dictated text |
US20030212554A1 (en) | 2002-05-09 | 2003-11-13 | Vatland Danny James | Method and apparatus for processing voice data |
US6654448B1 (en) | 1998-06-19 | 2003-11-25 | At&T Corp. | Voice messaging system |
US20030220798A1 (en) | 2002-05-24 | 2003-11-27 | Microsoft Corporation | Speech recognition status feedback user interface |
US20030220792A1 (en) | 2002-05-27 | 2003-11-27 | Pioneer Corporation | Speech recognition apparatus, speech recognition method, and computer-readable recording medium in which speech recognition program is recorded |
US20030223556A1 (en) | 2002-05-29 | 2003-12-04 | Yun-Cheng Ju | Electronic mail replies with speech recognition |
US20040005877A1 (en) | 2000-08-21 | 2004-01-08 | Vaananen Mikko Kalervo | Voicemail short massage service method and means and a subscriber terminal |
US20040015547A1 (en) | 2002-07-17 | 2004-01-22 | Griffin Chris Michael | Voice and text group chat techniques for wireless mobile terminals |
US20040019488A1 (en) | 2002-07-23 | 2004-01-29 | Netbytel, Inc. | Email address recognition using personal information |
US6687689B1 (en) | 2000-06-16 | 2004-02-03 | Nusuara Technologies Sdn. Bhd. | System and methods for document retrieval using natural language-based queries |
US6687339B2 (en) | 1997-12-31 | 2004-02-03 | Weblink Wireless, Inc. | Controller for use with communications systems for converting a voice message to a text message |
US6704034B1 (en) | 2000-09-28 | 2004-03-09 | International Business Machines Corporation | Method and apparatus for providing accessibility through a context sensitive magnifying glass |
US20040059712A1 (en) | 2002-09-24 | 2004-03-25 | Dean Jeffrey A. | Serving advertisements using information associated with e-mail |
US20040059708A1 (en) | 2002-09-24 | 2004-03-25 | Google, Inc. | Methods and apparatus for serving relevant advertisements |
US20040059632A1 (en) | 2002-09-23 | 2004-03-25 | International Business Machines Corporation | Method and system for providing an advertisement based on an URL and/or a search keyword entered by a user |
US20040107107A1 (en) | 2002-12-03 | 2004-06-03 | Philip Lenir | Distributed speech processing |
US20040133655A1 (en) | 1996-12-20 | 2004-07-08 | Liberate Technologies | Information retrieval system using an internet multiplexer to focus user selection |
US20040151358A1 (en) | 2003-01-31 | 2004-08-05 | Akiko Yanagita | Medical image processing system and method for processing medical image |
US6775360B2 (en) | 2000-12-28 | 2004-08-10 | Intel Corporation | Method and system for providing textual content along with voice messages |
US20040176906A1 (en) | 2002-03-15 | 2004-09-09 | Tsutomu Matsubara | Vehicular navigation device |
US20040193420A1 (en) * | 2002-07-15 | 2004-09-30 | Kennewick Robert A. | Mobile systems and methods for responding to natural language speech utterance |
US20040199595A1 (en) | 2003-01-16 | 2004-10-07 | Scott Banister | Electronic message delivery using a virtual gateway approach |
US6816578B1 (en) | 2001-11-27 | 2004-11-09 | Nortel Networks Limited | Efficient instant messaging using a telephony interface |
US6816468B1 (en) | 1999-12-16 | 2004-11-09 | Nortel Networks Limited | Captioning for tele-conferences |
US6820055B2 (en) | 2001-04-26 | 2004-11-16 | Speche Communications | Systems and methods for automated audio transcription, translation, and transfer with text display software for manipulating the text |
US20050004799A1 (en) | 2002-12-31 | 2005-01-06 | Yevgenly Lyudovyk | System and method for a spoken language interface to a large database of changing records |
US20050010641A1 (en) * | 2003-04-03 | 2005-01-13 | Jens Staack | Instant messaging context specific advertisements |
US20050021344A1 (en) | 2003-07-24 | 2005-01-27 | International Business Machines Corporation | Access to enhanced conferencing services using the tele-chat system |
US6850609B1 (en) | 1997-10-28 | 2005-02-01 | Verizon Services Corp. | Methods and apparatus for providing speech recording and speech transcription services |
US20050027538A1 (en) | 2003-04-07 | 2005-02-03 | Nokia Corporation | Method and device for providing speech-enabled input in an electronic device having a user interface |
US6856960B1 (en) | 1997-04-14 | 2005-02-15 | At & T Corp. | System and method for providing remote automatic speech recognition and text-to-speech services via a packet network |
US6859996B1 (en) | 1999-08-20 | 2005-03-01 | Seagate Technology Llc | Computer directed head stack assembly installation system |
US6865258B1 (en) | 1999-08-13 | 2005-03-08 | Intervoice Limited Partnership | Method and system for enhanced transcription |
US20050080786A1 (en) | 2003-10-14 | 2005-04-14 | Fish Edmund J. | System and method for customizing search results based on searcher's actual geographic location |
US20050101355A1 (en) | 2003-11-11 | 2005-05-12 | Microsoft Corporation | Sequential multimodal input |
US20050102142A1 (en) | 2001-02-13 | 2005-05-12 | Frederic Soufflet | Method, module, device and server for voice recognition |
US6895084B1 (en) | 1999-08-24 | 2005-05-17 | Microstrategy, Inc. | System and method for generating voice pages with included audio files for use in a voice page delivery system |
US20050149326A1 (en) | 2004-01-05 | 2005-07-07 | Kabushiki Kaisha Toshiba | Speech recognition system and technique |
US20050154587A1 (en) | 2003-09-11 | 2005-07-14 | Voice Signal Technologies, Inc. | Voice enabled phone book interface for speaker dependent name recognition and phone number categorization |
US20050165609A1 (en) | 1998-11-12 | 2005-07-28 | Microsoft Corporation | Speech recognition user interface |
US20050177376A1 (en) | 2004-02-05 | 2005-08-11 | Avaya Technology Corp. | Recognition results postprocessor for use in voice recognition systems |
US20050182628A1 (en) | 2004-02-18 | 2005-08-18 | Samsung Electronics Co., Ltd. | Domain-based dialog speech recognition method and apparatus |
US20050188029A1 (en) | 2003-12-18 | 2005-08-25 | Pauli Asikainen | Forming a message from information shown on display |
US20050187768A1 (en) | 2004-02-24 | 2005-08-25 | Godden Kurt S. | Dynamic N-best algorithm to reduce recognition errors |
US20050197840A1 (en) | 2004-03-05 | 2005-09-08 | Sunplus Technology Co., Ltd. | Device for event prediction on booting a motherboard |
US20050197145A1 (en) | 2004-03-03 | 2005-09-08 | Samsung Electro-Mechanics Co., Ltd. | Mobile phone capable of input of phone number without manipulating buttons and method of inputting phone number to the same |
US20050203751A1 (en) | 2000-05-02 | 2005-09-15 | Scansoft, Inc., A Delaware Corporation | Error correction in speech recognition |
US20050209868A1 (en) | 2004-03-19 | 2005-09-22 | Dadong Wan | Real-time sales support and learning tool |
US20050239495A1 (en) | 2004-04-12 | 2005-10-27 | Bayne Anthony J | System and method for the distribution of advertising and associated coupons via mobile media platforms |
US20050240406A1 (en) | 2004-04-21 | 2005-10-27 | David Carroll | Speech recognition computing device display with highlighted text |
US6961700B2 (en) | 1996-09-24 | 2005-11-01 | Allvoice Computing Plc | Method and apparatus for processing the output of a speech recognition engine |
US20050261907A1 (en) | 1999-04-12 | 2005-11-24 | Ben Franklin Patent Holding Llc | Voice integration platform |
US20050266884A1 (en) | 2003-04-22 | 2005-12-01 | Voice Genesis, Inc. | Methods and systems for conducting remote communications |
US6980954B1 (en) | 2000-09-30 | 2005-12-27 | Intel Corporation | Search method based on single triphone tree for large vocabulary continuous speech recognizer |
US20050288926A1 (en) | 2004-06-25 | 2005-12-29 | Benco David S | Network support for wireless e-mail using speech-to-text conversion |
US20060004570A1 (en) | 2004-06-30 | 2006-01-05 | Microsoft Corporation | Transcribing speech data with dialog context and/or recognition alternative information |
US20060009974A1 (en) | 2004-07-09 | 2006-01-12 | Matsushita Electric Industrial Co., Ltd. | Hands-free voice dialing for portable and remote devices |
US7007074B2 (en) | 2001-09-10 | 2006-02-28 | Yahoo! Inc. | Targeted advertisements using time-dependent key search terms |
US20060052127A1 (en) | 2004-09-07 | 2006-03-09 | Sbc Knowledge Ventures, L.P. | System and method for voice and text based service interworking |
US20060053016A1 (en) | 2002-02-04 | 2006-03-09 | Microsoft Corporation | Systems and methods for managing multiple grammars in a speech recognition system |
US20060074895A1 (en) | 2004-09-29 | 2006-04-06 | International Business Machines Corporation | Method and system for extracting and utilizing metadata to improve accuracy in speech to text conversions |
US20060075055A1 (en) | 2004-10-06 | 2006-04-06 | Andrew Littlefield | System and method for integration of instant messaging and virtual environment clients |
US7035901B1 (en) | 1999-12-06 | 2006-04-25 | Global Media Online, Inc. | SMTP server, POP server, mail server, mail processing system and web server |
US7039599B2 (en) | 1997-06-16 | 2006-05-02 | Doubleclick Inc. | Method and apparatus for automatic placement of advertising |
US20060111907A1 (en) | 2004-11-24 | 2006-05-25 | Microsoft Corporation | Generic spelling mnemonics |
US20060122834A1 (en) | 2004-12-03 | 2006-06-08 | Bennett Ian M | Emotion detection device & method for use in distributed systems |
US7062435B2 (en) | 1996-02-09 | 2006-06-13 | Canon Kabushiki Kaisha | Apparatus, method and computer readable memory medium for speech recognition using dynamic programming |
US20060129455A1 (en) | 2004-12-15 | 2006-06-15 | Kashan Shah | Method of advertising to users of text messaging |
US20060143007A1 (en) | 2000-07-24 | 2006-06-29 | Koh V E | User interaction with voice information services |
US20060149558A1 (en) | 2001-07-17 | 2006-07-06 | Jonathan Kahn | Synchronized pattern recognition source data processed by manual or automatic means for creation of shared speaker-dependent speech user profile |
US20060149630A1 (en) | 2004-11-16 | 2006-07-06 | Elliott Joseph F | Opt-in delivery of advertisements on mobile devices |
US20060159507A1 (en) | 2004-08-13 | 2006-07-20 | Bjorn Jawerth | One-row keyboard |
US7089194B1 (en) | 1999-06-17 | 2006-08-08 | International Business Machines Corporation | Method and apparatus for providing reduced cost online service and adaptive targeting of advertisements |
US7089184B2 (en) | 2001-03-22 | 2006-08-08 | Nurv Center Technologies, Inc. | Speech recognition for recognizing speaker-independent, continuous speech |
US20060195318A1 (en) | 2003-03-31 | 2006-08-31 | Stanglmayr Klaus H | System for correction of speech recognition results with confidence level indication |
US20060217159A1 (en) | 2005-03-22 | 2006-09-28 | Sony Ericsson Mobile Communications Ab | Wireless communications device with voice-to-text conversion |
US20060235684A1 (en) | 2005-04-14 | 2006-10-19 | Sbc Knowledge Ventures, Lp | Wireless device to access network-based voice-activated services using distributed speech recognition |
US20060235695A1 (en) | 1995-04-10 | 2006-10-19 | Thrift Philip R | Voice activated Hypermedia systems using grammatical metadata |
US7133513B1 (en) | 2004-07-21 | 2006-11-07 | Sprint Spectrum L.P. | Method and system for transcribing voice content of an on-going teleconference into human-readable notation |
US7136875B2 (en) | 2002-09-24 | 2006-11-14 | Google, Inc. | Serving advertisements based on content |
US7146615B1 (en) | 1999-07-09 | 2006-12-05 | France Telecom | System for fast development of interactive applications |
US20070005795A1 (en) | 1999-10-22 | 2007-01-04 | Activesky, Inc. | Object oriented video system |
US20070005368A1 (en) | 2003-08-29 | 2007-01-04 | Chutorash Richard J | System and method of operating a speech recognition system in a vehicle |
US20070033005A1 (en) * | 2005-08-05 | 2007-02-08 | Voicebox Technologies, Inc. | Systems and methods for responding to natural language speech utterance |
US20070038451A1 (en) | 2003-07-08 | 2007-02-15 | Laurent Cogne | Voice recognition for large dynamic vocabularies |
US20070038923A1 (en) | 2005-08-10 | 2007-02-15 | International Business Machines Corporation | Visual marker for speech enabled links |
US20070038740A1 (en) | 2005-08-10 | 2007-02-15 | Nortel Networks Limited | Notification service |
US7181387B2 (en) | 2004-06-30 | 2007-02-20 | Microsoft Corporation | Homonym processing in the context of voice-activated command systems |
US20070043569A1 (en) | 2005-08-19 | 2007-02-22 | Intervoice Limited Partnership | System and method for inheritance of advertised functionality in a user interactive system |
US20070061148A1 (en) | 2005-09-13 | 2007-03-15 | Cross Charles W Jr | Displaying speech command input state information in a multimodal browser |
US20070061146A1 (en) | 2005-09-12 | 2007-03-15 | International Business Machines Corporation | Retrieval and Presentation of Network Service Results for Mobile Device Using a Multimodal Browser |
US20070061300A1 (en) | 2005-09-14 | 2007-03-15 | Jorey Ramer | Mobile advertisement syndication |
US7200555B1 (en) | 2000-07-05 | 2007-04-03 | International Business Machines Corporation | Speech recognition correction for devices having limited or no display |
US20070079383A1 (en) | 2004-08-31 | 2007-04-05 | Gopalakrishnan Kumar C | System and Method for Providing Digital Content on Mobile Devices |
US7206932B1 (en) | 2003-02-14 | 2007-04-17 | Crystalvoice Communications | Firewall-tolerant voice-over-internet-protocol (VoIP) emulating SSL or HTTP sessions embedding voice data in cookies |
US20070086773A1 (en) | 2005-10-14 | 2007-04-19 | Fredrik Ramsten | Method for creating and operating a user interface |
US20070106506A1 (en) | 2005-11-07 | 2007-05-10 | Ma Changxue C | Personal synergic filtering of multimodal inputs |
US20070106507A1 (en) | 2005-11-09 | 2007-05-10 | International Business Machines Corporation | Noise playback enhancement of prerecorded audio for speech recognition operations |
US20070118374A1 (en) | 2005-11-23 | 2007-05-24 | Wise Gerald B | Method for generating closed captions |
US20070115845A1 (en) | 2005-10-24 | 2007-05-24 | Christian Hochwarth | Network time out handling |
US20070118592A1 (en) | 2004-07-24 | 2007-05-24 | Pixcall Gmbh | Method for the transmission of additional information in a communication system, exchange device and user station |
US20070118426A1 (en) | 2002-05-23 | 2007-05-24 | Barnes Jr Melvin L | Portable Communications Device and Method |
US7225125B2 (en) | 1999-11-12 | 2007-05-29 | Phoenix Solutions, Inc. | Speech recognition system trained with regional speech characteristics |
US7225224B2 (en) | 2002-03-26 | 2007-05-29 | Fujifilm Corporation | Teleconferencing server and teleconferencing system |
US20070123222A1 (en) | 2005-11-29 | 2007-05-31 | International Business Machines Corporation | Method and system for invoking push-to-service offerings |
US20070133771A1 (en) | 2005-12-12 | 2007-06-14 | Stifelman Lisa J | Providing missed call and message information |
US20070133769A1 (en) | 2005-12-08 | 2007-06-14 | International Business Machines Corporation | Voice navigation of a visual view for a session in a composite services enablement environment |
US7233655B2 (en) | 2001-10-03 | 2007-06-19 | Accenture Global Services Gmbh | Multi-modal callback |
US7236580B1 (en) | 2002-02-20 | 2007-06-26 | Cisco Technology, Inc. | Method and system for conducting a conference call |
US20070150275A1 (en) | 1999-10-28 | 2007-06-28 | Canon Kabushiki Kaisha | Pattern matching method and apparatus |
US20070156400A1 (en) | 2006-01-03 | 2007-07-05 | Wheeler Mark R | System and method for wireless dictation and transcription |
US7254384B2 (en) | 2001-10-03 | 2007-08-07 | Accenture Global Services Gmbh | Multi-modal messaging |
US20070180718A1 (en) | 2006-01-06 | 2007-08-09 | Tcl Communication Technology Holdings, Ltd. | Method for entering commands and/or characters for a portable communication device equipped with a tilt sensor |
US7260534B2 (en) | 2002-07-16 | 2007-08-21 | International Business Machines Corporation | Graphical user interface for determining speech recognition accuracy |
US20070233488A1 (en) | 2006-03-29 | 2007-10-04 | Dictaphone Corporation | System and method for applying dynamic contextual grammars and language models to improve automatic speech recognition accuracy |
US20070233487A1 (en) | 2006-04-03 | 2007-10-04 | Cohen Michael H | Automatic language model update |
US20070239837A1 (en) | 2006-04-05 | 2007-10-11 | Yap, Inc. | Hosted voice recognition system for wireless devices |
US20070255794A1 (en) | 2006-07-12 | 2007-11-01 | Marengo Intellectual Property Ltd. | Multi-conversation instant messaging |
US7302280B2 (en) | 2000-07-17 | 2007-11-27 | Microsoft Corporation | Mobile phone operation based upon context sensing |
US7310601B2 (en) | 2004-06-08 | 2007-12-18 | Matsushita Electric Industrial Co., Ltd. | Speech recognition apparatus and speech recognition method |
US7313526B2 (en) | 2001-09-05 | 2007-12-25 | Voice Signal Technologies, Inc. | Speech recognition using selectable recognition modes |
US7319957B2 (en) | 2004-02-11 | 2008-01-15 | Tegic Communications, Inc. | Handwriting and voice input with automatic correction |
US20080016142A1 (en) | 1999-03-22 | 2008-01-17 | Eric Schneider | Real-time communication processing method, product, and apparatus |
US7324942B1 (en) | 2002-01-29 | 2008-01-29 | Microstrategy, Incorporated | System and method for interactive voice services using markup language with N-best filter element |
US7328155B2 (en) | 2002-09-25 | 2008-02-05 | Toyota Infotechnology Center Co., Ltd. | Method and system for speech recognition using grammar weighted based upon location information |
US7330815B1 (en) | 1999-10-04 | 2008-02-12 | Globalenglish Corporation | Method and system for network-based speech recognition |
US20080040683A1 (en) | 2006-08-11 | 2008-02-14 | David Walsh | Multi-pane graphical user interface with common scroll control |
US20080037720A1 (en) | 2006-07-27 | 2008-02-14 | Speechphone, Llc | Voice Activated Communication Using Automatically Updated Address Books |
US20080052075A1 (en) | 2006-08-25 | 2008-02-28 | Microsoft Corporation | Incrementally regulated discriminative margins in MCE training for speech recognition |
US20080052073A1 (en) | 2004-11-22 | 2008-02-28 | National Institute Of Advanced Industrial Science And Technology | Voice Recognition Device and Method, and Program |
US20080065481A1 (en) | 2006-09-13 | 2008-03-13 | Microsoft Corporation | User-associated, interactive advertising monetization |
US20080063154A1 (en) * | 2006-08-09 | 2008-03-13 | Yossi Tamari | System and method of customized event notification |
US20080065737A1 (en) | 2006-08-03 | 2008-03-13 | Yahoo! Inc. | Electronic document information extraction |
US20080063155A1 (en) | 2006-02-10 | 2008-03-13 | Spinvox Limited | Mass-Scale, User-Independent, Device-Independent Voice Messaging System |
US20080077406A1 (en) | 2004-12-22 | 2008-03-27 | Nuance Communications Inc. | Mobile Dictation Correction User Interface |
US20080092168A1 (en) * | 1999-03-29 | 2008-04-17 | Logan James D | Audio and video program recording, editing and playback systems using metadata |
US20080091426A1 (en) | 2006-10-12 | 2008-04-17 | Rod Rempel | Adaptive context for automatic speech recognition systems |
US7376556B2 (en) | 1999-11-12 | 2008-05-20 | Phoenix Solutions, Inc. | Method for processing speech signal features for streaming transport |
US20080120375A1 (en) * | 2006-11-16 | 2008-05-22 | Benjamin Levy | Activity partner matching system and method |
US7379870B1 (en) | 2005-02-03 | 2008-05-27 | Hrl Laboratories, Llc | Contextual filtering |
US20080133232A1 (en) | 2006-02-10 | 2008-06-05 | Spinvox Limited | Mass-Scale, User-Independent, Device-Independent Voice Messaging System |
US20080147404A1 (en) | 2000-05-15 | 2008-06-19 | Nusuara Technologies Sdn Bhd | System and methods for accent classification and adaptation |
US7392185B2 (en) | 1999-11-12 | 2008-06-24 | Phoenix Solutions, Inc. | Speech based learning/training system using semantic decoding |
US20080154600A1 (en) | 2006-12-21 | 2008-06-26 | Nokia Corporation | System, Method, Apparatus and Computer Program Product for Providing Dynamic Vocabulary Prediction for Speech Recognition |
US20080154870A1 (en) | 2006-12-26 | 2008-06-26 | Voice Signal Technologies, Inc. | Collection and use of side information in voice-mediated mobile search |
US20080155060A1 (en) | 2006-12-22 | 2008-06-26 | Yahoo! Inc. | Exported overlays |
US7401122B2 (en) | 1999-12-03 | 2008-07-15 | Trend Micro, Inc. | Techniques for providing add-on services for an email system |
US20080177551A1 (en) | 2004-09-10 | 2008-07-24 | Atx Group, Inc. | Systems and Methods for Off-Board Voice-Automated Vehicle Navigation |
US20080172781A1 (en) | 2006-12-15 | 2008-07-24 | Terrance Popowich | System and method for obtaining and using advertising information |
US20080195588A1 (en) | 2005-05-06 | 2008-08-14 | Nhn Corporation | Personalized Search Method and System for Enabling the Method |
US20080200153A1 (en) | 2006-09-28 | 2008-08-21 | Dudley Fitzpatrick | Apparatuses, methods and systems for code triggered information querying and serving on mobile devices based on profiles |
US20080198981A1 (en) | 2007-02-21 | 2008-08-21 | Jens Ulrik Skakkebaek | Voicemail filtering and transcription |
US20080198898A1 (en) | 2007-02-21 | 2008-08-21 | Taylor John P | Apparatus, system and method for high resolution identification with temperature dependent resistive device |
US20080198980A1 (en) | 2007-02-21 | 2008-08-21 | Jens Ulrik Skakkebaek | Voicemail filtering and transcription |
US20080201139A1 (en) | 2007-02-20 | 2008-08-21 | Microsoft Corporation | Generic framework for large-margin MCE training in speech recognition |
US20080208582A1 (en) | 2002-09-27 | 2008-08-28 | Callminer, Inc. | Methods for statistical analysis of speech |
US20080208590A1 (en) | 2007-02-27 | 2008-08-28 | Cross Charles W | Disambiguating A Speech Recognition Grammar In A Multimodal Application |
US20080221897A1 (en) | 2007-03-07 | 2008-09-11 | Cerra Joseph P | Mobile environment speech processing facility |
US20080243500A1 (en) | 2007-03-30 | 2008-10-02 | Maximilian Bisani | Automatic Editing Using Probabilistic Word Substitution Models |
US20080243504A1 (en) | 2007-03-30 | 2008-10-02 | Verizon Data Services, Inc. | System and method of speech recognition training based on confirmed speaker utterances |
US20080261564A1 (en) | 2000-08-29 | 2008-10-23 | Logan James D | Communication and control system using location aware devices for audio message storage and transmission operating under rule-based control |
US20080275864A1 (en) | 2007-05-02 | 2008-11-06 | Yahoo! Inc. | Enabling clustered search processing via text messaging |
US20080275873A1 (en) * | 2002-04-05 | 2008-11-06 | Jason Bosarge | Method of enhancing emails with targeted ads |
US20080301250A1 (en) | 2007-05-29 | 2008-12-04 | Michael Thomas Hardy | Thread-based message prioritization |
US20080313039A1 (en) | 2007-06-18 | 2008-12-18 | Utbk, Inc. | Systems and Methods to Facilitate the Specification of a Complex Geographic Area |
US20080317219A1 (en) | 2007-06-21 | 2008-12-25 | Siemens Communications, Inc. | Method and apparatus for context based voice dialing |
US20090006194A1 (en) | 2007-06-27 | 2009-01-01 | Microsoft Corporation | Location, destination and other contextual information-based mobile advertisements |
US7475404B2 (en) | 2000-05-18 | 2009-01-06 | Maquis Techtrix Llc | System and method for implementing click-through for browser executed software including ad proxy and proxy cookie caching |
US20090012793A1 (en) | 2007-07-03 | 2009-01-08 | Dao Quyen C | Text-to-speech assist for portable communication devices |
US20090037255A1 (en) | 2006-12-06 | 2009-02-05 | Leo Chiu | Behavior aggregation |
US20090043855A1 (en) | 2007-08-08 | 2009-02-12 | Blake Bookstaff | System for providing information to originator of misdirected email |
US7496625B1 (en) | 2002-11-04 | 2009-02-24 | Cisco Technology, Inc. | System and method for communicating messages between a text-based client and a voice-based client |
US20090055538A1 (en) * | 2007-08-21 | 2009-02-26 | Microsoft Corporation | Content commentary |
US20090055175A1 (en) | 2007-08-22 | 2009-02-26 | Terrell Ii James Richard | Continuous speech transcription performance indication |
US20090055179A1 (en) | 2007-08-24 | 2009-02-26 | Samsung Electronics Co., Ltd. | Method, medium and apparatus for providing mobile voice web service |
US20090063151A1 (en) | 2007-08-28 | 2009-03-05 | Nexidia Inc. | Keyword spotting using a phoneme-sequence index |
US20090063268A1 (en) | 2007-09-04 | 2009-03-05 | Burgess David A | Targeting Using Historical Data |
US20090076917A1 (en) | 2007-08-22 | 2009-03-19 | Victor Roditis Jablokov | Facilitating presentation of ads relating to words of a message |
US20090076821A1 (en) | 2005-08-19 | 2009-03-19 | Gracenote, Inc. | Method and apparatus to control operation of a playback device |
US20090077493A1 (en) | 2006-03-10 | 2009-03-19 | Continental Automotive Gmbh | Method for the Selection of Functions with the Aid of a User Interface, and User Interface |
US20090086958A1 (en) | 2007-10-02 | 2009-04-02 | Utbk, Inc. | Systems and Methods to Provide Alternative Connections for Real Time Communications |
US20090100050A1 (en) | 2006-07-31 | 2009-04-16 | Berna Erol | Client device for interacting with a mixed media reality recognition system |
US20090117922A1 (en) | 2007-11-01 | 2009-05-07 | David Rowland Bell | Alerts based on significance of free format text messages |
US20090125299A1 (en) | 2007-11-09 | 2009-05-14 | Jui-Chang Wang | Speech recognition system |
US20090124272A1 (en) | 2006-04-05 | 2009-05-14 | Marc White | Filtering transcriptions of utterances |
US7539086B2 (en) | 2002-10-23 | 2009-05-26 | J2 Global Communications, Inc. | System and method for the secure, real-time, high accuracy conversion of general-quality speech into text |
US20090141875A1 (en) | 2007-01-10 | 2009-06-04 | Michael Demmitt | System and Method for Delivery of Voicemails to Handheld Devices |
US20090150405A1 (en) | 2007-07-13 | 2009-06-11 | Grouf Nicholas A | Systems and Methods for Expressing Data Using a Media Markup Language |
US20090150156A1 (en) | 2007-12-11 | 2009-06-11 | Kennewick Michael R | System and method for providing a natural language voice user interface in an integrated voice navigation services environment |
US20090163187A1 (en) | 2007-12-25 | 2009-06-25 | Yap, Inc. | Validation of mobile advertising from derived information |
US20090170478A1 (en) | 2003-04-22 | 2009-07-02 | Spinvox Limited | Method of providing voicemails to a wireless information device |
US20090182560A1 (en) | 2008-01-16 | 2009-07-16 | Yap, Inc. | Using a physical phenomenon detector to control operation of a speech recognition engine |
US20090182559A1 (en) | 2007-10-08 | 2009-07-16 | Franz Gerl | Context sensitive multi-stage speech recognition |
US20090199101A1 (en) | 2004-09-20 | 2009-08-06 | International Business Machines Corporation | Systems and methods for inputting graphical data into a graphical input field |
US20090204410A1 (en) | 2008-02-13 | 2009-08-13 | Sensory, Incorporated | Voice interface and search for electronic devices including bluetooth headsets and remote systems |
US7577569B2 (en) | 2001-09-05 | 2009-08-18 | Voice Signal Technologies, Inc. | Combined speech recognition and text-to-speech generation |
US20090210214A1 (en) | 2008-02-19 | 2009-08-20 | Jiang Qian | Universal Language Input |
US20090228274A1 (en) | 2008-03-07 | 2009-09-10 | Yap Inc. | Use of intermediate speech transcription results in editing final speech transcription results |
US20090240488A1 (en) | 2008-03-19 | 2009-09-24 | Yap, Inc. | Corrective feedback loop for automated speech recognition |
US20090248415A1 (en) | 2008-03-31 | 2009-10-01 | Yap, Inc. | Use of metadata to post process speech recognition output |
US20090276215A1 (en) | 2006-04-17 | 2009-11-05 | Hager Paul M | Methods and systems for correcting transcribed audio files |
US20090282363A1 (en) | 2006-09-15 | 2009-11-12 | Microsoft Corporation | Efficient navigation of search results |
US20090307090A1 (en) | 2008-06-05 | 2009-12-10 | Embarq Holdings Company, Llc | System and Method for Inserting Advertisements in Voicemail |
US7634403B2 (en) | 2001-09-05 | 2009-12-15 | Voice Signal Technologies, Inc. | Word recognition using word transformation commands |
US20090312040A1 (en) | 2008-06-13 | 2009-12-17 | Embarq Holdings Company, Llc | System and method for inserting advertisements into SMS messages |
US20090319187A1 (en) | 2008-06-23 | 2009-12-24 | Outside.In, Inc. | Generating Geocoded Targeted Web Advertisements |
US7640158B2 (en) | 2005-11-08 | 2009-12-29 | Multimodal Technologies, Inc. | Automatic detection and application of editing patterns in draft documents |
US7650284B2 (en) | 2004-11-19 | 2010-01-19 | Nuance Communications, Inc. | Enabling voice click in a multimodal page |
US20100017294A1 (en) | 2008-01-24 | 2010-01-21 | Mailmethods, Llc | Email advertisement system and method |
US20100049525A1 (en) | 2008-08-22 | 2010-02-25 | Yap, Inc. | Methods, apparatuses, and systems for providing timely user cues pertaining to speech recognition |
US20100058200A1 (en) | 2007-08-22 | 2010-03-04 | Yap, Inc. | Facilitating presentation by mobile device of additional content for a word or phrase upon utterance thereof |
US7680661B2 (en) | 2008-05-14 | 2010-03-16 | Nuance Communications, Inc. | Method and system for improved speech recognition |
US7685509B1 (en) | 1998-07-30 | 2010-03-23 | International Business Machines Corporation | Dividing a form field into areas associated with separate entry filters |
US7707163B2 (en) | 2005-05-25 | 2010-04-27 | Experian Marketing Solutions, Inc. | Software and metadata structures for distributed and interactive database architecture for parallel and asynchronous data processing of complex data and for real-time query processing |
US7716058B2 (en) | 2001-09-05 | 2010-05-11 | Voice Signal Technologies, Inc. | Speech recognition using automatic recognition turn off |
US20100121629A1 (en) | 2006-11-28 | 2010-05-13 | Cohen Sanford H | Method and apparatus for translating speech during a call |
US7729912B1 (en) | 2003-12-23 | 2010-06-01 | At&T Intellectual Property Ii, L.P. | System and method for latency reduction for automatic speech recognition using partial multi-pass results |
US20100146077A1 (en) | 2007-07-30 | 2010-06-10 | Nds Limited | Providing informatin about video content |
US7747437B2 (en) | 2004-12-16 | 2010-06-29 | Nuance Communications, Inc. | N-best list rescoring in speech recognition |
US7757162B2 (en) | 2003-03-31 | 2010-07-13 | Ricoh Co. Ltd. | Document collection manipulation |
US20100180202A1 (en) | 2005-07-05 | 2010-07-15 | Vida Software S.L. | User Interfaces for Electronic Devices |
US20100182325A1 (en) | 2002-01-22 | 2010-07-22 | Gizmoz Israel 2002 Ltd. | Apparatus and method for efficient animation of believable speaking 3d characters in real time |
US20100191619A1 (en) | 2002-10-07 | 2010-07-29 | Dicker Russell A | User interface and methods for recommending items to users |
US20100223056A1 (en) | 2009-02-27 | 2010-09-02 | Autonomy Corporation Ltd. | Various apparatus and methods for a speech recognition system |
US7796980B1 (en) | 2006-08-11 | 2010-09-14 | Sprint Communications Company L.P. | Remote mobile voice control of digital/personal video recorder |
US7809574B2 (en) | 2001-09-05 | 2010-10-05 | Voice Signal Technologies Inc. | Word recognition using choice lists |
US20100268726A1 (en) | 2005-11-30 | 2010-10-21 | Anchorfree, Inc. | Computerized system and method for advanced advertising |
US7822610B2 (en) | 2005-08-09 | 2010-10-26 | Mobile Voice Control, LLC | Use of multiple speech recognition software instances |
US20100278453A1 (en) | 2006-09-15 | 2010-11-04 | King Martin T | Capture and display of annotations in paper and electronic documents |
US20100279667A1 (en) | 2007-05-22 | 2010-11-04 | Wehrs Michael E | Keyword-based services for mobile device messages |
US20100286901A1 (en) | 2007-01-10 | 2010-11-11 | Pieter Geelen | Navigation device and method relating to an audible recognition mode |
US20100293242A1 (en) | 2004-03-31 | 2010-11-18 | Buchheit Paul T | Conversation-Based E-Mail Messaging |
US20100312619A1 (en) | 2007-05-23 | 2010-12-09 | Pekka Ala-Pietila | Method and a system for providing mobile communications services |
US20100312640A1 (en) | 2005-12-16 | 2010-12-09 | Apptera, Inc. | Call-Based Advertising |
US7852993B2 (en) | 2003-08-11 | 2010-12-14 | Microsoft Corporation | Speech recognition enhanced caller identification |
US20110029876A1 (en) | 2001-02-26 | 2011-02-03 | Benjamin Slotznick | Clickless navigation toolbar for clickless text-to-speech enabled browser |
US7890329B2 (en) | 2007-03-03 | 2011-02-15 | Industrial Technology Research Institute | Apparatus and method to reduce recognition errors through context relations among dialogue turns |
US7890586B1 (en) | 2004-11-01 | 2011-02-15 | At&T Mobility Ii Llc | Mass multimedia messaging |
US20110047452A1 (en) | 2006-12-06 | 2011-02-24 | Nuance Communications, Inc. | Enabling grammars in web page frame |
US7899670B1 (en) | 2006-12-21 | 2011-03-01 | Escription Inc. | Server-based speech recognition |
US20110054900A1 (en) | 2007-03-07 | 2011-03-03 | Phillips Michael S | Hybrid command and control between resident and remote speech recognition facilities in a mobile voice-to-speech application |
US7904301B2 (en) | 2002-09-06 | 2011-03-08 | Sony Europe Limited | Processing digital data |
US7908273B2 (en) | 2006-03-09 | 2011-03-15 | Gracenote, Inc. | Method and system for media navigation |
US7907705B1 (en) | 2006-10-10 | 2011-03-15 | Intuit Inc. | Speech to text for assisted form completion |
US20110064207A1 (en) | 2003-11-17 | 2011-03-17 | Apptera, Inc. | System for Advertisement Selection, Placement and Delivery |
US7925716B2 (en) | 2005-12-05 | 2011-04-12 | Yahoo! Inc. | Facilitating retrieval of information within a messaging environment |
US7949529B2 (en) | 2005-08-29 | 2011-05-24 | Voicebox Technologies, Inc. | Mobile systems and methods of supporting natural language human-machine interactions |
US20110144973A1 (en) | 2009-12-15 | 2011-06-16 | At&T Intellectual Property I, L.P. | System and method for combining geographic metadata in automatic speech recognition language and acoustic models |
US7970610B2 (en) | 2001-04-19 | 2011-06-28 | British Telecommunication Public Limited Company | Speech recognition |
US20110161072A1 (en) | 2008-08-20 | 2011-06-30 | Nec Corporation | Language model creation apparatus, language model creation method, speech recognition apparatus, speech recognition method, and recording medium |
US20110161276A1 (en) | 2005-06-30 | 2011-06-30 | Microsoft Corporation | Integration of location logs, gps signals, and spatial resources for identifying user activities, goals, and context |
US8010358B2 (en) | 2006-02-21 | 2011-08-30 | Sony Computer Entertainment Inc. | Voice recognition with parallel gender and age normalization |
US8027836B2 (en) | 2006-11-30 | 2011-09-27 | Nuance Communications, Inc. | Phonetic decoding and concatentive speech synthesis |
US8032372B1 (en) | 2005-09-13 | 2011-10-04 | Escription, Inc. | Dictation selection |
US8050918B2 (en) | 2003-12-11 | 2011-11-01 | Nuance Communications, Inc. | Quality evaluation tool for dynamic voice portals |
US8069047B2 (en) | 2007-02-12 | 2011-11-29 | Nuance Communications, Inc. | Dynamically defining a VoiceXML grammar in an X+V page of a multimodal application |
US20110296374A1 (en) | 2008-11-05 | 2011-12-01 | Google Inc. | Custom language models |
US20120022875A1 (en) | 2005-06-16 | 2012-01-26 | Nuance Communications, Inc. | Synchronizing visual and speech events in a multimodal application |
US8106285B2 (en) | 2006-02-10 | 2012-01-31 | Harman Becker Automotive Systems Gmbh | Speech-driven selection of an audio file |
US8121838B2 (en) | 2006-04-11 | 2012-02-21 | Nuance Communications, Inc. | Method and system for automatic transcription prioritization |
US20120059653A1 (en) | 2010-09-03 | 2012-03-08 | Adams Jeffrey P | Methods and systems for obtaining language models for transcribing communications |
US8135578B2 (en) | 2007-08-24 | 2012-03-13 | Nuance Communications, Inc. | Creation and use of application-generic class-based statistical language models for automatic speech recognition |
US8145493B2 (en) | 2006-09-11 | 2012-03-27 | Nuance Communications, Inc. | Establishing a preferred mode of interaction between a user and a multimodal application |
US8145485B2 (en) | 2007-12-17 | 2012-03-27 | Verizon Patent And Licensing Inc. | Grammar weighting voice recognition information |
US20120095831A1 (en) | 2007-03-09 | 2012-04-19 | Janne Aaltonen | Method and apparatus for controlling user communications |
US8209184B1 (en) | 1997-04-14 | 2012-06-26 | At&T Intellectual Property Ii, L.P. | System and method of providing generated speech via a network |
US20120166202A1 (en) * | 2000-03-21 | 2012-06-28 | Steven Jeromy Carriere | System and method for funneling user responses in an internet voice portal system to determine a desired item or servicebackground of the invention |
US8229743B2 (en) | 2009-06-23 | 2012-07-24 | Autonomy Corporation Ltd. | Speech recognition system |
US20120259729A1 (en) | 1998-09-18 | 2012-10-11 | Linden Gregory D | Discovery of behavior-based item relationships |
US8296139B2 (en) | 2006-12-22 | 2012-10-23 | International Business Machines Corporation | Adding real-time dictation capabilities for speech processing operations handled by a networked speech processing system |
US8311825B2 (en) | 2007-10-04 | 2012-11-13 | Kabushiki Kaisha Toshiba | Automatic speech recognition method and apparatus |
US20120324391A1 (en) | 2011-06-16 | 2012-12-20 | Microsoft Corporation | Predictive word completion |
US8355920B2 (en) | 2002-07-31 | 2013-01-15 | Nuance Communications, Inc. | Natural error handling in speech recognition |
US20130041667A1 (en) | 2004-06-02 | 2013-02-14 | Nuance Communications, Inc. | Multimodal disambiguation of speech recognition |
US8380511B2 (en) | 2007-02-20 | 2013-02-19 | Intervoice Limited Partnership | System and method for semantic categorization |
US8417530B1 (en) | 2010-08-20 | 2013-04-09 | Google Inc. | Accent-influenced search results |
US8510094B2 (en) | 2007-02-14 | 2013-08-13 | Google Inc. | Machine translation feedback |
US20130211815A1 (en) | 2003-09-05 | 2013-08-15 | Mark Seligman | Method and Apparatus for Cross-Lingual Communication |
US20130226894A1 (en) | 2006-03-30 | 2013-08-29 | Veveo, Inc. | Method and System for Incrementally Selecting and Providing Relevant Search Engines in Response to a User Query |
US20130281007A1 (en) | 2007-10-05 | 2013-10-24 | Qualcomm Incorporated | Location and time based filtering of broadcast information |
US8589164B1 (en) | 2012-10-18 | 2013-11-19 | Google Inc. | Methods and systems for speech recognition processing using search query information |
US8670977B2 (en) | 2004-08-23 | 2014-03-11 | At&T Intellectual Property Ii, L.P. | System and method of lattice-based search for spoken utterance retrieval |
US20140136199A1 (en) | 2006-04-17 | 2014-05-15 | Vovision, Llc | Correcting transcribed audio files with an email-client interface |
US8898065B2 (en) | 2011-01-07 | 2014-11-25 | Nuance Communications, Inc. | Configurable speech recognition system using multiple recognizers |
US9053489B2 (en) | 2007-08-22 | 2015-06-09 | Canyon Ip Holdings Llc | Facilitating presentation of ads relating to words of a message |
US9093061B1 (en) | 2011-04-14 | 2015-07-28 | Canyon IP Holdings, LLC. | Speech recognition with hierarchical networks |
US20150255067A1 (en) | 2006-04-05 | 2015-09-10 | Canyon IP Holding LLC | Filtering transcriptions of utterances using received information to correct transcription errors |
US9369581B2 (en) | 2001-06-12 | 2016-06-14 | At&T Intellectual Property Ii, L.P. | System and method for processing speech files |
US9436951B1 (en) | 2007-08-22 | 2016-09-06 | Amazon Technologies, Inc. | Facilitating presentation by mobile device of additional content for a word or phrase upon utterance thereof |
-
2008
- 2008-09-17 US US12/212,644 patent/US9973450B2/en active Active
Patent Citations (403)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US5036538A (en) | 1989-11-22 | 1991-07-30 | Telephonics Corporation | Multi-station voice recognition and processing system |
US5623609A (en) | 1993-06-14 | 1997-04-22 | Hal Trust, L.L.C. | Computer system and computer-implemented process for phonology-based automatic speech recognition |
US6100882A (en) | 1994-01-19 | 2000-08-08 | International Business Machines Corporation | Textual recording of contributions to audio conference using speech recognition |
US20060235695A1 (en) | 1995-04-10 | 2006-10-19 | Thrift Philip R | Voice activated Hypermedia systems using grammatical metadata |
US5675507A (en) | 1995-04-28 | 1997-10-07 | Bobo, Ii; Charles R. | Message storage and delivery system |
US5864603A (en) | 1995-06-02 | 1999-01-26 | Nokia Mobile Phones Limited | Method and apparatus for controlling a telephone with voice commands |
US6026368A (en) | 1995-07-17 | 2000-02-15 | 24/7 Media, Inc. | On-line interactive system and method for providing content and advertising information to a targeted set of viewers |
US5852801A (en) | 1995-10-04 | 1998-12-22 | Apple Computer, Inc. | Method and apparatus for automatically invoking a new word module for unrecognized user input |
US7062435B2 (en) | 1996-02-09 | 2006-06-13 | Canon Kabushiki Kaisha | Apparatus, method and computer readable memory medium for speech recognition using dynamic programming |
US5822730A (en) | 1996-08-22 | 1998-10-13 | Dragon Systems, Inc. | Lexical tree pre-filtering in speech recognition |
US6961700B2 (en) | 1996-09-24 | 2005-11-01 | Allvoice Computing Plc | Method and apparatus for processing the output of a speech recognition engine |
US5995928A (en) | 1996-10-02 | 1999-11-30 | Speechworks International, Inc. | Method and apparatus for continuous spelling speech recognition with early identification |
US5948061A (en) | 1996-10-29 | 1999-09-07 | Double Click, Inc. | Method of delivery, targeting, and measuring advertising over networks |
US20040133655A1 (en) | 1996-12-20 | 2004-07-08 | Liberate Technologies | Information retrieval system using an internet multiplexer to focus user selection |
US6173259B1 (en) | 1997-03-27 | 2001-01-09 | Speech Machines Plc | Speech to text conversion |
US6212498B1 (en) | 1997-03-28 | 2001-04-03 | Dragon Systems, Inc. | Enrollment in speech recognition |
US8209184B1 (en) | 1997-04-14 | 2012-06-26 | At&T Intellectual Property Ii, L.P. | System and method of providing generated speech via a network |
US6604077B2 (en) | 1997-04-14 | 2003-08-05 | At&T Corp. | System and method for providing remote automatic speech recognition and text to speech services via a packet network |
US6366886B1 (en) | 1997-04-14 | 2002-04-02 | At&T Corp. | System and method for providing remote automatic speech recognition services via a packet network |
US6856960B1 (en) | 1997-04-14 | 2005-02-15 | At & T Corp. | System and method for providing remote automatic speech recognition and text-to-speech services via a packet network |
US7039599B2 (en) | 1997-06-16 | 2006-05-02 | Doubleclick Inc. | Method and apparatus for automatic placement of advertising |
US6490561B1 (en) | 1997-06-25 | 2002-12-03 | Dennis L. Wilson | Continuous speech voice transcription |
US5974413A (en) | 1997-07-03 | 1999-10-26 | Activeword Systems, Inc. | Semantic user interface |
US6850609B1 (en) | 1997-10-28 | 2005-02-01 | Verizon Services Corp. | Methods and apparatus for providing speech recording and speech transcription services |
US6687339B2 (en) | 1997-12-31 | 2004-02-03 | Weblink Wireless, Inc. | Controller for use with communications systems for converting a voice message to a text message |
US6219407B1 (en) | 1998-01-16 | 2001-04-17 | International Business Machines Corporation | Apparatus and method for improved digit recognition and caller identification in telephone mail messaging |
US6654448B1 (en) | 1998-06-19 | 2003-11-25 | At&T Corp. | Voice messaging system |
US7685509B1 (en) | 1998-07-30 | 2010-03-23 | International Business Machines Corporation | Dividing a form field into areas associated with separate entry filters |
US20120259729A1 (en) | 1998-09-18 | 2012-10-11 | Linden Gregory D | Discovery of behavior-based item relationships |
US6219638B1 (en) | 1998-11-03 | 2001-04-17 | International Business Machines Corporation | Telephone messaging and editing system |
US20050165609A1 (en) | 1998-11-12 | 2005-07-28 | Microsoft Corporation | Speech recognition user interface |
US6571210B2 (en) | 1998-11-13 | 2003-05-27 | Microsoft Corporation | Confidence measure system using a near-miss pattern |
US6519562B1 (en) | 1999-02-25 | 2003-02-11 | Speechworks International, Inc. | Dynamic semantic control of a speech recognition system |
US6253177B1 (en) | 1999-03-08 | 2001-06-26 | International Business Machines Corp. | Method and system for automatically determining whether to update a language model based upon user amendments to dictated text |
US20080016142A1 (en) | 1999-03-22 | 2008-01-17 | Eric Schneider | Real-time communication processing method, product, and apparatus |
US20080092168A1 (en) * | 1999-03-29 | 2008-04-17 | Logan James D | Audio and video program recording, editing and playback systems using metadata |
US20050261907A1 (en) | 1999-04-12 | 2005-11-24 | Ben Franklin Patent Holding Llc | Voice integration platform |
US6298326B1 (en) | 1999-05-13 | 2001-10-02 | Alan Feller | Off-site data entry system |
US20030200093A1 (en) | 1999-06-11 | 2003-10-23 | International Business Machines Corporation | Method and system for proofreading and correcting dictated text |
US6760700B2 (en) | 1999-06-11 | 2004-07-06 | International Business Machines Corporation | Method and system for proofreading and correcting dictated text |
US7089194B1 (en) | 1999-06-17 | 2006-08-08 | International Business Machines Corporation | Method and apparatus for providing reduced cost online service and adaptive targeting of advertisements |
US7146615B1 (en) | 1999-07-09 | 2006-12-05 | France Telecom | System for fast development of interactive applications |
US6865258B1 (en) | 1999-08-13 | 2005-03-08 | Intervoice Limited Partnership | Method and system for enhanced transcription |
US6859996B1 (en) | 1999-08-20 | 2005-03-01 | Seagate Technology Llc | Computer directed head stack assembly installation system |
US6895084B1 (en) | 1999-08-24 | 2005-05-17 | Microstrategy, Inc. | System and method for generating voice pages with included audio files for use in a voice page delivery system |
US20020052781A1 (en) | 1999-09-10 | 2002-05-02 | Avantgo, Inc. | Interactive advertisement mechanism on a mobile device |
US7689415B1 (en) | 1999-10-04 | 2010-03-30 | Globalenglish Corporation | Real-time speech recognition over the internet |
US7330815B1 (en) | 1999-10-04 | 2008-02-12 | Globalenglish Corporation | Method and system for network-based speech recognition |
US8401850B1 (en) | 1999-10-04 | 2013-03-19 | Globalenglish Corporation | Processing packets of encoded speech using a plurality of processing levels based on values transmitted over a network |
US6453290B1 (en) | 1999-10-04 | 2002-09-17 | Globalenglish Corporation | Method and system for network-based speech recognition |
US20070005795A1 (en) | 1999-10-22 | 2007-01-04 | Activesky, Inc. | Object oriented video system |
US20070150275A1 (en) | 1999-10-28 | 2007-06-28 | Canon Kabushiki Kaisha | Pattern matching method and apparatus |
US7672841B2 (en) | 1999-11-12 | 2010-03-02 | Phoenix Solutions, Inc. | Method for processing speech data for a distributed recognition system |
US7555431B2 (en) | 1999-11-12 | 2009-06-30 | Phoenix Solutions, Inc. | Method for processing speech using dynamic grammars |
US7729904B2 (en) | 1999-11-12 | 2010-06-01 | Phoenix Solutions, Inc. | Partial speech processing device and method for use in distributed systems |
US7725321B2 (en) | 1999-11-12 | 2010-05-25 | Phoenix Solutions, Inc. | Speech based query system using semantic decoding |
US7725307B2 (en) | 1999-11-12 | 2010-05-25 | Phoenix Solutions, Inc. | Query engine for processing voice based queries including semantic decoding |
US7376556B2 (en) | 1999-11-12 | 2008-05-20 | Phoenix Solutions, Inc. | Method for processing speech signal features for streaming transport |
US7392185B2 (en) | 1999-11-12 | 2008-06-24 | Phoenix Solutions, Inc. | Speech based learning/training system using semantic decoding |
US7702508B2 (en) | 1999-11-12 | 2010-04-20 | Phoenix Solutions, Inc. | System and method for natural language processing of query answers |
US20090157401A1 (en) | 1999-11-12 | 2009-06-18 | Bennett Ian M | Semantic Decoding of User Queries |
US7657424B2 (en) | 1999-11-12 | 2010-02-02 | Phoenix Solutions, Inc. | System and method for processing sentence based queries |
US7225125B2 (en) | 1999-11-12 | 2007-05-29 | Phoenix Solutions, Inc. | Speech recognition system trained with regional speech characteristics |
US20030182113A1 (en) | 1999-11-22 | 2003-09-25 | Xuedong Huang | Distributed speech recognition for mobile communication devices |
US6532446B1 (en) | 1999-11-24 | 2003-03-11 | Openwave Systems Inc. | Server based speech recognition user interface for wireless devices |
US7401122B2 (en) | 1999-12-03 | 2008-07-15 | Trend Micro, Inc. | Techniques for providing add-on services for an email system |
US7035901B1 (en) | 1999-12-06 | 2006-04-25 | Global Media Online, Inc. | SMTP server, POP server, mail server, mail processing system and web server |
US6816468B1 (en) | 1999-12-16 | 2004-11-09 | Nortel Networks Limited | Captioning for tele-conferences |
US20010047294A1 (en) | 2000-01-06 | 2001-11-29 | Rothschild Anthony R. | System and method for adding an advertisement to a personal communication |
US6401075B1 (en) | 2000-02-14 | 2002-06-04 | Global Network, Inc. | Methods of placing, purchasing and monitoring internet advertising |
US20120166202A1 (en) * | 2000-03-21 | 2012-06-28 | Steven Jeromy Carriere | System and method for funneling user responses in an internet voice portal system to determine a desired item or servicebackground of the invention |
US20050203751A1 (en) | 2000-05-02 | 2005-09-15 | Scansoft, Inc., A Delaware Corporation | Error correction in speech recognition |
US20080147404A1 (en) | 2000-05-15 | 2008-06-19 | Nusuara Technologies Sdn Bhd | System and methods for accent classification and adaptation |
US7475404B2 (en) | 2000-05-18 | 2009-01-06 | Maquis Techtrix Llc | System and method for implementing click-through for browser executed software including ad proxy and proxy cookie caching |
US20020165773A1 (en) | 2000-05-31 | 2002-11-07 | Takeshi Natsuno | Method and system for distributing advertisements over network |
US20010056350A1 (en) | 2000-06-08 | 2001-12-27 | Theodore Calderone | System and method of voice recognition near a wireline node of a network supporting cable television and/or video delivery |
US6687689B1 (en) | 2000-06-16 | 2004-02-03 | Nusuara Technologies Sdn. Bhd. | System and methods for document retrieval using natural language-based queries |
US20010056369A1 (en) | 2000-06-16 | 2001-12-27 | Kuniharu Takayama | Advertisement posting system, advertisement-cost calculating method, and record medium storing advertisement-cost calculating program |
US7200555B1 (en) | 2000-07-05 | 2007-04-03 | International Business Machines Corporation | Speech recognition correction for devices having limited or no display |
US7302280B2 (en) | 2000-07-17 | 2007-11-27 | Microsoft Corporation | Mobile phone operation based upon context sensing |
US20020035474A1 (en) | 2000-07-18 | 2002-03-21 | Ahmet Alpdemir | Voice-interactive marketplace providing time and money saving benefits and real-time promotion publishing and feedback |
US20020016712A1 (en) | 2000-07-20 | 2002-02-07 | Geurts Lucas Jacobus Franciscus | Feedback of recognized command confidence level |
US20060143007A1 (en) | 2000-07-24 | 2006-06-29 | Koh V E | User interaction with voice information services |
US20040005877A1 (en) | 2000-08-21 | 2004-01-08 | Vaananen Mikko Kalervo | Voicemail short massage service method and means and a subscriber terminal |
US20080261564A1 (en) | 2000-08-29 | 2008-10-23 | Logan James D | Communication and control system using location aware devices for audio message storage and transmission operating under rule-based control |
US20020029101A1 (en) | 2000-09-05 | 2002-03-07 | Hunter Engineering Company | Method and apparatus for networked wheel alignment communications and services |
US20030093315A1 (en) | 2000-09-26 | 2003-05-15 | Kenji Sato | System and method for using e-mail as advertisement medium |
US6704034B1 (en) | 2000-09-28 | 2004-03-09 | International Business Machines Corporation | Method and apparatus for providing accessibility through a context sensitive magnifying glass |
US6980954B1 (en) | 2000-09-30 | 2005-12-27 | Intel Corporation | Search method based on single triphone tree for large vocabulary continuous speech recognizer |
US20020091570A1 (en) | 2000-12-01 | 2002-07-11 | Hiroaki Sakagawa | Electronic mail advertisement system, method, and program storage medium |
US6775360B2 (en) | 2000-12-28 | 2004-08-10 | Intel Corporation | Method and system for providing textual content along with voice messages |
US20020087330A1 (en) | 2001-01-03 | 2002-07-04 | Motorola, Inc. | Method of communicating a set of audio content |
US20050102142A1 (en) | 2001-02-13 | 2005-05-12 | Frederic Soufflet | Method, module, device and server for voice recognition |
US20110029876A1 (en) | 2001-02-26 | 2011-02-03 | Benjamin Slotznick | Clickless navigation toolbar for clickless text-to-speech enabled browser |
US7089184B2 (en) | 2001-03-22 | 2006-08-08 | Nurv Center Technologies, Inc. | Speech recognition for recognizing speaker-independent, continuous speech |
US7970610B2 (en) | 2001-04-19 | 2011-06-28 | British Telecommunication Public Limited Company | Speech recognition |
US7035804B2 (en) | 2001-04-26 | 2006-04-25 | Stenograph, L.L.C. | Systems and methods for automated audio transcription, translation, and transfer |
US20020161579A1 (en) | 2001-04-26 | 2002-10-31 | Speche Communications | Systems and methods for automated audio transcription, translation, and transfer |
US6820055B2 (en) | 2001-04-26 | 2004-11-16 | Speche Communications | Systems and methods for automated audio transcription, translation, and transfer with text display software for manipulating the text |
US20020165719A1 (en) | 2001-05-04 | 2002-11-07 | Kuansan Wang | Servers for web enabled speech recognition |
US9369581B2 (en) | 2001-06-12 | 2016-06-14 | At&T Intellectual Property Ii, L.P. | System and method for processing speech files |
EP1274222A2 (en) | 2001-07-02 | 2003-01-08 | Nortel Networks Limited | Instant messaging using a wireless interface |
US20030008661A1 (en) | 2001-07-03 | 2003-01-09 | Joyce Dennis P. | Location-based content delivery |
US7668718B2 (en) | 2001-07-17 | 2010-02-23 | Custom Speech Usa, Inc. | Synchronized pattern recognition source data processed by manual or automatic means for creation of shared speaker-dependent speech user profile |
US20060149558A1 (en) | 2001-07-17 | 2006-07-06 | Jonathan Kahn | Synchronized pattern recognition source data processed by manual or automatic means for creation of shared speaker-dependent speech user profile |
US20030028601A1 (en) | 2001-07-31 | 2003-02-06 | Rowe Lorin Bruce | Method and apparatus for providing interactive text messages during a voice call |
US7634403B2 (en) | 2001-09-05 | 2009-12-15 | Voice Signal Technologies, Inc. | Word recognition using word transformation commands |
US7716058B2 (en) | 2001-09-05 | 2010-05-11 | Voice Signal Technologies, Inc. | Speech recognition using automatic recognition turn off |
US7577569B2 (en) | 2001-09-05 | 2009-08-18 | Voice Signal Technologies, Inc. | Combined speech recognition and text-to-speech generation |
US7809574B2 (en) | 2001-09-05 | 2010-10-05 | Voice Signal Technologies Inc. | Word recognition using choice lists |
US7313526B2 (en) | 2001-09-05 | 2007-12-25 | Voice Signal Technologies, Inc. | Speech recognition using selectable recognition modes |
US20030126216A1 (en) | 2001-09-06 | 2003-07-03 | Avila J. Albert | Method and system for remote delivery of email |
US7007074B2 (en) | 2001-09-10 | 2006-02-28 | Yahoo! Inc. | Targeted advertisements using time-dependent key search terms |
US20030050778A1 (en) | 2001-09-13 | 2003-03-13 | Patrick Nguyen | Focused language models for improved speech input of structured documents |
US7233655B2 (en) | 2001-10-03 | 2007-06-19 | Accenture Global Services Gmbh | Multi-modal callback |
US7254384B2 (en) | 2001-10-03 | 2007-08-07 | Accenture Global Services Gmbh | Multi-modal messaging |
US20030101054A1 (en) | 2001-11-27 | 2003-05-29 | Ncc, Llc | Integrated system and method for electronic speech recognition and transcription |
US6816578B1 (en) | 2001-11-27 | 2004-11-09 | Nortel Networks Limited | Efficient instant messaging using a telephony interface |
US20090271194A1 (en) | 2001-11-27 | 2009-10-29 | Davis Michael K | Speech recognition and transcription among users having heterogeneous protocols |
US20030105630A1 (en) | 2001-11-30 | 2003-06-05 | Macginitie Andrew | Performance gauge for a distributed speech recognition system |
US20030139922A1 (en) | 2001-12-12 | 2003-07-24 | Gerhard Hoffmann | Speech recognition system and method for operating same |
US20030115060A1 (en) | 2001-12-13 | 2003-06-19 | Junqua Jean-Claude | System and interactive form filling with fusion of data from multiple unreliable information sources |
US20030125955A1 (en) | 2001-12-28 | 2003-07-03 | Arnold James F. | Method and apparatus for providing a dynamic speech-driven control and remote service access system |
US7013275B2 (en) | 2001-12-28 | 2006-03-14 | Sri International | Method and apparatus for providing a dynamic speech-driven control and remote service access system |
US20030149566A1 (en) | 2002-01-02 | 2003-08-07 | Esther Levin | System and method for a spoken language interface to a large database of changing records |
US20100182325A1 (en) | 2002-01-22 | 2010-07-22 | Gizmoz Israel 2002 Ltd. | Apparatus and method for efficient animation of believable speaking 3d characters in real time |
US7324942B1 (en) | 2002-01-29 | 2008-01-29 | Microstrategy, Incorporated | System and method for interactive voice services using markup language with N-best filter element |
US20030144906A1 (en) | 2002-01-31 | 2003-07-31 | Nissan Motor Co., Ltd. | Advertisement distribution method, advertisement distribution apparatus and advertisement displaying vehicle |
US20060053016A1 (en) | 2002-02-04 | 2006-03-09 | Microsoft Corporation | Systems and methods for managing multiple grammars in a speech recognition system |
US20060161429A1 (en) | 2002-02-04 | 2006-07-20 | Microsoft Corporation | Systems And Methods For Managing Multiple Grammars in a Speech Recognition System |
US7363229B2 (en) | 2002-02-04 | 2008-04-22 | Microsoft Corporation | Systems and methods for managing multiple grammars in a speech recognition system |
US7236580B1 (en) | 2002-02-20 | 2007-06-26 | Cisco Technology, Inc. | Method and system for conducting a conference call |
US20040176906A1 (en) | 2002-03-15 | 2004-09-09 | Tsutomu Matsubara | Vehicular navigation device |
US7225224B2 (en) | 2002-03-26 | 2007-05-29 | Fujifilm Corporation | Teleconferencing server and teleconferencing system |
US20030187643A1 (en) | 2002-03-27 | 2003-10-02 | Compaq Information Technologies Group, L.P. | Vocabulary independent speech decoder system and method using subword units |
US7181398B2 (en) | 2002-03-27 | 2007-02-20 | Hewlett-Packard Development Company, L.P. | Vocabulary independent speech recognition system and method using subword units |
US20030191639A1 (en) | 2002-04-05 | 2003-10-09 | Sam Mazza | Dynamic and adaptive selection of vocabulary and acoustic models based on a call context for speech recognition |
US20080275873A1 (en) * | 2002-04-05 | 2008-11-06 | Jason Bosarge | Method of enhancing emails with targeted ads |
US20030200086A1 (en) | 2002-04-17 | 2003-10-23 | Pioneer Corporation | Speech recognition apparatus, speech recognition method, and computer-readable recording medium in which speech recognition program is recorded |
US7590534B2 (en) | 2002-05-09 | 2009-09-15 | Healthsense, Inc. | Method and apparatus for processing voice data |
US20030212554A1 (en) | 2002-05-09 | 2003-11-13 | Vatland Danny James | Method and apparatus for processing voice data |
US20070118426A1 (en) | 2002-05-23 | 2007-05-24 | Barnes Jr Melvin L | Portable Communications Device and Method |
US7047200B2 (en) | 2002-05-24 | 2006-05-16 | Microsoft, Corporation | Voice recognition status display |
US20030220798A1 (en) | 2002-05-24 | 2003-11-27 | Microsoft Corporation | Speech recognition status feedback user interface |
US20030220792A1 (en) | 2002-05-27 | 2003-11-27 | Pioneer Corporation | Speech recognition apparatus, speech recognition method, and computer-readable recording medium in which speech recognition program is recorded |
US7146320B2 (en) | 2002-05-29 | 2006-12-05 | Microsoft Corporation | Electronic mail replies with speech recognition |
US7280966B2 (en) | 2002-05-29 | 2007-10-09 | Microsoft Corporation | Electronic mail replies with speech recognition |
US20030223556A1 (en) | 2002-05-29 | 2003-12-04 | Yun-Cheng Ju | Electronic mail replies with speech recognition |
US20060195541A1 (en) | 2002-05-29 | 2006-08-31 | Microsoft Corporation | Electronic mail replies with speech recognition |
US20040193420A1 (en) * | 2002-07-15 | 2004-09-30 | Kennewick Robert A. | Mobile systems and methods for responding to natural language speech utterance |
US20100145700A1 (en) | 2002-07-15 | 2010-06-10 | Voicebox Technologies, Inc. | Mobile systems and methods for responding to natural language speech utterance |
US7260534B2 (en) | 2002-07-16 | 2007-08-21 | International Business Machines Corporation | Graphical user interface for determining speech recognition accuracy |
US20040015547A1 (en) | 2002-07-17 | 2004-01-22 | Griffin Chris Michael | Voice and text group chat techniques for wireless mobile terminals |
US20040019488A1 (en) | 2002-07-23 | 2004-01-29 | Netbytel, Inc. | Email address recognition using personal information |
US8355920B2 (en) | 2002-07-31 | 2013-01-15 | Nuance Communications, Inc. | Natural error handling in speech recognition |
US7904301B2 (en) | 2002-09-06 | 2011-03-08 | Sony Europe Limited | Processing digital data |
US20040059632A1 (en) | 2002-09-23 | 2004-03-25 | International Business Machines Corporation | Method and system for providing an advertisement based on an URL and/or a search keyword entered by a user |
US20040059712A1 (en) | 2002-09-24 | 2004-03-25 | Dean Jeffrey A. | Serving advertisements using information associated with e-mail |
US20040059708A1 (en) | 2002-09-24 | 2004-03-25 | Google, Inc. | Methods and apparatus for serving relevant advertisements |
US7136875B2 (en) | 2002-09-24 | 2006-11-14 | Google, Inc. | Serving advertisements based on content |
US7328155B2 (en) | 2002-09-25 | 2008-02-05 | Toyota Infotechnology Center Co., Ltd. | Method and system for speech recognition using grammar weighted based upon location information |
US20080208582A1 (en) | 2002-09-27 | 2008-08-28 | Callminer, Inc. | Methods for statistical analysis of speech |
US20100191619A1 (en) | 2002-10-07 | 2010-07-29 | Dicker Russell A | User interface and methods for recommending items to users |
US7539086B2 (en) | 2002-10-23 | 2009-05-26 | J2 Global Communications, Inc. | System and method for the secure, real-time, high accuracy conversion of general-quality speech into text |
US7496625B1 (en) | 2002-11-04 | 2009-02-24 | Cisco Technology, Inc. | System and method for communicating messages between a text-based client and a voice-based client |
US7571100B2 (en) | 2002-12-03 | 2009-08-04 | Speechworks International, Inc. | Speech recognition and speaker verification using distributed speech processing |
US20040107107A1 (en) | 2002-12-03 | 2004-06-03 | Philip Lenir | Distributed speech processing |
US20050004799A1 (en) | 2002-12-31 | 2005-01-06 | Yevgenly Lyudovyk | System and method for a spoken language interface to a large database of changing records |
US20040199595A1 (en) | 2003-01-16 | 2004-10-07 | Scott Banister | Electronic message delivery using a virtual gateway approach |
US20040151358A1 (en) | 2003-01-31 | 2004-08-05 | Akiko Yanagita | Medical image processing system and method for processing medical image |
US7206932B1 (en) | 2003-02-14 | 2007-04-17 | Crystalvoice Communications | Firewall-tolerant voice-over-internet-protocol (VoIP) emulating SSL or HTTP sessions embedding voice data in cookies |
US7757162B2 (en) | 2003-03-31 | 2010-07-13 | Ricoh Co. Ltd. | Document collection manipulation |
US20060195318A1 (en) | 2003-03-31 | 2006-08-31 | Stanglmayr Klaus H | System for correction of speech recognition results with confidence level indication |
US20050010641A1 (en) * | 2003-04-03 | 2005-01-13 | Jens Staack | Instant messaging context specific advertisements |
US20050027538A1 (en) | 2003-04-07 | 2005-02-03 | Nokia Corporation | Method and device for providing speech-enabled input in an electronic device having a user interface |
US20090170478A1 (en) | 2003-04-22 | 2009-07-02 | Spinvox Limited | Method of providing voicemails to a wireless information device |
US20050266884A1 (en) | 2003-04-22 | 2005-12-01 | Voice Genesis, Inc. | Methods and systems for conducting remote communications |
US20070038451A1 (en) | 2003-07-08 | 2007-02-15 | Laurent Cogne | Voice recognition for large dynamic vocabularies |
US20050021344A1 (en) | 2003-07-24 | 2005-01-27 | International Business Machines Corporation | Access to enhanced conferencing services using the tele-chat system |
US7852993B2 (en) | 2003-08-11 | 2010-12-14 | Microsoft Corporation | Speech recognition enhanced caller identification |
US20070005368A1 (en) | 2003-08-29 | 2007-01-04 | Chutorash Richard J | System and method of operating a speech recognition system in a vehicle |
US20130211815A1 (en) | 2003-09-05 | 2013-08-15 | Mark Seligman | Method and Apparatus for Cross-Lingual Communication |
US20050154587A1 (en) | 2003-09-11 | 2005-07-14 | Voice Signal Technologies, Inc. | Voice enabled phone book interface for speaker dependent name recognition and phone number categorization |
US20050080786A1 (en) | 2003-10-14 | 2005-04-14 | Fish Edmund J. | System and method for customizing search results based on searcher's actual geographic location |
US20050101355A1 (en) | 2003-11-11 | 2005-05-12 | Microsoft Corporation | Sequential multimodal input |
US20110064207A1 (en) | 2003-11-17 | 2011-03-17 | Apptera, Inc. | System for Advertisement Selection, Placement and Delivery |
US8050918B2 (en) | 2003-12-11 | 2011-11-01 | Nuance Communications, Inc. | Quality evaluation tool for dynamic voice portals |
US20050188029A1 (en) | 2003-12-18 | 2005-08-25 | Pauli Asikainen | Forming a message from information shown on display |
US20110313764A1 (en) | 2003-12-23 | 2011-12-22 | At&T Intellectual Property Ii, L.P. | System and Method for Latency Reduction for Automatic Speech Recognition Using Partial Multi-Pass Results |
US7729912B1 (en) | 2003-12-23 | 2010-06-01 | At&T Intellectual Property Ii, L.P. | System and method for latency reduction for automatic speech recognition using partial multi-pass results |
US20050149326A1 (en) | 2004-01-05 | 2005-07-07 | Kabushiki Kaisha Toshiba | Speech recognition system and technique |
US20050177376A1 (en) | 2004-02-05 | 2005-08-11 | Avaya Technology Corp. | Recognition results postprocessor for use in voice recognition systems |
US7899671B2 (en) | 2004-02-05 | 2011-03-01 | Avaya, Inc. | Recognition results postprocessor for use in voice recognition systems |
US7319957B2 (en) | 2004-02-11 | 2008-01-15 | Tegic Communications, Inc. | Handwriting and voice input with automatic correction |
US20050182628A1 (en) | 2004-02-18 | 2005-08-18 | Samsung Electronics Co., Ltd. | Domain-based dialog speech recognition method and apparatus |
US20050187768A1 (en) | 2004-02-24 | 2005-08-25 | Godden Kurt S. | Dynamic N-best algorithm to reduce recognition errors |
US20050197145A1 (en) | 2004-03-03 | 2005-09-08 | Samsung Electro-Mechanics Co., Ltd. | Mobile phone capable of input of phone number without manipulating buttons and method of inputting phone number to the same |
US20050197840A1 (en) | 2004-03-05 | 2005-09-08 | Sunplus Technology Co., Ltd. | Device for event prediction on booting a motherboard |
US20050209868A1 (en) | 2004-03-19 | 2005-09-22 | Dadong Wan | Real-time sales support and learning tool |
US20100293242A1 (en) | 2004-03-31 | 2010-11-18 | Buchheit Paul T | Conversation-Based E-Mail Messaging |
US20050239495A1 (en) | 2004-04-12 | 2005-10-27 | Bayne Anthony J | System and method for the distribution of advertising and associated coupons via mobile media platforms |
US20050240406A1 (en) | 2004-04-21 | 2005-10-27 | David Carroll | Speech recognition computing device display with highlighted text |
US20130041667A1 (en) | 2004-06-02 | 2013-02-14 | Nuance Communications, Inc. | Multimodal disambiguation of speech recognition |
US7310601B2 (en) | 2004-06-08 | 2007-12-18 | Matsushita Electric Industrial Co., Ltd. | Speech recognition apparatus and speech recognition method |
US20050288926A1 (en) | 2004-06-25 | 2005-12-29 | Benco David S | Network support for wireless e-mail using speech-to-text conversion |
US20060004570A1 (en) | 2004-06-30 | 2006-01-05 | Microsoft Corporation | Transcribing speech data with dialog context and/or recognition alternative information |
US7181387B2 (en) | 2004-06-30 | 2007-02-20 | Microsoft Corporation | Homonym processing in the context of voice-activated command systems |
US20060009974A1 (en) | 2004-07-09 | 2006-01-12 | Matsushita Electric Industrial Co., Ltd. | Hands-free voice dialing for portable and remote devices |
US7133513B1 (en) | 2004-07-21 | 2006-11-07 | Sprint Spectrum L.P. | Method and system for transcribing voice content of an on-going teleconference into human-readable notation |
US20070118592A1 (en) | 2004-07-24 | 2007-05-24 | Pixcall Gmbh | Method for the transmission of additional information in a communication system, exchange device and user station |
US20060159507A1 (en) | 2004-08-13 | 2006-07-20 | Bjorn Jawerth | One-row keyboard |
US8670977B2 (en) | 2004-08-23 | 2014-03-11 | At&T Intellectual Property Ii, L.P. | System and method of lattice-based search for spoken utterance retrieval |
US20070079383A1 (en) | 2004-08-31 | 2007-04-05 | Gopalakrishnan Kumar C | System and Method for Providing Digital Content on Mobile Devices |
US20060052127A1 (en) | 2004-09-07 | 2006-03-09 | Sbc Knowledge Ventures, L.P. | System and method for voice and text based service interworking |
US20080177551A1 (en) | 2004-09-10 | 2008-07-24 | Atx Group, Inc. | Systems and Methods for Off-Board Voice-Automated Vehicle Navigation |
US20090199101A1 (en) | 2004-09-20 | 2009-08-06 | International Business Machines Corporation | Systems and methods for inputting graphical data into a graphical input field |
US20060074895A1 (en) | 2004-09-29 | 2006-04-06 | International Business Machines Corporation | Method and system for extracting and utilizing metadata to improve accuracy in speech to text conversions |
US7908141B2 (en) | 2004-09-29 | 2011-03-15 | International Business Machines Corporation | Extracting and utilizing metadata to improve accuracy in speech to text conversions |
US20060075055A1 (en) | 2004-10-06 | 2006-04-06 | Andrew Littlefield | System and method for integration of instant messaging and virtual environment clients |
US7890586B1 (en) | 2004-11-01 | 2011-02-15 | At&T Mobility Ii Llc | Mass multimedia messaging |
US20060149630A1 (en) | 2004-11-16 | 2006-07-06 | Elliott Joseph F | Opt-in delivery of advertisements on mobile devices |
US7650284B2 (en) | 2004-11-19 | 2010-01-19 | Nuance Communications, Inc. | Enabling voice click in a multimodal page |
US20080052073A1 (en) | 2004-11-22 | 2008-02-28 | National Institute Of Advanced Industrial Science And Technology | Voice Recognition Device and Method, and Program |
US20060111907A1 (en) | 2004-11-24 | 2006-05-25 | Microsoft Corporation | Generic spelling mnemonics |
US7418387B2 (en) | 2004-11-24 | 2008-08-26 | Microsoft Corporation | Generic spelling mnemonics |
US20060122834A1 (en) | 2004-12-03 | 2006-06-08 | Bennett Ian M | Emotion detection device & method for use in distributed systems |
US20060129455A1 (en) | 2004-12-15 | 2006-06-15 | Kashan Shah | Method of advertising to users of text messaging |
US7747437B2 (en) | 2004-12-16 | 2010-06-29 | Nuance Communications, Inc. | N-best list rescoring in speech recognition |
US20080077406A1 (en) | 2004-12-22 | 2008-03-27 | Nuance Communications Inc. | Mobile Dictation Correction User Interface |
US7379870B1 (en) | 2005-02-03 | 2008-05-27 | Hrl Laboratories, Llc | Contextual filtering |
WO2006101528A1 (en) | 2005-03-22 | 2006-09-28 | Sony Ericsson Mobile Communications Ab | Wireless communications device with voice-to-text conversion |
US20060217159A1 (en) | 2005-03-22 | 2006-09-28 | Sony Ericsson Mobile Communications Ab | Wireless communications device with voice-to-text conversion |
US20060235684A1 (en) | 2005-04-14 | 2006-10-19 | Sbc Knowledge Ventures, Lp | Wireless device to access network-based voice-activated services using distributed speech recognition |
US20080195588A1 (en) | 2005-05-06 | 2008-08-14 | Nhn Corporation | Personalized Search Method and System for Enabling the Method |
US7707163B2 (en) | 2005-05-25 | 2010-04-27 | Experian Marketing Solutions, Inc. | Software and metadata structures for distributed and interactive database architecture for parallel and asynchronous data processing of complex data and for real-time query processing |
US20120022875A1 (en) | 2005-06-16 | 2012-01-26 | Nuance Communications, Inc. | Synchronizing visual and speech events in a multimodal application |
US20110161276A1 (en) | 2005-06-30 | 2011-06-30 | Microsoft Corporation | Integration of location logs, gps signals, and spatial resources for identifying user activities, goals, and context |
US20100180202A1 (en) | 2005-07-05 | 2010-07-15 | Vida Software S.L. | User Interfaces for Electronic Devices |
US7640160B2 (en) | 2005-08-05 | 2009-12-29 | Voicebox Technologies, Inc. | Systems and methods for responding to natural language speech utterance |
US20070033005A1 (en) * | 2005-08-05 | 2007-02-08 | Voicebox Technologies, Inc. | Systems and methods for responding to natural language speech utterance |
US7822610B2 (en) | 2005-08-09 | 2010-10-26 | Mobile Voice Control, LLC | Use of multiple speech recognition software instances |
US7957975B2 (en) | 2005-08-09 | 2011-06-07 | Mobile Voice Control, LLC | Voice controlled wireless communication device system |
US20070038923A1 (en) | 2005-08-10 | 2007-02-15 | International Business Machines Corporation | Visual marker for speech enabled links |
US20070038740A1 (en) | 2005-08-10 | 2007-02-15 | Nortel Networks Limited | Notification service |
US20070043569A1 (en) | 2005-08-19 | 2007-02-22 | Intervoice Limited Partnership | System and method for inheritance of advertised functionality in a user interactive system |
US20090076821A1 (en) | 2005-08-19 | 2009-03-19 | Gracenote, Inc. | Method and apparatus to control operation of a playback device |
US7949529B2 (en) | 2005-08-29 | 2011-05-24 | Voicebox Technologies, Inc. | Mobile systems and methods of supporting natural language human-machine interactions |
US20120046950A1 (en) | 2005-09-12 | 2012-02-23 | Nuance Communications, Inc. | Retrieval and presentation of network service results for mobile device using a multimodal browser |
US8073700B2 (en) | 2005-09-12 | 2011-12-06 | Nuance Communications, Inc. | Retrieval and presentation of network service results for mobile device using a multimodal browser |
US20070061146A1 (en) | 2005-09-12 | 2007-03-15 | International Business Machines Corporation | Retrieval and Presentation of Network Service Results for Mobile Device Using a Multimodal Browser |
US20130158994A1 (en) | 2005-09-12 | 2013-06-20 | Nuance Communications, Inc. | Retrieval and presentation of network service results for mobile device using a multimodal browser |
US8032372B1 (en) | 2005-09-13 | 2011-10-04 | Escription, Inc. | Dictation selection |
US20070061148A1 (en) | 2005-09-13 | 2007-03-15 | Cross Charles W Jr | Displaying speech command input state information in a multimodal browser |
US7769764B2 (en) | 2005-09-14 | 2010-08-03 | Jumptap, Inc. | Mobile advertisement syndication |
US20070061300A1 (en) | 2005-09-14 | 2007-03-15 | Jorey Ramer | Mobile advertisement syndication |
US20070086773A1 (en) | 2005-10-14 | 2007-04-19 | Fredrik Ramsten | Method for creating and operating a user interface |
US20070115845A1 (en) | 2005-10-24 | 2007-05-24 | Christian Hochwarth | Network time out handling |
US20070106506A1 (en) | 2005-11-07 | 2007-05-10 | Ma Changxue C | Personal synergic filtering of multimodal inputs |
US7640158B2 (en) | 2005-11-08 | 2009-12-29 | Multimodal Technologies, Inc. | Automatic detection and application of editing patterns in draft documents |
US20070106507A1 (en) | 2005-11-09 | 2007-05-10 | International Business Machines Corporation | Noise playback enhancement of prerecorded audio for speech recognition operations |
US20070118374A1 (en) | 2005-11-23 | 2007-05-24 | Wise Gerald B | Method for generating closed captions |
US20070123222A1 (en) | 2005-11-29 | 2007-05-31 | International Business Machines Corporation | Method and system for invoking push-to-service offerings |
US20100268726A1 (en) | 2005-11-30 | 2010-10-21 | Anchorfree, Inc. | Computerized system and method for advanced advertising |
US7925716B2 (en) | 2005-12-05 | 2011-04-12 | Yahoo! Inc. | Facilitating retrieval of information within a messaging environment |
US20070133769A1 (en) | 2005-12-08 | 2007-06-14 | International Business Machines Corporation | Voice navigation of a visual view for a session in a composite services enablement environment |
US8126120B2 (en) | 2005-12-12 | 2012-02-28 | Tellme Networks, Inc. | Providing missed call and message information |
US20070133771A1 (en) | 2005-12-12 | 2007-06-14 | Stifelman Lisa J | Providing missed call and message information |
US20100312640A1 (en) | 2005-12-16 | 2010-12-09 | Apptera, Inc. | Call-Based Advertising |
US20070156400A1 (en) | 2006-01-03 | 2007-07-05 | Wheeler Mark R | System and method for wireless dictation and transcription |
US20070180718A1 (en) | 2006-01-06 | 2007-08-09 | Tcl Communication Technology Holdings, Ltd. | Method for entering commands and/or characters for a portable communication device equipped with a tilt sensor |
US8106285B2 (en) | 2006-02-10 | 2012-01-31 | Harman Becker Automotive Systems Gmbh | Speech-driven selection of an audio file |
US20080063155A1 (en) | 2006-02-10 | 2008-03-13 | Spinvox Limited | Mass-Scale, User-Independent, Device-Independent Voice Messaging System |
US20080133232A1 (en) | 2006-02-10 | 2008-06-05 | Spinvox Limited | Mass-Scale, User-Independent, Device-Independent Voice Messaging System |
US8010358B2 (en) | 2006-02-21 | 2011-08-30 | Sony Computer Entertainment Inc. | Voice recognition with parallel gender and age normalization |
US7908273B2 (en) | 2006-03-09 | 2011-03-15 | Gracenote, Inc. | Method and system for media navigation |
US20090077493A1 (en) | 2006-03-10 | 2009-03-19 | Continental Automotive Gmbh | Method for the Selection of Functions with the Aid of a User Interface, and User Interface |
US20070233488A1 (en) | 2006-03-29 | 2007-10-04 | Dictaphone Corporation | System and method for applying dynamic contextual grammars and language models to improve automatic speech recognition accuracy |
US20130226894A1 (en) | 2006-03-30 | 2013-08-29 | Veveo, Inc. | Method and System for Incrementally Selecting and Providing Relevant Search Engines in Response to a User Query |
US20070233487A1 (en) | 2006-04-03 | 2007-10-04 | Cohen Michael H | Automatic language model update |
US8117268B2 (en) | 2006-04-05 | 2012-02-14 | Jablokov Victor R | Hosted voice recognition system for wireless devices |
US20070239837A1 (en) | 2006-04-05 | 2007-10-11 | Yap, Inc. | Hosted voice recognition system for wireless devices |
US9009055B1 (en) | 2006-04-05 | 2015-04-14 | Canyon Ip Holdings Llc | Hosted voice recognition system for wireless devices |
US9583107B2 (en) | 2006-04-05 | 2017-02-28 | Amazon Technologies, Inc. | Continuous speech transcription performance indication |
US8498872B2 (en) | 2006-04-05 | 2013-07-30 | Canyon Ip Holdings Llc | Filtering transcriptions of utterances |
US9542944B2 (en) | 2006-04-05 | 2017-01-10 | Amazon Technologies, Inc. | Hosted voice recognition system for wireless devices |
US20150255067A1 (en) | 2006-04-05 | 2015-09-10 | Canyon IP Holding LLC | Filtering transcriptions of utterances using received information to correct transcription errors |
US8433574B2 (en) | 2006-04-05 | 2013-04-30 | Canyon IP Holdings, LLC | Hosted voice recognition system for wireless devices |
US20090124272A1 (en) | 2006-04-05 | 2009-05-14 | Marc White | Filtering transcriptions of utterances |
US8121838B2 (en) | 2006-04-11 | 2012-02-21 | Nuance Communications, Inc. | Method and system for automatic transcription prioritization |
US20140136199A1 (en) | 2006-04-17 | 2014-05-15 | Vovision, Llc | Correcting transcribed audio files with an email-client interface |
US20090276215A1 (en) | 2006-04-17 | 2009-11-05 | Hager Paul M | Methods and systems for correcting transcribed audio files |
US20070255794A1 (en) | 2006-07-12 | 2007-11-01 | Marengo Intellectual Property Ltd. | Multi-conversation instant messaging |
US20080037720A1 (en) | 2006-07-27 | 2008-02-14 | Speechphone, Llc | Voice Activated Communication Using Automatically Updated Address Books |
US20090100050A1 (en) | 2006-07-31 | 2009-04-16 | Berna Erol | Client device for interacting with a mixed media reality recognition system |
US20080065737A1 (en) | 2006-08-03 | 2008-03-13 | Yahoo! Inc. | Electronic document information extraction |
US20080063154A1 (en) * | 2006-08-09 | 2008-03-13 | Yossi Tamari | System and method of customized event notification |
US20080040683A1 (en) | 2006-08-11 | 2008-02-14 | David Walsh | Multi-pane graphical user interface with common scroll control |
US7796980B1 (en) | 2006-08-11 | 2010-09-14 | Sprint Communications Company L.P. | Remote mobile voice control of digital/personal video recorder |
US20080052075A1 (en) | 2006-08-25 | 2008-02-28 | Microsoft Corporation | Incrementally regulated discriminative margins in MCE training for speech recognition |
US8145493B2 (en) | 2006-09-11 | 2012-03-27 | Nuance Communications, Inc. | Establishing a preferred mode of interaction between a user and a multimodal application |
US20080065481A1 (en) | 2006-09-13 | 2008-03-13 | Microsoft Corporation | User-associated, interactive advertising monetization |
US20100278453A1 (en) | 2006-09-15 | 2010-11-04 | King Martin T | Capture and display of annotations in paper and electronic documents |
US20090282363A1 (en) | 2006-09-15 | 2009-11-12 | Microsoft Corporation | Efficient navigation of search results |
US20080200153A1 (en) | 2006-09-28 | 2008-08-21 | Dudley Fitzpatrick | Apparatuses, methods and systems for code triggered information querying and serving on mobile devices based on profiles |
US7907705B1 (en) | 2006-10-10 | 2011-03-15 | Intuit Inc. | Speech to text for assisted form completion |
US20080091426A1 (en) | 2006-10-12 | 2008-04-17 | Rod Rempel | Adaptive context for automatic speech recognition systems |
US20080120375A1 (en) * | 2006-11-16 | 2008-05-22 | Benjamin Levy | Activity partner matching system and method |
US20100121629A1 (en) | 2006-11-28 | 2010-05-13 | Cohen Sanford H | Method and apparatus for translating speech during a call |
US8027836B2 (en) | 2006-11-30 | 2011-09-27 | Nuance Communications, Inc. | Phonetic decoding and concatentive speech synthesis |
US20110040629A1 (en) | 2006-12-06 | 2011-02-17 | Apptera, Inc. | Behavior aggregation |
US20090037255A1 (en) | 2006-12-06 | 2009-02-05 | Leo Chiu | Behavior aggregation |
US20110047452A1 (en) | 2006-12-06 | 2011-02-24 | Nuance Communications, Inc. | Enabling grammars in web page frame |
US20080172781A1 (en) | 2006-12-15 | 2008-07-24 | Terrance Popowich | System and method for obtaining and using advertising information |
US7899670B1 (en) | 2006-12-21 | 2011-03-01 | Escription Inc. | Server-based speech recognition |
US20080154600A1 (en) | 2006-12-21 | 2008-06-26 | Nokia Corporation | System, Method, Apparatus and Computer Program Product for Providing Dynamic Vocabulary Prediction for Speech Recognition |
US20080155060A1 (en) | 2006-12-22 | 2008-06-26 | Yahoo! Inc. | Exported overlays |
US8296139B2 (en) | 2006-12-22 | 2012-10-23 | International Business Machines Corporation | Adding real-time dictation capabilities for speech processing operations handled by a networked speech processing system |
US20080154870A1 (en) | 2006-12-26 | 2008-06-26 | Voice Signal Technologies, Inc. | Collection and use of side information in voice-mediated mobile search |
US20100286901A1 (en) | 2007-01-10 | 2010-11-11 | Pieter Geelen | Navigation device and method relating to an audible recognition mode |
US20090141875A1 (en) | 2007-01-10 | 2009-06-04 | Michael Demmitt | System and Method for Delivery of Voicemails to Handheld Devices |
US8069047B2 (en) | 2007-02-12 | 2011-11-29 | Nuance Communications, Inc. | Dynamically defining a VoiceXML grammar in an X+V page of a multimodal application |
US8510094B2 (en) | 2007-02-14 | 2013-08-13 | Google Inc. | Machine translation feedback |
US8380511B2 (en) | 2007-02-20 | 2013-02-19 | Intervoice Limited Partnership | System and method for semantic categorization |
US20080201139A1 (en) | 2007-02-20 | 2008-08-21 | Microsoft Corporation | Generic framework for large-margin MCE training in speech recognition |
US20080198980A1 (en) | 2007-02-21 | 2008-08-21 | Jens Ulrik Skakkebaek | Voicemail filtering and transcription |
US20080198981A1 (en) | 2007-02-21 | 2008-08-21 | Jens Ulrik Skakkebaek | Voicemail filtering and transcription |
US20080198898A1 (en) | 2007-02-21 | 2008-08-21 | Taylor John P | Apparatus, system and method for high resolution identification with temperature dependent resistive device |
US20080208590A1 (en) | 2007-02-27 | 2008-08-28 | Cross Charles W | Disambiguating A Speech Recognition Grammar In A Multimodal Application |
US7890329B2 (en) | 2007-03-03 | 2011-02-15 | Industrial Technology Research Institute | Apparatus and method to reduce recognition errors through context relations among dialogue turns |
US20110054900A1 (en) | 2007-03-07 | 2011-03-03 | Phillips Michael S | Hybrid command and control between resident and remote speech recognition facilities in a mobile voice-to-speech application |
US20080221897A1 (en) | 2007-03-07 | 2008-09-11 | Cerra Joseph P | Mobile environment speech processing facility |
US20120095831A1 (en) | 2007-03-09 | 2012-04-19 | Janne Aaltonen | Method and apparatus for controlling user communications |
US20080243504A1 (en) | 2007-03-30 | 2008-10-02 | Verizon Data Services, Inc. | System and method of speech recognition training based on confirmed speaker utterances |
US20080243500A1 (en) | 2007-03-30 | 2008-10-02 | Maximilian Bisani | Automatic Editing Using Probabilistic Word Substitution Models |
US9330401B2 (en) | 2007-04-05 | 2016-05-03 | Amazon Technologies, Inc. | Validation of mobile advertising from derived information |
US9384735B2 (en) | 2007-04-05 | 2016-07-05 | Amazon Technologies, Inc. | Corrective feedback loop for automated speech recognition |
US20170004831A1 (en) | 2007-04-05 | 2017-01-05 | Amazon Technologies, Inc. | Corrective feedback loop for automated speech recognition |
US20080275864A1 (en) | 2007-05-02 | 2008-11-06 | Yahoo! Inc. | Enabling clustered search processing via text messaging |
US20100279667A1 (en) | 2007-05-22 | 2010-11-04 | Wehrs Michael E | Keyword-based services for mobile device messages |
US20100312619A1 (en) | 2007-05-23 | 2010-12-09 | Pekka Ala-Pietila | Method and a system for providing mobile communications services |
US20080301250A1 (en) | 2007-05-29 | 2008-12-04 | Michael Thomas Hardy | Thread-based message prioritization |
US20080313039A1 (en) | 2007-06-18 | 2008-12-18 | Utbk, Inc. | Systems and Methods to Facilitate the Specification of a Complex Geographic Area |
US20080317219A1 (en) | 2007-06-21 | 2008-12-25 | Siemens Communications, Inc. | Method and apparatus for context based voice dialing |
US20090006194A1 (en) | 2007-06-27 | 2009-01-01 | Microsoft Corporation | Location, destination and other contextual information-based mobile advertisements |
US20090012793A1 (en) | 2007-07-03 | 2009-01-08 | Dao Quyen C | Text-to-speech assist for portable communication devices |
US20090150405A1 (en) | 2007-07-13 | 2009-06-11 | Grouf Nicholas A | Systems and Methods for Expressing Data Using a Media Markup Language |
US20100146077A1 (en) | 2007-07-30 | 2010-06-10 | Nds Limited | Providing informatin about video content |
US20090043855A1 (en) | 2007-08-08 | 2009-02-12 | Blake Bookstaff | System for providing information to originator of misdirected email |
US20090055538A1 (en) * | 2007-08-21 | 2009-02-26 | Microsoft Corporation | Content commentary |
US8335830B2 (en) | 2007-08-22 | 2012-12-18 | Canyon IP Holdings, LLC. | Facilitating presentation by mobile device of additional content for a word or phrase upon utterance thereof |
US8296377B1 (en) | 2007-08-22 | 2012-10-23 | Canyon IP Holdings, LLC. | Facilitating presentation by mobile device of additional content for a word or phrase upon utterance thereof |
US9436951B1 (en) | 2007-08-22 | 2016-09-06 | Amazon Technologies, Inc. | Facilitating presentation by mobile device of additional content for a word or phrase upon utterance thereof |
US20090055175A1 (en) | 2007-08-22 | 2009-02-26 | Terrell Ii James Richard | Continuous speech transcription performance indication |
US8140632B1 (en) | 2007-08-22 | 2012-03-20 | Victor Roditis Jablokov | Facilitating presentation by mobile device of additional content for a word or phrase upon utterance thereof |
US8510109B2 (en) | 2007-08-22 | 2013-08-13 | Canyon Ip Holdings Llc | Continuous speech transcription performance indication |
US8543396B2 (en) | 2007-08-22 | 2013-09-24 | Canyon Ip Holdings Llc | Continuous speech transcription performance indication |
US20090076917A1 (en) | 2007-08-22 | 2009-03-19 | Victor Roditis Jablokov | Facilitating presentation of ads relating to words of a message |
US20100058200A1 (en) | 2007-08-22 | 2010-03-04 | Yap, Inc. | Facilitating presentation by mobile device of additional content for a word or phrase upon utterance thereof |
US8335829B1 (en) | 2007-08-22 | 2012-12-18 | Canyon IP Holdings, LLC | Facilitating presentation by mobile device of additional content for a word or phrase upon utterance thereof |
US9053489B2 (en) | 2007-08-22 | 2015-06-09 | Canyon Ip Holdings Llc | Facilitating presentation of ads relating to words of a message |
US8135578B2 (en) | 2007-08-24 | 2012-03-13 | Nuance Communications, Inc. | Creation and use of application-generic class-based statistical language models for automatic speech recognition |
US20090055179A1 (en) | 2007-08-24 | 2009-02-26 | Samsung Electronics Co., Ltd. | Method, medium and apparatus for providing mobile voice web service |
US20090063151A1 (en) | 2007-08-28 | 2009-03-05 | Nexidia Inc. | Keyword spotting using a phoneme-sequence index |
US20090063268A1 (en) | 2007-09-04 | 2009-03-05 | Burgess David A | Targeting Using Historical Data |
US20090086958A1 (en) | 2007-10-02 | 2009-04-02 | Utbk, Inc. | Systems and Methods to Provide Alternative Connections for Real Time Communications |
US8311825B2 (en) | 2007-10-04 | 2012-11-13 | Kabushiki Kaisha Toshiba | Automatic speech recognition method and apparatus |
US20130281007A1 (en) | 2007-10-05 | 2013-10-24 | Qualcomm Incorporated | Location and time based filtering of broadcast information |
US20090182559A1 (en) | 2007-10-08 | 2009-07-16 | Franz Gerl | Context sensitive multi-stage speech recognition |
US20090117922A1 (en) | 2007-11-01 | 2009-05-07 | David Rowland Bell | Alerts based on significance of free format text messages |
US20090125299A1 (en) | 2007-11-09 | 2009-05-14 | Jui-Chang Wang | Speech recognition system |
US20090150156A1 (en) | 2007-12-11 | 2009-06-11 | Kennewick Michael R | System and method for providing a natural language voice user interface in an integrated voice navigation services environment |
US8145485B2 (en) | 2007-12-17 | 2012-03-27 | Verizon Patent And Licensing Inc. | Grammar weighting voice recognition information |
US8611871B2 (en) | 2007-12-25 | 2013-12-17 | Canyon Ip Holdings Llc | Validation of mobile advertising from derived information |
US20090163187A1 (en) | 2007-12-25 | 2009-06-25 | Yap, Inc. | Validation of mobile advertising from derived information |
US20090182560A1 (en) | 2008-01-16 | 2009-07-16 | Yap, Inc. | Using a physical phenomenon detector to control operation of a speech recognition engine |
US8326636B2 (en) | 2008-01-16 | 2012-12-04 | Canyon Ip Holdings Llc | Using a physical phenomenon detector to control operation of a speech recognition engine |
US20100017294A1 (en) | 2008-01-24 | 2010-01-21 | Mailmethods, Llc | Email advertisement system and method |
US20090204410A1 (en) | 2008-02-13 | 2009-08-13 | Sensory, Incorporated | Voice interface and search for electronic devices including bluetooth headsets and remote systems |
US20090210214A1 (en) | 2008-02-19 | 2009-08-20 | Jiang Qian | Universal Language Input |
US8352261B2 (en) | 2008-03-07 | 2013-01-08 | Canyon IP Holdings, LLC | Use of intermediate speech transcription results in editing final speech transcription results |
US20090228274A1 (en) | 2008-03-07 | 2009-09-10 | Yap Inc. | Use of intermediate speech transcription results in editing final speech transcription results |
US8352264B2 (en) | 2008-03-19 | 2013-01-08 | Canyon IP Holdings, LLC | Corrective feedback loop for automated speech recognition |
US8793122B2 (en) | 2008-03-19 | 2014-07-29 | Canyon IP Holdings, LLC | Corrective feedback loop for automated speech recognition |
US20090240488A1 (en) | 2008-03-19 | 2009-09-24 | Yap, Inc. | Corrective feedback loop for automated speech recognition |
US20090248415A1 (en) | 2008-03-31 | 2009-10-01 | Yap, Inc. | Use of metadata to post process speech recognition output |
US7680661B2 (en) | 2008-05-14 | 2010-03-16 | Nuance Communications, Inc. | Method and system for improved speech recognition |
US20090307090A1 (en) | 2008-06-05 | 2009-12-10 | Embarq Holdings Company, Llc | System and Method for Inserting Advertisements in Voicemail |
US20090312040A1 (en) | 2008-06-13 | 2009-12-17 | Embarq Holdings Company, Llc | System and method for inserting advertisements into SMS messages |
US20090319187A1 (en) | 2008-06-23 | 2009-12-24 | Outside.In, Inc. | Generating Geocoded Targeted Web Advertisements |
US20110161072A1 (en) | 2008-08-20 | 2011-06-30 | Nec Corporation | Language model creation apparatus, language model creation method, speech recognition apparatus, speech recognition method, and recording medium |
US20100049525A1 (en) | 2008-08-22 | 2010-02-25 | Yap, Inc. | Methods, apparatuses, and systems for providing timely user cues pertaining to speech recognition |
US8301454B2 (en) | 2008-08-22 | 2012-10-30 | Canyon Ip Holdings Llc | Methods, apparatuses, and systems for providing timely user cues pertaining to speech recognition |
US20110296374A1 (en) | 2008-11-05 | 2011-12-01 | Google Inc. | Custom language models |
US20100223056A1 (en) | 2009-02-27 | 2010-09-02 | Autonomy Corporation Ltd. | Various apparatus and methods for a speech recognition system |
US8229743B2 (en) | 2009-06-23 | 2012-07-24 | Autonomy Corporation Ltd. | Speech recognition system |
US20110144973A1 (en) | 2009-12-15 | 2011-06-16 | At&T Intellectual Property I, L.P. | System and method for combining geographic metadata in automatic speech recognition language and acoustic models |
US8417530B1 (en) | 2010-08-20 | 2013-04-09 | Google Inc. | Accent-influenced search results |
US9099087B2 (en) | 2010-09-03 | 2015-08-04 | Canyon IP Holdings, LLC | Methods and systems for obtaining language models for transcribing communications |
US20120059653A1 (en) | 2010-09-03 | 2012-03-08 | Adams Jeffrey P | Methods and systems for obtaining language models for transcribing communications |
US8898065B2 (en) | 2011-01-07 | 2014-11-25 | Nuance Communications, Inc. | Configurable speech recognition system using multiple recognizers |
US9093061B1 (en) | 2011-04-14 | 2015-07-28 | Canyon IP Holdings, LLC. | Speech recognition with hierarchical networks |
US20120324391A1 (en) | 2011-06-16 | 2012-12-20 | Microsoft Corporation | Predictive word completion |
US8589164B1 (en) | 2012-10-18 | 2013-11-19 | Google Inc. | Methods and systems for speech recognition processing using search query information |
Non-Patent Citations (48)
Title |
---|
"International Search Report" and "Written Opinion of the International Search Authority" (Korean Intellectual Property Office) in Yap, Inc. International Patent Application Serial No. PCT/US2007/008621 corresponding to current U.S. patent application, Dated Nov. 13, 2007, 8 pages. |
"International Search Report"and "Written Opinion of the International Search Authority" (Korean Intellectual Property Office) in Yap, Inc. International Patent Application Serial No. PCT/US2007/008621, dated Nov. 13, 2007, 13 pages total. |
Allauzen, C., et al., A Generalized Composition Algorithm for Weighted Finite-State Transducers, Interspeech, Brighton, U.K., Sep. 2009, pp. 1203-1206. |
Bisani, M., et al., Automatic Editing in a Back-End Speech-to-Text System, 2008, 7 pages. |
Board of Patent Appeals and Interferences Answer in U.S. Appl. No. 12/352,442 dated May 15, 2012. |
Brown, E., et al., Capitalization Recovery for Text, Springer-Verlag Berlin Heidelberg, 2002, 12 pages. |
David H. Kemsley, et al., A Survey of Neural Network Research and Fielded Applications, 1992, in International Journal of Neural Networks: Research and Applications, vol. 2, No. 2/3/4, pp. 123-133. Accessed on Oct. 25, 2007 at http://citeseer.ist.psu.edu/cache/papers/cs/25638/ftp:zSzzSzaxon.cs.byu.eduzSzpubzSzpaperszSzkemsley_92.pdf/kemsley92survey.pdf, 12 pages total. |
David H. Kemsley, et al., A Survey of Neural Network Research and Fielded Applications, 1992, in International Journal of Neural Networks: Research and Applications, vol. 2, No. 2/3/4, pp. 123-133. Accessed on Oct. 25, 2007 at http://citeseer.ist.psu.edu/cache/papers/cs/25638/ftp:zSzzSzaxon.cs.byu.eduzSzpubzSzpaperszSzkemsley_92.pdf/kemsley92survey.pdf. |
Desilets, A., et al., Extracting Keyphrases From Spoken Audio Documents, Springer-Verlag Berlin Heidelberg, 2002, 15 pages. |
Fielding, et al., Hypertext Transfer Protocol-HTTP/1.1, RFC 2616, Network Working Group, sections 7, 9.5, 14.30, 12 pages total. |
Fielding, et al., Hypertext Transfer Protocol—HTTP/1.1, RFC 2616, Network Working Group, sections 7, 9.5, 14.30, 12 pages total. |
Glaser, M., et al., Web-Based Telephony Bridges for the Deaf, proceedings of the South African Telecommunications Networks & Applications Conference (2001), Wild Coast Sun, South Africa, 5 pages. |
Gotoh, Y., et al., Sentence Boundary Detection in Broadcase Speech Transcripts, Proceedings of the ISCA Workshop, 2000, 8 pages. |
Hori, T., et al., Efficient WFST-Based One-Pass Decoding With On-The-Fly Hypothesis Rescoring in Extremely Large Vocabulary Continuous Speech Recognition, IEEE Transactions on Audio, Speech, and Language Processing, vol. 15, No. 4, May 2007, pp. 1352-1365. |
Huang, J., et al., Extracting Caller Information from Voicemail, IBM T.J. Watson Research Center, 2002, pp. 67-77. |
Huang, J., et al., Maximum Entropy Model for Punctuation Annotation From Speech, in ICSLP 2002, pp. 917-920. |
Huang, J., Zweig, G. Padmanabhan, M., 2002, Extracting caller information from voicemail, Springer-Verlag Berlin Heidelberg, 11 pages. |
Information Disclosure Statement (IDS) Letter Regarding Common Patent Application(s) dated Jun. 4, 2010. |
Information Disclosure Statement (IDS) Letter Regarding Common Patent Application(s), dated Dec. 6, 2010. |
Information Disclosure Statement (IDS) Letter Regarding Common Patent Application(s), dated Feb. 14, 2012. |
Information Disclosure Statement (IDS) Letter Regarding Common Patent Application(s), dated Jul. 21, 2011. |
Information Disclosure Statement (IDS) Letter Regarding Common Patent Application(s), dated Jun. 4, 2010. |
Information Disclosure Statement (IDS) Letter Regarding Common Patent Application(s), dated Mar. 17, 2011. |
Information Disclosure Statement (IDS) Letter Regarding Common Patent Application(s), dated Nov. 24, 2009. |
Information Disclosure Statement (IDS) Letter Regarding Common Patent Application(s), submitted by Applicant on Jul. 21, 2009. |
International Search Report and Written Opinion International Patent Application No. PCT/US2007/008621, dated Nov. 13, 2007. |
J2EE Application Overview, publicly available on http://www.orionserever.com/docs/j2eeoverview.html, Mar. 1, 2001. |
J2EE Application Overview, publicly available on http://www/orionserver.com/docs/j2eeoverview.html since Mar. 1, 2001. Retrieved on Oct. 26, 2007, 3 pages total. |
J2EE Application Overview, publicly available on http://www/orionserver.com/docs/j2eeoverview.html since Mar. 1, 2001. Retrieved on Oct. 26, 2007. |
Justo, R., et al., Phrase Classes in Two-Level Language Models for ASR, Springer-Verlag London Limited, 2008, 11 pages. |
Kimura, K., et al., 1992, Association-based natural language processing with neural networks, in proceedings of the 7th annual meeting of the Association of Computational Linguistics, pp. 223-231. |
Knudsen, Jonathan, Session Handling in MIDP, Jan. 2002, retrieved from http://developers.sun.com/mobility/midp/articles/sessions/ on Jul. 25, 2008, 7 pages total. |
Lewis, J., et al., SoftBridge: An Architecture for Building IP-Based Bridges Over the Digital Divide, Proceedings of the South African Telecommunications Networks & Applications Conference (SATNAC 2002), Drakensberg, South Africa, 5 pages. |
Li, X., et al., Time based language models, CIKM '03 Proceedings of the twelfth international conference on Information and knowledge management, pp. 469-475, 2003. |
Marshall, James, HTTP Made Really Easy, Aug. 15, 1997, retrieved from http://www.jmarshall.com/easy/http/ on Jul. 25, 2008, 15 pages total. |
Office Action in Canadian Application No. 2648617 dated Feb. 27, 2014. |
Ries, K., Segmenting conversations by topic, initiative, and style, Springer-Verlag Berlin Heidelberg, 2002, 16 pages. |
Schalkwyk, J., et al., Speech Recognition with Dynamic Grammars Using Finite-State Transducers, Eurospeech 2003-Geneva, pp. 1969-1972. |
Shriberg, E., et al., Prosody-based automatic segmentation of speech into sentences and topics, 2000, 31 pages. |
SoftBridge: An Architecture for Building IP-based Bridges over the Digital Divide, Lewis et al. |
Soltau, H., and G. Saon, Dynamic Network Decoding Revisited, Automatic Speech Recognition and Understanding, 2009, IEEE Workshop, pp. 276-281. |
Stent, A., et al., Geo-Centric Language Models for Local Business Voice Search, Human Language Technologies: The 2009 Annual Conference of the North American Chapter of the ACL, pp. 386-396, 2009. |
Thomae, M., et al., Hierarchical Language Models for One-Stage Speech Interpretation, in Interspeech, 2005, pp. 3425-3428. |
Transl8it! translation engine, publicly available on http://www.transl8it.com since May 30, 2002. Retrieved on Oct. 26, 2007, 6 pages total. |
Transl8it! translation engine, publicly available on http://www.transl8it.com since May 30, 2002. Retrieved on Oct. 26, 2007. |
vBulletin Community Forum, thread posted on Mar. 5, 2004. Page retrieved on Oct. 26, 2007 from http://www.vbulletin.com/forum/showthread.php?t=96976, 1 page total. |
vBulletin Community Forum, thread posted on Mar. 5, 2004. Page retrieved on Oct. 26, 2007 from http://www.vbulletin.com/forum/showthread.php?t=96976. |
Web-based Telephony Bridges for the Deaf, Glaser et al. |
Cited By (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US11803532B2 (en) | 2017-01-04 | 2023-10-31 | Palantir Technologies Inc. | Integrated data analysis |
US10714080B2 (en) * | 2017-02-10 | 2020-07-14 | Samsung Electronics Co., Ltd. | WFST decoding system, speech recognition system including the same and method for storing WFST data |
Also Published As
Publication number | Publication date |
---|---|
US20090083032A1 (en) | 2009-03-26 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US9973450B2 (en) | Methods and systems for dynamically updating web service profile information by parsing transcribed message strings | |
US9583107B2 (en) | Continuous speech transcription performance indication | |
US9099090B2 (en) | Timely speech recognition | |
US8498872B2 (en) | Filtering transcriptions of utterances | |
US9542944B2 (en) | Hosted voice recognition system for wireless devices | |
US9940931B2 (en) | Corrective feedback loop for automated speech recognition | |
US8352261B2 (en) | Use of intermediate speech transcription results in editing final speech transcription results | |
US9053489B2 (en) | Facilitating presentation of ads relating to words of a message | |
US8676577B2 (en) | Use of metadata to post process speech recognition output | |
US20090076917A1 (en) | Facilitating presentation of ads relating to words of a message | |
US8825770B1 (en) | Facilitating presentation by mobile device of additional content for a word or phrase upon utterance thereof | |
US20150255067A1 (en) | Filtering transcriptions of utterances using received information to correct transcription errors | |
US9436951B1 (en) | Facilitating presentation by mobile device of additional content for a word or phrase upon utterance thereof |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
AS | Assignment |
Owner name: YAP, INC., NORTH CAROLINA Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:JABLOKOV, VICTOR RODITIS;JABLOKOV, IGOR RODITIS;REEL/FRAME:021761/0300 Effective date: 20081021 |
|
AS | Assignment |
Owner name: VENTURE LENDING & LEASING V, INC., CALIFORNIA Free format text: SECURITY AGREEMENT;ASSIGNOR:YAP INC.;REEL/FRAME:025521/0513 Effective date: 20100924 Owner name: VENTURE LENDING & LEASING VI, INC., CALIFORNIA Free format text: SECURITY AGREEMENT;ASSIGNOR:YAP INC.;REEL/FRAME:025521/0513 Effective date: 20100924 |
|
AS | Assignment |
Owner name: YAP INC., NORTH CAROLINA Free format text: RELEASE BY SECURED PARTY;ASSIGNOR:VENTIRE LENDING & LEASING V, INC. AND VENTURE LENDING & LEASING VI, INC.;REEL/FRAME:027001/0859 Effective date: 20110908 |
|
AS | Assignment |
Owner name: CANYON IP HOLDINGS LLC, DELAWARE Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:YAP LLC;REEL/FRAME:027770/0733 Effective date: 20120223 |
|
AS | Assignment |
Owner name: AMAZON TECHNOLOGIES, INC., WASHINGTON Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:CANYON IP HOLDINGS LLC;REEL/FRAME:037083/0914 Effective date: 20151106 |
|
AS | Assignment |
Owner name: YAP LLC, WASHINGTON Free format text: ENTITY CONVERSION;ASSIGNOR:YAP INC.;REEL/FRAME:040549/0866 Effective date: 20110921 |
|
STCF | Information on status: patent grant |
Free format text: PATENTED CASE |
|
CC | Certificate of correction | ||
MAFP | Maintenance fee payment |
Free format text: PAYMENT OF MAINTENANCE FEE, 4TH YEAR, LARGE ENTITY (ORIGINAL EVENT CODE: M1551); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY Year of fee payment: 4 |