WO2012138587A2 - Audio-interactive message exchange - Google Patents

Audio-interactive message exchange Download PDF

Info

Publication number
WO2012138587A2
WO2012138587A2 PCT/US2012/031778 US2012031778W WO2012138587A2 WO 2012138587 A2 WO2012138587 A2 WO 2012138587A2 US 2012031778 W US2012031778 W US 2012031778W WO 2012138587 A2 WO2012138587 A2 WO 2012138587A2
Authority
WO
WIPO (PCT)
Prior art keywords
user
audio
message
text
input
Prior art date
Application number
PCT/US2012/031778
Other languages
French (fr)
Other versions
WO2012138587A3 (en
Inventor
Liane AIHARA
Shane LANDRY
Lisa Stifelman
Madhusudan Chinthakunta
Anne Sullivan
Kathleen LEE
Original Assignee
Microsoft Corporation
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Microsoft Corporation filed Critical Microsoft Corporation
Priority to JP2014503705A priority Critical patent/JP2014512049A/en
Priority to EP20120768271 priority patent/EP2695406A4/en
Priority to CN2012800164763A priority patent/CN103443852A/en
Priority to KR1020137026109A priority patent/KR20140022824A/en
Publication of WO2012138587A2 publication Critical patent/WO2012138587A2/en
Publication of WO2012138587A3 publication Critical patent/WO2012138587A3/en

Links

Classifications

    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS OR SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING; SPEECH OR AUDIO CODING OR DECODING
    • G10L15/00Speech recognition
    • G10L15/26Speech to text systems
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS OR SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING; SPEECH OR AUDIO CODING OR DECODING
    • G10L13/00Speech synthesis; Text to speech systems
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS OR SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING; SPEECH OR AUDIO CODING OR DECODING
    • G10L13/00Speech synthesis; Text to speech systems
    • G10L13/08Text analysis or generation of parameters for speech synthesis out of text, e.g. grapheme to phoneme translation, prosody generation or stress or intonation determination
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04MTELEPHONIC COMMUNICATION
    • H04M1/00Substation equipment, e.g. for use by subscribers
    • H04M1/26Devices for calling a subscriber
    • H04M1/27Devices whereby a plurality of signals may be stored simultaneously
    • H04M1/271Devices whereby a plurality of signals may be stored simultaneously controlled by voice recognition
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04MTELEPHONIC COMMUNICATION
    • H04M1/00Substation equipment, e.g. for use by subscribers
    • H04M1/72Mobile telephones; Cordless telephones, i.e. devices for establishing wireless links to base stations without route selection
    • H04M1/724User interfaces specially adapted for cordless or mobile telephones
    • H04M1/72403User interfaces specially adapted for cordless or mobile telephones with means for local support of applications that increase the functionality
    • H04M1/7243User interfaces specially adapted for cordless or mobile telephones with means for local support of applications that increase the functionality with interactive means for internal management of messages
    • H04M1/72436User interfaces specially adapted for cordless or mobile telephones with means for local support of applications that increase the functionality with interactive means for internal management of messages for text messaging, e.g. SMS or e-mail
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04WWIRELESS COMMUNICATION NETWORKS
    • H04W4/00Services specially adapted for wireless communication networks; Facilities therefor
    • H04W4/12Messaging; Mailboxes; Announcements
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04MTELEPHONIC COMMUNICATION
    • H04M2250/00Details of telephonic subscriber devices
    • H04M2250/74Details of telephonic subscriber devices with voice recognition means

Definitions

  • Multi-modal communications through fixed or portable computing devices such as desktop computers, vehicle mount computers, portable computers, smart phones, and similar devices are a common occurrence. Because many facets of communications are controlled through easily customizable software / hardware combinations, previously unheard-of features are available for use in daily life. For example, integration of presence information into communication applications enables people to communicate with each other more efficiently. Simultaneous reduction in size and increase in computing capabilities enables use of smart phones or similar handheld computing devices for multi-modal communications including, but not limited to, audio, video, text message exchange, email, instant messaging, social networking posts/updates, etc.
  • Embodiments are directed to providing a completely hands free exchange of messages, especially in portable devices through a combination of speech recognition, text-to-speech (TTS), and detection algorithms.
  • TTS text-to-speech
  • an incoming message may be read aloud to a user and the user enabled to respond to the sender with a reply message through audio input.
  • Users may also be provided with options for responding in a different communication mode (e.g., a call) or perform other actions.
  • users may be enabled to initiate a message exchange using natural language.
  • FIG. 1 is a conceptual diagram illustrating networked communications between different example devices in various modalities
  • FIG. 2 illustrates an example flow of operations in a system according to embodiments for initiating a message exchange through audio input
  • FIG. 3 illustrates an example flow of operations in a system according to embodiments for responding to an incoming a message through audio input
  • FIG. 4 illustrates an example user interface of a portable computing device for facilitating communications
  • FIG. 5 is a networked environment, where a system according to
  • FIG. 6 is a block diagram of an example computing operating environment, where embodiments may be implemented.
  • an incoming message may be read aloud to a user and the user enabled to respond to the sender with a reply message through audio input upon determining whether the audio interaction mode is proper. Users may also be provided with options for responding in a different communication mode (e.g., a call) or perform other actions. Users may further be enabled to initiate a message exchange using natural language.
  • a different communication mode e.g., a call
  • Users may further be enabled to initiate a message exchange using natural language.
  • program modules include routines, programs, components, data structures, and other types of structures that perform particular tasks or implement particular abstract data types.
  • embodiments may be practiced with other computer system configurations, including hand-held devices, multiprocessor systems, microprocessor-based or programmable consumer electronics, minicomputers, mainframe computers, and comparable computing devices.
  • Embodiments may also be practiced in distributed computing environments where tasks are performed by remote processing devices that are linked through a communications network.
  • program modules may be located in both local and remote memory storage devices.
  • Embodiments may be implemented as a computer-implemented process (method), a computing system, or as an article of manufacture, such as a computer program product or computer readable media.
  • the computer program product may be a computer storage medium readable by a computer system and encoding a computer program that comprises instructions for causing a computer or computing system to perform example process(es).
  • the computer-readable storage medium can for example be implemented via one or more of a volatile computer memory, a non-volatile memory, a hard drive, a flash drive, a floppy disk, or a compact disk, and comparable media.
  • platform may be a combination of software and hardware components for facilitating multi-modal communications.
  • Examples of platforms include, but are not limited to, a hosted service executed over a plurality of servers, an application executed on a single server, and comparable systems.
  • server generally refers to a computing device executing one or more software programs typically in a networked environment. However, a server may also be implemented as a virtual server (software programs) executed on one or more computing devices viewed as a server on the network.
  • FIG. 1 is a conceptual diagram illustrating networked communications between different example devices in various modalities.
  • Modern communication systems may include exchange of information over one or more wired and/or wireless networks managed by servers and other specialized equipment.
  • User interaction may be facilitated by specialized devices such as cellular phones, smart phones, dedicated devices, or by general purpose computing devices (fixed or portable) that executed communication applications.
  • Diagram 100 illustrates two example systems, one utilizing a cellular network, the other utilizing data networks.
  • a cellular communication system enables audio, video, or text base exchanges to occur through cellular networks 102 managed by a complex backbone system.
  • Cellular phones 112 and 122 may have varying capabilities. These days, it is not uncommon for a smart phone to be very similar to a desktop computing device in terms of capabilities.
  • Data network 104 based communication systems enable exchange of a broader set of data and communication modalities through portable (e.g. handheld computers 114, 124) or stationary (e.g. desktop computers 116, 126) computing devices.
  • Data network 104 based communication systems are typically managed by one or more servers (e.g. server 106).
  • Communication sessions may also be facilitates across networks. For example, a user connected to data network 104 may initiate a
  • a communication system employs a combination of speech recognition, dictation, and text-to-speech (audio output) technologies in enabling a user to send an outgoing text-based messages and to reply to an incoming text-based message (receive notification, have the message read to them, and craft a response) without having to press any buttons or even look at the device screen, thereby rendering minimal to no interaction with the communication device.
  • Text-based messages may include any form of textual messages including, but not limited to, instant messages (IMs), short message service (SMS) messages, multimedia messaging service (MMS) messages, social networking posts/updates, emails, and comparable ones.
  • Example embodiments also include methods. These methods can be implemented in any number of ways, including the structures described in this document. One such way is by machine operations, of devices of the type described in this document.
  • Another optional way is for one or more of the individual operations of the methods to be performed in conjunction with one or more human operators performing some. These human operators need not be collocated with each other, but each can be only with a machine that performs a portion of the program.
  • FIG. 2 illustrates an example flow of operations in a system according to embodiments for initiating a message exchange through audio input.
  • An audio input to a computing device facilitating communications may come through an integrated or distinct component (wired or wireless) such as a microphone, a headset, a car kit, or similar audio devices. While a variety of sequences of operations may be performed in a
  • the example operation flow 200 may begin with activation of messaging actions through a predefined keyword (e.g. "Start Messaging") or pressing of a button on the device (232).
  • the messaging actions may be launched through natural language.
  • the user may provide an indication by uttering "Send a message to John Doe.”
  • the system may confirm that the identifier is proper and wait for further voice input.
  • the user utters a name
  • one or more determination algorithms may be executed to associate the received name with a phone number of similar identifier (e.g., a SIP identifier). For example, the received name may be compared to a contacts list or similar database.
  • the system may prompt the user to specify which contact is intended to receive the message. Furthermore, if there are multiple identifiers associated with a contact (e.g., telephone number, SIP identifier, email address, social networking address, etc.), the system may again prompt the user to select (through audio input) the intended identifier. For example, the system may automatically determine that a text message is not to be sent to a fax number of regular phone number associated with a contact, but if the contact has two cellular phone numbers, the user may be prompted to select between the two numbers. [0027] Once the intended recipient's identifier is determined, the system may prompt the user through an audio prompt or earcon to speak the message (234).
  • identifiers associated with a contact e.g., telephone number, SIP identifier, email address, social networking address, etc.
  • the system may again prompt the user to select (through audio input) the intended identifier. For example, the system may automatically determine that a text message is not to be sent to a fax
  • An earcon is a brief, distinctive sound (usually a synthesized tone or sound pattern) used to represent a specific event. Earcons are a common feature of computer operating systems, where a warning or an error message is accompanied by a distinctive tone or combination of tones.
  • the system may perform speech recognition (236). Speech recognition and/or other processing may be performed entirely or partially at the communication device. For example, in some applications, the communication device may send the recorded audio to a server, which may perform the speech recognition and provide the results to the communication device.
  • the device / application may optionally read back the message and prompt the user to edit/append/confirm that message (238).
  • the message may be transmitted as a text-based message to the recipient (240) and the user optionally provided a confirmation that the text-based message has been sent (242).
  • the user interface of the communication device / application may also provide visual feedback to the user. For example, various icons and/or text may be displayed indicating an action being performed or its result (e.g. an animated icon indicating speech recognition in process or a confirmation icon / text).
  • FIG. 3 illustrates an example flow of operations in a system according to embodiments for responding to an incoming a message through audio input.
  • the operations in diagram 300 begin with receipt of a text-based message (352).
  • the system may make a determination (354) whether audio interaction mode is available or allowed. For example, the user may turn off audio interaction mode when he/she is in a meeting or in a public place.
  • the determination may be made automatically based on a number of factors. For example, the user's calendar indicating a meeting may be used to turn off the audio interaction mode or the device being mobile (e.g. through GPS or similar location service) may prompt the system to activate the audio interaction mode. Similarly, the device's position (e.g., the device being face down) or comparable circumstances may also be used to determine whether the audio interaction mode should be used or not.
  • Audio-interactive mode may include, but are not limited to, a mobile status of the user (e.g., is the user stationary, walking, driving), an availability status of the user (as indicated in the user's calendar or similar application), and a configuration of the communication device (e.g., connected input / output devices).
  • a mobile status of the user e.g., is the user stationary, walking, driving
  • an availability status of the user as indicated in the user's calendar or similar application
  • a configuration of the communication device e.g., connected input / output devices.
  • the received text-based message may be converted to audio content through text-to-speech conversion (356) at the device or at a server, and the audio message played to the user (358).
  • the device/application may prompt the user with options (360) such as recording a response message, initiating an audio call (or video call), or performing comparable actions.
  • options such as recording a response message, initiating an audio call (or video call), or performing comparable actions.
  • the user may request that contact details of the sender be provided through audio or an earlier message in a string of messages be played back.
  • the sender's name and/or identifier (e.g. phone number) may also be played to the user at the beginning or at the end of the message.
  • the device / application may switch to a listening mode and wait for audio input from the user.
  • speech recognition may be performed (362) on the received audio input and depending on the user's response, one of a number of actions such as placing a call to the sender (364), replying to the text message (366), or other actions (368) may be performed.
  • visual cues may be displayed during the audio interaction with the user such as icons, text, color warnings, etc.
  • the interactions in operation flows 200 and 300 may be completely automated allowing the user to provide audio input through natural language or prompted (e.g. the device providing audio prompts at various stages). Moreover, physical interaction (pressing of physical or virtual buttons, text prompts, etc.) may also be employed at different stages of the interaction. Furthermore, users may be provided with the option of editing outgoing messages upon recording of those (following optional playback).
  • Audio-interactive message exchange may be implemented by similar processes with fewer or additional steps, as well as in different order of operations using the principles described herein.
  • FIG. 4 illustrates an example user interface of a portable computing device for facilitating communications.
  • audio interaction for text messaging may be implemented in any device facilitating communications.
  • the user interface illustrated in diagram 300 is just an example user interface of a mobile communication device. Embodiments are not limited to this example user interface or others discussed above.
  • An example mobile communication device may include a speaker 472 and a microphone in addition to a number of physical control elements such as buttons, knobs, keys, etc. Such a device may also include a camera 474 or similar ancillary devices that may be used in conjunction with different communication modalities.
  • the example user interface displays date and time and a number of icons for different applications such as phone application 476, messaging application 478, camera application 480, file organization application 482, and web browser 484.
  • the user interface may further include a number of virtual buttons (not shown) such as Dual Tone Multi-frequency (DTMF) keys for placing a call.
  • DTMF Dual Tone Multi-frequency
  • a picture (or representative icon) 486 of the sender of the received message may be displayed along with a textual clue about the message 488 and additional icons 490 (e.g. indicating message category, sender's presence status, etc.)
  • the user interface of the communication device / application may also provide visual feedback to the user. For example, additional icons and/or text may be displayed indicating an action being performed or its result (e.g. an animated icon indicating speech recognition in process or a confirmation icon / text).
  • additional icons and/or text may be displayed indicating an action being performed or its result (e.g. an animated icon indicating speech recognition in process or a confirmation icon / text).
  • the communication device may also be equipped to determine whether the audio interaction mode should / can be used or not.
  • a location and / or motion determination system may detect whether the user is moving (e.g. in a car) based on Global Positioning Service (GPS) information, cellular tower triangulation, wireless data network node detection, compass, and acceleration sensors, matching of camera input to known geo-position photos, and similar methods.
  • GPS Global Positioning Service
  • Another approach may include determining the user's location (e.g. a meeting room or a public space) and activating the audio interaction based on that.
  • information about the user such as from a calendaring application or a currently executed application may be used to determine the user's availability for audio interaction.
  • the communication employing audio interaction may be facilitated through any computing device such as desktop computers, laptop computers, notebooks; mobile devices such as smart phones, handheld computers, wireless Personal Digital Assistants (PDAs), cellular phones, vehicle mount computing devices, and similar ones.
  • PDAs Personal Digital Assistants
  • FIG. 1 through 4 may be implemented using distinct hardware modules, software modules, or combinations of hardware and software. Furthermore, such modules may perform two or more of the processes in an integrated manner. While some embodiments have been provided with specific examples for audio-interactive message exchange, embodiments are not limited to those. Indeed, embodiments may be implemented in various communication systems using a variety of communication devices and applications and with additional or fewer features using the principles described herein.
  • FIG. 5 is an example networked environment, where embodiments may be implemented.
  • a platform for providing communication services with audio-interactive message exchange may be implemented via software executed over one or more servers 514 such as a hosted service.
  • the platform may communicate with client applications on individual mobile devices such as a smart phone 511, cellular phone 512, or similar devices ('client devices') through network(s) 510.
  • Client applications executed on any of the client devices 511-512 may interact with a hosted service providing communication services from the servers 514, or on individual server 516.
  • the hosted service may provide multi-modal communication services and ancillary services such as presence, location, etc.
  • text message exchange may be facilitated between users with audio-interactivity as described above.
  • Some or all of the processing associated with the audio-interactivity such as speech recognition or text-to-speech conversion may be performed at one of more of the servers 514 or 516.
  • Relevant data such as speech recognition, text-to-speech conversion, contact information, and similar data may be stored and / or retrieved at/from data store(s) 519 directly or through database server 518.
  • Network(s) 510 may comprise any topology of servers, clients, Internet service providers, and communication media.
  • a system according to embodiments may have a static or dynamic topology.
  • Network(s) 510 may include secure networks such as an enterprise network, an unsecure network such as a wireless open network, or the Internet.
  • Network(s) 510 may also include (especially between the servers and the mobile devices) cellular networks.
  • network(s) 510 may include short range wireless networks such as Bluetooth or similar ones.
  • Network(s) 510 provide communication between the nodes described herein.
  • network(s) 510 may include wireless media such as acoustic, RF, infrared and other wireless media.
  • FIG. 6 and the associated discussion are intended to provide a brief, general description of a suitable computing environment in which embodiments may be implemented.
  • computing device 600 may be a mobile computing device capable of facilitating multi-modal communication including text message exchange with audio interactivity according to embodiments and include at least one processing unit 602 and system memory 604.
  • Computing device 600 may also include a plurality of processing units that cooperate in executing programs.
  • the system memory 604 may be volatile (such as RAM), non-volatile (such as ROM, flash memory, etc.) or some combination of the two.
  • System memory 604 typically includes an operating system 605 suitable for controlling the operation of the platform, such as the WINDOWS MOBILE®, WINDOWS PHONE®, or similar operating systems from MICROSOFT
  • the system memory 604 may also include one or more software applications such as program modules 606, communication application 622, and audio interactivity module 624.
  • Communication application 622 may enable multi-modal communications including text messaging.
  • Audio interactivity module 624 may play an incoming message to a user and enable the user to respond to the sender with a reply message through audio input through a combination of speech recognition, text-to-speech (TTS), and detection algorithms.
  • Communication application 622 may also provide users with options for responding in a different communication mode (e.g., a call) or for performing other actions.
  • Audio interactivity module 624 may further enable users to initiate a message exchange using natural language. This basic configuration is illustrated in FIG. 6 by those components within dashed line 608.
  • Computing device 600 may have additional features or functionality.
  • the computing device 600 may also include additional data storage devices (removable and/or non-removable) such as, for example, magnetic disks, optical disks, or tape.
  • additional storage is illustrated in FIG. 6 by removable storage 609 and nonremovable storage 610.
  • Computer readable storage media may include volatile and nonvolatile, removable and non-removable media implemented in any method or technology for storage of information, such as computer readable instructions, data structures, program modules, or other data.
  • System memory 604, removable storage 609 and non-removable storage 610 are all examples of computer readable storage media.
  • Computer readable storage media includes, but is not limited to, RAM, ROM, EEPROM, flash memory or other memory technology, CD-ROM, digital versatile disks (DVD) or other optical storage, magnetic tape, magnetic disk storage or other magnetic storage devices, or any other medium which can be used to store the desired information and which can be accessed by computing device 600. Any such computer readable storage media may be part of computing device 600.
  • Computing device 600 may also have input device(s) 612 such as keyboard, mouse, pen, voice input device, touch input device, and comparable input devices.
  • Output device(s) 614 such as a display, speakers, printer, and other types of output devices may also be included. These devices are well known in the art and need not be discussed at length here.
  • Computing device 600 may also contain communication connections 616 that allow the device to communicate with other devices 618, such as over a wired or wireless network in a distributed computing environment, a satellite link, a cellular link, a short range network, and comparable mechanisms.
  • Other devices 618 may include computer device(s) that execute communication applications, other servers, and comparable devices.
  • Communication connection(s) 616 is one example of communication media.
  • Communication media can include therein computer readable instructions, data structures, program modules, or other data.
  • communication media includes wired media such as a wired network or direct- wired connection, and wireless media such as acoustic, RF, infrared and other wireless media.

Abstract

A completely hands free exchange of messages, especially in portable devices, is provided through a combination of speech recognition, text-to-speech (TTS), and detection algorithms. An incoming message may be read aloud to a user and the user enabled to respond to the sender with a reply message through audio input upon determining whether the audio interaction mode is proper. Users may also be provided with options for responding in a different communication mode (e.g., a call) or perform other actions. Users may further be enabled to initiate a message exchange using natural language.

Description

AUDIO-INTERACTIVE MESSAGE EXCHANGE
BACKGROUND
[0001] With the development and wide use of computing and networking
technologies, personal and business communications have proliferated in quantity and quality. Multi-modal communications through fixed or portable computing devices such as desktop computers, vehicle mount computers, portable computers, smart phones, and similar devices are a common occurrence. Because many facets of communications are controlled through easily customizable software / hardware combinations, previously unheard-of features are available for use in daily life. For example, integration of presence information into communication applications enables people to communicate with each other more efficiently. Simultaneous reduction in size and increase in computing capabilities enables use of smart phones or similar handheld computing devices for multi-modal communications including, but not limited to, audio, video, text message exchange, email, instant messaging, social networking posts/updates, etc.
[0002] One of the results of the proliferation of communication technologies is the information overload. It is not unusual for a person to exchange hundreds of emails, participate in numerous audio or video communication sessions, and exchange a high number of text messages every day. Given the expansive range of communications, text message exchange is increasingly becoming more popular in place of more formal emails and time consuming audio / video communications. Still, using conventional typing technologies - whether on physical keyboards or using touch technologies - even text messaging may be inefficient, impractical, or dangerous in some cases (e.g., while driving).
SUMMARY
[0003] This summary is provided to introduce a selection of concepts in a simplified form that are further described below in the Detailed Description. This summary is not intended to exclusively identify key features or essential features of the claimed subject matter, nor is it intended as an aid in determining the scope of the claimed subject matter.
[0004] Embodiments are directed to providing a completely hands free exchange of messages, especially in portable devices through a combination of speech recognition, text-to-speech (TTS), and detection algorithms. According to some embodiments, an incoming message may be read aloud to a user and the user enabled to respond to the sender with a reply message through audio input. Users may also be provided with options for responding in a different communication mode (e.g., a call) or perform other actions. According to other embodiments, users may be enabled to initiate a message exchange using natural language.
[0005] These and other features and advantages will be apparent from a reading of the following detailed description and a review of the associated drawings. It is to be understood that both the foregoing general description and the following detailed description are explanatory and do not restrict aspects as claimed.
BRIEF DESCRIPTION OF THE DRAWINGS
[0006] FIG. 1 is a conceptual diagram illustrating networked communications between different example devices in various modalities;
[0007] FIG. 2 illustrates an example flow of operations in a system according to embodiments for initiating a message exchange through audio input;
[0008] FIG. 3 illustrates an example flow of operations in a system according to embodiments for responding to an incoming a message through audio input;
[0009] FIG. 4 illustrates an example user interface of a portable computing device for facilitating communications;
[0010] FIG. 5 is a networked environment, where a system according to
embodiments may be implemented; and
[0011] FIG. 6 is a block diagram of an example computing operating environment, where embodiments may be implemented.
DETAILED DESCRIPTION
[0012] As briefly described above, an incoming message may be read aloud to a user and the user enabled to respond to the sender with a reply message through audio input upon determining whether the audio interaction mode is proper. Users may also be provided with options for responding in a different communication mode (e.g., a call) or perform other actions. Users may further be enabled to initiate a message exchange using natural language. In the following detailed description, references are made to the accompanying drawings that form a part hereof, and in which are shown by way of illustrations specific embodiments or examples. These aspects may be combined, other aspects may be utilized, and structural changes may be made without departing from the spirit or scope of the present disclosure. The following detailed description is therefore not to be taken in a limiting sense, and the scope of the present invention is defined by the appended claims and their equivalents.
[0013] While the embodiments will be described in the general context of program modules that execute in conjunction with an application program that runs on an operating system on a personal computer, those skilled in the art will recognize that aspects may also be implemented in combination with other program modules.
[0014] Generally, program modules include routines, programs, components, data structures, and other types of structures that perform particular tasks or implement particular abstract data types. Moreover, those skilled in the art will appreciate that embodiments may be practiced with other computer system configurations, including hand-held devices, multiprocessor systems, microprocessor-based or programmable consumer electronics, minicomputers, mainframe computers, and comparable computing devices. Embodiments may also be practiced in distributed computing environments where tasks are performed by remote processing devices that are linked through a communications network. In a distributed computing environment, program modules may be located in both local and remote memory storage devices.
[0015] Embodiments may be implemented as a computer-implemented process (method), a computing system, or as an article of manufacture, such as a computer program product or computer readable media. The computer program product may be a computer storage medium readable by a computer system and encoding a computer program that comprises instructions for causing a computer or computing system to perform example process(es). The computer-readable storage medium can for example be implemented via one or more of a volatile computer memory, a non-volatile memory, a hard drive, a flash drive, a floppy disk, or a compact disk, and comparable media.
[0016] Throughout this specification, the term "platform" may be a combination of software and hardware components for facilitating multi-modal communications.
Examples of platforms include, but are not limited to, a hosted service executed over a plurality of servers, an application executed on a single server, and comparable systems. The term "server" generally refers to a computing device executing one or more software programs typically in a networked environment. However, a server may also be implemented as a virtual server (software programs) executed on one or more computing devices viewed as a server on the network.
[0017] FIG. 1 is a conceptual diagram illustrating networked communications between different example devices in various modalities. Modern communication systems may include exchange of information over one or more wired and/or wireless networks managed by servers and other specialized equipment. User interaction may be facilitated by specialized devices such as cellular phones, smart phones, dedicated devices, or by general purpose computing devices (fixed or portable) that executed communication applications.
[0018] The diversity in capabilities and features offered by modern communication systems enables users to take advantage of a variety of communication modalities. For example, audio, video, email, text message, data sharing, application sharing, and similar modalities can be used individually or in combination through the same device. A user may exchange text messages through their portable device and then continue a conversation with the same person over a different modality.
[0019] Diagram 100 illustrates two example systems, one utilizing a cellular network, the other utilizing data networks. A cellular communication system enables audio, video, or text base exchanges to occur through cellular networks 102 managed by a complex backbone system. Cellular phones 112 and 122 may have varying capabilities. These days, it is not uncommon for a smart phone to be very similar to a desktop computing device in terms of capabilities.
[0020] Data network 104 based communication systems on the other hand enable exchange of a broader set of data and communication modalities through portable (e.g. handheld computers 114, 124) or stationary (e.g. desktop computers 116, 126) computing devices. Data network 104 based communication systems are typically managed by one or more servers (e.g. server 106). Communication sessions may also be facilitates across networks. For example, a user connected to data network 104 may initiate a
communication session (in any modality) through their desktop communication application with a cellular phone user connected to cellular network 102.
[0021] Conventional systems and communication devices are, however, mostly limited to physical interaction such as typing or activation of buttons or similar control elements on the communication device. While speech recognition based technologies are in use in some systems, the users typically have to activate those by pressing a button. Furthermore, the user has to place the device / application in the proper mode before using the speech-based features.
[0022] A communication system according to some embodiments employs a combination of speech recognition, dictation, and text-to-speech (audio output) technologies in enabling a user to send an outgoing text-based messages and to reply to an incoming text-based message (receive notification, have the message read to them, and craft a response) without having to press any buttons or even look at the device screen, thereby rendering minimal to no interaction with the communication device. Text-based messages may include any form of textual messages including, but not limited to, instant messages (IMs), short message service (SMS) messages, multimedia messaging service (MMS) messages, social networking posts/updates, emails, and comparable ones.
[0023] Example embodiments also include methods. These methods can be implemented in any number of ways, including the structures described in this document. One such way is by machine operations, of devices of the type described in this document.
[0024] Another optional way is for one or more of the individual operations of the methods to be performed in conjunction with one or more human operators performing some. These human operators need not be collocated with each other, but each can be only with a machine that performs a portion of the program.
[0025] FIG. 2 illustrates an example flow of operations in a system according to embodiments for initiating a message exchange through audio input. An audio input to a computing device facilitating communications may come through an integrated or distinct component (wired or wireless) such as a microphone, a headset, a car kit, or similar audio devices. While a variety of sequences of operations may be performed in a
communication system according to embodiments, two example flows are discussed in FIG. 2 and FIG. 3.
[0026] The example operation flow 200 may begin with activation of messaging actions through a predefined keyword (e.g. "Start Messaging") or pressing of a button on the device (232). According to some embodiments, the messaging actions may be launched through natural language. For example, the user may provide an indication by uttering "Send a message to John Doe." If the user utters a phone number or similar identifier as recipient, the system may confirm that the identifier is proper and wait for further voice input. If the user utters a name, one or more determination algorithms may be executed to associate the received name with a phone number of similar identifier (e.g., a SIP identifier). For example, the received name may be compared to a contacts list or similar database. If there are multiple names or similar sounding names, the system may prompt the user to specify which contact is intended to receive the message. Furthermore, if there are multiple identifiers associated with a contact (e.g., telephone number, SIP identifier, email address, social networking address, etc.), the system may again prompt the user to select (through audio input) the intended identifier. For example, the system may automatically determine that a text message is not to be sent to a fax number of regular phone number associated with a contact, but if the contact has two cellular phone numbers, the user may be prompted to select between the two numbers. [0027] Once the intended recipient's identifier is determined, the system may prompt the user through an audio prompt or earcon to speak the message (234). An earcon is a brief, distinctive sound (usually a synthesized tone or sound pattern) used to represent a specific event. Earcons are a common feature of computer operating systems, where a warning or an error message is accompanied by a distinctive tone or combination of tones. When the user is done speaking the message (determined either by a duration of silence at the end exceeding a predefined time interval or user audio prompt such as "end of message"), the system may perform speech recognition (236). Speech recognition and/or other processing may be performed entirely or partially at the communication device. For example, in some applications, the communication device may send the recorded audio to a server, which may perform the speech recognition and provide the results to the communication device.
[0028] Upon conclusion of the speech recognition process, the device / application may optionally read back the message and prompt the user to edit/append/confirm that message (238). Upon confirmation, the message may be transmitted as a text-based message to the recipient (240) and the user optionally provided a confirmation that the text-based message has been sent (242). At different stages of the processing, the user interface of the communication device / application may also provide visual feedback to the user. For example, various icons and/or text may be displayed indicating an action being performed or its result (e.g. an animated icon indicating speech recognition in process or a confirmation icon / text).
[0029] FIG. 3 illustrates an example flow of operations in a system according to embodiments for responding to an incoming a message through audio input.
[0030] The operations in diagram 300 begin with receipt of a text-based message (352). Next, the system may make a determination (354) whether audio interaction mode is available or allowed. For example, the user may turn off audio interaction mode when he/she is in a meeting or in a public place. According to some embodiments, the determination may be made automatically based on a number of factors. For example, the user's calendar indicating a meeting may be used to turn off the audio interaction mode or the device being mobile (e.g. through GPS or similar location service) may prompt the system to activate the audio interaction mode. Similarly, the device's position (e.g., the device being face down) or comparable circumstances may also be used to determine whether the audio interaction mode should be used or not. Further factors in determining audio-interactive mode may include, but are not limited to, a mobile status of the user (e.g., is the user stationary, walking, driving), an availability status of the user (as indicated in the user's calendar or similar application), and a configuration of the communication device (e.g., connected input / output devices).
[0031] If the audio interaction mode is allowed / available, the received text-based message may be converted to audio content through text-to-speech conversion (356) at the device or at a server, and the audio message played to the user (358). Upon completion of the playing of the message, the device/application may prompt the user with options (360) such as recording a response message, initiating an audio call (or video call), or performing comparable actions. For example, the user may request that contact details of the sender be provided through audio or an earlier message in a string of messages be played back. The sender's name and/or identifier (e.g. phone number) may also be played to the user at the beginning or at the end of the message.
[0032] Upon playing the options to the user, the device / application may switch to a listening mode and wait for audio input from the user. When the user's response is received, speech recognition may be performed (362) on the received audio input and depending on the user's response, one of a number of actions such as placing a call to the sender (364), replying to the text message (366), or other actions (368) may be performed. Similar to the flow of operations in FIG. 2, visual cues may be displayed during the audio interaction with the user such as icons, text, color warnings, etc.
[0033] The interactions in operation flows 200 and 300 may be completely automated allowing the user to provide audio input through natural language or prompted (e.g. the device providing audio prompts at various stages). Moreover, physical interaction (pressing of physical or virtual buttons, text prompts, etc.) may also be employed at different stages of the interaction. Furthermore, users may be provided with the option of editing outgoing messages upon recording of those (following optional playback).
[0034] The operations included in processes 200 and 300 are for illustration purposes. Audio-interactive message exchange may be implemented by similar processes with fewer or additional steps, as well as in different order of operations using the principles described herein.
[0035] FIG. 4 illustrates an example user interface of a portable computing device for facilitating communications. As discussed above, audio interaction for text messaging may be implemented in any device facilitating communications. The user interface illustrated in diagram 300 is just an example user interface of a mobile communication device. Embodiments are not limited to this example user interface or others discussed above.
[0036] An example mobile communication device may include a speaker 472 and a microphone in addition to a number of physical control elements such as buttons, knobs, keys, etc. Such a device may also include a camera 474 or similar ancillary devices that may be used in conjunction with different communication modalities. The example user interface displays date and time and a number of icons for different applications such as phone application 476, messaging application 478, camera application 480, file organization application 482, and web browser 484. The user interface may further include a number of virtual buttons (not shown) such as Dual Tone Multi-frequency (DTMF) keys for placing a call.
[0037] At the bottom portion of the example user interface icons and text associated with a messaging application are shown. For example, a picture (or representative icon) 486 of the sender of the received message may be displayed along with a textual clue about the message 488 and additional icons 490 (e.g. indicating message category, sender's presence status, etc.)
[0038] At different stages of the processing, the user interface of the communication device / application may also provide visual feedback to the user. For example, additional icons and/or text may be displayed indicating an action being performed or its result (e.g. an animated icon indicating speech recognition in process or a confirmation icon / text).
[0039] The communication device may also be equipped to determine whether the audio interaction mode should / can be used or not. As discussed above, a location and / or motion determination system may detect whether the user is moving (e.g. in a car) based on Global Positioning Service (GPS) information, cellular tower triangulation, wireless data network node detection, compass, and acceleration sensors, matching of camera input to known geo-position photos, and similar methods. Another approach may include determining the user's location (e.g. a meeting room or a public space) and activating the audio interaction based on that. Similarly, information about the user such as from a calendaring application or a currently executed application may be used to determine the user's availability for audio interaction.
[0040] The communication employing audio interaction may be facilitated through any computing device such as desktop computers, laptop computers, notebooks; mobile devices such as smart phones, handheld computers, wireless Personal Digital Assistants (PDAs), cellular phones, vehicle mount computing devices, and similar ones. [0041] The different processes and systems discussed in FIG. 1 through 4 may be implemented using distinct hardware modules, software modules, or combinations of hardware and software. Furthermore, such modules may perform two or more of the processes in an integrated manner. While some embodiments have been provided with specific examples for audio-interactive message exchange, embodiments are not limited to those. Indeed, embodiments may be implemented in various communication systems using a variety of communication devices and applications and with additional or fewer features using the principles described herein.
[0042] FIG. 5 is an example networked environment, where embodiments may be implemented. A platform for providing communication services with audio-interactive message exchange may be implemented via software executed over one or more servers 514 such as a hosted service. The platform may communicate with client applications on individual mobile devices such as a smart phone 511, cellular phone 512, or similar devices ('client devices') through network(s) 510.
[0043] Client applications executed on any of the client devices 511-512 may interact with a hosted service providing communication services from the servers 514, or on individual server 516. The hosted service may provide multi-modal communication services and ancillary services such as presence, location, etc. As part of the multi-modal services, text message exchange may be facilitated between users with audio-interactivity as described above. Some or all of the processing associated with the audio-interactivity such as speech recognition or text-to-speech conversion may be performed at one of more of the servers 514 or 516. Relevant data such as speech recognition, text-to-speech conversion, contact information, and similar data may be stored and / or retrieved at/from data store(s) 519 directly or through database server 518.
[0044] Network(s) 510 may comprise any topology of servers, clients, Internet service providers, and communication media. A system according to embodiments may have a static or dynamic topology. Network(s) 510 may include secure networks such as an enterprise network, an unsecure network such as a wireless open network, or the Internet. Network(s) 510 may also include (especially between the servers and the mobile devices) cellular networks. Furthermore, network(s) 510 may include short range wireless networks such as Bluetooth or similar ones. Network(s) 510 provide communication between the nodes described herein. By way of example, and not limitation, network(s) 510 may include wireless media such as acoustic, RF, infrared and other wireless media. [0045] Many other configurations of computing devices, applications, data sources, and data distribution systems may be employed to implement a platform providing audio- interactive message exchange services. Furthermore, the networked environments discussed in FIG. 5 are for illustration purposes only. Embodiments are not limited to the example applications, modules, or processes.
[0046] FIG. 6 and the associated discussion are intended to provide a brief, general description of a suitable computing environment in which embodiments may be implemented. With reference to FIG. 6, a block diagram of an example computing operating environment for an application according to embodiments is illustrated, such as computing device 600. In a basic configuration, computing device 600 may be a mobile computing device capable of facilitating multi-modal communication including text message exchange with audio interactivity according to embodiments and include at least one processing unit 602 and system memory 604. Computing device 600 may also include a plurality of processing units that cooperate in executing programs. Depending on the exact configuration and type of computing device, the system memory 604 may be volatile (such as RAM), non-volatile (such as ROM, flash memory, etc.) or some combination of the two. System memory 604 typically includes an operating system 605 suitable for controlling the operation of the platform, such as the WINDOWS MOBILE®, WINDOWS PHONE®, or similar operating systems from MICROSOFT
CORPORATION of Redmond, Washington or similar ones. The system memory 604 may also include one or more software applications such as program modules 606, communication application 622, and audio interactivity module 624.
[0047] Communication application 622 may enable multi-modal communications including text messaging. Audio interactivity module 624 may play an incoming message to a user and enable the user to respond to the sender with a reply message through audio input through a combination of speech recognition, text-to-speech (TTS), and detection algorithms. Communication application 622 may also provide users with options for responding in a different communication mode (e.g., a call) or for performing other actions. Audio interactivity module 624 may further enable users to initiate a message exchange using natural language. This basic configuration is illustrated in FIG. 6 by those components within dashed line 608.
[0048] Computing device 600 may have additional features or functionality. For example, the computing device 600 may also include additional data storage devices (removable and/or non-removable) such as, for example, magnetic disks, optical disks, or tape. Such additional storage is illustrated in FIG. 6 by removable storage 609 and nonremovable storage 610. Computer readable storage media may include volatile and nonvolatile, removable and non-removable media implemented in any method or technology for storage of information, such as computer readable instructions, data structures, program modules, or other data. System memory 604, removable storage 609 and non-removable storage 610 are all examples of computer readable storage media. Computer readable storage media includes, but is not limited to, RAM, ROM, EEPROM, flash memory or other memory technology, CD-ROM, digital versatile disks (DVD) or other optical storage, magnetic tape, magnetic disk storage or other magnetic storage devices, or any other medium which can be used to store the desired information and which can be accessed by computing device 600. Any such computer readable storage media may be part of computing device 600. Computing device 600 may also have input device(s) 612 such as keyboard, mouse, pen, voice input device, touch input device, and comparable input devices. Output device(s) 614 such as a display, speakers, printer, and other types of output devices may also be included. These devices are well known in the art and need not be discussed at length here.
[0049] Computing device 600 may also contain communication connections 616 that allow the device to communicate with other devices 618, such as over a wired or wireless network in a distributed computing environment, a satellite link, a cellular link, a short range network, and comparable mechanisms. Other devices 618 may include computer device(s) that execute communication applications, other servers, and comparable devices. Communication connection(s) 616 is one example of communication media.
Communication media can include therein computer readable instructions, data structures, program modules, or other data. By way of example, and not limitation, communication media includes wired media such as a wired network or direct- wired connection, and wireless media such as acoustic, RF, infrared and other wireless media.
[0050] The above specification, examples and data provide a complete description of the manufacture and use of the composition of the embodiments. Although the subject matter has been described in language specific to structural features and/or methodological acts, it is to be understood that the subject matter defined in the appended claims is not necessarily limited to the specific features or acts described above. Rather, the specific features and acts described above are disclosed as example forms of implementing the claims and embodiments.

Claims

CLAIMS WHAT IS CLAIMED IS:
1. A method executed at least in part in a computing device for facilitating audio- interactive message exchange, the method comprising:
receiving an indication from a user to send a message;
enabling the user to provide a recipient of the message and an audio content of the message through audio input;
performing speech recognition on the received audio input; determining the recipient from the speech recognized audio input; and transmitting the speech recognized content of the message to the recipient as a text-based message.
2. The method of claim 1, further comprising:
receiving a text-based message from a sender;
generating an audio content from the received message by text-to-speech conversion;
playing the audio content to the user;
providing at least one option to the user associated with the played audio content; and
in response to receiving another audio input from the user, performing an action associated with the at least one option.
3. The method of claim 2, further comprising:
enabling the user to provide the indication to send the text-based message and the audio inputs using natural language.
4. The method of claim 2, further comprising:
upon receiving the audio inputs, playing back the received audio inputs; and
enabling the user to one of: edit the provided audio input and confirm the provided audio input.
5. The method of claim 2, wherein the action includes one from a set of: initiating an audio communication session with the sender, initiating a video communication session with the sender, replying with a text-based message, playing back a previous message, and providing information associated with the sender.
6. A computing device capable of facilitating audio-interactive message exchange, the computing device comprising: a communication module;
an audio input/output module;
a memory; and
a processor coupled to the communication module, the audio input/output module, and the memory adapted to execute a communication application that is configured to:
receive a text-based message from a sender;
generate an audio content from the received message by text-to- speech conversion;
play the audio content and one of a name and an identifier associated with the sender to the user;
provide at least one option to the user associated with the played
audio content; and
in response to receiving an audio input from the user, perform an action associated with the at least one option.
The computing device of claim 6, wherein the communication application is further configured to:
receive an audio indication from the user to send a text-based
message;
enable the user to provide a recipient of the text-based message and an audio content of the message through natural language input; perform speech recognition on the received input;
enable the user to one of: confirm and edit the message by playing back the received input;
determine the recipient from the speech recognized content of the input; and
transmit the speech recognized content of the text-based message to the recipient.
The computing device of claim 6, further comprising a display, wherein the communication application is further configured to provide a visual feedback to the user through the display including at least one of a text, a graphic, an animated graphic, and an icon representing an operation associated with the audio-interactive message exchange. A computer-readable storage medium with instructions stored thereon for facilitating audio-interactive message exchange, the instructions comprising: activating an audio interaction mode automatically based on at least one from a set of: a setting of a communication device facilitating the message exchange, a location of a user, a status of the user, and a user input;
receiving an audio indication from the user to send a text-based message; enabling the user to provide a recipient of the text-based message and an audio content of the message through natural language input; performing speech recognition on the received input;
determining the recipient from the speech recognized content of the input; transmitting the speech recognized content of the message to the recipient as a text-based message;
receiving a text-based message from a sender;
generating an audio content from the received message by text-to-speech conversion;
playing the audio content to the user;
providing at least one option to the user associated with the played audio content; and
in response to receiving another audio input from the user, performing an action associated with the other audio input.
The computer-readable medium of claim 9, wherein the status of the user includes at least one from a set of: a mobile status of the user, an availability status of the user, a position of the communication device, and a configuration of the communication device.
PCT/US2012/031778 2011-04-07 2012-04-02 Audio-interactive message exchange WO2012138587A2 (en)

Priority Applications (4)

Application Number Priority Date Filing Date Title
JP2014503705A JP2014512049A (en) 2011-04-07 2012-04-02 Voice interactive message exchange
EP20120768271 EP2695406A4 (en) 2011-04-07 2012-04-02 Audio-interactive message exchange
CN2012800164763A CN103443852A (en) 2011-04-07 2012-04-02 Audio-interactive message exchange
KR1020137026109A KR20140022824A (en) 2011-04-07 2012-04-02 Audio-interactive message exchange

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
US13/081,679 US20120259633A1 (en) 2011-04-07 2011-04-07 Audio-interactive message exchange
US13/081,679 2011-04-07

Publications (2)

Publication Number Publication Date
WO2012138587A2 true WO2012138587A2 (en) 2012-10-11
WO2012138587A3 WO2012138587A3 (en) 2012-11-29

Family

ID=46966786

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/US2012/031778 WO2012138587A2 (en) 2011-04-07 2012-04-02 Audio-interactive message exchange

Country Status (6)

Country Link
US (1) US20120259633A1 (en)
EP (1) EP2695406A4 (en)
JP (1) JP2014512049A (en)
KR (1) KR20140022824A (en)
CN (1) CN103443852A (en)
WO (1) WO2012138587A2 (en)

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US9653077B2 (en) 2012-11-16 2017-05-16 Honda Motor Co., Ltd. Message processing device

Families Citing this family (41)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20170169700A9 (en) * 2005-09-01 2017-06-15 Simplexgrinnell Lp System and method for emergency message preview and transmission
US8938677B2 (en) 2009-03-30 2015-01-20 Avaya Inc. System and method for mode-neutral communications with a widget-based communications metaphor
US9788349B2 (en) 2011-09-28 2017-10-10 Elwha Llc Multi-modality communication auto-activation
US20130079029A1 (en) * 2011-09-28 2013-03-28 Royce A. Levien Multi-modality communication network auto-activation
US9794209B2 (en) 2011-09-28 2017-10-17 Elwha Llc User interface for multi-modality communication
US9699632B2 (en) 2011-09-28 2017-07-04 Elwha Llc Multi-modality communication with interceptive conversion
US9906927B2 (en) 2011-09-28 2018-02-27 Elwha Llc Multi-modality communication initiation
US9204267B2 (en) * 2012-01-04 2015-12-01 Truvu Mobile, Llc Method and system for controlling mobile communication device interactions
US9961249B2 (en) * 2012-09-17 2018-05-01 Gregory Thomas Joao Apparatus and method for providing a wireless, portable, and/or handheld, device with safety features
CN103455530A (en) * 2012-10-25 2013-12-18 河南省佰腾电子科技有限公司 Portable-type device for creating textual word databases corresponding to personized voices
EP2926271A4 (en) * 2012-11-30 2016-07-06 Nokia Technologies Oy A method and a technical equipment for analysing message content
CN103001859B (en) * 2012-12-14 2016-06-29 上海量明科技发展有限公司 The method and system of stream of reply media information in instant messaging
US9271111B2 (en) * 2012-12-14 2016-02-23 Amazon Technologies, Inc. Response endpoint selection
CN103001858B (en) * 2012-12-14 2015-09-09 上海量明科技发展有限公司 The method of message, client and system is replied in instant messaging
JP6423673B2 (en) * 2014-09-26 2018-11-14 京セラ株式会社 Communication terminal and control method thereof
US10235996B2 (en) * 2014-10-01 2019-03-19 XBrain, Inc. Voice and connection platform
US20160157074A1 (en) 2014-11-30 2016-06-02 Raymond Anthony Joao Personal monitoring apparatus and method
CN112152906B (en) * 2015-02-16 2023-04-07 钉钉控股(开曼)有限公司 Communication method and server
CN104869497B (en) * 2015-03-24 2018-12-11 广东欧珀移动通信有限公司 A kind of the wireless network setting method and device of WIFI speaker
US9430949B1 (en) * 2015-03-25 2016-08-30 Honeywell International Inc. Verbal taxi clearance system
CN105427856B (en) * 2016-01-12 2020-05-19 北京光年无限科技有限公司 Appointment data processing method and system for intelligent robot
US9912800B2 (en) 2016-05-27 2018-03-06 International Business Machines Corporation Confidentiality-smart voice delivery of text-based incoming messages
ES2644887B1 (en) * 2016-05-31 2018-09-07 Xesol I Mas D Mas I, S.L. INTERACTION METHOD BY VOICE FOR COMMUNICATION DURING VEHICLE DRIVING AND DEVICE THAT IMPLEMENTS IT
CN106230698A (en) * 2016-08-07 2016-12-14 深圳市小马立行科技有限公司 A kind of social contact method based on vehicle intelligent terminal
US10580404B2 (en) 2016-09-01 2020-03-03 Amazon Technologies, Inc. Indicator for voice-based communications
US10074369B2 (en) 2016-09-01 2018-09-11 Amazon Technologies, Inc. Voice-based communications
US10453449B2 (en) 2016-09-01 2019-10-22 Amazon Technologies, Inc. Indicator for voice-based communications
KR20190032557A (en) * 2016-09-01 2019-03-27 아마존 테크놀로지스, 인크. Voice-based communication
US20180088969A1 (en) * 2016-09-28 2018-03-29 Lenovo (Singapore) Pte. Ltd. Method and device for presenting instructional content
CN106791015A (en) * 2016-11-29 2017-05-31 维沃移动通信有限公司 A kind of message is played and answering method and device
CN106601254B (en) * 2016-12-08 2020-11-06 阿里巴巴(中国)有限公司 Information input method and device and computing equipment
KR20180101063A (en) * 2017-03-03 2018-09-12 삼성전자주식회사 Electronic apparatus for processing user input and method for processing user input
CN109725798B (en) * 2017-10-25 2021-07-27 腾讯科技(北京)有限公司 Intelligent role switching method and related device
CN107734193A (en) * 2017-11-22 2018-02-23 深圳悉罗机器人有限公司 Smart machine system and smart machine control method
CN110048928B (en) * 2018-01-17 2022-07-05 阿里巴巴集团控股有限公司 Information submitting, obtaining and interacting method, device, equipment and system
KR102508677B1 (en) 2018-03-08 2023-03-13 삼성전자주식회사 System for processing user utterance and controlling method thereof
US10891939B2 (en) * 2018-11-26 2021-01-12 International Business Machines Corporation Sharing confidential information with privacy using a mobile phone
CN110211589B (en) * 2019-06-05 2022-03-15 广州小鹏汽车科技有限公司 Awakening method and device of vehicle-mounted system, vehicle and machine readable medium
US11765547B2 (en) 2019-07-30 2023-09-19 Raymond Anthony Joao Personal monitoring apparatus and methods
US11775780B2 (en) 2021-03-01 2023-10-03 Raymond Anthony Joao Personal monitoring apparatus and methods
CN114007130A (en) * 2021-10-29 2022-02-01 维沃移动通信有限公司 Data transmission method and device, electronic equipment and storage medium

Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2006133547A1 (en) 2005-06-13 2006-12-21 E-Lane Systems Inc. Vehicle immersive communication system
EP1879000A1 (en) 2006-07-10 2008-01-16 Harman Becker Automotive Systems GmbH Transmission of text messages by navigation systems
EP2224705A1 (en) 2009-02-27 2010-09-01 Research In Motion Limited Mobile wireless communications device with speech to text conversion and related method
US20100222086A1 (en) 2009-02-28 2010-09-02 Karl Schmidt Cellular Phone and other Devices/Hands Free Text Messaging

Family Cites Families (33)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US5475738A (en) * 1993-10-21 1995-12-12 At&T Corp. Interface between text and voice messaging systems
US7562392B1 (en) * 1999-05-19 2009-07-14 Digimarc Corporation Methods of interacting with audio and ambient music
CA2242065C (en) * 1997-07-03 2004-12-14 Henry C.A. Hyde-Thomson Unified messaging system with automatic language identification for text-to-speech conversion
FI115868B (en) * 2000-06-30 2005-07-29 Nokia Corp speech synthesis
US6925154B2 (en) * 2001-05-04 2005-08-02 International Business Machines Corproation Methods and apparatus for conversational name dialing systems
ITFI20010199A1 (en) * 2001-10-22 2003-04-22 Riccardo Vieri SYSTEM AND METHOD TO TRANSFORM TEXTUAL COMMUNICATIONS INTO VOICE AND SEND THEM WITH AN INTERNET CONNECTION TO ANY TELEPHONE SYSTEM
DE50104036D1 (en) * 2001-12-12 2004-11-11 Siemens Ag Speech recognition system and method for operating such a system
KR100450319B1 (en) * 2001-12-24 2004-10-01 한국전자통신연구원 Apparatus and Method for Communication with Reality in Virtual Environments
KR100788652B1 (en) * 2002-02-19 2007-12-26 삼성전자주식회사 Apparatus and method for dialing auto sound
DE10211777A1 (en) * 2002-03-14 2003-10-02 Philips Intellectual Property Creation of message texts
US7917581B2 (en) * 2002-04-02 2011-03-29 Verizon Business Global Llc Call completion via instant communications client
US7123695B2 (en) * 2002-05-21 2006-10-17 Bellsouth Intellectual Property Corporation Voice message delivery over instant messaging
GB0327416D0 (en) * 2003-11-26 2003-12-31 Ibm Directory dialler name recognition
ATE546966T1 (en) * 2003-12-23 2012-03-15 Kirusa Inc TECHNIQUES FOR COMBINING VOICE AND WIRELESS TEXT MESSAGING SERVICES
JP2007534278A (en) * 2004-04-20 2007-11-22 ボイス シグナル テクノロジーズ インコーポレイテッド Voice through short message service
US7583974B2 (en) * 2004-05-27 2009-09-01 Alcatel-Lucent Usa Inc. SMS messaging with speech-to-text and text-to-speech conversion
US8224647B2 (en) * 2005-10-03 2012-07-17 Nuance Communications, Inc. Text-to-speech user's voice cooperative server for instant messaging clients
CA2527813A1 (en) * 2005-11-24 2007-05-24 9160-8083 Quebec Inc. System, method and computer program for sending an email message from a mobile communication device based on voice input
US7929672B2 (en) * 2006-04-18 2011-04-19 Cisco Technology, Inc. Constrained automatic speech recognition for more reliable speech-to-text conversion
US8015010B2 (en) * 2006-06-13 2011-09-06 E-Lane Systems Inc. Vehicle communication system with news subscription service
US8634788B2 (en) * 2007-03-02 2014-01-21 Aegis Mobility, Inc. System and methods for monitoring the context associated with a mobile communication device
US9066199B2 (en) * 2007-06-28 2015-06-23 Apple Inc. Location-aware mobile device
WO2009073806A2 (en) * 2007-12-05 2009-06-11 Johnson Controls Technology Company Vehicle user interface systems and methods
US8538376B2 (en) * 2007-12-28 2013-09-17 Apple Inc. Event-based modes for electronic devices
US8131118B1 (en) * 2008-01-31 2012-03-06 Google Inc. Inferring locations from an image
US8364486B2 (en) * 2008-03-12 2013-01-29 Intelligent Mechatronic Systems Inc. Speech understanding method and system
US8248237B2 (en) * 2008-04-02 2012-08-21 Yougetitback Limited System for mitigating the unauthorized use of a device
US8417720B2 (en) * 2009-03-10 2013-04-09 Nokia Corporation Method and apparatus for accessing content based on user geolocation
US10540976B2 (en) * 2009-06-05 2020-01-21 Apple Inc. Contextual voice commands
US9978272B2 (en) * 2009-11-25 2018-05-22 Ridetones, Inc Vehicle to vehicle chatting and communication system
CN102117614B (en) * 2010-01-05 2013-01-02 索尼爱立信移动通讯有限公司 Personalized text-to-speech synthesis and personalized speech feature extraction
US8655965B2 (en) * 2010-03-05 2014-02-18 Qualcomm Incorporated Automated messaging response in wireless communication systems
EP2619911A4 (en) * 2010-09-21 2015-05-06 Cellepathy Ltd System and method for sensor-based determination of user role, location, and/or state of one of more in-vehicle mobile devices and enforcement of usage thereof

Patent Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2006133547A1 (en) 2005-06-13 2006-12-21 E-Lane Systems Inc. Vehicle immersive communication system
EP1879000A1 (en) 2006-07-10 2008-01-16 Harman Becker Automotive Systems GmbH Transmission of text messages by navigation systems
EP2224705A1 (en) 2009-02-27 2010-09-01 Research In Motion Limited Mobile wireless communications device with speech to text conversion and related method
US20100222086A1 (en) 2009-02-28 2010-09-02 Karl Schmidt Cellular Phone and other Devices/Hands Free Text Messaging

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
See also references of EP2695406A4

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US9653077B2 (en) 2012-11-16 2017-05-16 Honda Motor Co., Ltd. Message processing device

Also Published As

Publication number Publication date
KR20140022824A (en) 2014-02-25
EP2695406A2 (en) 2014-02-12
US20120259633A1 (en) 2012-10-11
JP2014512049A (en) 2014-05-19
WO2012138587A3 (en) 2012-11-29
CN103443852A (en) 2013-12-11
EP2695406A4 (en) 2014-09-03

Similar Documents

Publication Publication Date Title
US20120259633A1 (en) Audio-interactive message exchange
JP5362034B2 (en) Use advanced voicemail through automatic voicemail preview
CN108337380B (en) Automatically adjusting user interface for hands-free interaction
CN102427493B (en) Communication session is expanded with application
US10496753B2 (en) Automatically adapting user interfaces for hands-free interaction
US9111538B2 (en) Genius button secondary commands
US20140195252A1 (en) Systems and methods for hands-free notification summaries
US9418649B2 (en) Method and apparatus for phonetic character conversion
US20120201362A1 (en) Posting to social networks by voice
KR20190012255A (en) Providing a personal assistance module with an optionally steerable state machine
KR101834624B1 (en) Automatically adapting user interfaces for hands-free interaction
EP2327235B1 (en) Pre-determined responses for wireless devices
KR102217301B1 (en) Contact control of artificial intelligence reflecting personal schedule and lifestyle
CN102045462B (en) Method and apparatus for unified interface for heterogeneous session management
KR102017544B1 (en) Interactive ai agent system and method for providing seamless chatting service among users using multiple messanger program, computer readable recording medium
KR20140093170A (en) Method for transmitting information in voicemail and electronic device thereof
CN111818224A (en) Audio data processing method, mobile terminal and storage medium
KR20190104853A (en) Interactive ai agent system and method for providing seamless chatting service among users using multiple messanger program, computer readable recording medium
WO2015005938A1 (en) Interactive voicemail system

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 12768271

Country of ref document: EP

Kind code of ref document: A2

REEP Request for entry into the european phase

Ref document number: 2012768271

Country of ref document: EP

WWE Wipo information: entry into national phase

Ref document number: 2012768271

Country of ref document: EP

ENP Entry into the national phase

Ref document number: 20137026109

Country of ref document: KR

Kind code of ref document: A

ENP Entry into the national phase

Ref document number: 2014503705

Country of ref document: JP

Kind code of ref document: A

NENP Non-entry into the national phase

Ref country code: DE