US9672208B2 - Automatic selection of language for voice interface - Google Patents

Automatic selection of language for voice interface Download PDF

Info

Publication number
US9672208B2
US9672208B2 US14/193,685 US201414193685A US9672208B2 US 9672208 B2 US9672208 B2 US 9672208B2 US 201414193685 A US201414193685 A US 201414193685A US 9672208 B2 US9672208 B2 US 9672208B2
Authority
US
United States
Prior art keywords
language
wireless
source device
audio source
interface
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active, expires
Application number
US14/193,685
Other versions
US20150248399A1 (en
Inventor
Naganagouda B. Patil
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Bose Corp
Original Assignee
Bose Corp
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Bose Corp filed Critical Bose Corp
Priority to US14/193,685 priority Critical patent/US9672208B2/en
Assigned to BOSE CORPORATION reassignment BOSE CORPORATION ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: Patil, Naganagouda B.
Assigned to BOSE CORPORATION reassignment BOSE CORPORATION CORRECTIVE ASSIGNMENT TO CORRECT THE EXECUTION DATES OF CONVEYING PARTIES PREVIOUSLY RECORDED AT REEL: 032823 FRAME: 0626. ASSIGNOR(S) HEREBY CONFIRMS THE ASSIGNMENT. Assignors: Patil, Naganagouda B.
Publication of US20150248399A1 publication Critical patent/US20150248399A1/en
Application granted granted Critical
Publication of US9672208B2 publication Critical patent/US9672208B2/en
Application status is Active legal-status Critical
Adjusted expiration legal-status Critical

Links

Images

Classifications

    • G06F17/28
    • GPHYSICS
    • G06COMPUTING; CALCULATING; COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F40/00Handling natural language data
    • G06F40/40Processing or translation of natural language
    • GPHYSICS
    • G06COMPUTING; CALCULATING; COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/16Sound input; Sound output
    • G06F3/167Audio in a user interface, e.g. using voice commands for navigating, audio feedback
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04MTELEPHONIC COMMUNICATION
    • H04M1/00Substation equipment, e.g. for use by subscribers; Analogous equipment at exchanges
    • H04M1/72Substation extension arrangements; Cordless telephones, i.e. devices for establishing wireless links to base stations without route selecting
    • H04M1/725Cordless telephones
    • H04M1/72519Portable communication terminals with improved user interface to control a main telephone operation mode or to indicate the communication status
    • H04M1/72522With means for supporting locally a plurality of applications to increase the functionality
    • H04M1/72527With means for supporting locally a plurality of applications to increase the functionality provided by interfacing with an external accessory
    • H04M1/7253With means for supporting locally a plurality of applications to increase the functionality provided by interfacing with an external accessory using a two-way short-range wireless interface
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04MTELEPHONIC COMMUNICATION
    • H04M1/00Substation equipment, e.g. for use by subscribers; Analogous equipment at exchanges
    • H04M1/72Substation extension arrangements; Cordless telephones, i.e. devices for establishing wireless links to base stations without route selecting
    • H04M1/725Cordless telephones
    • H04M1/72519Portable communication terminals with improved user interface to control a main telephone operation mode or to indicate the communication status
    • H04M1/72583Portable communication terminals with improved user interface to control a main telephone operation mode or to indicate the communication status for operating the terminal by selecting telephonic functions from a plurality of displayed items, e.g. menus, icons
    • H04W4/005
    • H04W4/008
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04WWIRELESS COMMUNICATION NETWORKS
    • H04W4/00Services specially adapted for wireless communication networks; Facilities therefor
    • H04W4/02Services making use of location information
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04WWIRELESS COMMUNICATION NETWORKS
    • H04W4/00Services specially adapted for wireless communication networks; Facilities therefor
    • H04W4/02Services making use of location information
    • H04W4/029Location-based management or tracking services
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04WWIRELESS COMMUNICATION NETWORKS
    • H04W4/00Services specially adapted for wireless communication networks; Facilities therefor
    • H04W4/70Services for machine-to-machine communication [M2M] or machine type communication [MTC]
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04WWIRELESS COMMUNICATION NETWORKS
    • H04W4/00Services specially adapted for wireless communication networks; Facilities therefor
    • H04W4/80Services using short range communication, e.g. near-field communication [NFC], radio-frequency identification [RFID] or low energy communication
    • G06F9/4448
    • GPHYSICS
    • G06COMPUTING; CALCULATING; COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F9/00Arrangements for program control, e.g. control units
    • G06F9/06Arrangements for program control, e.g. control units using stored programs, i.e. using an internal store of processing equipment to receive or retain programs
    • G06F9/44Arrangements for executing specific programs
    • G06F9/451Execution arrangements for user interfaces
    • G06F9/454Multi-language systems; Localisation; Internationalisation
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04MTELEPHONIC COMMUNICATION
    • H04M1/00Substation equipment, e.g. for use by subscribers; Analogous equipment at exchanges
    • H04M1/72Substation extension arrangements; Cordless telephones, i.e. devices for establishing wireless links to base stations without route selecting
    • H04M1/725Cordless telephones
    • H04M1/72519Portable communication terminals with improved user interface to control a main telephone operation mode or to indicate the communication status
    • H04M1/72563Portable communication terminals with improved user interface to control a main telephone operation mode or to indicate the communication status with means for adapting by the user the functionality or the communication capability of the terminal under specific circumstances
    • H04M1/72572Portable communication terminals with improved user interface to control a main telephone operation mode or to indicate the communication status with means for adapting by the user the functionality or the communication capability of the terminal under specific circumstances according to a geographic location
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04MTELEPHONIC COMMUNICATION
    • H04M2250/00Details of telephonic subscriber devices
    • H04M2250/02Details of telephonic subscriber devices including a Bluetooth interface
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04MTELEPHONIC COMMUNICATION
    • H04M2250/00Details of telephonic subscriber devices
    • H04M2250/58Details of telephonic subscriber devices including a multilanguage function
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04WWIRELESS COMMUNICATION NETWORKS
    • H04W8/00Network data management
    • H04W8/22Processing or transfer of terminal data, e.g. status or physical capabilities

Abstract

A wireless device includes a processor, a memory, a wireless interface, and a voice interface. The processor is configured to select a language for the voice interface by detecting a second wireless device, establishing ongoing communications with the second wireless device, identifying a language used by a user interface of the second wireless device, and configuring the voice interface to use the same language as the user interface of the second wireless device.

Description

BACKGROUND

The present invention relates to user interface technology, and more specifically to automatic selection of a language for a voice interface.

A voice interface uses speech to interact with a user of an electronic device or service. In general, a voice interface may work in one or both of two directions. A text-to-speech interface converts textual data to spoken words to provide verbal output to the user. A voice recognition interface converts the user's spoken input into instructions usable by the electronics.

In general, Bluetooth® wireless technology is a wireless communications system intended to replace the cables connecting electronic devices. In order for two devices to communicate over Bluetooth, they must first be paired. A pairing sequence between two Bluetooth devices typically requires user setup actions on both devices and time for the Bluetooth devices to discover each other. Devices may be connected for various purposes, including remote control, exchange of data, or playback of audio signals generated by one device on a speaker in another device. In the context of this application, when we refer to “Bluetooth” we are referring to protocols and connections operating according to the version of the Bluetooth specification released by the Bluetooth Special Interest Group (Bluetooth SIG) as of the filing date of this application, which is Version 4.0 [Vol 0], 30 Jun. 2010. Other wireless technologies may provide similar features and capabilities and may be used in the same manner.

SUMMARY

The following presents a simplified summary in order to provide a basic understanding of some aspects of the invention. This summary is intended to neither identify key or critical elements of the invention nor delineate the scope of the invention. Its sole purpose is to present some concepts of the invention in a simplified form as a prelude to the more detailed description that is presented later.

In general, in one aspect, a wireless device includes a processor, a memory, a wireless interface, and a voice interface. The processor is configured to select a language for the voice interface by detecting a second wireless device, establishing ongoing communications with the second wireless device, identifying a language used by a user interface of the second wireless device, and configuring the voice interface to use the same language as the user interface of the second wireless device.

Implementations may include one or more of the following, in any combination. The processor may identify the language used by the user interface of the second wireless device by transmitting to the second wireless device a query requesting the identification of the language, and receiving a response identifying the language. The processor may identify the language used by the user interface of the second wireless device by receiving an identification of the language as part of establishing ongoing communications. The processor may identify the language used by the user interface of the second wireless device by receiving geographic information from the second wireless device, and selecting a language corresponding to the geographic information. The processor may be configured to output, via the speaker, an identification of the identified language, and receive through a user interface instructions to configure a different language for use by the voice interface. The wireless interface includes a Bluetooth wireless interface. The processor may configure the voice interface by loading a language library from the memory. The processor may configure the voice interface by receiving a language library from the second wireless device. The processor may configure the voice interface by receiving a language library from a remote server. The processor may receive the language library from the remote server using the wireless interface to access a network interface of the second wireless device.

In general, in one aspect, a system includes a first wireless device with a first processor, a first memory, a first wireless interface, and a voice interface, and a second wireless device with a second processor, a second memory, a second wireless interface, and a user interface. The first processor is configured to select a language for the voice interface by detecting the second wireless device, establishing ongoing communications with the second wireless device, identifying a language used by the user interface of the second wireless device, and configuring the voice interface to use the same language as the user interface of the second wireless device. The first processor transmits to the second wireless device a query requesting the identification of the language, the second processor receives the query and transmits to the first wireless device an identification of the language, and the first processor receives the identification.

Implementations may include one or more of the following, in any combination. The second memory may include a language library corresponding to the language used by the user interface of the second wireless device, the second wireless device providing the language library to the first wireless device, and the first processor configuring the voice interface using the language library from the second wireless device. The second wireless device may include a navigation system, the second wireless device providing geographic information from the navigation system to the first wireless device, and the first processor using the geographic information to identify the language used by the user interface of the second wireless device.

These and other features and advantages will be apparent from a reading of the following detailed description and a review of the associated drawings. It is to be understood that both the foregoing general description and the following detailed description are explanatory only and are not restrictive of aspects as claimed.

BRIEF DESCRIPTION OF THE DRAWINGS

FIG. 1 is a block diagram of an exemplary wireless entertainment system.

FIG. 2 is a flow diagram.

DETAILED DESCRIPTION

An increasingly common feature in consumer electronic products is a voice interface. A voice interface allows a user to interact with the electronics through the spoken word—by speaking instructions to be used as input, and by having the device provide its output as spoken language. Some devices provide only one or the other of these features—speaking output, but not accepting voice input, or vice-versa. For a consumer product targeting a large population, it may be desirable to include more than one language in the product's voice interface, so that a user can speak or hear his preferred language. When the product has no other interface, e.g., no graphical user interface, a challenge in such a product is how to initially select the language to use for the voice interface, given that users may not understand instructions given in a default language that explain how to select a different language. Even if they understand the instructions, the limited physical interfaces provided by some products may make it difficult to select a language from a large inventory of choices.

As shown in FIG. 1, an exemplary wireless entertainment system 10 includes a wireless playback device 15 and one or more wireless audio sources 20 a, 20 b and 20 c. Example wireless playback devices include, but are not limited to, a headset, a portable speaker, an automobile entertainment system, a home theater system, and so forth. Example wireless audio sources 20 a, 20 b, 20 c include, but are not limited to, an entertainment system, a communication system, a personal computer, a smart watch, and so forth. In some examples, Bluetooth wireless technology is used to provide wireless communication between the components of the system. Although FIG. 1 shows audio sources and a playback device, the interfaces described below are applicable to other types of devices, such as network access points, health monitors, and similar devices. In general, we will refer to the device 15 that is the primary subject of this disclosure as an output device, and the device to which is paired as a source device, but both of these roles may be reversed or entirely different in a given system.

In some examples, the output device 15 includes a processor 25, a memory 30, a wireless communication module 35, and a speaker 40. The wireless communication module may include, for example, a Bluetooth wireless communication interface or other similar interface. The output device may also include some form of physical user interface 50, ranging from simple switches and lights to fully interactive touchscreen displays, though the system described herein is generally most valuable when other interfaces are limited.

The processor 25 controls the general operation of the output device. In some examples, where Bluetooth or a similar wireless technology is used, the processor 25 initiates a Bluetooth-specific function implemented in the wireless module 35 upon detecting certain events, as described below. The processor 25 initiates an operation (e.g., pairing) necessary to establish communication between the output device 15 and the one or more wireless audio sources 20 a, 20 b, 20 c using Bluetooth wireless technology.

The memory 30 may include any suitable memory technology, such as Read Only Memory (ROM), Random Access Memory (RAM), or flash ROM, to name a few examples. Among other things, the memory stores a microcode of a program for processing and controlling the processor 25 and a variety of reference data, data generated during execution of any of the variety of programs performed by the processor 25, and various updateable data for safekeeping such as a phone book, outgoing messages, incoming messages and the like. Different types of data may be stored in different types of memory. The memory includes executable code for general operation of the wireless playback device, including the voice user interface In particular, the memory may include libraries of words, phoneme-to-letter mappings, and grammar, spelling and pronunciation rules for multiple languages. The memory also includes executable code for selecting one of the stored languages to use in the voice interface.

The wireless module 35 enables a wireless connection using Radio Frequency (RF) communication between the wireless playback device 15 and the wireless audio sources 20 a, 20 b, 20 c. The wireless module 35 exchanges a radio signal including data input/output through an antenna (not shown). For example, when using Bluetooth wireless technology in a transmission mode, the wireless module 35 processes data by channel coding and spreading, converts the processed data into a Radio Frequency (RF) signal and transmits the RF signal. In a reception mode, the wireless module 35 converts a received RF signal into a baseband signal, processes the baseband signal by de-spreading and channel decoding and restores the processed signal to data.

Each of the source devices, audio source 20 a for example, can include a processor, a memory, and a wireless communication module implementing Bluetooth or other appropriate wireless technology. The particular construction and technology of the source devices is not within the scope of this disclosure, except to note that wireless communication module is compatible with the wireless module 35 of the output device 15, such that it interacts with the output device in the manner described below. In some examples, the voice interface code and language libraries are contained within one of the connected source devices and available for use by the output device, or are provided by an on-line service that the output device can connect to through one of the source devices which is also connected to a wide-area network.

As shown in FIG. 2, a process 200 is performed by the output device 15 in communication with the source device 20 to automatically select a language for the voice interface. The process 200 begins when the output device is powered on (205). The output device then enters a pairing state (207), in which it broadcasts (210) that it is available for pairing, and waits for a source device 20 to respond to the broadcast by initiating pairing (215). In general, a source device will respond to the broadcast when instructed to do so by its user, e.g., as part of a pairing process within the source device's user interface. When a source device responds to the pairing broadcast, it will provide several pieces of information, including, for example, its name and the address it uses in the wireless protocol used by the two devices. In response to that, the output device confirms (220) that the pairing was successful and queries (225) the source device about what language it is using for its own interface. For the operation of process 200, it is assumed that the source device itself has a user interface that includes multiple languages, whether text-based or voice-based. The source devices replies (230) to the query with an identification of its language.

The output device sets (235) its own voice interface to use the same language, for example by loading a language library from the memory 30. In some examples, the output device then outputs (240) an audible indication of what language was selected, and offers (245) to let the user change it (250), initially using the automatically selected language to provide an interface for this. While the user may not want the output device to use the same language as the source device, for example, if the source device does not actually offer the user's preferred language, it may be assumed that the user will at least understand the language that the source device is using well enough to then use that language in the voice interface of the output device to select his preferred language.

In another example, the output device may not have the required language pre-installed. In this case, once it learns what language the source device is using, it downloads (255) the appropriate language, either from the source device itself, or using a network interface (internal or within the source device) to obtain the language resources from an on-line source. Especially if the availability of a network is highly likely, the output device may ship with few languages installed (saving on memory resources), and rely on the network to get whatever language it needs.

If the source device does not know or cannot communicate what language it is using, shown by optional branch point 260, other data within the source device may be used (265) to make a guess about what language should be used. For example, if the source device has access to and can share global positioning system (GPS) or other navigation data, the output device can select the majority language for the region where the devices are located. Similarly, the radio frequencies supported by any wireless interfaces in the source device may allow the output device to make an educated guess about its geographic location and therefore the likely language preference of its user.

The direct selection of audio source process 200 is illustrated in the context of the BLUETOOTH SPECIFICATION Version 4.0 [Vol 0], 30 Jun. 2010. However, it should be understood that the approach described above is equally applicable to other wireless protocols.

While the above describes a particular order of operations performed by certain embodiments of the invention, it should be understood that such order is exemplary, as alternative embodiments may perform the operations in a different order, combine certain operations, overlap certain operations, or the like. References in the specification to a given embodiment indicate that the embodiment described may include a particular feature, structure, or characteristic, but every embodiment may not necessarily include the particular feature, structure, or characteristic.

While given components of the system have been described separately, one of ordinary skill will appreciate that some of the functions may be combined or shared in given instructions, program sequences, code portions, and the like.

The foregoing description does not represent an exhaustive list of all possible implementations consistent with this disclosure or of all possible variations of the implementations described. A number of implementations have been described. Nevertheless, it will be understood that various modifications may be made without departing from the spirit and scope of the systems, devices, methods and techniques described here. Accordingly, other implementations are within the scope of the following claims.

Claims (24)

What is claimed is:
1. A method comprising, in a wireless headset comprising at least a processor, a memory, a speaker, a wireless interface, and a voice interface:
detecting a wireless audio source device;
establishing ongoing communications with the wireless audio source device;
identifying a suggested language for the voice interface by receiving geographic information from the wireless audio source device, wherein the suggested language is based on the geographic information;
configuring the voice interface of the wireless headset to provide the voice interface in the suggested language by receiving a language library for the suggested language from a remote server by using the wireless interface to access a network interface of the wireless audio source device, and using the network interface of the wireless audio source device to download the language library from the remote server;
outputting, via the speaker, an identification of the suggested language used to configure the voice interface; and
making available to the user an option to configure a different language for use by the voice interface.
2. The method of claim 1 wherein identifying the suggested language for the voice interface comprises transmitting to the wireless audio source device a query requesting the geographic information, and receiving a response identifying the geographic information.
3. The method of claim 2 wherein identifying the geographic information comprises receiving an identification of the geographic information as part of establishing ongoing communications.
4. The method of claim 1 wherein the wireless interface comprises a Bluetooth wireless interface.
5. The method of claim 1 wherein configuring the voice interface comprises loading the language library from the memory.
6. The method of claim 1 wherein configuring the voice interface comprises receiving the language library from the wireless audio source device.
7. The method of claim 1, further comprising determining a location of the wireless audio source device based on the geographic information.
8. The method of claim 7, wherein the suggested language is the majority language for the location of the wireless audio source device.
9. The method of claim 1, wherein the geographic information is based on at least one of: data from a global positioning system (GPS) in the wireless audio source device and a radio frequency supported by the wireless audio source device.
10. A wireless headset comprising:
a processor,
a memory,
a speaker,
a wireless interface, and
a voice interface, wherein
the processor is configured to select a language for the voice interface by:
detecting a wireless audio source device;
establishing ongoing communications with the wireless audio source device;
identifying a suggested language for the voice interface by receiving geographic information from the wireless audio source device, wherein the suggested language is based on the geographic information;
configuring the voice interface to use the suggested language by receiving a language library for the suggested language from a remote server by using the wireless interface to access a network interface of the wireless audio source device, and using the network interface of the wireless audio source device to download the language library from the remote server;
outputting, via the speaker, an identification of the suggested language used to configure the voice interface; and
making available to the user an option to configure a different language for use by the voice interface.
11. The wireless headset of claim 10 wherein the processor identifies the suggested language for the voice interface by transmitting to the wireless audio source device a query requesting the geographic information, and receiving a response identifying the geographic information.
12. The wireless headset of claim 11 wherein the processor identifies the geographic information by receiving an identification of the geographic information as part of establishing ongoing communications.
13. The wireless headset of claim 10 wherein the wireless interface comprises a Bluetooth wireless interface.
14. The wireless headset of claim 10 wherein the processor configures the voice interface by loading the language library from the memory.
15. The wireless headset of claim 10 wherein the processor configures the voice interface by receiving the language library from the wireless audio source device.
16. The wireless headset of claim 10, wherein the processor is further configured to determine a location of the wireless audio source device based on the geographic information.
17. The wireless headset of claim 16, wherein the suggested language is the majority language of the location of the wireless audio source device.
18. The wireless headset of claim 10, wherein the geographic information is based on at least one of: data from a global positioning system (GPS) in the wireless audio source device and a radio frequency supported by the wireless audio source device.
19. A system comprising:
a wireless headset comprising
a first processor,
a first memory,
a first speaker,
a first wireless interface, and
a voice interface; and
a wireless audio source device comprising
a second processor,
a second memory,
a second wireless interface, and
a user interface; wherein
the first processor is configured to select a language for the voice interface by:
detecting the wireless audio source device;
establishing ongoing communications with the wireless audio source device;
identifying a suggested language for the voice interface by receiving geographic information from the wireless audio source device, wherein the suggested language is based on the geographic information;
configuring the voice interface to use the suggested language by receiving a language library for the suggested language from a remote server by using the wireless interface to access a network interface of the wireless audio source device, and using the network interface of the wireless audio source device to download the language library from the remote server;
outputting, via the first speaker, an identification of the suggested language used to configure the voice interface; and
making available to the user an option to configure a different language for use by the voice interface.
20. The system of claim 19 wherein:
the first processor transmits to the wireless audio source device a query requesting the identification of the geographic information,
the second processor receives the query and transmits to the wireless headset an identification of the geographic information, and
the first processor receives the identification.
21. The system of claim 19 wherein:
the wireless audio source device includes a navigation system;
the wireless audio source device provides geographic information from the navigation system to the wireless headset; and
the first processor uses the geographic information to identify the suggested language.
22. The system of claim 19 wherein the first processor configures the voice interface by loading the language library from the first memory.
23. The system of claim 19, wherein the first processor configures the voice interface by receiving the language library from the wireless audio source device.
24. The wireless headset of claim 19, wherein the first processor is further configured to determine a location of the wireless audio source device based on the geographic information, and the suggested language is the majority language for the location of the wireless audio source device.
US14/193,685 2014-02-28 2014-02-28 Automatic selection of language for voice interface Active 2034-08-03 US9672208B2 (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
US14/193,685 US9672208B2 (en) 2014-02-28 2014-02-28 Automatic selection of language for voice interface

Applications Claiming Priority (3)

Application Number Priority Date Filing Date Title
US14/193,685 US9672208B2 (en) 2014-02-28 2014-02-28 Automatic selection of language for voice interface
CN201580010913.4A CN106134166A (en) 2014-02-28 2015-02-26 Automatically selecting of the language of speech interface
PCT/US2015/017656 WO2015130887A1 (en) 2014-02-28 2015-02-26 Automatic selection of language for voice interface

Publications (2)

Publication Number Publication Date
US20150248399A1 US20150248399A1 (en) 2015-09-03
US9672208B2 true US9672208B2 (en) 2017-06-06

Family

ID=52630519

Family Applications (1)

Application Number Title Priority Date Filing Date
US14/193,685 Active 2034-08-03 US9672208B2 (en) 2014-02-28 2014-02-28 Automatic selection of language for voice interface

Country Status (3)

Country Link
US (1) US9672208B2 (en)
CN (1) CN106134166A (en)
WO (1) WO2015130887A1 (en)

Families Citing this family (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US9998334B1 (en) * 2017-08-17 2018-06-12 Chengfu Yu Determining a communication language for internet of things devices

Citations (18)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20050131685A1 (en) * 2003-11-14 2005-06-16 Voice Signal Technologies, Inc. Installing language modules in a mobile communication device
US20060184356A1 (en) * 2005-02-16 2006-08-17 Samsung Electronics Co., Ltd. Multiple language support system and method for application
US20070073530A1 (en) * 2003-12-19 2007-03-29 Juha Iso-Sipila Electronic device equipped with a voice user interface and a method in an electronic device for performing language configurations of a user interface
US7203486B2 (en) * 2003-05-19 2007-04-10 France Telecom Wireless system having a dynamically configured multimodal user interface based on user preferences
US20070124675A1 (en) * 2005-11-29 2007-05-31 Ban Oliver K Methods and systems for changing language characters of graphical and application interfaces
US7340389B2 (en) * 2001-02-16 2008-03-04 Microsoft Corporation Multilanguage UI with localized resources
US20090178010A1 (en) * 2008-01-06 2009-07-09 Apple Inc. Specifying Language and Other Preferences for Mobile Device Applications
US20090279532A1 (en) * 2008-05-06 2009-11-12 Custom Teleconnect, Inc. Tcp/ip based voice communication system
US20090325630A1 (en) * 2008-06-27 2009-12-31 Nokia Corporation Method, apparatus, and computer program product for personalizing a device
US20100169074A1 (en) * 2008-12-31 2010-07-01 Hon Hai Precision Industry Co., Ltd. Method of configuring user preferences on electronic device
US20100267376A1 (en) * 2007-12-17 2010-10-21 Nokia Corporation Accessory Configuration and Management
US20130162523A1 (en) * 2011-12-27 2013-06-27 Advanced Micro Devices, Inc. Shared wireless computer user interface
EP2635058A1 (en) 2010-10-28 2013-09-04 ZTE Corporation Method and mobile terminal for setting language through mobile terminal
US8775157B2 (en) * 2011-04-21 2014-07-08 Blackberry Limited Methods and systems for sharing language capabilities
US20140317523A1 (en) * 2013-04-19 2014-10-23 Microsoft Corporation User experience mode transitioning
US9030409B2 (en) * 2013-01-11 2015-05-12 Lg Electronics Inc. Device for transmitting and receiving data using earphone and method for controlling the same
US20150220512A1 (en) * 2014-02-05 2015-08-06 Marco Álvarez Heinemeyer Language interface system, method and computer readable medium
US9106721B2 (en) * 2012-10-02 2015-08-11 Nextbit Systems Application state synchronization across multiple devices

Patent Citations (19)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US7340389B2 (en) * 2001-02-16 2008-03-04 Microsoft Corporation Multilanguage UI with localized resources
US7203486B2 (en) * 2003-05-19 2007-04-10 France Telecom Wireless system having a dynamically configured multimodal user interface based on user preferences
US20050131685A1 (en) * 2003-11-14 2005-06-16 Voice Signal Technologies, Inc. Installing language modules in a mobile communication device
US20070073530A1 (en) * 2003-12-19 2007-03-29 Juha Iso-Sipila Electronic device equipped with a voice user interface and a method in an electronic device for performing language configurations of a user interface
US20060184356A1 (en) * 2005-02-16 2006-08-17 Samsung Electronics Co., Ltd. Multiple language support system and method for application
US20070124675A1 (en) * 2005-11-29 2007-05-31 Ban Oliver K Methods and systems for changing language characters of graphical and application interfaces
US20100267376A1 (en) * 2007-12-17 2010-10-21 Nokia Corporation Accessory Configuration and Management
US20090178010A1 (en) * 2008-01-06 2009-07-09 Apple Inc. Specifying Language and Other Preferences for Mobile Device Applications
US20090279532A1 (en) * 2008-05-06 2009-11-12 Custom Teleconnect, Inc. Tcp/ip based voice communication system
US20090325630A1 (en) * 2008-06-27 2009-12-31 Nokia Corporation Method, apparatus, and computer program product for personalizing a device
US20100169074A1 (en) * 2008-12-31 2010-07-01 Hon Hai Precision Industry Co., Ltd. Method of configuring user preferences on electronic device
EP2635058A1 (en) 2010-10-28 2013-09-04 ZTE Corporation Method and mobile terminal for setting language through mobile terminal
US9049557B2 (en) * 2010-10-28 2015-06-02 Zte Corporation Language setting method for mobile terminal and mobile terminal
US8775157B2 (en) * 2011-04-21 2014-07-08 Blackberry Limited Methods and systems for sharing language capabilities
US20130162523A1 (en) * 2011-12-27 2013-06-27 Advanced Micro Devices, Inc. Shared wireless computer user interface
US9106721B2 (en) * 2012-10-02 2015-08-11 Nextbit Systems Application state synchronization across multiple devices
US9030409B2 (en) * 2013-01-11 2015-05-12 Lg Electronics Inc. Device for transmitting and receiving data using earphone and method for controlling the same
US20140317523A1 (en) * 2013-04-19 2014-10-23 Microsoft Corporation User experience mode transitioning
US20150220512A1 (en) * 2014-02-05 2015-08-06 Marco Álvarez Heinemeyer Language interface system, method and computer readable medium

Non-Patent Citations (3)

* Cited by examiner, † Cited by third party
Title
"User guide SmartWatch 2 SW2" , 1-3, 5-13, Jul. 31, 2013 (Jul. 31, 2013) , pp. 1-18, 15-23 XP055188454, Retrieved from the Internet: URL:http://data.manualslib.com/pdf3/58/5712/571145-sony/sw2.pdf?f2608d9182230308a674a77f717147dO [retrieved on May 11, 2015].
"User guide SmartWatch 2 SW2", SONY MOBILE COMMUNICATIONS AB, 31 July 2013 (2013-07-31), pages 1 - 18, XP055188454, Retrieved from the Internet <URL:http://data.manualslib.com/pdf3/58/5712/571145-sony/sw2.pdf?f2608d9182230308a674a77f717147d0> [retrieved on 20150511]
International Search Report and Written Opinion dated May 22, 2015 for International application No. PCT/US2015/017656.

Also Published As

Publication number Publication date
CN106134166A (en) 2016-11-16
WO2015130887A1 (en) 2015-09-03
US20150248399A1 (en) 2015-09-03

Similar Documents

Publication Publication Date Title
US9865259B1 (en) Speech-responsive portable speaker
US8793138B2 (en) Method and apparatus for smart voice recognition
CN103136928B (en) Terminal device, remote control equipment and its method of automatic matching
US8892175B2 (en) Connected phone system and method with associated bandwidth connections in a vehicle environment
US9507772B2 (en) Instant translation system
US9736587B2 (en) Smart tool for headphones
JP6480568B2 (en) Voice application architecture
KR101972955B1 (en) Method and apparatus for connecting service between user devices using voice
US20130332168A1 (en) Voice activated search and control for applications
EP3300341A1 (en) Volume adjustment method, system and equipment, and computer storage medium
US10313504B2 (en) Wireless handsfree headset method and system with handsfree applications
US6931463B2 (en) Portable companion device only functioning when a wireless link established between the companion device and an electronic device and providing processed data to the electronic device
US9711160B2 (en) Smart dock for activating a voice recognition mode of a portable electronic device
US9648153B2 (en) Method and device for incoming call notification
CN101641932B (en) Apparatus for providing location information of hand-held device and method thereof
KR101027548B1 (en) Voice browser dialog enabler for a communication system
US6753790B2 (en) Method and an apparatus for an adaptive remote controller
US20170126680A1 (en) Enabling and inhibiting synchronization of privacy settings
JP5355591B2 (en) Navigation device
US9363352B2 (en) Duplex audio for mobile communication device and accessory
US9641625B2 (en) Method and system for executing an internet radio application within a vehicle
US9674331B2 (en) Transmitting data from an automated assistant to an accessory
JP4536140B2 (en) Mobile terminal with media tuning, method of operating the same, and computer program
DE102011075066A1 (en) Method and systems for interfacing with a vehicle data processing system via multiple data transport channels
CN103997366A (en) Wireless network audio system having auto-pairing function and auto-pairing method using the same

Legal Events

Date Code Title Description
AS Assignment

Owner name: BOSE CORPORATION, MASSACHUSETTS

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:PATIL, NAGANAGOUDA B.;REEL/FRAME:032823/0626

Effective date: 20140502

AS Assignment

Owner name: BOSE CORPORATION, MASSACHUSETTS

Free format text: CORRECTIVE ASSIGNMENT TO CORRECT THE EXECUTION DATES OF CONVEYING PARTIES PREVIOUSLY RECORDED AT REEL: 032823 FRAME: 0626. ASSIGNOR(S) HEREBY CONFIRMS THE ASSIGNMENT;ASSIGNOR:PATIL, NAGANAGOUDA B.;REEL/FRAME:035485/0936

Effective date: 20150123

STCF Information on status: patent grant

Free format text: PATENTED CASE