CN107819929A - It is preferred that the identification and generation of emoticon - Google Patents

It is preferred that the identification and generation of emoticon Download PDF

Info

Publication number
CN107819929A
CN107819929A CN201710813706.3A CN201710813706A CN107819929A CN 107819929 A CN107819929 A CN 107819929A CN 201710813706 A CN201710813706 A CN 201710813706A CN 107819929 A CN107819929 A CN 107819929A
Authority
CN
China
Prior art keywords
emoticon
vehicle
user
voice
storehouse
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN201710813706.3A
Other languages
Chinese (zh)
Inventor
X·F·赵
G·塔瓦尔
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
GM Global Technology Operations LLC
Original Assignee
GM Global Technology Operations LLC
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by GM Global Technology Operations LLC filed Critical GM Global Technology Operations LLC
Publication of CN107819929A publication Critical patent/CN107819929A/en
Pending legal-status Critical Current

Links

Classifications

    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
    • G10L13/00Speech synthesis; Text to speech systems
    • G10L13/08Text analysis or generation of parameters for speech synthesis out of text, e.g. grapheme to phoneme translation, prosody generation or stress or intonation determination
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04MTELEPHONIC COMMUNICATION
    • H04M1/00Substation equipment, e.g. for use by subscribers
    • H04M1/72Mobile telephones; Cordless telephones, i.e. devices for establishing wireless links to base stations without route selection
    • H04M1/724User interfaces specially adapted for cordless or mobile telephones
    • H04M1/72403User interfaces specially adapted for cordless or mobile telephones with means for local support of applications that increase the functionality
    • H04M1/7243User interfaces specially adapted for cordless or mobile telephones with means for local support of applications that increase the functionality with interactive means for internal management of messages
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/048Interaction techniques based on graphical user interfaces [GUI]
    • G06F3/0481Interaction techniques based on graphical user interfaces [GUI] based on specific properties of the displayed interaction object or a metaphor-based environment, e.g. interaction with desktop elements like windows or icons, or assisted by a cursor's changing behaviour or appearance
    • G06F3/0482Interaction with lists of selectable items, e.g. menus
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/50Information retrieval; Database structures therefor; File system structures therefor of still image data
    • G06F16/51Indexing; Data structures therefor; Storage structures
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/011Arrangements for interaction with the human body, e.g. for user immersion in virtual reality
    • G06F3/012Head tracking input arrangements
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/017Gesture based interaction, e.g. based on a set of recognized hand gestures
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/048Interaction techniques based on graphical user interfaces [GUI]
    • G06F3/0481Interaction techniques based on graphical user interfaces [GUI] based on specific properties of the displayed interaction object or a metaphor-based environment, e.g. interaction with desktop elements like windows or icons, or assisted by a cursor's changing behaviour or appearance
    • G06F3/04817Interaction techniques based on graphical user interfaces [GUI] based on specific properties of the displayed interaction object or a metaphor-based environment, e.g. interaction with desktop elements like windows or icons, or assisted by a cursor's changing behaviour or appearance using icons
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/048Interaction techniques based on graphical user interfaces [GUI]
    • G06F3/0487Interaction techniques based on graphical user interfaces [GUI] using specific features provided by the input device, e.g. functions controlled by the rotation of a mouse with dual sensing arrangements, or of the nature of the input device, e.g. tap gestures based on pressure sensed by a digitiser
    • G06F3/0488Interaction techniques based on graphical user interfaces [GUI] using specific features provided by the input device, e.g. functions controlled by the rotation of a mouse with dual sensing arrangements, or of the nature of the input device, e.g. tap gestures based on pressure sensed by a digitiser using a touch-screen or digitiser, e.g. input of commands through traced gestures
    • G06F3/04883Interaction techniques based on graphical user interfaces [GUI] using specific features provided by the input device, e.g. functions controlled by the rotation of a mouse with dual sensing arrangements, or of the nature of the input device, e.g. tap gestures based on pressure sensed by a digitiser using a touch-screen or digitiser, e.g. input of commands through traced gestures for inputting data by handwriting, e.g. gesture or text
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/16Sound input; Sound output
    • G06F3/167Audio in a user interface, e.g. using voice commands for navigating, audio feedback
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
    • G10L15/00Speech recognition
    • G10L15/08Speech classification or search
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
    • G10L15/00Speech recognition
    • G10L15/08Speech classification or search
    • G10L15/18Speech classification or search using natural language modelling
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
    • G10L15/00Speech recognition
    • G10L15/24Speech recognition using non-acoustical features
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
    • G10L15/00Speech recognition
    • G10L15/26Speech to text systems
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L51/00User-to-user messaging in packet-switching networks, transmitted according to store-and-forward or real-time protocols, e.g. e-mail
    • H04L51/07User-to-user messaging in packet-switching networks, transmitted according to store-and-forward or real-time protocols, e.g. e-mail characterised by the inclusion of specific contents
    • H04L51/10Multimedia information
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04MTELEPHONIC COMMUNICATION
    • H04M1/00Substation equipment, e.g. for use by subscribers
    • H04M1/72Mobile telephones; Cordless telephones, i.e. devices for establishing wireless links to base stations without route selection
    • H04M1/724User interfaces specially adapted for cordless or mobile telephones
    • H04M1/72403User interfaces specially adapted for cordless or mobile telephones with means for local support of applications that increase the functionality
    • H04M1/7243User interfaces specially adapted for cordless or mobile telephones with means for local support of applications that increase the functionality with interactive means for internal management of messages
    • H04M1/72436User interfaces specially adapted for cordless or mobile telephones with means for local support of applications that increase the functionality with interactive means for internal management of messages for text messaging, e.g. short messaging services [SMS] or e-mails
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04MTELEPHONIC COMMUNICATION
    • H04M1/00Substation equipment, e.g. for use by subscribers
    • H04M1/72Mobile telephones; Cordless telephones, i.e. devices for establishing wireless links to base stations without route selection
    • H04M1/724User interfaces specially adapted for cordless or mobile telephones
    • H04M1/72403User interfaces specially adapted for cordless or mobile telephones with means for local support of applications that increase the functionality
    • H04M1/7243User interfaces specially adapted for cordless or mobile telephones with means for local support of applications that increase the functionality with interactive means for internal management of messages
    • H04M1/72439User interfaces specially adapted for cordless or mobile telephones with means for local support of applications that increase the functionality with interactive means for internal management of messages for image or video messaging
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F2203/00Indexing scheme relating to G06F3/00 - G06F3/048
    • G06F2203/01Indexing scheme relating to G06F3/01
    • G06F2203/011Emotion or mood input determined on the basis of sensed human body parameters such as pulse, heart rate or beat, temperature of skin, facial expressions, iris, voice pitch, brain activity patterns
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
    • G10L13/00Speech synthesis; Text to speech systems
    • G10L13/08Text analysis or generation of parameters for speech synthesis out of text, e.g. grapheme to phoneme translation, prosody generation or stress or intonation determination
    • G10L2013/083Special characters, e.g. punctuation marks
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L51/00User-to-user messaging in packet-switching networks, transmitted according to store-and-forward or real-time protocols, e.g. e-mail
    • H04L51/07User-to-user messaging in packet-switching networks, transmitted according to store-and-forward or real-time protocols, e.g. e-mail characterised by the inclusion of specific contents
    • H04L51/08Annexed information, e.g. attachments

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Human Computer Interaction (AREA)
  • Physics & Mathematics (AREA)
  • General Engineering & Computer Science (AREA)
  • Multimedia (AREA)
  • Health & Medical Sciences (AREA)
  • General Physics & Mathematics (AREA)
  • Audiology, Speech & Language Pathology (AREA)
  • Computational Linguistics (AREA)
  • Acoustics & Sound (AREA)
  • Signal Processing (AREA)
  • Computer Networks & Wireless Communication (AREA)
  • General Business, Economics & Management (AREA)
  • Business, Economics & Management (AREA)
  • Artificial Intelligence (AREA)
  • Software Systems (AREA)
  • Data Mining & Analysis (AREA)
  • Databases & Information Systems (AREA)
  • General Health & Medical Sciences (AREA)
  • Navigation (AREA)
  • Telephone Function (AREA)
  • Telephonic Communication Services (AREA)

Abstract

A kind of system and method for being used to identify and generate preferred emoticon include:The emoticon of multiple selections is detected in wireless devices;It is determined that select the frequency of each emoticon;Based on the frequency for selecting each emoticon, the emoticon of identification specified quantity from the emoticon of multiple selections;And create conventional emoticon storehouse for the emoticon of identification.

Description

It is preferred that the identification and generation of emoticon
Technical field
The present invention relates to using emoticon, relate more specifically to identify and generate the emoticon that user most often sends.
Background technology
Electronic device user sends more complicated electronic information with its equipment.Past, electronic information only use comprising user The content of text of keyboard addition.Electronic information is had developed into comprising content in addition to text.For example, electronic device user can To be selected from a large amount of emoticon arrays that can be included in electronic information.Emoticon is to express to graphically The small-sized art pattern of idea, may be embodied in electronic information.Many electronic equipments, which all have, includes many emoticons Storehouse, user be may browse through and selected in the message for them.Even if user can access many different expressions, but use Family send message generally only comprising equipment storehouse in available emoticon sub-fraction.It is the most frequently used in identification and selection storehouse Emoticon can be efficiently completed.
The content of the invention
According to one embodiment, there is provided a kind of method for identifying and generating preferred emoticon.This method is included in nothing The emoticon of multiple selections is detected at line equipment;It is determined that select the frequency of each emoticon;Based on each emoticon of selection Number frequency, from the emoticon of multiple selections identify specified quantity emoticon;And the emoticon wound for identification Build a conventional emoticon storehouse.
According to another embodiment, there is provided a kind of method for identifying and generating preferred emoticon.This method is included in Wireless devices initiate electronic information;The voice received to contain the emoticon in electronic information describes;By reception Voice is contrasted with the emoticon description being stored in conventional emoticon storehouse;Emoticon is identified based on the contrast; And emoticon is inserted in electronic information.
According to another embodiment, there is provided a kind of method for identifying and generating preferred emoticon.This method is included in Wireless devices initiate electronic information;Receive for identifying that it is defeated that the user for the emoticon being included in electronic information specifies Enter;The input that the input that the user received is specified is specified with the previously stored user being stored in conventional emoticon storehouse Association with emoticon is contrasted;Emoticon is identified based on the contrast;And emoticon is inserted into electronic information In.
Brief description of the drawings
One or more embodiments of the invention is described below with reference to accompanying drawing, wherein identical reference represents identical Element, and wherein:
Fig. 1 is the block diagram for the embodiment for showing the communication system that can utilize method disclosed herein;And
Fig. 2 is the block diagram for the embodiment for showing text-to-speech (TTS) system that can utilize method disclosed herein;
Fig. 3 is the frame for the embodiment for showing automatic speech recognition (ASR) system that can utilize method disclosed herein Figure;And
Fig. 4 is the flow chart of the embodiment for the method for showing identification and generating preferred emoticon.
Embodiment
Systems described below and method identify the emoticon most often selected by the user of wireless device, and contribute to These emoticons are inserted into electronic information.Wireless device can monitor what wireless device users were sent within a period of time Emoticon.Wireless device can calculate the number that user sends particular emotion symbol, and after the period of time has elapsed, nothing Line equipment can determine the emoticon most often sent.Then, wireless device creates the conventional emoticon of these emoticons Storehouse.Conventional emoticon storehouse can be associated with the text description of these emoticons by the most frequently used emoticon.Text is retouched State the description generally accepted that can include each emoticon and added by specific user or user based on user investigation The definition specified.
Emoticon in conventional emoticon storehouse can be related to the input that the user that wireless device can detect specifies Connection.For example, the input that user specifies can be the facial expression that can be identified by camera.User can be in conventional emoticon storehouse Emoticon is identified, and emoticon and specific facial expression are matched.When user wants the emoticon being added to electricity During sub- message, user can make the facial expression associated with emoticon, and camera will detect the facial expression, access with The associated emoticon of expression, and the emoticon is added in electronic information.In another example, user specifies Input can be specific pattern movement of the finger of user on touch pad or touch-screen.User can be in conventional emoticon storehouse Middle identification emoticon, and emoticon is matched with user with the specific pattern of its finger tracking.When user wish by When emoticon is added to electronic information, user can draw the specific pattern associated with emoticon on the touchscreen, with The associated wireless device of touch-screen will detect the pattern, access the emoticon associated with the pattern, and by emoticon It is added in electronic information.
Communication system
With reference to figure 1, show including mobile vehicle communication system 10 and can be used for realizing method disclosed herein Operating environment.Communication system 10 generally includes vehicle 12, one or more wireless carrier systems 14, terrestrial communications network 16, meter Calculation machine 18 and call center 20.It should be appreciated that disclosed method can be used together with any amount of different system, and It is not particularly limited to operating environment shown in this article.In addition, the framework of system 10 and its all parts, construction, setting and operation exist It is commonly known in this area.Therefore, paragraphs below simply provides the general introduction of this communication system 10;However, herein Unshowned other systems can also use disclosed method.
Vehicle 12 is described as car in the embodiment shown, but it is to be understood that any other vehicle can also be used, Including motorcycle, truck, sport vehicle (SUV), leisure vehicle (RV), ship, aircraft etc..Generally shown in Fig. 1 Some vehicle electronics 28, it includes telematics unit 30, microphone 32, one or more buttons or other controls System input 34, audio system 36, visual displays 38, GPS module 40 and a number of other vehicle system modules (VSM) 42.This Some in a little equipment can be directly connected to telematics unit, such as microphone 32 and button 34, and miscellaneous equipment Then use one or more network connections, such as communication bus 44 or entertainment bus 46.The example of suitable network connection includes Controller zone network (CAN), media guidance system transmission (MOST), local interconnection network (LIN), LAN (LAN) and Other appropriate connections, for example Ethernet or meet known ISO, SAE and ieee standard and other networks of specification, only lift number Example.
Telematics unit 30 is a vehicle system module (VSM) in itself, and can be implemented as being arranged on vehicle Interior OEM installations (embedded) or equipment, and by wireless carrier system 14 and nothing can be realized via wireless network after sale Line voice and/or data communication.This is allowed the vehicle to and call center 20, other telematics vehicles or other entities Or equipment is communicated.It is logical with wireless carrier system 14 to establish that wireless radio transmission is preferably used in telematics unit Believe channel (voice channel and/or data channel) so that voice and/or data transfer can be sent and received by the channel. By providing voice communications versus data communications, telematics unit 30 allows the vehicle to provide a variety of different services, including The service related to navigation, phone, emergency aid, diagnosis, Infotainment etc..Data can connect via data (for example to be passed through Packet data transmission on data channel) or via the voice channel using techniques known in the art.Lead to for being related to voice Letter (such as with the Field Adviser at call center 20 or voice response unit) and data communicate (for example to be carried to call center 20 For GPS location data or vehicle diagnostic data) composite services, system can be by voice channel using individual calling and as needed Switch between being transmitted by the voice-and-data of voice channel, and can be come using technology well known by persons skilled in the art Complete.
According to embodiment, telematics unit 30 utilizes the cellular communication according to GSM, CDMA or LTE standard, and Therefore including standard cellular chipset 50, the wireless modulation-demodulation for data transfer for voice communication (hands free calls) Device, electronic processing equipment 52, one or more digital storage equipments 54 and double antenna 56.It should be appreciated that modem can To be realized by the software for being stored in telematics unit and being performed by processor 52, or it can be positioned at remote Single hardware component inside or outside journey information process unit 30.Modem can use any amount of different marks Accurate or agreement (such as LTE, EVDO, CDMA, GPRS and EDGE) is operated.Wireless network between vehicle and other networked devices Network can also be performed using telematics unit 30.Therefore, telematics unit 30 can be configured to according to one Kind or multiple wireless protocols progress radio communication, including short-distance wireless communication (SRWC), such as IEEE802.11 agreements, WiMAX、ZigBeeTM, Wi-Fi directly, bluetoothTMOr near-field communication (NFC).When for packet switched data communication (such as T/ When IP), telematics unit can be configured to static ip address, or could be arranged to set from another on network automatically Standby (such as router or from network address server) receives the IP address of distribution.
One of networked devices that can be communicated with telematics unit 30 are wireless devices, such as smart phone 57. Smart phone 57 can include computer process ability, can use the transceiver that is communicated of short-range wireless protocol and Visible intelligent telephone displays 59.In some embodiments, smart phone display 59 also includes touch-screen graphical user circle Face.Smart phone 57 can also include that gps satellite signal can be received and based on the GPS module of those signal generation gps coordinates. Smart phone 57 also includes performing machine code to generate the one or more microprocessors of logic output.One or more cameras It can be included in smart phone 57.Camera can be located at the opposite side of smart phone display 59.But in some configurations In, smart phone can have multiple cameras, and one of camera is adjacent with display 59.The example of smart phone 57 include by The iPhone and the Galaxy and other by Samsung manufacture of Apple Inc.'s manufacture.Although smart phone 57 can be including the use of nothing The ability that line carrier system 14 is communicated by cellular communication, but simultaneously not always this situation.For example, Apple Inc.'s production is all IPad the and iPod Touch equipment of such as various models, it includes disposal ability, display screen 59 and passes through short-distance wireless The ability that communication link is communicated.However, iPod TouchTMWith some iPadTMDo not possess cellular communication capability.Even if such as This, for method described herein, these and other similar equipment can also be used or be considered as that one kind is wirelessly set Standby type, such as smart phone 57.
Processor 52 can be any kind of equipment that can handle e-command, including microprocessor, microcontroller, Primary processor, controller, vehicle communication processor and application specific integrated circuit (ASIC).It can be only for telematics The application specific processor of unit 30, or can be shared with other Vehicular systems.Processor 52 performs various types of stored digitals Instruction, such as the software or firmware program being stored in memory 54, it enables telematics unit to provide respectively The service of kind various kinds.For example, processor 52 can be with configuration processor or processing data, to perform at least the one of the method being discussed herein Part.
Telematics unit 30 may be used to provide the various each of the radio communication that is related to reaching and/or from vehicle The vehicle service of sample.Such service includes:The turn direction provided together with the automobile navigation module 40 based on GPS and its Its navigation related service;Be connected cloth with one or more crash sensor interface modules (such as car body control module (not shown)) The air-bag deployment notice put and other urgent or roadside auxiliary related services;Use the diagnosis of one or more diagnostic modules Report;And by Infotainment module (not shown) download and be stored for the music currently or later played, webpage, Film, TV programme, electronic game and/or the Infotainment related service of other information.Above-mentioned service is definitely not remote information Institute's functional full list of processing unit 30, but simply list some that telematics unit can provide Service.Moreover, it will be appreciated that at least some aforementioned modules can be with being stored in inside or outside telematics unit 30 The form of software instruction realizes that they can be the hardware component inside or outside telematics unit 30, or They can be each other or together with other system integrations of whole vehicle and/or shared by person, only quotes some possible feelings here Condition.In the case of the VSMs42 that module is implemented within outside telematics unit 30, they can utilize vehicle Bus 44 exchanges data and order with telematics unit.
GPS module 40 receives radio signal from the constellation 60 of gps satellite.According to these signals, module 40 can determine For providing the vehicle location of navigation and other positions related service to vehicle driver.Navigation information can be presented on display On 38 (or other displays in vehicle), or can be with oral presentation, such as the navigation that carries out when providing pavement branch sections navigation. Navigation Service can be provided using special vehicle mounted guidance module (it can be a part for GPS module 40), or can be with Some or all of navigation Services are completed by telematics unit 30, wherein, positional information is sent to remote location, with Just navigation map is provided for vehicle, map annotates (point of interest, restaurant etc.), route calculation etc..Positional information can be supplied to and exhale Center 20 or other remote computer systems (such as computer 18) are, for other purposes, such as fleet management.In addition, New or renewal map datum can download to GPS module 40 via telematics unit 30 from call center 20.
In addition to audio system 36 and GPS module 40, vehicle 12 can include the electronic hardware being located in whole vehicle Other vehicle system modules (VSM) 42 of component form, it generally receives from one or more sensors and inputted, and uses detection To input come perform diagnosis, monitoring, control, report and/or other functions.Each vehicle system module 42 is preferably by logical Letter bus 44 is connected to other vehicle system modules and telematics unit 30, and can be programmed to run vehicle System and subsystem diagnostic test.As example, a vehicle system module 42 can be each side for controlling power operation The engine control module (ECM) in face (such as fuel ignition and ignition timing etc.), another vehicle system module 42 can be The power train control module power assembly of the operation of one or more parts of regulation vehicle powertrain, and another Vehicular system Module 42 can be car body control module, and for controlling the various electric components on the whole vehicle, for example vehicle is electronic Door lock and headlight.According to embodiment, equipped with OBD (OBD) feature, it provides countless real-time engine control module Data, such as the real time data received from the various sensors including vehicle emission sensor, and provide standardization is Row DTC (DTC), so that technical staff quickly identifies and the failure in reconditioned vehicle.Such as those skilled in the art institute Understand, above-mentioned vehicle system module is only the example for some modules that can be used in vehicle 12, because being also likely to be perhaps More other modules.
Vehicle electronics 28 also include multiple vehicle user interfaces, and it provides offer to automotive occupant and/or received The device of information, including microphone 32, button 34, audio system 36 and visual displays 38.As it is used herein, term " vehicle user interface " broadly includes the electronic equipment of any suitable form, including on vehicle and enables vehicle user Enough hardware and software parts for being communicated with vehicle part or being communicated by vehicle part.Microphone 32 is to remote information Processing unit provides audio input so that driver or other occupants can provide voice command simultaneously by wireless carrier system 14 Perform hands free calls.Therefore, it can be connected at vehicle-mounted automatic speech using man-machine interface known in the art (HMI) technology Manage unit.Button 34 allows user to be manually entered into telematics unit 30 to initiate radiotelephone call and provide other Data, response or control input.Single button can be used for initiating urgent call, rather than carry out routine to call center 20 The help of service calls.Audio system 36 provides audio output to automotive occupant, and can be special autonomous system or A part for main vehicle audio system.According to particular embodiments illustrated herein, audio system 36 is operatively coupled to car Bus 44 and entertainment bus 46, and AM, FM and satelline radio, CD, DVD and other multimedia functions can be provided.Should Function can be combined or be provided independently with above- mentioned information entertainment modules.In some embodiments, audio system 36 can make Realized with Infotainment body.Infotainment body can include one or more computer processors, its can operate with The transceiver that Infotainment body is included together.Transceiver can be in its own and vehicle remote information process unit 30, intelligence Can between phone 57 or both between perform the short-distance wireless communications of data.Infotainment body can provide known in the art Audio and video information entertainment content.Visual displays 38 be preferably image display, such as instrument board on touch-screen or From the head-up display of windshield reflection, and it may be used to provide a variety of inputs and output function.It can also use various Other vehicle user interfaces, because the interface in Fig. 1 is only the example of a particular implementation.
Wireless carrier system 14 is preferably cell phone system, and it includes multiple celi towers 70 and (only shows one It is individual), one or more mobile switching centres (MSC) 72 and wireless carrier system 14 and land network 16 are attached institute Any other network components needed.Each celi towers 70 include sending and receiving antenna and base station, wherein from different small The base station of area's launching tower is direct or is connected to MSC 72 via intermediate equipment (such as base station controller).Cellular system 14 can be real What incumbent suitable communication technology, including such as analogue technique (such as AMPS) or newer digital technology (such as CDMA (ratios Such as, CDMA2000 or 1xEV-DO) or GSM/GPRS (such as 4GLTE)).As it will appreciated by a person of ordinary skill, cell is sent out The arrangement for penetrating tower/base station/MSC can be various, and can be used together with wireless system 14.For example, base station and Celi towers can be located at identical website, or they can away from each other, and each base station can be responsible for single subdistrict hair Tower is penetrated, or single base station can service each celi towers, and each base station may be coupled to single MSC, only lifts several The possible arrangement of kind.
Except using in addition to wireless carrier system 14, can using the different radio carrier system of satellite communication form come for Vehicle provides unidirectional or two-way communication.This can be complete using one or more telecommunication satellites 62 and up-link cell site 64 Into.One-way communication can be such as satellite radio services, wherein programming content (message, music etc.) is received by dispatching station 64, Packing uploads, and is subsequently sent to satellite 62, from satellite 62 to user's broadcast program.Two-way communication can be for example using satellite 62 The satellite phone service of the telephone communication come between relay vehicle 12 and station 64.If you are using, this satellite phone can To be used outside wireless carrier system 14 or in replacement wireless carrier system 14.
Land network 16 can be attached to one or more land line phones and wireless carrier system 14 is connected into calling The conventional continental rise communication network at center 20.For example, land network 16 can include PSTN (PSTN), for example use In that of offer hardwire phone, packet switched data communication and the Internet infrastructure.One or more of land network 16 Individual part can be by using standard wired network, optical fiber or other optical-fiber networks, cable network, power line, other wireless networks (such as WLAN (WLAN)) or provide the network of broadband wireless access (BWA) or its any combinations is realized.In addition, exhale Make center 20 to be attached via land network 16, but radiotelephone installation can be included so that it can be direct Communicated with the wireless network of such as wireless carrier system 14 etc.
Computer 18 can be one in the multiple computers accessed by special or public network (such as internet). Each such computer 18 can be used for one or more purposes, for example can pass through telematics unit 30 by vehicle The webserver accessed with wireless carrier 14.Other such addressable computers 18 can be for example:Service centre calculates Machine, wherein diagnostic message and other vehicle datas can be uploaded from vehicle via telematics unit 30;By car owner or its The client computer that his user uses, for accessing or receiving vehicle data or user preference or control vehicle is set or configured Function;By with vehicle 12 or call center 20 or both communicated to provide the 3rd of vehicle data or other information the Square thesaurus.Computer 18 can be also used for providing the Internet connection of such as DNS service etc, or be taken as network address Business device, IP address is distributed for vehicle 12 using the suitable agreements of DHCP or other.
Call center 20 is configured to vehicle electronics 28 and provides multiple different system back-end functions, and according to Exemplary embodiment illustrated herein, generally include one or more interchangers 80, server 82, database 84, Field Adviser 86 and automatic voice responsion system (VRS) 88, all these is all known in the art.These various call center's parts It is preferably coupled to each other via wired or wireless LAN 90.Interchanger 80 can be PBX (PBX) interchanger, Input signal is route by it so that voice transfer is generally sent to Field Adviser 86 by routine call, or uses VoIP is sent to automatic speed response system 88.Field Adviser's phone can also use VoIP as shown in phantom in Figure 1.Pass through The VoIP of interchanger 80 and the communication of other data (are not shown by the modem being connected between interchanger 80 and network 90 Go out) realize.Data transfer is sent to server 82 and/or database 84 by modem.Database 84 can store Account information, such as user authentication information, vehicle identifiers, profile record, behavior pattern and other relevant user informations.Data Transmission can also be carried out by wireless systems such as 802.11x, GPRS.Although illustrated embodiment has described as it and will adopted It is used together with Field Adviser 86 with someone call center 20, but it is to be understood that call center can alternatively use VRS 88 are used as Field Adviser, or can use the combination of VRS 88 and Field Adviser 86.
As described in the background section, text-to-speech (TTS) system is generally known to those skilled in the art.But It is that Fig. 2 shows the example of the improved tts system according to the present invention.According to embodiment, some or all of system 210 It can reside on Fig. 1 telematics unit 30 and be handled by it.According to the exemplary embodiment of replacement, tts system Some or all of 210 can reside on the computing device of the opening position (such as call center 20) away from vehicle 12 simultaneously Handled by it.For example, language model, acoustic model etc. can be stored in server 82 and/or database in call center 20 In one of 84 memory, and it is sent to vehicle remote information process unit 30 and is used for vehicle-mounted TTS processing.Similarly, may be used TTS software is handled with processor using a server 82 in call center 20.In other words, tts system 210 can be with Reside in telematics unit 30, or be distributed in any desired manner on whole call center 20 and vehicle 12.
System 210 can include one or more text sources 212 and memory (such as telematics memory 54), For storing the text from text source 212 and storing TTS software and data.System 210 can also include processor, such as far Journey message handler 52, to handle text and function together with following system module with memory.Preprocessor 214 is from text Source 212 receives text, and converts text to suitable word etc..Compositing Engine 216 turns the output from preprocessor 214 Appropriate linguistic unit is changed into, such as phrase, subordinate clause and/or sentence.The language of one or more stored records of speech database 218 Sound.Unit selection device 220 selects to best correspond to the unit of the storaged voice of the output of Compositing Engine 216 from database 218. One or more of selected unit of voice of storage is changed or adjusted to preprocessor 222.One or more language moulds Type 224 is used as the input of Compositing Engine 216, and one or more acoustic models 226 are used as the defeated of unit selectors 220 Enter.System 210 can also include being used for the acoustic interface 228 and loudspeaker that selected voice unit is converted to audio signal 230 (such as telematics audio systems), to convert audio signals into audible voice.System 210 can also wrap Include microphone, such as telematics microphone 32 and acoustic interface 232, speech digit turned into acoustic data, with Make the feedback to preprocessor 222.
Text source 212 can be in any suitable medium, and can include any suitable content.For example, text Source 212 can be one or more scanned documents, text or application data file, or any other suitable computer literal Part etc..Text source 212 can include it is to be synthesized for voice and be output to the word of text converter 214, numeral, symbol and/ Or punctuation mark.The text source of any suitable number amount and type can be used.
Text from text source 212 is converted into word, identifier etc. by preprocessor 214.For example, it is numeral in text In the case of form, numeral can be converted into corresponding word by preprocessor 214.In another example, it is mark in text Point symbol, with representing the appropriate capitalization emphasized with intonation or other spcial characters (such as cedilla), underscore or runic In the case of being emphasized, preprocessor 214 can be converted into Compositing Engine 216 and/or Unit selection device 220.
Compositing Engine 216 receives the output from text converter 214, and can be arranged in output and may include one Or multiple sentences, subordinate clause, phrase, word, sub- word and/or similar linguistic unit.Engine 216 can use language model 224 To assist the most probable arrangement of linguistic unit.The output of text converter 214 is being arranged in linguistic unit by language model 224 When rule, grammer and/or semanteme are provided.Model 224 can also define system 210 under any TTS patterns specified any The total domain for the time desired linguistic unit specified, and/or rule etc. can be provided, to provide the linguistic unit of which type And/or the rhythm can logically follow other types of linguistic unit and/or the rhythm, so as to form natural sound.Language list Member can be made up of voice equivalent, such as phone string etc., and can be the form of phoneme HMM.
Speech database 218 includes the pre-recorded voice from one or more people.Voice can include note in advance The sentence of record, subordinate clause, phrase, word, the sub- word etc. for word of pre-recording.Speech database 218 can also include the language with prerecording The associated data of sound, such as metadata, to identify the voice segments recorded used by Unit selection device 220.It can use The speech database of any suitable type and quantity.
Unit selection device 220 selects exporting compared with the speech data of storage from Compositing Engine 216 most The good storaged voice corresponding to Compositing Engine output.The voice selected by Unit selection device 220 can include pre-recorded sentence Son, subordinate clause, phrase, word, the sub- word for the word prerecorded and/or similar content.Selector 220 can use acoustic model 226 come help to compare and select storage voice most probable or it is optimal corresponding to candidate.Acoustic model 226 can be with selector 220 are used together, to compare and contrast the speech data of Compositing Engine output data and storage, assess difference between them or The size of similitude, and finally identify the voice data of best match and the corresponding record language of output using decision logic Sound.
Generally, the speech data of best match be with any of various technologies well known by persons skilled in the art institute really Speech data of the output with minimum dissimilarity or maximum probability fixed, with Compositing Engine 216.Such technology can wrap Dynamic time warping grader, artificial intelligence technology, neutral net, free phoneme recognizer and/or conceptual schema adaptation are included, Such as hidden Markov model (HMM) engine.Hidden Markov model engine known to those skilled in the art is multiple for producing TTS model candidate or hypothesis.Identify and select to represent Compositing Engine output in the acoustic character eventually through voice Consider to assume during the voice data that most probable is correctly explained.More specifically, the confidence calculated according to hidden Markov model The probability of angle value or the acoustic data sequence observed for one or the other linguistic unit specified is (for example, by using shellfish Leaf this theorem), the statistical model of " N is optimal " tabular form that hidden Markov model engine generation linguistic unit is assumed.
In one embodiment, the output from Unit selection device 220 can be directly passed to acoustic interface 228 or pass through Preprocessor 222, without post processing.In another embodiment, preprocessor 222 can be received from Unit selection device 220 Output, for further processing.
In any case, digital audio-frequency data is converted to simulated audio signal by acoustic interface 228.Interface 228 can be with It is digital-to-analogue conversion equipment, circuit and/or software etc..Loudspeaker 230 is electroacoustic transducer, and simulated audio signal is converted into using by it The audible voice received with microphone 32 in family.
Perform, and wrap in the computer program product that this method or part thereof can embody in computer-readable medium The instruction that can be used by the one or more processors of one or more computers of one or more systems is included, so that system is real Existing one or more method and steps.Computer program product can include by source code, object code, executable code or other One or more software programs of programmed instruction composition in the code of form;One or more firmware programs;Or hardware description Language (HDL) file;And any program related data.Data can include data structure, inquiry table or any other suitable The data of form.Programmed instruction can include program module, routine, program, object, component etc..Computer program can be one Performed on individual computer or on the multiple computers to communicate with one another.
Program can embody on a computer-readable medium, and it can be temporary with right and wrong and can include one or more Storage device, product etc..Exemplary computer-readable medium includes computer system memory, for example RAM (deposit by arbitrary access Reservoir), ROM (read-only storage);Semiconductor memory, such as EPROM (erasable programmable ROM), EEPROM (electric erasables Programming ROM), flash memory;Disk or CD or tape;And/or analog.Computer-readable medium can also arrive including computer The connection of computer, such as when data are transmitted or carried by network or other communication connections (wired, wireless or its combination) For when.Any combinations of examples detailed above are also included within the range of computer-readable medium.It will thus be appreciated that this method can With at least in part by being able to carry out any electronic article of the instruction corresponding with the one or more steps of disclosed method And/or equipment performs.
Turning now to reference to figure 3, the exemplary frame available for the ASR system 310 for realizing presently disclosed method is shown Structure.Generally, automotive occupant carries out interactive voice with automatic speech recognition system (ASR), following one or more basic to carry out Purpose:Training system, to understand the specific sound of automotive occupant;Store discrete voice, such as oral name label or oral control Word processed, such as numeral or keyword;Or the voice of identification automotive occupant, for any appropriate purpose, such as phonetic dialing, Menu navigation, transcription, service request, vehicle arrangement or functions of the equipments control etc..Generally, ASR system extracts from human speech Acoustic data, this acoustic data compared with the sub- word data stored and is contrasted, selection can be with other selected sons The appropriate sub- word of word connection, and the sub- word or word that export connection are dialled for post processing, such as dictation or transcription, address book Number, storage to memory, training ASR system model or adaptation parameter etc..
ASR system is generally known to those skilled in the art, and Fig. 3 illustrate only a specific exemplary ASR system 310.System 310 includes being used for the equipment for receiving such as voice of telematics microphone 32, and such as with modulus The acoustic interface 33 of the sound card of the telematics unit 30 of converter, acoustic data is turned to by speech digit.System 310 Also include the memory of such as telematics memory 54 etc, for storing acoustic data and storaged voice identification software And database, and such as processor of remote information processor 52 etc, to handle acoustic data.Processor connects with memory Worked together with lower module:One or more FEPs, preprocessor or pre-processor software module 312, are used for The acoustic data stream of voice is parsed into parameter to represent, such as acoustic feature;One or more decoders or decoder software mould Block 314, correspond to the digital sub- word or word output data that input speech utterance for decoding acoustic feature to produce;And one Individual or multiple back-end processors, preprocessor or post-processor software module 316, for being used for any suitable purpose Output data from decoder module 314.
System 310 can also receive voice from any other suitable audio-source 31, its can as shown by the solid line directly with Pre-processor software module 312 communicates, or communicates with 33 indirectly via acoustic interface.Audio-source 31 can be included for example Telephone audio source (such as voice-mail system) or any kind of other telephone services.
One or more modules or model may be used as the input of decoder module 314.First, grammer and/or dictionary mould Type 318 can provide rule, it is specified which word can logically follow other words to form effective sentence.Broadly Say, desired vocabulary is complete at any given time for system 310 defined in any given ASR mode for vocabulary or grammer Domain.For example, if system 310 is in the training mode for training order, then dictionary or syntactic model 318 can include The known and used all orders of system 310.In another example, if system 310 is in main menu mode, then activity Dictionary or syntactic model 318 can include the desired all main menu commands of system 310, for example call, dial, exiting, deleting, Inquiry etc..Second, acoustic model 320 helps to select the most probable son corresponding to the input from watermark pre-processor 312 Word or word.3rd, word model 322 and sentence/language model 324 provide by selected sub- word or word be put into word or Rule, grammer and/or semanteme in sentence linguistic context.In addition, sentence/language model 324 can be in any given ASR mode The total domain of desired sentence, and/or can provide and provide which sentence can be logically at any given time of definition system 310 Rule of other sentences etc. is followed, to form effective extended voice.
According to the exemplary embodiment of replacement, some or all of ASR system 310 can reside on the position away from vehicle 12 Put on the computing device at (such as call center 20) place and be processed.For example, syntactic model, acoustic model etc. can be stored in Server 82 in call center 20 and/or in the memory of one of database 84, and send it at vehicle remote information Unit 30 is managed for vehicle-mounted voice processing.Similarly, the processor of a server 82 in call center 20 can be used To handle speech recognition software.In other words, ASR system 310 can be resided in telematics unit 30, or with any Desired mode is distributed on whole call center 20 and vehicle 12, and/or is resided at call center 20.
First, acoustic data is extracted from human speech, wherein automotive occupant is spoken against microphone 32, and microphone 32 will Language is converted into electric signal, and these signals are sent into acoustic interface 33.Voice response element in microphone 32 is by occupant Speech utterance be captured as the change of air pressure, and language is converted into the corresponding of analog electrical signal (such as direct current or voltage) Change.Acoustic interface 33 receives analog electrical signal, and these analog electrical signals are sampled first so that the value of analog signal is discrete Moment is captured, and is then quantized so that the amplitude of analog signal is converted into continuous digital speech in each sampling instant Data flow.In other words, analog electrical signal is converted to digital electronic signal by acoustic interface 33.Numerical data is binary digit, They are buffered in telematics memory 54, are then handled by remote information processor 52, or can be initial It is processed when being received in real time by processor 52.
Second, continuous voice digital data stream is converted into the discrete series of parameters,acoustic by watermark pre-processor 312.More Specifically, processor 52 performs watermark pre-processor 312, so that digital voice data is segmented into such as duration as 10- 30ms overlapping voice or voiced frame.These frames correspond to the sub- words of acoustics such as syllable, half syllable, phone, double-tone, phoneme. Watermark pre-processor 312 also performs phonemic analysis, to extract parameters,acoustic, such as time-varying from the voice of occupant out of each frame Characteristic vector.Sounding in occupant's voice can be expressed as the sequence of these characteristic vectors.For example, and such as people in the art Known to member, characteristic vector can be extracted, and its can include for example can be by carrying out the Fourier transform and use of frame Sound spacing, Energy distribution, spectral properties and/or the cepstrum coefficient that cosine transform decorrelation acousto-optic is composed to obtain.Continue specific The voice of time, its voiced frame and corresponding parameter are cascaded into the unknown tone testing pattern that will be decoded.
3rd, computing device decoder module 314, to handle the input feature value of each test pattern.Decoder Module 314 also referred to as identifies engine or grader, and the voice of the known reference model using storage.With test pattern Equally, reference model is defined as the cascade of associated acoustic frame and corresponding parameter.The sub- word that decoder module 314 will identify is surveyed The acoustic feature vector of die trial formula compared with the sub- word reference model stored and contrasts, and assesses the difference or phase between them Like the size of property, and the sub- word of best match finally is selected as the word of accreditation using decision logic.Generally, best match Sub- word correspond to storage it is known reference pattern, have with it is well known by persons skilled in the art be used for analyze and identify sub- word The minimum dissimilarity of test pattern that is determined of any of various technologies or the sub- word of maximum probability.Such technology Dynamic time warping grader, artificial intelligence technology, neutral net, free phoneme recognizer and/or conceptual schema can be included Adaptation, such as hidden Markov model (HMM) engine.
Hidden Markov model engine known to those skilled in the art is used for the multiple speech recognition modelings for producing vocal input Assuming that.In view of these it is assumed that identifying and selecting to represent that the most probable of vocal input is correct eventually through the signature analysis of voice The identification output of decoding.More specifically, according to hidden Markov model calculate confidence value or specify one or the other Probability (for example, by using Bayes' theorem) sequence of the acoustic data sequence observed of sub- word, hidden Markov model Engine generates the statistical model of " N is optimal " tabular form of sub- word model hypothesis.
For the observation sequence of given acoustic feature vector, the identification of Bayes's hidden Markov model process corresponds to most Possible pronunciation or sub- word sequence optimal it is assumed that and its confidence value can depend on many factors, including with it is incoming The associated sound signal to noise ratio of acoustic data.Hidden Markov model can also include the statistical for being referred to as diagonal Gaussian Mixture Cloth, it is directed to the generation possibility fraction of each characteristic vector observed of every sub- word, and the fraction can be used for arranging again Sequence N most preferably assumes list.Hidden Markov model engine can also identify and the sub- word of preference pattern possibility highest scoring.
In a similar way, a series of single hidden Markov model for sub- words can be cascaded with establish it is single or The hidden Markov model of multiple words.Hereafter, it can generate and further assess single or multiple word reference models and phase The N-best of related parameter value.
In an example, speech recognition decoder 314 carrys out processing feature using appropriate acoustic model, grammer and algorithm Vector, to generate the N-best of reference model.As it is used herein, term reference pattern can be with model, waveform, mould Plate, abundant signal model, example, hypothesis or other types of with reference to mutually replacement.Reference model can include representing one Or the series of features of multiple words or sub- word vector, and can be based on specific speaker, locution and audible Environmental condition.It would be recognized by those skilled in the art that reference model can train life by the suitable reference model of ASR system Into, and store in memory.It will also be appreciated by the skilled artisan that the reference model of storage can be manipulated, wherein being based on Reference model trains the difference of the voice input signal between the actual use of ASR system to change the parameter of reference model Value.For example, training data or different acoustic conditions based on the limited quantity from different vehicle occupant, for a vehicle One group of reference model that occupant or some acoustic conditions are trained go for and save as different vehicle passenger or Another group of reference model of different acoustic conditions.In other words, reference model is not necessarily fixed, can be in the speech recognition phase Between be adjusted.
Explain and survey from memory access using lexicon grammar and any suitable decoder algorithm and acoustic model, processor Several reference models of die trial formula.For example, processor can generate N number of optimal vocabulary result or reference model list and Corresponding parameter value, and store onto memory.Exemplary parameter value can include the N-best of vocabulary and associated The confidence level fraction of each reference model in segment durations, likelihood score, signal to noise ratio (SNR) value etc..It can pass through The descending of parameter value size is arranged to be ranked up to the N-best of vocabulary.For example, the vocabulary with highest confidence score Reference model is the first optimal reference pattern, and the rest may be inferred.Once establish the sub- word of a string of identification, it is possible to be used to structure The input of word model 322 is built, and for constructing sentence from the input of language model 324.
Finally, post-processor software module 316 is from the Rreceive output data of decoder module 314, for any suitable Purpose.In an example, post-processor software module 316 can be from the N-best of single or multiple word reference models The voice of middle identification or one reference model of selection as identification.In another example, postprocessor module 316 can be used for Text or the numeral that acoustic data is converted into being used together with the other side of ASR system or other Vehicular systems.Another In individual example, postprocessor module 316 can be used for providing training feedback to decoder 314 or preprocessor 312.More specifically, Preprocessor 316 can be used for the acoustic model for training decoder module 314, or train for the suitable of watermark pre-processor 312 With parameter.
Method
Turning now to reference to figure 4, show identification and generate the method (400) of preferred emoticon.By wirelessly setting The emoticon of multiple selections is detected at standby place, and method 400 starts at step 410.In this embodiment, will be relative to intelligence Phone 57 describes wireless device.It is understood that the other types of wireless device that can send electronic information can be into Perform this method work(.These equipment include above-mentioned vehicle remote information process unit 30 or information entertainment main frame.Electronics Message generally includes the message for allowing the emoticon that will be transmitted between wireless device and remote destination to be inserted.Electronics disappears Breath can be email message, by character message (SMS) or messaging software application program, multi-media SMS (MMS) Or the short message that other similar message agreements are sent.
The emoticon identity of the equipment user included in electronic information can be monitored using software application, and The frequency of the selected each available emoticon of record., can be from the expression of acquiescence when equipment user creates electronic information One or more emoticons are selected in symbolic library, and are included into message.The emoticon storehouse of acquiescence can deliver It is loaded into before on wireless device to final terminal user, and it is available comprising hundreds of different emoticons.Each table Feelings symbol all depicts a kind of Art Design or image, and passes on thought or sensation according to its design.For example, one is described as The emoticon of " being so happy as to weep " depicts the smiling face that tear is carried beside facial eyes.Many other emoticon also be present Number, and had been left out detailed description thereof.Each emoticon can be related to identifying its hexadecimal code Connection.For example, emoticon " being so happy as to weep " can be represented with 1F600-1F64F codes.The technical specification of emoticon is by establishing Unicode consortium's definition of Unicode standard.Method 400 proceeds to step 420.
In step 420, it may be determined that select the frequency of each emoticon.Software application can be monitored and record is set Standby user selects each emoticon in default library to be contained in the number in electronic information.Software application can be with It is loaded on smart phone 57, wherein the disposal ability of smart phone 57 performs the function of software application, and records one section The selection frequency of emoticon in time included in the memory of phone 57.In one embodiment, this period can To be one month, but shorter or longer other time section can be used.It is remote the software application can also to be stored in vehicle In journey information process unit 30.Position regardless of software application, it can monitor using smart phone 57 or vehicle electronics The electronic information that equipment creates.Vehicle remote information process unit 30 and smart phone 57 can be via short range wireless communication protocols To transmit the data for representing selection emoticon.Therefore, when equipment user uses vehicle remote information process unit 30 or information When entertainment systems main frame creates message, when software application is stored in smart phone 57, the equipment based on vehicle can be to intelligence Phone 57 sends the identity and frequency of selection emoticon, and vice versa.Method 400 proceeds to step 430.
In step 430, based on the frequency for selecting each emoticon, identified from the emoticon of multiple selections and specify number The emoticon of amount.After selection a period of time of monitoring emoticon, software application can be most normal with identification equipment user The emoticon of selection.For example, software application may be configured to identify 30 or 40 measured within a period of time most The scene frequently selected.This method proceeds to step 440.
In step 440, conventional emoticon storehouse is created for emoticon.Smart phone 57 can establish one individually Storehouse, wherein including the information related to the emoticon identified.Conventional emoticon storehouse can include each emoticon Text based default description, one or more text based descriptions substituted of each emoticon and each expression The hex identifiers of symbol.The text based description of the replacement of each emoticon can be specified by user so that be used Family can input his or she description using display 59, to add them to conventional emoticon storehouse.Or can To provide the text based default description of replacement by inquiry.Investigation result can show as statistically valid number, And receive the description of each emoticon.Conventional emoticon storehouse can add the FAQs received from census of population and retouch State.Although describe emoticon identifier using hexadecimal code, but it is to be understood that other codes can be used Form, such as binary code.Conventional emoticon storehouse can be configured to and above-mentioned tts system 210 and ASR system 310 1 Rise and use.Method 400 proceeds to step 450.
In step 450, electronic information is initiated at smart phone 57, user selects emoticon from conventional emoticon storehouse Number to be included in electronic information.Equipment user can create electronic information, and include one or more expressions within the message Symbol.Equipment user can write message with oral, and its content can be by smart phone 57 or vehicle remote information process unit 30 Receive.For example, smart phone 57 can receive the voice from equipment user, and by short-distance wireless communication technology by voice The ASR system 310 being sent at vehicle 12.Or in another embodiment, vehicle 12 can receive via microphone 32 Voice from user, and handle voice using ASR system 310.Conventional emoticon storehouse can be loaded as by ASR system 310 Text sentence or language model 324.When handling voice to include emoticon in electronic information, commonly using emoticon storehouse can To provide higher efficiency.Conventional emoticon storehouse is not the text source for including defining with each possible emoticon, and It is that can be provided the emoticon occurred most frequently related information according to the behavior of particular device user.
As a part for regulation electronic message content, equipment user can recite him or she and want to include in the message Emoticon.ASR system 310 can be handled voice and be identified that oral emoticon describes using conventional emoticon storehouse.After Continuous example started above, equipment user can recite the text included in message text, and can also say word " happiness Pole and weep ".ASR system 310 can identify this description of emoticon, and emoticon " is so happy as to weep " together with word Insert in electronic information.
In another embodiment, wireless device can receive the input that user specifies, and it identifies that being included in electronics disappears Emoticon in breath.Conventional emoticon storehouse can be configured to the input that user specifies is associated with particular emotion symbol. When smart phone 57 or vehicle remote information process unit 30 detect that equipment user carries out the input that user specifies, with input Associated emoticon can be inserted into electronic information.The input that user specifies can be the face made by equipment user Portion's expression or the specific pattern drawn by equipment user using his or her finger.Smart phone 57 can be placed in training mode In, during training mode, its camera can be with the facial expression of recording equipment user.When equipment user establishes specific face During expression, user is also an option that specific emoticon with associated with specific facial expression.Then, smart phone 57 can To record associating for facial expression and emoticon in conventional emoticon storehouse.Different facial expressions can be distributed to conventional Different emoticons in emoticon storehouse.Then, smart phone 57 can terminate training mode and return to normal operating.Just Often during operation, equipment user can create electronic information and by the camera of smart phone towards his or her face.Then, use Face can be configured in the associated facial expression of the emoticon to be selected by family.Smart phone 57 will can connect from camera In the image of receipts and conventional emoticon storehouse compared with the image that emoticon is associated.Upon finding the match, intelligence electricity Associated emoticon can be inserted into electronic information by words 57.
The input specified for other users, different emoticons can be related to the specific pattern that equipment user draws Connection.Smart phone 57 can be placed in training mode, and during the training mode, its display 59 can record to be used by equipment The specific pattern that family is drawn.For example, equipment user can be drawn " X " along the surface of display 59 with finger.Work as equipment user When drawing specific pattern, user is also an option that particular emotion symbol with associated with pattern.Then, smart phone 57 can incite somebody to action Draw a design with the associated record of emoticon in conventional emoticon storehouse.Different patterns can distribute to conventional emoticon Different expressions in number storehouse.Then, smart phone 57 can terminate training mode and return to normal operating.In the normal operating phase Between, when equipment user thinks to insert particular emotion symbol in the message, he or she can create electronic information, then in display Pattern is followed the trail of on 59.The pattern that smart phone 57 can be detected is associated with emoticon with conventional emoticon storehouse Pattern be compared.Upon finding the match, associated emoticon can be inserted into electronic information by smart phone 57.
Conventional emoticon storehouse can also be used to generate the electronic information included in wireless device processes by tts system 210 In emoticon word picture.For example, wireless device can be by associated with emoticon and be also included within message In unique hexadecimal code identify emoticon that message includes.Can be by the hexadecimal code of identification and knowledge The hexadecimal code of emoticon that Chang Yong be in emoticon storehouse is compared.Tts system 210 can by with matching The associated description of emoticon and generate voice;Conventional emoticon storehouse may be used as text source 212, and expression is represented to produce The voice of symbol.Then, method 400 terminates.
It should be appreciated that foregoing teachings are the descriptions to one or more embodiments of the invention.The invention is not restricted to herein Disclosed specific embodiment, but be limited only by the following claims.In addition, the statement included in being described above be related to it is specific Embodiment, and the limitation of the definition of the term to being used in the scope of the present invention or claims is not necessarily to be construed as, remove The non-term or word are particularly limited to above.The various change of various other embodiments and the disclosed embodiments and Modification will be apparent for those skilled in the art.All these other embodiments, deformation and modification are intended to fall under institute In the range of attached claim.
In the specification and claims term used " such as ", " such as ", " for example ", " such as " with it is " similar In " and verb " comprising ", " having ", " include " and other verb forms, when them and one or more elements or other items When purpose list is used together, it all should be considered as open, it is meant that list is not considered as eliminating other add ons Or project.Other terms are explained using its widest rationally implication, unless it is used for the context that requirement has different explanations In.

Claims (10)

1. a kind of method for identifying and generating preferred emoticon, it comprises the following steps:
(a) emoticon of multiple selections is detected in wireless devices;
(b) frequency of each emoticon of selection is determined;
(c) based on the frequency for selecting each emoticon, the table of identification specified quantity from the emoticon of the multiple selection Feelings symbol;And
(d) emoticon for identification in step (c) creates conventional emoticon storehouse.
2. according to the method for claim 1, wherein the wireless device includes vehicle remote information process unit.
3. according to the method for claim 1, wherein the wireless device includes smart phone.
4. according to the method for claim 1, it further comprises knowing the conventional emoticon storehouse as automatic speech The step of model of not (ASR) system is loaded.
5. according to the method for claim 1, it further comprises the steps:The description that one or more users are specified It is associated with each emoticon, and those descriptions are stored in the conventional emoticon storehouse.
6. a kind of method for identifying and generating preferred emoticon, it comprises the following steps:
(a) electronic information is initiated in wireless device;
(b) voice for the emoticon that description is included in the electronic information is received;
(c) by the voice of the reception compared with the emoticon description being stored in conventional emoticon storehouse;
(d) emoticon is relatively identified according to described;And
(e) emoticon is inserted in the electronic information.
7. according to the method for claim 6, it further comprises knowing the conventional emoticon storehouse as automatic speech The step of model of not (ASR) system is loaded.
8. according to the method for claim 6, wherein emoticon description includes the description that user specifies and acquiescence is retouched State.
9. according to the method for claim 6, wherein emoticon description includes the description generated by inquiry.
10. a kind of method for identifying and generating preferred emoticon, it comprises the following steps:
(a) electronic information is initiated in wireless device;
(b) input that identification is specified to contain the user of the emoticon in the electronic information is received;
(c) input for specifying the user of the reception is specified with the user prestored being stored in conventional emoticon storehouse Input and the association of emoticon be compared;
(d) emoticon is relatively identified according to described;And
(e) emoticon is inserted in the electronic information.
CN201710813706.3A 2016-09-14 2017-09-11 It is preferred that the identification and generation of emoticon Pending CN107819929A (en)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
US15/265522 2016-09-14
US15/265,522 US20180074661A1 (en) 2016-09-14 2016-09-14 Preferred emoji identification and generation

Publications (1)

Publication Number Publication Date
CN107819929A true CN107819929A (en) 2018-03-20

Family

ID=61247271

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201710813706.3A Pending CN107819929A (en) 2016-09-14 2017-09-11 It is preferred that the identification and generation of emoticon

Country Status (3)

Country Link
US (1) US20180074661A1 (en)
CN (1) CN107819929A (en)
DE (1) DE102017121059A1 (en)

Cited By (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN108809817A (en) * 2018-07-06 2018-11-13 上海博泰悦臻电子设备制造有限公司 Vehicle, vehicle device equipment, Cloud Server and the communication means of vehicle-mounted instant chat
CN109977925A (en) * 2019-04-22 2019-07-05 北京字节跳动网络技术有限公司 Expression determines method, apparatus and electronic equipment
CN112118359A (en) * 2020-09-22 2020-12-22 珠海格力电器股份有限公司 Text information processing method and device, electronic equipment and electronic system

Families Citing this family (11)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US11355250B2 (en) 2016-03-30 2022-06-07 Livecare Corp. Gateway device and system for service to individuals with diminishing dexterity and neurological physiological functionality
US20190261153A1 (en) * 2016-03-30 2019-08-22 Livecare Corp. Emergency response device and system for service to individuals with diminishing dexterity and neurological physiological functionality
US10586623B2 (en) 2016-03-30 2020-03-10 Livecare Corp. Patient care device and system for service to individuals with diminishing dexterity and neurological physiological functionality
US10593087B2 (en) 2017-10-23 2020-03-17 Paypal, Inc. System and method for generating emoji mashups with machine learning
US11496425B1 (en) * 2018-05-10 2022-11-08 Whatsapp Llc Modifying message content based on user preferences
DE102019202662B4 (en) 2019-02-27 2021-01-14 Volkswagen Aktiengesellschaft Method for checking the functionality of an emergency call device of a motor vehicle and motor vehicle for carrying out the method
US10825449B1 (en) * 2019-09-27 2020-11-03 CrowdAround Inc. Systems and methods for analyzing a characteristic of a communication using disjoint classification models for parsing and evaluation of the communication
WO2021076166A1 (en) * 2019-10-15 2021-04-22 Google Llc Voice-controlled entry of content into graphical user interfaces
CN112306224A (en) * 2020-04-30 2021-02-02 北京字节跳动网络技术有限公司 Man-machine interaction method and device and electronic equipment
US11657558B2 (en) 2021-09-16 2023-05-23 International Business Machines Corporation Context-based personalized communication presentation
CN114092608B (en) * 2021-11-17 2023-06-13 广州博冠信息科技有限公司 Expression processing method and device, computer readable storage medium and electronic equipment

Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN101146056A (en) * 2007-09-24 2008-03-19 腾讯科技(深圳)有限公司 A display method and system for emotion icons
CN103294220A (en) * 2012-02-28 2013-09-11 联想(北京)有限公司 Input method and device
CN103809766A (en) * 2012-11-06 2014-05-21 夏普株式会社 Method and electronic device for converting characters into emotion icons
CN104063683A (en) * 2014-06-06 2014-09-24 北京搜狗科技发展有限公司 Expression input method and device based on face identification
US9043196B1 (en) * 2014-07-07 2015-05-26 Machine Zone, Inc. Systems and methods for identifying and suggesting emoticons

Family Cites Families (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US8494859B2 (en) * 2002-10-15 2013-07-23 Gh, Llc Universal processing system and methods for production of outputs accessible by people with disabilities
EP1603116A1 (en) * 2003-02-19 2005-12-07 Matsushita Electric Industrial Co., Ltd. Speech recognition device and speech recognition method
US20120113011A1 (en) * 2009-03-20 2012-05-10 Genqing Wu Ime text entry assistance
KR101419764B1 (en) * 2013-06-07 2014-07-17 정영민 Mobile terminal control method for voice emoticon
KR102063766B1 (en) * 2013-09-17 2020-01-08 엘지전자 주식회사 Mobile terminal and control method for the mobile terminal
US9576175B2 (en) * 2014-05-16 2017-02-21 Verizon Patent And Licensing Inc. Generating emoticons based on an image of a face

Patent Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN101146056A (en) * 2007-09-24 2008-03-19 腾讯科技(深圳)有限公司 A display method and system for emotion icons
CN103294220A (en) * 2012-02-28 2013-09-11 联想(北京)有限公司 Input method and device
CN103809766A (en) * 2012-11-06 2014-05-21 夏普株式会社 Method and electronic device for converting characters into emotion icons
CN104063683A (en) * 2014-06-06 2014-09-24 北京搜狗科技发展有限公司 Expression input method and device based on face identification
US9043196B1 (en) * 2014-07-07 2015-05-26 Machine Zone, Inc. Systems and methods for identifying and suggesting emoticons

Cited By (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN108809817A (en) * 2018-07-06 2018-11-13 上海博泰悦臻电子设备制造有限公司 Vehicle, vehicle device equipment, Cloud Server and the communication means of vehicle-mounted instant chat
CN109977925A (en) * 2019-04-22 2019-07-05 北京字节跳动网络技术有限公司 Expression determines method, apparatus and electronic equipment
CN112118359A (en) * 2020-09-22 2020-12-22 珠海格力电器股份有限公司 Text information processing method and device, electronic equipment and electronic system

Also Published As

Publication number Publication date
DE102017121059A1 (en) 2018-03-15
US20180074661A1 (en) 2018-03-15

Similar Documents

Publication Publication Date Title
CN107819929A (en) It is preferred that the identification and generation of emoticon
US9202465B2 (en) Speech recognition dependent on text message content
CN101071564B (en) Distinguishing out-of-vocabulary speech from in-vocabulary speech
US9570066B2 (en) Sender-responsive text-to-speech processing
CN101272416B (en) Voice dialing using a rejection reference
US8639508B2 (en) User-specific confidence thresholds for speech recognition
CN101354887B (en) Ambient noise injection method for use in speech recognition
US10255913B2 (en) Automatic speech recognition for disfluent speech
US8438028B2 (en) Nametag confusability determination
CN106816149A (en) The priorization content loading of vehicle automatic speech recognition system
CN102097096B (en) Using pitch during speech recognition post-processing to improve recognition accuracy
CN103124318B (en) Start the method for public conference calling
US8756062B2 (en) Male acoustic model adaptation based on language-independent female speech data
US9997155B2 (en) Adapting a speech system to user pronunciation
US20190122661A1 (en) System and method to detect cues in conversational speech
US8762151B2 (en) Speech recognition for premature enunciation
US20150056951A1 (en) Vehicle telematics unit and method of operating the same
CN109671424B (en) Responsive activation of vehicle features
US9911408B2 (en) Dynamic speech system tuning
CN105609109A (en) Hybridized automatic speech recognition
CN107818788A (en) Remote speech identification on vehicle
CN108447488A (en) Enhance voice recognition tasks to complete
US20120197643A1 (en) Mapping obstruent speech energy to lower frequencies
US8050928B2 (en) Speech to DTMF generation
US20160267901A1 (en) User-modified speech output in a vehicle

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
WD01 Invention patent application deemed withdrawn after publication

Application publication date: 20180320

WD01 Invention patent application deemed withdrawn after publication