US9665567B2 - Suggesting emoji characters based on current contextual emotional state of user - Google Patents
Suggesting emoji characters based on current contextual emotional state of user Download PDFInfo
- Publication number
- US9665567B2 US9665567B2 US14/859,959 US201514859959A US9665567B2 US 9665567 B2 US9665567 B2 US 9665567B2 US 201514859959 A US201514859959 A US 201514859959A US 9665567 B2 US9665567 B2 US 9665567B2
- Authority
- US
- United States
- Prior art keywords
- user
- emotional state
- emotional
- current
- current contextual
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Expired - Fee Related
Links
- 230000002996 emotional effect Effects 0.000 title claims abstract description 262
- 238000000034 method Methods 0.000 claims description 47
- 230000008451 emotion Effects 0.000 claims description 22
- 230000001815 facial effect Effects 0.000 claims description 10
- 238000004590 computer program Methods 0.000 claims description 7
- 230000008569 process Effects 0.000 description 16
- 238000010586 diagram Methods 0.000 description 11
- 238000012545 processing Methods 0.000 description 8
- 238000013459 approach Methods 0.000 description 7
- 230000006870 function Effects 0.000 description 7
- 230000006855 networking Effects 0.000 description 6
- 230000005540 biological transmission Effects 0.000 description 4
- 238000001514 detection method Methods 0.000 description 4
- 238000004458 analytical method Methods 0.000 description 3
- 238000013473 artificial intelligence Methods 0.000 description 3
- 238000005516 engineering process Methods 0.000 description 3
- 230000004044 response Effects 0.000 description 3
- 238000003491 array Methods 0.000 description 2
- 238000004891 communication Methods 0.000 description 2
- 230000000994 depressogenic effect Effects 0.000 description 2
- 238000005259 measurement Methods 0.000 description 2
- 230000007246 mechanism Effects 0.000 description 2
- 230000003287 optical effect Effects 0.000 description 2
- 230000001902 propagating effect Effects 0.000 description 2
- 238000011160 research Methods 0.000 description 2
- 238000012552 review Methods 0.000 description 2
- RYGMFSIKBFXOCR-UHFFFAOYSA-N Copper Chemical compound [Cu] RYGMFSIKBFXOCR-UHFFFAOYSA-N 0.000 description 1
- 241000369592 Platycephalus richardsoni Species 0.000 description 1
- 230000006399 behavior Effects 0.000 description 1
- 230000003542 behavioural effect Effects 0.000 description 1
- 230000001149 cognitive effect Effects 0.000 description 1
- 229910052802 copper Inorganic materials 0.000 description 1
- 239000010949 copper Substances 0.000 description 1
- 230000001419 dependent effect Effects 0.000 description 1
- 230000008921 facial expression Effects 0.000 description 1
- 239000000835 fiber Substances 0.000 description 1
- 230000014509 gene expression Effects 0.000 description 1
- 230000006698 induction Effects 0.000 description 1
- 238000009434 installation Methods 0.000 description 1
- 238000004519 manufacturing process Methods 0.000 description 1
- 238000013507 mapping Methods 0.000 description 1
- 238000012544 monitoring process Methods 0.000 description 1
- 230000036651 mood Effects 0.000 description 1
- 230000000926 neurological effect Effects 0.000 description 1
- 238000003909 pattern recognition Methods 0.000 description 1
- 230000029058 respiratory gaseous exchange Effects 0.000 description 1
- 239000004065 semiconductor Substances 0.000 description 1
- 230000005236 sound signal Effects 0.000 description 1
- 230000003068 static effect Effects 0.000 description 1
- 230000001960 triggered effect Effects 0.000 description 1
Images
Classifications
-
- G06F17/279—
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F40/00—Handling natural language data
- G06F40/30—Semantic analysis
- G06F40/35—Discourse or dialogue representation
-
- G06F17/21—
-
- G06F17/24—
-
- G06F17/2765—
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
- G06F3/011—Arrangements for interaction with the human body, e.g. for user immersion in virtual reality
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
- G06F3/02—Input arrangements using manually operated switches, e.g. using keyboards or dials
- G06F3/023—Arrangements for converting discrete items of information into a coded form, e.g. arrangements for interpreting keyboard generated codes as alphanumeric codes, operand codes or instruction codes
- G06F3/0233—Character input methods
- G06F3/0237—Character input methods using prediction or retrieval techniques
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
- G06F3/048—Interaction techniques based on graphical user interfaces [GUI]
- G06F3/0484—Interaction techniques based on graphical user interfaces [GUI] for the control of specific functions or operations, e.g. selecting or manipulating an object, an image or a displayed text element, setting a parameter value or selecting a range
- G06F3/04842—Selection of displayed objects or displayed text elements
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/16—Sound input; Sound output
- G06F3/167—Audio in a user interface, e.g. using voice commands for navigating, audio feedback
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F40/00—Handling natural language data
- G06F40/10—Text processing
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F40/00—Handling natural language data
- G06F40/10—Text processing
- G06F40/166—Editing, e.g. inserting or deleting
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F40/00—Handling natural language data
- G06F40/20—Natural language analysis
- G06F40/279—Recognition of textual entities
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F40/00—Handling natural language data
- G06F40/30—Semantic analysis
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V40/00—Recognition of biometric, human-related or animal-related patterns in image or video data
- G06V40/10—Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
- G06V40/16—Human faces, e.g. facial parts, sketches or expressions
- G06V40/174—Facial expression recognition
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
- G10L13/00—Speech synthesis; Text to speech systems
- G10L13/02—Methods for producing synthetic speech; Speech synthesisers
- G10L13/033—Voice editing, e.g. manipulating the voice of the synthesiser
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
- G10L15/00—Speech recognition
- G10L15/08—Speech classification or search
- G10L15/18—Speech classification or search using natural language modelling
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F2203/00—Indexing scheme relating to G06F3/00 - G06F3/048
- G06F2203/01—Indexing scheme relating to G06F3/01
- G06F2203/011—Emotion or mood input determined on the basis of sensed human body parameters such as pulse, heart rate or beat, temperature of skin, facial expressions, iris, voice pitch, brain activity patterns
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V40/00—Recognition of biometric, human-related or animal-related patterns in image or video data
- G06V40/10—Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
- G06V40/15—Biometric patterns based on physiological signals, e.g. heartbeat, blood flow
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
- G10L25/00—Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00
- G10L25/48—Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00 specially adapted for particular use
- G10L25/51—Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00 specially adapted for particular use for comparison or discrimination
- G10L25/63—Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00 specially adapted for particular use for comparison or discrimination for estimating an emotional state
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04M—TELEPHONIC COMMUNICATION
- H04M1/00—Substation equipment, e.g. for use by subscribers
- H04M1/72—Mobile telephones; Cordless telephones, i.e. devices for establishing wireless links to base stations without route selection
- H04M1/724—User interfaces specially adapted for cordless or mobile telephones
- H04M1/72403—User interfaces specially adapted for cordless or mobile telephones with means for local support of applications that increase the functionality
- H04M1/7243—User interfaces specially adapted for cordless or mobile telephones with means for local support of applications that increase the functionality with interactive means for internal management of messages
- H04M1/72436—User interfaces specially adapted for cordless or mobile telephones with means for local support of applications that increase the functionality with interactive means for internal management of messages for text messaging, e.g. short messaging services [SMS] or e-mails
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04M—TELEPHONIC COMMUNICATION
- H04M2250/00—Details of telephonic subscriber devices
- H04M2250/12—Details of telephonic subscriber devices including a sensor for measuring a physical value, e.g. temperature or motion
Definitions
- emoji characters have become a popular way by which users enhance text messages and posts, within private text messages and within social networking posts.
- An emoji is literally a picture character or pictograph, and is a single character with a unique code point value as part of a text string that can provide additional meaning to text or provide contextual information to the text to assist in interpretation of the text, among other purposes.
- a user may switch from an alphabetic keyboard to an emoji keyboard and select the desired emoji.
- An example method includes determining, by a computing device, a current perceived emotional state of a user.
- the method includes determining, by the computing device, a current contextual emotional state of the user based on text inputted by the user and based on the current perceived emotional state of the user.
- the method includes, from emoji characters mapped to different contextual emotional states, determining, by the computing device, one or more selected emoji characters that are mapped to the current contextual emotional state of the user.
- the method includes suggesting, by the computing device, the selected emoji characters to the user to add to the text inputted by the user.
- An example computer program product includes a computer readable storage medium having stored thereon program instructions.
- the instructions are executable by a computing device to cause the computing device to determine a current perceived emotional state of a user, and determine a semantic content of text inputted by the user.
- the instructions are executable by the computing device to cause the computing device to determine a current contextual emotional state of the user based on the semantic content of the text inputted by the user and based on the current perceived emotional state of the user.
- the instructions are executable by the computing device to cause the computing device to, from emoji characters mapped to different contextual emotional states, determine one or more selected emoji characters that are mapped to the current contextual emotional state of the user, and to suggest the selected emoji characters to the user to add to the text inputted by the user.
- An example computing device includes a processor, a memory, and program instructions stored in the memory and executable by the processor.
- the instructions are executable by the processor to determine a current perceived emotional state of a user, determine semantic content of text inputted by the user, and determine an ongoing context in which the text inputted by the user pertains.
- the instructions are executable by the processor to determine a current contextual emotional state of the user based on the ongoing context in which the text inputted by the user pertains, based on the semantic content of the text inputted by the user, and based on the current perceived emotional state of the user.
- the instructions are executable by the processor to, from emoji characters mapped to different contextual emotional states, determine one or more selected emoji characters that are mapped to the current contextual emotional state of the user, and to suggest the selected emoji characters to the user to add to the text inputted by the user.
- FIG. 1 is a diagram of a process flow by which emoji characters are suggested to the user.
- FIGS. 2 and 3 are flowcharts of example methods to ascertain a current contextual emotional state of a user in different implementations.
- FIG. 4 is a flowchart of an example computing device.
- emoji characters are single picture characters that have become popular to add to text messages and posts within private text messages and social network posts.
- text messages include those transmitted over the short message service (SMS) that is ubiquitous among mobile phone operators, as well as those transmitted over proprietary message services that require a particular type of smartphone or installation of a particular type of computer program or “app.”
- SMS short message service
- social networking posts include those posted using social networking services like those run by Facebook, Inc., of Menlo Park, Calif., and Twitter, Inc., of San Francisco, Calif.
- a difficulty with using emoji characters is that there are multitudes of different such characters, making selection of an appropriate character difficult to accomplish, particularly on a mobile computing device like a smartphone. Many users also have difficulty knowing what particular emoji characters, a problem borne at least in part in the initial usage of such characters in a country, Japan, which culturally associates certain meanings to certain characters in ways that people in other countries do not. As a result of these downsides, a good number of users use a limited number of emoji characters, if any at all.
- emoji characters are suggested to the user.
- the user can select a desired emoji character or characters to add to his or her message or post.
- entry is more easily accomplished by simply selecting one of the suggested characters, instead of having to scroll over multiple pages or screens of emoji characters to locate the most appropriate character.
- Users may find themselves as a result using a richer set of emoji characters, since they will know that the emoji characters suggested to them are appropriate to add. That is, the users will be able to intrinsically discern the meaning of emoji characters because just relevant and appropriate such characters are suggested for adding to their text messages and posts.
- FIG. 1 shows an example process flow 100 by which appropriate emoji characters can be suggested for selection by the user.
- Ovals and rounded rectangles in FIG. 1 indicate data that is not generated within the process flow 100 itself, but rather are received within the flow 100 .
- Rectangles indicate data that is generated or determined within the process flow 100 .
- Arrows indicate the order of process flow and the inputs on which basis the data in the rectangles is determined.
- Parallelograms indicate parts, steps, or acts that are also performed within the process flow 100 .
- a user inputs text 102 in an ongoing context 104 .
- the text 102 may be a text message within a text messaging service or a post within a social networking service.
- the ongoing context 104 is the context in which the text 102 has been input, and may include other texts, by the same and other users, that have been entered. That is, the context 104 is the context to which the text 102 pertains.
- the text 102 may be entered as part of an ongoing conversation between the user and one or more other users, where the prior texts by the same and the other users form the ongoing context 104 .
- the text 102 may be a comment, response, or reply to a post within a social networking service that has other comments, responses, or replies that have already been entered, where the original post and the other comments, responses, or replies form the ongoing context 104 .
- the emotional inputs 106 can include one or more of the following.
- the emotional inputs 106 can include biometric information of the user that entered the text 104 , as detected by a biometric sensing device, such as a heart rate sensor, a breathing rate sensor, a perspiration sensor, and so on. Such biometric information is thus physiological measurements of the user on which basis an emotional state of the user can be determined.
- the emotional inputs 106 can include a facial image of the user as detected by a camera device, where the facial image can exhibit frowns, furrowed brows, smiles, smirks, and so on, on which basis an emotional state of the user can be determined.
- the emotional inputs 106 can include recorded audio of the user as detected by an audio recorded device, where the audio can exhibit laughter, yelling, snickering, and so on, on which basis an emotional state of the user can be determined.
- the emotional inputs 106 can further include the user-inputted text 102 itself.
- a current perceived emotional state 108 which is a first emotional state, of the user is determined from the emotional inputs 106 .
- the emotional state 108 may not reflect the semantic content, or meaning of the user-inputted text 102 . Rather, the emotional state 108 reflects a perceived emotional state from the user's biometric information, facial image, recorded audio, and the text 102 (the latter in one implementation without reflecting the semantic content of the text 102 ).
- Determining the emotional state 108 from biometric information of the user can be achieved in a number of different ways. Several approaches are described, for instance, in C. Chandler et al., “Biometric Measurement of Human Emotions,” TECHNIA—International Journal of Computing Science and Communication Technologies, vol. 4, no. 2, January 2012; C. Conati, “A Study on Using Biometric Sensors for Monitoring User Emotions in Educational Games,” published at http://www.cs.ubc.ca/ ⁇ conati/my-papers/um03affect-camera-conati.pdf, and published no later than 2004; and M. H.
- Determining the emotional state 108 from facial images of the user can be achieved in a number of different ways. Several approaches are described, for instance, in R. Adolphs, “Recognizing Emotion from Facial Expressions: Psychological and Neurological Mechanisms,” Behavioral and Cognitive Neuroscience Reviews, vol. 1, no. 1, March 2002; G. A. Ramirez et al., “Color Analysis of Facial Skin: Detection of Emotional State,” Computer Vision and Pattern Recognition Workshops (CVPRW), 2014 IEEE Conference, Jun. 23-28, 2014; and Surbhi, “A Face Identification Technique for Human Facial Image,” International Journal of Computer Science and Information Technologies, vol. 3(6), 2012.
- Determining the emotional state 108 from recorded audio (e.g., speech) of the user can be achieved in a number of different ways.
- Several approaches are described, for instance, in E. S. Erdem et al., “Efficient Recognition of Human Emotional States from Audio Signals,” 2014 IEEE International Symposium on Multimedia; C. Cullen et al, “Generation of High Quality Audio Natural Emotional Speech Corpus using Task Based Mood Induction,” International Conference on Multidisciplinary Information Sciences and Technologies Extremadura, Merida, Spain, Oct. 25-28, 2006; and T. S.
- Determining the emotional state 108 from text entered by the user can be achieved in a number of different ways. Several approaches are described, for instance, in S. N. Shivare, “Emotion Detection from Text,” published at http://arxiv.org/ftp/arxiv/papers/1205/1205.4944.pdf, 2012; C. Strapparava et al., “Learning to Identify Emotions in Text,” SAC'08, Mar. 16-20, 2008; and “Recognizing Emotions and Sentiments in Text,” S. M. Kim, Masters Thesis at the University of Sydney, April 2011.
- the semantic content 110 of the text 102 is determined.
- the semantic content 110 of the text 102 can include the meaning of the text 102 in a manner in which computational processing can understand. Determining the semantic content 110 of the text 102 can be achieved in a number of different ways. Several approaches are described, for instance, in the references noted in the previous paragraph, as well as in P. W. Foltz, “Latent Semantic Analysis for Text-Based Research,” Behavior Research Methods, Instruments and Computers, vol. 28(2), 1996; M. Yazdani et al., “Computing Text Semantic Relatedness . . . ,” Artificial Intelligence, Nov. 19, 2012; and R. Mihalcea et al., “Corpus-based and Knowledge-based Measures of Text Semantic Similarity,” American Association for Artificial Intelligence, 2006.
- a second emotional state 112 of the user is determined from the semantic content 110 of the text 102 .
- This emotional state 112 of the user is thus another measure of the user's emotional state, in addition to the first emotional state 108 determined from the emotional inputs 108 .
- Determining the second emotional state 112 from the semantic content 110 of the text 102 can be achieved in a number of different ways. Several approaches are described, for instance, in the references noted with respect to determining the first emotional state 108 from text, as well as in S. Aman, “Recognizing Emotions in Text,” Masters Thesis at the University of Ottawa, 2007; S. Aman et al., “Identifying Expressions of Emotion in Text,” TSD 2007; and C. Kalyan et al., “Detecting emotional scenes using Semantic Analysis on Subtitles,” Stanford University CS224N Final Project, Jun. 4, 2009.
- the semantic content 110 of the text 102 is first explicitly determined from the text 102 , and then the second emotional state 112 is determined from this semantic content 110 .
- the second emotional state 112 is determined from the text 102 , where the semantic content 110 may be implicitly determined as part of this process.
- the second emotional state 112 may be the same as the first emotional state 108 if the latter is generated from the text 102 , and therefore in this implementation the first emotional state 108 is generated from emotional inputs 106 other than the text 102 .
- a third emotional state 114 of the user may be determined from the ongoing context 104 within which the text 102 has been input by the user. This emotional state 114 of the user is thus another measure of the user's emotional state, in addition to the first emotional state 108 and the second emotional state 112 of the user. Determining the third emotional state 114 from the ongoing context 104 can be achieved in a number of different ways. For instance, the same methodology used to determine the second emotional state 112 may be used to determine the third emotional state 114 , but where the entire corpus of text entered by the user (and in one implementation, other users' entered text in the same context) is considered. Other approaches that can be employed are described in, for instance, A. H.
- the current contextual emotional state 116 of the user is determined.
- the emotional state 116 can be considered as an aggregate of the emotional states 108 , 112 , and 114 in one implementation, to most accurately reflect the actual emotional state of the user while he or she inputted the text 102 within the ongoing context 104 .
- Different mechanisms by which the emotional state 116 can be determined can be employed.
- the emotional state 116 is one of a predetermined number of emotional states, such as happy, sad, angry, depressed, ironic, and so on.
- the emotional states 108 , 112 , and 114 may each be determined as weighting of one or more different emotions. Therefore, these weights are combined to determine the current contextual emotional state 116 . If the combination of the weights yields an emotion with a weight greater than a threshold, then it can be concluded that a current contextual emotion state 116 has been successfully determined.
- the emotional state 108 may be determined as 0.8 happy and 0.2 sad.
- the emotional state 112 may be determined as 0.7 happy and 0.1 angry.
- the emotional state 114 may be determined as inapposite—that is, no significant emotion may have been determined as part of the state 114 . Therefore, the current contextual emotional state 116 is the summation of these three states 108 , 112 , and 114 , where the state 114 is not effectively considered since it is inapposite.
- the current contextual emotional state 116 is initially 1.5 happy, 0.2 sad, and 0.1 angry. If 1.5 is greater than a threshold, then the current contextual emotional state 116 is determined as happy.
- one or more selected emoji characters 118 are determined that map to the current contextual emotional state 116 .
- this is achieved by referencing a predetermined database 120 that maps each of a number of emoji characters to a corresponding emotional state. Therefore, for a given emotional state, there may be more than one emoji character mapped thereto. Different emotional states can be differing degrees of the same base emotion, such as happy, happier, and happiest, where each such state has one or more emoji characters mapped thereto.
- the mapping of the predetermined database 120 can be defined, and redefined as desired, by users and/or vendors. Furthermore, the database 120 can vary by the country in which a user or the recipient of the user's text is located, since the meaning of emoji characters can differ by culture.
- the selected emoji characters 118 that have been determined are suggested to the user in part 121 .
- the emoji characters 118 are selected emoji characters in that they are particular emoji characters that have been determined that map to the current contextual emotional state 116 , as opposed to being selected by the user.
- a small bubble, window, or other graphical user interface (GUI) element may be displayed that shows the selected emoji characters 118 for selection by the user.
- the GUI element may be automatically displayed, or may be displayed by the user selecting an option corresponding to showing emoji characters. For example, if the user selects a virtual keyboard corresponding to emoji characters, the selected emoji characters 118 may be displayed towards the beginning of the list of such characters, either before or immediately after emoji characters that the user previously entered.
- the process flow 100 can be repeated each time a user enters text. For example, each time a user completes a sentence, ending in punctuation such as a period, question mark, or exclamation point, the process flow 100 can be performed. As another example, just prior to the user posting or sending the text 102 the process flow 100 can be performed. As a third example, after the user has entered the text 102 and has paused for a predetermined length of time, it may be concluded that the user has finished entering the text 102 such that the process flow 100 is performed upon detection of this pause. As a final example, the process flow 100 can be manually triggered, when the users requests a suggested emoji character or switches to an emoji character keyboard, as noted above.
- FIG. 2 shows an example method 200 for determining the current contextual emotional state 116 in one implementation, where the third emotional state 114 is not determined. That is, the method 200 determines the current contextual emotional state 116 based on just the first emotional state 108 and the second emotional state 112 .
- the emotional states 108 and 112 are each either an emotional state without weighting, such as happy, sad, angry, depressed, and so on, or an inapposite emotional state.
- an emotional state may have been determined as being inapposite, and if the emotional state has a weighting greater than the threshold, then the state is determined as being this state.
- the threshold may be 0.6.
- the first emotional state 108 may have been determined as 0.2 angry, 0.4 sad, and 0.1 happy. In this case, the first emotional state 108 is deemed as being inapposite, since none of these emotions is greater than 0.6.
- the second emotional state 112 may have been determined as 0.9 angry and 0.7 sad. In this case, the second emotional state is deemed as being angry, since angry is the highest emotion, and has a weighting greater than the threshold.
- the first emotional state 108 is compared in the method 200 to the second emotional state 112 ( 202 ). Based on this comparison, and assuming that at least one of the emotional states 108 and 112 is not inapposite, the current contextual emotional state 116 can be determined as follows. If the two emotional states 108 and 112 are consistent with one another, then the current contextual emotional state 116 is ascertained as a high degree of these states 108 and 112 ( 204 ). For example, if both the emotional states 108 and 112 are happy, then the current contextual emotional state 116 is ascertained as very happy (i.e., happier).
- the current contextual emotional state 116 is ascertained as a baseline degree of the second emotional state 112 ( 206 ). For example, if the emotional state 108 is inapposite and the emotional state 112 is happy, then the current contextual emotional state 116 is ascertained as happy (as opposed to very happy or happier). Similarly, if the second emotional state 112 is neither inconsistent nor consistent with the first emotional state 108 (viz., the second emotional state 108 is inapposite), then the current contextual emotional state 116 is ascertained as a baseline degree of the first emotional state 108 ( 208 ).
- the current contextual emotional state 116 can be ascertained as the user being ironic—that is, irony ( 210 ). For instance, if the emotional state 108 is happy and the emotional state 112 is angry, then the current contextual emotional state 116 is irony. As a concrete example, the user may be laughing, but writing something angry in tone, such that it is ascertained that the user's current contextual emotional state 116 is one of being ironic.
- the method 200 of FIG. 2 thus provides granularity in the emoji characters that are suggested to the user.
- a happy emoji character may be a simple smiley face, whereas a very happy emoji character may be a smiley face that has a very pronounced smile. It is noted that in one implementation, the emoji character corresponding to irony may be a winking smiley face.
- FIG. 3 shows an example method 300 for determining the current contextual emotional state 116 in another implementation, in which the third emotional state 114 is determined. That is, the method 300 determines the current contextual emotional state 116 based on the emotional states 108 , 112 , and 114 . As in the method 200 , the emotional states 108 , 112 , and 114 are each either an emotional state without weighting, or an inapposite emotional state.
- the emotional states 108 , 112 , and 114 are compared to one another ( 302 ). Based on this comparison, and assuming that at least two of the emotional states 108 , 112 , and 114 is not inapposite, the current contextual emotional state 116 can be determined as follows. If the three emotional states 108 , 112 , and 114 are consistent with one another, then the current contextual emotional state 116 is ascertained as a highest degree of these states 108 , 112 , and 114 ( 304 ). For example, if all three emotional states 108 , 112 , and 114 are happy, then the current contextual emotional state 116 is ascertained as happiest.
- the current contextual emotional state 116 is ascertained as a high degree of the two emotional states ( 306 ). For example, if the emotional state 112 is inapposite and the emotional states 108 and 114 are each happy, then the current contextual emotional state 116 is ascertained as happier or very happy (as opposed to just happy, or happiest). By comparison, if two of the emotional states 108 , 112 , and 114 are neither inconsistent nor inconsistent with the remaining emotional state, then the current contextual emotional state 116 is ascertained as a baseline degree of the non-inapposite emotional state ( 308 ). For example, if the emotional state 112 is happy and the emotional states 108 and 114 are each inapposite, then the current contextual emotional state 116 is ascertained as happy (as opposed to happier or happiest).
- the current contextual emotional state 116 can be ascertained as irony ( 310 ).
- the method 300 of FIG. 3 provides even greater granularity in the emoji characters that are suggested to the user than the method 200 of FIG. 2 .
- a happy emoji character may be a simple smiley face
- a happier emoji character may be a smiley face that has a very pronounced smile
- a happiest emoji character may be a smiley face with a toothy smile (i.e., in which the face's teeth can be seen), as an example.
- FIG. 4 shows an example computing device 400 that can implement the techniques that have been described.
- the computing device 400 may be a mobile computing device, such as smartphone or tablet computing device.
- the computing device 500 may be a computer, such a desktop computer, or a laptop or notebook computer.
- the computing device 400 can include a processor 402 , a memory 404 , a display 406 , an input device 408 , and network hardware 410 .
- the memory 404 can be a volatile or non-volatile memory device, and stores program instructions 412 that the processor 402 executes to perform the process flow 100 and/or the methods 200 and 300 that have been described.
- the display 406 displays the emoji characters suggested by the techniques disclosed herein and the text input by the user, and can be a flat-screen display device.
- the input device 408 may be or include a physical keyboard, a touchscreen, a pointing device such as a mouse or touchpad, and so on. The user enters text via the input device 408 and can select one of the suggested emoji characters via the device 408 .
- the network hardware 410 permits the computing device 400 to communicate with communication networks, such as mobile phone networks, the Internet, wireless and/or wired networks, and so on. Via the network hardware 410 , then, the computing device 400 permits the user to send and receive text messages and/or access social networking services.
- the present invention may be a system, a method, and/or a computer program product.
- the computer program product may include a computer readable storage medium (or media) having computer readable program instructions thereon for causing a processor to carry out aspects of the present invention.
- the computer readable storage medium can be a tangible device that can retain and store instructions for use by an instruction execution device.
- the computer readable storage medium may be, for example, but is not limited to, an electronic storage device, a magnetic storage device, an optical storage device, an electromagnetic storage device, a semiconductor storage device, or any suitable combination of the foregoing.
- a non-exhaustive list of more specific examples of the computer readable storage medium includes the following: a portable computer diskette, a hard disk, a random access memory (RAM), a read-only memory (ROM), an erasable programmable read-only memory (EPROM or Flash memory), a static random access memory (SRAM), a portable compact disc read-only memory (CD-ROM), a digital versatile disk (DVD), a memory stick, a floppy disk, a mechanically encoded device such as punch-cards or raised structures in a groove having instructions recorded thereon, and any suitable combination of the foregoing.
- RAM random access memory
- ROM read-only memory
- EPROM or Flash memory erasable programmable read-only memory
- SRAM static random access memory
- CD-ROM compact disc read-only memory
- DVD digital versatile disk
- memory stick a floppy disk
- a mechanically encoded device such as punch-cards or raised structures in a groove having instructions recorded thereon
- a computer readable storage medium is not to be construed as being transitory signals per se, such as radio waves or other freely propagating electromagnetic waves, electromagnetic waves propagating through a waveguide or other transmission media (e.g., light pulses passing through a fiber-optic cable), or electrical signals transmitted through a wire.
- Computer readable program instructions described herein can be downloaded to respective computing/processing devices from a computer readable storage medium or to an external computer or external storage device via a network, for example, the Internet, a local area network, a wide area network and/or a wireless network.
- the network may comprise copper transmission cables, optical transmission fibers, wireless transmission, routers, firewalls, switches, gateway computers and/or edge servers.
- a network adapter card or network interface in each computing/processing device receives computer readable program instructions from the network and forwards the computer readable program instructions for storage in a computer readable storage medium within the respective computing/processing device.
- Computer readable program instructions for carrying out operations of the present invention may be assembler instructions, instruction-set-architecture (ISA) instructions, machine instructions, machine dependent instructions, microcode, firmware instructions, state-setting data, or either source code or object code written in any combination of one or more programming languages, including an object oriented programming language such as Smalltalk, C++ or the like, and conventional procedural programming languages, such as the “C” programming language or similar programming languages.
- the computer readable program instructions may execute entirely on the user's computer, partly on the user's computer, as a stand-alone software package, partly on the user's computer and partly on a remote computer or entirely on the remote computer or server.
- the remote computer may be connected to the user's computer through any type of network, including a local area network (LAN) or a wide area network (WAN), or the connection may be made to an external computer (for example, through the Internet using an Internet Service Provider).
- electronic circuitry including, for example, programmable logic circuitry, field-programmable gate arrays (FPGA), or programmable logic arrays (PLA) may execute the computer readable program instructions by utilizing state information of the computer readable program instructions to personalize the electronic circuitry, in order to perform aspects of the present invention.
- These computer readable program instructions may be provided to a processor of a general purpose computer, special purpose computer, or other programmable data processing apparatus to produce a machine, such that the instructions, which execute via the processor of the computer or other programmable data processing apparatus, create means for implementing the functions/acts specified in the flowchart and/or block diagram block or blocks.
- These computer readable program instructions may also be stored in a computer readable storage medium that can direct a computer, a programmable data processing apparatus, and/or other devices to function in a particular manner, such that the computer readable storage medium having instructions stored therein comprises an article of manufacture including instructions which implement aspects of the function/act specified in the flowchart and/or block diagram block or blocks.
- the computer readable program instructions may also be loaded onto a computer, other programmable data processing apparatus, or other device to cause a series of operational steps to be performed on the computer, other programmable apparatus or other device to produce a computer implemented process, such that the instructions which execute on the computer, other programmable apparatus, or other device implement the functions/acts specified in the flowchart and/or block diagram block or blocks.
- each block in the flowchart or block diagrams may represent a module, segment, or portion of instructions, which comprises one or more executable instructions for implementing the specified logical function(s).
- the functions noted in the block may occur out of the order noted in the figures.
- two blocks shown in succession may, in fact, be executed substantially concurrently, or the blocks may sometimes be executed in the reverse order, depending upon the functionality involved.
Landscapes
- Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- General Engineering & Computer Science (AREA)
- Physics & Mathematics (AREA)
- General Physics & Mathematics (AREA)
- Human Computer Interaction (AREA)
- Health & Medical Sciences (AREA)
- Audiology, Speech & Language Pathology (AREA)
- General Health & Medical Sciences (AREA)
- Computational Linguistics (AREA)
- Artificial Intelligence (AREA)
- Multimedia (AREA)
- Acoustics & Sound (AREA)
- Oral & Maxillofacial Surgery (AREA)
- User Interface Of Digital Computer (AREA)
Abstract
Description
Claims (8)
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US14/859,959 US9665567B2 (en) | 2015-09-21 | 2015-09-21 | Suggesting emoji characters based on current contextual emotional state of user |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US14/859,959 US9665567B2 (en) | 2015-09-21 | 2015-09-21 | Suggesting emoji characters based on current contextual emotional state of user |
Publications (2)
Publication Number | Publication Date |
---|---|
US20170083506A1 US20170083506A1 (en) | 2017-03-23 |
US9665567B2 true US9665567B2 (en) | 2017-05-30 |
Family
ID=58282429
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US14/859,959 Expired - Fee Related US9665567B2 (en) | 2015-09-21 | 2015-09-21 | Suggesting emoji characters based on current contextual emotional state of user |
Country Status (1)
Country | Link |
---|---|
US (1) | US9665567B2 (en) |
Cited By (108)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US10063929B1 (en) | 2016-09-01 | 2018-08-28 | Nufbee Llc | Community controlled audio entertainment system |
US10311144B2 (en) * | 2017-05-16 | 2019-06-04 | Apple Inc. | Emoji word sense disambiguation |
US10348659B1 (en) * | 2017-12-21 | 2019-07-09 | International Business Machines Corporation | Chat message processing |
US10390213B2 (en) | 2014-09-30 | 2019-08-20 | Apple Inc. | Social reminders |
US10388034B2 (en) * | 2017-04-24 | 2019-08-20 | International Business Machines Corporation | Augmenting web content to improve user experience |
US10395654B2 (en) | 2017-05-11 | 2019-08-27 | Apple Inc. | Text normalization based on a data-driven learning network |
US10403283B1 (en) | 2018-06-01 | 2019-09-03 | Apple Inc. | Voice interaction at a primary device to access call functionality of a companion device |
US10417405B2 (en) | 2011-03-21 | 2019-09-17 | Apple Inc. | Device access using voice authentication |
US10417344B2 (en) | 2014-05-30 | 2019-09-17 | Apple Inc. | Exemplar-based natural language processing |
US10417266B2 (en) | 2017-05-09 | 2019-09-17 | Apple Inc. | Context-aware ranking of intelligent response suggestions |
US10438595B2 (en) | 2014-09-30 | 2019-10-08 | Apple Inc. | Speaker identification and unsupervised speaker adaptation techniques |
US10453443B2 (en) | 2014-09-30 | 2019-10-22 | Apple Inc. | Providing an indication of the suitability of speech recognition |
US10474753B2 (en) | 2016-09-07 | 2019-11-12 | Apple Inc. | Language identification using recurrent neural networks |
US10496705B1 (en) | 2018-06-03 | 2019-12-03 | Apple Inc. | Accelerated task performance |
US10529332B2 (en) | 2015-03-08 | 2020-01-07 | Apple Inc. | Virtual assistant activation |
US10558757B2 (en) | 2017-03-11 | 2020-02-11 | International Business Machines Corporation | Symbol management |
US10580409B2 (en) | 2016-06-11 | 2020-03-03 | Apple Inc. | Application integration with a digital assistant |
US10592604B2 (en) | 2018-03-12 | 2020-03-17 | Apple Inc. | Inverse text normalization for automatic speech recognition |
US10657966B2 (en) | 2014-05-30 | 2020-05-19 | Apple Inc. | Better resolution when referencing to concepts |
US10666735B2 (en) | 2014-05-19 | 2020-05-26 | Auerbach Michael Harrison Tretter | Dynamic computer systems and uses thereof |
US10681212B2 (en) | 2015-06-05 | 2020-06-09 | Apple Inc. | Virtual assistant aided communication with 3rd party service in a communication session |
US10692504B2 (en) | 2010-02-25 | 2020-06-23 | Apple Inc. | User profiling for voice input processing |
US10699717B2 (en) | 2014-05-30 | 2020-06-30 | Apple Inc. | Intelligent assistant for home automation |
US20200218781A1 (en) * | 2019-01-04 | 2020-07-09 | International Business Machines Corporation | Sentiment adapted communication |
US10714117B2 (en) | 2013-02-07 | 2020-07-14 | Apple Inc. | Voice trigger for a digital assistant |
US10741185B2 (en) | 2010-01-18 | 2020-08-11 | Apple Inc. | Intelligent automated assistant |
US10741181B2 (en) | 2017-05-09 | 2020-08-11 | Apple Inc. | User interface for correcting recognition errors |
US10748546B2 (en) | 2017-05-16 | 2020-08-18 | Apple Inc. | Digital assistant services based on device capabilities |
US10769385B2 (en) | 2013-06-09 | 2020-09-08 | Apple Inc. | System and method for inferring user intent from speech inputs |
US10798044B1 (en) | 2016-09-01 | 2020-10-06 | Nufbee Llc | Method for enhancing text messages with pre-recorded audio clips |
US10839159B2 (en) | 2018-09-28 | 2020-11-17 | Apple Inc. | Named entity normalization in a spoken dialog system |
US10878809B2 (en) | 2014-05-30 | 2020-12-29 | Apple Inc. | Multi-command single utterance input method |
US10892996B2 (en) | 2018-06-01 | 2021-01-12 | Apple Inc. | Variable latency device coordination |
US10909171B2 (en) | 2017-05-16 | 2021-02-02 | Apple Inc. | Intelligent automated assistant for media exploration |
US10930282B2 (en) | 2015-03-08 | 2021-02-23 | Apple Inc. | Competing devices responding to voice triggers |
US10942703B2 (en) | 2015-12-23 | 2021-03-09 | Apple Inc. | Proactive assistance based on dialog communication between devices |
US10956666B2 (en) | 2015-11-09 | 2021-03-23 | Apple Inc. | Unconventional virtual assistant interactions |
US11010561B2 (en) | 2018-09-27 | 2021-05-18 | Apple Inc. | Sentiment prediction from textual data |
US11009970B2 (en) | 2018-06-01 | 2021-05-18 | Apple Inc. | Attention aware virtual assistant dismissal |
US11010127B2 (en) | 2015-06-29 | 2021-05-18 | Apple Inc. | Virtual assistant for media playback |
US11037565B2 (en) | 2016-06-10 | 2021-06-15 | Apple Inc. | Intelligent digital assistant in a multi-tasking environment |
US11048473B2 (en) | 2013-06-09 | 2021-06-29 | Apple Inc. | Device, method, and graphical user interface for enabling conversation persistence across two or more instances of a digital assistant |
US11070949B2 (en) | 2015-05-27 | 2021-07-20 | Apple Inc. | Systems and methods for proactively identifying and surfacing relevant content on an electronic device with a touch-sensitive display |
US11074398B2 (en) | 2018-10-12 | 2021-07-27 | International Business Machines Corporation | Tracking and managing emoji annotations |
US11120372B2 (en) | 2011-06-03 | 2021-09-14 | Apple Inc. | Performing actions associated with task items that represent tasks to perform |
US11126400B2 (en) | 2015-09-08 | 2021-09-21 | Apple Inc. | Zero latency digital assistant |
US11127397B2 (en) | 2015-05-27 | 2021-09-21 | Apple Inc. | Device voice control |
US11133008B2 (en) | 2014-05-30 | 2021-09-28 | Apple Inc. | Reducing the need for manual start/end-pointing and trigger phrases |
US11140099B2 (en) | 2019-05-21 | 2021-10-05 | Apple Inc. | Providing message response suggestions |
US11169616B2 (en) | 2018-05-07 | 2021-11-09 | Apple Inc. | Raise to speak |
US11170166B2 (en) | 2018-09-28 | 2021-11-09 | Apple Inc. | Neural typographical error modeling via generative adversarial networks |
US11217251B2 (en) | 2019-05-06 | 2022-01-04 | Apple Inc. | Spoken notifications |
US11227589B2 (en) | 2016-06-06 | 2022-01-18 | Apple Inc. | Intelligent list reading |
US11231904B2 (en) | 2015-03-06 | 2022-01-25 | Apple Inc. | Reducing response latency of intelligent automated assistants |
US11237797B2 (en) | 2019-05-31 | 2022-02-01 | Apple Inc. | User activity shortcut suggestions |
US11269678B2 (en) | 2012-05-15 | 2022-03-08 | Apple Inc. | Systems and methods for integrating third party services with a digital assistant |
US11283751B1 (en) | 2020-11-03 | 2022-03-22 | International Business Machines Corporation | Using speech and facial bio-metrics to deliver text messages at the appropriate time |
US11289073B2 (en) | 2019-05-31 | 2022-03-29 | Apple Inc. | Device text to speech |
US11301477B2 (en) | 2017-05-12 | 2022-04-12 | Apple Inc. | Feedback analysis of a digital assistant |
US11307752B2 (en) | 2019-05-06 | 2022-04-19 | Apple Inc. | User configurable task triggers |
US11314370B2 (en) | 2013-12-06 | 2022-04-26 | Apple Inc. | Method for extracting salient dialog usage from live data |
US11348582B2 (en) | 2008-10-02 | 2022-05-31 | Apple Inc. | Electronic devices with voice command and contextual data processing capabilities |
US11348573B2 (en) | 2019-03-18 | 2022-05-31 | Apple Inc. | Multimodality in digital assistant systems |
US11360641B2 (en) | 2019-06-01 | 2022-06-14 | Apple Inc. | Increasing the relevance of new available information |
US11380310B2 (en) | 2017-05-12 | 2022-07-05 | Apple Inc. | Low-latency intelligent automated assistant |
US11388291B2 (en) | 2013-03-14 | 2022-07-12 | Apple Inc. | System and method for processing voicemail |
US11405466B2 (en) | 2017-05-12 | 2022-08-02 | Apple Inc. | Synchronization and task delegation of a digital assistant |
US11423908B2 (en) | 2019-05-06 | 2022-08-23 | Apple Inc. | Interpreting spoken requests |
US11423886B2 (en) | 2010-01-18 | 2022-08-23 | Apple Inc. | Task flow identification based on user intent |
US11462215B2 (en) | 2018-09-28 | 2022-10-04 | Apple Inc. | Multi-modal inputs for voice commands |
US11467802B2 (en) | 2017-05-11 | 2022-10-11 | Apple Inc. | Maintaining privacy of personal information |
US11468282B2 (en) | 2015-05-15 | 2022-10-11 | Apple Inc. | Virtual assistant in a communication session |
US11475884B2 (en) | 2019-05-06 | 2022-10-18 | Apple Inc. | Reducing digital assistant latency when a language is incorrectly determined |
US11475898B2 (en) | 2018-10-26 | 2022-10-18 | Apple Inc. | Low-latency multi-speaker speech recognition |
US11488406B2 (en) | 2019-09-25 | 2022-11-01 | Apple Inc. | Text detection using global geometry estimators |
US11495218B2 (en) | 2018-06-01 | 2022-11-08 | Apple Inc. | Virtual assistant operation in multi-device environments |
US11496600B2 (en) | 2019-05-31 | 2022-11-08 | Apple Inc. | Remote execution of machine-learned models |
US11500672B2 (en) | 2015-09-08 | 2022-11-15 | Apple Inc. | Distributed personal assistant |
US11516539B2 (en) | 2021-03-01 | 2022-11-29 | Comcast Cable Communications, Llc | Systems and methods for providing contextually relevant information |
US11516537B2 (en) | 2014-06-30 | 2022-11-29 | Apple Inc. | Intelligent automated assistant for TV user interactions |
US11526368B2 (en) | 2015-11-06 | 2022-12-13 | Apple Inc. | Intelligent automated assistant in a messaging environment |
US11532306B2 (en) | 2017-05-16 | 2022-12-20 | Apple Inc. | Detecting a trigger of a digital assistant |
US11550751B2 (en) * | 2016-11-18 | 2023-01-10 | Microsoft Technology Licensing, Llc | Sequence expander for data entry/information retrieval |
WO2023009323A1 (en) * | 2021-07-29 | 2023-02-02 | Snap Inc. | Emoji recommendation system using user context and biosignals |
US11580990B2 (en) | 2017-05-12 | 2023-02-14 | Apple Inc. | User-specific acoustic models |
US11599331B2 (en) | 2017-05-11 | 2023-03-07 | Apple Inc. | Maintaining privacy of personal information |
US11638059B2 (en) | 2019-01-04 | 2023-04-25 | Apple Inc. | Content playback on multiple devices |
US11656884B2 (en) | 2017-01-09 | 2023-05-23 | Apple Inc. | Application integration with a digital assistant |
US11657813B2 (en) | 2019-05-31 | 2023-05-23 | Apple Inc. | Voice identification in digital assistant systems |
US11657558B2 (en) | 2021-09-16 | 2023-05-23 | International Business Machines Corporation | Context-based personalized communication presentation |
US11671920B2 (en) | 2007-04-03 | 2023-06-06 | Apple Inc. | Method and system for operating a multifunction portable electronic device using voice-activation |
US11696060B2 (en) | 2020-07-21 | 2023-07-04 | Apple Inc. | User identification using headphones |
US11710482B2 (en) | 2018-03-26 | 2023-07-25 | Apple Inc. | Natural assistant interaction |
US11755276B2 (en) | 2020-05-12 | 2023-09-12 | Apple Inc. | Reducing description length based on confidence |
US11765209B2 (en) | 2020-05-11 | 2023-09-19 | Apple Inc. | Digital assistant hardware abstraction |
US11790914B2 (en) | 2019-06-01 | 2023-10-17 | Apple Inc. | Methods and user interfaces for voice-based control of electronic devices |
US11798547B2 (en) | 2013-03-15 | 2023-10-24 | Apple Inc. | Voice activated device for use with a voice-based digital assistant |
WO2023212259A1 (en) * | 2022-04-28 | 2023-11-02 | Theai, Inc. | Artificial intelligence character models with modifiable behavioral characteristics |
US11809483B2 (en) | 2015-09-08 | 2023-11-07 | Apple Inc. | Intelligent automated assistant for media search and playback |
US11809783B2 (en) | 2016-06-11 | 2023-11-07 | Apple Inc. | Intelligent device arbitration and control |
US11838734B2 (en) | 2020-07-20 | 2023-12-05 | Apple Inc. | Multi-device audio adjustment coordination |
US11853536B2 (en) | 2015-09-08 | 2023-12-26 | Apple Inc. | Intelligent automated assistant in a media environment |
US11854539B2 (en) | 2018-05-07 | 2023-12-26 | Apple Inc. | Intelligent automated assistant for delivering content from user experiences |
US11914848B2 (en) | 2020-05-11 | 2024-02-27 | Apple Inc. | Providing relevant data items based on context |
US11928604B2 (en) | 2005-09-08 | 2024-03-12 | Apple Inc. | Method and apparatus for building an intelligent automated assistant |
US12010262B2 (en) | 2013-08-06 | 2024-06-11 | Apple Inc. | Auto-activating smart responses based on activities from remote devices |
US12014118B2 (en) | 2017-05-15 | 2024-06-18 | Apple Inc. | Multi-modal interfaces having selection disambiguation and text modification capability |
US12051413B2 (en) | 2015-09-30 | 2024-07-30 | Apple Inc. | Intelligent device identification |
Families Citing this family (32)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20170344224A1 (en) * | 2016-05-27 | 2017-11-30 | Nuance Communications, Inc. | Suggesting emojis to users for insertion into text-based messages |
US11321890B2 (en) * | 2016-11-09 | 2022-05-03 | Microsoft Technology Licensing, Llc | User interface for generating expressive content |
US20180176168A1 (en) * | 2016-11-30 | 2018-06-21 | Fujitsu Limited | Visual feedback system |
CN107103900B (en) * | 2017-06-06 | 2020-03-31 | 西北师范大学 | Cross-language emotion voice synthesis method and system |
KR102485253B1 (en) * | 2017-11-10 | 2023-01-06 | 현대자동차주식회사 | Dialogue processing system(apparatus), and method for controlling thereof |
US10441886B2 (en) * | 2018-01-29 | 2019-10-15 | Sony Interactive Entertainment LLC | Dynamic allocation of contextual assistance during game play |
US10610783B2 (en) | 2018-01-31 | 2020-04-07 | Sony Interactive Entertainment LLC | Assignment of contextual game play assistance to player reaction |
GB2572984A (en) * | 2018-04-18 | 2019-10-23 | Sony Corp | Method and data processing apparatus |
US10719968B2 (en) * | 2018-04-18 | 2020-07-21 | Snap Inc. | Augmented expression system |
US20190325201A1 (en) * | 2018-04-19 | 2019-10-24 | Microsoft Technology Licensing, Llc | Automated emotion detection and keyboard service |
US11573679B2 (en) * | 2018-04-30 | 2023-02-07 | The Trustees of the California State University | Integration of user emotions for a smartphone or other communication device environment |
CN110910898B (en) * | 2018-09-15 | 2022-12-30 | 华为技术有限公司 | Voice information processing method and device |
CN113330476A (en) * | 2018-09-21 | 2021-08-31 | 史蒂夫·柯蒂斯 | System and method for allocating revenue among users based on quantified and qualified mood data |
US10346541B1 (en) * | 2018-10-05 | 2019-07-09 | Capital One Services, Llc | Typifying emotional indicators for digital messaging |
US11579589B2 (en) * | 2018-10-25 | 2023-02-14 | International Business Machines Corporation | Selectively activating a resource by detecting emotions through context analysis |
GB2581328A (en) * | 2019-02-01 | 2020-08-19 | Sony Europe Ltd | Method and data processing apparatus |
US20220121817A1 (en) * | 2019-02-14 | 2022-04-21 | Sony Group Corporation | Information processing device, information processing method, and information processing program |
US11232407B1 (en) * | 2019-03-06 | 2022-01-25 | Anthem, Inc. | System and method of assessing sentiment of an organization |
CN114402273A (en) * | 2019-04-30 | 2022-04-26 | 跃进公司 | Electronic system and method for emotional state assessment |
US11373039B2 (en) * | 2019-09-26 | 2022-06-28 | International Business Machines Corporation | Content context aware message intent checker |
CN111128190B (en) * | 2019-12-31 | 2023-03-21 | 恒信东方文化股份有限公司 | Expression matching method and system |
US11593984B2 (en) * | 2020-02-07 | 2023-02-28 | Apple Inc. | Using text for avatar animation |
CN111772648A (en) * | 2020-06-10 | 2020-10-16 | 南京七岩电子科技有限公司 | Method and device for judging emotion by combining HRV signal and facial expression |
US20220269354A1 (en) * | 2020-06-19 | 2022-08-25 | Talent Unlimited Online Services Private Limited | Artificial intelligence-based system and method for dynamically predicting and suggesting emojis for messages |
CN113158656B (en) * | 2020-12-25 | 2024-05-14 | 北京中科闻歌科技股份有限公司 | Ironic content recognition method, ironic content recognition device, electronic device, and storage medium |
KR20220130952A (en) * | 2021-03-19 | 2022-09-27 | 현대자동차주식회사 | Apparatus for generating emojies, vehicle and method for generating emojies |
US11888797B2 (en) | 2021-04-20 | 2024-01-30 | Snap Inc. | Emoji-first messaging |
US11593548B2 (en) | 2021-04-20 | 2023-02-28 | Snap Inc. | Client device processing received emoji-first messages |
US11531406B2 (en) * | 2021-04-20 | 2022-12-20 | Snap Inc. | Personalized emoji dictionary |
US11676317B2 (en) | 2021-04-27 | 2023-06-13 | International Business Machines Corporation | Generation of custom composite emoji images based on user-selected input feed types associated with Internet of Things (IoT) device input feeds |
US11902231B2 (en) * | 2022-02-14 | 2024-02-13 | International Business Machines Corporation | Dynamic display of images based on textual content |
US20230316812A1 (en) * | 2022-03-31 | 2023-10-05 | Matrixcare, Inc. | Sign language sentiment analysis |
Citations (68)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US5029214A (en) * | 1986-08-11 | 1991-07-02 | Hollander James F | Electronic speech control apparatus and methods |
US5860064A (en) * | 1993-05-13 | 1999-01-12 | Apple Computer, Inc. | Method and apparatus for automatic generation of vocal emotion in a synthetic text-to-speech system |
US20020158599A1 (en) * | 2000-03-31 | 2002-10-31 | Masahiro Fujita | Robot device, robot device action control method, external force detecting device and external force detecting method |
US6578019B1 (en) | 1994-11-08 | 2003-06-10 | Canon Kabushiki Kaisha | Information processing system which understands information and acts accordingly and method therefor |
US6601057B1 (en) * | 1999-11-05 | 2003-07-29 | Decentrix Inc. | Method and apparatus for generating and modifying multiple instances of an element of a web site |
US20030182123A1 (en) * | 2000-09-13 | 2003-09-25 | Shunji Mitsuyoshi | Emotion recognizing method, sensibility creating method, device, and software |
US20030216919A1 (en) * | 2002-05-13 | 2003-11-20 | Roushar Joseph C. | Multi-dimensional method and apparatus for automated language interpretation |
US20040158454A1 (en) * | 2003-02-11 | 2004-08-12 | Livia Polanyi | System and method for dynamically determining the attitude of an author of a natural language document |
US20040249634A1 (en) * | 2001-08-09 | 2004-12-09 | Yoav Degani | Method and apparatus for speech analysis |
JP2004538543A (en) | 2001-02-05 | 2004-12-24 | インターナショナル・ビジネス・マシーンズ・コーポレーション | System and method for multi-mode focus detection, reference ambiguity resolution and mood classification using multi-mode input |
US20050027525A1 (en) * | 2003-07-29 | 2005-02-03 | Fuji Photo Film Co., Ltd. | Cell phone having an information-converting function |
US20050038657A1 (en) * | 2001-09-05 | 2005-02-17 | Voice Signal Technologies, Inc. | Combined speech recongnition and text-to-speech generation |
US20050054381A1 (en) | 2003-09-05 | 2005-03-10 | Samsung Electronics Co., Ltd. | Proactive user interface |
US20060036585A1 (en) * | 2004-02-15 | 2006-02-16 | King Martin T | Publishing techniques for adding value to a rendered document |
US20060229889A1 (en) * | 2005-03-30 | 2006-10-12 | Ianywhere Solutions, Inc. | Context proposed items mechanism for natural language user interface |
US20060247915A1 (en) * | 1998-12-04 | 2006-11-02 | Tegic Communications, Inc. | Contextual Prediction of User Words and User Actions |
US20070071206A1 (en) * | 2005-06-24 | 2007-03-29 | Gainsboro Jay L | Multi-party conversation analyzer & logger |
US20070100603A1 (en) * | 2002-10-07 | 2007-05-03 | Warner Douglas K | Method for routing electronic correspondence based on the level and type of emotion contained therein |
US20070208569A1 (en) * | 2006-03-03 | 2007-09-06 | Balan Subramanian | Communicating across voice and text channels with emotion preservation |
US7392475B1 (en) * | 2003-05-23 | 2008-06-24 | Microsoft Corporation | Method and system for automatic insertion of context information into an application program module |
US20080151038A1 (en) | 2006-12-20 | 2008-06-26 | Cisco Technology, Inc. | Video contact center facial expression analyzer module |
US20080221892A1 (en) | 2007-03-06 | 2008-09-11 | Paco Xander Nathan | Systems and methods for an autonomous avatar driver |
US20100111375A1 (en) | 2008-10-31 | 2010-05-06 | Michael Jeffrey Jones | Method for Determining Atributes of Faces in Images |
US20100125811A1 (en) | 2008-11-19 | 2010-05-20 | Bradford Allen Moore | Portable Touch Screen Device, Method, and Graphical User Interface for Entering and Using Emoji Characters |
US7751623B1 (en) * | 2002-06-28 | 2010-07-06 | Microsoft Corporation | Writing guide for a free-form document editor |
US20110004624A1 (en) * | 2009-07-02 | 2011-01-06 | International Business Machines Corporation | Method for Customer Feedback Measurement in Public Places Utilizing Speech Recognition Technology |
US20110041153A1 (en) | 2008-01-03 | 2011-02-17 | Colin Simon | Content management and delivery method, system and apparatus |
KR20110026218A (en) | 2009-09-07 | 2011-03-15 | 동국대학교 산학협력단 | Apparatus and method for inputting text message and its program stored in recording medium |
US20110099006A1 (en) * | 2009-10-27 | 2011-04-28 | Cisco Technology, Inc. | Automated and enhanced note taking for online collaborative computing sessions |
US20110273455A1 (en) * | 2010-05-04 | 2011-11-10 | Shazam Entertainment Ltd. | Systems and Methods of Rendering a Textual Animation |
US8065150B2 (en) * | 2002-11-29 | 2011-11-22 | Nuance Communications, Inc. | Application of emotion-based intonation and prosody to speech in text-to-speech systems |
US20120030227A1 (en) * | 2010-07-30 | 2012-02-02 | Microsoft Corporation | System of providing suggestions based on accessible and contextual information |
US20120035924A1 (en) * | 2010-08-06 | 2012-02-09 | Google Inc. | Disambiguating input based on context |
US20120072217A1 (en) * | 2010-09-17 | 2012-03-22 | At&T Intellectual Property I, L.P | System and method for using prosody for voice-enabled search |
US20120166180A1 (en) * | 2009-03-23 | 2012-06-28 | Lawrence Au | Compassion, Variety and Cohesion For Methods Of Text Analytics, Writing, Search, User Interfaces |
US20120278064A1 (en) | 2011-04-29 | 2012-11-01 | Adam Leary | System and method for determining sentiment from text content |
US20120304074A1 (en) | 2011-05-23 | 2012-11-29 | Microsoft Corporation | Device user interface to input emoji and other symbols |
US20120317046A1 (en) * | 2011-06-10 | 2012-12-13 | Myslinski Lucas J | Candidate fact checking method and system |
JP2013000300A (en) | 2011-06-15 | 2013-01-07 | Nissan Motor Co Ltd | Mood determining apparatus and mood determining method |
US20130019163A1 (en) * | 2010-03-26 | 2013-01-17 | British Telecommunications Public Limited Company | System |
US20130103386A1 (en) * | 2011-10-24 | 2013-04-25 | Lei Zhang | Performing sentiment analysis |
US20130151237A1 (en) * | 2011-12-09 | 2013-06-13 | Chrysler Group Llc | Dynamic method for emoticon translation |
US20130151508A1 (en) * | 2011-12-12 | 2013-06-13 | Empire Technology Development Llc | Content-based automatic input protocol selection |
US20130159919A1 (en) * | 2011-12-19 | 2013-06-20 | Gabriel Leydon | Systems and Methods for Identifying and Suggesting Emoticons |
US8554701B1 (en) * | 2011-03-18 | 2013-10-08 | Amazon Technologies, Inc. | Determining sentiment of sentences from customer reviews |
US20130275899A1 (en) * | 2010-01-18 | 2013-10-17 | Apple Inc. | Application Gateway for Providing Different User Interfaces for Limited Distraction and Non-Limited Distraction Contexts |
US8588825B2 (en) | 2010-05-25 | 2013-11-19 | Sony Corporation | Text enhancement |
US8620850B2 (en) | 2010-09-07 | 2013-12-31 | Blackberry Limited | Dynamically manipulating an emoticon or avatar |
US20140035823A1 (en) * | 2012-08-01 | 2014-02-06 | Apple Inc. | Dynamic Context-Based Language Determination |
US20140046660A1 (en) | 2012-08-10 | 2014-02-13 | Yahoo! Inc | Method and system for voice based mood analysis |
US20140088954A1 (en) * | 2012-09-27 | 2014-03-27 | Research In Motion Limited | Apparatus and method pertaining to automatically-suggested emoticons |
US20140095150A1 (en) * | 2012-10-03 | 2014-04-03 | Kanjoya, Inc. | Emotion identification system and method |
US20140108006A1 (en) * | 2012-09-07 | 2014-04-17 | Grail, Inc. | System and method for analyzing and mapping semiotic relationships to enhance content recommendations |
WO2014102722A1 (en) | 2012-12-26 | 2014-07-03 | Sia Technology Ltd. | Device, system, and method of controlling electronic devices via thought |
US20140236596A1 (en) * | 2013-02-21 | 2014-08-21 | Nuance Communications, Inc. | Emotion detection in voicemail |
US8922481B1 (en) * | 2012-03-16 | 2014-12-30 | Google Inc. | Content annotation |
US20150025403A1 (en) | 2013-04-15 | 2015-01-22 | Yonglin Biotech Corp. | Mood analysis method, system, and apparatus |
US20150046371A1 (en) * | 2011-04-29 | 2015-02-12 | Cbs Interactive Inc. | System and method for determining sentiment from text content |
US20150052462A1 (en) | 2013-08-15 | 2015-02-19 | Yahoo! Inc. | Capture and retrieval of a personalized mood icon |
US8972424B2 (en) * | 2009-05-29 | 2015-03-03 | Peter Snell | Subjective linguistic analysis |
US9014364B1 (en) * | 2014-03-31 | 2015-04-21 | Noble Systems Corporation | Contact center speech analytics system having multiple speech analytics engines |
US20150121285A1 (en) * | 2013-10-24 | 2015-04-30 | Fleksy, Inc. | User interface for text input and virtual keyboard manipulation |
US9043196B1 (en) * | 2014-07-07 | 2015-05-26 | Machine Zone, Inc. | Systems and methods for identifying and suggesting emoticons |
US20150186369A1 (en) * | 2013-12-31 | 2015-07-02 | Abbyy Development Llc | Method and System for Dossiers for Data Units |
US20150248424A1 (en) * | 2014-02-28 | 2015-09-03 | International Business Machines Corporation | Sorting and displaying documents according to sentiment level in an online community |
US20150350118A1 (en) * | 2014-05-30 | 2015-12-03 | Apple Inc. | Canned answers in messages |
US20160042359A1 (en) * | 2014-08-11 | 2016-02-11 | 24/7 Customer, Inc. | Methods and apparatuses for modeling customer interaction experiences |
US20160071119A1 (en) * | 2013-04-11 | 2016-03-10 | Longsand Limited | Sentiment feedback |
-
2015
- 2015-09-21 US US14/859,959 patent/US9665567B2/en not_active Expired - Fee Related
Patent Citations (70)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US5029214A (en) * | 1986-08-11 | 1991-07-02 | Hollander James F | Electronic speech control apparatus and methods |
US5860064A (en) * | 1993-05-13 | 1999-01-12 | Apple Computer, Inc. | Method and apparatus for automatic generation of vocal emotion in a synthetic text-to-speech system |
US6578019B1 (en) | 1994-11-08 | 2003-06-10 | Canon Kabushiki Kaisha | Information processing system which understands information and acts accordingly and method therefor |
US20060247915A1 (en) * | 1998-12-04 | 2006-11-02 | Tegic Communications, Inc. | Contextual Prediction of User Words and User Actions |
US6601057B1 (en) * | 1999-11-05 | 2003-07-29 | Decentrix Inc. | Method and apparatus for generating and modifying multiple instances of an element of a web site |
US20020158599A1 (en) * | 2000-03-31 | 2002-10-31 | Masahiro Fujita | Robot device, robot device action control method, external force detecting device and external force detecting method |
US20030182123A1 (en) * | 2000-09-13 | 2003-09-25 | Shunji Mitsuyoshi | Emotion recognizing method, sensibility creating method, device, and software |
JP2004538543A (en) | 2001-02-05 | 2004-12-24 | インターナショナル・ビジネス・マシーンズ・コーポレーション | System and method for multi-mode focus detection, reference ambiguity resolution and mood classification using multi-mode input |
US20040249634A1 (en) * | 2001-08-09 | 2004-12-09 | Yoav Degani | Method and apparatus for speech analysis |
US20050038657A1 (en) * | 2001-09-05 | 2005-02-17 | Voice Signal Technologies, Inc. | Combined speech recongnition and text-to-speech generation |
US20030216919A1 (en) * | 2002-05-13 | 2003-11-20 | Roushar Joseph C. | Multi-dimensional method and apparatus for automated language interpretation |
US7751623B1 (en) * | 2002-06-28 | 2010-07-06 | Microsoft Corporation | Writing guide for a free-form document editor |
US20070100603A1 (en) * | 2002-10-07 | 2007-05-03 | Warner Douglas K | Method for routing electronic correspondence based on the level and type of emotion contained therein |
US8065150B2 (en) * | 2002-11-29 | 2011-11-22 | Nuance Communications, Inc. | Application of emotion-based intonation and prosody to speech in text-to-speech systems |
US20040158454A1 (en) * | 2003-02-11 | 2004-08-12 | Livia Polanyi | System and method for dynamically determining the attitude of an author of a natural language document |
US7392475B1 (en) * | 2003-05-23 | 2008-06-24 | Microsoft Corporation | Method and system for automatic insertion of context information into an application program module |
US20050027525A1 (en) * | 2003-07-29 | 2005-02-03 | Fuji Photo Film Co., Ltd. | Cell phone having an information-converting function |
US20050054381A1 (en) | 2003-09-05 | 2005-03-10 | Samsung Electronics Co., Ltd. | Proactive user interface |
US20060036585A1 (en) * | 2004-02-15 | 2006-02-16 | King Martin T | Publishing techniques for adding value to a rendered document |
US20060229889A1 (en) * | 2005-03-30 | 2006-10-12 | Ianywhere Solutions, Inc. | Context proposed items mechanism for natural language user interface |
US20070071206A1 (en) * | 2005-06-24 | 2007-03-29 | Gainsboro Jay L | Multi-party conversation analyzer & logger |
US20070208569A1 (en) * | 2006-03-03 | 2007-09-06 | Balan Subramanian | Communicating across voice and text channels with emotion preservation |
US20080151038A1 (en) | 2006-12-20 | 2008-06-26 | Cisco Technology, Inc. | Video contact center facial expression analyzer module |
US20080221892A1 (en) | 2007-03-06 | 2008-09-11 | Paco Xander Nathan | Systems and methods for an autonomous avatar driver |
US20110041153A1 (en) | 2008-01-03 | 2011-02-17 | Colin Simon | Content management and delivery method, system and apparatus |
US20100111375A1 (en) | 2008-10-31 | 2010-05-06 | Michael Jeffrey Jones | Method for Determining Atributes of Faces in Images |
US20100125811A1 (en) | 2008-11-19 | 2010-05-20 | Bradford Allen Moore | Portable Touch Screen Device, Method, and Graphical User Interface for Entering and Using Emoji Characters |
US20120166180A1 (en) * | 2009-03-23 | 2012-06-28 | Lawrence Au | Compassion, Variety and Cohesion For Methods Of Text Analytics, Writing, Search, User Interfaces |
US9213687B2 (en) * | 2009-03-23 | 2015-12-15 | Lawrence Au | Compassion, variety and cohesion for methods of text analytics, writing, search, user interfaces |
US8972424B2 (en) * | 2009-05-29 | 2015-03-03 | Peter Snell | Subjective linguistic analysis |
US20110004624A1 (en) * | 2009-07-02 | 2011-01-06 | International Business Machines Corporation | Method for Customer Feedback Measurement in Public Places Utilizing Speech Recognition Technology |
KR20110026218A (en) | 2009-09-07 | 2011-03-15 | 동국대학교 산학협력단 | Apparatus and method for inputting text message and its program stored in recording medium |
US20110099006A1 (en) * | 2009-10-27 | 2011-04-28 | Cisco Technology, Inc. | Automated and enhanced note taking for online collaborative computing sessions |
US20130275899A1 (en) * | 2010-01-18 | 2013-10-17 | Apple Inc. | Application Gateway for Providing Different User Interfaces for Limited Distraction and Non-Limited Distraction Contexts |
US20130019163A1 (en) * | 2010-03-26 | 2013-01-17 | British Telecommunications Public Limited Company | System |
US20110273455A1 (en) * | 2010-05-04 | 2011-11-10 | Shazam Entertainment Ltd. | Systems and Methods of Rendering a Textual Animation |
US8588825B2 (en) | 2010-05-25 | 2013-11-19 | Sony Corporation | Text enhancement |
US20120030227A1 (en) * | 2010-07-30 | 2012-02-02 | Microsoft Corporation | System of providing suggestions based on accessible and contextual information |
US20120035924A1 (en) * | 2010-08-06 | 2012-02-09 | Google Inc. | Disambiguating input based on context |
US8620850B2 (en) | 2010-09-07 | 2013-12-31 | Blackberry Limited | Dynamically manipulating an emoticon or avatar |
US20120072217A1 (en) * | 2010-09-17 | 2012-03-22 | At&T Intellectual Property I, L.P | System and method for using prosody for voice-enabled search |
US8554701B1 (en) * | 2011-03-18 | 2013-10-08 | Amazon Technologies, Inc. | Determining sentiment of sentences from customer reviews |
US20120278064A1 (en) | 2011-04-29 | 2012-11-01 | Adam Leary | System and method for determining sentiment from text content |
US20150046371A1 (en) * | 2011-04-29 | 2015-02-12 | Cbs Interactive Inc. | System and method for determining sentiment from text content |
US20120304074A1 (en) | 2011-05-23 | 2012-11-29 | Microsoft Corporation | Device user interface to input emoji and other symbols |
US20120317046A1 (en) * | 2011-06-10 | 2012-12-13 | Myslinski Lucas J | Candidate fact checking method and system |
JP2013000300A (en) | 2011-06-15 | 2013-01-07 | Nissan Motor Co Ltd | Mood determining apparatus and mood determining method |
US20130103386A1 (en) * | 2011-10-24 | 2013-04-25 | Lei Zhang | Performing sentiment analysis |
US20130151237A1 (en) * | 2011-12-09 | 2013-06-13 | Chrysler Group Llc | Dynamic method for emoticon translation |
US20130151508A1 (en) * | 2011-12-12 | 2013-06-13 | Empire Technology Development Llc | Content-based automatic input protocol selection |
US20130159919A1 (en) * | 2011-12-19 | 2013-06-20 | Gabriel Leydon | Systems and Methods for Identifying and Suggesting Emoticons |
US20150095020A1 (en) | 2011-12-19 | 2015-04-02 | Machine Zone, Inc. | Systems and Methods for Identifying and Suggesting Emoticons |
US8922481B1 (en) * | 2012-03-16 | 2014-12-30 | Google Inc. | Content annotation |
US20140035823A1 (en) * | 2012-08-01 | 2014-02-06 | Apple Inc. | Dynamic Context-Based Language Determination |
US20140046660A1 (en) | 2012-08-10 | 2014-02-13 | Yahoo! Inc | Method and system for voice based mood analysis |
US20140108006A1 (en) * | 2012-09-07 | 2014-04-17 | Grail, Inc. | System and method for analyzing and mapping semiotic relationships to enhance content recommendations |
US20140088954A1 (en) * | 2012-09-27 | 2014-03-27 | Research In Motion Limited | Apparatus and method pertaining to automatically-suggested emoticons |
US20140095150A1 (en) * | 2012-10-03 | 2014-04-03 | Kanjoya, Inc. | Emotion identification system and method |
WO2014102722A1 (en) | 2012-12-26 | 2014-07-03 | Sia Technology Ltd. | Device, system, and method of controlling electronic devices via thought |
US20140236596A1 (en) * | 2013-02-21 | 2014-08-21 | Nuance Communications, Inc. | Emotion detection in voicemail |
US20160071119A1 (en) * | 2013-04-11 | 2016-03-10 | Longsand Limited | Sentiment feedback |
US20150025403A1 (en) | 2013-04-15 | 2015-01-22 | Yonglin Biotech Corp. | Mood analysis method, system, and apparatus |
US20150052462A1 (en) | 2013-08-15 | 2015-02-19 | Yahoo! Inc. | Capture and retrieval of a personalized mood icon |
US20150121285A1 (en) * | 2013-10-24 | 2015-04-30 | Fleksy, Inc. | User interface for text input and virtual keyboard manipulation |
US20150186369A1 (en) * | 2013-12-31 | 2015-07-02 | Abbyy Development Llc | Method and System for Dossiers for Data Units |
US20150248424A1 (en) * | 2014-02-28 | 2015-09-03 | International Business Machines Corporation | Sorting and displaying documents according to sentiment level in an online community |
US9014364B1 (en) * | 2014-03-31 | 2015-04-21 | Noble Systems Corporation | Contact center speech analytics system having multiple speech analytics engines |
US20150350118A1 (en) * | 2014-05-30 | 2015-12-03 | Apple Inc. | Canned answers in messages |
US9043196B1 (en) * | 2014-07-07 | 2015-05-26 | Machine Zone, Inc. | Systems and methods for identifying and suggesting emoticons |
US20160042359A1 (en) * | 2014-08-11 | 2016-02-11 | 24/7 Customer, Inc. | Methods and apparatuses for modeling customer interaction experiences |
Non-Patent Citations (37)
Title |
---|
Adolphs, Ralph, "Recognizing Emotion From Facial Expressions: Psychological and Neurological Mechanisms", University of Iowa College of Medicine, Behavioral and Cognitive Neuroscience Reviews vol. 1 No. 1, Mar. 2002, pp. 21-62. |
Aman, Saima et al., "Identifying Expressions of Emotion in Text", TSD 2007, LNAI 4629, Springer-Verlag Berlin Heidelberg 2007, pp. 196-205. |
Aman, Saima, "Recognizing Emotions in Text", Master of Computer Science Thesis, University of Ottawa, 2007, 105 pp. |
Blagdon, Jeff, "How emoji conquered the world", The Verge, Mar. 4, 2013, online <http://www.theverge.com/2013/3/4/3966140/how-emoji-conquered-the-world>. |
Chalabi, Mona, "The 100 Most-Used Emojis", FiveThirtyEight, Jun. 5, 2014, online <http://fivethirtyeight.com/datalab/the-100-most-used-emojis/>, 7 pp. |
Chandler, Clive et al., "Biometric Measurement of Human Emotions", International Journal of Computing Science and Communication Technologies, vol. 4 No. 2, Jan. 2012, 5 pp. |
Conati, Cristina et al., "A Study on Using Biometric Sensors for Monitoring User Emotions in Educational Games", Workshop on Assessing and Adapting to User Attitudes and Affect: Why, When and How. 2003. 7 pp. |
Cullen, Charlie et al., "Generation of High Quality Audio Natural Emotional Speech Corpus using Task Based Mood Induction", International Conference on Multidisciplinary Information Sciences and Technologies Extremadura (InSciT), Merida, Spain. Oct. 25-28, 2006, 10 pp. |
Devillers et al. "Annotation and Detection of Emotion in a Task-oreiented Human-Human Dialog Corpus," ISLE workshop, Dec. 2002, 10 pages. * |
Erdem, Ernur Sonat et al., "Efficient Recognition of Human Emotional States from Audio Signals", 2014 IEEE International Symposium on Multimedia, Dec. 10-12, 2014, 4 pp. |
Fischer, Agneta H. et al., "Social influences on the emotion process", Chapter in Stroebe, W. & Hewstone, M. (2003). European Review of Social Psychology, 14, pp. 171-202. |
Foltz, Peter W., "Latent Semantic Analysis for Text-Based Research", online <http://www-psych.nmsu.edu/˜pfoltz/reprints/BRMIC96.html>, Behavior Research Methods, Instruments and Computers. 28(2), pp. 197-202. |
Ismael, Chris, "List of 20+ Sentiment Analysis APIs", Apr. 24, 2013, online <http://blog.mashape.com/list-of-20-sentiment-analysis-apis/>, 4 pp. |
Kalyan, Chetan et al., "Detecting emotional scenes using Semantic Analysis on Subtitles", Stanford University, CS224N Final Project, Jun. 4, 2009, 11 pp. |
Kim, Sunghwan Mac, "Recognising Emotions and Sentiments in Text", Thesis-University of Sydney, Apr. 2011, 128 pp. |
Kim, Sunghwan Mac, "Recognising Emotions and Sentiments in Text", Thesis—University of Sydney, Apr. 2011, 128 pp. |
Klingensmith, Ashton, "The Capacity to Delineate and Interpret Emotion in Text Messages", Thesis-Liberty University, 2012, 34 pp. |
Klingensmith, Ashton, "The Capacity to Delineate and Interpret Emotion in Text Messages", Thesis—Liberty University, 2012, 34 pp. |
LiKamWa, Robert et al., "MoodScope: Building a Mood Sensor from Smartphone Usage Patterns", MobiSys'13, Jun. 25-28, 2013, Taipei, Taiwan. 13 pp. |
List of IBM Patents or Applications Treated as Related, dated Jan. 7, 2015. |
Mihalcea, Rada et al., "Corpus-based and Knowledge-based Measures of Text Semantic Similarity", 2006, American Association for Artificial Intelligence (www.aaai.org), pp. 775-780. |
Neviarouskaya et al., "Textual Affect Sensing for Sociable and Expressive Online Communication, " ACII 2007, LNCS 4738, 2007 Springer-Verlag, pp. 218-229. * |
Polzin, Thomas S. et al., "Detecting Emotions in Speech", Proceedings of the CMC, 1998, 7 pp. |
Ptaszynski, Michael, "Towards Context Aware Emotional Intelligence in Machines: Computing Contextual Appropriateness of Affective States", International Joint Conferences on Artificial Intelligence, Jul. 2009, pp. 1469-1474. |
Ramirez, Geovany et al., "Color Analysis of Facial Skin: Detection of Emotional State", 2014 IEEE Conference on Computer Vision and Pattern Recognition Workshops, Jun. 23-28, 2014, pp. 474-479. |
Reyes et al., "A multidimensional approach for detecting irony in Twitter," Lang Resources & Evaluation, 2013, vol. 47, pp. 239-268. * |
Schut, Marleen et al., "Biometrics for Emotion Detection (BED): Exploring the combination of Speech and ECG", 1st International Workshop on Bio-inspired Human-Machine Interfaces and Healthcare Applications-B-Interface 2010, Jan. 21, 2010, Valencia, Spain , pp. 51-58. |
Schut, Marleen et al., "Biometrics for Emotion Detection (BED): Exploring the combination of Speech and ECG", 1st International Workshop on Bio-inspired Human-Machine Interfaces and Healthcare Applications—B-Interface 2010, Jan. 21, 2010, Valencia, Spain , pp. 51-58. |
Semantria Web Demo, accessed online Sep. 21, 2015, <https://semantria.com/demo>, 2 pp. |
Sentiment Analysis, wikipedia.com, Jun. 27, 2014, online <https://en.wikipedia.org/wiki/Sentiment-analysis>, 7 pp. |
Sentiment Analysis, wikipedia.com, Jun. 27, 2014, online <https://en.wikipedia.org/wiki/Sentiment—analysis>, 7 pp. |
Shivhare, Shiv Naresh et al., "Emotion Detection from Text", Department of CSE and IT, Maulana Azad National Institute of Technology, Bhopal, Madhya Pradesh, India, 2012, 7 pp. |
Strapparava, Carlo et al., "Learning to Identify Emotions in Text", SAC'08 Mar. 16-20, 2008, Fortaleza, Cear'a, Brazil, 5 pp. |
Surbhi, Vishal Arora, "A Face Identification Technique for Human Facial Image", International Journal of Computer Science and Information Technologies, vol. 3 (6), 2012, pp. 5390-5393. |
Ulinski, Morgan et al., "Finding Emotion in Image Descriptions", WISDOM'12, Aug. 12, 2012, Beijing, China, 7 pp. |
Weiss-Meyer, Amy, "A Peek Inside the Non-Profit Consortium That Makes Emoji Possible", The New Republic, Jun. 27, 2014, online <http://www.newrepublic.com/article/118421/emoji-made-possible-non-profit-consortium>, 3 pp. |
Yazdani, Majid et al., "Computing Text Semantic Relatedness using the Contents and Links of a Hypertext Encyclopedia", Artificial Intelligence, Nov. 19, 2012, 32 pp. |
Cited By (170)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US11928604B2 (en) | 2005-09-08 | 2024-03-12 | Apple Inc. | Method and apparatus for building an intelligent automated assistant |
US11671920B2 (en) | 2007-04-03 | 2023-06-06 | Apple Inc. | Method and system for operating a multifunction portable electronic device using voice-activation |
US11979836B2 (en) | 2007-04-03 | 2024-05-07 | Apple Inc. | Method and system for operating a multi-function portable electronic device using voice-activation |
US11348582B2 (en) | 2008-10-02 | 2022-05-31 | Apple Inc. | Electronic devices with voice command and contextual data processing capabilities |
US11900936B2 (en) | 2008-10-02 | 2024-02-13 | Apple Inc. | Electronic devices with voice command and contextual data processing capabilities |
US11423886B2 (en) | 2010-01-18 | 2022-08-23 | Apple Inc. | Task flow identification based on user intent |
US12087308B2 (en) | 2010-01-18 | 2024-09-10 | Apple Inc. | Intelligent automated assistant |
US10741185B2 (en) | 2010-01-18 | 2020-08-11 | Apple Inc. | Intelligent automated assistant |
US10692504B2 (en) | 2010-02-25 | 2020-06-23 | Apple Inc. | User profiling for voice input processing |
US10417405B2 (en) | 2011-03-21 | 2019-09-17 | Apple Inc. | Device access using voice authentication |
US11120372B2 (en) | 2011-06-03 | 2021-09-14 | Apple Inc. | Performing actions associated with task items that represent tasks to perform |
US11321116B2 (en) | 2012-05-15 | 2022-05-03 | Apple Inc. | Systems and methods for integrating third party services with a digital assistant |
US11269678B2 (en) | 2012-05-15 | 2022-03-08 | Apple Inc. | Systems and methods for integrating third party services with a digital assistant |
US11636869B2 (en) | 2013-02-07 | 2023-04-25 | Apple Inc. | Voice trigger for a digital assistant |
US12009007B2 (en) | 2013-02-07 | 2024-06-11 | Apple Inc. | Voice trigger for a digital assistant |
US11557310B2 (en) | 2013-02-07 | 2023-01-17 | Apple Inc. | Voice trigger for a digital assistant |
US10978090B2 (en) | 2013-02-07 | 2021-04-13 | Apple Inc. | Voice trigger for a digital assistant |
US11862186B2 (en) | 2013-02-07 | 2024-01-02 | Apple Inc. | Voice trigger for a digital assistant |
US10714117B2 (en) | 2013-02-07 | 2020-07-14 | Apple Inc. | Voice trigger for a digital assistant |
US11388291B2 (en) | 2013-03-14 | 2022-07-12 | Apple Inc. | System and method for processing voicemail |
US11798547B2 (en) | 2013-03-15 | 2023-10-24 | Apple Inc. | Voice activated device for use with a voice-based digital assistant |
US11048473B2 (en) | 2013-06-09 | 2021-06-29 | Apple Inc. | Device, method, and graphical user interface for enabling conversation persistence across two or more instances of a digital assistant |
US10769385B2 (en) | 2013-06-09 | 2020-09-08 | Apple Inc. | System and method for inferring user intent from speech inputs |
US11727219B2 (en) | 2013-06-09 | 2023-08-15 | Apple Inc. | System and method for inferring user intent from speech inputs |
US12073147B2 (en) | 2013-06-09 | 2024-08-27 | Apple Inc. | Device, method, and graphical user interface for enabling conversation persistence across two or more instances of a digital assistant |
US12010262B2 (en) | 2013-08-06 | 2024-06-11 | Apple Inc. | Auto-activating smart responses based on activities from remote devices |
US11314370B2 (en) | 2013-12-06 | 2022-04-26 | Apple Inc. | Method for extracting salient dialog usage from live data |
US11172026B2 (en) | 2014-05-19 | 2021-11-09 | Michael H. Auerbach | Dynamic computer systems and uses thereof |
US10666735B2 (en) | 2014-05-19 | 2020-05-26 | Auerbach Michael Harrison Tretter | Dynamic computer systems and uses thereof |
US11257504B2 (en) | 2014-05-30 | 2022-02-22 | Apple Inc. | Intelligent assistant for home automation |
US11699448B2 (en) | 2014-05-30 | 2023-07-11 | Apple Inc. | Intelligent assistant for home automation |
US12118999B2 (en) | 2014-05-30 | 2024-10-15 | Apple Inc. | Reducing the need for manual start/end-pointing and trigger phrases |
US11670289B2 (en) | 2014-05-30 | 2023-06-06 | Apple Inc. | Multi-command single utterance input method |
US11133008B2 (en) | 2014-05-30 | 2021-09-28 | Apple Inc. | Reducing the need for manual start/end-pointing and trigger phrases |
US10878809B2 (en) | 2014-05-30 | 2020-12-29 | Apple Inc. | Multi-command single utterance input method |
US10417344B2 (en) | 2014-05-30 | 2019-09-17 | Apple Inc. | Exemplar-based natural language processing |
US10699717B2 (en) | 2014-05-30 | 2020-06-30 | Apple Inc. | Intelligent assistant for home automation |
US11810562B2 (en) | 2014-05-30 | 2023-11-07 | Apple Inc. | Reducing the need for manual start/end-pointing and trigger phrases |
US12067990B2 (en) | 2014-05-30 | 2024-08-20 | Apple Inc. | Intelligent assistant for home automation |
US10714095B2 (en) | 2014-05-30 | 2020-07-14 | Apple Inc. | Intelligent assistant for home automation |
US10657966B2 (en) | 2014-05-30 | 2020-05-19 | Apple Inc. | Better resolution when referencing to concepts |
US11838579B2 (en) | 2014-06-30 | 2023-12-05 | Apple Inc. | Intelligent automated assistant for TV user interactions |
US11516537B2 (en) | 2014-06-30 | 2022-11-29 | Apple Inc. | Intelligent automated assistant for TV user interactions |
US10390213B2 (en) | 2014-09-30 | 2019-08-20 | Apple Inc. | Social reminders |
US10453443B2 (en) | 2014-09-30 | 2019-10-22 | Apple Inc. | Providing an indication of the suitability of speech recognition |
US10438595B2 (en) | 2014-09-30 | 2019-10-08 | Apple Inc. | Speaker identification and unsupervised speaker adaptation techniques |
US11231904B2 (en) | 2015-03-06 | 2022-01-25 | Apple Inc. | Reducing response latency of intelligent automated assistants |
US11842734B2 (en) | 2015-03-08 | 2023-12-12 | Apple Inc. | Virtual assistant activation |
US10930282B2 (en) | 2015-03-08 | 2021-02-23 | Apple Inc. | Competing devices responding to voice triggers |
US10529332B2 (en) | 2015-03-08 | 2020-01-07 | Apple Inc. | Virtual assistant activation |
US11087759B2 (en) | 2015-03-08 | 2021-08-10 | Apple Inc. | Virtual assistant activation |
US11468282B2 (en) | 2015-05-15 | 2022-10-11 | Apple Inc. | Virtual assistant in a communication session |
US12001933B2 (en) | 2015-05-15 | 2024-06-04 | Apple Inc. | Virtual assistant in a communication session |
US11127397B2 (en) | 2015-05-27 | 2021-09-21 | Apple Inc. | Device voice control |
US11070949B2 (en) | 2015-05-27 | 2021-07-20 | Apple Inc. | Systems and methods for proactively identifying and surfacing relevant content on an electronic device with a touch-sensitive display |
US10681212B2 (en) | 2015-06-05 | 2020-06-09 | Apple Inc. | Virtual assistant aided communication with 3rd party service in a communication session |
US11947873B2 (en) | 2015-06-29 | 2024-04-02 | Apple Inc. | Virtual assistant for media playback |
US11010127B2 (en) | 2015-06-29 | 2021-05-18 | Apple Inc. | Virtual assistant for media playback |
US11126400B2 (en) | 2015-09-08 | 2021-09-21 | Apple Inc. | Zero latency digital assistant |
US11954405B2 (en) | 2015-09-08 | 2024-04-09 | Apple Inc. | Zero latency digital assistant |
US11550542B2 (en) | 2015-09-08 | 2023-01-10 | Apple Inc. | Zero latency digital assistant |
US11853536B2 (en) | 2015-09-08 | 2023-12-26 | Apple Inc. | Intelligent automated assistant in a media environment |
US11809483B2 (en) | 2015-09-08 | 2023-11-07 | Apple Inc. | Intelligent automated assistant for media search and playback |
US11500672B2 (en) | 2015-09-08 | 2022-11-15 | Apple Inc. | Distributed personal assistant |
US12051413B2 (en) | 2015-09-30 | 2024-07-30 | Apple Inc. | Intelligent device identification |
US11809886B2 (en) | 2015-11-06 | 2023-11-07 | Apple Inc. | Intelligent automated assistant in a messaging environment |
US11526368B2 (en) | 2015-11-06 | 2022-12-13 | Apple Inc. | Intelligent automated assistant in a messaging environment |
US10956666B2 (en) | 2015-11-09 | 2021-03-23 | Apple Inc. | Unconventional virtual assistant interactions |
US11886805B2 (en) | 2015-11-09 | 2024-01-30 | Apple Inc. | Unconventional virtual assistant interactions |
US11853647B2 (en) | 2015-12-23 | 2023-12-26 | Apple Inc. | Proactive assistance based on dialog communication between devices |
US10942703B2 (en) | 2015-12-23 | 2021-03-09 | Apple Inc. | Proactive assistance based on dialog communication between devices |
US11227589B2 (en) | 2016-06-06 | 2022-01-18 | Apple Inc. | Intelligent list reading |
US11657820B2 (en) | 2016-06-10 | 2023-05-23 | Apple Inc. | Intelligent digital assistant in a multi-tasking environment |
US11037565B2 (en) | 2016-06-10 | 2021-06-15 | Apple Inc. | Intelligent digital assistant in a multi-tasking environment |
US11749275B2 (en) | 2016-06-11 | 2023-09-05 | Apple Inc. | Application integration with a digital assistant |
US11809783B2 (en) | 2016-06-11 | 2023-11-07 | Apple Inc. | Intelligent device arbitration and control |
US11152002B2 (en) | 2016-06-11 | 2021-10-19 | Apple Inc. | Application integration with a digital assistant |
US10580409B2 (en) | 2016-06-11 | 2020-03-03 | Apple Inc. | Application integration with a digital assistant |
US10063929B1 (en) | 2016-09-01 | 2018-08-28 | Nufbee Llc | Community controlled audio entertainment system |
US10798044B1 (en) | 2016-09-01 | 2020-10-06 | Nufbee Llc | Method for enhancing text messages with pre-recorded audio clips |
US10474753B2 (en) | 2016-09-07 | 2019-11-12 | Apple Inc. | Language identification using recurrent neural networks |
US11550751B2 (en) * | 2016-11-18 | 2023-01-10 | Microsoft Technology Licensing, Llc | Sequence expander for data entry/information retrieval |
US11656884B2 (en) | 2017-01-09 | 2023-05-23 | Apple Inc. | Application integration with a digital assistant |
US10558757B2 (en) | 2017-03-11 | 2020-02-11 | International Business Machines Corporation | Symbol management |
US10388034B2 (en) * | 2017-04-24 | 2019-08-20 | International Business Machines Corporation | Augmenting web content to improve user experience |
US10741181B2 (en) | 2017-05-09 | 2020-08-11 | Apple Inc. | User interface for correcting recognition errors |
US10417266B2 (en) | 2017-05-09 | 2019-09-17 | Apple Inc. | Context-aware ranking of intelligent response suggestions |
US11599331B2 (en) | 2017-05-11 | 2023-03-07 | Apple Inc. | Maintaining privacy of personal information |
US10395654B2 (en) | 2017-05-11 | 2019-08-27 | Apple Inc. | Text normalization based on a data-driven learning network |
US11467802B2 (en) | 2017-05-11 | 2022-10-11 | Apple Inc. | Maintaining privacy of personal information |
US11580990B2 (en) | 2017-05-12 | 2023-02-14 | Apple Inc. | User-specific acoustic models |
US11837237B2 (en) | 2017-05-12 | 2023-12-05 | Apple Inc. | User-specific acoustic models |
US11862151B2 (en) | 2017-05-12 | 2024-01-02 | Apple Inc. | Low-latency intelligent automated assistant |
US11538469B2 (en) | 2017-05-12 | 2022-12-27 | Apple Inc. | Low-latency intelligent automated assistant |
US11405466B2 (en) | 2017-05-12 | 2022-08-02 | Apple Inc. | Synchronization and task delegation of a digital assistant |
US11380310B2 (en) | 2017-05-12 | 2022-07-05 | Apple Inc. | Low-latency intelligent automated assistant |
US11301477B2 (en) | 2017-05-12 | 2022-04-12 | Apple Inc. | Feedback analysis of a digital assistant |
US12014118B2 (en) | 2017-05-15 | 2024-06-18 | Apple Inc. | Multi-modal interfaces having selection disambiguation and text modification capability |
US10909171B2 (en) | 2017-05-16 | 2021-02-02 | Apple Inc. | Intelligent automated assistant for media exploration |
US10311144B2 (en) * | 2017-05-16 | 2019-06-04 | Apple Inc. | Emoji word sense disambiguation |
US11675829B2 (en) | 2017-05-16 | 2023-06-13 | Apple Inc. | Intelligent automated assistant for media exploration |
US12026197B2 (en) | 2017-05-16 | 2024-07-02 | Apple Inc. | Intelligent automated assistant for media exploration |
US10748546B2 (en) | 2017-05-16 | 2020-08-18 | Apple Inc. | Digital assistant services based on device capabilities |
US11532306B2 (en) | 2017-05-16 | 2022-12-20 | Apple Inc. | Detecting a trigger of a digital assistant |
US10348659B1 (en) * | 2017-12-21 | 2019-07-09 | International Business Machines Corporation | Chat message processing |
US10592604B2 (en) | 2018-03-12 | 2020-03-17 | Apple Inc. | Inverse text normalization for automatic speech recognition |
US11710482B2 (en) | 2018-03-26 | 2023-07-25 | Apple Inc. | Natural assistant interaction |
US11854539B2 (en) | 2018-05-07 | 2023-12-26 | Apple Inc. | Intelligent automated assistant for delivering content from user experiences |
US11169616B2 (en) | 2018-05-07 | 2021-11-09 | Apple Inc. | Raise to speak |
US11907436B2 (en) | 2018-05-07 | 2024-02-20 | Apple Inc. | Raise to speak |
US11487364B2 (en) | 2018-05-07 | 2022-11-01 | Apple Inc. | Raise to speak |
US11900923B2 (en) | 2018-05-07 | 2024-02-13 | Apple Inc. | Intelligent automated assistant for delivering content from user experiences |
US11431642B2 (en) | 2018-06-01 | 2022-08-30 | Apple Inc. | Variable latency device coordination |
US10892996B2 (en) | 2018-06-01 | 2021-01-12 | Apple Inc. | Variable latency device coordination |
US10403283B1 (en) | 2018-06-01 | 2019-09-03 | Apple Inc. | Voice interaction at a primary device to access call functionality of a companion device |
US11009970B2 (en) | 2018-06-01 | 2021-05-18 | Apple Inc. | Attention aware virtual assistant dismissal |
US12080287B2 (en) | 2018-06-01 | 2024-09-03 | Apple Inc. | Voice interaction at a primary device to access call functionality of a companion device |
US11495218B2 (en) | 2018-06-01 | 2022-11-08 | Apple Inc. | Virtual assistant operation in multi-device environments |
US10984798B2 (en) | 2018-06-01 | 2021-04-20 | Apple Inc. | Voice interaction at a primary device to access call functionality of a companion device |
US11630525B2 (en) | 2018-06-01 | 2023-04-18 | Apple Inc. | Attention aware virtual assistant dismissal |
US12067985B2 (en) | 2018-06-01 | 2024-08-20 | Apple Inc. | Virtual assistant operations in multi-device environments |
US11360577B2 (en) | 2018-06-01 | 2022-06-14 | Apple Inc. | Attention aware virtual assistant dismissal |
US12061752B2 (en) | 2018-06-01 | 2024-08-13 | Apple Inc. | Attention aware virtual assistant dismissal |
US10720160B2 (en) | 2018-06-01 | 2020-07-21 | Apple Inc. | Voice interaction at a primary device to access call functionality of a companion device |
US10944859B2 (en) | 2018-06-03 | 2021-03-09 | Apple Inc. | Accelerated task performance |
US10504518B1 (en) | 2018-06-03 | 2019-12-10 | Apple Inc. | Accelerated task performance |
US10496705B1 (en) | 2018-06-03 | 2019-12-03 | Apple Inc. | Accelerated task performance |
US11010561B2 (en) | 2018-09-27 | 2021-05-18 | Apple Inc. | Sentiment prediction from textual data |
US11462215B2 (en) | 2018-09-28 | 2022-10-04 | Apple Inc. | Multi-modal inputs for voice commands |
US11893992B2 (en) | 2018-09-28 | 2024-02-06 | Apple Inc. | Multi-modal inputs for voice commands |
US11170166B2 (en) | 2018-09-28 | 2021-11-09 | Apple Inc. | Neural typographical error modeling via generative adversarial networks |
US10839159B2 (en) | 2018-09-28 | 2020-11-17 | Apple Inc. | Named entity normalization in a spoken dialog system |
US11074398B2 (en) | 2018-10-12 | 2021-07-27 | International Business Machines Corporation | Tracking and managing emoji annotations |
US11475898B2 (en) | 2018-10-26 | 2022-10-18 | Apple Inc. | Low-latency multi-speaker speech recognition |
US10909328B2 (en) * | 2019-01-04 | 2021-02-02 | International Business Machines Corporation | Sentiment adapted communication |
US11638059B2 (en) | 2019-01-04 | 2023-04-25 | Apple Inc. | Content playback on multiple devices |
US20200218781A1 (en) * | 2019-01-04 | 2020-07-09 | International Business Machines Corporation | Sentiment adapted communication |
US11783815B2 (en) | 2019-03-18 | 2023-10-10 | Apple Inc. | Multimodality in digital assistant systems |
US11348573B2 (en) | 2019-03-18 | 2022-05-31 | Apple Inc. | Multimodality in digital assistant systems |
US11307752B2 (en) | 2019-05-06 | 2022-04-19 | Apple Inc. | User configurable task triggers |
US11217251B2 (en) | 2019-05-06 | 2022-01-04 | Apple Inc. | Spoken notifications |
US11675491B2 (en) | 2019-05-06 | 2023-06-13 | Apple Inc. | User configurable task triggers |
US11475884B2 (en) | 2019-05-06 | 2022-10-18 | Apple Inc. | Reducing digital assistant latency when a language is incorrectly determined |
US11705130B2 (en) | 2019-05-06 | 2023-07-18 | Apple Inc. | Spoken notifications |
US11423908B2 (en) | 2019-05-06 | 2022-08-23 | Apple Inc. | Interpreting spoken requests |
US11888791B2 (en) | 2019-05-21 | 2024-01-30 | Apple Inc. | Providing message response suggestions |
US11140099B2 (en) | 2019-05-21 | 2021-10-05 | Apple Inc. | Providing message response suggestions |
US11237797B2 (en) | 2019-05-31 | 2022-02-01 | Apple Inc. | User activity shortcut suggestions |
US11360739B2 (en) | 2019-05-31 | 2022-06-14 | Apple Inc. | User activity shortcut suggestions |
US11657813B2 (en) | 2019-05-31 | 2023-05-23 | Apple Inc. | Voice identification in digital assistant systems |
US11496600B2 (en) | 2019-05-31 | 2022-11-08 | Apple Inc. | Remote execution of machine-learned models |
US11289073B2 (en) | 2019-05-31 | 2022-03-29 | Apple Inc. | Device text to speech |
US11360641B2 (en) | 2019-06-01 | 2022-06-14 | Apple Inc. | Increasing the relevance of new available information |
US11790914B2 (en) | 2019-06-01 | 2023-10-17 | Apple Inc. | Methods and user interfaces for voice-based control of electronic devices |
US11488406B2 (en) | 2019-09-25 | 2022-11-01 | Apple Inc. | Text detection using global geometry estimators |
US11924254B2 (en) | 2020-05-11 | 2024-03-05 | Apple Inc. | Digital assistant hardware abstraction |
US11765209B2 (en) | 2020-05-11 | 2023-09-19 | Apple Inc. | Digital assistant hardware abstraction |
US11914848B2 (en) | 2020-05-11 | 2024-02-27 | Apple Inc. | Providing relevant data items based on context |
US11755276B2 (en) | 2020-05-12 | 2023-09-12 | Apple Inc. | Reducing description length based on confidence |
US11838734B2 (en) | 2020-07-20 | 2023-12-05 | Apple Inc. | Multi-device audio adjustment coordination |
US11750962B2 (en) | 2020-07-21 | 2023-09-05 | Apple Inc. | User identification using headphones |
US11696060B2 (en) | 2020-07-21 | 2023-07-04 | Apple Inc. | User identification using headphones |
US11283751B1 (en) | 2020-11-03 | 2022-03-22 | International Business Machines Corporation | Using speech and facial bio-metrics to deliver text messages at the appropriate time |
US11516539B2 (en) | 2021-03-01 | 2022-11-29 | Comcast Cable Communications, Llc | Systems and methods for providing contextually relevant information |
US12003811B2 (en) | 2021-03-01 | 2024-06-04 | Comcast Cable Communications, Llc | Systems and methods for providing contextually relevant information |
WO2023009323A1 (en) * | 2021-07-29 | 2023-02-02 | Snap Inc. | Emoji recommendation system using user context and biosignals |
US11765115B2 (en) | 2021-07-29 | 2023-09-19 | Snap Inc. | Emoji recommendation system using user context and biosignals |
US11657558B2 (en) | 2021-09-16 | 2023-05-23 | International Business Machines Corporation | Context-based personalized communication presentation |
WO2023212259A1 (en) * | 2022-04-28 | 2023-11-02 | Theai, Inc. | Artificial intelligence character models with modifiable behavioral characteristics |
US12033086B2 (en) | 2022-04-28 | 2024-07-09 | Theai, Inc. | Artificial intelligence character models with modifiable behavioral characteristics |
Also Published As
Publication number | Publication date |
---|---|
US20170083506A1 (en) | 2017-03-23 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US9665567B2 (en) | Suggesting emoji characters based on current contextual emotional state of user | |
US10783711B2 (en) | Switching realities for better task efficiency | |
US9299268B2 (en) | Tagging scanned data with emotional tags, predicting emotional reactions of users to data, and updating historical user emotional reactions to data | |
US10373273B2 (en) | Evaluating an impact of a user's content utilized in a social network | |
Bragg et al. | The fate landscape of sign language ai datasets: An interdisciplinary perspective | |
US20190251638A1 (en) | Identification of life events within social media conversations | |
US11756567B2 (en) | Autocreation of conversational image representation | |
US11182447B2 (en) | Customized display of emotionally filtered social media content | |
US11861493B2 (en) | Machine learning models based on altered data and systems and methods for training and using the same | |
US10223440B2 (en) | Question and answer system emulating people and clusters of blended people | |
US10067935B2 (en) | Prediction and optimized prevention of bullying and other counterproductive interactions in live and virtual meeting contexts | |
US20170132208A1 (en) | Personalized paraphrasing for reading improvement | |
US20200135039A1 (en) | Content pre-personalization using biometric data | |
US9661474B2 (en) | Identifying topic experts among participants in a conference call | |
US20200125671A1 (en) | Altering content based on machine-learned topics of interest | |
US20190139447A1 (en) | Cognitive real-time feedback speaking coach on a mobile device | |
JP6776310B2 (en) | User-Real-time feedback information provision methods and systems associated with input content | |
US11169667B2 (en) | Profile picture management tool on social media platform | |
Catling et al. | The effects of age of acquisition on an object classification task | |
US11138367B2 (en) | Dynamic interaction behavior commentary | |
US11822599B2 (en) | Visualization resonance for collaborative discourse | |
US20220108624A1 (en) | Reader assistance method and system for comprehension checks | |
US10296723B2 (en) | Managing companionship data | |
US11558471B1 (en) | Multimedia content differentiation | |
US11734588B2 (en) | Managing domain competence during a computing session |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
AS | Assignment |
Owner name: INTERNATIONAL BUSINESS MACHINES CORPORATION, NEW Y Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:LIU, SU;ROZNER, ERIC J.;SZE, CHIN NGAI;AND OTHERS;REEL/FRAME:036612/0835 Effective date: 20150828 |
|
STCF | Information on status: patent grant |
Free format text: PATENTED CASE |
|
FEPP | Fee payment procedure |
Free format text: MAINTENANCE FEE REMINDER MAILED (ORIGINAL EVENT CODE: REM.); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY |
|
LAPS | Lapse for failure to pay maintenance fees |
Free format text: PATENT EXPIRED FOR FAILURE TO PAY MAINTENANCE FEES (ORIGINAL EVENT CODE: EXP.); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY |
|
STCH | Information on status: patent discontinuation |
Free format text: PATENT EXPIRED DUE TO NONPAYMENT OF MAINTENANCE FEES UNDER 37 CFR 1.362 |
|
FP | Lapsed due to failure to pay maintenance fee |
Effective date: 20210530 |