WO2010151700A1 - System and method for creating, editing, and accessing an intelligent avatar - Google Patents
System and method for creating, editing, and accessing an intelligent avatar Download PDFInfo
- Publication number
- WO2010151700A1 WO2010151700A1 PCT/US2010/039867 US2010039867W WO2010151700A1 WO 2010151700 A1 WO2010151700 A1 WO 2010151700A1 US 2010039867 W US2010039867 W US 2010039867W WO 2010151700 A1 WO2010151700 A1 WO 2010151700A1
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- data
- avatar
- logic
- user
- computer representation
- Prior art date
Links
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F16/00—Information retrieval; Database structures therefor; File system structures therefor
- G06F16/30—Information retrieval; Database structures therefor; File system structures therefor of unstructured textual data
- G06F16/33—Querying
- G06F16/332—Query formulation
- G06F16/3329—Natural language query formulation or dialogue systems
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F16/00—Information retrieval; Database structures therefor; File system structures therefor
- G06F16/30—Information retrieval; Database structures therefor; File system structures therefor of unstructured textual data
- G06F16/33—Querying
- G06F16/3331—Query processing
-
- A—HUMAN NECESSITIES
- A63—SPORTS; GAMES; AMUSEMENTS
- A63F—CARD, BOARD, OR ROULETTE GAMES; INDOOR GAMES USING SMALL MOVING PLAYING BODIES; VIDEO GAMES; GAMES NOT OTHERWISE PROVIDED FOR
- A63F2300/00—Features of games using an electronically generated display having two or more dimensions, e.g. on a television screen, showing representations related to the game
- A63F2300/50—Features of games using an electronically generated display having two or more dimensions, e.g. on a television screen, showing representations related to the game characterized by details of game servers
- A63F2300/55—Details of game data or player data management
- A63F2300/5546—Details of game data or player data management using player registration data, e.g. identification, account, preferences, game history
- A63F2300/5553—Details of game data or player data management using player registration data, e.g. identification, account, preferences, game history user representation in the game field, e.g. avatar
Definitions
- avatar refers to a computer representation of person or thing in the form of a visual model, which can be a two-dimensional or three-dimensional model. Oftentimes, avatars are used in computer games or on social-networking Internet forums.
- Some avatars have a limited ability to move and speak. In this regard, some avatars can answer a question that is asked by a user of a computer. Further, such an avatar may make slight movements during communication with the computer user.
- FIG. 1 depicts an intelligent avatar creating, editing, and runtime system 100 in accordance with an embodiment of the present disclosure.
- the system 100 comprises a client computing device 101 and a server computing device 102.
- the client computing device 101 and the server computing device 102 communicate over a network 103.
- the network 103 may be any type of network known in the art or future-developed.
- the server 101 may be connected to the computing devices 102 via the Internet, a local area network (LAN), or a wide area network (WAN).
- LAN local area network
- WAN wide area network
- the client computing device 101 comprises user interface logic 104.
- the client computing device 101 is a personal computer (PC).
- the client computing device 101 may be, for example, a mobile device or a personal digital assistant (PDA).
- PDA personal digital assistant
- the user interface logic 104 may be, for example, a web browser and/or a Flash Player, as in a thin client configuration, which is described further with reference to FIG. 3.
- the system 100 may also be implemented via a thick client configuration, which is described further with reference to FIG. 4.
- the present disclosure contemplates any type of web browser (not shown) that allows a user A to display and interact with information located on a web page (not shown), which may be accessible via the network 103.
- the present disclosure contemplates any type of flash player that allows the user A to view animations and movies in the web browser.
- the server 102 comprises intelligent avatar creation/editing logic 105, intelligent avatar data 106, and intelligent avatar runtime logic 107.
- the intelligent avatar creation logic 105 creates intelligent avatar data 106 indicative of an intelligent avatar (not shown) based upon input from the user A received via the user interface logic 104, which is described in more detail herein.
- avatar refers to a computer representation (not shown) of a human being that can be personalized based upon desired characteristics of a user A of the client computing device 101 of the system 100.
- the intelligent avatar data 106 may be data indicative of a two-dimensional or three-dimensional graphical representation of the avatar, a voice font, a brain type, a personality type, or the like. Such is described in more detail hereafter.
- the intelligent avatar runtime logic 107 creates, upon request, an avatar (not shown) based on the intelligent avatar data 106.
- the avatar created exhibits those desired characteristics identified by user A in creating the avatar.
- the intelligent avatar runtime logic 107 enables communication with the avatar over the network 103 and can allow the intelligent avatar to learn new and different information based upon data received during such communication.
- the avatar comprises a text component, a visual component and an audio component.
- the intelligent avatar runtime logic 107 displays the generated avatar, through the user interface logic 104 to user A.
- User A can type in questions for the avatar, and the intelligent avatar runtime logic 107 transmits animation and voice data to the user interface logic 104, such that the avatar appears to answer questions asked by the user A. This is described further herein.
- the user types in the questions for the avatar.
- the user may speak into a microphone, for example, the questions the user has for the avatar.
- the intelligent avatar creating, editing, and runtime system 100 in accordance with another embodiment of the present disclosure further comprises a third-party computing device 108.
- the server computing device 102 communicates with the third-party computing device 108 over the network 103.
- the third-party computing device 108 may be, for example, a Web server for transmitting content upon request to other users on the network 103.
- the third-party computing device 108 may host the Wikipedia ® Internet Web site, and the third-party computing device 108 may transmit content related to a particular subject to a user on the network 103 upon request.
- the third-party computing device 108 comprises third-party data 109.
- the third-party data 109 can be any data known in the art or future-developed.
- the third-party data 109 comprises data indicative of an identifier of a subject and data describing the subject.
- User A may transmit data textually or verbally to the intelligent avatar runtime logic 107 indicative of a question related to a particular topic.
- the intelligent avatar runtime logic 107 may search the intelligent avatar data 106 and not find any information relating to the particular topic. In such a situation, the intelligent avatar runtime logic 107 automatically transmits a request to the third-party computing device 108 requesting information related to the particular subject, and the third-party computing device 108 searches the third-party data 109, retrieves any information relating to the subject, and transmits the data retrieved back to the server computing device 102.
- the intelligent avatar runtime logic 107 may communicate the received data to the client computing device 101 so that User A can see information relating to the subject about which User A asked a question.
- the intelligent avatar runtime logic 107 may process the received data through the intelligent avatar creation/editing logic 105 such that the received data is then incorporated into the intelligent avatar data 106 for future retrieval.
- FIG. 2 depicts the server 102 and further illustrates the intelligent avatar creation/editing logic 105 in accordance with an embodiment of the present disclosure.
- the server 102 further comprises animation creation logic 200, voice font data 201, and brain and personality data 202.
- the brain and personality data 202 comprises data indicative of particular brain and/or personality types.
- the intelligent avatar creation logic 105 may display data indicative of the brain and/or personality types to the user A (FIG. 1), and the user A may select which brain and/or personality type that the user A desires to be incorporated into the avatar being created.
- Exemplary types of brains may be, for example, an extrovert brain type or an introvert brain type.
- the brain and personality data 202 comprises a logical set of rules that define behaviors of the avatar being created.
- a rule may equate, textually, the word "likes” with a list of words indicating what the avatar (and hence the user A) enjoys, e.g., hiking, reading, swimming, golfing, etc.
- the voice font data 201 comprises data indicative of a plurality of particular voice fonts.
- the intelligent avatar creation/editing logic 105 may display data indicative of the voice fonts to user A (FIG. 1), and user A may select which voice font that user A desires to be incorporated into the avatar being created.
- the intelligent avatar creation/editing logic 105 may allow user A to preselect and listen to each of the different voice fonts prior to the user A selecting one of the voice fonts for incorporation into the avatar.
- the server computing device 102 may comprise some type of voice synthesis logic (not shown) that generates digital representations of the voice of user A.
- the digital representation (not shown) may be generated, for example, from audio of the voice of user A and uploaded as particular voice font data 201.
- the voice synthesis logic may be any type of voice synthesis software known in the art or future-developed for generating a digital representation of the voice of User A.
- the voice synthesis logic 306 may be Nuance's RealSpeak, DragonSpeak, or Cepstral's VoiceForge.
- the animation creation logic 200 receives as input a photograph of a person, and generates animation data based upon the photograph submitted.
- the animation data may be data indicative of the person moving his/her lips and/or moving his/her head as the person moves his/her lips.
- the intelligent avatar creation/editing logic 105 may request a photograph from user A, which is used by the animation creation logic 200 to create the animation data 205.
- the intelligent avatar creation/editing logic 105 stores the data selected by user A during the avatar creation process as avatar A data 203 in the intelligent avatar data 106.
- the intelligent avatar data 106 is a compilation of data describing a plurality of avatars, e.g., avatar A data 203 and avatar B data 204.
- avatar A data 203 relates the selected brain and personality data, the voice font, and the animation data selected by user A.
- the avatar A data 203 further comprises conversation data and owner data.
- Conversation data is data indicative of conversations that third parties have with the created avatar, which is described further herein with reference to FIG. 3.
- the owner data can be data indicative of any additional information that the user desires to become a part of the avatar A data.
- user A can store as owner data e-mails, blogs, photographs,
- the owner data can be any type of data known in the art or future-developed that describes or is related to User A.
- owner data may be still images in any format known in the art or future-developed, including Joint Photographic Experts Group (JPEG), Graphic Interchange Format (GIF), or Portable Network Graphics (PNG).
- JPEG Joint Photographic Experts Group
- GIF Graphic Interchange Format
- PNG Portable Network Graphics
- the avatar A data 203 may be video, for example Windows Media Video (WMV) or Moving Pictures Expert Group (MPEG).
- the avatar A data 203 may be audio, including Windows Media Video (WMV) or Windows Wave (WAV).
- the owner data may be multimedia, including Adobe Flash Movie (SWF).
- FIG. 3 depicts the system 100 in accordance with an embodiment of the present disclosure having a plurality of users, including user A, user B, and user C.
- FIG. 3 depicts a more detailed view of the intelligent avatar runtime logic 107 and the intelligent avatar creation/editing logic 105.
- the intelligent avatar runtime logic 107 is invoked when user A, B, or C desires to interact with an avatar created using the intelligent avatar creation logic 105.
- the user A, B, C may desire to simply carry on a conversation with an avatar.
- the system 100 is a "thick client” configuration.
- the term “thick client” refers to a client/server system wherein the client of the system provides some of the functionality of the system independent of the server.
- the thick client configuration will be contrasted with a “thin client” configuration in the discussion related to FIG. 4.
- the term “thin client” refers to a system wherein the functionality of the system is resident on the server side of the system.
- the intelligent avatar runtime logic 107 comprises artificial intelligence (AI) logic 300, automated speech recognition logic 301, animation/gesture logic 302, and text to speech logic 303. Each of these logical components is used in interaction with an avatar stored on the server computing device 102. Note that each of the logical components 300-303 are shown for clarity as separate logical components; however, this should in no way limit the intelligent avatar runtime logic 107. Initially, the user A, B, C selects a desired avatar via a user interface (not shown).
- the user A, B, C has selected the avatar A corresponding with the avatar A data 203.
- the AI logic 300 is described hereinafter as interfacing with the avatar A data 203 of the intelligent avatar data 106.
- the intelligent avatar runtime logic 107 would interface with the data describing the selected other avatar.
- the avatar A data 203 is shown in the runtime example for exemplary purposes only.
- the intelligent avatar runtime logic 107 loads into random access memory (RAM) the data indicative of the avatar A data 203. By reading the data into RAM, the process of accessing the avatar A data 203 is more efficient. However, in another embodiment, the intelligent avatar runtime logic 107 accesses data as needed, for example, if the avatar A data is stored in a relational database ⁇
- the intelligent avatar runtime logic 107 receives text and/or audio input 304 from the user A, B, C.
- the user A, B, C may ask a question via a microphone or the user A, B, C may type in a question using a keyboard.
- data indicative of the question is transmitted via the network 103 to the server computing device 102.
- the received data 304 is transmitted to the automated speech recognition logic 301.
- the automated speech recognition logic 301 translates the received data into text, and provides textual data indicative of the audio received to the AI logic 300. If the text/ Audio input 304 comprises textual data, the textual data is transmitted to the AI logic 300.
- the AI logic 300 parses the text and identifies key words within the text. The
- AI logic 300 may identify additional concepts and/or text related to the key words identified. As an example, the question received from the user A, B, C may be "What do you like?" The AI logic 300 parses the text and identifies the word “like” as a key word. In addition, the AI logic 300 may associate the word "like” with the word “love.”
- the AI logic 300 then accesses the brain and personality data of the avatar A data 203.
- the AI intelligence logic 300 then generates a textual response to the question, which can be "I like swimming, but I love Chihuahuas!
- the textual response generated by the AI logic is transmitted to the text to speech logic 303.
- the text to speech logic generates an audio signal indicative of the response and based upon the voice font data associated with avatar A data 203.
- the animation/gesture logic 302 generates data indicative of the movement of the avatar as it recites the response. Such data indicative of movement of the avatar is generated based upon the animation data corresponding to avatar A data 203.
- text/audio/animation output 305 is transmitted to the user A, B, C that submitted the question "What do you like?"
- the user interface 104 of the client computing device 101 receives the text/audio/animation output 305, and plays for the user A, B, C an animation of the avatar saying "I like swimming, but I love Chihuahuas!
- the AI logic 300 stores conversation tracking data 306 indicative of most of the text received during the conversation that user A, B, C has with the avatar indicative of the avatar A data. This conversation data is stored in relation to the avatar A data 203.
- user A, B, C may desire to edit his/her own avatar.
- the intelligent avatar creation/editing logic 105 receives data indicative of additional characteristics related to the avatar A data 203.
- the intelligent avatar creation/editing logic 105 adds such additional information provided by user A to the owner data in the avatar A data 203.
- the AI logic 300 may read the owner data into the brain and personality data. In so doing, the AI logic 300 may use the newly added owner data to respond to queries made from other users, e.g., user B, C. In addition, the AI logic 300 may look within the user data to retrieve additional information that may be relevant to queries.
- the user data may have a photograph of user A's child.
- the AI logic 300 may transmit, in response to the query, text, audio, and animation data that includes "Yes, I have one male child named Michael,” and the AI logic 300 may display a photograph or video of the child.
- FIG. 4 depicts the system 100 in a thick client configuration. In such a configuration, each of the components behaves substantially similar to the like components of FIG. 3.
- the difference in the configuration of system 100 in FIG. 3 and 4 is that some of the functionality related to the system 100 is moved to the client computing device 101.
- the automated speech recognition logic 301, the text to speech logic 303, and the animation /gesture logic 302 are resident on the client computing device 101.
- text generated by the automated speech recognition logic 301 is provided over network 103 to the server 102. This is shown in FIG. 4 as text input 400. Accordingly, what is transmitted by the intelligent avatar runtime logic 107 is text or symbol output 401.
- the text to speech logic 303 receives the text and converts it to an audio signal.
- the animation/gesture logic 302 receives the symbol data indicative of the animation/gesture.
- the user interface logic 104 displays the avatar reciting a response using generated gestures and animation.
- FIG. 5 depicts an exemplary client computing device 101 in accordance with an embodiment of the present disclosure.
- the exemplary client computing device 101 comprises processor 500, display device 504, input device 505, and network device 503. Each of these components communicates over local interface 502, which can include one or more buses.
- the exemplary client computing device 101 further comprises the user interface logic 104, which is described hereinabove.
- the client 101 shown in FIG. 5 is the client 101 implemented in a thin client configuration, which was described in detail hereinabove.
- the user interface logic 104 can be software, hardware, or a combination thereof.
- user interface logic 104 is shown as software stored in memory 501.
- Memory 501 may be of any type of memory known in the art, including, but not limited to random access memory (RAM), read-only memory (ROM), flash memory, and the like.
- Processor 500 may be a digital processor or other type of circuitry configured to run the user interface logic 104 by processing and executing the instructions of the user interface logic 104.
- the processor 500 may be a processor having digital signal processor (DSP) architecture.
- DSP digital signal processor
- the processor 500 communicates to and drives the other elements within the client computing device 101 via the local interface 502.
- the network device 503 receives and/or transmits data over network 103
- the network device 503 may be, for example, an Ethernet card (not shown), which connects the client computing device 101 to the network 103.
- the display device 504 is a device for visually displaying an avatar (not shown) indicative of the text, audio, and animation data 305 (FIG. 3) received from the server computing device 102 (FIG. 3).
- the display device 504 displays one or more graphical user interfaces (not shown) for use by the user to interact with the avatar displayed to the display device 504, which is described further herein.
- the display device 504 may be, for example, a backlit liquid crystal display (LCD) screen. Other types of display devices may be used in other embodiments of the present disclosure.
- the input device 505 enables is for entering data into the client computing device 101.
- the input device 505 is a keyboard, and the user A (FIG. 3) uses the keyboard to input data relating to the avatar displayed.
- the input device may be a microphone, and user A can audibly enter data into the client computing device 102 related to the avatar.
- FIG. 6 depicts an exemplary server computing device 102 in accordance with an embodiment of the present disclosure.
- the exemplary server computing device 102 comprises processor 600 and network device 603. Each of these components communicates over local interface 602, which can include one or more buses.
- the exemplary server computing device 102 further comprises the intelligent avatar creation/editing logic 105 and the intelligent avatar runtime logic 107, as described hereinabove.
- the intelligent avatar creation/editing logic 105 and the intelligent avatar runtime logic 107 can be software, hardware, or a combination thereof.
- intelligent avatar creation logic 105 and the intelligent avatar runtime logic 107 are shown as software stored in memory 601.
- Memory 601 may be of any type of memory known in the art, including, but not limited to random access memory (RAM), read-only memory (ROM), flash memory, and the like.
- intelligent avatar creation/editing logic 105 and the intelligent avatar runtime logic 107 can be stored and transported on any computer-readable medium for use by or in connection with an instruction execution system, apparatus, or device, such as a computer-based system, processor-containing system, or other system that can fetch the instructions from the instruction execution system, apparatus, or device and execute the instructions.
- a "computer-readable medium” can be any means that can contain, store, communicate, propagate, or transport the program for use by or in connection with the instruction execution system, apparatus, or device.
- the computer readable medium can be, for example but not limited to, an electronic, magnetic, optical, electromagnetic, infrared, or semiconductor system, apparatus, device, or propagation medium
- Processor 600 may be a digital processor or other type of circuitry configured to run the intelligent avatar creation/editing logic 105 and the intelligent avatar runtime logic 107 by processing and executing the instructions of the intelligent avatar creation/editing logic 105 and the intelligent avatar runtime logic 107.
- the processor 600 may be a processor having digital signal processor (DSP) architecture.
- DSP digital signal processor
- the processor 600 communicates to and drives the other elements within the server computing device 102 via the local interface 602.
- the network device 603 receives and/or transmits data over network 103
- the network device 603 may be, for example, an Ethernet card (not shown), which connects the server computing device 102 to the network 103.
- the intelligent avatar creation/editing logic 105 may display one or more graphical user interfaces (GUIs) (not shown) that requests particular information. Such information, as described hereinabove, is used to generate the intelligent avatar data 106, which is stored in memory 601. An avatar embodying the intelligent avatar data 106 may then be accessed by a user A, B, C (FIG. 4) via the intelligent avatar runtime logic 107 or the intelligent avatar creation/editing logic 105.
- GUIs graphical user interfaces
- the intelligent avatar creation/editing logic 105 may display a basic personality test via a GUI (not shown) to user A (FIG. 4).
- the basic personality test would comprise, for example, questions indicative of a Jung-Myers- Briggs personality typing test.
- the intelligent avatar creation/editing logic 105 stores data indicative of the answers input by user A.
- the intelligent avatar creation/editing logic 105 may provide a list of brain and personality types from the brain and personality types data 202 (FIG. 2) from which the user could select a particular brain and personality type that most closely resembles the type of personality of the user based upon the test results.
- the intelligent avatar creation/editing logic 105 may provide additional questions to user A via the display device 506 (FIG. 5) in order to obtain additional information about User A, which can be used to create the intelligent avatar data 106.
- the intelligent avatar creation/editing logic 105 could obtain addition information related to topics such as birth line and genealogy, birth and childhood, the first day of school, the first date... etc. Such information would be obtained from the user via a set of questions targeted at the particular topic and data indicative of the answers stored as intelligent avatar data 106.
- FIG. 7 depicts an exemplary client computing device 402 in accordance with an embodiment of the present disclosure.
- the exemplary client computing device 402 comprises processor 700, display device 704, input device 705, and network device 703. Each of these components communicates over local interface 702, which can include one or more buses.
- the exemplary client computing device 402 further comprises the user interface logic 104, text to speech logic 303, automated speech recognition logic 301, and animation/gesture logic 302 which are described hereinabove.
- the client 402 shown in FIG. 7 is the client 402 implemented in a thick client configuration, which was described in detail hereinabove with reference to FIG. 4.
- the user interface logic 104, text to speech logic 303, automated speech recognition logic 301, and animation/gesture logic 302 can be software, hardware, or a combination thereof.
- user interface logic 104, text to speech logic 303, automated speech recognition logic 301, and animation/gesture logic 302 are shown as software stored in memory 701.
- Memory 701 may be of any type of memory known in the art, including, but not limited to random access memory (RAM), read-only memory (ROM), flash memory, and the like.
- Processor 700 may be a digital processor or other type of circuitry configured to run the user interface logic 104, text to speech logic 303, automated speech recognition logic 301, and animation/gesture logic 302 by processing and executing the instructions of the user interface logic 104, text to speech logic 303, automated speech recognition logic 301, and animation/gesture logic 302.
- the processor 700 may be a processor having digital signal processor (DSP) architecture.
- DSP digital signal processor
- the processor 700 communicates to and drives the other elements within the client computing device 402 via the local interface 702.
- the network device 703 receives and/or transmits data over network 103
- the network device 703 may be, for example, an Ethernet card (not shown), which connects the client computing device 402 to the network 103.
- the 302 receives text/symbol data 401 from the server computing device 102.
- the user interface logic 104 displays to the display device 704 an avatar (not shown) indicative of the text/symbol data 401 (FIG. 4) received from the server computing device 102 (FIG. 4).
- the display device 704 displays one or more graphical user interfaces (not shown) for use by the user to interact with the avatar displayed to the display device 604, which is described further herein.
- the display device 704 may be, for example, a backlit liquid crystal display (LCD) screen. Other types of display devices may be used in other embodiments of the present disclosure.
- LCD liquid crystal display
- the input device 705 enables is for entering data into the client computing device 402.
- the input device 705 is a keyboard, and the user A (FIG. 4) uses the keyboard to input data relating to the avatar displayed.
- the input device may be a microphone, and user A can audibly enter data into the client computing device 402 related to the avatar.
- FIG. 8 depicts a main graphical user interface (GUI) 800 in accordance with an embodiment of the present disclosure that is displayed by the intelligent avatar creation/editing logic 105 (FIG. 3).
- GUI graphical user interface
- an avatar 806 is being displayed in the GUI 800. Therefore, it is assumed for purposes of the following discussion that the avatar 806 has already been created based upon particularly selected brain and personality data 202 (FIG. 2), voice font data 201 (FIG. 2), and animation data 205 (FIG. X).
- the intelligent avatar creation logic 105 (FIG. 3) is displaying the GUI 800 and the avatar 906 to the display device 604 (FIG. 6) of a client computing device 101 (FIG. 6).
- the GUI 800 comprises a text input field 801.
- the user can enter textual data into the text input field 801.
- the user can query the avatar with specific questions.
- the query may be "What do you like to do?"
- the AI logic 300 (FIG. 3) interacts with the avatars corresponding avatar data 106, e.g., the avatar A data 203 (FIG. 3), and returns a related response.
- the response returned consists of text, audio, and animation data 305 (FIG. 3).
- the text portion of the data 305 may be displayed in the transcript text field 802.
- the transcript text field may say, "I like to swim and play golf.”
- the user interface logic 104 uses the audio data and the animation data 305 to illicit movement in the avatar's gestures, e.g., the avatar moves its mouth, and the message is audibly played for the user.
- the GUI 800 comprises a display box 809.
- the AI logic 300 may also display in the display box 809 information related to the query.
- information related to the query may be photographs, web sites, video, or any other data related to the query.
- the GUI 800 further comprises a push buttons 803, 804, and 805.
- the client computing device 101 when selected, turns on the microphone so that audible sound or speech from the user may be received by the client computing device 101 and transmitted to the server computing device 102.
- Push button 704 is the “Book of Life” push button, and when selected displays
- Push button 805 is the personal diary push button, and when selected displays GUI 1000 of FIG. 10, which is described further herein.
- FIG. 9 depicts a book of life GUI 900 in accordance with an embodiment of the present disclosure.
- the user can continue to communicate with the AI logic 300 (FIG. 3) via the text input box 902 and the transcript box 901.
- a push button 903 provides for the microphone implementation, or a "QUIET! push button 904 can quiet the application.
- the user selects "SUBMIT" push button 905 to submit text within the text input box 902.
- the intelligent avatar creation/editing logic 105 displays a list of hyperlinks 1. - 10.
- the intelligent avatar creation/editing logic 105 may display questions such as "What are you grandparents' names?" or "Name each of your siblings.”
- the intelligent avatar creation/editing logic 105 stores such additional information as owner data. This owner data may be accessed by the AI logic 300 when answering questions by third parties or it may be read into the brain and/or personality data as permanent additions to the brain and personality data.
- FIG. 10 depicts a personal diary GUI 1000 in accordance with an embodiment of the present disclosure.
- the GUI 1000 comprises one or more text fields 1001 in which the user A (FIG. 3) can enter daily occurrences or records special events.
- the user A can select a "SUBMIT" push button 1002, and the intelligent avatar creation/editing logic 105 (FIG. 3) stores the information in owner data of the avatar A data 203.
- the user A can select a "CANCEL" push button 1003 to cancel an entry so that it is not written to owner data.
- User A can enter periodic data indicative of life experiences of user A, which the intelligent avatar creation/editing logic 105 (FIG. 3) stores as owner data in avatar A data 203 (FIG. 3).
- FIG. 11 is a flowchart depicting exemplary architecture and functionality of the intelligent avatar creation/editing logic 105 (FIG. 3).
- the intelligent avatar creation/editing logic 105 inquires of the user A (FIG. 3) whether the user A would like to take a personality test. Such inquiry can be performed, for example, via a web page displayed via the user interface 104 (FIG. 3).
- the intelligent avatar creation/editing logic 105 displays a personality test (not shown) via the user interface 104, as indicated in step 1107.
- the personality test may be a Jung- Myers-Briggs personality type test. If a personality test is taken, user A answers a series of queries related to the personality test and the intelligent avatar creation/editing logic 105 receives data in response to the queries, as indicated in step 1108. Based upon the received data, the intelligent avatar creation/editing logic 105 displays a list of brains obtained from the brain and personality data 202 (FIG. 2) that correspond to the responses to the personality test, as indicated in step 1109.
- the responses to the personality test indicate an extroverted personality
- one or more brains exhibiting extroverted characteristics is displayed.
- User A selects from the list which brain user A desires to be used in the avatar being created, and the intelligent avatar creation/editing logic 105 receives data indicative of the user's brain/personality selection, as indicated in steep 1102.
- step 1100 user A does not desire to take a personality test, then the intelligent avatar creation/editing logic 105 displays a list a available brains with associated personality characteristics, obtained from the brain and personality data 202, via the user interface 104 to user A. User A then selects from the list of brains/personalities which brain user A desires to be used in the avatar being created, and the intelligent avatar creation/editing logic 105 receives data indicative of the user's brain/personality selection, as indicated in step 1102.
- the intelligent avatar creation/editing logic 105 displays a list of voice fonts in step 1103 from which user A can select a voice font for use in the avatar being created.
- the system 100 may comprise voice synthesis logic (not shown) that allows a user to generate a voice font based upon an audible signal from user A. The user selects a particular voice font from the displayed list of voice fonts, and the intelligent avatar creation/editing logic 105 receives data indicative of the voice font selected in step 1104.
- voice synthesis logic may also be used.
- the voice synthesis logic could digitally capture and synthesize user A' s voice via, for example, a microphone, to create a customized voice font. This customized voice font could then be used for the avatar.
- the intelligent avatar creation/editing logic 105 can query user A for a photograph.
- the intelligent avatar creation/editing logic 105 receives a photograph in step 1105.
- the photograph can be one of a person or a caricature.
- animation creation logic 200 (FIG. 2) generates data and logic for animating the received data indicative of the photograph, as indicated in step 1106.
- the intelligent avatar creation/editing logic 105 obtains brain and personality data, the voice font, and the animation data corresponding to the generated avatar.
- the intelligent avatar creation/editing logic 105 then stores the brain and personality data, the voice font, and the animation data for the created avatar in memory, e.g., avatar A data 203 (FIG. 2).
- FIG. 12 is a flowchart depicting exemplary architecture and functionality of intelligent avatar runtime logic 107 (FIG. 3). Note that the flowchart described is described with reference to the thin client configuration. However, the functionality for the thick client configuration is substantially similar.
- the intelligent avatar runtime logic 107 displays a list of already existing avatars to a user, e.g., user A, via the user interface 104 (FIG. 3).
- a user e.g., user A
- this can be, for example, a web page having a plurality of hyperlinks wherein each hyperlink is associated with a particular avatar.
- the intelligent avatar runtime logic 107 receives an avatar selection, as indicated in step 1200.
- the intelligent avatar runtime logic 107 displays the graphical user interface similar to GUI 800 depicted in FIG. 8.
- the avatar 806 (FIG. 8) is displayed to the screen, and a text box 801
- FIG. 8 is provided in which one can type questions to interact with the avatar.
- User A enters data indicative of a question, which can be entered in the text box 801 or spoken into a microphone, and the intelligent avatar runtime logic 107 receives data indicative of an interaction input from User A.
- the intelligent avatar runtime logic 107 retrieves information from the avatar data 106 (FIG. 3) related to the text data. Note that the intelligent avatar runtime logic 107 parses the interaction input text data for keywords. The intelligent avatar runtime logic 107 then contextually searches the brain and personality data of the avatar A data 203 (FIG. 3) in search of text that may be conceptually bound to the keywords being searched.
- the intelligent avatar runtime logic 107 converts the audio data to text data in step 1208.
- the intelligent avatar runtime logic 107 retrieves information from the avatar data 106 (FIG. 3) related to the text data. As described hereinabove, the intelligent avatar runtime logic 107 parses the interaction input text data for keywords. The intelligent avatar runtime logic 107 then contextually searches the brain and personality data of the avatar A data 203 (FIG. 3) in search of text that may be conceptually bound to the keywords being searched.
- the intelligent avatar runtime logic 107 Based upon the interaction input received and text found from the search of the avatar A data 203, the intelligent avatar runtime logic 107 generates a text response.
- the text response generated can be in the form of an answer to the question received as the interaction input from the user in step 1201.
- intelligent avatar runtime logic 107 converts the text response to audio data. This conversion can be performed with any type of software and/or hardware for converting text to audio known in the art or future-developed.
- intelligent avatar runtime logic 107 generates animation data that when applied to the avatar 806 makes the avatar 806 appear to be moving his/her lips and/or moving other parts of his/her body.
- step 1207 intelligent avatar runtime logic 107 displays the text response to the transcript text field 802 (FIG. 8).
- the intelligent avatar runtime logic 107 plays the audio and the animation so that the avatar 806 appears to be answering user A' s previously asked interaction input question.
- FIG. 13 depicts another system 1300 in accordance with another exemplary embodiment of the present disclosure.
- the system 1300 comprises processor 1302, display device 1304, input device 1305, and network device 1306.
- processor 1302 display device 1304, input device 1305, and network device 1306.
- display device 1304 input device 1305, and network device 1306.
- network device 1306 Each of these components communicates over local interface 602, which can include one or more buses.
- the exemplary system 1300 further comprises the user interface logic 1306, intelligent avatar runtime logic 107, and intelligent avatar data 106 having at least one set of data indicative of an avatar, e.g., avatar A data 203.
- the system 1300 is a standalone system and operates independent of a network connection.
- the user interface logic 1306 and the intelligent avatar runtime logic 107 can be software, hardware, or a combination thereof.
- user interface logic 1306 and intelligent avatar runtime logic 107 is shown as software stored in memory 1301.
- Memory 1301 may be of any type of memory known in the art, including, but not limited to random access memory (RAM), read-only memory (ROM), flash memory, and the like.
- Processor 1302 may be a digital processor or other type of circuitry configured to run the user interface logic 1306 and intelligent avatar runtime logic 107 by processing and executing the instructions of the user interface logic 1306 and the intelligent avatar runtime logic 107.
- the processor 1302 may be a processor having digital signal processor (DSP) architecture.
- DSP digital signal processor
- the processor 1302 communicates to and drives the other elements within the system 1300 via the local interface 1303.
- the network device 1305 receives and/or transmits data over a network (not shown).
- the network device 1305 may be, for example, an Ethernet card (not shown), which connects the system 1300 to the network.
- the network device 1305 is configured to receive data indicative of avatar A data 203 or other intelligent avatar data 106 for implementation on the system 1300.
- the display device 1304 is a device for visually displaying an avatar (not shown) indicative of the text, audio, and animation data 305.
- the display device 1304 displays one or more graphical user interfaces (not shown) for use by the user to interact with the avatar displayed to the display device 1305, which is described further herein.
- the display device 1304 may be, for example, a backlit liquid crystal display (LCD) screen. Other types of display devices may be used in other embodiments of the present disclosure.
- the input device 1305 is for entering data into the system 1300 and/or otherwise responding or querying the avatar displayed to the display device 1304.
- the input device 1305 is a keyboard, and the user A (FIG. 3) uses the keyboard to input data relating to the avatar displayed.
- the input device may be a microphone, and user A can audibly enter data into the system 1300 related to the avatar.
- user interface logic 1306 displays an GUI, similar to the
- a user (not shown) of the system 1300 may enter data into the text box 801 (FIG. 8) indicative of a question for the avatar 806 (FIG. 8). Note that the user may also use a microphone to audibly communicate with the avatar 806.
- the user interface logic 1306 transmits the text/audio data to the intelligent avatar runtime logic 107.
- the intelligent avatar runtime logic 107 processes the data received substantially similar to the way the data is processed in the thick client system 100 with reference to FIG. 3.
- the user interface logic 1306 then displays text, plays audio, and plays animation data 305 via the GUI 800.
- response text is displayed to the text box 802 (FIG. 8), audios indicative of the response if audibly played for the user, and the avatar 806 moves in accordance with the animation data received from the intelligent avatar runtime logic 107.
- intelligent avatar data 106 is legacy data related to user A.
- the intelligent avatar data 106 comprises data indicative of narrative stories about particular memories or experiences.
- elements of the stories or experiences may be varied in real-time so that the same narrative of a memory can be expressed in limitless combinations.
- Video and/or photographs associated with the story may also be stored as intelligent avatar data 106 and retrieved while the story is being relayed to a user of the systems of the present disclosure.
- the technology described in the present disclosure may be implemented in a dating application.
- a user could provide input and ask questions of an avatar 806 (FIG. 8) to learn more about the user that created the avatar 806. For example, a user could ask about his/her likes/dislikes, their favorite foods, their favorite activities, or the like.
- an owner of an avatar 806 is a subject matter expert on a particular subject matter.
- an owner of an avatar 806 may be a doctor or a lawyer that has specialized knowledge in a particular field.
- the owner can modify the brain and personality data to include information related to questions other users may have about their particular field, including the answers to such questions.
- the avatar 806 may be created indicative of a famous historical figure, e.g., George Washington.
- the brain and personality data associated with the avatar can comprise information related to facts or particular stories related to the historical figures.
- a user could query the historical figures and the avatar would answer the queries based upon the information in the brain and personality data.
- the avatar 806 may be created indicative of a customer service representative.
- the brain and personality data associated with the avatar can comprise information related to facts corresponding to particular products or services. For example, user A may ask the avatar 806 how to install a particular part on a washing machine, and the avatar 806, would give user A detailed directions on how to undertake such a task.
- the avatar 806 may be created indicative of a virtual concierge. In such an embodiment, the brain and personality data associated with the avatar 806 may be indicative of local events, restaurants, hotels, etc.
Landscapes
- Engineering & Computer Science (AREA)
- Physics & Mathematics (AREA)
- Theoretical Computer Science (AREA)
- Mathematical Physics (AREA)
- Computational Linguistics (AREA)
- Data Mining & Analysis (AREA)
- Databases & Information Systems (AREA)
- General Engineering & Computer Science (AREA)
- General Physics & Mathematics (AREA)
- Artificial Intelligence (AREA)
- Human Computer Interaction (AREA)
- User Interface Of Digital Computer (AREA)
Abstract
An system of the present disclosure has memory and logic that receives creation input from a user. The creation input comprises avatar data defining a computer representation, and the avatar data comprises brain data defining characteristics of the computer representation. The avatar data is stored in the memory. The logic further displays the computer representation to a user, receives a query input from the user related to the computer representation, performs a contextual search of the avatar data and dynamically generates a response to the user based upon the brain data based upon the contextual search of the avatar data.
Description
SYSTEM AND METHOD FOR CREATING, EDITING, AND ACCESSING AN INTELLIGENT AVATAR
CROSS REFERENCE TO RELATED APPLICATION
[0001] This application claims priority to U.S. Provisional Patent Application Serial
Number 61/219,961 entitled "System and Method for Creating, Editing, and Accessing an Intelligent Avatar," and filed on June 24, 2009, which is incorporated herein by reference. This application is related to U.S. Patent Application Serial No. 12/823,018 filed on June 24, 2010, which is also incorporated herein by reference.
BACKGROUND
[0002] The term "avatar" refers to a computer representation of person or thing in the form of a visual model, which can be a two-dimensional or three-dimensional model. Oftentimes, avatars are used in computer games or on social-networking Internet forums.
[0003] Some avatars have a limited ability to move and speak. In this regard, some avatars can answer a question that is asked by a user of a computer. Further, such an avatar may make slight movements during communication with the computer user.
DESCRIPTION OF THE DRAWINGS
[0004] FIG. 1 depicts an intelligent avatar creating, editing, and runtime system 100 in accordance with an embodiment of the present disclosure. The system 100 comprises a client computing device 101 and a server computing device 102. The client computing device 101 and the server computing device 102 communicate over a network 103. Note that the network 103 may be any type of network known in the art or future-developed. For example, the server 101 may be connected to the computing devices 102 via the Internet, a local area network (LAN), or a wide area network (WAN).
[0005] The client computing device 101 comprises user interface logic 104. In one embodiment, the client computing device 101 is a personal computer (PC). In other embodiment, the client computing device 101 may be, for example, a mobile device or a personal digital assistant (PDA).
[0006] In addition, the user interface logic 104 may be, for example, a web browser and/or a Flash Player, as in a thin client configuration, which is described further with reference to FIG. 3. The system 100 may also be implemented via a thick client configuration, which is described further with reference to FIG. 4.
[0007] The present disclosure contemplates any type of web browser (not shown) that allows a user A to display and interact with information located on a web page (not shown), which may be accessible via the network 103. In addition, the present disclosure contemplates any type of flash player that allows the user A to view animations and movies in the web browser.
[0008] The server 102 comprises intelligent avatar creation/editing logic 105, intelligent avatar data 106, and intelligent avatar runtime logic 107. The intelligent avatar creation logic 105 creates intelligent avatar data 106 indicative of an intelligent avatar (not shown) based upon input from the user A received via the user interface logic 104, which is described in more detail herein. Note that the term "avatar" refers to a computer representation (not shown) of a human being that can be personalized based upon desired characteristics of a user A of the client computing device 101 of the system 100. As examples, the intelligent avatar data 106 may be data indicative of a two-dimensional or three-dimensional graphical representation of the avatar, a voice font, a brain type, a personality type, or the like. Such is described in more detail hereafter.
[0009] Once the intelligent avatar data 106 is created, the intelligent avatar runtime logic 107 creates, upon request, an avatar (not shown) based on the intelligent avatar data 106. The avatar created exhibits those desired characteristics identified by user A in creating the avatar. The intelligent avatar runtime logic 107 enables communication with the avatar over the network 103 and can allow the intelligent avatar to learn new and different information based upon data received during such communication.
[0010] In this regard, the avatar comprises a text component, a visual component and an audio component. The intelligent avatar runtime logic 107 displays the generated avatar, through the user interface logic 104 to user A. User A can type in questions for the avatar, and the intelligent avatar runtime logic 107 transmits animation and voice data to the user interface logic 104, such that the avatar appears to answer questions asked by the user A. This is described further herein.
[0011] Note that in such an embodiment, the user types in the questions for the avatar. However, in other embodiments, the user may speak into a microphone, for example, the questions the user has for the avatar.
[0012] The intelligent avatar creating, editing, and runtime system 100 in accordance with another embodiment of the present disclosure further comprises a third-party computing device 108. The server computing device 102 communicates with the third-party computing device 108 over the network 103. The third-party computing device 108 may be, for example, a Web server for transmitting content upon request to other users on the network 103. For example, the third-party computing device 108 may host the Wikipedia® Internet Web site, and the third-party computing device 108 may transmit content related to a particular subject to a user on the network 103 upon request.
[0013] The third-party computing device 108 comprises third-party data 109. The third-party data 109 can be any data known in the art or future-developed. In the example of the Wikipedia® Internet Web site, the third-party data 109 comprises data indicative of an identifier of a subject and data describing the subject.
[0014] During operation of the server computing device 102, User A may transmit data textually or verbally to the intelligent avatar runtime logic 107 indicative of a question related to a particular topic. The intelligent avatar runtime logic 107 may search the intelligent avatar data 106 and not find any information relating to the particular topic. In such a situation, the intelligent avatar runtime logic 107 automatically transmits a request to the third-party computing device 108 requesting information related to the particular subject, and the third-party computing device 108 searches the third-party data 109, retrieves any information relating to the subject, and transmits the data retrieved back to the server computing device 102.
[0015] Upon receipt, the intelligent avatar runtime logic 107 may communicate the received data to the client computing device 101 so that User A can see information relating to the subject about which User A asked a question. In addition, the intelligent avatar runtime logic 107 may process the received data through the intelligent avatar creation/editing logic 105 such that the received data is then incorporated into the intelligent avatar data 106 for future retrieval.
[0016] FIG. 2 depicts the server 102 and further illustrates the intelligent avatar creation/editing logic 105 in accordance with an embodiment of the present
disclosure. The server 102 further comprises animation creation logic 200, voice font data 201, and brain and personality data 202.
[0017] In one embodiment, the brain and personality data 202 comprises data indicative of particular brain and/or personality types. The intelligent avatar creation logic 105 may display data indicative of the brain and/or personality types to the user A (FIG. 1), and the user A may select which brain and/or personality type that the user A desires to be incorporated into the avatar being created. Exemplary types of brains may be, for example, an extrovert brain type or an introvert brain type.
[0018] In such an embodiment, the brain and personality data 202 comprises a logical set of rules that define behaviors of the avatar being created. For example, a rule may equate, textually, the word "likes" with a list of words indicating what the avatar (and hence the user A) enjoys, e.g., hiking, reading, swimming, golfing, etc. There may be other key words associated with additional information. For example, the word "born" may be textually associated with data identifying the place or date of birth, e.g., "born = Columbus, Indiana, March, 22, 1968.
[0019] In one embodiment, the voice font data 201 comprises data indicative of a plurality of particular voice fonts. The intelligent avatar creation/editing logic 105 may display data indicative of the voice fonts to user A (FIG. 1), and user A may select which voice font that user A desires to be incorporated into the avatar being created. In such an embodiment, the intelligent avatar creation/editing logic 105 may allow user A to preselect and listen to each of the different voice fonts prior to the user A selecting one of the voice fonts for incorporation into the avatar.
[0020] In another embodiment, the server computing device 102 may comprise some type of voice synthesis logic (not shown) that generates digital representations of the voice of user A. The digital representation (not shown) may be generated, for example, from audio of the voice of user A and uploaded as particular voice font data 201. The voice synthesis logic may be any type of voice synthesis software known in the art or future-developed for generating a digital representation of the voice of User A. For example, the voice synthesis logic 306 may be Nuance's RealSpeak, DragonSpeak, or Cepstral's VoiceForge.
[0021] In one embodiment, the animation creation logic 200 receives as input a photograph of a person, and generates animation data based upon the photograph submitted. As an example, the animation data may be data indicative of the person moving his/her lips and/or moving his/her head as the person moves his/her lips.
Thus, the intelligent avatar creation/editing logic 105 may request a photograph from user A, which is used by the animation creation logic 200 to create the animation data 205.
[0022] The intelligent avatar creation/editing logic 105 stores the data selected by user A during the avatar creation process as avatar A data 203 in the intelligent avatar data 106. Note that the intelligent avatar data 106 is a compilation of data describing a plurality of avatars, e.g., avatar A data 203 and avatar B data 204. Such avatar A data 203 relates the selected brain and personality data, the voice font, and the animation data selected by user A.
[0023] The avatar A data 203 further comprises conversation data and owner data.
Conversation data is data indicative of conversations that third parties have with the created avatar, which is described further herein with reference to FIG. 3. The owner data can be data indicative of any additional information that the user desires to become a part of the avatar A data.
[0024] As an example, user A can store as owner data e-mails, blogs, photographs,
MySpace page, narrative or text based accounts data, etc. The owner data can be any type of data known in the art or future-developed that describes or is related to User A. For example, owner data may be still images in any format known in the art or future-developed, including Joint Photographic Experts Group (JPEG), Graphic Interchange Format (GIF), or Portable Network Graphics (PNG). The avatar A data 203 may be video, for example Windows Media Video (WMV) or Moving Pictures Expert Group (MPEG). The avatar A data 203 may be audio, including Windows Media Video (WMV) or Windows Wave (WAV). Further, the owner data may be multimedia, including Adobe Flash Movie (SWF).
[0025] FIG. 3 depicts the system 100 in accordance with an embodiment of the present disclosure having a plurality of users, including user A, user B, and user C. In addition, FIG. 3 depicts a more detailed view of the intelligent avatar runtime logic 107 and the intelligent avatar creation/editing logic 105. As described hereinabove, the intelligent avatar runtime logic 107 is invoked when user A, B, or C desires to interact with an avatar created using the intelligent avatar creation logic 105. In this regard, the user A, B, C may desire to simply carry on a conversation with an avatar.
[0026] Note that the system 100 is a "thick client" configuration. The term "thick client" refers to a client/server system wherein the client of the system provides some of the functionality of the system independent of the server. The thick client
configuration will be contrasted with a "thin client" configuration in the discussion related to FIG. 4. The term "thin client" refers to a system wherein the functionality of the system is resident on the server side of the system.
[0027] The intelligent avatar runtime logic 107 comprises artificial intelligence (AI) logic 300, automated speech recognition logic 301, animation/gesture logic 302, and text to speech logic 303. Each of these logical components is used in interaction with an avatar stored on the server computing device 102. Note that each of the logical components 300-303 are shown for clarity as separate logical components; however, this should in no way limit the intelligent avatar runtime logic 107. Initially, the user A, B, C selects a desired avatar via a user interface (not shown).
[0028] In the exemplary runtime environment depicted in FIG. 3, the user A, B, C has selected the avatar A corresponding with the avatar A data 203. Thus, the AI logic 300 is described hereinafter as interfacing with the avatar A data 203 of the intelligent avatar data 106. However, if the user A, B, C selected a different avatar, the intelligent avatar runtime logic 107 would interface with the data describing the selected other avatar. The avatar A data 203 is shown in the runtime example for exemplary purposes only.
[0029] In one embodiment, the intelligent avatar runtime logic 107 loads into random access memory (RAM) the data indicative of the avatar A data 203. By reading the data into RAM, the process of accessing the avatar A data 203 is more efficient. However, in another embodiment, the intelligent avatar runtime logic 107 accesses data as needed, for example, if the avatar A data is stored in a relational databaseΛ
[0030] During operation, the intelligent avatar runtime logic 107 receives text and/or audio input 304 from the user A, B, C. In this regard, the user A, B, C may ask a question via a microphone or the user A, B, C may type in a question using a keyboard. Regardless of the input method used, data indicative of the question is transmitted via the network 103 to the server computing device 102.
[0031] If the text and/or audio input 304 received comprises audio data, then the received data 304 is transmitted to the automated speech recognition logic 301. The automated speech recognition logic 301 translates the received data into text, and provides textual data indicative of the audio received to the AI logic 300. If the text/ Audio input 304 comprises textual data, the textual data is transmitted to the AI logic 300.
[0032] The AI logic 300 parses the text and identifies key words within the text. The
AI logic 300 may identify additional concepts and/or text related to the key words identified. As an example, the question received from the user A, B, C may be "What do you like?" The AI logic 300 parses the text and identifies the word "like" as a key word. In addition, the AI logic 300 may associate the word "like" with the word "love."
[0033] The AI logic 300 then accesses the brain and personality data of the avatar A data 203. The AI logic 300 may perform a contextual search of the brain and personality data for the words like and/or love. Rules in the brain and personality data may be "like = swimming" and "love = Chihuahuas." The AI intelligence logic 300 then generates a textual response to the question, which can be "I like swimming, but I love Chihuahuas!"
[0034] The textual response generated by the AI logic is transmitted to the text to speech logic 303. The text to speech logic generates an audio signal indicative of the response and based upon the voice font data associated with avatar A data 203.
[0035] In addition, the animation/gesture logic 302 generates data indicative of the movement of the avatar as it recites the response. Such data indicative of movement of the avatar is generated based upon the animation data corresponding to avatar A data 203.
[0036] Thereafter, text/audio/animation output 305 is transmitted to the user A, B, C that submitted the question "What do you like?" The user interface 104 of the client computing device 101 receives the text/audio/animation output 305, and plays for the user A, B, C an animation of the avatar saying "I like swimming, but I love Chihuahuas!"
[0037] In one embodiment, the AI logic 300 stores conversation tracking data 306 indicative of most of the text received during the conversation that user A, B, C has with the avatar indicative of the avatar A data. This conversation data is stored in relation to the avatar A data 203.
[0038] In addition, user A, B, C may desire to edit his/her own avatar. Thus, assuming for illustrative purposes that user A desires to edit his/her avatar A data 203, the intelligent avatar creation/editing logic 105 receives data indicative of additional characteristics related to the avatar A data 203. The intelligent avatar creation/editing logic 105 adds such additional information provided by user A to the owner data in the avatar A data 203.
[0039] During operation, the AI logic 300 may read the owner data into the brain and personality data. In so doing, the AI logic 300 may use the newly added owner data to respond to queries made from other users, e.g., user B, C. In addition, the AI logic 300 may look within the user data to retrieve additional information that may be relevant to queries.
[0040] As an example, the user data may have a photograph of user A's child. User
B may ask of the avatar A, "Do you have any children?" The brain and personality data in the avatar A data 203 may have relational text "children = one, male, Michael," and the owner data may have associated with the one male child a photograph. Therefore, the AI logic 300 may transmit, in response to the query, text, audio, and animation data that includes "Yes, I have one male child named Michael," and the AI logic 300 may display a photograph or video of the child.
[0041] FIG. 4 depicts the system 100 in a thick client configuration. In such a configuration, each of the components behaves substantially similar to the like components of FIG. 3.
[0042] However, the difference in the configuration of system 100 in FIG. 3 and 4 is that some of the functionality related to the system 100 is moved to the client computing device 101. In an exemplary embodiment, the automated speech recognition logic 301, the text to speech logic 303, and the animation /gesture logic 302 are resident on the client computing device 101.
[0043] Thus, text generated by the automated speech recognition logic 301 is provided over network 103 to the server 102. This is shown in FIG. 4 as text input 400. Accordingly, what is transmitted by the intelligent avatar runtime logic 107 is text or symbol output 401. The text to speech logic 303 receives the text and converts it to an audio signal. In addition, the animation/gesture logic 302 receives the symbol data indicative of the animation/gesture. The user interface logic 104 displays the avatar reciting a response using generated gestures and animation.
[0044] FIG. 5 depicts an exemplary client computing device 101 in accordance with an embodiment of the present disclosure. The exemplary client computing device 101 comprises processor 500, display device 504, input device 505, and network device 503. Each of these components communicates over local interface 502, which can include one or more buses.
[0045] The exemplary client computing device 101 further comprises the user interface logic 104, which is described hereinabove. Note that the client 101 shown
in FIG. 5 is the client 101 implemented in a thin client configuration, which was described in detail hereinabove.
[0046] The user interface logic 104 can be software, hardware, or a combination thereof. In the exemplary client computing device 101 depicted in FIG. 5, user interface logic 104 is shown as software stored in memory 501. Memory 501 may be of any type of memory known in the art, including, but not limited to random access memory (RAM), read-only memory (ROM), flash memory, and the like.
[0047] Processor 500 may be a digital processor or other type of circuitry configured to run the user interface logic 104 by processing and executing the instructions of the user interface logic 104. By way of example, the processor 500 may be a processor having digital signal processor (DSP) architecture. The processor 500 communicates to and drives the other elements within the client computing device 101 via the local interface 502.
[0048] The network device 503 receives and/or transmits data over network 103
(FIG. 1). Thus, the network device 503 may be, for example, an Ethernet card (not shown), which connects the client computing device 101 to the network 103.
[0049] The display device 504 is a device for visually displaying an avatar (not shown) indicative of the text, audio, and animation data 305 (FIG. 3) received from the server computing device 102 (FIG. 3). In addition, the display device 504 displays one or more graphical user interfaces (not shown) for use by the user to interact with the avatar displayed to the display device 504, which is described further herein. The display device 504 may be, for example, a backlit liquid crystal display (LCD) screen. Other types of display devices may be used in other embodiments of the present disclosure.
[0050] The input device 505 enables is for entering data into the client computing device 101. In one embodiment, the input device 505 is a keyboard, and the user A (FIG. 3) uses the keyboard to input data relating to the avatar displayed. In addition, the input device may be a microphone, and user A can audibly enter data into the client computing device 102 related to the avatar.
[0051] FIG. 6 depicts an exemplary server computing device 102 in accordance with an embodiment of the present disclosure. The exemplary server computing device 102 comprises processor 600 and network device 603. Each of these components communicates over local interface 602, which can include one or more buses.
[0052] The exemplary server computing device 102 further comprises the intelligent avatar creation/editing logic 105 and the intelligent avatar runtime logic 107, as described hereinabove. The intelligent avatar creation/editing logic 105 and the intelligent avatar runtime logic 107 can be software, hardware, or a combination thereof. In the exemplary server computing device 102 depicted in FIG. 2, intelligent avatar creation logic 105 and the intelligent avatar runtime logic 107 are shown as software stored in memory 601. Memory 601 may be of any type of memory known in the art, including, but not limited to random access memory (RAM), read-only memory (ROM), flash memory, and the like.
[0053] When stored in memory 601, intelligent avatar creation/editing logic 105 and the intelligent avatar runtime logic 107 can be stored and transported on any computer-readable medium for use by or in connection with an instruction execution system, apparatus, or device, such as a computer-based system, processor-containing system, or other system that can fetch the instructions from the instruction execution system, apparatus, or device and execute the instructions.
[0054] In the context of the present disclosure, a "computer-readable medium" can be any means that can contain, store, communicate, propagate, or transport the program for use by or in connection with the instruction execution system, apparatus, or device. The computer readable medium can be, for example but not limited to, an electronic, magnetic, optical, electromagnetic, infrared, or semiconductor system, apparatus, device, or propagation medium
[0055] Processor 600 may be a digital processor or other type of circuitry configured to run the intelligent avatar creation/editing logic 105 and the intelligent avatar runtime logic 107 by processing and executing the instructions of the intelligent avatar creation/editing logic 105 and the intelligent avatar runtime logic 107. By way of example, the processor 600 may be a processor having digital signal processor (DSP) architecture. The processor 600 communicates to and drives the other elements within the server computing device 102 via the local interface 602.
[0056] The network device 603 receives and/or transmits data over network 103
(FIG. 1). Thus, the network device 603 may be, for example, an Ethernet card (not shown), which connects the server computing device 102 to the network 103.
[0057] During operation, the intelligent avatar creation/editing logic 105 may display one or more graphical user interfaces (GUIs) (not shown) that requests particular information. Such information, as described hereinabove, is used to generate the
intelligent avatar data 106, which is stored in memory 601. An avatar embodying the intelligent avatar data 106 may then be accessed by a user A, B, C (FIG. 4) via the intelligent avatar runtime logic 107 or the intelligent avatar creation/editing logic 105.
[0058] In one embodiment, the intelligent avatar creation/editing logic 105 may display a basic personality test via a GUI (not shown) to user A (FIG. 4). The basic personality test would comprise, for example, questions indicative of a Jung-Myers- Briggs personality typing test. The intelligent avatar creation/editing logic 105 stores data indicative of the answers input by user A. Based upon the information provided by user A, the intelligent avatar creation/editing logic 105 may provide a list of brain and personality types from the brain and personality types data 202 (FIG. 2) from which the user could select a particular brain and personality type that most closely resembles the type of personality of the user based upon the test results.
[0059] In addition, the intelligent avatar creation/editing logic 105 may provide additional questions to user A via the display device 506 (FIG. 5) in order to obtain additional information about User A, which can be used to create the intelligent avatar data 106. In this regard, the intelligent avatar creation/editing logic 105 could obtain addition information related to topics such as birth line and genealogy, birth and childhood, the first day of school, the first date... etc. Such information would be obtained from the user via a set of questions targeted at the particular topic and data indicative of the answers stored as intelligent avatar data 106.
[0060] FIG. 7 depicts an exemplary client computing device 402 in accordance with an embodiment of the present disclosure. The exemplary client computing device 402 comprises processor 700, display device 704, input device 705, and network device 703. Each of these components communicates over local interface 702, which can include one or more buses.
[0061] The exemplary client computing device 402 further comprises the user interface logic 104, text to speech logic 303, automated speech recognition logic 301, and animation/gesture logic 302 which are described hereinabove. Note that the client 402 shown in FIG. 7 is the client 402 implemented in a thick client configuration, which was described in detail hereinabove with reference to FIG. 4.
[0062] The user interface logic 104, text to speech logic 303, automated speech recognition logic 301, and animation/gesture logic 302 can be software, hardware, or a combination thereof. In the exemplary client computing device 402 depicted in FIG. 7, user interface logic 104, text to speech logic 303, automated speech
recognition logic 301, and animation/gesture logic 302 are shown as software stored in memory 701. Memory 701 may be of any type of memory known in the art, including, but not limited to random access memory (RAM), read-only memory (ROM), flash memory, and the like.
[0063] Processor 700 may be a digital processor or other type of circuitry configured to run the user interface logic 104, text to speech logic 303, automated speech recognition logic 301, and animation/gesture logic 302 by processing and executing the instructions of the user interface logic 104, text to speech logic 303, automated speech recognition logic 301, and animation/gesture logic 302. By way of example, the processor 700 may be a processor having digital signal processor (DSP) architecture. The processor 700 communicates to and drives the other elements within the client computing device 402 via the local interface 702.
[0064] The network device 703 receives and/or transmits data over network 103
(FIG. 4). Thus, the network device 703 may be, for example, an Ethernet card (not shown), which connects the client computing device 402 to the network 103.
[0065] During operation, the text to speech logic 303 and the animation/gesture logic
302 receives text/symbol data 401 from the server computing device 102. The user interface logic 104 then displays to the display device 704 an avatar (not shown) indicative of the text/symbol data 401 (FIG. 4) received from the server computing device 102 (FIG. 4). In addition, the display device 704 displays one or more graphical user interfaces (not shown) for use by the user to interact with the avatar displayed to the display device 604, which is described further herein. The display device 704 may be, for example, a backlit liquid crystal display (LCD) screen. Other types of display devices may be used in other embodiments of the present disclosure.
[0066] The input device 705 enables is for entering data into the client computing device 402. In one embodiment, the input device 705 is a keyboard, and the user A (FIG. 4) uses the keyboard to input data relating to the avatar displayed. In addition, the input device may be a microphone, and user A can audibly enter data into the client computing device 402 related to the avatar.
[0067] FIG. 8 depicts a main graphical user interface (GUI) 800 in accordance with an embodiment of the present disclosure that is displayed by the intelligent avatar creation/editing logic 105 (FIG. 3). In such an embodiment, an avatar 806 is being displayed in the GUI 800. Therefore, it is assumed for purposes of the following discussion that the avatar 806 has already been created based upon particularly
selected brain and personality data 202 (FIG. 2), voice font data 201 (FIG. 2), and animation data 205 (FIG. X). In such an embodiment, the intelligent avatar creation logic 105 (FIG. 3) is displaying the GUI 800 and the avatar 906 to the display device 604 (FIG. 6) of a client computing device 101 (FIG. 6).
[0068] The GUI 800 comprises a text input field 801. The user can enter textual data into the text input field 801. In this regard, the user can query the avatar with specific questions. For example, the query may be "What do you like to do?" In such a query, the AI logic 300 (FIG. 3) interacts with the avatars corresponding avatar data 106, e.g., the avatar A data 203 (FIG. 3), and returns a related response.
[0069] The response returned consists of text, audio, and animation data 305 (FIG. 3).
The text portion of the data 305 may be displayed in the transcript text field 802. For example, the transcript text field may say, "I like to swim and play golf." In addition to the text being displayed, the user interface logic 104 (FIG. 3) uses the audio data and the animation data 305 to illicit movement in the avatar's gestures, e.g., the avatar moves its mouth, and the message is audibly played for the user.
[0070] In addition, the GUI 800 comprises a display box 809. In response to the query, the AI logic 300 may also display in the display box 809 information related to the query. Such information may be photographs, web sites, video, or any other data related to the query.
[0071] The GUI 800 further comprises a push buttons 803, 804, and 805. Push button
803, when selected, turns on the microphone so that audible sound or speech from the user may be received by the client computing device 101 and transmitted to the server computing device 102.
[0072] Push button 704 is the "Book of Life" push button, and when selected displays
GUI 900 of FIG. 9, which is described further herein. Push button 805 is the personal diary push button, and when selected displays GUI 1000 of FIG. 10, which is described further herein.
[0073] FIG. 9 depicts a book of life GUI 900 in accordance with an embodiment of the present disclosure. When the user (not shown) is in the book of life GUI 900, the user can continue to communicate with the AI logic 300 (FIG. 3) via the text input box 902 and the transcript box 901. A push button 903 provides for the microphone implementation, or a "QUIET!" push button 904 can quiet the application. The user selects "SUBMIT" push button 905 to submit text within the text input box 902.
[0074] In the book of life GUI 900, the user can enter additional information that becomes associated with the avatar displayed. In one embodiment, the intelligent avatar creation/editing logic 105 displays a list of hyperlinks 1. - 10. that when selected displays a list of additional questions that the user can answer in order to associate additional information with the avatar. As an example, "1." In GUI 900 is "My Family History." When selected, the intelligent avatar creation/editing logic 105 may display questions such as "What are you grandparents' names?" or "Name each of your siblings." The intelligent avatar creation/editing logic 105 stores such additional information as owner data. This owner data may be accessed by the AI logic 300 when answering questions by third parties or it may be read into the brain and/or personality data as permanent additions to the brain and personality data.
[0075] FIG. 10 depicts a personal diary GUI 1000 in accordance with an embodiment of the present disclosure. The GUI 1000 comprises one or more text fields 1001 in which the user A (FIG. 3) can enter daily occurrences or records special events. The user A can select a "SUBMIT" push button 1002, and the intelligent avatar creation/editing logic 105 (FIG. 3) stores the information in owner data of the avatar A data 203. In addition, the user A can select a "CANCEL" push button 1003 to cancel an entry so that it is not written to owner data. In such a GUI 1000, User A can enter periodic data indicative of life experiences of user A, which the intelligent avatar creation/editing logic 105 (FIG. 3) stores as owner data in avatar A data 203 (FIG. 3).
[0076] FIG. 11 is a flowchart depicting exemplary architecture and functionality of the intelligent avatar creation/editing logic 105 (FIG. 3). In step 1100, the intelligent avatar creation/editing logic 105 inquires of the user A (FIG. 3) whether the user A would like to take a personality test. Such inquiry can be performed, for example, via a web page displayed via the user interface 104 (FIG. 3).
[0077] If user A desires to take a personality test, the intelligent avatar creation/editing logic 105 displays a personality test (not shown) via the user interface 104, as indicated in step 1107. As noted herein, the personality test may be a Jung- Myers-Briggs personality type test. If a personality test is taken, user A answers a series of queries related to the personality test and the intelligent avatar creation/editing logic 105 receives data in response to the queries, as indicated in step 1108. Based upon the received data, the intelligent avatar creation/editing logic 105 displays a list of brains obtained from the brain and personality data 202 (FIG. 2) that
correspond to the responses to the personality test, as indicated in step 1109. For example, if the responses to the personality test indicate an extroverted personality, then one or more brains exhibiting extroverted characteristics is displayed. User A then selects from the list which brain user A desires to be used in the avatar being created, and the intelligent avatar creation/editing logic 105 receives data indicative of the user's brain/personality selection, as indicated in steep 1102.
[0078] If in step 1100 user A does not desire to take a personality test, then the intelligent avatar creation/editing logic 105 displays a list a available brains with associated personality characteristics, obtained from the brain and personality data 202, via the user interface 104 to user A. User A then selects from the list of brains/personalities which brain user A desires to be used in the avatar being created, and the intelligent avatar creation/editing logic 105 receives data indicative of the user's brain/personality selection, as indicated in step 1102.
[0079] In addition, user A can select the voice font the user desires to be used in the avatar being created. In this regard, the intelligent avatar creation/editing logic 105 displays a list of voice fonts in step 1103 from which user A can select a voice font for use in the avatar being created. Note that in one embodiment, the system 100 (FIG. 3) may comprise voice synthesis logic (not shown) that allows a user to generate a voice font based upon an audible signal from user A. The user selects a particular voice font from the displayed list of voice fonts, and the intelligent avatar creation/editing logic 105 receives data indicative of the voice font selected in step 1104.
[0080] Note that in one embodiment, voice synthesis logic (not shown) may also be used. In this regard, the voice synthesis logic could digitally capture and synthesize user A' s voice via, for example, a microphone, to create a customized voice font. This customized voice font could then be used for the avatar.
[0081] Furthermore, the intelligent avatar creation/editing logic 105 can query user A for a photograph. The intelligent avatar creation/editing logic 105 receives a photograph in step 1105. Note that the photograph can be one of a person or a caricature. In step 1106, animation creation logic 200 (FIG. 2) generates data and logic for animating the received data indicative of the photograph, as indicated in step 1106.
[0082] Thus, corresponding to an avatar, the intelligent avatar creation/editing logic
105 obtains brain and personality data, the voice font, and the animation data
corresponding to the generated avatar. The intelligent avatar creation/editing logic 105 then stores the brain and personality data, the voice font, and the animation data for the created avatar in memory, e.g., avatar A data 203 (FIG. 2).
[0083] FIG. 12 is a flowchart depicting exemplary architecture and functionality of intelligent avatar runtime logic 107 (FIG. 3). Note that the flowchart described is described with reference to the thin client configuration. However, the functionality for the thick client configuration is substantially similar.
[0084] Initially, the intelligent avatar runtime logic 107 displays a list of already existing avatars to a user, e.g., user A, via the user interface 104 (FIG. 3). Note that this can be, for example, a web page having a plurality of hyperlinks wherein each hyperlink is associated with a particular avatar.
[0085] User A selects an avatar from the list and the intelligent avatar runtime logic
107 receives an avatar selection, as indicated in step 1200. When an avatar is selected the intelligent avatar runtime logic 107 displays the graphical user interface similar to GUI 800 depicted in FIG. 8.
[0086] Notably, the avatar 806 (FIG. 8) is displayed to the screen, and a text box 801
(FIG. 8) is provided in which one can type questions to interact with the avatar. User A enters data indicative of a question, which can be entered in the text box 801 or spoken into a microphone, and the intelligent avatar runtime logic 107 receives data indicative of an interaction input from User A.
[0087] If the interaction input received from user A in step 1201 is text data, the intelligent avatar runtime logic 107 retrieves information from the avatar data 106 (FIG. 3) related to the text data. Note that the intelligent avatar runtime logic 107 parses the interaction input text data for keywords. The intelligent avatar runtime logic 107 then contextually searches the brain and personality data of the avatar A data 203 (FIG. 3) in search of text that may be conceptually bound to the keywords being searched.
[0088] If the interaction input received from user A in step 1201 is audio data, the intelligent avatar runtime logic 107 converts the audio data to text data in step 1208. In step 1203, the intelligent avatar runtime logic 107 retrieves information from the avatar data 106 (FIG. 3) related to the text data. As described hereinabove, the intelligent avatar runtime logic 107 parses the interaction input text data for keywords. The intelligent avatar runtime logic 107 then contextually searches the
brain and personality data of the avatar A data 203 (FIG. 3) in search of text that may be conceptually bound to the keywords being searched.
[0089] Based upon the interaction input received and text found from the search of the avatar A data 203, the intelligent avatar runtime logic 107 generates a text response. The text response generated can be in the form of an answer to the question received as the interaction input from the user in step 1201.
[0090] In step 1204, intelligent avatar runtime logic 107 converts the text response to audio data. This conversion can be performed with any type of software and/or hardware for converting text to audio known in the art or future-developed. In addition, intelligent avatar runtime logic 107 generates animation data that when applied to the avatar 806 makes the avatar 806 appear to be moving his/her lips and/or moving other parts of his/her body.
[0091] In step 1207, intelligent avatar runtime logic 107 displays the text response to the transcript text field 802 (FIG. 8). In addition, the intelligent avatar runtime logic 107 plays the audio and the animation so that the avatar 806 appears to be answering user A' s previously asked interaction input question.
[0092] FIG. 13 depicts another system 1300 in accordance with another exemplary embodiment of the present disclosure. In such an embodiment, the system 1300 comprises processor 1302, display device 1304, input device 1305, and network device 1306. Each of these components communicates over local interface 602, which can include one or more buses.
[0093] The exemplary system 1300 further comprises the user interface logic 1306, intelligent avatar runtime logic 107, and intelligent avatar data 106 having at least one set of data indicative of an avatar, e.g., avatar A data 203. In such an embodiment, the system 1300 is a standalone system and operates independent of a network connection.
[0094] The user interface logic 1306 and the intelligent avatar runtime logic 107can be software, hardware, or a combination thereof. In the exemplary system 1300 depicted in FIG. 13, user interface logic 1306 and intelligent avatar runtime logic 107 is shown as software stored in memory 1301. Memory 1301 may be of any type of memory known in the art, including, but not limited to random access memory (RAM), read-only memory (ROM), flash memory, and the like.
[0095] Processor 1302 may be a digital processor or other type of circuitry configured to run the user interface logic 1306 and intelligent avatar runtime logic 107 by
processing and executing the instructions of the user interface logic 1306 and the intelligent avatar runtime logic 107. By way of example, the processor 1302 may be a processor having digital signal processor (DSP) architecture. The processor 1302 communicates to and drives the other elements within the system 1300 via the local interface 1303.
[0096] The network device 1305 receives and/or transmits data over a network (not shown). Thus, the network device 1305 may be, for example, an Ethernet card (not shown), which connects the system 1300 to the network. In such an embodiment, the network device 1305 is configured to receive data indicative of avatar A data 203 or other intelligent avatar data 106 for implementation on the system 1300.
[0097] The display device 1304 is a device for visually displaying an avatar (not shown) indicative of the text, audio, and animation data 305. In addition, the display device 1304 displays one or more graphical user interfaces (not shown) for use by the user to interact with the avatar displayed to the display device 1305, which is described further herein. The display device 1304 may be, for example, a backlit liquid crystal display (LCD) screen. Other types of display devices may be used in other embodiments of the present disclosure.
[0098] The input device 1305 is for entering data into the system 1300 and/or otherwise responding or querying the avatar displayed to the display device 1304. In one embodiment, the input device 1305 is a keyboard, and the user A (FIG. 3) uses the keyboard to input data relating to the avatar displayed. In addition, the input device may be a microphone, and user A can audibly enter data into the system 1300 related to the avatar.
[0099] During operation, user interface logic 1306 displays an GUI, similar to the
GUI 800 depicted in FIG. 8 to display device 1304. A user (not shown) of the system 1300 may enter data into the text box 801 (FIG. 8) indicative of a question for the avatar 806 (FIG. 8). Note that the user may also use a microphone to audibly communicate with the avatar 806.
[00100] The user interface logic 1306 transmits the text/audio data to the intelligent avatar runtime logic 107. The intelligent avatar runtime logic 107processes the data received substantially similar to the way the data is processed in the thick client system 100 with reference to FIG. 3. The user interface logic 1306 then displays text, plays audio, and plays animation data 305 via the GUI 800. In this regard, response text is displayed to the text box 802 (FIG. 8), audios indicative of the response if
audibly played for the user, and the avatar 806 moves in accordance with the animation data received from the intelligent avatar runtime logic 107.
[00101] There are a variety of ways in which the present invention can be used. In one embodiment of the present disclosure, intelligent avatar data 106 (FIG. 3) is legacy data related to user A. In this regard, the intelligent avatar data 106 comprises data indicative of narrative stories about particular memories or experiences. In such an embodiment, elements of the stories or experiences may be varied in real-time so that the same narrative of a memory can be expressed in limitless combinations. Video and/or photographs associated with the story may also be stored as intelligent avatar data 106 and retrieved while the story is being relayed to a user of the systems of the present disclosure.
[00102] In another embodiment, the technology described in the present disclosure may be implemented in a dating application. In such an example, a user could provide input and ask questions of an avatar 806 (FIG. 8) to learn more about the user that created the avatar 806. For example, a user could ask about his/her likes/dislikes, their favorite foods, their favorite activities, or the like.
[00103] Another application may be that an owner of an avatar 806 is a subject matter expert on a particular subject matter. For example, an owner of an avatar 806 may be a doctor or a lawyer that has specialized knowledge in a particular field. In such an example, the owner can modify the brain and personality data to include information related to questions other users may have about their particular field, including the answers to such questions.
[00104] In another embodiment, the avatar 806 may be created indicative of a famous historical figure, e.g., George Washington. The brain and personality data associated with the avatar can comprise information related to facts or particular stories related to the historical figures. Thus, a user could query the historical figures and the avatar would answer the queries based upon the information in the brain and personality data.
[00105] In another embodiment, the avatar 806 may be created indicative of a customer service representative. The brain and personality data associated with the avatar can comprise information related to facts corresponding to particular products or services. For example, user A may ask the avatar 806 how to install a particular part on a washing machine, and the avatar 806, would give user A detailed directions on how to undertake such a task.
In another embodiment, the avatar 806 may be created indicative of a virtual concierge. In such an embodiment, the brain and personality data associated with the avatar 806 may be indicative of local events, restaurants, hotels, etc.
Claims
1. An system, comprising: memory; and logic configured to receive creation input, the creation input comprising avatar data defining a computer representation, the avatar data comprising brain data defining characteristics of the computer representation, the avatar data stored in the memory, the logic further configured to display the computer representation to a user, receive a query input from the user related to the computer representation, perform a contextual search of the avatar data and dynamically generate a response to the user based upon the brain data based upon the contextual search of the avatar data.
2. The system of claim 1, wherein the avatar data comprises voice font data indicative of an audible response to be produced by the computer representation when the logic responds to the user.
3. The system of claim 1, wherein the avatar data comprises animation data indicative of motion to be applied to the computer representation when the logic responds to the user.
4. The system of claim 1, wherein the avatar data comprises conversation data indicative of the input received from the user.
5. The system of claim 1, wherein the logic is further configured to modify the brain data based upon new input.
6. The system of claim 1, wherein the logic displays one or more personality types and generates the brain data based upon a selection of the one or more personality types.
7. The system of claim 1, wherein the brain data comprises a logical set of textual rules.
8. The system of claim 7, wherein the logic is further configured to search the logical set of textual rules based upon the query input.
9. The system of claim 9, wherein the logic is further configured to generate the response based upon results of the search of the logical set of textual rules.
10. The system of claim 1 , wherein the computer representation is indicative of a person.
11. The system of claim 1 , wherein the computer representation is indicative of an animated character.
12. An method, comprising: receiving creation input, the creation input comprising avatar data defining a computer representation, the avatar data comprising brain data defining characteristics of the computer representation; storing the avatar data stored in memory; displaying the computer representation to a user; receiving a query input from the user related to the computer representation; and generating a response to the user based upon the brain data.
13. The method of claim 12, wherein the generating step further comprises the computer representation audibly responding to the query input.
14. The method of claim 12, wherein the generating step further comprises the computer representation moving, based upon animation data indicative of motion to be applied to the computer representation, when responding to query input.
15. The method of claim 12, further comprising modifying the brain data based upon new input.
16. The method of claim 12, further comprising displaying one or more personality types; and generating the brain data based upon a selection of the one or more personality types.
17. The method of claim 12, wherein the brain data comprises a logical set of textual rules further comprising searching the logical set of textual rules based upon the query input.
18. The method of claim 17, wherein the generating step further comprising generating the response based upon the searching step.
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US21996109P | 2009-06-24 | 2009-06-24 | |
US61/219,961 | 2009-06-24 |
Publications (1)
Publication Number | Publication Date |
---|---|
WO2010151700A1 true WO2010151700A1 (en) | 2010-12-29 |
Family
ID=43386897
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
PCT/US2010/039867 WO2010151700A1 (en) | 2009-06-24 | 2010-06-24 | System and method for creating, editing, and accessing an intelligent avatar |
Country Status (1)
Country | Link |
---|---|
WO (1) | WO2010151700A1 (en) |
Cited By (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
FR3074947A1 (en) * | 2017-12-07 | 2019-06-14 | Leaproad, Inc. | SOCIAL PERSONALITY PROFILING FOR PRODUCTIVITY ENHANCEMENT |
US10614081B2 (en) | 2017-12-07 | 2020-04-07 | Leap Road, Inc. | Social personality profiling for productivity enhancement |
Citations (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US6545682B1 (en) * | 2000-05-24 | 2003-04-08 | There, Inc. | Method and apparatus for creating and customizing avatars using genetic paradigm |
US20050125229A1 (en) * | 2003-12-08 | 2005-06-09 | Kurzweil Raymond C. | Use of avatar with event processing |
US20070168863A1 (en) * | 2003-03-03 | 2007-07-19 | Aol Llc | Interacting avatars in an instant messaging communication session |
US20080120558A1 (en) * | 2006-11-16 | 2008-05-22 | Paco Xander Nathan | Systems and methods for managing a persistent virtual avatar with migrational ability |
-
2010
- 2010-06-24 WO PCT/US2010/039867 patent/WO2010151700A1/en active Application Filing
Patent Citations (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US6545682B1 (en) * | 2000-05-24 | 2003-04-08 | There, Inc. | Method and apparatus for creating and customizing avatars using genetic paradigm |
US20070168863A1 (en) * | 2003-03-03 | 2007-07-19 | Aol Llc | Interacting avatars in an instant messaging communication session |
US20050125229A1 (en) * | 2003-12-08 | 2005-06-09 | Kurzweil Raymond C. | Use of avatar with event processing |
US20080120558A1 (en) * | 2006-11-16 | 2008-05-22 | Paco Xander Nathan | Systems and methods for managing a persistent virtual avatar with migrational ability |
Cited By (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
FR3074947A1 (en) * | 2017-12-07 | 2019-06-14 | Leaproad, Inc. | SOCIAL PERSONALITY PROFILING FOR PRODUCTIVITY ENHANCEMENT |
US10614081B2 (en) | 2017-12-07 | 2020-04-07 | Leap Road, Inc. | Social personality profiling for productivity enhancement |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US11861674B1 (en) | Method, one or more computer-readable non-transitory storage media, and a system for generating comprehensive information for products of interest by assistant systems | |
KR102496452B1 (en) | Accelerated task performance | |
US10536402B2 (en) | Context-sensitive generation of conversational responses | |
US20210117214A1 (en) | Generating Proactive Content for Assistant Systems | |
CN110998725B (en) | Generating a response in a dialog | |
CN107491929B (en) | The natural language event detection of data-driven and classification | |
US20200395008A1 (en) | Personality-Based Conversational Agents and Pragmatic Model, and Related Interfaces and Commercial Models | |
CN108352006A (en) | Intelligent automation assistant in instant message environment | |
CN111831798A (en) | Information processing method, information processing device, electronic equipment and computer readable storage medium | |
JP7300074B2 (en) | Facilitated task execution | |
Dasgupta et al. | Voice user interface design | |
US20230164106A1 (en) | Generating Context-Aware Rendering of Media Contents for Assistant Systems | |
Wilks | Artificial companions as a new kind of interface to the future internet | |
Liu et al. | Visual captions: augmenting verbal communication with on-the-fly visuals | |
Foster | State of the art review: Multimodal fission | |
WO2010151700A1 (en) | System and method for creating, editing, and accessing an intelligent avatar | |
Sridharan et al. | Computational models for experiences in the arts, and multimedia | |
CN113168354A (en) | System and method for selecting and providing available actions to a user from one or more computer applications | |
Liu | Knowledge Graph Driven Conversational Virtual Museum Guide | |
Ma | Confucius: An intelligent multimedia storytelling interpretation and presentation system | |
US20240347041A1 (en) | Digital assistant for providing and modifying an output of an electronic document | |
US20240045704A1 (en) | Dynamically Morphing Virtual Assistant Avatars for Assistant Systems | |
EP4343493A1 (en) | Presenting attention states associated with voice commands for assistant systems | |
US12125297B2 (en) | Task execution based on real-world text detection for assistant systems | |
US20240219720A1 (en) | Reading Out Scene Analysis for Users Wearing Head-mounted Devices |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
121 | Ep: the epo has been informed by wipo that ep was designated in this application |
Ref document number: 10792684 Country of ref document: EP Kind code of ref document: A1 |
|
NENP | Non-entry into the national phase |
Ref country code: DE |
|
122 | Ep: pct application non-entry in european phase |
Ref document number: 10792684 Country of ref document: EP Kind code of ref document: A1 |