US20110143632A1 - Figure interactive systems and methods - Google Patents

Figure interactive systems and methods Download PDF

Info

Publication number
US20110143632A1
US20110143632A1 US12782733 US78273310A US2011143632A1 US 20110143632 A1 US20110143632 A1 US 20110143632A1 US 12782733 US12782733 US 12782733 US 78273310 A US78273310 A US 78273310A US 2011143632 A1 US2011143632 A1 US 2011143632A1
Authority
US
Grant status
Application
Patent type
Prior art keywords
interactive
instruction set
data corresponding
scenario data
according
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Abandoned
Application number
US12782733
Inventor
Sheng-Chun Lin
Yu-Chuan Chang
Yu-Shiang Hung
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Institute for Information Industry
Original Assignee
Institute for Information Industry
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date

Links

Images

Classifications

    • AHUMAN NECESSITIES
    • A63SPORTS; GAMES; AMUSEMENTS
    • A63HTOYS, e.g. TOPS, DOLLS, HOOPS, BUILDING BLOCKS
    • A63H3/00Dolls
    • A63H3/28Arrangements of sound-producing means in dolls; Means in dolls for producing sounds
    • AHUMAN NECESSITIES
    • A63SPORTS; GAMES; AMUSEMENTS
    • A63HTOYS, e.g. TOPS, DOLLS, HOOPS, BUILDING BLOCKS
    • A63H2200/00Computerized interactive toys, e.g. dolls

Abstract

Figure interactive systems and methods are provided. The system includes at least a base device. The base device includes a storage unit, a detecting unit, and a processing unit. The storage unit stores a content database recording scenario data corresponding to a plurality of figures. The detecting unit respectively detects identification data of at least a first figure and a second figure. The processing unit respectively retrieves the scenario data of the first figure and the second figure from the content database, dynamically generates an interactive instruction set for the first figure and the second figure, and enables the first figure and the second figure to interact according to the interactive instruction set.

Description

    CROSS REFERENCE TO RELATED APPLICATIONS
  • This application claims priority of Taiwan Patent Application No. 098142210, filed on Dec. 10, 2009, the entirety of which is incorporated by reference herein.
  • BACKGROUND OF THE INVENTION
  • 1. Field of the Invention
  • The disclosure relates generally to figure interactive systems and methods, and more particularly, to systems and methods that detect a plurality of figures and dynamically generate an interactive instruction set for the detected figures.
  • 2. Description of the Related Art
  • Figures or dolls are popular items. In addition to static figures, electronic figures have been developed. Electronic figures can be manipulated by electronic signals to increase applications thereof.
  • As an example, a figure device or a electronic figure, supporting multiple instant communication software, can connect to a personal computer, such that notifications can be performed when messages or new email messages are received, and when the statuses of friends becomes on-line statuses in the instant communication software. For example, required functions of an electronic rabbit figure can be set via a computer, and a server can transmit related data, such as weather forecasts, or head-line news to the electronic rabbit figure, so that data is displayed via the electronic rabbit figure.
  • Generally, conventional electronic figures can only receive fixed messages, and perform related operations according to the received messages. Some electronic figures can perform related operations, such as music playback and dancing based on predefined programs. However, since these programs are fixed and burned into the electronic figures, operating flexibility of the electronic figures is limited, thus hindering popularity among users and development of the electronic figures. Accordingly, with limited variability, users often quickly lose interest in the electronic figures. Currently, there is no technology to automatically detect a plurality of figures and dynamically generate interactive content (not the fixed programs/operations in conventional electronic figures) in the field.
  • BRIEF SUMMARY OF THE INVENTION
  • Figure interactive systems and methods are provided.
  • An embodiment of a figure interactive system includes at least a base device. The base device includes a storage unit, a detecting unit, and a processing unit. The storage unit stores a content database. The detecting unit respectively detects identification data of at least a first figure and a second figure. The processing unit respectively retrieves scenario data corresponding to the first figure and the second figure from the content database, dynamically generates an interactive instruction set for the first figure and the second figure according to the scenario data corresponding to the first figure and the second figure, and enables the first figure and the second figure to interact with each other according to the interactive instruction set.
  • Another embodiment of a figure interactive system includes at least a first figure and a second figure, a base device, and an electronic device. The base device includes at least a detecting unit, and a first communication unit, wherein the detecting unit respectively detects identification data of the first figure and the second figure. The electronic device at least includes a second communication unit which can communicate with the first communication unit via a communication connection, a storage unit storing a content database, and a processing unit which respectively retrieves scenario data corresponding to the first figure and the second figure from the content database. Also, the processing unit dynamically generates an interactive instruction set for the first figure and the second figure according to the scenario data corresponding to the first figure and the second figure, and transmits the interactive instruction set to the base device. The base device enables the first figure and the second figure to interact with each other according to the interactive instruction set.
  • In an embodiment of a figure interactive method, identification data of at least a first figure and a second figure is respectively detected. Then, scenario data corresponding to the first figure and the second figure are respectively retrieved from a content database. The interactive instruction set for the first figure and the second figure is dynamically generated according to the scenario data corresponding to the first figure and the second figure, and the first figure and the second figure are enabled to interact with each other according to the interactive instruction set.
  • In some embodiments, the identification data of the first figure and the second figure can be transmitted to a server via a network. The server can respectively retrieve the scenario data corresponding to the first figure and the second figure according to the identification data of the first figure and the second figure, and transmit the scenario data corresponding to the first figure and the second figure to the base device as the content database via the network. In some embodiments, renewed scenario data corresponding to the first figure and the second figure can be also received via the network, and stored to the content database.
  • In some embodiments, during the interactive instruction set is generated, the first figure can be selected from the start, and first interactive data for a specific topic is retrieved for the first figure from the scenario data corresponding to the first figure. Then, second interactive data corresponding to the first interactive data for the specific topic is retrieved for the second figure from the scenario data corresponding to the second figure. The first interactive data and the second interactive data can be added to the interactive instruction set. In some embodiments, the selected figure, the specific topic, and/or the first interactive data can be randomly selected, or selected according to a specific order.
  • Figure interactive methods may take the form of a program code embodied in a tangible media. When the program code is loaded into and executed by a machine, the machine becomes an apparatus for practicing the disclosed method.
  • BRIEF DESCRIPTION OF THE DRAWINGS
  • The invention will become more fully understood by referring to the following detailed description with reference to the accompanying drawings, wherein:
  • FIG. 1A is a schematic diagram illustrating an embodiment of a figure interactive system of the invention;
  • FIG. 1B is a schematic diagram illustrating another embodiment of a figure interactive system of the invention;
  • FIG. 2 is a schematic diagram illustrating an embodiment of the structure of a server of the invention;
  • FIG. 3 is a schematic diagram illustrating an embodiment of the structure of a base device of the invention;
  • FIG. 4 is a flowchart of an embodiment of a figure interactive method of the invention; and
  • FIG. 5 is a flowchart of an embodiment of a method for generating an interactive instruction set of the invention.
  • DETAILED DESCRIPTION OF THE INVENTION
  • Figure interactive systems and methods are provided.
  • FIG. 1A is a schematic diagram illustrating an embodiment of a figure interactive system of the invention.
  • The structure of the figure interactive system comprises a server 1000 and a base device 2000. The base device 2000 can simultaneously detect identification data of a plurality of figures (such as F1 and F2), and connect to the server 1000 via a network 3000. It is noted that, only two figures are disclosed in this embodiment, however, the invention is not limited thereto.
  • FIG. 2 is a schematic diagram illustrating an embodiment of the structure of a server of the invention.
  • The server 1000 may be a processor-based electronic device, such as general purpose computer, a personal computer, a notebook, or a workstation. The server 1000 at least comprises a scenario database 1100. The scenario database 1100 can comprise scenario data corresponding to a plurality of figures respectively. It is understood that, in some embodiments, the scenario data can comprise dialogues, images, sound effects, music, light signals, and/or actions of the figures, such as swinging, vibrating, rotating, beating and movements, among others.
  • FIG. 3 is a schematic diagram illustrating an embodiment of the structure of a base device of the invention.
  • The base device 2000 can comprise a detecting unit 2100, a storage unit 2200, and a processing unit 2300. It is understood that, in some embodiments, the identification data of the figure can be detected by an RFID (Radio-Frequency Identification), an IR (Infrared) communication recognition system, a USB (Universal Serial Bus) wired/wireless communication recognition system, a 2-dimension/3-dimension barcode recognition system, recognition software and related communication interfaces, or other recognition systems/manners. When several figures are placed on or close to the base device 2000, the detecting unit 2100 can simultaneously detect the identification data of the figures. The storage unit 2200 can at least comprise a content database 2210. The content database 2210 can store the scenario data (such as 2211 and 2212) corresponding to the respective figures. Similarly, in some embodiments, the scenario data can comprise dialogues, images, sound effects, music, light signals, and/or actions. The content database 2210 can further store an interactive instruction set 2220 corresponding to at least two figures. It is noted that, the interactive instruction set 2220 can be dynamically generated according to the scenario data in the content database 2210. The generation and use of the interactive instruction set 2220 are discussed later. The processing unit 2300 performs the figure interactive method of the invention, which will be discussed further in the following paragraphs.
  • FIG. 1B is a schematic diagram illustrating another embodiment of a figure interactive system of the invention. The figure interactive system comprises a server 1000, a base device 2000, and an electronic device 4000. The base device 2000 can simultaneously detect the identification data of several figures, such as F1 and F2, and communicate with the electronic device 4000 via a communication connection. The electronic device 4000 couples to the server 1000 via a network 3000. It is noted that, in this embodiment, the base device 2000 at least comprises the detecting unit 2100 in FIG. 3, and a first communication unit (not shown), and the electronic device 4000 at least comprises the storage unit 2200 and the processing unit 2300 in FIG. 3, and a second communication unit (not shown). The functions and features of the detecting unit 2100, the storage unit 2200, and the processing unit 2300 are similar to that disclosed in FIG. 3, and omitted herefrom. The electronic device 4000 may be a general purpose computer, a personal computer, a notebook, a netbook, a handheld computer, or a PDA (Personal Digital Assistant). The communication connection may be an RS232 connection, a USB communication connection, or an RFID communication connection. The first/second communication unit corresponding to the above communication connections may be an RS232 interface, a USB communication interface, or an RFID communication interface.
  • FIG. 4 is a flowchart of an embodiment of a figure interactive method of the invention. The figure interactive method of the invention can enable multiple figures to interact with each other. It is understood that, in this embodiment, a first figure and a second figure are used for explanation, but the invention is not limited thereto.
  • In step S4100, the identification data of the first figure and the second figure is respectively detected by the base device. Similarly, when several figures are placed on or close to the base device, the detecting unit of the base device can simultaneously detect the identification data of the figures. In step S4200, the scenario data corresponding to the first figure and the second figure are respectively retrieved from a content database according to the identification data of the first figure and the second figure. For example, the scenario data corresponding to the first figure and the second figure can be retrieved from the content database 2210 via the base device 2000 or the electronic device 4000. When the scenario data is retrieved via the electronic device 4000, the electronic device 4000 can transmit the scenario data corresponding to the first figure and the second figure to the base device 2000. In step S4300, the interactive instruction set for the first figure and the second figure is dynamically generated according to the scenario data corresponding to the first figure and the second figure, and in step S4400, the first figure and the second figure are enabled to interact with each other according to the interactive instruction set.
  • In other embodiments, the identification data of the first figure and the second figure can be transmitted to the server via the network from the base device or the electronic device before step S4200. After the identification data of the first figure and the second figure is received, the server can respectively retrieve the corresponding scenario data from the scenario database according to the identification data of the first figure and the second figure, and transmits the scenario data corresponding to the first figure and the second figure to the base device or the electronic device, such that the scenario data corresponding to the first figure and the second figure is stored into the content database.
  • It is understood that, in some embodiments, each figure may have at least a drive component (not shown). The drive component can receive part of the interactive instruction set relating to the figure from the base device, and drive commands according to the actions in the received interactive instruction set, such that the figure and/or at least one component of the figure can be accordingly driven to perform an operation. In some embodiments, the base device or the figure may comprise a display unit (not shown in FIG. 3), for displaying the dialogues (ex. texts), symbols, animations, colors, and/or images in the interactive instruction set. In some embodiments, the base device or the figure may comprise a speaker (not shown in FIG. 3), to play the dialogues by voices, music, and/or sound effects in the interactive instruction set.
  • It is noted that, in some embodiments, when the base device or the electronic device detects and recognize the figures, the base device or the electronic device can immediately transmit the identification data of the figures to the server, and the server searches, receives or generates the scenario data corresponding to the figures, dynamically generates the interactive instruction set according to the scenario data, and transmits the interactive instruction set back to the base device, such that the first figure and the second figure can interact with each other according to the interactive instruction set. In some embodiments, the base device or the electronic device can store the scenario data corresponding to the figures in advance. When the identification data of the figures is detected, the scenario data corresponding to the figures can be directly retrieved from the content database of the storage unit in the base device or the electronic device, and the interactive instruction set can be dynamically generated according to the retrieved scenario data. In some embodiments, the base device or the electronic device can periodically or randomly receive renewed scenario data corresponding to the figures via the network, and store the renewed scenario data to the content database.
  • FIG. 5 is a flowchart of an embodiment of a method for generating an interactive instruction set of the invention.
  • In step S5100, a specific topic is determined. It is understood that, in some embodiments, the specific topic may be a different classification of scenario, such as a narrative, an emotional, or a combat scenario. In some embodiments, the specific topic can be determined randomly. In step S5200, a figure is selected from the figures for interaction. The selected figure may be an initial figure. Similarly, in some embodiments, the initial figure can be randomly selected from the figures. Then, in step S5300, interactive data for the specific topic is retrieved for the initial figure from the scenario data corresponding to the initial figure. Similarly, in some embodiments, the interactive data for the initial figure can be randomly selected from the scenario data corresponding to the initial figure. As described, in some embodiments, the scenario data/interaction data can comprise dialogues, images, sound effects, and/or actions. After the interactive data for the initial figure is determined, in step S5400, interactive data, corresponding to the interactive data for the initial figure, for the specific topic is retrieved for another figure (called associated figure) from the scenario data corresponding to the associated figure. It is understood that, in some embodiments, the scenario data/interaction data can respectively define a tag. Relationships among the scenario data/interaction data can be established via the tags. In step S5500, it is determined whether the generation of the interactive instruction set is complete. It is noted that, the determination of whether the generation of the interactive instruction set is complete may be different based on different requirements and applications. In some embodiments, the scenario data corresponding to the initial figure and/or the associated figure may define a terminal tag. When the interactive data corresponding to the initial figure and/or the associated figure has the terminal tag, the generation of the interactive instruction set is complete. When the generation of the interactive instruction set is not complete (No in step S5500), in step S5600, interactive data, corresponding to the interactive data for the associated figure, for the specific topic is retrieved for the initial figure from the scenario data corresponding to the initial figure, and the procedure returns to step S5400. When the generation of the interactive instruction set is complete (Yes in step S5500), in step S5700, the interactive data corresponding to the initial figure and the associated figure are combined as the interactive instruction set.
  • It is noted that, the embodiment of FIG. 5 is used to generate the interactive instruction set for a specific topic. However, in some embodiments, the specific topic is not needed to be determined in advance, and the interactive instruction set corresponding to the respective figures can be directly determined according to the respective scenario data.
  • Therefore, the figure interactive systems and methods can dynamically generate an interactive instruction set for multiple figures, and enable the figures to interact with each other according to the interactive instruction set. Thus operating flexibility of the figures is increased.
  • Figure interactive methods, or certain aspects or portions thereof, may take the form of a program code (i.e., executable instructions) embodied in tangible media, such as floppy diskettes, CD-ROMS, hard drives, or any other machine-readable storage medium, wherein, when the program code is loaded into and executed by a machine, such as a computer, the machine thereby becomes an apparatus for practicing the methods. When implemented on a general-purpose processor, the program code combines with the processor to provide a unique apparatus that operates analogously to the application of specific logic circuits.
  • While the invention has been described by way of example and in terms of preferred embodiment, it is to be understood that the invention is not limited thereto. Those who are skilled in this technology can still make various alterations and modifications without departing from the scope and spirit of this invention. Therefore, the scope of the present invention shall be defined and protected by the following claims and their equivalents.

Claims (20)

  1. 1. A figure interactive system, comprising:
    at least a first figure and a second figure; and
    a base device, comprising:
    a storage unit storing a content database;
    a detecting unit respectively detecting identification data of the first figure and the second figure; and
    a processing unit respectively retrieving scenario data corresponding to the first figure and the second figure from the content database according to the identification data of the first figure and the second figure, dynamically generating an interactive instruction set for the first figure and the second figure according to the scenario data corresponding to the first figure and the second figure, and enabling the first figure and the second figure to interact with each other according to the interactive instruction set.
  2. 2. The system of claim 1, wherein the processing unit further transmits the identification data of the first figure and the second figure to a server via a network, and the server respectively retrieves the scenario data corresponding to the first figure and the second figure according to the identification data of the first figure and the second figure, and transmits the scenario data corresponding to the first figure and the second figure to the content database of the storage unit via the network.
  3. 3. The system of claim 2, wherein the processing unit further receives renewed scenario data corresponding to the first figure and the second figure from the server via the network, and stores the renewed scenario data to the content database.
  4. 4. The system of claim 1, wherein the interactive instruction set further comprises an action drive command, and the first figure or the second figure respectively has at least a drive component for receiving the interactive instruction set from the base device, and driving the first figure or the second figure to perform an operation according to the action drive command in the interactive instruction set.
  5. 5. The system of claim 1, wherein the base device further comprises a display unit for displaying dialogues or images in the interactive instruction set.
  6. 6. The system of claim 1, wherein the base device further comprises a speaker for playing dialogues or sound effects in the interactive instruction set.
  7. 7. The system of claim 1, wherein during the interactive instruction set is generated, the processing unit selects the first figure from the start, retrieves first interactive data for a specific topic for the first figure from the scenario data corresponding to the first figure, retrieves second interactive data corresponding to the first interactive data for the specific topic for the second figure from the scenario data corresponding to the second figure, and combines the first interactive data and the second interactive data as the interactive instruction set.
  8. 8. A figure interactive system, comprising:
    at least a first figure and a second figure;
    a base device, comprising a detecting unit and a first communication unit, wherein the detecting unit respectively detects identification data of the first figure and the second figure; and
    an electronic device, comprising a second communication unit for communicating with the first communication unit via a communication connection, a storage unit for storing a content database, and a processing unit for respectively retrieving scenario data corresponding to the first figure and the second figure from the content database, dynamically generating an interactive instruction set for the first figure and the second figure according to the scenario data corresponding to the first figure and the second figure, and transmitting the interactive instruction set to the base device,
    wherein the base device enables the first figure and the second figure to interact with each other according to the interactive instruction set.
  9. 9. The system of claim 8, wherein the processing unit transmits the identification data of the first figure and the second figure to a server via a network, and the server respectively retrieves the scenario data corresponding to the first figure and the second figure according to the identification data of the first figure and the second figure, and transmits the scenario data corresponding to the first figure and the second figure to the content database of the storage unit via the network.
  10. 10. The system of claim 9, wherein the processing unit further receives renewed scenario data corresponding to the first figure and the second figure from the server via the network, and stores the renewed scenario data to the content database.
  11. 11. The system of claim 8, wherein the interactive instruction set further comprises an action drive command, and the first figure or the second figure respectively has at least a drive component for receiving the interactive instruction set from the base device, and driving the first figure or the second figure to perform an operation according to the action drive command in the interactive instruction set.
  12. 12. The system of claim 8, wherein during the interactive instruction set is generated, the processing unit selects the first figure from the start, retrieves first interactive data for a specific topic for the first figure from the scenario data corresponding to the first figure, retrieves second interactive data corresponding to the first interactive data for the specific topic for the second figure from the scenario data corresponding to the second figure, and combines the first interactive data and the second interactive data as the interactive instruction set.
  13. 13. A figure interactive method, comprising:
    respectively detecting identification data of a first figure and a second figure;
    respectively retrieving scenario data corresponding to the first figure and the second figure from a content database according to the identification data of the first figure and the second figure;
    dynamically generating an interactive instruction set for the first figure and the second figure according to the scenario data corresponding to the first figure and the second figure; and
    enabling the first figure and the second figure to interact with each other according to the interactive instruction set.
  14. 14. The method of claim 13, further comprising:
    transmitting the identification data of the first figure and the second figure to a server via a network, wherein the server respectively retrieves the scenario data corresponding to the first figure and the second figure according to the identification data of the first figure and the second figure; and
    receiving the scenario data corresponding to the first figure and the second figure via the network as the content database.
  15. 15. The method of claim 14, further comprising:
    receiving renewed scenario data corresponding to the first figure and the second figure from the server via the network; and
    storing the renewed scenario data to the content database.
  16. 16. The method of claim 13, further comprising:
    respectively transmitting the interactive instruction set to at least a drive component of the first figure or the second figure; and
    respectively driving the first figure or the second figure to perform an operation according to an action drive command in the interactive instruction set.
  17. 17. The method of claim 13, further comprising displaying dialogues or images in the interactive instruction set via a display unit.
  18. 18. The method of claim 13, further comprising playing dialogues or sound effects in the interactive instruction set via a speaker.
  19. 19. The method of claim 13, wherein during the interactive instruction set is generated, the method further comprises the steps of:
    selecting the first figure from the start;
    retrieving first interactive data for a specific topic for the first figure from the scenario data corresponding to the first figure;
    retrieving second interactive data corresponding to the first interactive data for the specific topic for the second figure from the scenario data corresponding to the second figure; and
    combining the first interactive data and the second interactive data as the interactive instruction set.
  20. 20. A machine-readable storage medium comprising a computer program, which, when executed, causes a device to perform a figure interactive method, and the method comprises:
    respectively detecting identification data of a first figure and a second figure;
    respectively retrieving scenario data corresponding to the first figure and the second figure from a content database according to the identification data of the first figure and the second figure;
    dynamically generating an interactive instruction set for the first figure and the second figure according to the scenario data corresponding to the first figure and the second figure; and
    enabling the first figure and the second figure to interact with each other according to the interactive instruction set.
US12782733 2009-12-10 2010-05-19 Figure interactive systems and methods Abandoned US20110143632A1 (en)

Priority Applications (2)

Application Number Priority Date Filing Date Title
TW98142210 2009-12-10
TW98142210 2009-12-10

Publications (1)

Publication Number Publication Date
US20110143632A1 true true US20110143632A1 (en) 2011-06-16

Family

ID=44143459

Family Applications (1)

Application Number Title Priority Date Filing Date
US12782733 Abandoned US20110143632A1 (en) 2009-12-10 2010-05-19 Figure interactive systems and methods

Country Status (1)

Country Link
US (1) US20110143632A1 (en)

Citations (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US5636994A (en) * 1995-11-09 1997-06-10 Tong; Vincent M. K. Interactive computer controlled doll
US5752880A (en) * 1995-11-20 1998-05-19 Creator Ltd. Interactive doll
US6089942A (en) * 1998-04-09 2000-07-18 Thinking Technology, Inc. Interactive toys
US6110000A (en) * 1998-02-10 2000-08-29 T.L. Products Promoting Co. Doll set with unidirectional infrared communication for simulating conversation
US6309275B1 (en) * 1997-04-09 2001-10-30 Peter Sui Lun Fong Interactive talking dolls
US6352478B1 (en) * 1997-08-18 2002-03-05 Creator, Ltd. Techniques and apparatus for entertainment sites, amusement parks and other information and/or entertainment dispensing sites
US20080160877A1 (en) * 2005-04-26 2008-07-03 Steven Lipman Toys
US20090275408A1 (en) * 2008-03-12 2009-11-05 Brown Stephen J Programmable interactive talking device

Patent Citations (10)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US5636994A (en) * 1995-11-09 1997-06-10 Tong; Vincent M. K. Interactive computer controlled doll
US5752880A (en) * 1995-11-20 1998-05-19 Creator Ltd. Interactive doll
US6309275B1 (en) * 1997-04-09 2001-10-30 Peter Sui Lun Fong Interactive talking dolls
US20060009113A1 (en) * 1997-04-09 2006-01-12 Fong Peter S L Interactive talking dolls
US7068941B2 (en) * 1997-04-09 2006-06-27 Peter Sui Lun Fong Interactive talking dolls
US6352478B1 (en) * 1997-08-18 2002-03-05 Creator, Ltd. Techniques and apparatus for entertainment sites, amusement parks and other information and/or entertainment dispensing sites
US6110000A (en) * 1998-02-10 2000-08-29 T.L. Products Promoting Co. Doll set with unidirectional infrared communication for simulating conversation
US6089942A (en) * 1998-04-09 2000-07-18 Thinking Technology, Inc. Interactive toys
US20080160877A1 (en) * 2005-04-26 2008-07-03 Steven Lipman Toys
US20090275408A1 (en) * 2008-03-12 2009-11-05 Brown Stephen J Programmable interactive talking device

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
Copy of of the WO 01/012285 A1 (22 February 2001), Liu, Dexter 69 pages. *

Similar Documents

Publication Publication Date Title
US8055708B2 (en) Multimedia spaces
US20110061017A1 (en) Systems and Methods for Haptically-Enhanced Text Interfaces
US20080177627A1 (en) Electronic advertisement delivery system and method
US20080201222A1 (en) Blog advertising
US20120246343A1 (en) Synchronizing digital content
US20140250219A1 (en) Synchronizing translated digital content
US8551186B1 (en) Audible alert for stolen user devices
US20050209868A1 (en) Real-time sales support and learning tool
US8904304B2 (en) Creation and exposure of embedded secondary content data relevant to a primary content page of an electronic book
US20130076788A1 (en) Apparatus, method and software products for dynamic content management
US20100205529A1 (en) Device, system, and method for creating interactive guidance with execution of operations
US8209223B2 (en) Video object tag creation and processing
US20120030015A1 (en) Automatic abstracted creative generation from a web site
US20130111408A1 (en) Method and apparatus for increasing the functionality of a user device in a locked state
US20140236596A1 (en) Emotion detection in voicemail
US20150325236A1 (en) Context specific language model scale factors
CN101917512A (en) Method and system for displaying head picture of contact person and mobile terminal
US20040056878A1 (en) Digital assistants
US20140089826A1 (en) System and method for a universal resident scalable navigation and content display system compatible with any digital device using scalable transparent adaptable resident interface design and picto-overlay interface enhanced trans-snip technology
CN103442001A (en) Information recommendation method, device and server
US20150281401A1 (en) Hybrid Client/Server Architecture for Parallel Processing
US20120030014A1 (en) Automatic Ad Creative Generation
US20130159001A1 (en) Satisfying specified intent(s) based on multimodal request(s)
CN101238508A (en) Redictation of misrecognized words using a list of alternatives
CN102622400A (en) Electronic-book extended reading mark generating method and relevant equipment

Legal Events

Date Code Title Description
AS Assignment

Owner name: INSTITUTE FOR INFORMATION INDUSTRY, TAIWAN

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:LIN, SHENG-CHUN;CHANG, YU-CHUAN;HUNG, YU-SHIANG;REEL/FRAME:024425/0708

Effective date: 20100302