WO2010029557A1 - Content personalization - Google Patents
Content personalization Download PDFInfo
- Publication number
- WO2010029557A1 WO2010029557A1 PCT/IL2009/000895 IL2009000895W WO2010029557A1 WO 2010029557 A1 WO2010029557 A1 WO 2010029557A1 IL 2009000895 W IL2009000895 W IL 2009000895W WO 2010029557 A1 WO2010029557 A1 WO 2010029557A1
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- content
- viewer
- presentation device
- identifier
- content presentation
- Prior art date
Links
- 238000000034 method Methods 0.000 claims abstract description 21
- 238000004891 communication Methods 0.000 claims description 3
- 238000010586 diagram Methods 0.000 description 2
- 230000008901 benefit Effects 0.000 description 1
- 230000008859 change Effects 0.000 description 1
- 238000005516 engineering process Methods 0.000 description 1
- 238000011156 evaluation Methods 0.000 description 1
- 230000001815 facial effect Effects 0.000 description 1
- 238000012986 modification Methods 0.000 description 1
- 230000004048 modification Effects 0.000 description 1
- 230000008520 organization Effects 0.000 description 1
- 230000002093 peripheral effect Effects 0.000 description 1
- 230000004044 response Effects 0.000 description 1
- 230000000007 visual effect Effects 0.000 description 1
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06Q—INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES; SYSTEMS OR METHODS SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES, NOT OTHERWISE PROVIDED FOR
- G06Q30/00—Commerce
- G06Q30/02—Marketing; Price estimation or determination; Fundraising
Definitions
- the field of the present invention is digital media players.
- multiple users share the same entertainment device.
- Family members may share the same television, the same computer, and the same stereo deck.
- Some entertainment devices are programmable to automatically play content selected by a user.
- a user may manually select content to be automatically played, or indicate preferences such as artist, title and genre.
- preferences such as artist, title and genre.
- they may define multiple profiles, each profile indicating preferences of a corresponding user.
- a Microsoft Windows user profile for example, is used to configure personal computer parameter settings for a user, including settings for the Windows environment and settings relating to pictures, video and other such media.
- the user defines his profile and stores his preferred content files in a designated directory.
- the user wants to have his computer screensaver present a slideshow of his pictures, he must store his pictures in the designated directory.
- the screensaver present a slideshow of different pictures then the other user must store his pictures in a different directory, and configure his profile accordingly, and change the currently active profile of the computer.
- aspects of the present invention relate to a method and apparatus for automatically personalizing presentation of media content based on the identity of the person enjoying the presentation, without manual intervention, so that the content being presented is the person's preferred content.
- Embodiments of the present invention may be implemented in a variety of presentation devices, including inter alia digital picture frames, stereo decks, video decks, radios, televisions, computers, and other such entertainment appliances which often play content continuously and uninterruptedly over periods of time.
- Embodiments of the present invention detect identities of one or more people enjoying content on a player device at any given time, from IDs received from their transmitters, from biometrics, from voice recognition, or from facial or other images captured by one or more cameras.
- Embodiments of the present invention associate media content with people based on their playlists, based on their preferences, based on metadata tags in content files, and by applying face and voice recognition to images and videos.
- a method for dynamic real-time content personalization and display including collecting data about at least one viewer in the vicinity of a content presentation device, identifying the at least one viewer from the collected data, locating content associated with the identified at least one viewer, and automatically presenting the located content.
- a content presentation device with content personalization functionality including a data collector, for collecting data about at least one viewer in the vicinity of a content presentation device, a viewer identifier communicatively coupled with the data collector, for identifying the at least one viewer from the data collected by the data collector, a content locator communicatively coupled with the viewer identifier, for locating content associated with the at least one viewer identified by the viewer identifier, and a media player communicatively coupled with the content locator, for automatically presenting the content located by the content locator.
- CONTENT refers broadly to media content including inter alia e- books, games, pictures, songs, slide shows, television shows, video clips and movies.
- ENJOYING CONTENT ⁇ refers broadly to listening, watching or interacting with content.
- PRESENTATION DEVICE - refers broadly to a content player including inter alia an audio player, a video player, an electronic picture frame, a radio, a television and a game play station.
- VIEWER - refers broadly a person who enjoys content presented by a presentation device. Generally, the viewer is in the vicinity of the presentation device.
- FIG. 1 is a simplified flowchart of a method for dynamic real-time content personalization and display, in accordance with an embodiment of the present invention.
- FIG. 2 is a simplified block diagram of a presentation device having content personalization functionality, in accordance with an embodiment of the present invention.
- Embodiments of the present invention relate to a media presentation device with functionality for automatically presenting content that is associated with one or more viewers of the content.
- the presentation device identifies the viewer, and in turn presents content that is determined to be associated with the identified viewer.
- FIG. 1 is a simplified flowchart of a method for dynamic real-time content personalization and display, in accordance with an embodiment of the present invention.
- FIG. 2 is a simplified block diagram of a presentation device 200 having content personalization functionality, in accordance with an embodiment of the present invention. Steps 110, 120, 130 and 140 shown in FIG. 1 are performed by correspondingly numbered components of presentation device 200 shown in FIG. 2, which include a data collector 210, a viewer identifier 220, a content locator 230, and a media player 240.
- player device 200 is a passive device, which automatically presents content uninterruptedly without manual intervention.
- Presentation device 200 may be a digital picture frame, which automatically presents a slide show of pictures. Presentation device 200 may be a stereo or video deck, which automatically plays music or movies. Presentation device 200 may be a radio, which plays broadcast sound. Presentation device 200 may be a television, which automatically plays broadcast TV shows. Presentation device 200 may be a computer, which automatically presents a Screensaver when it is in an idle state.
- Step 110 data collector 210 collects data about viewers 250 in its vicinity.
- Step 110 may be implemented in many different ways.
- Step 110 may be implemented by receiving electronic IDs from viewers' devices.
- viewers 250 may have cell phones with Bluetooth IDs, near-field communication (NFC) IDs, radio frequency IDs (RFID), bar code IDs, or other such identifiers.
- data collector 210 is a corresponding Bluetooth receiver, NFC receiver, RFID receiver, bar code scanner, or such other receiver or scanner.
- Step 110 may be implemented by scanning viewer biometrics; e.g., by scanning an eye iris, scanning a fingerprint, or scanning a palm.
- step 110 may be implemented by recording a voice.
- data collector 210 is a corresponding iris scanner, fingerprint scanner, palm scanner, or voice recorder.
- Step 110 may be implemented by analyzing images captured by a still or video camera located on or near the player device.
- data collector 210 is a camera.
- viewer identifier 220 identifies the viewers 250 in its vicinity from the data collected at step 110. For example, viewer identifier 220 may look up an electronic ID in a viewer database 260. Alternatively, viewer identifier 220 may employ iris recognition, fingerprint recognition, palm recognition or voice recognition software. Alternatively, viewer identifier 220 may employ face recognition software to identify one or more persons in captured images. E.g., viewer identifier 220 may use the OKAO VISIONTM face sensing software developed and marketed by OMRON Corporation of Kyoto, Japan, or the Face Sensing Engine developed and marketed by Oki Electric Industry Co., Ltd. of Tokyo, Japan.
- content locator 230 locates content associated with the viewers 250 identified at step 120.
- Content locator 230 may consult a content database 270 that indexes content according to viewer association. Association of content with viewers may be performed in many different ways.
- One or more tags may be associated with a viewer, including inter alia a tag for the viewer himself, and tags for topics of interest. For example, a viewer may wish to see pictures of himself, pictures of his family, pictures of sunsets, and/or pictures from designated folders.
- Audio files may be associated with viewers based on existing playlists associated with viewers, and based on preset viewer preferences such as viewer preferred genres, and by identifying the viewer's voice in the files.
- Image and video files may be associated with viewers by cataloging the files according to people included in the images and videos, based on face recognition and other such recognition techniques.
- Software such as the content-based image organization application developed and marketed by Picporta, Inc. of Ahmedabad, India, and the visual search application developed by Riya, Inc. of Bangalore, India, may be used to do the cataloging.
- audio, image and video files may be associated with viewers based on informational metadata tags in the files.
- the FACEBOOK ® system developed and marketed by Facebook, Inc. of Palo Alto, CA, enables users to tag people in photos by marking areas in the photos.
- viewer database 260 may be local to presentation device 200, as indicated in FIG. 2, or may be remotely accessible via a network, or may be partially local and partially remote.
- Viewer identifier 220 may be an internal component of presentation device 200, or an external component communicatively coupled with presentation device 200.
- content database 270 may be local to presentation device 200, as indicated in FIG. 2, or may be remotely accessible via a network, or may be partially local and partially remote.
- content database 270 may be local to presentation device 200, as indicated in FIG. 2, or may be remotely accessible via a network, or may be partially local and partially remote.
- the Kodak EasyShare EX-1011 Digital Picture Frame developed and manufactured by the Eastman Kodak Company of Rochester, NY, presents content that is stored remotely on a PC or at an online photo-sharing service.
- Content locator 230 may be an internal component of presentation device 200, or an external component communicatively coupled with presentation device 200.
- the content located at step 130 by content locator 230 is automatically presented by media player 240.
- media player 240 gives priority to content that is associated with the multiple identified viewers.
- media player 240 rotates its presentation between content associated with each identified viewer. As such, the presentation time is divided between content presented for the multiple viewers.
- specific content is designated as default content, and when one or more viewers are identified, media player 240 rotates its presentation between default content and content associated with each identified viewer.
- media player 240 uses predefined rules for content to be presented, based on viewers identified at step 120. For example, pre-designated content is prevented from being presented if one or more pre-designated viewers are identified as being in the vicinity of media player 240.
- the method of FIG. 1 periodically returns to step 110, in order to regularly determine if one or more previously identified viewers leave the vicinity of presentation device 200, and if one or more new viewers enter the vicinity.
- embodiments of the present invention dynamically search for new viewers in the vicinity of presentation device 200, and present relevant content in real-time, quickly in response to identification of such new viewers.
Landscapes
- Business, Economics & Management (AREA)
- Engineering & Computer Science (AREA)
- Accounting & Taxation (AREA)
- Development Economics (AREA)
- Strategic Management (AREA)
- Finance (AREA)
- Game Theory and Decision Science (AREA)
- Entrepreneurship & Innovation (AREA)
- Economics (AREA)
- Marketing (AREA)
- Physics & Mathematics (AREA)
- General Business, Economics & Management (AREA)
- General Physics & Mathematics (AREA)
- Theoretical Computer Science (AREA)
- Two-Way Televisions, Distribution Of Moving Picture Or The Like (AREA)
Abstract
A method for dynamic real-time content personalization and display, including collecting data about at least one viewer in the vicinity of a content presentation device, identifying the at least one viewer from the collected data, locating content associated with the identified at least one viewer, and automatically presenting the located content. A system is also presented and claimed.
Description
CONTENT PERSONALIZATION
FIELD OF THE INVENTION The field of the present invention is digital media players.
BACKGROUND OF THE INVENTION
Often, multiple users share the same entertainment device. Family members, for example, may share the same television, the same computer, and the same stereo deck. Some entertainment devices are programmable to automatically play content selected by a user. A user may manually select content to be automatically played, or indicate preferences such as artist, title and genre. When there are multiple users, they may define multiple profiles, each profile indicating preferences of a corresponding user.
A Microsoft Windows user profile, for example, is used to configure personal computer parameter settings for a user, including settings for the Windows environment and settings relating to pictures, video and other such media. To configure his parameter settings, the user defines his profile and stores his preferred content files in a designated directory. Thus if the user wants to have his computer screensaver present a slideshow of his pictures, he must store his pictures in the designated directory. If another user wants to have the screensaver present a slideshow of different pictures, then the other user must store his pictures in a different directory, and configure his profile accordingly, and change the currently active profile of the computer.
It would thus be of advantage for a shared entertainment device to be able to automatically personalize its content presentation according to the preferences of the person viewing the content, without manual intervention.
SUMMARY OF THE DESCRIPTION
Aspects of the present invention relate to a method and apparatus for automatically personalizing presentation of media content based on the identity of the person enjoying the presentation, without manual intervention, so that the content being presented is the person's preferred content. Embodiments of the present invention may be implemented in a variety of presentation devices, including inter alia digital picture frames, stereo decks, video decks, radios, televisions, computers, and other such entertainment appliances which often play content continuously and uninterruptedly over periods of time. Embodiments of the present invention detect identities of one or more people enjoying content on a player device at any given time, from IDs received from their transmitters, from biometrics, from voice recognition, or from facial or other images captured by one or more cameras.
Embodiments of the present invention associate media content with people based on their playlists, based on their preferences, based on metadata tags in content files, and by applying face and voice recognition to images and videos.
There is thus provided in accordance with an embodiment of the present invention a method for dynamic real-time content personalization and display, including collecting data about at least one viewer in the vicinity of a content presentation device, identifying the at least one viewer from the collected data, locating content associated with the identified at least one viewer, and automatically presenting the located content.
There is further provided in accordance with an embodiment of the present invention a content presentation device with content personalization functionality, including a data collector, for collecting data about at least one viewer in the vicinity of a content presentation device, a viewer identifier communicatively coupled with the data collector, for identifying the at least one viewer from the data collected by the data collector, a content locator communicatively coupled with the viewer identifier, for locating content associated with the at least one viewer identified by the viewer identifier, and a media player communicatively coupled with the content locator, for automatically presenting the content located by the content locator.
The following definitions are employed throughout the specification.
CONTENT — refers broadly to media content including inter alia e- books, games, pictures, songs, slide shows, television shows, video clips and movies.
ENJOYING CONTENT ~ refers broadly to listening, watching or interacting with content. PRESENTATION DEVICE - refers broadly to a content player including inter alia an audio player, a video player, an electronic picture frame, a radio, a television and a game play station.
VIEWER - refers broadly a person who enjoys content presented by a presentation device. Generally, the viewer is in the vicinity of the presentation device.
BRIEF DESCRIPTION OF THE DRAWINGS
The present invention will be more fully understood and appreciated from the following detailed description, taken in conjunction with the drawings in which: FIG. 1 is a simplified flowchart of a method for dynamic real-time content personalization and display, in accordance with an embodiment of the present invention; and
FIG. 2 is a simplified block diagram of a presentation device having content personalization functionality, in accordance with an embodiment of the present invention.
DETAILED DESCRIPTION
Embodiments of the present invention relate to a media presentation device with functionality for automatically presenting content that is associated with one or more viewers of the content. The presentation device identifies the viewer, and in turn presents content that is determined to be associated with the identified viewer.
Reference is now made to FIG. 1, which is a simplified flowchart of a method for dynamic real-time content personalization and display, in accordance with an embodiment of the present invention. In conjunction with FIG. 1, reference is also made to FIG. 2, which is a simplified block diagram of a presentation device 200 having content personalization functionality, in accordance with an embodiment of the present invention. Steps 110, 120, 130 and 140 shown in FIG. 1 are performed by correspondingly numbered components of presentation device 200 shown in FIG. 2, which include a data collector 210, a viewer identifier 220, a content locator 230, and a media player 240. According to an embodiment of the present invention, player device 200 is a passive device, which automatically presents content uninterruptedly without manual intervention. Presentation device 200 may be a digital picture frame, which automatically presents a slide show of pictures. Presentation device 200 may be a stereo or video deck, which automatically plays music or movies. Presentation device 200 may be a radio, which plays broadcast sound. Presentation device 200 may be a television, which automatically plays broadcast TV shows. Presentation device 200 may be a computer, which automatically presents a Screensaver when it is in an idle state.
At step 110, data collector 210 collects data about viewers 250 in its vicinity. Step 110 may be implemented in many different ways.
Step 110 may be implemented by receiving electronic IDs from viewers' devices. For example, viewers 250 may have cell phones with Bluetooth IDs, near-field communication (NFC) IDs, radio frequency IDs (RFID), bar code IDs, or other such identifiers. For such implementation, data collector 210 is a corresponding Bluetooth receiver, NFC receiver, RFID receiver, bar code scanner, or such other receiver or scanner.
Step 110 may be implemented by scanning viewer biometrics; e.g., by scanning an eye iris, scanning a fingerprint, or scanning a palm. Alternatively, step 110 may be implemented by recording a voice. For such implementation, data collector 210 is a corresponding iris scanner, fingerprint scanner, palm scanner, or voice recorder. Step 110 may be implemented by analyzing images captured by a still or video camera located on or near the player device. For such implementation, data collector 210 is a camera.
If no viewers are detected at step 110, then the method repeats step 110 periodically, until one or more viewers are detected. At step 120 viewer identifier 220 identifies the viewers 250 in its vicinity from the data collected at step 110. For example, viewer identifier 220 may look up an electronic ID in a viewer database 260. Alternatively, viewer identifier 220 may employ iris recognition, fingerprint recognition, palm recognition or voice recognition software. Alternatively, viewer identifier 220 may employ face recognition software to identify one or more persons in captured images. E.g., viewer identifier 220 may use the OKAO VISION™ face sensing software developed and marketed by OMRON Corporation of Kyoto, Japan, or the Face Sensing Engine developed and marketed by Oki Electric Industry Co., Ltd. of Tokyo, Japan.
At step 130 content locator 230 locates content associated with the viewers 250 identified at step 120. Content locator 230 may consult a content database 270 that indexes content according to viewer association. Association of content with viewers may be performed in many different ways. One or more tags may be associated with a viewer, including inter alia a tag for the viewer himself, and tags for topics of interest. For example, a viewer may wish to see pictures of himself, pictures of his family, pictures of sunsets, and/or pictures from designated folders.
Audio files may be associated with viewers based on existing playlists associated with viewers, and based on preset viewer preferences such as viewer preferred genres, and by identifying the viewer's voice in the files. Image and video files may be associated with viewers by cataloging the files according to people included in the images and videos, based on face recognition and other such recognition techniques. Software, such as the content-based image organization application developed and marketed by Picporta, Inc. of Ahmedabad, India, and the visual search
application developed by Riya, Inc. of Bangalore, India, may be used to do the cataloging. Alternatively, audio, image and video files may be associated with viewers based on informational metadata tags in the files. The FACEBOOK® system, developed and marketed by Facebook, Inc. of Palo Alto, CA, enables users to tag people in photos by marking areas in the photos.
A summary and evaluation of face recognition technologies that may be used for automatic cataloging of image collections is presented in Corcoran, P. and Costache, G., "The automated sorting of consumer image collections using face and peripheral region image classifiers", IEEE Trans. Consumer Electronics, Vol. 51, No. 3, August 2005, pgs. 747 - 754.
It will be appreciated by those skilled in the art that viewer database 260 may be local to presentation device 200, as indicated in FIG. 2, or may be remotely accessible via a network, or may be partially local and partially remote. Viewer identifier 220 may be an internal component of presentation device 200, or an external component communicatively coupled with presentation device 200.
It will further be appreciated by those skilled in the art that content database 270 may be local to presentation device 200, as indicated in FIG. 2, or may be remotely accessible via a network, or may be partially local and partially remote. For example, the Kodak EasyShare EX-1011 Digital Picture Frame, developed and manufactured by the Eastman Kodak Company of Rochester, NY, presents content that is stored remotely on a PC or at an online photo-sharing service. Content locator 230 may be an internal component of presentation device 200, or an external component communicatively coupled with presentation device 200.
At step 140 the content located at step 130 by content locator 230, is automatically presented by media player 240. hi case more than one viewer was identified at step 120, media player 240 gives priority to content that is associated with the multiple identified viewers. Additionally, or alternatively, media player 240 rotates its presentation between content associated with each identified viewer. As such, the presentation time is divided between content presented for the multiple viewers. In accordance with another embodiment of the present invention, specific content is designated as default content, and when one or more viewers are
identified, media player 240 rotates its presentation between default content and content associated with each identified viewer.
In an embodiment of the present invention, media player 240 uses predefined rules for content to be presented, based on viewers identified at step 120. For example, pre-designated content is prevented from being presented if one or more pre-designated viewers are identified as being in the vicinity of media player 240.
The method of FIG. 1 periodically returns to step 110, in order to regularly determine if one or more previously identified viewers leave the vicinity of presentation device 200, and if one or more new viewers enter the vicinity. As such it will be appreciated by those skilled in the art that embodiments of the present invention dynamically search for new viewers in the vicinity of presentation device 200, and present relevant content in real-time, quickly in response to identification of such new viewers.
[0001] In the foregoing specification, the invention has been described with reference to specific exemplary embodiments thereof. It will, however, be evident that various modifications and changes may be made to the specific exemplary embodiments without departing from the broader spirit and scope of the invention as set forth in the appended claims. Accordingly, the specification and drawings are to be regarded in an illustrative rather than a restrictive sense.
Claims
1. A method for dynamic real-time content personalization and display, 5 comprising: collecting data about at least one viewer in the vicinity of a content presentation device; identifying the at least one viewer from the collected data; locating content associated with the identified at least one viewer; and l o automatically presenting the located content.
2. The method of claim 1 wherein said collecting data comprises receiving at least one ID from at least one electronic device in possession of the at least one viewer, and wherein said identifying comprises identifying the at least one received ID.
15
3. The method of claim 2 wherein the at least one ID includes a Bluetooth ID, or a radio frequency ID (RFID), or a near-field communication (NFC) ID.
4. The method of claim 1 wherein said collecting data comprises scanning 0 at least one bar code corresponding to the at least one viewer, and wherein said identifying comprises identifying the at least one scanned bar code.
5. The method of claim 1 wherein said collecting data comprises scanning at least one eye iris, or at least one fingerprint, or at least one palm of the at least one 5 viewer, and wherein said identifying comprises identifying the at least one scanned Ms, or the at least one scanned fingerprint, or the at least one scanned palm, respectively.
6. The method of claim 1 wherein said collecting data comprises recording at least one voice of the at least one viewer, and wherein said identifying comprises 0 identifying the at least one recorded voice.
7. The method of claim 1 wherein said collecting data comprises capturing at least one photograph of the at least one viewer, and wherein said identifying comprises recognizing at least one face in the at least one captured photograph.
5 8. The method of claim 1 wherein said collecting data comprises capturing at least one video frame of the at least one viewer, and wherein said identifying comprises recognizing at least one face in the at least one captured video frame.
9. The method of claim 1 wherein said locating content comprises reading o playlists.
10. The method of claim 1 where said locating content comprises reading content file metadata tags. 5
11. The method of claim 1 wherein the at least one viewer comprises a plurality of viewers, and wherein said automatically presenting comprises rotating presenting between content associated with each one of the plurality of viewers.
12. The method of claim 1 wherein the at least one viewer comprises a o plurality of viewers, and wherein said automatically presenting comprises presenting content that is associated with more than one of the plurality of viewers.
13. The method of claim 1 wherein said automatically presenting comprises preventing presentation of pre-designated content if the at least one viewer comprises one or more pre-designated viewers.
14. A content presentation device with content personalization functionality, comprising: a data collector, for collecting data about at least one viewer in the vicinity of a content presentation device; a viewer identifier communicatively coupled with said data collector, for identifying the at least one viewer from the data collected by said data collector; a content locator communicatively coupled with said viewer identifier, for locating content associated with the at least one viewer identified by said viewer identifier; and a media player communicatively coupled with said content locator, for automatically presenting the content located by said content locator.
15. The content presentation device of claim 14 wherein said viewer identifier is physically coupled with said data collector.
16. The content presentation device of claim 14 wherein said viewer identifier is remote from said data collector, and is wirelessly coupled with said data collector.
17. The content presentation device of claim 14 further comprising a viewer database for storing viewer information, and wherein said viewer identifier accesses said viewer database to identify the at least one viewer.
18. The content presentation device of claim 17 wherein said viewer database is physically coupled with said viewer identifier.
19. The content presentation device of claim 17 wherein said viewer database is wirelessly coupled with said viewer identifier.
20. The content presentation device of claim 17 wherein said viewer database is remote from said viewer identifier.
21. The content presentation device of claim 14 wherein said content locator is physically coupled with said viewer identifier.
22. The content presentation device of claim 14 wherein said content locator is remote from said viewer identifier, and is wirelessly coupled with said viewer identifier.
23. The content presentation device of claim 14 wherein said content locator is physically coupled with said media player.
24. The content presentation device of claim 14 wherein said content locator is remote from said media player, and is wirelessly coupled with said media player.
25. The content presentation device of claim 14 further comprising a content database for storing content information, and wherein said content locator accesses said content database to locate the content associated with the at least one viewer.
26. The content presentation device of claim 25 wherein said content database is physically coupled with said content locator.
27. The content presentation device of claim 25 wherein said content database is wirelessly coupled with said content locator.
28. The content presentation device of claim 25 wherein said viewer database is remote from said content locator.
29. The content presentation device of claim 14 wherein said data collector comprises a receiver for receiving at least one ID from at least one electronic device in possession of the at least one viewer, and wherein said viewer identifier comprises an ID identifier.
30. The content presentation device of claim 29 wherein said receiver is a Bluetooth receiver, or a radio frequency ID (RFID) receiver, or a near-field communication (NFC) receiver.
31. The content presentation device of claim 14 wherein said data collector comprises a barcode scanner, and wherein said viewer identifier comprises a barcode identifier.
32. The content presentation device of claim 14 wherein said data collector comprises an eye iris scanner, or a fingerprint scanner, or a palm scanner, and wherein said viewer identifier comprise an iris identifier, a fingerprint identifier, or a palm
5 identifier, respectively.
33. The content presentation device of claim 14 wherein said data collector comprises a voice recorder, and wherein said viewer identifier comprises a voice recognizer. 0
34. The content presentation device of claim 14 wherein said data collector comprises a camera, and wherein said viewer identifier comprises a face recognizer.
35. The content presentation device of claim 14 wherein said media player is5 an audio player or a video player or a slideshow presenter.
36. The content presentation device of claim 14 wherein said media player is a digital picture frame.
o 37. The content presentation device of claim 14 wherein said media player is a computer Screensaver renderer.
38. The content presentation device of claim 14 wherein said media player is a radio or a television. 5
39. The content presentation device of claim 14 wherein said media player is a game play station.
40. The content presentation device of claim 14 wherein said media player is an e-book reader.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
EP09812786.3A EP2377030A4 (en) | 2008-09-14 | 2009-09-13 | Content personalization |
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US12/210,199 | 2008-09-14 | ||
US12/210,199 US20100071003A1 (en) | 2008-09-14 | 2008-09-14 | Content personalization |
Publications (1)
Publication Number | Publication Date |
---|---|
WO2010029557A1 true WO2010029557A1 (en) | 2010-03-18 |
Family
ID=42004855
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
PCT/IL2009/000895 WO2010029557A1 (en) | 2008-09-14 | 2009-09-13 | Content personalization |
Country Status (3)
Country | Link |
---|---|
US (1) | US20100071003A1 (en) |
EP (1) | EP2377030A4 (en) |
WO (1) | WO2010029557A1 (en) |
Cited By (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US9542083B2 (en) | 2014-12-04 | 2017-01-10 | Comcast Cable Communications, Llc | Configuration responsive to a device |
Families Citing this family (141)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US8677377B2 (en) | 2005-09-08 | 2014-03-18 | Apple Inc. | Method and apparatus for building an intelligent automated assistant |
US9318108B2 (en) | 2010-01-18 | 2016-04-19 | Apple Inc. | Intelligent automated assistant |
US10002189B2 (en) | 2007-12-20 | 2018-06-19 | Apple Inc. | Method and apparatus for searching using an active ontology |
US8270303B2 (en) * | 2007-12-21 | 2012-09-18 | Hand Held Products, Inc. | Using metadata tags in video recordings produced by portable encoded information reading terminals |
US9330720B2 (en) | 2008-01-03 | 2016-05-03 | Apple Inc. | Methods and apparatus for altering audio output signals |
US8996376B2 (en) | 2008-04-05 | 2015-03-31 | Apple Inc. | Intelligent text-to-speech conversion |
US20100030549A1 (en) | 2008-07-31 | 2010-02-04 | Lee Michael M | Mobile device having human language translation capability with positional feedback |
US8676904B2 (en) | 2008-10-02 | 2014-03-18 | Apple Inc. | Electronic devices with voice command and contextual data processing capabilities |
US20100198876A1 (en) | 2009-02-02 | 2010-08-05 | Honeywell International, Inc. | Apparatus and method of embedding meta-data in a captured image |
US10706373B2 (en) | 2011-06-03 | 2020-07-07 | Apple Inc. | Performing actions associated with task items that represent tasks to perform |
US10241752B2 (en) | 2011-09-30 | 2019-03-26 | Apple Inc. | Interface for a virtual digital assistant |
US10241644B2 (en) | 2011-06-03 | 2019-03-26 | Apple Inc. | Actionable reminder entries |
US9519814B2 (en) | 2009-06-12 | 2016-12-13 | Hand Held Products, Inc. | Portable data terminal |
US9431006B2 (en) | 2009-07-02 | 2016-08-30 | Apple Inc. | Methods and apparatuses for automatic speech recognition |
US10276170B2 (en) | 2010-01-18 | 2019-04-30 | Apple Inc. | Intelligent automated assistant |
US8269813B2 (en) * | 2010-02-25 | 2012-09-18 | Coby Neuenschwander | Enterprise system and computer program product for inter-connecting multiple parties in an interactive environment exhibiting virtual picture books |
US8682667B2 (en) | 2010-02-25 | 2014-03-25 | Apple Inc. | User profiling for selecting user specific voice input processing information |
JP5528318B2 (en) * | 2010-03-23 | 2014-06-25 | パナソニック株式会社 | Display device |
US8849199B2 (en) * | 2010-11-30 | 2014-09-30 | Cox Communications, Inc. | Systems and methods for customizing broadband content based upon passive presence detection of users |
TWI422504B (en) * | 2010-12-31 | 2014-01-11 | Altek Corp | Vehicle apparatus control system and method thereof |
US9262612B2 (en) | 2011-03-21 | 2016-02-16 | Apple Inc. | Device access using voice authentication |
US10057736B2 (en) | 2011-06-03 | 2018-08-21 | Apple Inc. | Active transport based notifications |
US10134385B2 (en) | 2012-03-02 | 2018-11-20 | Apple Inc. | Systems and methods for name pronunciation |
US10417037B2 (en) | 2012-05-15 | 2019-09-17 | Apple Inc. | Systems and methods for integrating third party services with a digital assistant |
US9721563B2 (en) | 2012-06-08 | 2017-08-01 | Apple Inc. | Name recognition system |
US9547647B2 (en) * | 2012-09-19 | 2017-01-17 | Apple Inc. | Voice-based media searching |
CN113470641B (en) | 2013-02-07 | 2023-12-15 | 苹果公司 | Voice trigger of digital assistant |
US10652394B2 (en) | 2013-03-14 | 2020-05-12 | Apple Inc. | System and method for processing voicemail |
US10748529B1 (en) | 2013-03-15 | 2020-08-18 | Apple Inc. | Voice activated device for use with a voice-based digital assistant |
WO2014197334A2 (en) | 2013-06-07 | 2014-12-11 | Apple Inc. | System and method for user-specified pronunciation of words for speech synthesis and recognition |
WO2014197335A1 (en) | 2013-06-08 | 2014-12-11 | Apple Inc. | Interpreting and acting upon commands that involve sharing information with remote devices |
KR101772152B1 (en) | 2013-06-09 | 2017-08-28 | 애플 인크. | Device, method, and graphical user interface for enabling conversation persistence across two or more instances of a digital assistant |
US10176167B2 (en) | 2013-06-09 | 2019-01-08 | Apple Inc. | System and method for inferring user intent from speech inputs |
CN105453026A (en) | 2013-08-06 | 2016-03-30 | 苹果公司 | Auto-activating smart responses based on activities from remote devices |
US10296160B2 (en) | 2013-12-06 | 2019-05-21 | Apple Inc. | Method for extracting salient dialog usage from live data |
US20150248702A1 (en) * | 2014-03-03 | 2015-09-03 | Ebay Inc. | Proximity-based visual notifications |
US9715875B2 (en) | 2014-05-30 | 2017-07-25 | Apple Inc. | Reducing the need for manual start/end-pointing and trigger phrases |
US10170123B2 (en) | 2014-05-30 | 2019-01-01 | Apple Inc. | Intelligent assistant for home automation |
US9633004B2 (en) | 2014-05-30 | 2017-04-25 | Apple Inc. | Better resolution when referencing to concepts |
US9430463B2 (en) | 2014-05-30 | 2016-08-30 | Apple Inc. | Exemplar-based natural language processing |
WO2015184186A1 (en) | 2014-05-30 | 2015-12-03 | Apple Inc. | Multi-command single utterance input method |
US9338493B2 (en) | 2014-06-30 | 2016-05-10 | Apple Inc. | Intelligent automated assistant for TV user interactions |
US9818400B2 (en) | 2014-09-11 | 2017-11-14 | Apple Inc. | Method and apparatus for discovering trending terms in speech requests |
US10074360B2 (en) | 2014-09-30 | 2018-09-11 | Apple Inc. | Providing an indication of the suitability of speech recognition |
US10127911B2 (en) | 2014-09-30 | 2018-11-13 | Apple Inc. | Speaker identification and unsupervised speaker adaptation techniques |
US9668121B2 (en) | 2014-09-30 | 2017-05-30 | Apple Inc. | Social reminders |
CN105657483B (en) * | 2014-11-10 | 2019-06-04 | 扬智科技股份有限公司 | Multimedia play system, multimedia file sharing method and its control method |
US10152299B2 (en) | 2015-03-06 | 2018-12-11 | Apple Inc. | Reducing response latency of intelligent automated assistants |
US9721566B2 (en) | 2015-03-08 | 2017-08-01 | Apple Inc. | Competing devices responding to voice triggers |
US9886953B2 (en) | 2015-03-08 | 2018-02-06 | Apple Inc. | Virtual assistant activation |
US10567477B2 (en) | 2015-03-08 | 2020-02-18 | Apple Inc. | Virtual assistant continuity |
US10460227B2 (en) | 2015-05-15 | 2019-10-29 | Apple Inc. | Virtual assistant in a communication session |
US10083688B2 (en) | 2015-05-27 | 2018-09-25 | Apple Inc. | Device voice control for selecting a displayed affordance |
US9578173B2 (en) | 2015-06-05 | 2017-02-21 | Apple Inc. | Virtual assistant aided communication with 3rd party service in a communication session |
US11025565B2 (en) | 2015-06-07 | 2021-06-01 | Apple Inc. | Personalized prediction of responses for instant messaging |
US20160378747A1 (en) | 2015-06-29 | 2016-12-29 | Apple Inc. | Virtual assistant for media playback |
US10671428B2 (en) | 2015-09-08 | 2020-06-02 | Apple Inc. | Distributed personal assistant |
US10747498B2 (en) | 2015-09-08 | 2020-08-18 | Apple Inc. | Zero latency digital assistant |
US10366158B2 (en) | 2015-09-29 | 2019-07-30 | Apple Inc. | Efficient word encoding for recurrent neural network language models |
US11010550B2 (en) | 2015-09-29 | 2021-05-18 | Apple Inc. | Unified language modeling framework for word prediction, auto-completion and auto-correction |
US10691473B2 (en) | 2015-11-06 | 2020-06-23 | Apple Inc. | Intelligent automated assistant in a messaging environment |
US10049668B2 (en) | 2015-12-02 | 2018-08-14 | Apple Inc. | Applying neural network language models to weighted finite state transducers for automatic speech recognition |
US10223066B2 (en) | 2015-12-23 | 2019-03-05 | Apple Inc. | Proactive assistance based on dialog communication between devices |
US10446143B2 (en) | 2016-03-14 | 2019-10-15 | Apple Inc. | Identification of voice inputs providing credentials |
US9934775B2 (en) | 2016-05-26 | 2018-04-03 | Apple Inc. | Unit-selection text-to-speech synthesis based on predicted concatenation parameters |
US9972304B2 (en) | 2016-06-03 | 2018-05-15 | Apple Inc. | Privacy preserving distributed evaluation framework for embedded personalized systems |
US11227589B2 (en) | 2016-06-06 | 2022-01-18 | Apple Inc. | Intelligent list reading |
US10249300B2 (en) | 2016-06-06 | 2019-04-02 | Apple Inc. | Intelligent list reading |
US10049663B2 (en) | 2016-06-08 | 2018-08-14 | Apple, Inc. | Intelligent automated assistant for media exploration |
DK179309B1 (en) | 2016-06-09 | 2018-04-23 | Apple Inc | Intelligent automated assistant in a home environment |
US10586535B2 (en) | 2016-06-10 | 2020-03-10 | Apple Inc. | Intelligent digital assistant in a multi-tasking environment |
US10490187B2 (en) | 2016-06-10 | 2019-11-26 | Apple Inc. | Digital assistant providing automated status report |
US10509862B2 (en) | 2016-06-10 | 2019-12-17 | Apple Inc. | Dynamic phrase expansion of language input |
US10067938B2 (en) | 2016-06-10 | 2018-09-04 | Apple Inc. | Multilingual word prediction |
US10192552B2 (en) | 2016-06-10 | 2019-01-29 | Apple Inc. | Digital assistant providing whispered speech |
DK179049B1 (en) | 2016-06-11 | 2017-09-18 | Apple Inc | Data driven natural language event detection and classification |
DK179415B1 (en) | 2016-06-11 | 2018-06-14 | Apple Inc | Intelligent device arbitration and control |
DK179343B1 (en) | 2016-06-11 | 2018-05-14 | Apple Inc | Intelligent task discovery |
DK201670540A1 (en) | 2016-06-11 | 2018-01-08 | Apple Inc | Application integration with a digital assistant |
US10474753B2 (en) | 2016-09-07 | 2019-11-12 | Apple Inc. | Language identification using recurrent neural networks |
US10043516B2 (en) | 2016-09-23 | 2018-08-07 | Apple Inc. | Intelligent automated assistant |
US11281993B2 (en) | 2016-12-05 | 2022-03-22 | Apple Inc. | Model and ensemble compression for metric learning |
US10593346B2 (en) | 2016-12-22 | 2020-03-17 | Apple Inc. | Rank-reduced token representation for automatic speech recognition |
US11204787B2 (en) | 2017-01-09 | 2021-12-21 | Apple Inc. | Application integration with a digital assistant |
DK201770383A1 (en) | 2017-05-09 | 2018-12-14 | Apple Inc. | User interface for correcting recognition errors |
US10417266B2 (en) | 2017-05-09 | 2019-09-17 | Apple Inc. | Context-aware ranking of intelligent response suggestions |
US10395654B2 (en) | 2017-05-11 | 2019-08-27 | Apple Inc. | Text normalization based on a data-driven learning network |
DK201770439A1 (en) | 2017-05-11 | 2018-12-13 | Apple Inc. | Offline personal assistant |
US10726832B2 (en) | 2017-05-11 | 2020-07-28 | Apple Inc. | Maintaining privacy of personal information |
DK201770429A1 (en) | 2017-05-12 | 2018-12-14 | Apple Inc. | Low-latency intelligent automated assistant |
DK179745B1 (en) | 2017-05-12 | 2019-05-01 | Apple Inc. | SYNCHRONIZATION AND TASK DELEGATION OF A DIGITAL ASSISTANT |
DK179496B1 (en) | 2017-05-12 | 2019-01-15 | Apple Inc. | USER-SPECIFIC Acoustic Models |
US11301477B2 (en) | 2017-05-12 | 2022-04-12 | Apple Inc. | Feedback analysis of a digital assistant |
DK201770432A1 (en) | 2017-05-15 | 2018-12-21 | Apple Inc. | Hierarchical belief states for digital assistants |
DK201770431A1 (en) | 2017-05-15 | 2018-12-20 | Apple Inc. | Optimizing dialogue policy decisions for digital assistants using implicit feedback |
US20180336892A1 (en) | 2017-05-16 | 2018-11-22 | Apple Inc. | Detecting a trigger of a digital assistant |
US10311144B2 (en) | 2017-05-16 | 2019-06-04 | Apple Inc. | Emoji word sense disambiguation |
US10303715B2 (en) | 2017-05-16 | 2019-05-28 | Apple Inc. | Intelligent automated assistant for media exploration |
US10403278B2 (en) | 2017-05-16 | 2019-09-03 | Apple Inc. | Methods and systems for phonetic matching in digital assistant services |
DK179560B1 (en) | 2017-05-16 | 2019-02-18 | Apple Inc. | Far-field extension for digital assistant services |
US10657328B2 (en) | 2017-06-02 | 2020-05-19 | Apple Inc. | Multi-task recurrent neural network architecture for efficient morphology handling in neural language modeling |
US10445429B2 (en) | 2017-09-21 | 2019-10-15 | Apple Inc. | Natural language understanding using vocabularies with compressed serialized tries |
US10755051B2 (en) | 2017-09-29 | 2020-08-25 | Apple Inc. | Rule-based natural language processing |
US10636424B2 (en) | 2017-11-30 | 2020-04-28 | Apple Inc. | Multi-turn canned dialog |
US10733982B2 (en) | 2018-01-08 | 2020-08-04 | Apple Inc. | Multi-directional dialog |
US10733375B2 (en) | 2018-01-31 | 2020-08-04 | Apple Inc. | Knowledge-based framework for improving natural language understanding |
US10789959B2 (en) | 2018-03-02 | 2020-09-29 | Apple Inc. | Training speaker recognition models for digital assistants |
US10592604B2 (en) | 2018-03-12 | 2020-03-17 | Apple Inc. | Inverse text normalization for automatic speech recognition |
US10818288B2 (en) | 2018-03-26 | 2020-10-27 | Apple Inc. | Natural assistant interaction |
US10909331B2 (en) | 2018-03-30 | 2021-02-02 | Apple Inc. | Implicit identification of translation payload with neural machine translation |
US11145294B2 (en) | 2018-05-07 | 2021-10-12 | Apple Inc. | Intelligent automated assistant for delivering content from user experiences |
US10928918B2 (en) | 2018-05-07 | 2021-02-23 | Apple Inc. | Raise to speak |
US10984780B2 (en) | 2018-05-21 | 2021-04-20 | Apple Inc. | Global semantic word embeddings using bi-directional recurrent neural networks |
DK180639B1 (en) | 2018-06-01 | 2021-11-04 | Apple Inc | DISABILITY OF ATTENTION-ATTENTIVE VIRTUAL ASSISTANT |
DK179822B1 (en) | 2018-06-01 | 2019-07-12 | Apple Inc. | Voice interaction at a primary device to access call functionality of a companion device |
US11386266B2 (en) | 2018-06-01 | 2022-07-12 | Apple Inc. | Text correction |
US10892996B2 (en) | 2018-06-01 | 2021-01-12 | Apple Inc. | Variable latency device coordination |
DK201870355A1 (en) | 2018-06-01 | 2019-12-16 | Apple Inc. | Virtual assistant operation in multi-device environments |
US10496705B1 (en) | 2018-06-03 | 2019-12-03 | Apple Inc. | Accelerated task performance |
US11615134B2 (en) | 2018-07-16 | 2023-03-28 | Maris Jacob Ensing | Systems and methods for generating targeted media content |
US10831817B2 (en) | 2018-07-16 | 2020-11-10 | Maris Jacob Ensing | Systems and methods for generating targeted media content |
US10484818B1 (en) | 2018-09-26 | 2019-11-19 | Maris Jacob Ensing | Systems and methods for providing location information about registered user based on facial recognition |
US11010561B2 (en) | 2018-09-27 | 2021-05-18 | Apple Inc. | Sentiment prediction from textual data |
US10839159B2 (en) | 2018-09-28 | 2020-11-17 | Apple Inc. | Named entity normalization in a spoken dialog system |
US11170166B2 (en) | 2018-09-28 | 2021-11-09 | Apple Inc. | Neural typographical error modeling via generative adversarial networks |
US11462215B2 (en) | 2018-09-28 | 2022-10-04 | Apple Inc. | Multi-modal inputs for voice commands |
US11475898B2 (en) | 2018-10-26 | 2022-10-18 | Apple Inc. | Low-latency multi-speaker speech recognition |
US11638059B2 (en) | 2019-01-04 | 2023-04-25 | Apple Inc. | Content playback on multiple devices |
US11348573B2 (en) | 2019-03-18 | 2022-05-31 | Apple Inc. | Multimodality in digital assistant systems |
US11307752B2 (en) | 2019-05-06 | 2022-04-19 | Apple Inc. | User configurable task triggers |
US11423908B2 (en) | 2019-05-06 | 2022-08-23 | Apple Inc. | Interpreting spoken requests |
DK201970509A1 (en) | 2019-05-06 | 2021-01-15 | Apple Inc | Spoken notifications |
US11475884B2 (en) | 2019-05-06 | 2022-10-18 | Apple Inc. | Reducing digital assistant latency when a language is incorrectly determined |
US11140099B2 (en) | 2019-05-21 | 2021-10-05 | Apple Inc. | Providing message response suggestions |
US11496600B2 (en) | 2019-05-31 | 2022-11-08 | Apple Inc. | Remote execution of machine-learned models |
DK201970510A1 (en) | 2019-05-31 | 2021-02-11 | Apple Inc | Voice identification in digital assistant systems |
DK180129B1 (en) | 2019-05-31 | 2020-06-02 | Apple Inc. | User activity shortcut suggestions |
US11289073B2 (en) | 2019-05-31 | 2022-03-29 | Apple Inc. | Device text to speech |
US11360641B2 (en) | 2019-06-01 | 2022-06-14 | Apple Inc. | Increasing the relevance of new available information |
WO2021056255A1 (en) | 2019-09-25 | 2021-04-01 | Apple Inc. | Text detection using global geometry estimators |
US12081632B2 (en) * | 2021-12-23 | 2024-09-03 | Apple Inc. | Content customization and presentation based on user presence and identification |
Citations (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20030164268A1 (en) | 2002-03-01 | 2003-09-04 | Thomas Meyer | Procedures, system and computer program product for the presentation of multimedia contents in elevator installations |
US20030222134A1 (en) | 2001-02-17 | 2003-12-04 | Boyd John E | Electronic advertising device and method of using the same |
US6916244B2 (en) * | 2002-06-05 | 2005-07-12 | Cyberscan Technology, Inc. | Server-less cashless gaming systems and methods |
US20060106726A1 (en) * | 2004-11-18 | 2006-05-18 | Contentguard Holdings, Inc. | Method, system, and device for license-centric content consumption |
US20070024580A1 (en) | 2005-07-29 | 2007-02-01 | Microsoft Corporation | Interactive display device, such as in context-aware environments |
US7346687B2 (en) * | 1999-10-05 | 2008-03-18 | Zapmedia Services, Inc. | GUI driving media playback device |
Family Cites Families (11)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US4931865A (en) * | 1988-08-24 | 1990-06-05 | Sebastiano Scarampi | Apparatus and methods for monitoring television viewers |
US5550928A (en) * | 1992-12-15 | 1996-08-27 | A.C. Nielsen Company | Audience measurement system and method |
US7107605B2 (en) * | 2000-09-19 | 2006-09-12 | Simple Devices | Digital image frame and method for using the same |
US20020194586A1 (en) * | 2001-06-15 | 2002-12-19 | Srinivas Gutta | Method and system and article of manufacture for multi-user profile generation |
US20030237093A1 (en) * | 2002-06-19 | 2003-12-25 | Marsh David J. | Electronic program guide systems and methods for handling multiple users |
US20040194128A1 (en) * | 2003-03-28 | 2004-09-30 | Eastman Kodak Company | Method for providing digital cinema content based upon audience metrics |
US7490340B2 (en) * | 2003-04-21 | 2009-02-10 | International Business Machines Corporation | Selectively de-scrambling media signals |
US7564994B1 (en) * | 2004-01-22 | 2009-07-21 | Fotonation Vision Limited | Classification system for consumer digital images using automatic workflow and face detection and recognition |
GB0618266D0 (en) * | 2006-09-18 | 2006-10-25 | Dosanjh Harkamaljit | Mobile devices and systems for using them |
US20090037949A1 (en) * | 2007-02-22 | 2009-02-05 | Birch James R | Integrated and synchronized cross platform delivery system |
US8285006B2 (en) * | 2007-04-13 | 2012-10-09 | Mira Electronics Co., Ltd. | Human face recognition and user interface system for digital camera and video camera |
-
2008
- 2008-09-14 US US12/210,199 patent/US20100071003A1/en not_active Abandoned
-
2009
- 2009-09-13 WO PCT/IL2009/000895 patent/WO2010029557A1/en active Application Filing
- 2009-09-13 EP EP09812786.3A patent/EP2377030A4/en not_active Withdrawn
Patent Citations (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US7346687B2 (en) * | 1999-10-05 | 2008-03-18 | Zapmedia Services, Inc. | GUI driving media playback device |
US20030222134A1 (en) | 2001-02-17 | 2003-12-04 | Boyd John E | Electronic advertising device and method of using the same |
US20030164268A1 (en) | 2002-03-01 | 2003-09-04 | Thomas Meyer | Procedures, system and computer program product for the presentation of multimedia contents in elevator installations |
US6916244B2 (en) * | 2002-06-05 | 2005-07-12 | Cyberscan Technology, Inc. | Server-less cashless gaming systems and methods |
US20060106726A1 (en) * | 2004-11-18 | 2006-05-18 | Contentguard Holdings, Inc. | Method, system, and device for license-centric content consumption |
US20070024580A1 (en) | 2005-07-29 | 2007-02-01 | Microsoft Corporation | Interactive display device, such as in context-aware environments |
Non-Patent Citations (2)
Title |
---|
RHODES B J ET AL.: "WEARABLE COMPUTERS, 1999. DIGEST OF PAPERS. THE THIRD INTERNATIONAL SY MPOSIUM ON SAN FRANCISCO", 18 October 1999, IEEE COMPUT. SOC, article "Wearable computing meets ubiquitous computing: reaping the best of both worlds", pages: 141 - 149 |
See also references of EP2377030A4 |
Cited By (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US9542083B2 (en) | 2014-12-04 | 2017-01-10 | Comcast Cable Communications, Llc | Configuration responsive to a device |
Also Published As
Publication number | Publication date |
---|---|
EP2377030A4 (en) | 2013-09-25 |
US20100071003A1 (en) | 2010-03-18 |
EP2377030A1 (en) | 2011-10-19 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US20100071003A1 (en) | Content personalization | |
US9788066B2 (en) | Information processing apparatus, information processing method, computer program, and information sharing system | |
US9241195B2 (en) | Searching recorded or viewed content | |
KR100424848B1 (en) | Television receiver | |
US8745024B2 (en) | Techniques for enhancing content | |
US9230151B2 (en) | Method, apparatus, and system for searching for image and image-related information using a fingerprint of a captured image | |
EP3528199B1 (en) | Method and apparatus for collecting content | |
US20080146274A1 (en) | Method and apparatus for storing image file in mobile terminal | |
WO2014206147A1 (en) | Method and device for recommending multimedia resource | |
US8943020B2 (en) | Techniques for intelligent media show across multiple devices | |
JP2006236218A (en) | Electronic album display system, electronic album display method, and electronic album display program | |
WO2007029918A1 (en) | Method and apparatus for encoding multimedia contents and method and system for applying encoded multimedia contents | |
JP2010272077A (en) | Method and device for reproducing information | |
JP6046393B2 (en) | Information processing apparatus, information processing system, information processing method, and recording medium | |
CN110879944A (en) | Anchor recommendation method, storage medium, equipment and system based on face similarity | |
JP4723901B2 (en) | Television display device | |
KR101351818B1 (en) | A method for playback of contents appropriate to context of mobile communication terminal | |
WO2013124528A1 (en) | Method, apparatus and computer program product for management of media files | |
KR20150108562A (en) | Image processing apparatus, control method thereof and computer readable medium having computer program recorded therefor | |
EP1935183A1 (en) | Method and apparatus for encoding multimedia contents and method and system for applying encoded multimedia contents | |
US9451321B2 (en) | Content management with biometric feature recognition | |
US20140189769A1 (en) | Information management device, server, and control method | |
US20190095468A1 (en) | Method and system for identifying an individual in a digital image displayed on a screen | |
US20170208358A1 (en) | Device for and method of tv streaming and downloading for personal photos and videos presentation on tv that seamlessly integrates with mobile application and cloud media server | |
EP1569448A1 (en) | Image description system and method thereof |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
121 | Ep: the epo has been informed by wipo that ep was designated in this application |
Ref document number: 09812786 Country of ref document: EP Kind code of ref document: A1 |
|
NENP | Non-entry into the national phase |
Ref country code: DE |
|
REEP | Request for entry into the european phase |
Ref document number: 2009812786 Country of ref document: EP |
|
WWE | Wipo information: entry into national phase |
Ref document number: 2009812786 Country of ref document: EP |