WO2013045123A1 - Personalised augmented a/v stream creation - Google Patents

Personalised augmented a/v stream creation Download PDF

Info

Publication number
WO2013045123A1
WO2013045123A1 PCT/EP2012/061655 EP2012061655W WO2013045123A1 WO 2013045123 A1 WO2013045123 A1 WO 2013045123A1 EP 2012061655 W EP2012061655 W EP 2012061655W WO 2013045123 A1 WO2013045123 A1 WO 2013045123A1
Authority
WO
WIPO (PCT)
Prior art keywords
stream
data
personalised
user
data streams
Prior art date
Application number
PCT/EP2012/061655
Other languages
French (fr)
Inventor
Daniel Edward WOULD
Matthew Whitbourne
Katherine Marie SHANN
Shanna Xu
Christopher Phillips
Original Assignee
International Business Machines Corporation
Ibm United Kingdom Limited
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by International Business Machines Corporation, Ibm United Kingdom Limited filed Critical International Business Machines Corporation
Priority to US14/346,899 priority Critical patent/US9332313B2/en
Publication of WO2013045123A1 publication Critical patent/WO2013045123A1/en

Links

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/45Management operations performed by the client for facilitating the reception of or the interaction with the content or administrating data related to the end-user or to the client device itself, e.g. learning user preferences for recommending movies, resolving scheduling conflicts
    • H04N21/462Content or additional data management, e.g. creating a master electronic program guide from data received from the Internet and a Head-end, controlling the complexity of a video stream by scaling the resolution or bit-rate based on the client capabilities
    • H04N21/4622Retrieving content or additional data from different sources, e.g. from a broadcast channel and the Internet
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/40Information retrieval; Database structures therefor; File system structures therefor of multimedia data, e.g. slideshows comprising image and additional audio data
    • G06F16/41Indexing; Data structures therefor; Storage structures
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/40Information retrieval; Database structures therefor; File system structures therefor of multimedia data, e.g. slideshows comprising image and additional audio data
    • G06F16/43Querying
    • G06F16/435Filtering based on additional data, e.g. user or group profiles
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/40Information retrieval; Database structures therefor; File system structures therefor of multimedia data, e.g. slideshows comprising image and additional audio data
    • G06F16/48Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/40Information retrieval; Database structures therefor; File system structures therefor of multimedia data, e.g. slideshows comprising image and additional audio data
    • G06F16/48Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually
    • G06F16/489Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually using time information
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/70Information retrieval; Database structures therefor; File system structures therefor of video data
    • G06F16/78Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/20Servers specifically adapted for the distribution of content, e.g. VOD servers; Operations thereof
    • H04N21/23Processing of content or additional data; Elementary server operations; Server middleware
    • H04N21/235Processing of additional data, e.g. scrambling of additional data or processing content descriptors
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/20Servers specifically adapted for the distribution of content, e.g. VOD servers; Operations thereof
    • H04N21/25Management operations performed by the server for facilitating the content distribution or administrating data related to end-users or client devices, e.g. end-user or client device authentication, learning user preferences for recommending movies
    • H04N21/266Channel or content management, e.g. generation and management of keys and entitlement messages in a conditional access system, merging a VOD unicast channel into a multicast channel
    • H04N21/2668Creating a channel for a dedicated end-user group, e.g. insertion of targeted commercials based on end-user profiles
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/45Management operations performed by the client for facilitating the reception of or the interaction with the content or administrating data related to the end-user or to the client device itself, e.g. learning user preferences for recommending movies, resolving scheduling conflicts
    • H04N21/462Content or additional data management, e.g. creating a master electronic program guide from data received from the Internet and a Head-end, controlling the complexity of a video stream by scaling the resolution or bit-rate based on the client capabilities
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/47End-user applications
    • H04N21/478Supplemental services, e.g. displaying phone caller identification, shopping application
    • H04N21/4782Web browsing, e.g. WebTV
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/47End-user applications
    • H04N21/478Supplemental services, e.g. displaying phone caller identification, shopping application
    • H04N21/4788Supplemental services, e.g. displaying phone caller identification, shopping application communicating with other users, e.g. chatting
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/80Generation or processing of content or additional data by content creator independently of the distribution process; Content per se
    • H04N21/83Generation or processing of protective or descriptive data associated with content; Content structuring
    • H04N21/84Generation or processing of descriptive data, e.g. content descriptors

Landscapes

  • Engineering & Computer Science (AREA)
  • Multimedia (AREA)
  • Databases & Information Systems (AREA)
  • Theoretical Computer Science (AREA)
  • Signal Processing (AREA)
  • General Engineering & Computer Science (AREA)
  • Data Mining & Analysis (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Library & Information Science (AREA)
  • Software Systems (AREA)
  • Television Signal Processing For Recording (AREA)
  • Two-Way Televisions, Distribution Of Moving Picture Or The Like (AREA)

Abstract

A method and apparatus for creating a personalised audio and/or video (AV) stream for a user, comprising selecting a first stream of AV data (S.1) and one or more further data streams (S.3 – S.5) having a predetermined relation to the first stream. Monitoring (3) and recording (2) means identify and capture portions of the first stream by reference to identifying predetermined indicia in one or more of the streams, together with capturing one or more time stamps (T.0) to form a master timeline (MTL 2a). The personalised AV stream (1) comprises one or more captured portions of the first stream (S.1) packaged with selected data from the or each further stream.

Description

PERSONALISED AUGMENTED A/V STREAM CREATION
DESCRIPTION
FIELD OF THE INVENTION
The present invention relates generally to the field of audio and/or video (hereinafter "AV") data stream capture and processing. More particularly, the present invention relates to the identification and handling of AV stream components by reference to, or in conjunction with, ancillary data from one or more further data streams linked to the AV data stream, especially social media streams.
BACKGROUND ART
Recent years have seen a huge growth in the area of web-based and mobile technologies referred to generally as social media by which users can form online or networked groups for the exchange of comments and creation and sharing of media comment such as AV clips. Well-known examples include Facebook, flickr, YouTube and Twitter.
A more recent development is Social TV which has attempted to integrate television broadcasts with social media feeds to provide greater interactivity through enabling viewers to post comments as the broadcast happens: an example of this from the UK is the political debate programme "Question Time" which hosts a live Twitter feed as the broadcast is aired to allow viewers to join the debate.
As part of the shared experience, users may wish to identify (tag) individual segments of a broadcast as being of potential interest to others, with a resulting stream of tagged portions creating a user-defined highlights package of the broadcast One method to enable such tagging is described in United States Patent Application 2009/0228492A1 (Valdez et al) wherein, in response to a user request to mark a portion of media content, a tag is inserted in the stream which tag may subsequently be used to access the tagged portion of content. A problem arises when it is desired not only to capture AV segments, but also the related social media data (comments, links to related content and so forth). Devices such as a personal video recorder (PVR) are commonly used to record TV programs. This allows a user to watch a football match, for instance, at a later time and skip through adverts. This is not very sociable in the sense that the user misses out on discussions/highlights that are posted on the social media at the time when the event is broadcasting live. The user could surf related websites and watch highlights/discussions before or after watching the recorded match. However this is time consuming and the user may end up watching parts of the match many times. Sometimes it also spoils the fun, if the user wanted to avoid but then accidentally saw the results. There is no currently satisfactory mechanism to embed the comments/highlights into a recorded program.
A problem that must be solved is how to correctly associate the comments in the social media stream with the moment in the recording that they are associated with. This problem arises because people watching a program 'live' are not all necessarily watching the program at the same time. Viewers experience different levels of lag depending on whether they are watching the program through normal terrestrial TV, a satellite broadcast, or an internet streaming service (for example the BBC's iPlayer service). Comments may also arrive significantly after a given event, depending on the user's behaviour.
It is therefore an object of the present invention to provide a method and supporting system to improve the integration of AV streams with data from associated streams, such as social media feeds. SUMMARY OF THE INVENTION
In accordance with a first aspect of the present invention there is provided a method for creating a personalised audio and/or video (AV) stream for a user, comprising:
- selecting a first stream of AV data by reference to predetermined user preferences; - selecting one or more further data streams having a predetermined relation to the first stream; - identifying and capturing portions of the first stream by reference to identifying predetermined indicia in said first stream or one of said one or more further data streams, including capturing one or more time stamps from the first or further stream; and
- generating a personalised AV stream comprising a captured portion of the first stream and data selected by reference to said predetermined indicia from at least one of said one or more further data streams, said selection being based on the or each captured time stamp. By reference to the captured time stamp, the data from the ancillary data streams can be linked to the captured AV content such that, for example, later posted comments on a social media feed can be matched up with the AV content that lead to their creation.
Suitably, the method includes defining a master timeline by reference to a time signal containing a sequence of timestamps in the first stream or a selected one of the one or more further data streams. As will be described in further detail by reference to potential embodiments of the invention, the source of the master timeline may comprise a public timeline, a master component stream, or a subtitle track accompanying the AV stream.
The timestamp at which a predetermined indica is matched may be used to determine the extent of the portion of the first stream that is captured. For example, a digital recording device may be continuously capturing the AV content, and the captured portion added to the personalised AV stream may comprise a certain number of minutes before and/or after the time stamp.
The user may be enabled to specify an extent for the period of capture for at least one of the first stream and the or each of the other data streams, and user-operable means may be provided to indicate the start and/or end of a portion of the first stream to be captured. In this way, a user may construct their own highlights package from an AV stream, with the captured supporting data (comments for example) just relating to those highlights identified by the user. Playback suitably comprises presenting the personalised AV stream to a user via an audio and video playback device, including presenting on a part of the screen the data captured from the or each further data stream. The screen layout is an important consideration: the ancillary data should be clearly readable but not intrusive. Past experiments with live comment feeds being pasted on top of AV content have shown this arrangement to be unpopular with users.
Where the first stream includes an audio component, the predetermined indicia may be a predetermined audio pattern in the first stream, for example individual keywords or a sudden increase in volume (perhaps indicative of a goal being scored). If the first (AV) stream is a video data stream including an associated subtitle data stream, the master timeline may be defined by timestamps in the subtitle date stream. Furthermore, particular text streams within the subtitle data stream (such as particular keywords or phrases) may provide the predetermined indicia.
In embodiments where the triggering for identification and capture is based on the ancillary data (rather than the AV stream), the predetermined indicia may be a predetermined keyword or combination of keywords in the one or more further data streams. As an example, if a contributor on a social media stream were to enter the phrase "CATCH THIS", the previous few minutes of AV would be captured and linked to the contributor comment for inclusion in the personalised stream.
In order to avoid so-called "spoilers", the method may further comprise the identification of a second predetermined indicia or group of indicia, whereby the identification of such causes exclusion from the personalised AV stream. For example, where the personalised stream comprises the highlights of a sporting event, any detection of a phrase such as "the final score" may trigger exclusion to avoid the user being inadvertently notified of the outcome before it is reached in the AV stream. As will be described below with reference to an embodiment of the invention, use of indicia to block comments may be applied to a users general browser so that they may continue to browse social media streams whilst the recording and capture is ongoing whilst reducing the risk of spoilers associated with the recorded content.
In the case of particularly popular items of AV content, the volume of ancillary data may become excessive. To address this, the user may suitably be provided with means to restrict the volume of captured data. This may comprise enabling user selection of the one or more further data streams from a list of available data streams (such as to choose only the comment streams of interest), or a set of pre-stored user preferences may be used to select one or more of the one or more further data streams for contribution to the personalised AV stream, for example only accepting comments from a "friends" list, perhaps triggered by the exceeding of a predetermined threshold on the volume of ancillary data to be accepted and added to the personalised AV stream.
The present invention further provides an apparatus operable to create a personalised audio and/or video (AV) stream for a user, comprising:
- a data recorder connectable to, and operable to record data from, a first stream of AV data and one or more further data streams having a predetermined relation to the first stream;
- monitoring means coupled with the data recorder and configured to identify predetermined indicia in said first stream or one of said one or more further data streams and, in response thereto, cause the data recorder to capture a portion of the first stream, including capturing one or more time stamps from the first or further stream;
- wherein the data recorder is further operable to generate and play back a personalised AV stream comprising a captured portion of the first stream and data selected by reference to said predetermined indicia from at least one of said one or more further data streams, said selection being based on the or each captured time stamp. The data recorder may be further operable to create and store a master timeline by reference to a time signal containing a sequence of timestamps in said first stream or a selected one of said one or more further data streams.
The invention further provides a computer program stored on a computer readable medium and loadable into the internal memory of a digital computer, comprising software code portions, when said program is run on a computer, for performing the method according to the invention and as described above.
The summary of the present invention does not recite all the necessary features of the invention, and sub-combinations of those features may also encompass the invention. BRIEF DESCRIPTION OF THE DRAWINGS
The present invention will now be described, by way of example only, with reference to preferred embodiments, as illustrated in the following figures, in which:
Figure 1 schematically represents a series of data streams and functional components of a system for personalised AV stream creation;
Figure 2 shows in greater detail components of an apparatus suitable for use in the system of Figure 1;
Figure 3 represents the presentation of a created personalised AV stream on a personal computer;
Figure 4 is a flow chart showing steps in a method for creation of a personalised AV stream embodying aspects of the present invention; and
Figure 5 schematically represents an additional feature linking a system such as that if Figure 1 to a general purpose browser.
Figure 1 schematically represents the components of an apparatus operable to create a personalised audio and/or video stream (PAVS) 1 for a user, comprising a data recorder 2 connectable to, and operable to record data from, a first stream of AV data S.1 and one or more further data streams S.3 - S.5 having a predetermined relation to the first stream. In this embodiment, the first stream S.l is a television broadcast, and further streams S.3 - S.5 are social media streams carrying comments relating to the television broadcast. A further stream S.2 provides a reference time signal source. As will be described below, this reference timeline may be derived from the AV stream or one of the comment streams.
The apparatus includes a monitoring stage 3 coupled with the data recorder 2 and further coupled 8 to access the AV and comment streams. The monitoring stage 3 is configured to identify predetermined indicia (as will be described below) in the AV stream S.1 or one of the comment streams S.3 - S.5 and, in response thereto, cause the data recorder 2 to capture a portion 100 of the AV stream S.l . Capture of the AV portion 100 further includes capturing at least one timestamp T.O from the reference timeline S.2, and subsequent capture of portions 102, 104, 106, 108 of the comment streams S.3 - S.5, which capture is generally indicated in the Figure by the reference numeral 9. The indicia to be used by the monitoring stage 3 are stored in associated memory 4, and are suitably provided by user-operable input means 7. The indicia may be simple keywords to be identified in one of the comment streams S.3 - S.5 or in the soundtrack or an associated subtitle stream (not shown) of the AV stream S.1. Also, a threshold test may be
implemented, requiring two or more detected instances of an indicia before capture is triggered. Alternately, the user input may comprise direct "start capture" and "stop capture" commands which trigger the recording stage 2 to start/stop recording from the AV stream. Suitably, the recording stage is a digital AV recorder running on a loop (e.g. perpetually recording the last 5 minutes from the AV stream) such that a "start capture" command will include, say, the previous 60 seconds of recorded AV: this will allow a user who, for example, spots himself in the audience of a TV broadcast, to capture that appearance in portion 100 provided that the "start capture" command is issued within 60 seconds of the appearance. Preferentially, the captured timestamp T.O is that coinciding with the start of the captured portion, rather than the instant at which the user command is input.
The data recorder is further operable to generate (by means of collation stage 5) and play back (stage 6) the personalised AV stream 1 comprising a captured portion of the AV stream S.l, combined with the selected and captured data from the comment streams. As will be described below, the selection from the comment streams is based (at least partly) on the or each captured time stamp.
In operation, the apparatus will typically create a sequence of captured AV segments 100 and associated comment in the personalised stream 1. In order to tie these together in a logical sequence, the data recorder 2 creates and stores a master timeline MTL 2a by reference to the time signal S.2 containing the sequence of timestamps captured in conjunction with the AV segment.
Figure 2 schematically represents the components of a computer system suitable to embody the present invention. A first processor CPU 10 is coupled with random access memory RAM 12 and read only memory ROM 14 by an address and data bus 16. As will be understood, CPU 10 may comprise a cluster of processors (nodes) with individual processes and/or process threads being handled by individual nodes. Also connected to CPU 10 via the address and data bus 16 is at least one further processor 42 (or cluster of nodes), which may be a further CPU sharing tasks with the first CPU 10, or may be a coprocessor device supplementing the function of the CPU 10, handling processes such as floating point arithmetic, graphics processing, signal processing and encryption. Each of these internal hardware devices 10, 12, 14, 42 includes a respective interface (not shown) supporting connection to the bus 16. These interfaces are conventional in form and need not be described in further detail
Also connected to the CPU 10 via bus 16 are a number of external hardware device interface stages (generally denoted 18). A first interface stage 20 supports the connection of external input/output devices, such as a mouse 22 and/or keyboard 24 (equivalent to input stage 7 in Figure 1 through which the user enters indicia and/or start/stop commands). A second interface stage 26 supports the connection of external output devices such as a display screen 28 (and/or audio output device 30, such as headphones or speakers) through which the PAVS 1 may be presented to the user. A third interface stage 32 supports the connection to external data storage devices in the form of computer readable media: such external storage may as shown be provided by a removable optical or magnetic disc 34 (accessed by a suitably configured disc reader 36). Alternatively or additionally the external storage may be in the form of a solid state memory device such as an extension drive or memory stick. The external storage may contain a computer program, containing program software code portions which, when run by the CPU 10 and/or further processor 42, perform the method according to the present invention.
A fourth interface stage 38 supports connection of the system to remote devices or systems via wired or wireless networks 40, for example over a local area network LAN, via the internet, or another cloud computing source. The network 40 will provide the source for the streams S. l - S.5 of Figure 1, with the CPU 10 and interface 38 providing the monitoring means 3, and the local 12 and offline 32, 34, 36 storage providing the recording stage 2.
It will be readily understood that there are many physical forms that the apparatus may take, such as an interlinked combination of a television, DVR (digital video recorder) or PVR (personal video recorder), and personal computer. Another possible arrangement is shown in Figure 3, in which all of the functional components are provided within a personal computer 200, either coupled to an external AV feed (not shown) or accessing AV content over the internet. Suitably the screen space is divided into separate windows to display the AV content 202, to display the captured comments 204, and to provide an instruction menu space 206 (perhaps including clickable buttons for the start/stop capture options described above, a field for entry of specific indicia, means for inputting default selections such as the duration of capture, and so forth). As will be described below, the window 206 may provide a menu list of comment streams associated with a particular AV stream, such that a user can make a selection of the sources of comments to be included in the PAVS.
The basic steps of a method embodying various aspects of the invention will now be described with reference to the flow chart of Figure 4. The process starts at step 402 with the selection of a first stream of AV data by reference to predetermined user preferences. This may comprise an analysis of a users past viewing history to identify a suitable AV stream of potential interest to a user, or it may be as simple as a user directly inputting 428 a selection (e.g. choosing a television programme).
The next step 404 is selecting one or more further data streams (ancillary sources) having a predetermined relation to the first stream, for example social media feeds directly associated with the selected AV stream. As before, this selection made be made by reference to user input 428, perhaps to select the feeds to be monitored and optionally, if there are likely to be an excessive number of posted comments, to limit the captured comments.
The issue of scalability arises when, for example, there are a large number of people in a users "friends" stream all commenting on the programme. The number of comments on an event may exceed the ability of the embedding system to display them in the space and time appropriate. One preferred solution to this is to use existing algorithms (such as those used by a Facebook friends stream) to prioritise which friends a user wishes to hear from. For example, selection criteria may include how often a user interacts with them, or by whether their comments are part of a conversation or stand-alone. At step 406, the indicia determining capture are specified, again potentially as a result of direct user input 428. As outlined above, these indicia may comprise specific keywords or groups of keywords in one of the comment streams or a subtitle stream accompanying the AV stream, or may comprise sudden changes of volume in the audio component of the AV stream. Where a speech-to-text utility is incorporated, then spoken keywords in the audio component of the AV stream may also be used as indicia. Additionally, a second indicia or group of indicia may be specified that trigger exclusion from the PAVS, thereby allowing a user to refine the selection further.
At step 408 the process of monitoring the selected stream or streams commences, with step 410 (looping back to 408 in the case of a negative result) determining when a predetermined indicia is matched. As an optional step 412, the number of matches is compared to a pre- stored threshold level: if the threshold has not been reached, the number of detected instances of the indicia is incremented by 1 at step 414 and the monitoring continues from step 408.
In the absence of a threshold check, or if the threshold has been reached, the process moves to step 416 at which a timestamp for the match is captured (e.g. from timeline S.2), following which a portion of the AV stream is captured at step 418. As indicated by dashed line 426, the capture may be by reference to user input 428, whether in terms of a specified duration of AV stream to capture or, effectively bypassing the monitoring process, through the user input of a direct "capture" command.
At step 420, data from the ancillary (comment) streams is captured and stored with the AV data. Through the timestamp captured at 416, input on comment streams may be tied to an AV stream event, even if the comment is posted at a later time so that, although the monitoring stage may stop synchronizing with the recorder for AV content at the end of a broadcast, it may continue search on social media for a pre-defined length of time.
At step 422, the personalised AV stream is created, bringing together in sequence each captured AV segment with the comments associated with that segment. In order to facilitate the sequencing and playback, a master timeline is suitably generated (step 424) by reference to the captured timestamps from the selected stream S.2, thereby enabling a user to subsequently switch on the PVR (or other form of apparatus used for the recording) and watch the highlights or the entire video stream with tagged comments appended to the appropriate points of the AV stream.
As identified above, a particular feature is to identify a reference stream which provides the source of the timestamps and which defines the times at which events have occurred. This reference stream could be one of the following:
a) The public timeline. Here the system uses a comment stream (for example containing friends' updates) to retrieve the content which is embedded. It also monitors a public timeline for comments which are related, by hashtag or by content, to those of friends.
The generally large number of people in the sample will give a smooth distribution with a peak at the moment when most people were commenting on the event. This suitably defines the timestamp that the comments embedded from friends are keyed from.
b) A master comment stream. Some programmes have an associated Twitter account that tweets along with the programme, or a live commentary in the case of any live events.
The timing defined by this stream could be used as the reference.
c) An already embedded subtitles track. Most programmes on broadcast television, even those which are live, contain an embedded subtitles track. The text in this track can be compared to the content of the comments on the social media stream to judge the point at which an event occurred, (for example "GOAL!", "You're fired!").
As illustrated in Figure 5, the aspects of the foregoing embodiments relating to the selection of appropriate comments for later embedding in an item of replayed media, also allow the extraction or filtering of comments from a social media feed to happen. As represented at
502 a user may initiate the content plus comment capture process but then desire to continue browsing the social media (at 504) whilst the recording is ongoing. Those indicia selected for capture of ancillary data at 506 in the recording process may also be ported to the users general browser at 508 to block user access to comments matching those indicia. This will allow a user to continue to browse their social feeds with the removal of comments that could act as spoilers for the recorded media, by using the same mechanism as described previously. The block may be directly cancelled by a user or it may be linked to the recording system such that only when playback has occurred at 510 does the system issue a release command at 512 which unblocks the general browser at 514.
As comment threads stemming from the original event continue they are likely to diverge from the original subject matter. The content of the comments can be compared with the content of the master stream. Through this comparison it can be judged whether the comment in the thread is relevant to the initial event, irrelevant, or relevant to a new event. Based on this comparison one of the following actions could be performed on the comments: a) The comments can be embedded at the time they occur relative to the comment that started the thread.
b) The comments can be not embedded.
c) The comments can be embedded at the time related to the new event, and be considered to begin a new thread.
In terms of deciding at what point the embedding process occurs, three possible options are: a) The embedding happens as the programme is being recorded. Because of the time it takes people to react to an event and process their thoughts into what they want to enter as a comment, this embedding would not happen immediately, but would more likely be on a 5 minute delay. For example, if a goal is scored in a football match the viewer would possibly first celebrate with the people in the room, have a drink, and then watch the replay before making a comment on social media. The advantage of this method is that the person who wishes to watch the programme with the embedded comment stream could watch it on a short time lag while the programme is still on.
b) The embedding happens after the programme has finished, possibly at the point the user requests playback. After the programme is over the social media stream and the reference stream are compared and the comments are embedded to create the master timeline and the PAVS as at step 422 in Figure 4. The advantage of this is that there may be a lot of post-programme analysis and discussion. This discussion can also be included in the embedded comments, taking place over the credits or after the programme has finished. c) Both, the embedding happens as the programme goes along, but is further refined by comments that occur after the program is finished. This provides a recorded programme with embedded comments which can be watched on a small lag from the live event, but if the person recording the programme chooses to watch the recording much later then the PVR can continue to monitor the social media stream and additional comments from the social media stream can be added in at the appropriate point - including comments from other viewers who may have time-shifted the programme.
Whilst embodiments of the present invention have been described above, the technical scope of the invention is not limited to the scope of the above-described embodiments. It should be apparent to those skilled in the art that various changes or improvements can be made to the embodiments. It is apparent from the description of the appended claims that implementations including such changes or improvements are encompassed in the technical scope of the invention.

Claims

1. A method for creating a personalised audio and/or video (AV) stream for a user, comprising:
- selecting a first stream of AV data by reference to predetermined user preferences;
- selecting one or more further data streams having a predetermined relation to the first stream;
- identifying and capturing portions of the first stream by reference to identifying predetermined indicia in said first stream or one of said one or more further data streams, including capturing one or more time stamps from the first or further stream; and
- generating a personalised AV stream comprising a captured portion of the first stream and data selected by reference to said predetermined indicia from at least one of said one or more further data streams, said selection being based on the or each captured time stamp.
2. A method as claimed in Claim 1, further comprising the step of defining a master timeline by reference to a time signal containing a sequence of timestamps in said first stream or a selected one of said one or more further data streams.
3. A method as claimed in Claim 2, wherein the timestamp at which a predetermined indicia is matched is used to determine the extent of the portion of the first stream that is captured.
4. A method as claimed in Claim 2, wherein the user is enabled to specify an extent for the period of capture for at least one of the first stream and the or each of the other data streams.
5. A method as claimed in Claim 1, further comprising presenting the personalised AV stream to a user via an audio and video playback device, including presenting on a part of the screen the data captured from the or each further data stream.
6. A method as claimed in Claim 1, wherein the first stream includes an audio component, and said predetermined indicia is a predetermined audio pattern in said first stream.
7. A method as claimed in Claim 2, wherein the first stream is a video data stream including a subtitle data stream, wherein the master timeline is defined by the subtitle date stream, and text streams within the subtitle data stream provide the predetermined indicia.
8. A method as claimed in Claim 1, wherein the predetermined indicia is a
predetermined keyword or combination of keywords in said one or more further data streams.
9. A method as claimed in Claim 1, further comprising the identification of a second predetermined indicia or group of indicia, whereby the identification of such causes exclusion from the personalised AV stream.
10. A method as claimed in Claim 1, further comprising providing user-operable means to indicate the start and/or end of a portion of the first stream to be captured.
11. A method as claimed in Claim 1 , further comprising enabling user selection of said one or more further data streams from a list of available data streams.
12. A method as claimed in Claim 1, wherein a set of pre-stored user preferences are used to select one or more of said one or more further data streams for contribution to said personalised AV stream.
13. Apparatus operable to create a personalised audio and/or video (AV) stream for a user, comprising:
- a data recorder connectable to, and operable to record data from, a first stream of AV data and one or more further data streams having a predetermined relation to the first stream; - monitoring means coupled with the data recorder and configured to identify predetermined indicia in said first stream or one of said one or more further data streams and, in response thereto, cause the data recorder to capture a portion of the first stream, including capturing one or more time stamps from the first or further stream;
- wherein the data recorder is further operable to generate and play back a personalised AV stream comprising a captured portion of the first stream and data selected by reference to said predetermined indicia from at least one of said one or more further data streams, said selection being based on the or each captured time stamp.
14. Apparatus as claimed in Claim 13, wherein the data recorder is further operable to create and store a master timeline by reference to a time signal containing a sequence of timestamps in said first stream or a selected one of said one or more further data streams.
15. A computer program stored on a computer readable medium and loadable into the internal memory of a digital computer, comprising software code portions, when said program is run on a computer, for performing the method of any of claims 1 to 12.
PCT/EP2012/061655 2011-09-28 2012-06-19 Personalised augmented a/v stream creation WO2013045123A1 (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
US14/346,899 US9332313B2 (en) 2011-09-28 2012-06-19 Personalized augmented A/V stream creation

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
EP11183186.3 2011-09-28
EP11183186 2011-09-28

Publications (1)

Publication Number Publication Date
WO2013045123A1 true WO2013045123A1 (en) 2013-04-04

Family

ID=46354289

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/EP2012/061655 WO2013045123A1 (en) 2011-09-28 2012-06-19 Personalised augmented a/v stream creation

Country Status (2)

Country Link
US (1) US9332313B2 (en)
WO (1) WO2013045123A1 (en)

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
EP2811752A1 (en) * 2013-06-03 2014-12-10 Alcatel Lucent Synchronization between multimedia flows and social network threads
WO2017204836A1 (en) * 2016-05-27 2017-11-30 Thomson Licensing Method and apparatus for personal multimedia content distribution

Families Citing this family (12)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US10410222B2 (en) 2009-07-23 2019-09-10 DISH Technologies L.L.C. Messaging service for providing updates for multimedia content of a live event delivered over the internet
EP3340575A1 (en) 2011-12-06 2018-06-27 EchoStar Technologies L.L.C. Remote storage digital video recorder and related operating methods
US9237386B2 (en) * 2012-08-31 2016-01-12 Google Inc. Aiding discovery of program content by providing deeplinks into most interesting moments via social media
US10971191B2 (en) * 2012-12-12 2021-04-06 Smule, Inc. Coordinated audiovisual montage from selected crowd-sourced content with alignment to audio baseline
US10051025B2 (en) 2012-12-31 2018-08-14 DISH Technologies L.L.C. Method and apparatus for estimating packet loss
US10104141B2 (en) 2012-12-31 2018-10-16 DISH Technologies L.L.C. Methods and apparatus for proactive multi-path routing
US10708319B2 (en) * 2012-12-31 2020-07-07 Dish Technologies Llc Methods and apparatus for providing social viewing of media content
US9401947B1 (en) 2013-02-08 2016-07-26 Google Inc. Methods, systems, and media for presenting comments based on correlation with content
US10986063B2 (en) * 2013-02-27 2021-04-20 Comcast Cable Communications, Llc Methods and systems for providing supplemental data
US10057651B1 (en) * 2015-10-05 2018-08-21 Twitter, Inc. Video clip creation using social media
WO2017117264A1 (en) 2015-12-29 2017-07-06 Echostar Technologies L.L.C Remote storage digital video recorder streaming and related methods
US10771848B1 (en) * 2019-01-07 2020-09-08 Alphonso Inc. Actionable contents of interest

Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20060149781A1 (en) * 2004-12-30 2006-07-06 Massachusetts Institute Of Technology Techniques for relating arbitrary metadata to media files
US20090228492A1 (en) 2008-03-10 2009-09-10 Verizon Data Services Inc. Apparatus, system, and method for tagging media content
US20110058101A1 (en) * 2009-09-03 2011-03-10 Fox Broadcasting Company Method and apparatus for concurrent broadcast of media program and social networking derived information exchange
US20110225178A1 (en) * 2010-03-11 2011-09-15 Apple Inc. Automatic discovery of metadata
US20110221962A1 (en) * 2010-03-10 2011-09-15 Microsoft Corporation Augmented reality via a secondary channel

Family Cites Families (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
AU1782001A (en) 1999-11-24 2001-06-04 Visual Data Corporation Method for manipulating a live audio media stream
WO2006078365A1 (en) 2005-01-21 2006-07-27 Thomson Licensing Method and apparatus for skipping commercials
US20090162024A1 (en) 2007-12-20 2009-06-25 Geevee, Inc. Intra-Media Demarcation
US9788043B2 (en) 2008-11-07 2017-10-10 Digimarc Corporation Content interaction methods and systems employing portable devices
US20100153984A1 (en) 2008-12-12 2010-06-17 Microsoft Corporation User Feedback Based Highlights of Recorded Programs
US20100318520A1 (en) * 2009-06-01 2010-12-16 Telecordia Technologies, Inc. System and method for processing commentary that is related to content
US8504373B2 (en) 2009-07-02 2013-08-06 Nuance Communications, Inc. Processing verbal feedback and updating digital video recorder (DVR) recording patterns
US8526609B1 (en) * 2011-06-23 2013-09-03 Bby Solutions, Inc. Method for real-time compositing of encrypted video streams without decryption

Patent Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20060149781A1 (en) * 2004-12-30 2006-07-06 Massachusetts Institute Of Technology Techniques for relating arbitrary metadata to media files
US20090228492A1 (en) 2008-03-10 2009-09-10 Verizon Data Services Inc. Apparatus, system, and method for tagging media content
US20110058101A1 (en) * 2009-09-03 2011-03-10 Fox Broadcasting Company Method and apparatus for concurrent broadcast of media program and social networking derived information exchange
US20110221962A1 (en) * 2010-03-10 2011-09-15 Microsoft Corporation Augmented reality via a secondary channel
US20110225178A1 (en) * 2010-03-11 2011-09-15 Apple Inc. Automatic discovery of metadata

Cited By (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
EP2811752A1 (en) * 2013-06-03 2014-12-10 Alcatel Lucent Synchronization between multimedia flows and social network threads
WO2014195261A1 (en) * 2013-06-03 2014-12-11 Alcatel Lucent Synchronization between multimedia flows and social network threads
CN105284121A (en) * 2013-06-03 2016-01-27 阿尔卡特朗讯 Synchronization between multimedia flows and social network threads
CN105284121B (en) * 2013-06-03 2019-03-12 阿尔卡特朗讯 Synchronization between media stream and social networks thread
US10291968B2 (en) 2013-06-03 2019-05-14 Alcatel Lucent Synchronization between multimedia flows and social network threads
WO2017204836A1 (en) * 2016-05-27 2017-11-30 Thomson Licensing Method and apparatus for personal multimedia content distribution
US11057661B2 (en) 2016-05-27 2021-07-06 Interdigital Ce Patent Holdings Method and apparatus for personal multimedia content distribution

Also Published As

Publication number Publication date
US20140237510A1 (en) 2014-08-21
US9332313B2 (en) 2016-05-03

Similar Documents

Publication Publication Date Title
US9332313B2 (en) Personalized augmented A/V stream creation
US11490136B2 (en) Systems and methods for providing a slow motion video stream concurrently with a normal-speed video stream upon detection of an event
CA2924065C (en) Content based video content segmentation
US9578366B2 (en) Companion device services based on the generation and display of visual codes on a display device
US9430115B1 (en) Storyline presentation of content
EP2901631B1 (en) Enriching broadcast media related electronic messaging
US9396761B2 (en) Methods and systems for generating automatic replays in a media asset
US20100005084A1 (en) Method and system for prefetching internet content for video recorders
US8453179B2 (en) Linking real time media context to related applications and services
JP5868978B2 (en) Method and apparatus for providing community-based metadata
US10659842B2 (en) Integral program content distribution
US20150012946A1 (en) Methods and systems for presenting tag lines associated with media assets
US20130125188A1 (en) Multimedia presentation processing
US9959298B2 (en) Method, apparatus and system for indexing content based on time information
TWI554090B (en) Method and system for multimedia summary generation
US20160127807A1 (en) Dynamically determined audiovisual content guidebook
EP3044728A1 (en) Content based video content segmentation

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 12729539

Country of ref document: EP

Kind code of ref document: A1

WWE Wipo information: entry into national phase

Ref document number: 14346899

Country of ref document: US

NENP Non-entry into the national phase

Ref country code: DE

122 Ep: pct application non-entry in european phase

Ref document number: 12729539

Country of ref document: EP

Kind code of ref document: A1