WO2022086507A1 - Storage of remote-presented media content - Google Patents

Storage of remote-presented media content Download PDF

Info

Publication number
WO2022086507A1
WO2022086507A1 PCT/US2020/056497 US2020056497W WO2022086507A1 WO 2022086507 A1 WO2022086507 A1 WO 2022086507A1 US 2020056497 W US2020056497 W US 2020056497W WO 2022086507 A1 WO2022086507 A1 WO 2022086507A1
Authority
WO
WIPO (PCT)
Prior art keywords
media content
time
triggering event
electronic device
point
Prior art date
Application number
PCT/US2020/056497
Other languages
French (fr)
Inventor
Rafael DAL ZOTTO
Gareth R. WESTLAKE
Original Assignee
Hewlett-Packard Development Company, L.P.
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Hewlett-Packard Development Company, L.P. filed Critical Hewlett-Packard Development Company, L.P.
Priority to PCT/US2020/056497 priority Critical patent/WO2022086507A1/en
Publication of WO2022086507A1 publication Critical patent/WO2022086507A1/en

Links

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/20Servers specifically adapted for the distribution of content, e.g. VOD servers; Operations thereof
    • H04N21/23Processing of content or additional data; Elementary server operations; Server middleware
    • H04N21/231Content storage operation, e.g. caching movies for short term storage, replicating data over plural servers, prioritizing data for deletion
    • H04N21/23106Content storage operation, e.g. caching movies for short term storage, replicating data over plural servers, prioritizing data for deletion involving caching operations
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/43Processing of content or additional data, e.g. demultiplexing additional data from a digital video stream; Elementary client operations, e.g. monitoring of home network or synchronising decoder's clock; Client middleware
    • H04N21/4302Content synchronisation processes, e.g. decoder synchronisation
    • H04N21/4307Synchronising the rendering of multiple content streams or additional data on devices, e.g. synchronisation of audio on a mobile phone with the video output on the TV screen
    • H04N21/43074Synchronising the rendering of multiple content streams or additional data on devices, e.g. synchronisation of audio on a mobile phone with the video output on the TV screen of additional data with content streams on the same device, e.g. of EPG data or interactive icon with a TV program
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/80Generation or processing of content or additional data by content creator independently of the distribution process; Content per se
    • H04N21/83Generation or processing of protective or descriptive data associated with content; Content structuring
    • H04N21/845Structuring of content, e.g. decomposing content into time segments
    • H04N21/8456Structuring of content, e.g. decomposing content into time segments by decomposing the content in the time domain, e.g. in time segments
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/80Generation or processing of content or additional data by content creator independently of the distribution process; Content per se
    • H04N21/85Assembly of content; Generation of multimedia applications
    • H04N21/858Linking data to content, e.g. by linking an URL to a video object, by creating a hotspot

Definitions

  • Events such as meetings, classes, lectures, trainings, social gatherings, and other events can be conducted in a virtual or online environment.
  • Such events can be presented on a display screen by a graphical user interface that includes a combination of visual and audible components, such as audio, video, a shared screen, etc. While these events can facilitate interaction between attendees and provide information to the attendees, the interactions may be different in nature than if the events took place in an in-person or face-to-face nature rather than in a virtual or online environment.
  • FIG. 1 is a computing environment in accordance with various examples.
  • FIG. 2 is an action timeline in accordance with various examples.
  • FIG. 3 is an electronic device in accordance with various examples.
  • FIG. 4 is a flowchart of a method in accordance with various examples.
  • FIG. 5 is a flowchart of a method in accordance with various examples.
  • FIG. 6 is a flowchart of a method in accordance with various examples.
  • FIG. 7 is a flowchart of a method in accordance with various examples.
  • events such as meetings, classes, lectures, trainings, social gatherings, and other events can be conducted in a virtual or online environment.
  • Effectively consuming media content provided through these events can be challenging because of the volume of content provided, distractions around a viewer of the content, or other various challenges.
  • Understanding a context in which notes are taken during these events can aid in effective consumption of the content of the events, as well as later recall of the events and usefulness of the notes.
  • retaining portions of the events relating to particular subject matter can be useful based on perceived importance or significance of that subject matter.
  • This disclosure facilitates the composition of enhanced notes that include both notes and contextual information of events pertaining to those notes.
  • the creation of the enhanced notes is performed via executable code executing locally on a device on which a user is viewing content of the event.
  • the executable code in some examples, further enables the capturing of topics of interest in the events and/or tagging of content from the virtual events.
  • the executable code is independent of a provider of the event. For example, the executable code is usable across multiple event providers, media formats, etc. and is not an integrated component of an interface or system of the event provider.
  • the executable code implements a sliding buffer or a preprogrammed first duration of time.
  • the first duration of time may be any suitable duration selected and preprogrammed by a user.
  • the buffer may store any of audio, video, and/or other media content being presented to the user.
  • the user may select which of the audio, video, and/or other media content being presented is captured by the buffer.
  • the executable code may store the contents of the buffer, the note, and audio, video, and/or other media content being presented to the user for a second duration of time beginning with and following a time at which the note was generated.
  • the audio, video, and/or other media content being presented to the user for the first duration of time, preceding creation of the note, and the second duration of time, following creation of the note is stored to provide context to the note, aiding in understanding and usefulness of the note.
  • similar functionality of storing the audio, video, and/or other media content being presented to the user for the first duration of time and the second duration of time is performed responsive to a user input to tag the content being provided during the first and second durations of time.
  • similar functionality of storing the audio, video, and/or other media content being presented to the user for the first duration of time and the second duration of time is performed responsive to the executable code detecting the presence of a keyword or key image in the audio, video, and/or other media content being presented to the user.
  • FIG. 1 is a block diagram depicting an example computing environment 100.
  • the computing environment 100 includes a content distribution device 102, a network 104, and an electronic device 106.
  • the content distribution device 102 includes, in some examples, content distribution executable instructions 108.
  • Executable instructions in at least some examples, are also referred to as executable code.
  • the electronic device 106 includes at least contextual recording executable instructions 110.
  • the content distribution device 102 distributes media content via the network 104.
  • the content distribution device 102 may be a server or any other suitable computing device that is a remote presenter of the media content.
  • a remote presenter is a person, entity, or device that transmits media content to the electronic device 106 for presentation by the electronic device 106.
  • the remote presenter provides a particular graphical user interface or other executable code element through which the electronic device 106 accesses the media content and which is unique to the remote presenter.
  • the content distribution device 102 may distribute or present the media content according to any suitable technology, such as broadcast, unicast, simulcast, multicast, on-demand, subscription, streaming, etc.
  • the media content may be audio content, video content, screen sharing or other computer-generated graphics or content, an output of a document camera, an at least partially digital whiteboard or sharing space, a digitally created environment (e.g., such as an augmented reality or virtual reality environment), or any combination thereof or other media content suitable for distribution electronically.
  • the electronic device 106 receives the media content distributed by the content distribution device 102. Accordingly, in at least some examples the electronic device 106 is a local device in the computing environment 100.
  • the media content may be presented by the electronic device 106, such as via a graphical user interface implemented at least partially according to the contextual recording executable instructions 110 and presented on a display (not shown) and/or via speakers (not shown) of, or coupled to, the electronic device 106.
  • the graphical user interface may include multiple elements, such as a display element that displays the media content received from the content distribution device 102, a notes portion for addition of written, typed, spoken, video, or other notes by a user viewing the media content in the graphical user interface, and any suitable number of interactive elements.
  • the interactive elements may include a button to create a note, a button to add a marker, a button or entry field for modifying a setting associated with the contextual recording executable instructions 110 (e.g., such as a buffer or recording duration), or any other suitable preference or settings buttons or entry fields.
  • the contextual recording executable instructions 110 may implement a sliding buffer window that records a programmed duration of the media content.
  • the contextual recording executable instructions 110 records a programmed duration of the media content in a buffer in a first in, first out method.
  • the contextual recording executable instructions 110 may begin recording the media content to the buffer.
  • the contextual recording executable instructions 110 discards a unit time of media content from a beginning of the buffer.
  • the unit time of media content discarded from the buffer represents an oldest portion of the media content that is stored in the buffer.
  • the contextual recording executable instructions 110 writes the contents of the buffer to a persistent storage location.
  • the contextual recording executable instructions 110 further writes a second programmed duration of the media content that follows the triggering event to the persistent storage location.
  • the first programmed duration and the second programmed duration are stored together in a single, or aggregated, file.
  • the same file in at least some examples, includes an indication of, or information related to, the triggering event (e.g., such as a marker, tag, bookmark, identified keyword, audio transcript, user-created note(s), etc.).
  • a user may program which portions of the media content are written to the buffer and/or the persistent storage location. For example, the user may provide preference setting inputs indicating whether audio, video, a shared screen, or any other components of the media content are written to the buffer and/or the persistent storage location or are not written to the buffer and/or the persistent storage location.
  • the timeline 200 is representative of actions associated with the media content from a perspective of the contextual recording executable instructions 110.
  • the media content begins.
  • the triggering event discussed above occurs.
  • the media content ends.
  • the first programmed duration is y such that the buffer includes the media content from time T x-y to time T x .
  • the second programmed duration is z such that after the triggering event occurs, the contextual recording executable instructions 110 writes the media content from the time T x to T x+Z to the persistent location with the media content from the buffer.
  • the contextual recording executable instructions 110 writes media content from time T x-y to time T x+Z to the persistent storage location.
  • the buffer includes the portion of the media content from time T o to time T x .
  • T f occurs prior to T x+Z
  • the portion of the media content from time T x to time T f is written to the persistent storage location with the contents of the buffer.
  • the first programmed duration y and the second programed duration z may be the same value or may be different values, and each may be programmable (e.g., adjustable or selectable) by a user to any value(s) desired by the user. In at least some examples, when left unprogrammed by the user the first programmed duration y and the second programmed duration z may each have default durations of 30 seconds.
  • the triggering event may take various forms.
  • the triggering event may be the creation of a note by a user viewing the media content.
  • the note may be generated by the user typing or otherwise entering the note into a predefined entry field or via a predefined user entry button (e.g., such as a button to begin an audio recording).
  • a predefined user entry button e.g., such as a button to begin an audio recording.
  • the note may be provided by the user via an input device of the electronic device.
  • the triggering event may be a user providing an input to create a tag or bookmark in the media content.
  • the input may be the user pressing a soft button (e.g., an executable instructions based button, generated based on the contextual recording executable instructions 110 or an operating system of the electronic device 106) or a hardware button (e.g., such as a key or button of an input device of the electronic).
  • a soft button e.g., an executable instructions based button, generated based on the contextual recording executable instructions 110 or an operating system of the electronic device 106
  • a hardware button e.g., such as a key or button of an input device of the electronic.
  • the user may provide a name or other form of label for the tag or bookmark.
  • the triggering event is automatically generated according to the contextual recording executable instructions 110.
  • the user may program any number of topics of interest, such as words or phrases.
  • the electronic device 106 may monitor the media content in real-time or substantially real-time. Based on the monitoring, the electronic device 106 may activate the triggering event responsive to detection of a topic of interest.
  • contents of the buffer may be processed by the electronic device 106, such as via the contextual recording executable instructions 110, according to voice recognition and/or image recognition processes. These voice and/or image recognition processes may detect words contained in audio of the media content and/or words contained in video of the media content and transcribe these words into a text transcript.
  • the text transcript may be searched based on the contextual recording executable instructions 110 to determine or detect whether one of the topics of interest exists in the text transcript.
  • the electronic device 106 activates the triggering event based on the contextual recording executable instructions 110.
  • any number of triggering events, of any type may occur over the course of the media content and each triggering event is handled in substantially a same manner as described herein.
  • an indication of the triggering event may be written to the persistent storage location with the portions of the media content written to the persistent storage location.
  • the note created by the user, the marker or bookmark created by the user (and, in some examples, a label for the tag or bookmark), or the topic of interest may be written to the persistent storage location with the portions of the media content written to the persistent storage location.
  • a text transcript is determined for audio of the portion of the media content written to the persistent location and that text transcript is written to the persistent storage location as the indication of the triggering event with the portions of the media content written to the persistent storage location.
  • the indication of the triggering event is written to the persistent storage location with an association to a timecode of the media content. In this way, when the portion of the media content written to the persistent location is played back, the indication of the triggering event is presented at a contextually accurate time of the media content.
  • the portion of the media content written to the persistent location and the indication of the triggering event may, in some examples, be stored locally on the electronic device 106. In such examples, the portion of the media content written to the persistent location and the indication of the triggering event may be private to the user of the electronic device 106. In other examples, the portion of the media content written to the persistent location and the indication of the triggering event may be stored remotely, such as in a cloud-based database (not shown) or cloud-computing device (not shown).
  • the user may access a dashboard that includes content written based on the contextual recording executable instructions 110.
  • the dashboard may include content locally on the electronic device 106 or that is stored remotely. Via the dashboard, the user may delete, view, search for, modify, or share (e.g., such as with other users and/or via electronic messaging) the content presented via the dashboard.
  • FIG. 3 is a block diagram depicting an example of the electronic device 106 in more detail.
  • Electronic device 300 may be any suitable computing or processing device capable of performing the functions disclosed herein such as a computer system, a server, a laptop device, a tablet device, a personal computer, a cloud-computing node or device, etc.
  • Electronic device 300 implements at least some of the features/methods disclosed herein, for example, as described above with respect to the computing environment 100 and/or as described below with respect to any of the method 400, method 500, method 600, and/or method 700.
  • the electronic device 300 comprises input devices 310. Some of the input devices 310 may be microphones, keyboards, touchscreens, buttons, toggle switches, cameras, sensors, and/or other devices that allow a user to interact with, and provide input to, the electronic device 300. Some other of the input devices 310 may be downstream ports coupled to a transceiver (Tx/Rx) 320, which are transmitters, receivers, or combinations thereof. The Tx/Rx 320 transmits and/or receives data to and/or from other computing devices via at least some of the input devices 310. Similarly, the electronic device 300 comprises a plurality of output devices 340.
  • Tx/Rx transceiver
  • Some of the output devices 340 may be speakers, a display screen (which may also be an input device such as a touchscreen), lights, or any other device that allows a user to interact with, and receive output from, the electronic device 300.
  • At least some of the output devices 340 may be upstream ports coupled to another Tx/Rx 320, wherein the Tx/Rx 320 transmits and/or receives data from other nodes via the upstream ports.
  • the downstream ports and/or the upstream ports may include electrical and/or optical transmitting and/or receiving components.
  • the electronic device 300 comprises antennas (not shown) coupled to the Tx/Rx 320.
  • the Tx/Rx 320 transmits and/or receives data from other computing or storage devices wirelessly via the antennas.
  • the electronic device 300 may include additional Tx/Rx 320 such that the electronic device 300 has multiple networking or communication interfaces, for example, such that the electronic device 300 may communicate with a first device using a first communication interface (e.g., such as via the Internet) and may communicate with a second device using a second communication interface (e.g., such as another electronic device 300 without using the Internet).
  • a first communication interface e.g., such as via the Internet
  • a second communication interface e.g., such as another electronic device 300 without using the Internet.
  • a processor 330 is coupled to the Tx/Rx 320 and at least some of the input devices 310 and/or output devices 340 and implements the contextual recording executable instructions 110 described herein, such as via an contextual recording executable computer program product 360.
  • the processor 330 comprises multi-core processors and/or memory modules 350, which functions as data stores, buffers, etc.
  • the processor 330 is implemented as a general processor or as part of application specific integrated circuits (ASICs), field-programmable gate arrays (FPGAs), and/or digital signal processors (DSPs). Although illustrated as a single processor, the processor 330 is not so limited and may comprises multiple processors.
  • FIG. 3 also illustrates that a memory module 350 is coupled to the processor 330 and is a non-transitory medium to store various types of data.
  • Memory module 350 comprises memory devices including secondary storage, read-only memory (ROM), and random-access memory (RAM).
  • the secondary storage may comprise of one or more disk drives, optical drives, solid-state drives (SSDs), and/or tape drives and is used for non-volatile storage of data and as an over-flow storage device if the RAM is not large enough to hold all working data.
  • the secondary storage is used to store programs that are loaded into the RAM when such programs are selected for execution.
  • the ROM is used to store instructions and perhaps data that are read during program execution.
  • the ROM is a non-volatile memory device that may have a small memory capacity relative to the larger memory capacity of the secondary storage.
  • the RAM is used to store volatile data and perhaps to store instructions. Access to both the ROM and RAM may be faster than to the secondary storage.
  • the memory module 350 may be used to house the instructions for carrying out the various examples described herein.
  • the memory module 350 may comprise the contextual recording executable instructions 110 as a contextual recording executable computer program product 360, which is executed by processor 330.
  • the memory module 350 further includes a persistent storage location 365 and a buffer 370, as described elsewhere herein.
  • FIG. 4 is a flowchart of an example method 400 for contextual media content storage.
  • the method 400 is suitable for implementation on an electronic device, such as the electronic device 106 of the computing environment 100 of FIG. 1.
  • the method 400 may be embodied as the contextual recording executable instructions 110 executed by the electronic device 106.
  • the method 400 is implemented by the electronic device, in some examples, to record a portion of a media content with contextual data.
  • the electronic device receives media content.
  • the media content is, in some examples, streaming media content streamed according to any suitable protocol.
  • the media content may include any number of audio components and any number of video components.
  • the electronic device buffers a first portion of the media content.
  • the electronic device may implement a sliding buffer window that records a programmed duration of the media content as a first portion of the media content. For example, the electronic device records a programmed duration of the media content to a buffer in a first in, first out method.
  • the electronic device may begin recording the media content to the buffer.
  • the electronic device discards a unit time of media content from a beginning of the buffer. In this way, the unit time of media content discarded from the buffer represents an oldest portion of the media content that is stored in the buffer.
  • the electronic device detects a triggering event associated with the media content.
  • the triggering event may be user-generated or may be generated by the electronic device.
  • the triggering event may be a user creating a note, a user creating a marker or bookmark, or the electronic device automatically detecting a topic of interest or keyword in the media content.
  • the electronic device stores the contents of the buffer and a second portion of the media content to a persistent storage location. Responsive to detection of the triggering event, the electronic device stores the contents of the buffer and the second portion of the media content.
  • the first portion of the media content, stored in the buffer may be a first duration of the media content.
  • the second portion of the media content may be a second duration of the media content that is the same as the first duration or is different than the first duration.
  • the electronic device stores data associated with the triggering event to the persistent storage location.
  • the electronic device may store the note, marker, or topic of interest to the persistent storage location, as described elsewhere herein.
  • the data associated with the triggering event and stored in the persistent storage location may be associated with the stored first or second portions of the media content. In this way, the first and second portions of the media content provide context to the data associated with the triggering event, creating, for example, an enhanced note when the data associated with the triggering event is a note.
  • the method 400 includes additional, other operations.
  • the method 400 may further include analyzing the first and/or second portions of the media content stored by the electronic device to generate a text transcript of the first and/or second portions of the media content.
  • this text transcript is stored by the electronic device with the first and/or second portions of the media content.
  • FIG. 5 is a flowchart of an example method 500 for contextual media content storage.
  • the method 500 is suitable for implementation on an electronic device, such as the electronic device 106 of the computing environment 100 of FIG. 1 and FIG. 3.
  • the method 500 may be embodied as the contextual recording executable instructions 110 executed by the electronic device 106 as the contextual recording executable computer program product 360.
  • the method 500 may be implemented as computer-executable instructions or code, stored on a computer-readable medium, such as the memory module 350 of FIG. 3, which, when executed by a processor such as the processor 330 of FIG. 3, causes the processor to execute the computer-executable instructions to perform operations.
  • the method 500 is implemented by the electronic device, in some examples, to record a portion of a media content with contextual data.
  • the electronic device which includes a computer- readable medium, buffers a first portion of a media content presented on the electronic device by a remote presenter.
  • the electronic device may implement a sliding buffer window that records a programmed duration of the media content as a first portion of the media content. For example, the electronic device records a programmed duration of the media content to a buffer in a first in, first out method.
  • the electronic device may begin recording the media content to the buffer.
  • the electronic device discards a unit time of media content from a beginning of the buffer. In this way, the unit time of media content discarded from the buffer represents an oldest portion of the media content that is stored in the buffer.
  • the electronic device detects an occurrence of a triggering event associated with the media content.
  • the triggering event may be user-generated or may be generated by the electronic device.
  • the triggering event may be a user creating a note, a user creating a marker or bookmark, or the electronic device automatically detecting a topic of interest or keyword in the media content.
  • the electronic device stores in the computer-readable medium the first portion of the media content and a second portion of the media content that follows a point in time at which the triggering event was detected, along with a reference to the triggering event.
  • the method 500 includes additional, other operations.
  • the method 500 may further include automatically analyzing the first portion of the media content and the second portion of the media content to generate a text transcript of audio included in the first portion of the media content and the second portion of the media content.
  • this text transcript is stored in the computer-readable medium with an association to the first portion of the media content and the second portion of the media content.
  • FIG. 6 is a flowchart of an example method 600 for contextual media content storage.
  • the method 600 is suitable for implementation on an electronic device, such as the electronic device 106 of the computing environment 100 of FIG. 1 and FIG. 3.
  • the method 600 may be embodied as the contextual recording executable instructions 110 executed by the electronic device 106 as the contextual recording executable computer program product 360.
  • the method 600 is implemented by the electronic device, in some examples, to record a portion of a media content with contextual data.
  • the electronic device presenting, on the electron device, media content provided by a remote presenter in a graphical user interface other than a graphical user interface provided by the remote presenter.
  • the media content is, in some examples, streaming media content streamed according to any suitable protocol.
  • the media content may include any number of audio components and any number of video components.
  • the electronic device detects an occurrence of a triggering event associated with the media content at a point in time.
  • the triggering event may be user-generated or may be generated by the electronic device.
  • the triggering event may be a user creating a note, a user creating a marker or bookmark, or the electronic device automatically detecting a topic of interest or keyword in the media content.
  • the electronic device stores a first portion of the media content preceding the point in time for a first duration of time and a second portion of the media content following the point in time for a second duration of time.
  • the first portion of the media content may be a first duration of the media content and may have previously been stored by the electronic device in a sliding window buffer.
  • the second portion of the media content may be a second duration of the media content that is the same as the first duration or is different than the first duration.
  • the electronic device stores an association between the triggering event and the stored first and second portions of the media content.
  • the electronic device may store a note, marker or tag, or topic of interest or keyword, as described elsewhere herein. In this way, the first and second portions of the media content provide context to the triggering event, creating, for example, an enhanced note when the triggering 600 is a note.
  • the method 600 includes additional, other operations.
  • the method 600 may further include automatically analyzing the first portion of the media content and the second portion of the media content to generate a text transcript of audio included in the first portion of the media content and the second portion of the media content.
  • this text transcript is stored in the computer-readable medium with an association to the first portion of the media content and the second portion of the media content.
  • FIG. 7 is a flowchart of an example method 700 for contextual media content storage.
  • the method 700 is suitable for implementation on an electronic device, such as the electronic device 106 of the computing environment 100 of FIG. 1 and FIG. 3. Accordingly, reference is made in FIG. 7 to components of FIG. 3, but those components of FIG. 3 are not separately described herein with respect to FIG. 7.
  • the method 700 may be embodied as the contextual recording executable instructions 110 executed by the electronic device 106 as the contextual recording executable computer program product 360.
  • the method 700 is implemented by the electronic device, in some examples, to record a portion of a media content with contextual data, such that the method 700 is stored in the memory 350 and is executable by the processor 330.
  • the electronic device stores a first portion of a media content presented on the electronic device by a remote presenter.
  • the media content is, in some examples, streaming media content streamed according to any suitable protocol.
  • the media content may include any number of audio components and any number of video components.
  • the first portion of the media content is stored, in some sliding buffer window of a memory of the electronic device.
  • the electronic device receives user input associated with the media content at a point in time occurring at an end of the first portion of the media content, the user input comprising a triggering event.
  • the triggering event may be user-generated or may be generated by the electronic device.
  • the triggering event may be a user creating a note, a user creating a marker or bookmark, or the electronic device automatically detecting a topic of interest or keyword in the media content.
  • the electronic device stores, in the memory of the electronic device, the user input with a context of the media content corresponding to the point in time, the context of the media provided by the electronic device storing the first portion of the media content and a second portion of the media content following the point in time.
  • the method 700 includes additional, other operations.
  • the method 700 may further include automatically analyzing the first portion of the media content and the second portion of the media content to generate a text transcript of audio included in the first portion of the media content and the second portion of the media content.
  • this text transcript is stored in the computer-readable medium with an association to the first portion of the media content and the second portion of the media content.

Landscapes

  • Engineering & Computer Science (AREA)
  • Multimedia (AREA)
  • Signal Processing (AREA)
  • Information Transfer Between Computers (AREA)

Abstract

A method is described in which media content provided by a remote presenter is presented on a local device in a graphical user interface other than a graphical user interface provided by the remote presenter; an occurrence of a triggering event associated with the media content at a point in time is detected on the local device; a first portion of the media content preceding the point in time for a first duration of time and a second portion of the media content following the point in time for a second duration of time is stored on the local device responsive to detecting the occurrence of the triggering event; and an association between the triggering event and the stored first and second portions of the media content is stored on the local device.

Description

STORAGE OF REMOTE-PRESENTED MEDIA CONTENT
BACKGROUND
[0001] Events such as meetings, classes, lectures, trainings, social gatherings, and other events can be conducted in a virtual or online environment. Such events can be presented on a display screen by a graphical user interface that includes a combination of visual and audible components, such as audio, video, a shared screen, etc. While these events can facilitate interaction between attendees and provide information to the attendees, the interactions may be different in nature than if the events took place in an in-person or face-to-face nature rather than in a virtual or online environment.
BRIEF DESCRIPTION OF THE DRAWINGS
[0002] Various examples will be described below referring to the following figures:
[0003] FIG. 1 is a computing environment in accordance with various examples.
[0004] FIG. 2 is an action timeline in accordance with various examples.
[0005] FIG. 3 is an electronic device in accordance with various examples.
[0006] FIG. 4 is a flowchart of a method in accordance with various examples.
[0007] FIG. 5 is a flowchart of a method in accordance with various examples.
[0008] FIG. 6 is a flowchart of a method in accordance with various examples.
[0009] FIG. 7 is a flowchart of a method in accordance with various examples.
DETAILED DESCRIPTION
[0010] As explained above, events such as meetings, classes, lectures, trainings, social gatherings, and other events can be conducted in a virtual or online environment. Effectively consuming media content provided through these events, such as presented through a graphical user interface associated with the events can be challenging because of the volume of content provided, distractions around a viewer of the content, or other various challenges. Understanding a context in which notes are taken during these events can aid in effective consumption of the content of the events, as well as later recall of the events and usefulness of the notes. Further, retaining portions of the events relating to particular subject matter can be useful based on perceived importance or significance of that subject matter.
[0011] This disclosure facilitates the composition of enhanced notes that include both notes and contextual information of events pertaining to those notes. In at least some examples, the creation of the enhanced notes is performed via executable code executing locally on a device on which a user is viewing content of the event. The executable code, in some examples, further enables the capturing of topics of interest in the events and/or tagging of content from the virtual events. In at least some examples, the executable code is independent of a provider of the event. For example, the executable code is usable across multiple event providers, media formats, etc. and is not an integrated component of an interface or system of the event provider.
[0012] In some examples, the executable code implements a sliding buffer or a preprogrammed first duration of time. The first duration of time may be any suitable duration selected and preprogrammed by a user. The buffer may store any of audio, video, and/or other media content being presented to the user. The user may select which of the audio, video, and/or other media content being presented is captured by the buffer. When the user generates a note, the executable code may store the contents of the buffer, the note, and audio, video, and/or other media content being presented to the user for a second duration of time beginning with and following a time at which the note was generated. In this way, the audio, video, and/or other media content being presented to the user for the first duration of time, preceding creation of the note, and the second duration of time, following creation of the note, is stored to provide context to the note, aiding in understanding and usefulness of the note. In other examples, similar functionality of storing the audio, video, and/or other media content being presented to the user for the first duration of time and the second duration of time is performed responsive to a user input to tag the content being provided during the first and second durations of time. In yet other examples, similar functionality of storing the audio, video, and/or other media content being presented to the user for the first duration of time and the second duration of time is performed responsive to the executable code detecting the presence of a keyword or key image in the audio, video, and/or other media content being presented to the user.
[0013] FIG. 1 is a block diagram depicting an example computing environment 100. In at least some examples, the computing environment 100 includes a content distribution device 102, a network 104, and an electronic device 106. The content distribution device 102 includes, in some examples, content distribution executable instructions 108. Executable instructions, in at least some examples, are also referred to as executable code. The electronic device 106 includes at least contextual recording executable instructions 110.
[0014] In some examples, the content distribution device 102, according to the content distribution executable instructions 108, distributes media content via the network 104. The content distribution device 102 may be a server or any other suitable computing device that is a remote presenter of the media content. As used herein, a remote presenter is a person, entity, or device that transmits media content to the electronic device 106 for presentation by the electronic device 106. In some examples, the remote presenter provides a particular graphical user interface or other executable code element through which the electronic device 106 accesses the media content and which is unique to the remote presenter. The content distribution device 102 may distribute or present the media content according to any suitable technology, such as broadcast, unicast, simulcast, multicast, on-demand, subscription, streaming, etc. The media content may be audio content, video content, screen sharing or other computer-generated graphics or content, an output of a document camera, an at least partially digital whiteboard or sharing space, a digitally created environment (e.g., such as an augmented reality or virtual reality environment), or any combination thereof or other media content suitable for distribution electronically.
[0015] In at least some examples, the electronic device 106, according to the contextual recording executable instructions 110, receives the media content distributed by the content distribution device 102. Accordingly, in at least some examples the electronic device 106 is a local device in the computing environment 100. The media content may be presented by the electronic device 106, such as via a graphical user interface implemented at least partially according to the contextual recording executable instructions 110 and presented on a display (not shown) and/or via speakers (not shown) of, or coupled to, the electronic device 106. The graphical user interface may include multiple elements, such as a display element that displays the media content received from the content distribution device 102, a notes portion for addition of written, typed, spoken, video, or other notes by a user viewing the media content in the graphical user interface, and any suitable number of interactive elements. The interactive elements may include a button to create a note, a button to add a marker, a button or entry field for modifying a setting associated with the contextual recording executable instructions 110 (e.g., such as a buffer or recording duration), or any other suitable preference or settings buttons or entry fields.
[0016] Based on the contextual recording executable instructions 110, a portion of the media content is recorded. For example, the contextual recording executable instructions 110 may implement a sliding buffer window that records a programmed duration of the media content. For example, the contextual recording executable instructions 110 records a programmed duration of the media content in a buffer in a first in, first out method. At the beginning of the media content, the contextual recording executable instructions 110 may begin recording the media content to the buffer. After a first programmed duration of the media content has been recorded to the buffer, for each unit time of media content subsequently recorded to the buffer, the contextual recording executable instructions 110 discards a unit time of media content from a beginning of the buffer. In this way, the unit time of media content discarded from the buffer represents an oldest portion of the media content that is stored in the buffer. When a triggering event occurs, the contextual recording executable instructions 110 writes the contents of the buffer to a persistent storage location. The contextual recording executable instructions 110 further writes a second programmed duration of the media content that follows the triggering event to the persistent storage location. In at least some implementations, the first programmed duration and the second programmed duration are stored together in a single, or aggregated, file. The same file, in at least some examples, includes an indication of, or information related to, the triggering event (e.g., such as a marker, tag, bookmark, identified keyword, audio transcript, user-created note(s), etc.). In at least some examples, a user may program which portions of the media content are written to the buffer and/or the persistent storage location. For example, the user may provide preference setting inputs indicating whether audio, video, a shared screen, or any other components of the media content are written to the buffer and/or the persistent storage location or are not written to the buffer and/or the persistent storage location.
[0017] Turning for a moment to FIG. 2, an example action timeline 200 is shown. The timeline 200 is representative of actions associated with the media content from a perspective of the contextual recording executable instructions 110. At time To, the media content begins. At a point in time Tx, the triggering event discussed above occurs. At time Tf, the media content ends. The first programmed duration is y such that the buffer includes the media content from time Tx-y to time Tx. The second programmed duration is z such that after the triggering event occurs, the contextual recording executable instructions 110 writes the media content from the time Tx to Tx+Z to the persistent location with the media content from the buffer. In this way, when a triggering event occurs, the contextual recording executable instructions 110 writes media content from time Tx-y to time Tx+Z to the persistent storage location. Although not illustrated in FIG. 2, when Tx occurs prior to T0+y, the buffer includes the portion of the media content from time To to time Tx. Although similarly not illustrated in FIG. 2, when Tf occurs prior to Tx+Z, the portion of the media content from time Tx to time Tf is written to the persistent storage location with the contents of the buffer. The first programmed duration y and the second programed duration z may be the same value or may be different values, and each may be programmable (e.g., adjustable or selectable) by a user to any value(s) desired by the user. In at least some examples, when left unprogrammed by the user the first programmed duration y and the second programmed duration z may each have default durations of 30 seconds.
[0018] Returning now to FIG. 1 , in various examples the triggering event may take various forms. In some examples, the triggering event may be the creation of a note by a user viewing the media content. The note may be generated by the user typing or otherwise entering the note into a predefined entry field or via a predefined user entry button (e.g., such as a button to begin an audio recording). Generally, the note may be provided by the user via an input device of the electronic device.
[0019] In other examples, the triggering event may be a user providing an input to create a tag or bookmark in the media content. The input may be the user pressing a soft button (e.g., an executable instructions based button, generated based on the contextual recording executable instructions 110 or an operating system of the electronic device 106) or a hardware button (e.g., such as a key or button of an input device of the electronic). In some examples, the user may provide a name or other form of label for the tag or bookmark.
[0020] In yet other examples, the triggering event is automatically generated according to the contextual recording executable instructions 110. For example, the user may program any number of topics of interest, such as words or phrases. The electronic device 106, based at least partially on the contextual recording executable instructions 110, may monitor the media content in real-time or substantially real-time. Based on the monitoring, the electronic device 106 may activate the triggering event responsive to detection of a topic of interest. For example, contents of the buffer may be processed by the electronic device 106, such as via the contextual recording executable instructions 110, according to voice recognition and/or image recognition processes. These voice and/or image recognition processes may detect words contained in audio of the media content and/or words contained in video of the media content and transcribe these words into a text transcript. The text transcript may be searched based on the contextual recording executable instructions 110 to determine or detect whether one of the topics of interest exists in the text transcript. When a topic of interest is detected in the text transcript, the electronic device 106 activates the triggering event based on the contextual recording executable instructions 110. In various examples, any number of triggering events, of any type, may occur over the course of the media content and each triggering event is handled in substantially a same manner as described herein. [0021] In some examples, an indication of the triggering event may be written to the persistent storage location with the portions of the media content written to the persistent storage location. For example, the note created by the user, the marker or bookmark created by the user (and, in some examples, a label for the tag or bookmark), or the topic of interest may be written to the persistent storage location with the portions of the media content written to the persistent storage location. In yet other examples, a text transcript is determined for audio of the portion of the media content written to the persistent location and that text transcript is written to the persistent storage location as the indication of the triggering event with the portions of the media content written to the persistent storage location. In at least some examples, the indication of the triggering event is written to the persistent storage location with an association to a timecode of the media content. In this way, when the portion of the media content written to the persistent location is played back, the indication of the triggering event is presented at a contextually accurate time of the media content.
[0022] The portion of the media content written to the persistent location and the indication of the triggering event may, in some examples, be stored locally on the electronic device 106. In such examples, the portion of the media content written to the persistent location and the indication of the triggering event may be private to the user of the electronic device 106. In other examples, the portion of the media content written to the persistent location and the indication of the triggering event may be stored remotely, such as in a cloud-based database (not shown) or cloud-computing device (not shown). The user may access a dashboard that includes content written based on the contextual recording executable instructions 110. The dashboard may include content locally on the electronic device 106 or that is stored remotely. Via the dashboard, the user may delete, view, search for, modify, or share (e.g., such as with other users and/or via electronic messaging) the content presented via the dashboard.
[0023] FIG. 3 is a block diagram depicting an example of the electronic device 106 in more detail. Electronic device 300 may be any suitable computing or processing device capable of performing the functions disclosed herein such as a computer system, a server, a laptop device, a tablet device, a personal computer, a cloud-computing node or device, etc. Electronic device 300 implements at least some of the features/methods disclosed herein, for example, as described above with respect to the computing environment 100 and/or as described below with respect to any of the method 400, method 500, method 600, and/or method 700.
[0024] The electronic device 300 comprises input devices 310. Some of the input devices 310 may be microphones, keyboards, touchscreens, buttons, toggle switches, cameras, sensors, and/or other devices that allow a user to interact with, and provide input to, the electronic device 300. Some other of the input devices 310 may be downstream ports coupled to a transceiver (Tx/Rx) 320, which are transmitters, receivers, or combinations thereof. The Tx/Rx 320 transmits and/or receives data to and/or from other computing devices via at least some of the input devices 310. Similarly, the electronic device 300 comprises a plurality of output devices 340. Some of the output devices 340 may be speakers, a display screen (which may also be an input device such as a touchscreen), lights, or any other device that allows a user to interact with, and receive output from, the electronic device 300. At least some of the output devices 340 may be upstream ports coupled to another Tx/Rx 320, wherein the Tx/Rx 320 transmits and/or receives data from other nodes via the upstream ports. The downstream ports and/or the upstream ports may include electrical and/or optical transmitting and/or receiving components. In another example, the electronic device 300 comprises antennas (not shown) coupled to the Tx/Rx 320. The Tx/Rx 320 transmits and/or receives data from other computing or storage devices wirelessly via the antennas. In yet other examples, the electronic device 300 may include additional Tx/Rx 320 such that the electronic device 300 has multiple networking or communication interfaces, for example, such that the electronic device 300 may communicate with a first device using a first communication interface (e.g., such as via the Internet) and may communicate with a second device using a second communication interface (e.g., such as another electronic device 300 without using the Internet).
[0025] A processor 330 is coupled to the Tx/Rx 320 and at least some of the input devices 310 and/or output devices 340 and implements the contextual recording executable instructions 110 described herein, such as via an contextual recording executable computer program product 360. In an example, the processor 330 comprises multi-core processors and/or memory modules 350, which functions as data stores, buffers, etc. The processor 330 is implemented as a general processor or as part of application specific integrated circuits (ASICs), field-programmable gate arrays (FPGAs), and/or digital signal processors (DSPs). Although illustrated as a single processor, the processor 330 is not so limited and may comprises multiple processors.
[0026] FIG. 3 also illustrates that a memory module 350 is coupled to the processor 330 and is a non-transitory medium to store various types of data. Memory module 350 comprises memory devices including secondary storage, read-only memory (ROM), and random-access memory (RAM). The secondary storage may comprise of one or more disk drives, optical drives, solid-state drives (SSDs), and/or tape drives and is used for non-volatile storage of data and as an over-flow storage device if the RAM is not large enough to hold all working data. The secondary storage is used to store programs that are loaded into the RAM when such programs are selected for execution. The ROM is used to store instructions and perhaps data that are read during program execution. The ROM is a non-volatile memory device that may have a small memory capacity relative to the larger memory capacity of the secondary storage. The RAM is used to store volatile data and perhaps to store instructions. Access to both the ROM and RAM may be faster than to the secondary storage.
[0027] The memory module 350 may be used to house the instructions for carrying out the various examples described herein. For example, the memory module 350 may comprise the contextual recording executable instructions 110 as a contextual recording executable computer program product 360, which is executed by processor 330. The memory module 350 further includes a persistent storage location 365 and a buffer 370, as described elsewhere herein.
[0028] It is understood that by programming and/or loading executable instructions onto the electronic device 300, at least one of the processor 330 and/or the memory module 350 are changed, transforming the electronic device 300 in part into a particular machine or apparatus, for example, a contextual recording system having the novel functionality taught by the present disclosure. [0029] FIG. 4 is a flowchart of an example method 400 for contextual media content storage. In at least some examples, the method 400 is suitable for implementation on an electronic device, such as the electronic device 106 of the computing environment 100 of FIG. 1. For example, in at least some implementations the method 400 may be embodied as the contextual recording executable instructions 110 executed by the electronic device 106. The method 400 is implemented by the electronic device, in some examples, to record a portion of a media content with contextual data.
[0030] At operation 402, the electronic device receives media content. The media content is, in some examples, streaming media content streamed according to any suitable protocol. The media content may include any number of audio components and any number of video components.
[0031] At operation 404, the electronic device buffers a first portion of the media content. The electronic device may implement a sliding buffer window that records a programmed duration of the media content as a first portion of the media content. For example, the electronic device records a programmed duration of the media content to a buffer in a first in, first out method. At the beginning of the media content, the electronic device may begin recording the media content to the buffer. After a first programmed duration of the media content has been recorded to the buffer, for each unit time of media content subsequently recorded to the buffer, the electronic device discards a unit time of media content from a beginning of the buffer. In this way, the unit time of media content discarded from the buffer represents an oldest portion of the media content that is stored in the buffer.
[0032] At operation 406, the electronic device detects a triggering event associated with the media content. The triggering event may be user-generated or may be generated by the electronic device. For example, as described elsewhere herein, the triggering event may be a user creating a note, a user creating a marker or bookmark, or the electronic device automatically detecting a topic of interest or keyword in the media content.
[0033] At operation 408, the electronic device stores the contents of the buffer and a second portion of the media content to a persistent storage location. Responsive to detection of the triggering event, the electronic device stores the contents of the buffer and the second portion of the media content. The first portion of the media content, stored in the buffer, may be a first duration of the media content. The second portion of the media content may be a second duration of the media content that is the same as the first duration or is different than the first duration.
[0034] At operation 410, the electronic device stores data associated with the triggering event to the persistent storage location. For example, the electronic device may store the note, marker, or topic of interest to the persistent storage location, as described elsewhere herein. The data associated with the triggering event and stored in the persistent storage location may be associated with the stored first or second portions of the media content. In this way, the first and second portions of the media content provide context to the data associated with the triggering event, creating, for example, an enhanced note when the data associated with the triggering event is a note.
[0035] In at least some examples, the method 400 includes additional, other operations. For example, in various implementations the method 400 may further include analyzing the first and/or second portions of the media content stored by the electronic device to generate a text transcript of the first and/or second portions of the media content. In some examples, this text transcript is stored by the electronic device with the first and/or second portions of the media content.
[0036] FIG. 5 is a flowchart of an example method 500 for contextual media content storage. In at least some examples, the method 500 is suitable for implementation on an electronic device, such as the electronic device 106 of the computing environment 100 of FIG. 1 and FIG. 3. For example, in at least some implementations the method 500 may be embodied as the contextual recording executable instructions 110 executed by the electronic device 106 as the contextual recording executable computer program product 360. Accordingly, the method 500 may be implemented as computer-executable instructions or code, stored on a computer-readable medium, such as the memory module 350 of FIG. 3, which, when executed by a processor such as the processor 330 of FIG. 3, causes the processor to execute the computer-executable instructions to perform operations. The method 500 is implemented by the electronic device, in some examples, to record a portion of a media content with contextual data.
[0037] At operation 502, the electronic device, which includes a computer- readable medium, buffers a first portion of a media content presented on the electronic device by a remote presenter. The electronic device may implement a sliding buffer window that records a programmed duration of the media content as a first portion of the media content. For example, the electronic device records a programmed duration of the media content to a buffer in a first in, first out method. At the beginning of the media content, the electronic device may begin recording the media content to the buffer. After a first programmed duration of the media content has been recorded to the buffer, for each unit time of media content subsequently recorded to the buffer, the electronic device discards a unit time of media content from a beginning of the buffer. In this way, the unit time of media content discarded from the buffer represents an oldest portion of the media content that is stored in the buffer.
[0038] At operation 504, the electronic device detects an occurrence of a triggering event associated with the media content. The triggering event may be user-generated or may be generated by the electronic device. For example, as described elsewhere herein, the triggering event may be a user creating a note, a user creating a marker or bookmark, or the electronic device automatically detecting a topic of interest or keyword in the media content.
[0039] At operation 506, responsive to detecting the occurrence of the triggering event, the electronic device stores in the computer-readable medium the first portion of the media content and a second portion of the media content that follows a point in time at which the triggering event was detected, along with a reference to the triggering event.
[0040] In at least some examples, the method 500 includes additional, other operations. For example, in various implementations the method 500 may further include automatically analyzing the first portion of the media content and the second portion of the media content to generate a text transcript of audio included in the first portion of the media content and the second portion of the media content. In some examples, this text transcript is stored in the computer-readable medium with an association to the first portion of the media content and the second portion of the media content.
[0041] FIG. 6 is a flowchart of an example method 600 for contextual media content storage. In at least some examples, the method 600 is suitable for implementation on an electronic device, such as the electronic device 106 of the computing environment 100 of FIG. 1 and FIG. 3. For example, in at least some implementations the method 600 may be embodied as the contextual recording executable instructions 110 executed by the electronic device 106 as the contextual recording executable computer program product 360. The method 600 is implemented by the electronic device, in some examples, to record a portion of a media content with contextual data.
[0042] At operation 602, the electronic device presenting, on the electron device, media content provided by a remote presenter in a graphical user interface other than a graphical user interface provided by the remote presenter. The media content is, in some examples, streaming media content streamed according to any suitable protocol. The media content may include any number of audio components and any number of video components.
[0043] At operation 604, the electronic device detects an occurrence of a triggering event associated with the media content at a point in time. The triggering event may be user-generated or may be generated by the electronic device. For example, as described elsewhere herein, the triggering event may be a user creating a note, a user creating a marker or bookmark, or the electronic device automatically detecting a topic of interest or keyword in the media content. [0044] At operation 606, responsive to detecting the occurrence of the triggering event, the electronic device stores a first portion of the media content preceding the point in time for a first duration of time and a second portion of the media content following the point in time for a second duration of time. The first portion of the media content may be a first duration of the media content and may have previously been stored by the electronic device in a sliding window buffer. The second portion of the media content may be a second duration of the media content that is the same as the first duration or is different than the first duration. [0045] At operation 608, the electronic device stores an association between the triggering event and the stored first and second portions of the media content. For example, the electronic device may store a note, marker or tag, or topic of interest or keyword, as described elsewhere herein. In this way, the first and second portions of the media content provide context to the triggering event, creating, for example, an enhanced note when the triggering 600 is a note.
[0046] In at least some examples, the method 600 includes additional, other operations. For example, in various implementations the method 600 may further include automatically analyzing the first portion of the media content and the second portion of the media content to generate a text transcript of audio included in the first portion of the media content and the second portion of the media content. In some examples, this text transcript is stored in the computer-readable medium with an association to the first portion of the media content and the second portion of the media content.
[0047] FIG. 7 is a flowchart of an example method 700 for contextual media content storage. In at least some examples, the method 700 is suitable for implementation on an electronic device, such as the electronic device 106 of the computing environment 100 of FIG. 1 and FIG. 3. Accordingly, reference is made in FIG. 7 to components of FIG. 3, but those components of FIG. 3 are not separately described herein with respect to FIG. 7. For example, in at least some implementations the method 700 may be embodied as the contextual recording executable instructions 110 executed by the electronic device 106 as the contextual recording executable computer program product 360. The method 700 is implemented by the electronic device, in some examples, to record a portion of a media content with contextual data, such that the method 700 is stored in the memory 350 and is executable by the processor 330.
[0048] At operation 702, the electronic device stores a first portion of a media content presented on the electronic device by a remote presenter. The media content is, in some examples, streaming media content streamed according to any suitable protocol. The media content may include any number of audio components and any number of video components. The first portion of the media content is stored, in some sliding buffer window of a memory of the electronic device.
[0049] At operation 704, the electronic device receives user input associated with the media content at a point in time occurring at an end of the first portion of the media content, the user input comprising a triggering event. The triggering event may be user-generated or may be generated by the electronic device. For example, as described elsewhere herein, the triggering event may be a user creating a note, a user creating a marker or bookmark, or the electronic device automatically detecting a topic of interest or keyword in the media content.
[0050] At operation 706, responsive to receiving the user input as the triggering event, the electronic device stores, in the memory of the electronic device, the user input with a context of the media content corresponding to the point in time, the context of the media provided by the electronic device storing the first portion of the media content and a second portion of the media content following the point in time.
[0051] In at least some examples, the method 700 includes additional, other operations. For example, in various implementations the method 700 may further include automatically analyzing the first portion of the media content and the second portion of the media content to generate a text transcript of audio included in the first portion of the media content and the second portion of the media content. In some examples, this text transcript is stored in the computer-readable medium with an association to the first portion of the media content and the second portion of the media content.
[0052] The above discussion is meant to be illustrative of the principles and various examples of the present disclosure. Numerous variations and modifications will become apparent to those skilled in the art once the above disclosure is fully appreciated. It is intended that the following claims be interpreted to embrace all such variations and modifications.

Claims

CLAIMS What is claimed is:
1. A computer-readable medium storing executable code, which, when executed by a processor, causes the processor to: buffer, on a local device including the computer-readable medium, a first portion of a media content presented on the local device by a remote presenter; detect an occurrence of a triggering event associated with the media content; and responsive to detecting the occurrence of the triggering event, store in the computer-readable medium the first portion of the media content and a second portion of the media content that follows a point in time at which the triggering event was detected, along with a reference to the triggering event.
2. The computer-readable medium of claim 1 , wherein the triggering event comprises a user generating a note at the point in time.
3. The computer-readable medium of claim 1 , wherein the triggering event comprises: analyzing the first portion of the media content to detect a programmed topic of interest included in the first portion of the media content; and detecting, at the point in time, the topic of interest in the media content.
4. The computer-readable medium of claim 1 , wherein executing the executable code causes the processor to: automatically analyze the first portion of the media content and the second portion of the media content to generate a text transcript of audio included in the first portion of the media content and the second portion of the media content; and store the text transcript in the computer-readable medium with an association to the first portion of the media content and the second portion of the media content.
5. A method, comprising: presenting, on a local device, media content provided by a remote presenter in a graphical user interface other than a graphical user interface provided by the remote presenter; detecting, on the local device, an occurrence of a triggering event associated with the media content at a point in time; responsive to detecting the occurrence of the triggering event, storing, by the local device, a first portion of the media content preceding the point in time for a first duration of time and a second portion of the media content following the point in time for a second duration of time; and storing, on the local device, an association between the triggering event and the stored first and second portions of the media content.
6. The method of claim 5, wherein detecting the occurrence of the triggering event associated with the media content at the point in time comprises receiving user input from a user beginning at the point in time.
7. The method of claim 5, wherein detecting the occurrence of the triggering event associated with the media content at the point in time comprises: analyzing the first portion of the media content to detect a programmed topic of interest included in the first portion of the media content; and detecting, at the point in time, the topic of interest in the media content. 18
8. The method of claim 5, comprising: automatically analyzing the first and second portions of the media content to generate a text transcript of audio included in the first and second portions of the media content; and storing the text transcript with an association to the triggering event and the stored first and second portions of the media content.
9. The method of claim 5, wherein the first duration of time is the same as the second duration of time.
10. The method of claim 5, wherein the first duration of time is different from the second duration of time.
11 . The method of claim 5, comprising, prior to detecting the occurrence of the triggering event associated with the media content, storing the first portion of the media content in a sliding buffer window.
12. An electronic device, comprising: a memory; and a processor coupled to the memory, the processor to: store, in the memory, a first portion of a media content presented on the electronic device by a remote presenter; receive user input associated with the media content at a point in time occurring at an end of the first portion of the media content, the user input comprising a triggering event; and responsive to receiving the user input as the triggering event, store, in the memory, the user input with a context of the media content corresponding to the point in time, the context of the media provided by the electronic device storing the first portion of the media content and a second portion of the media content following the point in time. 19
13. The electronic device of claim 12, wherein the user input is a text-based note generated by a user beginning at the point in time in a user interface in which the media content is presented on the electronic device.
14. The electronic device of claim 12, wherein only one of audio or video of the media content is stored in the memory.
15. The electronic device of claim 12, wherein the processor is to: automatically analyze the first and second portions of the media content to generate a text transcript of audio included in the first and second portions of the media content; and store the text transcript with an association to the user input and the context of the media content.
PCT/US2020/056497 2020-10-20 2020-10-20 Storage of remote-presented media content WO2022086507A1 (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
PCT/US2020/056497 WO2022086507A1 (en) 2020-10-20 2020-10-20 Storage of remote-presented media content

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
PCT/US2020/056497 WO2022086507A1 (en) 2020-10-20 2020-10-20 Storage of remote-presented media content

Publications (1)

Publication Number Publication Date
WO2022086507A1 true WO2022086507A1 (en) 2022-04-28

Family

ID=81289340

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/US2020/056497 WO2022086507A1 (en) 2020-10-20 2020-10-20 Storage of remote-presented media content

Country Status (1)

Country Link
WO (1) WO2022086507A1 (en)

Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20070094405A1 (en) * 2005-10-21 2007-04-26 Zhang Xinyan System and method for presenting streaming media content
US20140046955A1 (en) * 2012-08-07 2014-02-13 Google Inc. Media content receiving device and distribution of media content utilizing social networks and social circles
US20150289023A1 (en) * 2014-04-07 2015-10-08 Spotify Ab System and method for providing watch-now functionality in a media content environment
US20170188065A1 (en) * 2015-12-29 2017-06-29 Echostar Technologies L.L.C. Methods and systems for assisted content delivery

Patent Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20070094405A1 (en) * 2005-10-21 2007-04-26 Zhang Xinyan System and method for presenting streaming media content
US20140046955A1 (en) * 2012-08-07 2014-02-13 Google Inc. Media content receiving device and distribution of media content utilizing social networks and social circles
US20150289023A1 (en) * 2014-04-07 2015-10-08 Spotify Ab System and method for providing watch-now functionality in a media content environment
US20170188065A1 (en) * 2015-12-29 2017-06-29 Echostar Technologies L.L.C. Methods and systems for assisted content delivery

Similar Documents

Publication Publication Date Title
WO2021196890A1 (en) Method and device for multimedia processing, electronic device, and storage medium
US10531044B2 (en) Intelligent virtual assistant system and method
US20170371496A1 (en) Rapidly skimmable presentations of web meeting recordings
US10320876B2 (en) Media production system with location-based feature
US10613702B2 (en) Rendering video over multiple windows
US20190246165A1 (en) Messaging and commenting for videos
US20150007055A1 (en) Multi-User Collaboration Tracking Methods and Systems
US11843567B2 (en) Shared reactions within a video communication session
US10062413B2 (en) Media-production system with social media content interface feature
US10990828B2 (en) Key frame extraction, recording, and navigation in collaborative video presentations
WO2024002072A1 (en) Information collection method and apparatus, and electronic device
US11582269B2 (en) Systems and methods for establishing a virtual shared experience for media playback
WO2022086507A1 (en) Storage of remote-presented media content
WO2012141856A1 (en) Augmented reality system for re-casting a seminar with private calculations
KR101328270B1 (en) Annotation method and augmenting video process in video stream for smart tv contents and system thereof
CN112601129B (en) Video interaction system, method and receiving terminal
CN114422745A (en) Method and device for rapidly arranging conference summary of audio and video conference and computer equipment
US20230205808A1 (en) Presentation systems and methods
CN118283379A (en) Method, apparatus, device, medium and product for multimedia content control
WO2018132514A1 (en) Intelligent virtual assistant system and method

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 20958858

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: DE

122 Ep: pct application non-entry in european phase

Ref document number: 20958858

Country of ref document: EP

Kind code of ref document: A1