US20140115621A1 - Intra-Video Ratings - Google Patents
Intra-Video Ratings Download PDFInfo
- Publication number
- US20140115621A1 US20140115621A1 US14/145,710 US201314145710A US2014115621A1 US 20140115621 A1 US20140115621 A1 US 20140115621A1 US 201314145710 A US201314145710 A US 201314145710A US 2014115621 A1 US2014115621 A1 US 2014115621A1
- Authority
- US
- United States
- Prior art keywords
- video
- segment
- timeline
- segments
- user
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Abandoned
Links
Images
Classifications
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B5/00—Measuring for diagnostic purposes; Identification of persons
- A61B5/16—Devices for psychotechnics; Testing reaction times ; Devices for evaluating the psychological state
- A61B5/165—Evaluating the state of mind, e.g. depression, anxiety
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/40—Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
- H04N21/47—End-user applications
- H04N21/475—End-user interface for inputting end-user data, e.g. personal identification number [PIN], preference data
- H04N21/4756—End-user interface for inputting end-user data, e.g. personal identification number [PIN], preference data for rating content, e.g. scoring a recommended movie
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B5/00—Measuring for diagnostic purposes; Identification of persons
- A61B5/02—Detecting, measuring or recording pulse, heart rate, blood pressure or blood flow; Combined pulse/heart-rate/blood pressure determination; Evaluating a cardiovascular condition not otherwise provided for, e.g. using combinations of techniques provided for in this group with electrocardiography or electroauscultation; Heart catheters for measuring blood pressure
- A61B5/0205—Simultaneously evaluating both cardiovascular conditions and different types of body conditions, e.g. heart and respiratory condition
- A61B5/02055—Simultaneously evaluating both cardiovascular condition and temperature
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F16/00—Information retrieval; Database structures therefor; File system structures therefor
- G06F16/70—Information retrieval; Database structures therefor; File system structures therefor of video data
- G06F16/74—Browsing; Visualisation therefor
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F16/00—Information retrieval; Database structures therefor; File system structures therefor
- G06F16/90—Details of database functions independent of the retrieved data types
- G06F16/95—Retrieval from the web
- G06F16/953—Querying, e.g. by the use of web search engines
- G06F16/9535—Search customisation based on user profiles and personalisation
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F16/00—Information retrieval; Database structures therefor; File system structures therefor
- G06F16/90—Details of database functions independent of the retrieved data types
- G06F16/95—Retrieval from the web
- G06F16/958—Organisation or management of web site content, e.g. publishing, maintaining pages or automatic linking
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04H—BROADCAST COMMUNICATION
- H04H60/00—Arrangements for broadcast applications with a direct linking to broadcast information or broadcast space-time; Broadcast-related systems
- H04H60/29—Arrangements for monitoring broadcast services or broadcast-related services
- H04H60/33—Arrangements for monitoring the users' behaviour or opinions
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04H—BROADCAST COMMUNICATION
- H04H60/00—Arrangements for broadcast applications with a direct linking to broadcast information or broadcast space-time; Broadcast-related systems
- H04H60/61—Arrangements for services using the result of monitoring, identification or recognition covered by groups H04H60/29-H04H60/54
- H04H60/65—Arrangements for services using the result of monitoring, identification or recognition covered by groups H04H60/29-H04H60/54 for using the result on users' side
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/20—Servers specifically adapted for the distribution of content, e.g. VOD servers; Operations thereof
- H04N21/25—Management operations performed by the server for facilitating the content distribution or administrating data related to end-users or client devices, e.g. end-user or client device authentication, learning user preferences for recommending movies
- H04N21/251—Learning process for intelligent management, e.g. learning user preferences for recommending movies
- H04N21/252—Processing of multiple end-users' preferences to derive collaborative data
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/40—Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
- H04N21/47—End-user applications
- H04N21/482—End-user interface for program selection
- H04N21/4826—End-user interface for program selection using recommendation lists, e.g. of programs or channels sorted out according to their score
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B5/00—Measuring for diagnostic purposes; Identification of persons
- A61B5/02—Detecting, measuring or recording pulse, heart rate, blood pressure or blood flow; Combined pulse/heart-rate/blood pressure determination; Evaluating a cardiovascular condition not otherwise provided for, e.g. using combinations of techniques provided for in this group with electrocardiography or electroauscultation; Heart catheters for measuring blood pressure
- A61B5/024—Detecting, measuring or recording pulse rate or heart rate
- A61B5/02405—Determining heart rate variability
Definitions
- the present invention generally relates to the management and display of votes associated with video.
- Video hosting websites allow users to upload, view, and rate videos. Users browsing a video hosting website can locate videos of interest by, for example, searching for videos, browsing directories, or sorting by ratings.
- the present invention includes systems and methods for receiving and displaying voting data for a plurality of viewers of online videos hosted in a video hosting service.
- a video hosted on the service is provided to the viewers via their client devices.
- the viewers vote as to their opinion on segments of the videos as they are being displayed on the viewers' devices. For example, the viewers can vote on whether they agree or disagree with the content of each segment.
- Each voter thus provides a series of votes corresponding to the segments of the video. For example, if a 1 minute video has 10 six-second segments, each viewer may cast 10 votes; or the video may have 60 one-second segments in which case a viewer can cast sixty votes.
- segments in a video can have variable length, with some segments being longer or shorter than other segments.
- Each vote has a value that is selected from a unipolar or multipolar set of potential values, such as bipolar (e.g., agree/disagree, favorable/unfavorable, funny/not funny), or categorical votes.
- bipolar e.g., agree/disagree, favorable/unfavorable, funny/not funny
- categorical votes e.g., categorical votes.
- the votes are transmitted from the client devices and received by the video hosting service.
- the votes can be aggregated according to their vote values; this aggregation is preferably, though not necessarily, done while the video is being provided to the client devices.
- the viewers can provide either a favorable or unfavorable vote
- the number of favorable and unfavorable votes for each segment can be determined.
- the aggregated information reveals the aggregate opinion or judgment of the viewers for each segment of the video, and thus how that opinion may have changed throughout the video.
- the video hosting service transmits information about the aggregated votes for the previously displayed segments, such as a graphical or numerical representation (e.g., totals for each vote value), to the viewer's client device, where it can be displayed in a number of ways, such as gradations of color, or charts such as bar graphs.
- the display of the aggregated information is preferably done while the video is still playing. Thus, as the viewer is watching the video and voting, she also sees the aggregated voting information for segments of the video that have already been shown. In one embodiment, the aggregated information is displayed on a timeline representation.
- the voting data may be associated with demographic categories, such as a viewer's age, sex, or political affiliation; the votes can be aggregated according those demographic categories, and information about the demographic aggregation can be provided to the client devices for display thereon. Further, a video may be selectively played to display segments of most interest to given demographic categories, as indicated by voting patterns of those categories.
- a computer-implemented method of associating a series of votes with a video comprises transmitting the video for display to a client device of a viewer, whereby the client device plays back the video for viewing by the viewer, and receiving a first series of votes from the viewer while the video is being played back on the client device, each vote associated with one of the segments of the video and having a vote value, the vote value selected by the viewer from a multipolar set of potential values.
- the method further comprises storing the first series of votes in a storage medium in association with information identifying the video.
- FIG. 1 is a block diagram of a system architecture for receiving and displaying voting data associated with online hosted videos in accordance with one embodiment of the present invention.
- FIG. 2 illustrates steps performed by a client related to voting on videos, according to one embodiment.
- FIGS. 3A-3G illustrate user interfaces for use with voting on videos, according to different embodiments.
- FIG. 1 is a block diagram of a system architecture in accordance with one embodiment.
- a video hosting website 108 includes a front end server 124 , a video server 126 , a network interface 122 , a video database 128 , a video access log 129 , a user database 140 , a vote database 142 , a vote display module 118 , and a vote storage module 144 .
- Other conventional features, such as firewalls, load balancers, application servers, failover servers, site management tools, and so forth are not shown so as to more clearly illustrate the features of the system.
- Examples of a suitable website 108 for implementation of the system include the YouTubeTM and Google VideoTM websites; other video hosting sites are known as well, and can be adapted to operate according the teaching disclosed herein.
- the term “website” represents any system and method of providing content and is not intended to be limited to systems that support content provided via the Internet or the HTTP protocol.
- the various servers are conventionally implemented, whether as a single piece of software or hardware or as multiple pieces of software or hardware and can couple to the network 105 via the network interface 122 . In general, functions described in one embodiment as being performed on the server side can also be performed on the client side in other embodiments if appropriate.
- a client 130 executes a browser 132 , and connects to the front end server 124 via a network 105 , which is typically the Internet, but may also be any network, including but not limited to a LAN, a MAN, a WAN, a mobile, wired or wireless network, a private network, or a virtual private network. While only a single client 130 and browser 132 are shown, it is understood that very large numbers (e.g., millions) of clients are supported and can be in communication with the website 108 at any time.
- the client 130 may include a variety of different computing devices. Examples of client devices 130 are personal computers, digital assistants, personal digital assistants, cellular phones, mobile phones, smart phones or laptop computers. As will be obvious to one of ordinary skill in the art, the present invention is not limited to the devices listed above.
- the browser 132 includes an embedded video player 134 such as, for example, the FlashTM player from Adobe Systems, Inc. or any other player adapted for the video file formats used in the video hosting website 108 .
- an embedded video player 134 such as, for example, the FlashTM player from Adobe Systems, Inc. or any other player adapted for the video file formats used in the video hosting website 108 .
- a user can access a video from the video hosting website 108 by browsing a catalog of videos, conducting searches on keywords, reviewing play lists from other users or the system administrator (e.g., collections of videos forming channels), or viewing videos associated with particular user group (e.g., communities).
- Video server 126 receives uploaded media content, such as videos, from content providers and allows content to be viewed by the client 130 .
- Content providers may be users, commercial enterprises, and the like.
- Content may be uploaded to video server 126 via the Internet from a personal computer, through a cellular network from a telephone or PDA, or by other means for transferring data over network 105 known to those of ordinary skill in the art.
- Content may be downloaded from video server 126 in a similar manner; in one embodiment media content is provided as a file download to a client 130 ; in an alternative embodiment, media content is streamed to client 130 .
- the means by which media content is received by video server 126 need not match the means by which it is delivered to client 130 .
- a content provider may upload a video via a browser on a personal computer, whereas client 130 may view that video as a stream sent to a PDA.
- video server 126 may itself serve as the content provider.
- Users of clients 130 can also search for videos based on keywords, tags or other metadata. These requests are received as queries by the front end server 124 and provided to the video server 126 , which is responsible for searching the video database 128 for videos that satisfy the user queries.
- the video server 126 supports searching on any fielded data for a video, including its title, description, tags, author, category and so forth. User can also browse for videos of interest via categories, channels, lists or other means.
- the uploaded content can include, for example, video, audio or a combination of video and audio.
- the uploaded content is processed and stored in the video database 128 . This processing can include format conversion (transcoding), compression, metadata tagging, and other data processing.
- An uploaded content file is associated with the uploading user, and so the user's account record is updated in the user database 140 as needed.
- Each uploaded video is assigned a video identifier when it is processed.
- the uploaded content will be referred to as “videos”, “video files”, or “video items”, but no limitation on the types of content that can be uploaded are intended by this terminology.
- the video database 128 is used to store the received videos.
- the video database 128 stores video content and associated metadata, provided by their respective content owners.
- the video files have metadata associated with each file such as a video ID, artist, video title, label, genre, and time length.
- the video file data may be viewed as comprising a number of video segments with which votes may be associated.
- segments are coherent portions of the video, e.g. a 5-second clip in a video of a debate of a candidate from a particular viewing angle, as determined by a video segmentation algorithm.
- the segmentation could be determined based on characteristics of the video data itself, such as scene changes, or on voting data, such as that created by clustering votes nearby in time to find the boundaries of the various segments.
- a segment's start and end points can be determined based on, e.g., scene changes, the various segments can have variable lengths with respect to each other.
- a video can be segmented into a set of fixed-length segments, such as 1-second segments, or 10-second segments, or individual frames.
- the user database 140 is responsible for maintaining a record of all users viewing videos on the website. Each individual user is assigned a user ID. The user ID can be based on any identifying information, such as the user's IP address, user name, or the like.
- the user database may also contain information about the reputation of the user in the video context, as well as through other applications, such as the use of email or text messaging.
- a video access log 129 within video database 128 stores each instance of video access. Each entry in the access log 129 identifies a video being accessed, a time of access, an IP address of the user, a user ID if available, cookies, search queries, data identifying the type of interaction with the video, and the time of every interaction with the video. Interaction types can include any user interactions in the user interface of the website 108 , such as playing, pausing, rewinding, forwarding and submitting votes or ratings for a video.
- a vote database 142 stores votes entered by users for the various videos of the video database 128 .
- Votes may be represented in the vote database 142 as a vote value associated with a segment or segments of a video, such as a single frame or segment.
- Vote values may comprise various types of data in different embodiments, representing the different types of votes and what data is tracked.
- a vote value type is “unipolar,” merely representing a single concept, such as interest or favorability, and thus need have no associated value.
- a vote value type is “multipolar,” i.e. having a plurality of potential values.
- multipolar vote is “bipolar,” representing two distinct possibilities, such as “like”/“dislike”, “favorable”/“unfavorable”, “positive”/“negative”, “interesting”/“not interesting”, and the like.
- Another type of multipolar vote is a “categorical” vote, in which, a vote can take on one of a number of values for a given category; as one example, a vote describing the user's subjective emotional reaction to the video could have possible values representing reactions such as “clever”, “boring”, “interesting”, “inspiring”, “frightening” and the like.
- a multipolar vote may represent data from any one of a number of different domains, such as a set of integers, or a continuous range of real numbers.
- the values may be orderable, such as votes representing levels of favorability, such as “poor”, “OK”, “good”, and “great”, or they may not be orderable, such as categories representing different unrelated emotions such as “clever”, “inspiring”, “frightening”, etc.
- Votes also have an associated temporal indicator indicating when the vote was submitted, which allows the derivation of statistics on patterns of voting over time within a given video.
- the temporal indicator may take different forms in different embodiments. For example, in one embodiment the temporal indicator indicates a particular time at which the vote was specified, e.g. a number of seconds into the video. In this case, the video hosting website 108 can map this time to a particular segment, through either a function or a stored map that identifies the times at which the segment boundaries of the video occur.
- the temporal indicator instead is an index of a segment; for example, for an embodiment having fixed length segments of 10 seconds, votes specified 45 seconds and 108 seconds into the video could be given zero-based indices of 5 and 11, indicating that they are in the 5th and 11th of the 10-second segments, respectively.
- the calculation of the index is performed by the client at the time the vote is specified, e.g. by performing integer division of the second offset into the video by the fixed number of seconds per segment.
- the temporal indicators can be specified as absolute values relative to the start of the video as described above, or can be relative offsets from the previous indicator; with the first temporal indicator being an offset from the beginning of the video, either in time (e.g., milliseconds) or frames, and then each following temporal indicator being the amount of time or number of frames from the prior indicator.
- the server can use it to determine a particular segment with which the vote is associated.
- Votes may also have associated data such as demographic categories associated with the user who submitted the vote, e.g., gender, age group, political party affiliation, or other categories of interest. This information allows for even finer grained analysis of voting patterns for a given video.
- a vote storage module 144 stores vote data provided by the client within the vote database 142 .
- the vote storage module 144 can additionally store vote-related information, such as a vote timestamp, vote totals for a given segment of a video, and the like.
- FIG. 2 illustrates steps performed by the client 130 of FIG. 1 .
- the client 130 displays 210 a video. While the video is being displayed, the client 130 receives 220 voting data from the user as inputs to the client 130 . Client 130 also displays 230 the received voting data in association with the video. Note that these steps need not be performed in a linear sequence. For example, prior votes for a video have typically been received 220 by the time that the video is displayed 210 . Further, any prior voting data may be displayed 230 before the receipt of additional voting data at step 220 . These steps are now discussed in further detail in conjunction with FIGS. 3A-3G , which illustrate different aspects of user interfaces for receiving and displaying voting data according to various embodiments.
- a video is displayed 210 .
- the video server 126 of the video hosting website 108 retrieves data for the video from the video database 128 and provides it (e.g., via streaming) to the client 130 , which then displays it, e.g. via the embedded player 134 .
- Also displayed are controls for playback of the video, and also optionally for specifying voting data.
- FIG. 3A illustrates one sample user interface for entering unipolar voting data.
- a video is displayed in video area 305 , and an associated timeline 310 corresponds to the different segments of the video over time, with a marker 315 indicating the current position of the video being displayed.
- the timeline 310 provides a representation of votes over time, using visual indications such as colors, bars, and the like positioned at locations along the timeline to illustrate the vote values.
- the timeline 310 displays color gradients, such as gradient 325 , to depict voting information from prior votes.
- the changes in color show changes in voting across the different video segments, e.g. showing a change in total number of votes for each segment.
- the displayed prior votes may be only those of the viewer currently viewing and voting on the video, or they may include those historical votes previously submitted by other viewers before the viewer began to view the video (e.g.
- votes submitted in the previous 30 days, or some other set time frame may include votes of other viewers who are watching the same video as the viewer, either synchronously (e.g., all viewers are watching a broadcast) or asynchronously (different viewers watching the video but starting at different times), or any combination of the foregoing (e.g., showing the aggregate of historical and current synchronous votes).
- voting data is received 220 at the client 130 from the user.
- Voting can be done in real time while the video is being displayed, i.e. the user inputs a series of votes for a video on the client 130 substantially contemporaneously while the video is being played, so that the user's votes are received at substantially the same time that the user is making a judgment regarding the video content, rather than the user being obliged to halt playback of the video to vote, and/or to specify only a single vote representing an overall impression of the video content.
- each vote is received and stored at the video hosting website 108 substantially immediately after if it is entered at the client, including typical delays associated with transmissions over the network 105 .
- the votes can be first cached at the client 130 , and then transmitted in batch to the website 108 , for example when the user finishes watching the video.
- voting icon 320 may both provide a visual description of the types of votes associated with the video, and also serve as a means of entering votes.
- the voting icon 320 indicates that the votes are unipolar (i.e., that there is only one type of vote, and that it is favorable, as illustrated with the thumbs-up image). Further, clicking on the voting icon 320 specifies a unipolar vote associated with a segment of the video corresponding to the marker 315 at the time the vote is input.
- Votes can be submitted in a number of different manners in different embodiments. For example, a user could submit a vote using a graphical user interface provided by, for example, a web browser, such as clicking on voting icon 320 of FIG. 3 A. Similarly, the user could click on a point on the timeline 310 to designate a unipolar vote for the segment of the video corresponding to the clicked point. Users may also click and drag on the timeline 310 to specify a unipolar vote for a longer interval of video, possibly comprising several distinct segments of the video. Users may also submit votes via a typical input device, such as a keyboard, where the press of a particular key corresponds to a given vote value.
- a typical input device such as a keyboard
- a user could specify a unipolar vote by pressing any key, or a bipolar or categorical vote by pressing a key associated with an accepted value (e.g. “f” or “u”, or “+” or “ ⁇ ”, or up and down arrows for favorable or unfavorable values of a bipolar vote, or “b”, “i”, or “s” to designate “boring”, “interesting”, or “stimulating” in a categorical vote).
- the vote value is then associated with a time corresponding to the position of the marker 315 .
- the client 130 After a user has submitted voting data, the client 130 then communicates the voting data to the vote storage module 144 of the video hosting website 108 , which converts any raw values received into their corresponding semantic values and stores them in the vote database 142 .
- a raw value is the data associated with input action, such as the value “f” of a keypress, and a semantic value describes the meaning associated with the raw value, such as a bipolar vote specifying a favorable value.
- the raw values can be converted to semantic values on the client 130 prior to being transmitted to the video hosting website 108 .
- Voting information specified by other users may or may not be displayed to the user at the time that the user is entering his or her own votes. Displaying votes of other users provides the user with a maximum of information, but runs the risk of introducing bias, leading the user to emulate the votes of the other users. If the potential for bias is a concern, the vote display module 118 can refrain altogether from displaying the data from other users in contexts when the user can vote, or it can display the voting data from other users only at times before the currently playing segment of the video. For example, the user interface of FIG. 3B displays only voting data in timeline 310 of other users associated with times prior to the current play time in the video, as indicated by video playback marker 315 .
- the vote storage module 144 may additionally store whether or not a given vote was specified when the voting data of other users was available; such information could be used to optionally filter out possibly biased data when displaying voting data, or it could be analyzed to estimate the effects of bias on voting.
- the vote storage module 144 additionally associates a temporal indicator, such as a timestamp or a segment index, with the vote.
- a temporal indicator such as a timestamp or a segment index
- This temporal indicator allows determining—directly or indirectly—the segment to which the vote corresponds.
- the timestamp corresponding to the time that the vote was specified on the client 130 , or received by the video hosting website 108 need not correspond to the precise point in the video that prompted the user to submit the vote.
- the vote storage module 144 can apply clustering algorithms to the votes, grouping votes nearby in time into a single common group, and then identifying a segment of the video associated with that group. For example, the vote storage module 144 could calculate an average vote time of the votes in the group, then adjust that time by subtracting a known estimate of the reaction time of the average user, e.g. 1 second. Each vote in the group is then associated with the segment containing that adjusted time, e.g. by associating the vote with the segment index of the segment.
- the vote storage module 144 may also derive and store additional data in association with videos and/or segments of a given video. For example, the vote storage module 144 may compute a total number of votes for a particular segment of a video and store it in association with that segment, updating this stored total when additional votes are specified for that segment by additional users who watch the video and vote on that segment. Alternatively, the vote storage module 144 may compute a total value of the votes for a particular segment, rather than a raw count of the number of votes for that segment.
- weights may be assigned to votes based on, for example, a reputation score of the user, the reputation score being based on factors such as whether the user is new or unrecognized, the usefulness of votes previously submitted by that user as rated by other users, or other information about the user within the user database 140 .
- the vote storage module 144 additionally identifies any other copies of the same video that exist within the video database 128 and aggregates the voting data for the various copies of the video into a single set of data for that video.
- the vote storage module 144 can accomplish the identification of other copies of the video by a variety of means, such as by computing a similarity score between a set of representative feature vectors of each video. Each of the copies of the video is then associated with this aggregate vote data.
- Each vote can also be associated with a set of demographic attributes, such as age, sex, political affiliation, and the like.
- the demographic attributes are specified via a questionnaire provided along with the video, such as that depicted in FIG. 3C , which depicts a questionnaire related to a political video.
- selection menu 360 allows the user to specify one of a set of political parties
- selection menu 365 allows the user to specify a particular candidate.
- Profile update checkboxes 370 allow the user to specify that the specified answers should be reflected in the user's profile on the video hosting website 108 .
- the questionnaire should include some data that specifies how the particular questions of the questionnaire correspond to the fields of the global profile of the video hosting website.
- the questionnaire may encode, for each field to be updated, an identifier that is standard for profiles on the video hosting website 108 .
- the update checkbox 370 for the political party selection menu 360 encodes an identifier specifying that the political party field of the user profile should be updated.
- Checkbox 370 may be omitted for questions not tied to a field of the user profile.
- the questionnaire could be part of the video itself, including interactive elements for specifying and providing input, or it could be a separate interactive multimedia unit that is linked to by the embedded video player 134 on the client 130 if the user provided any votes.
- the demographic attributes need not be explicitly specified by the user as in the interface of FIG. 3C , but can be automatically extracted from the profile of the user if the user has a profile that specifies values for the attributes in question. For example, if the user profile specifies a value of “Independent” for a “political party affiliation” attribute, then that value could be automatically set in selection menu 360 of FIG. 3C , for example, or the interface of FIG. 3C could alternatively be omitted entirely.
- the vote storage module 144 associates each vote received with respect to a video with the provided demographic attributes. For example, if the system-wide identity of the user—such as a user ID—is known, the votes can simply be associated with a system-wide entry for that user identity, the entry comprising the specified demographic attributes or the user ID for subsequent retrieval of the attributes. Alternatively, if no system-wide identity of the user is known, the votes by that user can still be grouped and stored in association with that particular video and with the specified demographic attributes.
- voting data is displayed 230 .
- the timeline 310 may be graphically depicted to illustrate the voting data associated with the video, as is further described below; for example, the timeline of FIG. 3A contains color gradations indicating different numbers of votes for different segments.
- the vote display module 118 provides for the display of an indication of previously stored voting data on a timeline associated with the videos. This provides a user with a visual summary of how users voted across the various segments of a video.
- the voting values can be displayed in different ways in different embodiments. For example, for unipolar votes, times corresponding to the different segments of the video can be colored differently, such as by associating a first color with a maximum vote value and a second color with a minimum vote value and computing a linear blend of the colors for a given segment of the video based on the vote value of that segment.
- a segment with no associated votes could be colored yellow, and a segment having a number (or value) of votes at or exceeding the maximum value could be colored red, and a segment with a vote value between those extremes could be colored an appropriate blend of yellow and red.
- This is depicted in grayscale form in the timeline 310 of FIG. 3A , where darker shades represent the color red, and lighter shades the color yellow.
- the display of the voting data is updated in real time as votes are entered.
- the user's own votes for earlier segments of the video can be displayed substantially immediately after the votes are specified.
- the votes of other users made at substantially the same time e.g. after the user began to view the user interface 300 for the video, can be incorporated into the display.
- the votes of other clients 130 for a given video can be stored in the vote database 142 , and those votes, or updated graphical data reflecting those votes, can then be provided to each of the other clients 130 that is currently viewing that same video, and the display of the clients can be updated accordingly.
- each can be assigned a color and the net value computed for a given segment of the video.
- color gradients 311 - 313 are indicated on the timeline 310 corresponding to segments where the voting is largely favorable, largely unfavorable, and a roughly equal, respectively.
- the colors for the gradients 311 - 313 represent the color associated with favorable icon 321 , the color associated with the unfavorable icon 322 , and a blend thereof, respectively.
- the color of a particular color gradient can be a linear blend of the colors representing the number of favorable and unfavorable votes.
- a segment with 60% favorable votes and 40% unfavorable votes could be represented with a linear blend having 60% red and 40% blue, resulting in a shade of purple.
- the overall color could be that corresponding to the majority of the votes, and the saturation of the color could be based on how strong the majority is. For example, for a segment of the video for which there are 150 red (favorable) votes and 100 blue (unfavorable) votes, the segment could be colored red, with the 50-vote difference between the total of 250 favorable and unfavorable votes used to determine the intensity of the red.
- votes can be represented with a graph, such as the bar charts depicted in FIG. 3D , which comprises one bar chart area 391 representing favorable votes over time, and another bar chart area 392 representing unfavorable votes over time, with longer bars representing more votes for the corresponding segments, and the horizontal position of each bar corresponding to the approximate time of the votes contributing to the bar.
- FIG. 3D also depicts a button 393 for the selection of particular demographics, e.g., filtering the voting data to show just that voting data associated with users of a particular political party.
- such “controversial” segments are detected by computing the ratio of favorable to unfavorable votes, and determining the distance of this ratio from 0.5 (indicating split votes), with distances less than some specified threshold distance (e.g., 0.1) indicating a controversy. In other embodiments, it is additionally preliminarily determined whether there is a sufficient total number of votes (e.g., 100 votes); if not, then it is not deemed important enough to represent a controversy.
- controversies can be difficult to visually detect for some types of visual display, such as color blending, where an equal number of positive and negative values produces the same blended color, regardless of whether votes are unanimously neutral or strongly but equally divided.
- the system can additionally provide a visual indicator to clearly identify such segments of controversy.
- FIG. 3E depicts three star-shaped controversy indicators 394 corresponding to three segments in which there are a large number of both positive and negative votes. Users can then move playback to those controversial segments of the video, e.g. by clicking at those locations or dragging the marker 315 to them.
- Controversies may be shown between different demographic groups, e.g. all voters, as in FIG. 3E , or within a single group.
- FIG. 3F illustrates showing separate voting data and separate controversy indicators 394 on separate timelines 310 for each, e.g. within the Democratic, Independent, and Democratic groups.
- FIG. 3G depicts a user interface for specifying watching a video segmented according to a given demographic attribute or attributes, e.g., political party.
- FIG. 3G shows selection lists 395 and 396 for selecting votes by those of a particular party, and who mostly favor a particular candidate, respectively.
- the selected attribute values are then applied as filters, selecting only votes associated with the selected demographic attribute values.
- the identified votes are then displayed on the timeline 310 , e.g., as a group visually distinct from votes not associated with the selected demographic attribute values.
- votes associated with the selected demographic attribute values may be displayed, or they may be displayed in addition to but separately from other votes, such as on a separate timeline, as in FIG. 3F , where each value of the political party demographic attribute (e.g., Republicans, Independents, Democrats) is depicted on a separate timeline.
- the political party demographic attribute e.g., Republicans, Independents, Democrats
- the client player 134 can be set to play only those segments of the video that are considered of interest to the selected demographics, e.g. those segments having particularly high voting values, or segments that are determined to be controversial.
- the user interface of FIG. 3G could be shown at different times in different embodiments, such as at the start of the video, at the end of the video, or in response to a user request to select segments of interest to given demographics, e.g. via selection of the button 393 of FIG. 3D .
- the client 130 transmits a designation of the selected demographic categories to the video hosting website 108 , which then determines which segments of the video have at least a threshold number of votes that are associated with a demographic attribute value(s) matching the designated demographic attribute value(s), and then provides an indication of the determined segments (e.g., a segment identifier) to the client 130 , which can then display only those segments. For example, if a viewer selected “Independent” as a political party demographic attribute value, then the video hosting website 108 would provide to the client 130 an indication of only the segments with at least a threshold number of votes associated with the Independent political party, e.g. those votes by viewers with profiles specifying that party.
- process steps and instructions of the present invention can be embodied in software, firmware or hardware, and when embodied in software, can be downloaded to reside on and be operated from different platforms used by a variety of operating systems.
- the present invention also relates to an apparatus for performing the operations herein.
- This apparatus may be specially constructed for the required purposes, or it may comprise a general-purpose computer selectively activated or reconfigured by a computer program stored in the computer.
- a computer program may be stored in a computer readable storage medium, such as, but is not limited to, any type of disk including floppy disks, optical disks, CD-ROMs, magnetic-optical disks, read-only memories (ROMs), random access memories (RAMs), EPROMs, EEPROMs, magnetic or optical cards, application specific integrated circuits (ASICs), or any type of media suitable for storing electronic instructions, and each coupled to a computer system bus.
- the computers referred to in the specification may include a single processor or may be architectures employing multiple processor designs for increased computing capability.
Landscapes
- Engineering & Computer Science (AREA)
- Databases & Information Systems (AREA)
- Health & Medical Sciences (AREA)
- Theoretical Computer Science (AREA)
- Multimedia (AREA)
- Signal Processing (AREA)
- Physics & Mathematics (AREA)
- Life Sciences & Earth Sciences (AREA)
- Human Computer Interaction (AREA)
- Data Mining & Analysis (AREA)
- General Engineering & Computer Science (AREA)
- General Physics & Mathematics (AREA)
- General Health & Medical Sciences (AREA)
- Cardiology (AREA)
- Social Psychology (AREA)
- Pathology (AREA)
- Veterinary Medicine (AREA)
- Biophysics (AREA)
- Physiology (AREA)
- Biomedical Technology (AREA)
- Heart & Thoracic Surgery (AREA)
- Medical Informatics (AREA)
- Molecular Biology (AREA)
- Surgery (AREA)
- Animal Behavior & Ethology (AREA)
- Public Health (AREA)
- Psychiatry (AREA)
- Computing Systems (AREA)
- Software Systems (AREA)
- Pulmonology (AREA)
- Child & Adolescent Psychology (AREA)
- Developmental Disabilities (AREA)
- Educational Technology (AREA)
- Hospice & Palliative Care (AREA)
- Psychology (AREA)
- Two-Way Televisions, Distribution Of Moving Picture Or The Like (AREA)
Abstract
A system and method is provided for receiving and displaying voting data with respect to online hosted videos. Viewers vote on portions within various videos stored in an online video hosting website. Voting data can represent different types of votes, such as unipolar, bipolar, or categorical. The prior voting data, including that submitted by other viewers, can be displayed in a number of ways, such as gradations of color, or charts such as bar graphs. The voting data may be associated with demographic categories, such as a voter's age, sex, or political affiliation. Further, a video may be selectively played to display portions of most interest to given demographic categories, as indicated by voting patterns of those categories.
Description
- This application is a continuation of U.S. patent application Ser. No. 13/363,358, filed on Jan. 31, 2012, which in turn is a continuation of U.S. patent application Ser. No. 12/414,553 (now U.S. Pat. No. 8,132,200), filed on Mar. 30, 2009, both of which are incorporated herein by reference.
- The present invention generally relates to the management and display of votes associated with video.
- The proliferation of video sharing through websites provides numerous opportunities for users to collaborate and experience videos in online communities. Video hosting websites allow users to upload, view, and rate videos. Users browsing a video hosting website can locate videos of interest by, for example, searching for videos, browsing directories, or sorting by ratings.
- User rating of videos provides a way to complement video with useful information. However, because conventional systems allow user rating only of a video as a whole, it cannot be determined which segments of the video the users found particularly interesting. This is especially problematic in light of increasingly large video length maximums allowed by video sharing websites, which leads to the presence of long videos, only a small portion of which may be interesting to most users. In addition, there is no way to navigate videos to find the most popular segments of a video. Although a user may wish to view only the most popular segments of the video, the user is obliged to view the entire video because the most popular parts of the video have not been determined. This results in considerable waste of time and user frustration.
- The present invention includes systems and methods for receiving and displaying voting data for a plurality of viewers of online videos hosted in a video hosting service. In one embodiment, a video hosted on the service is provided to the viewers via their client devices. The viewers vote as to their opinion on segments of the videos as they are being displayed on the viewers' devices. For example, the viewers can vote on whether they agree or disagree with the content of each segment. Each voter thus provides a series of votes corresponding to the segments of the video. For example, if a 1 minute video has 10 six-second segments, each viewer may cast 10 votes; or the video may have 60 one-second segments in which case a viewer can cast sixty votes. Alternatively, segments in a video can have variable length, with some segments being longer or shorter than other segments.
- Each vote has a value that is selected from a unipolar or multipolar set of potential values, such as bipolar (e.g., agree/disagree, favorable/unfavorable, funny/not funny), or categorical votes.
- The votes are transmitted from the client devices and received by the video hosting service. For each segment of the video, the votes can be aggregated according to their vote values; this aggregation is preferably, though not necessarily, done while the video is being provided to the client devices. For example, where the viewers can provide either a favorable or unfavorable vote, the number of favorable and unfavorable votes for each segment can be determined. The aggregated information reveals the aggregate opinion or judgment of the viewers for each segment of the video, and thus how that opinion may have changed throughout the video.
- The video hosting service transmits information about the aggregated votes for the previously displayed segments, such as a graphical or numerical representation (e.g., totals for each vote value), to the viewer's client device, where it can be displayed in a number of ways, such as gradations of color, or charts such as bar graphs. The display of the aggregated information is preferably done while the video is still playing. Thus, as the viewer is watching the video and voting, she also sees the aggregated voting information for segments of the video that have already been shown. In one embodiment, the aggregated information is displayed on a timeline representation.
- The voting data may be associated with demographic categories, such as a viewer's age, sex, or political affiliation; the votes can be aggregated according those demographic categories, and information about the demographic aggregation can be provided to the client devices for display thereon. Further, a video may be selectively played to display segments of most interest to given demographic categories, as indicated by voting patterns of those categories.
- In one embodiment, a computer-implemented method of associating a series of votes with a video, wherein the video has a plurality of segments, comprises transmitting the video for display to a client device of a viewer, whereby the client device plays back the video for viewing by the viewer, and receiving a first series of votes from the viewer while the video is being played back on the client device, each vote associated with one of the segments of the video and having a vote value, the vote value selected by the viewer from a multipolar set of potential values. The method further comprises storing the first series of votes in a storage medium in association with information identifying the video.
- The features and advantages described in this summary and the following detailed description are not all-inclusive. Many additional features and advantages will be apparent to one of ordinary skill in the art in view of the drawings, specification, and claims presented herein.
-
FIG. 1 is a block diagram of a system architecture for receiving and displaying voting data associated with online hosted videos in accordance with one embodiment of the present invention. -
FIG. 2 illustrates steps performed by a client related to voting on videos, according to one embodiment. -
FIGS. 3A-3G illustrate user interfaces for use with voting on videos, according to different embodiments. - The figures depict various embodiments of the present invention for purposes of illustration only. One skilled in the art will readily recognize from the following discussion that alternative embodiments of the structures and methods illustrated herein may be employed without departing from the principles of the invention described herein.
-
FIG. 1 is a block diagram of a system architecture in accordance with one embodiment. As illustrated inFIG. 1 , avideo hosting website 108 includes afront end server 124, avideo server 126, anetwork interface 122, avideo database 128, avideo access log 129, auser database 140, avote database 142, avote display module 118, and avote storage module 144. Other conventional features, such as firewalls, load balancers, application servers, failover servers, site management tools, and so forth are not shown so as to more clearly illustrate the features of the system. Examples of asuitable website 108 for implementation of the system include the YouTube™ and Google Video™ websites; other video hosting sites are known as well, and can be adapted to operate according the teaching disclosed herein. It will be understood that the term “website” represents any system and method of providing content and is not intended to be limited to systems that support content provided via the Internet or the HTTP protocol. The various servers are conventionally implemented, whether as a single piece of software or hardware or as multiple pieces of software or hardware and can couple to thenetwork 105 via thenetwork interface 122. In general, functions described in one embodiment as being performed on the server side can also be performed on the client side in other embodiments if appropriate. - A
client 130 executes abrowser 132, and connects to thefront end server 124 via anetwork 105, which is typically the Internet, but may also be any network, including but not limited to a LAN, a MAN, a WAN, a mobile, wired or wireless network, a private network, or a virtual private network. While only asingle client 130 andbrowser 132 are shown, it is understood that very large numbers (e.g., millions) of clients are supported and can be in communication with thewebsite 108 at any time. Theclient 130 may include a variety of different computing devices. Examples ofclient devices 130 are personal computers, digital assistants, personal digital assistants, cellular phones, mobile phones, smart phones or laptop computers. As will be obvious to one of ordinary skill in the art, the present invention is not limited to the devices listed above. - In some embodiments, the
browser 132 includes an embeddedvideo player 134 such as, for example, the Flash™ player from Adobe Systems, Inc. or any other player adapted for the video file formats used in thevideo hosting website 108. A user can access a video from thevideo hosting website 108 by browsing a catalog of videos, conducting searches on keywords, reviewing play lists from other users or the system administrator (e.g., collections of videos forming channels), or viewing videos associated with particular user group (e.g., communities). -
Video server 126 receives uploaded media content, such as videos, from content providers and allows content to be viewed by theclient 130. Content providers may be users, commercial enterprises, and the like. Content may be uploaded tovideo server 126 via the Internet from a personal computer, through a cellular network from a telephone or PDA, or by other means for transferring data overnetwork 105 known to those of ordinary skill in the art. Content may be downloaded fromvideo server 126 in a similar manner; in one embodiment media content is provided as a file download to aclient 130; in an alternative embodiment, media content is streamed toclient 130. The means by which media content is received byvideo server 126 need not match the means by which it is delivered toclient 130. For example, a content provider may upload a video via a browser on a personal computer, whereasclient 130 may view that video as a stream sent to a PDA. Note also thatvideo server 126 may itself serve as the content provider. - Users of
clients 130 can also search for videos based on keywords, tags or other metadata. These requests are received as queries by thefront end server 124 and provided to thevideo server 126, which is responsible for searching thevideo database 128 for videos that satisfy the user queries. Thevideo server 126 supports searching on any fielded data for a video, including its title, description, tags, author, category and so forth. User can also browse for videos of interest via categories, channels, lists or other means. - Users of the
clients 130 andbrowser 132 can upload content to thevideo hosting website 108 vianetwork 105. The uploaded content can include, for example, video, audio or a combination of video and audio. The uploaded content is processed and stored in thevideo database 128. This processing can include format conversion (transcoding), compression, metadata tagging, and other data processing. An uploaded content file is associated with the uploading user, and so the user's account record is updated in theuser database 140 as needed. Each uploaded video is assigned a video identifier when it is processed. - For purposes of convenience and the description of one embodiment, the uploaded content will be referred to as “videos”, “video files”, or “video items”, but no limitation on the types of content that can be uploaded are intended by this terminology.
- The
video database 128 is used to store the received videos. Thevideo database 128 stores video content and associated metadata, provided by their respective content owners. The video files have metadata associated with each file such as a video ID, artist, video title, label, genre, and time length. The video file data may be viewed as comprising a number of video segments with which votes may be associated. For example, in one embodiment segments are coherent portions of the video, e.g. a 5-second clip in a video of a debate of a candidate from a particular viewing angle, as determined by a video segmentation algorithm. The segmentation could be determined based on characteristics of the video data itself, such as scene changes, or on voting data, such as that created by clustering votes nearby in time to find the boundaries of the various segments. Since a segment's start and end points can be determined based on, e.g., scene changes, the various segments can have variable lengths with respect to each other. In another embodiment, a video can be segmented into a set of fixed-length segments, such as 1-second segments, or 10-second segments, or individual frames. - Information about the votes received is stored in the
user database 140. Theuser database 140 is responsible for maintaining a record of all users viewing videos on the website. Each individual user is assigned a user ID. The user ID can be based on any identifying information, such as the user's IP address, user name, or the like. The user database may also contain information about the reputation of the user in the video context, as well as through other applications, such as the use of email or text messaging. - A
video access log 129 withinvideo database 128 stores each instance of video access. Each entry in the access log 129 identifies a video being accessed, a time of access, an IP address of the user, a user ID if available, cookies, search queries, data identifying the type of interaction with the video, and the time of every interaction with the video. Interaction types can include any user interactions in the user interface of thewebsite 108, such as playing, pausing, rewinding, forwarding and submitting votes or ratings for a video. - A
vote database 142 stores votes entered by users for the various videos of thevideo database 128. Votes may be represented in thevote database 142 as a vote value associated with a segment or segments of a video, such as a single frame or segment. Vote values may comprise various types of data in different embodiments, representing the different types of votes and what data is tracked. In one embodiment, a vote value type is “unipolar,” merely representing a single concept, such as interest or favorability, and thus need have no associated value. In other embodiments, a vote value type is “multipolar,” i.e. having a plurality of potential values. One type of multipolar vote is “bipolar,” representing two distinct possibilities, such as “like”/“dislike”, “favorable”/“unfavorable”, “positive”/“negative”, “interesting”/“not interesting”, and the like. Another type of multipolar vote is a “categorical” vote, in which, a vote can take on one of a number of values for a given category; as one example, a vote describing the user's subjective emotional reaction to the video could have possible values representing reactions such as “clever”, “boring”, “interesting”, “inspiring”, “frightening” and the like. In general, a multipolar vote may represent data from any one of a number of different domains, such as a set of integers, or a continuous range of real numbers. The values may be orderable, such as votes representing levels of favorability, such as “poor”, “OK”, “good”, and “great”, or they may not be orderable, such as categories representing different unrelated emotions such as “clever”, “inspiring”, “frightening”, etc. - Votes also have an associated temporal indicator indicating when the vote was submitted, which allows the derivation of statistics on patterns of voting over time within a given video. The temporal indicator may take different forms in different embodiments. For example, in one embodiment the temporal indicator indicates a particular time at which the vote was specified, e.g. a number of seconds into the video. In this case, the
video hosting website 108 can map this time to a particular segment, through either a function or a stored map that identifies the times at which the segment boundaries of the video occur. - In another embodiment, the temporal indicator instead is an index of a segment; for example, for an embodiment having fixed length segments of 10 seconds, votes specified 45 seconds and 108 seconds into the video could be given zero-based indices of 5 and 11, indicating that they are in the 5th and 11th of the 10-second segments, respectively. In one embodiment, the calculation of the index is performed by the client at the time the vote is specified, e.g. by performing integer division of the second offset into the video by the fixed number of seconds per segment.
- The temporal indicators can be specified as absolute values relative to the start of the video as described above, or can be relative offsets from the previous indicator; with the first temporal indicator being an offset from the beginning of the video, either in time (e.g., milliseconds) or frames, and then each following temporal indicator being the amount of time or number of frames from the prior indicator.
- Regardless of which type of temporal identifier is provided to the
video hosting website 108, the server can use it to determine a particular segment with which the vote is associated. - Votes may also have associated data such as demographic categories associated with the user who submitted the vote, e.g., gender, age group, political party affiliation, or other categories of interest. This information allows for even finer grained analysis of voting patterns for a given video.
- A
vote storage module 144 stores vote data provided by the client within thevote database 142. Thevote storage module 144 can additionally store vote-related information, such as a vote timestamp, vote totals for a given segment of a video, and the like. -
FIG. 2 illustrates steps performed by theclient 130 ofFIG. 1 . At one point, theclient 130 displays 210 a video. While the video is being displayed, theclient 130 receives 220 voting data from the user as inputs to theclient 130.Client 130 also displays 230 the received voting data in association with the video. Note that these steps need not be performed in a linear sequence. For example, prior votes for a video have typically been received 220 by the time that the video is displayed 210. Further, any prior voting data may be displayed 230 before the receipt of additional voting data atstep 220. These steps are now discussed in further detail in conjunction withFIGS. 3A-3G , which illustrate different aspects of user interfaces for receiving and displaying voting data according to various embodiments. - Referring again to
FIG. 2 , a video is displayed 210. Thevideo server 126 of thevideo hosting website 108 retrieves data for the video from thevideo database 128 and provides it (e.g., via streaming) to theclient 130, which then displays it, e.g. via the embeddedplayer 134. Also displayed are controls for playback of the video, and also optionally for specifying voting data. For example,FIG. 3A illustrates one sample user interface for entering unipolar voting data. A video is displayed invideo area 305, and an associatedtimeline 310 corresponds to the different segments of the video over time, with amarker 315 indicating the current position of the video being displayed. In general, thetimeline 310 provides a representation of votes over time, using visual indications such as colors, bars, and the like positioned at locations along the timeline to illustrate the vote values. InFIG. 3A , thetimeline 310 displays color gradients, such asgradient 325, to depict voting information from prior votes. The changes in color (depicted in black and white inFIG. 3A as shades of gray of greater or lesser intensity) show changes in voting across the different video segments, e.g. showing a change in total number of votes for each segment. The displayed prior votes may be only those of the viewer currently viewing and voting on the video, or they may include those historical votes previously submitted by other viewers before the viewer began to view the video (e.g. votes submitted in the previous 30 days, or some other set time frame), and/or they may include votes of other viewers who are watching the same video as the viewer, either synchronously (e.g., all viewers are watching a broadcast) or asynchronously (different viewers watching the video but starting at different times), or any combination of the foregoing (e.g., showing the aggregate of historical and current synchronous votes). - Referring back to
FIG. 2 , while the video is being displayed, voting data is received 220 at theclient 130 from the user. (Such voting data may be in addition to voting data received at a previous time.) Voting can be done in real time while the video is being displayed, i.e. the user inputs a series of votes for a video on theclient 130 substantially contemporaneously while the video is being played, so that the user's votes are received at substantially the same time that the user is making a judgment regarding the video content, rather than the user being obliged to halt playback of the video to vote, and/or to specify only a single vote representing an overall impression of the video content. Further, in one embodiment, each vote is received and stored at thevideo hosting website 108 substantially immediately after if it is entered at the client, including typical delays associated with transmissions over thenetwork 105. Alternatively, the votes can be first cached at theclient 130, and then transmitted in batch to thewebsite 108, for example when the user finishes watching the video. - In
FIG. 3A , for example, votingicon 320 may both provide a visual description of the types of votes associated with the video, and also serve as a means of entering votes. For example, thevoting icon 320 indicates that the votes are unipolar (i.e., that there is only one type of vote, and that it is favorable, as illustrated with the thumbs-up image). Further, clicking on thevoting icon 320 specifies a unipolar vote associated with a segment of the video corresponding to themarker 315 at the time the vote is input. - Votes can be submitted in a number of different manners in different embodiments. For example, a user could submit a vote using a graphical user interface provided by, for example, a web browser, such as clicking on
voting icon 320 of FIG. 3A. Similarly, the user could click on a point on thetimeline 310 to designate a unipolar vote for the segment of the video corresponding to the clicked point. Users may also click and drag on thetimeline 310 to specify a unipolar vote for a longer interval of video, possibly comprising several distinct segments of the video. Users may also submit votes via a typical input device, such as a keyboard, where the press of a particular key corresponds to a given vote value. For example, a user could specify a unipolar vote by pressing any key, or a bipolar or categorical vote by pressing a key associated with an accepted value (e.g. “f” or “u”, or “+” or “−”, or up and down arrows for favorable or unfavorable values of a bipolar vote, or “b”, “i”, or “s” to designate “boring”, “interesting”, or “stimulating” in a categorical vote). The vote value is then associated with a time corresponding to the position of themarker 315. - After a user has submitted voting data, the
client 130 then communicates the voting data to thevote storage module 144 of thevideo hosting website 108, which converts any raw values received into their corresponding semantic values and stores them in thevote database 142. A raw value is the data associated with input action, such as the value “f” of a keypress, and a semantic value describes the meaning associated with the raw value, such as a bipolar vote specifying a favorable value. Alternatively, the raw values can be converted to semantic values on theclient 130 prior to being transmitted to thevideo hosting website 108. - Voting information specified by other users may or may not be displayed to the user at the time that the user is entering his or her own votes. Displaying votes of other users provides the user with a maximum of information, but runs the risk of introducing bias, leading the user to emulate the votes of the other users. If the potential for bias is a concern, the
vote display module 118 can refrain altogether from displaying the data from other users in contexts when the user can vote, or it can display the voting data from other users only at times before the currently playing segment of the video. For example, the user interface ofFIG. 3B displays only voting data intimeline 310 of other users associated with times prior to the current play time in the video, as indicated byvideo playback marker 315. Thevote storage module 144 may additionally store whether or not a given vote was specified when the voting data of other users was available; such information could be used to optionally filter out possibly biased data when displaying voting data, or it could be analyzed to estimate the effects of bias on voting. - In one embodiment, the
vote storage module 144 additionally associates a temporal indicator, such as a timestamp or a segment index, with the vote. This temporal indicator allows determining—directly or indirectly—the segment to which the vote corresponds. With regard to determining a segment index based on a timestamp, note that the timestamp corresponding to the time that the vote was specified on theclient 130, or received by thevideo hosting website 108, need not correspond to the precise point in the video that prompted the user to submit the vote. For example, there will be some delay in the user's mental processes between the time at which the point of interest was played and when the user decides to vote on it, and some delay in the user's physical actions between the time that the user decides to vote on it and when he or she actually submits the vote, and these delays will vary among different users. However, as long as there is a large amount of vote data available, thevote storage module 144 can apply clustering algorithms to the votes, grouping votes nearby in time into a single common group, and then identifying a segment of the video associated with that group. For example, thevote storage module 144 could calculate an average vote time of the votes in the group, then adjust that time by subtracting a known estimate of the reaction time of the average user, e.g. 1 second. Each vote in the group is then associated with the segment containing that adjusted time, e.g. by associating the vote with the segment index of the segment. - The
vote storage module 144 may also derive and store additional data in association with videos and/or segments of a given video. For example, thevote storage module 144 may compute a total number of votes for a particular segment of a video and store it in association with that segment, updating this stored total when additional votes are specified for that segment by additional users who watch the video and vote on that segment. Alternatively, thevote storage module 144 may compute a total value of the votes for a particular segment, rather than a raw count of the number of votes for that segment. For example, weights may be assigned to votes based on, for example, a reputation score of the user, the reputation score being based on factors such as whether the user is new or unrecognized, the usefulness of votes previously submitted by that user as rated by other users, or other information about the user within theuser database 140. - Since multiple copies of a video may exist in the
video database 128, each of which could receive separate voting data, voting data can become fragmented across different copies of the video, with some users viewing and voting on one particular copy, and other users viewing and voting on another copy. Thus, in one embodiment thevote storage module 144 additionally identifies any other copies of the same video that exist within thevideo database 128 and aggregates the voting data for the various copies of the video into a single set of data for that video. Thevote storage module 144 can accomplish the identification of other copies of the video by a variety of means, such as by computing a similarity score between a set of representative feature vectors of each video. Each of the copies of the video is then associated with this aggregate vote data. - Each vote can also be associated with a set of demographic attributes, such as age, sex, political affiliation, and the like. In one embodiment, the demographic attributes are specified via a questionnaire provided along with the video, such as that depicted in
FIG. 3C , which depicts a questionnaire related to a political video. InFIG. 3C ,selection menu 360 allows the user to specify one of a set of political parties, andselection menu 365 allows the user to specify a particular candidate. Profile updatecheckboxes 370 allow the user to specify that the specified answers should be reflected in the user's profile on thevideo hosting website 108. Since the questionnaire is video-specific and can be created by the owner of the video, as well as by thevideo hosting website 108 itself, the questionnaire should include some data that specifies how the particular questions of the questionnaire correspond to the fields of the global profile of the video hosting website. Thus, the questionnaire may encode, for each field to be updated, an identifier that is standard for profiles on thevideo hosting website 108. For example, theupdate checkbox 370 for the politicalparty selection menu 360 encodes an identifier specifying that the political party field of the user profile should be updated.Checkbox 370 may be omitted for questions not tied to a field of the user profile. The questionnaire could be part of the video itself, including interactive elements for specifying and providing input, or it could be a separate interactive multimedia unit that is linked to by the embeddedvideo player 134 on theclient 130 if the user provided any votes. - In another embodiment, the demographic attributes need not be explicitly specified by the user as in the interface of
FIG. 3C , but can be automatically extracted from the profile of the user if the user has a profile that specifies values for the attributes in question. For example, if the user profile specifies a value of “Independent” for a “political party affiliation” attribute, then that value could be automatically set inselection menu 360 ofFIG. 3C , for example, or the interface ofFIG. 3C could alternatively be omitted entirely. - Regardless of the source of the attributes, the
vote storage module 144 associates each vote received with respect to a video with the provided demographic attributes. For example, if the system-wide identity of the user—such as a user ID—is known, the votes can simply be associated with a system-wide entry for that user identity, the entry comprising the specified demographic attributes or the user ID for subsequent retrieval of the attributes. Alternatively, if no system-wide identity of the user is known, the votes by that user can still be grouped and stored in association with that particular video and with the specified demographic attributes. - Referring again back to
FIG. 2 , voting data is displayed 230. InFIG. 3A , for example, thetimeline 310 may be graphically depicted to illustrate the voting data associated with the video, as is further described below; for example, the timeline ofFIG. 3A contains color gradations indicating different numbers of votes for different segments. - The
vote display module 118 provides for the display of an indication of previously stored voting data on a timeline associated with the videos. This provides a user with a visual summary of how users voted across the various segments of a video. The voting values can be displayed in different ways in different embodiments. For example, for unipolar votes, times corresponding to the different segments of the video can be colored differently, such as by associating a first color with a maximum vote value and a second color with a minimum vote value and computing a linear blend of the colors for a given segment of the video based on the vote value of that segment. For instance, a segment with no associated votes could be colored yellow, and a segment having a number (or value) of votes at or exceeding the maximum value could be colored red, and a segment with a vote value between those extremes could be colored an appropriate blend of yellow and red. This is depicted in grayscale form in thetimeline 310 ofFIG. 3A , where darker shades represent the color red, and lighter shades the color yellow. - In one embodiment, the display of the voting data, such as the
timeline 310, is updated in real time as votes are entered. For example, the user's own votes for earlier segments of the video can be displayed substantially immediately after the votes are specified. Additionally, in some embodiments the votes of other users made at substantially the same time, e.g. after the user began to view theuser interface 300 for the video, can be incorporated into the display. For example, the votes ofother clients 130 for a given video can be stored in thevote database 142, and those votes, or updated graphical data reflecting those votes, can then be provided to each of theother clients 130 that is currently viewing that same video, and the display of the clients can be updated accordingly. - In the case of bipolar voting, such as favorable versus unfavorable votes, each can be assigned a color and the net value computed for a given segment of the video. For example, referring back to
FIG. 3B , color gradients 311-313 are indicated on thetimeline 310 corresponding to segments where the voting is largely favorable, largely unfavorable, and a roughly equal, respectively. Although represented in grayscale inFIG. 3B , the colors for the gradients 311-313 represent the color associated withfavorable icon 321, the color associated with theunfavorable icon 322, and a blend thereof, respectively. The color of a particular color gradient can be a linear blend of the colors representing the number of favorable and unfavorable votes. For example, if favorable votes are represented by the color red, and unfavorable votes by the color blue, then a segment with 60% favorable votes and 40% unfavorable votes could be represented with a linear blend having 60% red and 40% blue, resulting in a shade of purple. Alternatively, the overall color could be that corresponding to the majority of the votes, and the saturation of the color could be based on how strong the majority is. For example, for a segment of the video for which there are 150 red (favorable) votes and 100 blue (unfavorable) votes, the segment could be colored red, with the 50-vote difference between the total of 250 favorable and unfavorable votes used to determine the intensity of the red. As an alternative example, votes can be represented with a graph, such as the bar charts depicted inFIG. 3D , which comprises onebar chart area 391 representing favorable votes over time, and anotherbar chart area 392 representing unfavorable votes over time, with longer bars representing more votes for the corresponding segments, and the horizontal position of each bar corresponding to the approximate time of the votes contributing to the bar.FIG. 3D also depicts abutton 393 for the selection of particular demographics, e.g., filtering the voting data to show just that voting data associated with users of a particular political party. - In general, there can be video segments for which there are strong differences of opinion among the users viewing the video, such as a large number of both favorable and unfavorable votes for the same segment in the case of bipolar voting; this could indicate an area of controversy or dispute. For example, this might occur in the context of a video of a political debate, with Democratic viewers voting strongly in favor of a particular segment of the video (e.g., in which a Democratic candidate made a statement appealing to Democratic viewers), and Republican viewers voting strongly against it. In general, controversies are detected by quantifying a level of divergence among the votes for the segment and comparing the quantified divergence to a threshold value. In one embodiment for bipolar voting, such “controversial” segments are detected by computing the ratio of favorable to unfavorable votes, and determining the distance of this ratio from 0.5 (indicating split votes), with distances less than some specified threshold distance (e.g., 0.1) indicating a controversy. In other embodiments, it is additionally preliminarily determined whether there is a sufficient total number of votes (e.g., 100 votes); if not, then it is not deemed important enough to represent a controversy. Such controversies can be difficult to visually detect for some types of visual display, such as color blending, where an equal number of positive and negative values produces the same blended color, regardless of whether votes are unanimously neutral or strongly but equally divided. Thus, in some embodiments the system can additionally provide a visual indicator to clearly identify such segments of controversy. For example,
FIG. 3E depicts three star-shapedcontroversy indicators 394 corresponding to three segments in which there are a large number of both positive and negative votes. Users can then move playback to those controversial segments of the video, e.g. by clicking at those locations or dragging themarker 315 to them. - Controversies may be shown between different demographic groups, e.g. all voters, as in
FIG. 3E , or within a single group. For example,FIG. 3F illustrates showing separate voting data andseparate controversy indicators 394 onseparate timelines 310 for each, e.g. within the Republican, Independent, and Democratic groups. - The previously stored voting data can also be filtered and used to focus on segments of interest within the video. For example,
FIG. 3G depicts a user interface for specifying watching a video segmented according to a given demographic attribute or attributes, e.g., political party. Specifically,FIG. 3G shows selection lists 395 and 396 for selecting votes by those of a particular party, and who mostly favor a particular candidate, respectively. In response to, for instance, selection ofbutton 397, the selected attribute values are then applied as filters, selecting only votes associated with the selected demographic attribute values. The identified votes are then displayed on thetimeline 310, e.g., as a group visually distinct from votes not associated with the selected demographic attribute values. For example, only the votes associated with the selected demographic attribute values may be displayed, or they may be displayed in addition to but separately from other votes, such as on a separate timeline, as inFIG. 3F , where each value of the political party demographic attribute (e.g., Republicans, Independents, Democrats) is depicted on a separate timeline. - Additionally, the
client player 134 can be set to play only those segments of the video that are considered of interest to the selected demographics, e.g. those segments having particularly high voting values, or segments that are determined to be controversial. The user interface ofFIG. 3G could be shown at different times in different embodiments, such as at the start of the video, at the end of the video, or in response to a user request to select segments of interest to given demographics, e.g. via selection of thebutton 393 ofFIG. 3D . In one embodiment, theclient 130 transmits a designation of the selected demographic categories to thevideo hosting website 108, which then determines which segments of the video have at least a threshold number of votes that are associated with a demographic attribute value(s) matching the designated demographic attribute value(s), and then provides an indication of the determined segments (e.g., a segment identifier) to theclient 130, which can then display only those segments. For example, if a viewer selected “Independent” as a political party demographic attribute value, then thevideo hosting website 108 would provide to theclient 130 an indication of only the segments with at least a threshold number of votes associated with the Independent political party, e.g. those votes by viewers with profiles specifying that party. - Reference in the specification to “one embodiment” or to “an embodiment” means that a particular feature, structure, or characteristic described in connection with the embodiments is included in at least one embodiment of the invention. The appearances of the phrase “in one embodiment” in various places in the specification are not necessarily all referring to the same embodiment.
- It should be noted that the process steps and instructions of the present invention can be embodied in software, firmware or hardware, and when embodied in software, can be downloaded to reside on and be operated from different platforms used by a variety of operating systems.
- The present invention also relates to an apparatus for performing the operations herein. This apparatus may be specially constructed for the required purposes, or it may comprise a general-purpose computer selectively activated or reconfigured by a computer program stored in the computer. Such a computer program may be stored in a computer readable storage medium, such as, but is not limited to, any type of disk including floppy disks, optical disks, CD-ROMs, magnetic-optical disks, read-only memories (ROMs), random access memories (RAMs), EPROMs, EEPROMs, magnetic or optical cards, application specific integrated circuits (ASICs), or any type of media suitable for storing electronic instructions, and each coupled to a computer system bus. Furthermore, the computers referred to in the specification may include a single processor or may be architectures employing multiple processor designs for increased computing capability.
- The algorithms and displays presented herein are not inherently related to any particular computer or other apparatus. Various general-purpose systems may also be used with programs in accordance with the teachings herein, or it may prove convenient to construct more specialized apparatus to perform the required method steps. The required structure for a variety of these systems will appear from the description below. In addition, the present invention is not described with reference to any particular programming language. It will be appreciated that a variety of programming languages may be used to implement the teachings of the present invention as described herein, and any references below to specific languages are provided for disclosure of enablement and best mode of the present invention.
- While the invention has been particularly shown and described with reference to a preferred embodiment and several alternate embodiments, it will be understood by persons skilled in the relevant art that various changes in form and details can be made therein without departing from the spirit and scope of the invention.
- Finally, it should be noted that the language used in the specification has been principally selected for readability and instructional purposes, and may not have been selected to delineate or circumscribe the inventive subject matter. Accordingly, the disclosure of the present invention is intended to be illustrative, but not limiting, of the scope of the invention, which is set forth in the following claims.
Claims (20)
1. A graphical display, comprising:
a playback portion configured to display a video comprising a plurality of segments to a user;
a timeline representing the video;
a marker indicating a segment of the video currently displayed in the playback portion; and
a visual indication of a popularity of at least some of the segments with other users.
2. The graphical display of claim 1 , further comprising a favorable vote icon for the user to select in order to indicate a favorable opinion of the segment of the video currently displayed in the playback portion.
3. The graphical display of claim 2 , further comprising an unfavorable vote icon for the user to select in order to indicate an unfavorable opinion of the segment of the video currently displayed in the playback portion.
4. The graphical display of claim 1 , wherein the visual indication of the popularity of at least some of the segments with other users comprises an intensity of shading of portions of the timeline, the intensity of shading of a portion indicating the popularity of the corresponding segment of the video.
5. The graphical display of claim 1 , wherein the visual indication of the popularity of at least some the segments with other users comprises a graph overlaid on the timeline.
6. The graphical display of claim 5 , wherein the graph comprises a plurality of bars at positions along the timeline, the height of each bar indicating a strength of favorability among the users regarding a segment of the video corresponding to the position of the bar along the timeline.
7. The graphical display of claim 5 , wherein the graph comprises a plurality of bars at positions along the timeline, the height of each bar indicating a strength of unfavorability among the users regarding a segment of the video corresponding to the position of the bar along the timeline.
8. The graphical display of claim 1 , wherein the visual indication of the popularity of at least some the segments with other users comprises at least one controversy marker at a position on the timeline, the controversy marker indicating a large number of users have indicated both a favorable and unfavorable response to a corresponding segment of the video.
9. A method of displaying a video and indicating a popularity of a plurality of segments of the video, the method comprising:
displaying the video comprising the plurality of segments to a user;
displaying a timeline representing the video;
displaying a marker indicating a segment of the video currently displayed; and
displaying a visual indication of the popularity of at least some of the segments with other users.
10. The method of claim 9 , further comprising displaying a favorable vote icon for the user to select in order to indicate a favorable opinion of the segment of the video currently displayed.
11. The method of claim 10 , further comprising displaying an unfavorable vote icon for the user to select in order to indicate an unfavorable opinion of the segment of the video currently displayed.
12. The method of claim 9 , wherein the visual indication of the popularity of at least some of the segments with other users comprises an intensity of shading of portions of the timeline, the intensity of shading of a portion indicating the popularity of the corresponding segment of the video.
13. The method of claim 9 , wherein the visual indication of the popularity of at least some of the segments with other users comprises a graph overlaid on the timeline.
14. The method of claim 13 , wherein the graph comprises a plurality of bars at positions along the timeline, the height of each bar indicating a strength of favorability among the users regarding a segment of the video corresponding to the position of the bar along the timeline.
15. The method of claim 13 , wherein the graph comprises a plurality of bars at positions along the timeline, the height of each bar indicating a strength of unfavorability among the users regarding a segment of the video corresponding to the position of the bar along the timeline.
16. The method of claim 9 , wherein the visual indication of the popularity of at least some of the segments with other users comprises at least one controversy marker at a position on the timeline, the controversy marker indicating a large number of users have indicated both a favorable and unfavorable response to a corresponding segment of the video.
17. A non-transitory computer-readable medium storing computer program instructions for displaying a video and indicating a popularity of a plurality of segments of the video, the computer program instructions configuring a computer to perform operations comprising:
displaying the video comprising the plurality of segments to a user;
displaying a timeline representing the video;
displaying a marker indicating a segment of the video currently displayed; and
displaying a visual indication of the popularity of at least some of the segments with other users.
18. The non-transitory computer-readable medium of claim 17 , wherein the operations further comprises:
displaying a favorable vote icon for the user to select in order to indicate a favorable opinion of the segment of the video currently displayed; and
displaying an unfavorable vote icon for the user to select in order to indicate an unfavorable opinion of the segment of the video currently displayed.
19. The non-transitory computer-readable medium of claim 17 , wherein the visual indication of the popularity of at least some of the segments with other users comprises an intensity of shading of portions of the timeline, the intensity of shading of a portion indicating the popularity of the corresponding segment of the video.
20. The non-transitory computer-readable medium of claim 17 , wherein the visual indication of the popularity of at least some of the segments with other users comprises a graph overlaid on the timeline, the graph comprising a plurality of bars at positions along the timeline, the height of each bar indicating a strength of favorability among the users regarding a segment of the video corresponding to the position of the bar along the timeline.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US14/145,710 US20140115621A1 (en) | 2009-03-30 | 2013-12-31 | Intra-Video Ratings |
Applications Claiming Priority (3)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US12/414,553 US8132200B1 (en) | 2009-03-30 | 2009-03-30 | Intra-video ratings |
US13/363,358 US9044183B1 (en) | 2009-03-30 | 2012-01-31 | Intra-video ratings |
US14/145,710 US20140115621A1 (en) | 2009-03-30 | 2013-12-31 | Intra-Video Ratings |
Related Parent Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US13/363,358 Continuation US9044183B1 (en) | 2009-03-30 | 2012-01-31 | Intra-video ratings |
Publications (1)
Publication Number | Publication Date |
---|---|
US20140115621A1 true US20140115621A1 (en) | 2014-04-24 |
Family
ID=45758067
Family Applications (3)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US12/414,553 Active 2030-04-10 US8132200B1 (en) | 2009-03-30 | 2009-03-30 | Intra-video ratings |
US13/363,358 Active 2031-02-11 US9044183B1 (en) | 2009-03-30 | 2012-01-31 | Intra-video ratings |
US14/145,710 Abandoned US20140115621A1 (en) | 2009-03-30 | 2013-12-31 | Intra-Video Ratings |
Family Applications Before (2)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US12/414,553 Active 2030-04-10 US8132200B1 (en) | 2009-03-30 | 2009-03-30 | Intra-video ratings |
US13/363,358 Active 2031-02-11 US9044183B1 (en) | 2009-03-30 | 2012-01-31 | Intra-video ratings |
Country Status (1)
Country | Link |
---|---|
US (3) | US8132200B1 (en) |
Cited By (9)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20140089801A1 (en) * | 2012-09-21 | 2014-03-27 | Comment Bubble, Inc. | Timestamped commentary system for video content |
US20140317647A1 (en) * | 2011-10-27 | 2014-10-23 | Yuichiro Itakura | Content evaluation/playback device |
CN105263056A (en) * | 2015-10-26 | 2016-01-20 | 天脉聚源(北京)科技有限公司 | Method and device for displaying the number of votes from the audience in real time on video program |
CN105282595A (en) * | 2015-10-26 | 2016-01-27 | 天脉聚源(北京)科技有限公司 | Method and device for displaying support rates in real time on video program |
US20190197566A1 (en) * | 2010-09-30 | 2019-06-27 | Rakuten, Inc. | Server apparatus for collecting a response of a questionnaire, questionnaire response collection method, questionnaire response collection program and computer-readable recording medium recorded with a questionnaire response collection program |
US10779044B1 (en) * | 2019-03-14 | 2020-09-15 | The Nielsen Company (Us), Llc | Viewer authentication |
US20200304841A1 (en) * | 2015-03-27 | 2020-09-24 | Twitter, Inc. | Live video streaming services |
US20210105510A1 (en) * | 2016-01-26 | 2021-04-08 | Twitter, Inc. | Live video streaming services using one or more external devices |
WO2023278256A1 (en) * | 2021-07-02 | 2023-01-05 | Datashapes, Inc. | Navigating content by relevance |
Families Citing this family (38)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US8112702B2 (en) | 2008-02-19 | 2012-02-07 | Google Inc. | Annotating video intervals |
WO2011071309A2 (en) * | 2009-12-09 | 2011-06-16 | Samsung Electronics Co., Ltd. | Method and apparatus for sharing comments regarding content |
US20120072845A1 (en) * | 2010-09-21 | 2012-03-22 | Avaya Inc. | System and method for classifying live media tags into types |
US9237297B1 (en) * | 2010-12-06 | 2016-01-12 | Kenneth M. Waddell | Jump view interactive video system |
US20120179557A1 (en) * | 2011-01-12 | 2012-07-12 | John Nicholas Gross | Performance Based Internet Reward System |
US9147194B1 (en) | 2011-09-23 | 2015-09-29 | Google Inc. | Aggregated performance information for video content items |
US20130247078A1 (en) * | 2012-03-19 | 2013-09-19 | Rawllin International Inc. | Emoticons for media |
US20130265314A1 (en) * | 2012-04-09 | 2013-10-10 | Buzzerandbell, Inc. | Method and apparatus for capturing and analyzing real-time user sentiment for an event |
US11011006B2 (en) * | 2012-06-29 | 2021-05-18 | Papalove Productions, Llc | Method and system for evaluating and sharing media |
US8843951B1 (en) | 2012-08-27 | 2014-09-23 | Google Inc. | User behavior indicator |
US9560332B2 (en) | 2012-09-10 | 2017-01-31 | Google Inc. | Media summarization |
KR20140062886A (en) * | 2012-11-15 | 2014-05-26 | 엘지전자 주식회사 | Mobile terminal and control method thereof |
WO2014120675A1 (en) * | 2013-01-31 | 2014-08-07 | Knotch, Inc. | Gradation coding to express sentiment |
CN104035697B (en) * | 2013-03-04 | 2018-02-27 | 联想(北京)有限公司 | A kind of control method and electronic equipment |
US20140344695A1 (en) * | 2013-05-17 | 2014-11-20 | Funtini, Inc. | Invitation to participate based on user generated content |
US20140379710A1 (en) * | 2013-06-19 | 2014-12-25 | International Business Machines Corporation | Pattern based video frame navigation aid |
KR102268052B1 (en) | 2013-11-11 | 2021-06-22 | 삼성전자주식회사 | Display apparatus, server apparatus and user interface screen providing method thereof |
USD764507S1 (en) | 2014-01-28 | 2016-08-23 | Knotch, Inc. | Display screen or portion thereof with animated graphical user interface |
KR20160035649A (en) * | 2014-09-23 | 2016-04-01 | 삼성전자주식회사 | Apparatus and method for displaying preference of contents in electronics device |
CN105307038A (en) * | 2015-10-26 | 2016-02-03 | 天脉聚源(北京)科技有限公司 | Method and device for displaying support rates of competing teams in real time |
CN105407386A (en) * | 2015-10-26 | 2016-03-16 | 天脉聚源(北京)科技有限公司 | Method of displaying approval rate in video program in real time and apparatus thereof |
CN105407374A (en) * | 2015-10-26 | 2016-03-16 | 天脉聚源(北京)科技有限公司 | Method and device for displaying participating team support rate in real time |
CN105307037A (en) * | 2015-10-26 | 2016-02-03 | 天脉聚源(北京)科技有限公司 | Method and device for displaying support rates on video program in real time |
CN105307035A (en) * | 2015-10-26 | 2016-02-03 | 天脉聚源(北京)科技有限公司 | Method and device for displaying audience voting numbers on video program in real time |
CN105407387A (en) * | 2015-10-26 | 2016-03-16 | 天脉聚源(北京)科技有限公司 | Method of displaying number of participating team audience votes in real time and apparatus thereof |
CN105307036A (en) * | 2015-10-26 | 2016-02-03 | 天脉聚源(北京)科技有限公司 | Method and device for displaying support rates of competing teams in real time |
CN105376631A (en) * | 2015-10-26 | 2016-03-02 | 天脉聚源(北京)科技有限公司 | Real-time audience vote number display method and device |
US10999614B2 (en) * | 2016-03-31 | 2021-05-04 | Rovi Guides, Inc. | Methods and systems for efficiently downloading media assets |
US11132716B2 (en) * | 2016-06-28 | 2021-09-28 | Gavin Washington Brown | System and method for promoting a talent of a user via a wireless network of mobile client devices |
US10721503B2 (en) * | 2017-06-09 | 2020-07-21 | Sony Interactive Entertainment LLC | Systems and methods for operating a streaming service to provide community spaces for media content items |
US11285372B2 (en) | 2017-10-03 | 2022-03-29 | Todd Wanke | Systems, devices, and methods employing the same for enhancing audience engagement in a competition or performance |
WO2019070351A1 (en) | 2017-10-03 | 2019-04-11 | Fanmountain Llc | Systems, devices, and methods employing the same for enhancing audience engagement in a competition or performance |
US10595083B2 (en) | 2018-04-20 | 2020-03-17 | The Nielsen Company (Us), Llc | Methods and apparatus to determine audio source impact on an audience of media |
US20230325857A1 (en) * | 2018-12-11 | 2023-10-12 | Hiwave Technologies Inc. | Method and system of sentiment-based selective user engagement |
US12047649B2 (en) | 2020-09-16 | 2024-07-23 | Fanmountain Llc | Devices, systems, and their methods of use in generating and distributing content |
US12003814B2 (en) | 2021-04-22 | 2024-06-04 | STE Capital, LLC | System for audience sentiment feedback and analysis |
USD1029795S1 (en) * | 2022-09-16 | 2024-06-04 | Microsoft Corporation | Headset |
USD1039512S1 (en) * | 2022-09-16 | 2024-08-20 | Microsoft Corporation | Headset |
Citations (8)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US6144375A (en) * | 1998-08-14 | 2000-11-07 | Praja Inc. | Multi-perspective viewer for content-based interactivity |
US20060218590A1 (en) * | 2005-03-10 | 2006-09-28 | Sbc Knowledge Ventures, L.P. | System and method for displaying an electronic program guide |
US20080092168A1 (en) * | 1999-03-29 | 2008-04-17 | Logan James D | Audio and video program recording, editing and playback systems using metadata |
US20080168055A1 (en) * | 2007-01-04 | 2008-07-10 | Wide Angle Llc | Relevancy rating of tags |
US20090007200A1 (en) * | 2007-06-29 | 2009-01-01 | At&T Knowledge Ventures, Lp | System and method of providing video content commentary |
US20090276805A1 (en) * | 2008-05-03 | 2009-11-05 | Andrews Ii James K | Method and system for generation and playback of supplemented videos |
US20130290996A1 (en) * | 2008-06-18 | 2013-10-31 | Ipowow! Ltd. | Assessing digital content across a communications network |
US8713618B1 (en) * | 2008-11-06 | 2014-04-29 | Google Inc. | Segmenting video based on timestamps in comments |
Family Cites Families (99)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
EP0526064B1 (en) | 1991-08-02 | 1997-09-10 | The Grass Valley Group, Inc. | Video editing system operator interface for visualization and interactive control of video material |
JPH05197573A (en) | 1991-08-26 | 1993-08-06 | Hewlett Packard Co <Hp> | Task controlling system with task oriented paradigm |
GB2270242A (en) | 1992-08-29 | 1994-03-02 | Ibm | A method of editing for an object oriented computer system |
US5339393A (en) | 1993-04-15 | 1994-08-16 | Sony Electronics, Inc. | Graphical user interface for displaying available source material for editing |
US5996121A (en) | 1993-07-28 | 1999-12-07 | Harris; Eunice | Convertible coat |
US5664216A (en) | 1994-03-22 | 1997-09-02 | Blumenau; Trevor | Iconic audiovisual data editing environment |
US5600775A (en) | 1994-08-26 | 1997-02-04 | Emotion, Inc. | Method and apparatus for annotating full motion video and other indexed data structures |
US5812642A (en) * | 1995-07-12 | 1998-09-22 | Leroy; David J. | Audience response monitor and analysis system and method |
US5708845A (en) | 1995-09-29 | 1998-01-13 | Wistendahl; Douglass A. | System for mapping hot spots in media content for interactive digital media program |
US5732184A (en) | 1995-10-20 | 1998-03-24 | Digital Processing Systems, Inc. | Video and audio cursor video editing system |
US6570587B1 (en) | 1996-07-26 | 2003-05-27 | Veon Ltd. | System and method and linking information to a video |
US20020120925A1 (en) | 2000-03-28 | 2002-08-29 | Logan James D. | Audio and video program recording, editing and playback systems using metadata |
US6006241A (en) | 1997-03-14 | 1999-12-21 | Microsoft Corporation | Production of a video stream with synchronized annotations over a computer network |
US6295092B1 (en) * | 1998-07-30 | 2001-09-25 | Cbs Corporation | System for analyzing television programs |
US6917965B2 (en) | 1998-09-15 | 2005-07-12 | Microsoft Corporation | Facilitating annotation creation and notification via electronic mail |
US6956593B1 (en) | 1998-09-15 | 2005-10-18 | Microsoft Corporation | User interface for creating, viewing and temporally positioning annotations for media content |
US6357042B2 (en) | 1998-09-16 | 2002-03-12 | Anand Srinivasan | Method and apparatus for multiplexing separately-authored metadata for insertion into a video data stream |
US20020059218A1 (en) * | 1999-01-26 | 2002-05-16 | Katherine Grace August | System and method for obtaining real time survey information for media programming using input device |
KR100326400B1 (en) | 1999-05-19 | 2002-03-12 | 김광수 | Method for generating caption location information, method for searching thereby, and reproducing apparatus using the methods |
US6415438B1 (en) | 1999-10-05 | 2002-07-02 | Webtv Networks, Inc. | Trigger having a time attribute |
US7055168B1 (en) | 2000-05-03 | 2006-05-30 | Sharp Laboratories Of America, Inc. | Method for interpreting and executing user preferences of audiovisual information |
KR20040041082A (en) | 2000-07-24 | 2004-05-13 | 비브콤 인코포레이티드 | System and method for indexing, searching, identifying, and editing portions of electronic multimedia files |
JP2002163143A (en) | 2000-07-28 | 2002-06-07 | Any One Wireless Co Ltd | Content reformatting system and its method for radio site |
US20020065678A1 (en) | 2000-08-25 | 2002-05-30 | Steven Peliotis | iSelect video |
US8020183B2 (en) | 2000-09-14 | 2011-09-13 | Sharp Laboratories Of America, Inc. | Audiovisual management system |
US6774908B2 (en) | 2000-10-03 | 2004-08-10 | Creative Frontier Inc. | System and method for tracking an object in a video and linking information thereto |
US7254605B1 (en) * | 2000-10-26 | 2007-08-07 | Austen Services Llc | Method of modulating the transmission frequency in a real time opinion research network |
US7032178B1 (en) | 2001-03-30 | 2006-04-18 | Gateway Inc. | Tagging content for different activities |
US20040138946A1 (en) | 2001-05-04 | 2004-07-15 | Markus Stolze | Web page annotation systems |
US20020188630A1 (en) | 2001-05-21 | 2002-12-12 | Autodesk, Inc. | Method and apparatus for annotating a sequence of frames |
AU2002314309A1 (en) | 2001-06-11 | 2002-12-23 | C-Burn Systems Ltd | Selecting tracks from a jukebox via a wireless communications device |
WO2003001413A1 (en) | 2001-06-22 | 2003-01-03 | Nosa Omoigui | System and method for knowledge retrieval, management, delivery and presentation |
TW520602B (en) | 2001-06-28 | 2003-02-11 | Ulead Systems Inc | Device and method of editing video program |
US20030018668A1 (en) | 2001-07-20 | 2003-01-23 | International Business Machines Corporation | Enhanced transcoding of structured documents through use of annotation techniques |
US20050160113A1 (en) | 2001-08-31 | 2005-07-21 | Kent Ridge Digital Labs | Time-based media navigation system |
US7343487B2 (en) | 2001-10-10 | 2008-03-11 | Nokia Corporation | Datacast distribution system |
US7180623B2 (en) | 2001-12-03 | 2007-02-20 | Canon Kabushiki Kaisha | Method and apparatus for print error recovery |
US20030107592A1 (en) | 2001-12-11 | 2003-06-12 | Koninklijke Philips Electronics N.V. | System and method for retrieving information related to persons in video programs |
US20030112276A1 (en) | 2001-12-19 | 2003-06-19 | Clement Lau | User augmentation of content |
US7137062B2 (en) | 2001-12-28 | 2006-11-14 | International Business Machines Corporation | System and method for hierarchical segmentation with latent semantic indexing in scale space |
US20040205482A1 (en) | 2002-01-24 | 2004-10-14 | International Business Machines Corporation | Method and apparatus for active annotation of multimedia content |
US6988245B2 (en) | 2002-06-18 | 2006-01-17 | Koninklijke Philips Electronics N.V. | System and method for providing videomarks for a video program |
US7149755B2 (en) | 2002-07-29 | 2006-12-12 | Hewlett-Packard Development Company, Lp. | Presenting a collection of media objects |
US7257774B2 (en) | 2002-07-30 | 2007-08-14 | Fuji Xerox Co., Ltd. | Systems and methods for filtering and/or viewing collaborative indexes of recorded media |
US6993347B2 (en) | 2002-12-17 | 2006-01-31 | International Business Machines Corporation | Dynamic media interleaving |
US8307273B2 (en) | 2002-12-30 | 2012-11-06 | The Board Of Trustees Of The Leland Stanford Junior University | Methods and apparatus for interactive network sharing of digital video content |
US7131059B2 (en) | 2002-12-31 | 2006-10-31 | Hewlett-Packard Development Company, L.P. | Scalably presenting a collection of media objects |
US7383497B2 (en) | 2003-01-21 | 2008-06-03 | Microsoft Corporation | Random access editing of media |
US7904797B2 (en) | 2003-01-21 | 2011-03-08 | Microsoft Corporation | Rapid media group annotation |
US7739597B2 (en) | 2003-02-24 | 2010-06-15 | Microsoft Corporation | Interactive media frame display |
US8392834B2 (en) | 2003-04-09 | 2013-03-05 | Hewlett-Packard Development Company, L.P. | Systems and methods of authoring a multimedia file |
AU2004246672B2 (en) | 2003-06-02 | 2009-02-26 | Disney Enterprises, Inc. | System and method of interactive video playback |
US8321470B2 (en) | 2003-06-20 | 2012-11-27 | International Business Machines Corporation | Heterogeneous multi-level extendable indexing for general purpose annotation systems |
US7418656B1 (en) | 2003-10-03 | 2008-08-26 | Adobe Systems Incorporated | Dynamic annotations for electronics documents |
US20050203892A1 (en) | 2004-03-02 | 2005-09-15 | Jonathan Wesley | Dynamically integrating disparate systems and providing secure data sharing |
WO2005091175A1 (en) | 2004-03-15 | 2005-09-29 | Yahoo! Inc. | Search systems and methods with integration of user annotations |
WO2005089291A2 (en) | 2004-03-15 | 2005-09-29 | Yahoo Inc. | Search system and methods with integration of user annotations from a trust network |
JP2005352933A (en) | 2004-06-14 | 2005-12-22 | Fuji Xerox Co Ltd | Display arrangement, system, and display method |
US20050289469A1 (en) | 2004-06-28 | 2005-12-29 | Chandler Roger D | Context tagging apparatus, systems, and methods |
US20060041564A1 (en) | 2004-08-20 | 2006-02-23 | Innovative Decision Technologies, Inc. | Graphical Annotations and Domain Objects to Create Feature Level Metadata of Images |
US20070118794A1 (en) | 2004-09-08 | 2007-05-24 | Josef Hollander | Shared annotation system and method |
US20060064733A1 (en) | 2004-09-20 | 2006-03-23 | Norton Jeffrey R | Playing an audiovisual work with dynamic choosing |
WO2006042159A2 (en) | 2004-10-05 | 2006-04-20 | Vectormax Corporation | Interactive video collaboration framework |
US7472341B2 (en) | 2004-11-08 | 2008-12-30 | International Business Machines Corporation | Multi-user, multi-timed collaborative annotation |
JP2006155384A (en) | 2004-11-30 | 2006-06-15 | Nippon Telegr & Teleph Corp <Ntt> | Video comment input/display method and device, program, and storage medium with program stored |
JP4353083B2 (en) | 2004-11-30 | 2009-10-28 | 日本電信電話株式会社 | Inter-viewer communication method, apparatus and program |
US9275052B2 (en) | 2005-01-19 | 2016-03-01 | Amazon Technologies, Inc. | Providing annotations of a digital work |
US20060286536A1 (en) | 2005-04-01 | 2006-12-21 | Sherman Mohler | System and method for regulating use of content and content styles in a distributed learning system |
US7636883B2 (en) | 2005-05-18 | 2009-12-22 | International Business Machines Corporation | User form based automated and guided data collection |
US20080005064A1 (en) | 2005-06-28 | 2008-01-03 | Yahoo! Inc. | Apparatus and method for content annotation and conditional annotation retrieval in a search context |
WO2007002820A2 (en) | 2005-06-28 | 2007-01-04 | Yahoo! Inc. | Search engine with augmented relevance ranking by community participation |
EP1908303A4 (en) | 2005-07-01 | 2011-04-06 | Sonic Solutions | Method, apparatus and system for use in multimedia signal encoding |
KR20070004153A (en) | 2005-07-04 | 2007-01-09 | 주식회사 다음커뮤니케이션 | System of providing contents preferred by a user and method thereof, system of selecting contents preferred by a user and method thereof, and system of selecting contents preferred by a group of users and method thereof |
EP1758398A1 (en) | 2005-08-23 | 2007-02-28 | Syneola SA | Multilevel semiotic and fuzzy logic user and metadata interface means for interactive multimedia system having cognitive adaptive capability |
EP1920546B1 (en) * | 2005-08-30 | 2014-04-16 | NDS Limited | Enhanced electronic program guides |
JP2007151057A (en) | 2005-10-25 | 2007-06-14 | Dainippon Printing Co Ltd | Video content browsing system using evaluation comment information |
US20070099684A1 (en) | 2005-11-03 | 2007-05-03 | Evans Butterworth | System and method for implementing an interactive storyline |
US7945653B2 (en) | 2006-10-11 | 2011-05-17 | Facebook, Inc. | Tagging digital media |
US7761436B2 (en) | 2006-01-03 | 2010-07-20 | Yahoo! Inc. | Apparatus and method for controlling content access based on shared annotations for annotated users in a folksonomy scheme |
WO2007082169A2 (en) | 2006-01-05 | 2007-07-19 | Eyespot Corporation | Automatic aggregation of content for use in an online video editing system |
US8214516B2 (en) | 2006-01-06 | 2012-07-03 | Google Inc. | Dynamic media serving infrastructure |
JP2007274090A (en) | 2006-03-30 | 2007-10-18 | Toshiba Corp | Content reproducing apparatus, method, and program |
JP5649303B2 (en) | 2006-03-30 | 2015-01-07 | エスアールアイ インターナショナルSRI International | Method and apparatus for annotating media streams |
US8701005B2 (en) | 2006-04-26 | 2014-04-15 | At&T Intellectual Property I, Lp | Methods, systems, and computer program products for managing video information |
US7954049B2 (en) | 2006-05-15 | 2011-05-31 | Microsoft Corporation | Annotating multimedia files along a timeline |
JP4769635B2 (en) | 2006-05-22 | 2011-09-07 | 日本電信電話株式会社 | Server apparatus and client apparatus and program thereof |
WO2008014408A1 (en) | 2006-07-28 | 2008-01-31 | Blue Lava Technologies | Method and system for displaying multimedia content |
EP2051509B1 (en) * | 2006-08-10 | 2016-07-27 | Panasonic Intellectual Property Management Co., Ltd. | Program recommendation system, program view terminal, program view program, program view method, program recommendation server, program recommendation program, and program recommendation method |
US8850464B2 (en) * | 2006-10-09 | 2014-09-30 | Verizon Patent And Licensing Inc. | Systems and methods for real-time interactive television polling |
US7559017B2 (en) | 2006-12-22 | 2009-07-07 | Google Inc. | Annotation framework for video |
US7707162B2 (en) | 2007-01-08 | 2010-04-27 | International Business Machines Corporation | Method and apparatus for classifying multimedia artifacts using ontology selection and semantic classification |
US20080250331A1 (en) * | 2007-04-04 | 2008-10-09 | Atul Tulshibagwale | Method and System of a Voting Based Wiki and Its Application to Internet Topic Directories |
WO2009046324A2 (en) | 2007-10-05 | 2009-04-09 | Flickbitz Corporation | Online search, storage, manipulation, and delivery of video content |
US8209223B2 (en) | 2007-11-30 | 2012-06-26 | Google Inc. | Video object tag creation and processing |
US20090172745A1 (en) | 2007-12-28 | 2009-07-02 | Motorola, Inc. | Method and Apparatus Regarding Receipt of Audio-Visual Content Information and Use of Such Information to Automatically Infer a Relative Popularity of That Content |
US8181197B2 (en) * | 2008-02-06 | 2012-05-15 | Google Inc. | System and method for voting on popular video intervals |
US8112702B2 (en) | 2008-02-19 | 2012-02-07 | Google Inc. | Annotating video intervals |
US8566353B2 (en) | 2008-06-03 | 2013-10-22 | Google Inc. | Web-based system for collaborative generation of interactive videos |
US8839327B2 (en) * | 2008-06-25 | 2014-09-16 | At&T Intellectual Property Ii, Lp | Method and apparatus for presenting media programs |
-
2009
- 2009-03-30 US US12/414,553 patent/US8132200B1/en active Active
-
2012
- 2012-01-31 US US13/363,358 patent/US9044183B1/en active Active
-
2013
- 2013-12-31 US US14/145,710 patent/US20140115621A1/en not_active Abandoned
Patent Citations (8)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US6144375A (en) * | 1998-08-14 | 2000-11-07 | Praja Inc. | Multi-perspective viewer for content-based interactivity |
US20080092168A1 (en) * | 1999-03-29 | 2008-04-17 | Logan James D | Audio and video program recording, editing and playback systems using metadata |
US20060218590A1 (en) * | 2005-03-10 | 2006-09-28 | Sbc Knowledge Ventures, L.P. | System and method for displaying an electronic program guide |
US20080168055A1 (en) * | 2007-01-04 | 2008-07-10 | Wide Angle Llc | Relevancy rating of tags |
US20090007200A1 (en) * | 2007-06-29 | 2009-01-01 | At&T Knowledge Ventures, Lp | System and method of providing video content commentary |
US20090276805A1 (en) * | 2008-05-03 | 2009-11-05 | Andrews Ii James K | Method and system for generation and playback of supplemented videos |
US20130290996A1 (en) * | 2008-06-18 | 2013-10-31 | Ipowow! Ltd. | Assessing digital content across a communications network |
US8713618B1 (en) * | 2008-11-06 | 2014-04-29 | Google Inc. | Segmenting video based on timestamps in comments |
Cited By (15)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20190197566A1 (en) * | 2010-09-30 | 2019-06-27 | Rakuten, Inc. | Server apparatus for collecting a response of a questionnaire, questionnaire response collection method, questionnaire response collection program and computer-readable recording medium recorded with a questionnaire response collection program |
US10535073B2 (en) * | 2010-09-30 | 2020-01-14 | Rakuten, Inc. | Server apparatus for collecting a response of a questionnaire, questionnaire response collection method, questionnaire response collection program and computer-readable recording medium recorded with a questionnaire response collection program |
US20140317647A1 (en) * | 2011-10-27 | 2014-10-23 | Yuichiro Itakura | Content evaluation/playback device |
US20140089801A1 (en) * | 2012-09-21 | 2014-03-27 | Comment Bubble, Inc. | Timestamped commentary system for video content |
US9332315B2 (en) | 2012-09-21 | 2016-05-03 | Comment Bubble, Inc. | Timestamped commentary system for video content |
US20200304841A1 (en) * | 2015-03-27 | 2020-09-24 | Twitter, Inc. | Live video streaming services |
US20220150572A1 (en) * | 2015-03-27 | 2022-05-12 | Twitter, Inc. | Live video streaming services |
CN105282595A (en) * | 2015-10-26 | 2016-01-27 | 天脉聚源(北京)科技有限公司 | Method and device for displaying support rates in real time on video program |
CN105263056A (en) * | 2015-10-26 | 2016-01-20 | 天脉聚源(北京)科技有限公司 | Method and device for displaying the number of votes from the audience in real time on video program |
US20210105510A1 (en) * | 2016-01-26 | 2021-04-08 | Twitter, Inc. | Live video streaming services using one or more external devices |
US10779044B1 (en) * | 2019-03-14 | 2020-09-15 | The Nielsen Company (Us), Llc | Viewer authentication |
US11463777B2 (en) * | 2019-03-14 | 2022-10-04 | The Nielsen Company (Us), Llc | Viewer authentication |
WO2023278256A1 (en) * | 2021-07-02 | 2023-01-05 | Datashapes, Inc. | Navigating content by relevance |
US11722739B2 (en) | 2021-07-02 | 2023-08-08 | Datashapes, Inc. | Navigating content by relevance |
US11968428B2 (en) | 2021-07-02 | 2024-04-23 | Datashapes, Inc. | Navigating content by relevance |
Also Published As
Publication number | Publication date |
---|---|
US9044183B1 (en) | 2015-06-02 |
US8132200B1 (en) | 2012-03-06 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US9044183B1 (en) | Intra-video ratings | |
US10795929B2 (en) | Interactive music feedback system | |
US11354368B2 (en) | Displaying information related to spoken dialogue in content playing on a device | |
US8181197B2 (en) | System and method for voting on popular video intervals | |
US9407974B2 (en) | Segmenting video based on timestamps in comments | |
US9253511B2 (en) | Systems and methods for performing multi-modal video datastream segmentation | |
US20210319475A1 (en) | Method and system for matching location-based content | |
EP3346718B1 (en) | Methods and systems for displaying contextually relevant information regarding a media asset | |
US20120102410A1 (en) | Media management for multi-user group | |
US11095930B2 (en) | Methods, systems, and media for indicating viewership of a video | |
WO2020157283A1 (en) | Method for recommending video content | |
US9946769B2 (en) | Displaying information related to spoken dialogue in content playing on a device | |
US20200183975A1 (en) | Video content optimization system | |
US20110179003A1 (en) | System for Sharing Emotion Data and Method of Sharing Emotion Data Using the Same | |
US8875177B1 (en) | Serving video content segments | |
US20140059452A1 (en) | Cumulative encyclopedia method,system and device | |
US10503794B2 (en) | Video content optimization system and method for content and advertisement placement improvement on a third party media content platform | |
CN112445921B (en) | Digest generation method and digest generation device | |
CN104427361A (en) | Television service system and method for providing video and audio service | |
TW201914305A (en) | System and method for video with personalized weighted rating scores | |
US11074596B1 (en) | System and method for identifying topic coverage for a distribution platform that provides access to online content items | |
CN116128324A (en) | Multimedia resource evaluation and pushing method and server |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
STCB | Information on status: application discontinuation |
Free format text: ABANDONED -- FAILURE TO RESPOND TO AN OFFICE ACTION |