US20180358049A1 - Multi-modal collaborative web-based video annotation system - Google Patents

Multi-modal collaborative web-based video annotation system Download PDF

Info

Publication number
US20180358049A1
US20180358049A1 US16/103,307 US201816103307A US2018358049A1 US 20180358049 A1 US20180358049 A1 US 20180358049A1 US 201816103307 A US201816103307 A US 201816103307A US 2018358049 A1 US2018358049 A1 US 2018358049A1
Authority
US
United States
Prior art keywords
video
segment
comment
user
electronic device
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Abandoned
Application number
US16/103,307
Inventor
Celine LATULIPE
Vikash Singh
David Wilson
Sybil HUSKEY
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
University of North Carolina at Charlotte
Original Assignee
University of North Carolina at Charlotte
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Priority claimed from US13/627,899 external-priority patent/US9354763B2/en
Application filed by University of North Carolina at Charlotte filed Critical University of North Carolina at Charlotte
Priority to US16/103,307 priority Critical patent/US20180358049A1/en
Publication of US20180358049A1 publication Critical patent/US20180358049A1/en
Priority to US16/850,176 priority patent/US20200286519A1/en
Abandoned legal-status Critical Current

Links

Images

Classifications

    • GPHYSICS
    • G11INFORMATION STORAGE
    • G11BINFORMATION STORAGE BASED ON RELATIVE MOVEMENT BETWEEN RECORD CARRIER AND TRANSDUCER
    • G11B27/00Editing; Indexing; Addressing; Timing or synchronising; Monitoring; Measuring tape travel
    • G11B27/002Programmed access in sequence to a plurality of record carriers or indexed parts, e.g. tracks, thereof, e.g. for editing
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/70Information retrieval; Database structures therefor; File system structures therefor of video data
    • G06F16/78Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually
    • G06F16/7867Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually using information manually generated, e.g. tags, keywords, comments, title and artist information, manually generated time, location and usage information, user ratings
    • G06F17/241
    • G06F17/3082
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/048Interaction techniques based on graphical user interfaces [GUI]
    • G06F3/0481Interaction techniques based on graphical user interfaces [GUI] based on specific properties of the displayed interaction object or a metaphor-based environment, e.g. interaction with desktop elements like windows or icons, or assisted by a cursor's changing behaviour or appearance
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/048Interaction techniques based on graphical user interfaces [GUI]
    • G06F3/0481Interaction techniques based on graphical user interfaces [GUI] based on specific properties of the displayed interaction object or a metaphor-based environment, e.g. interaction with desktop elements like windows or icons, or assisted by a cursor's changing behaviour or appearance
    • G06F3/0482Interaction with lists of selectable items, e.g. menus
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/048Interaction techniques based on graphical user interfaces [GUI]
    • G06F3/0484Interaction techniques based on graphical user interfaces [GUI] for the control of specific functions or operations, e.g. selecting or manipulating an object, an image or a displayed text element, setting a parameter value or selecting a range
    • G06F3/04842Selection of displayed objects or displayed text elements
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/048Interaction techniques based on graphical user interfaces [GUI]
    • G06F3/0484Interaction techniques based on graphical user interfaces [GUI] for the control of specific functions or operations, e.g. selecting or manipulating an object, an image or a displayed text element, setting a parameter value or selecting a range
    • G06F3/04847Interaction techniques to control parameter settings, e.g. interaction with sliders or dials
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F40/00Handling natural language data
    • G06F40/10Text processing
    • G06F40/166Editing, e.g. inserting or deleting
    • G06F40/169Annotation, e.g. comment data or footnotes
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06QINFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES; SYSTEMS OR METHODS SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES, NOT OTHERWISE PROVIDED FOR
    • G06Q10/00Administration; Management
    • G06Q10/10Office automation; Time management
    • G06Q10/101Collaborative creation, e.g. joint development of products or services
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06QINFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES; SYSTEMS OR METHODS SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES, NOT OTHERWISE PROVIDED FOR
    • G06Q10/00Administration; Management
    • G06Q10/10Office automation; Time management
    • G06Q10/103Workflow collaboration or project management
    • GPHYSICS
    • G11INFORMATION STORAGE
    • G11BINFORMATION STORAGE BASED ON RELATIVE MOVEMENT BETWEEN RECORD CARRIER AND TRANSDUCER
    • G11B27/00Editing; Indexing; Addressing; Timing or synchronising; Monitoring; Measuring tape travel
    • G11B27/02Editing, e.g. varying the order of information signals recorded on, or reproduced from, record carriers
    • G11B27/031Electronic editing of digitised analogue information signals, e.g. audio or video signals
    • G11B27/036Insert-editing
    • GPHYSICS
    • G11INFORMATION STORAGE
    • G11BINFORMATION STORAGE BASED ON RELATIVE MOVEMENT BETWEEN RECORD CARRIER AND TRANSDUCER
    • G11B27/00Editing; Indexing; Addressing; Timing or synchronising; Monitoring; Measuring tape travel
    • G11B27/10Indexing; Addressing; Timing or synchronising; Measuring tape travel
    • G11B27/102Programmed access in sequence to addressed parts of tracks of operating record carriers
    • G11B27/105Programmed access in sequence to addressed parts of tracks of operating record carriers of operating discs
    • GPHYSICS
    • G11INFORMATION STORAGE
    • G11BINFORMATION STORAGE BASED ON RELATIVE MOVEMENT BETWEEN RECORD CARRIER AND TRANSDUCER
    • G11B27/00Editing; Indexing; Addressing; Timing or synchronising; Monitoring; Measuring tape travel
    • G11B27/10Indexing; Addressing; Timing or synchronising; Measuring tape travel
    • G11B27/19Indexing; Addressing; Timing or synchronising; Measuring tape travel by using information detectable on the record carrier
    • G11B27/22Means responsive to presence or absence of recorded information signals
    • GPHYSICS
    • G11INFORMATION STORAGE
    • G11BINFORMATION STORAGE BASED ON RELATIVE MOVEMENT BETWEEN RECORD CARRIER AND TRANSDUCER
    • G11B27/00Editing; Indexing; Addressing; Timing or synchronising; Monitoring; Measuring tape travel
    • G11B27/10Indexing; Addressing; Timing or synchronising; Measuring tape travel
    • G11B27/34Indicating arrangements 
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L67/00Network arrangements or protocols for supporting network services or applications
    • H04L67/01Protocols
    • H04L67/02Protocols based on web technology, e.g. hypertext transfer protocol [HTTP]
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L67/00Network arrangements or protocols for supporting network services or applications
    • H04L67/01Protocols
    • H04L67/10Protocols in which an application is distributed across nodes in the network
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/47End-user applications
    • H04N21/478Supplemental services, e.g. displaying phone caller identification, shopping application
    • H04N21/4788Supplemental services, e.g. displaying phone caller identification, shopping application communicating with other users, e.g. chatting
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/80Generation or processing of content or additional data by content creator independently of the distribution process; Content per se
    • H04N21/85Assembly of content; Generation of multimedia applications
    • H04N21/854Content authoring
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/80Generation or processing of content or additional data by content creator independently of the distribution process; Content per se
    • H04N21/85Assembly of content; Generation of multimedia applications
    • H04N21/854Content authoring
    • H04N21/8547Content authoring involving timestamps for synchronizing content

Definitions

  • the present invention generally relates to collaborative video annotation.
  • Video annotation interfaces to facilitate communication between different stakeholders have been explored by different groups. Interfaces have been developed to support multimedia annotation as educational tools for the distance learning market, such as the classroom 2000 project and the Microsoft Research Annotation System.
  • the “WACTool” supports P2P capabilities that allow user collaboration and content sharing using voice and digital ink comments over individual frames of a video.
  • “Project Pad” is a standalone web application for adding text notes to media clips and sharing those notes with others.
  • Video Image Annotation” tool is a Windows application that provides an interface to manually annotate a video with text.
  • VideoANT is a web-based video annotation application developed for e-learning, to annotate videos with time point markers and text.
  • Advancedene allows people to design and exchange videos with annotations.
  • “Anvil” is another tool that allows annotation of audiovisual material by inserting time-anchored comments and gesture elements.
  • the applications discussed typically support annotations in only one form (mostly text) and lack multimodal capabilities. None of these applications support fine-grained access and navigational control, multimodal annotation grouping, rich annotation positioning and user-based color coding. All of the applications discussed allow some form of video annotation but none of them comprehensively address the issues of collaboratively working on a creative temporal project, such as dance.
  • the present invention includes many aspects and features. Moreover, while many aspects and features relate to, and are described in, the context of collaborative video annotation, the present invention is not limited to use only in this context, as will become apparent from the following summaries and detailed descriptions of aspects, features, and one or more embodiments of the present invention.
  • one aspect of the present invention relates to a method comprising displaying, to a user via an electronic display associated with an electronic device, a login interface; receiving, at the electronic device from the user, input corresponding to login credentials of a user account of the user; communicating, from the electronic device, data corresponding to the login credentials; displaying, to the user via the electronic display, a video selection interface configured to allow the user to a select a video for viewing; receiving, at the electronic device from the user, input corresponding to selection of a video for viewing; communicating, from the electronic device, data corresponding to the selected video; receiving, at the electronic device, data corresponding to the video, and data corresponding to a plurality of comments associated with the video, each of the plurality of comments being associated with a particular user account of a plurality of user accounts.
  • the method further includes displaying, to the user via the electronic display associated with the electronic device, a video annotation interface comprising a video pane configured to display the video, a video timeline bar including a video play-head indicating a current point of the video which is being played, a segment timeline bar including initial and final handles configured to define a segment of the video for playing, a plurality of comment markers displayed in connection with the video timeline bar, each of the plurality of comment markers corresponding to one of the plurality of comments associated with the video, a comment display pane displaying text corresponding to at least some of the plurality of comments associated with the video, and a comment button configured to allow the user to add a comment to the video.
  • a video annotation interface comprising a video pane configured to display the video, a video timeline bar including a video play-head indicating a current point of the video which is being played, a segment timeline bar including initial and final handles configured to define a segment of the video for playing, a plurality of comment markers displayed in connection with the video timeline bar,
  • the method still further includes receiving, at the electronic device from the user, input corresponding to selection of a displayed comment marker; and in response to receiving input corresponding to selection of a displayed comment marker, automatically moving the video play-head to a point on the video determined based on the selected comment marker, and displaying, in the video pane, a frame corresponding to the position of the video play-head, and displaying, in the video pane overlaid over the displayed frame of the video, one or more annotations associated with the comment marker.
  • automatically moving the video play-head to a point on the video determined based on the selected comment marker comprises automatically moving the video play-head to a point of the video a predetermined period of time prior to a point on the video associated with the comment marker.
  • the predetermined period of time is two seconds.
  • the method in addition to automatically moving the video play-head to a point on the video determined based on the selected comment marker, includes automatically defining a segment corresponding to the selected comment marker utilizing the segment timeline bar by moving the initial handle of the segment timeline bar to a beginning point corresponding to the point on the video determined based on the selected comment marker, and automatically moving the final handle of the segment timeline bar to an end point corresponding to a second point on the video determined based on the selected comment.
  • the method further includes receiving, at the electronic device from the user, input corresponding to engagement of a play button of the video annotation interface; and repeatedly playing the video in the video pane in a loop from a frame corresponding to the position of the initial handle to a frame corresponding to the position of the final handle.
  • the received data corresponding to a plurality of comments associated with the video includes data corresponding to a particular point comment associated with the later-selected comment marker, and wherein the data corresponding to the particular point comment includes an identification of a point of the video associated with the comment.
  • the identification of a point of the video comprises a timestamp.
  • the identification of a point of the video comprises a frame identification.
  • the received data corresponding to a plurality of comments associated with the video includes data corresponding to a particular segment comment associated with the later-selected comment marker, and wherein the data corresponding to the particular segment comment includes an identification of a starting point of a segment of the video associated with the particular segment comment.
  • the data corresponding to the particular segment comment includes an identification of a length of the segment comment.
  • the data corresponding to the particular segment comment includes an identification of an end point of a segment of the video associated with the particular segment comment.
  • the displayed video annotation interface further comprises a playback speed slider configured to allow the user to vary the speed of playback of the video in the video pane.
  • the step of receiving, at the electronic device, data corresponding to the video, and data corresponding to a plurality of comments associated with the video comprises receiving such data at the electronic device via a communication adapter.
  • the electronic display is part of the electronic device.
  • the electronic device is a tablet, slate computer, or smartphone.
  • the electronic display is connected to the electronic device.
  • the electronic display is a touchscreen display, and wherein at least one of the steps of receiving, at the electronic device from the user, input, comprises receiving input from the user input via the touchscreen display.
  • the method further includes, in response to receiving input corresponding to selection of a displayed comment marker, highlighting a portion of the video displayed in the video pane that corresponds to a focus area associated with a comment corresponding to the comment marker.
  • the method further includes, in response to receiving input corresponding to selection of a displayed comment marker, zooming in on a portion of the video displayed in the video pane that corresponds to a zoom area associated with a comment corresponding to the comment marker.
  • the method further includes receiving, at the electronic device from the user, input corresponding to selection of a baseline definition tool; receiving, at the electronic device from the user, input corresponding to drawing a baseline line segment on the video pane; receiving, at the electronic device from the user, input representing a baseline measurement for the user-drawn baseline line segment; receiving, at the electronic device from the user, input corresponding to drawing a measurement line segment on the video pane; automatically determining a measurement for the measurement line segment based on a length of the user-drawn baseline line segment, the user input representing a baseline measurement for the user-drawn baseline line segment, and a length of the user-drawn measurement line segment.
  • the method still further includes displaying, to the user via the electronic display, the automatically determined measurement for the measurement line segment.
  • each of the plurality of comment markers is color coded based on the particular user account associated with a comment the comment marker is associated with.
  • the one or more annotations includes a text annotation.
  • the one or more annotations includes a video annotation.
  • the one or more annotations includes an audio annotation.
  • the one or more annotations includes a drawing annotation.
  • the one or more annotations includes annotations in a plurality of different modalities.
  • annotations in two or more of the following modalities: text, drawing, video, audio.
  • Another aspect of the present invention relates to a method comprising receiving, at an electronic device, data corresponding to a video, and data corresponding to a plurality of comments associated with the video, each of the plurality of comments being associated with a particular user account of a plurality of user accounts; displaying, to a user via an electronic display associated with the electronic device, a video annotation interface comprising a video pane configured to display the video, a video timeline bar including a video play-head indicating a current point of the video which is being played, a segment timeline bar including initial and final handles configured to define a segment of the video for playing, and a plurality of comment markers displayed in connection with the video timeline bar, each of the plurality of comment markers corresponding to one of the plurality of comments associated with the video.
  • the method further includes receiving, at the electronic device from the user, input corresponding to selection of a displayed comment marker; and in response to receiving input corresponding to selection of a displayed comment marker, automatically moving the video play-head to a point on the video determined based on the selected comment marker, and displaying, in the video pane, a frame corresponding to the position of the video play-head, and displaying, in the video pane overlaid over the displayed frame of the video, one or more comments associated with the comment marker.
  • automatically moving the video play-head to a point on the video determined based on the selected comment marker comprises automatically moving the video play-head to a point of the video a predetermined period of time prior to a point on the video associated with the comment marker.
  • the predetermined period of time is two seconds.
  • the method in addition to automatically moving the video play-head to a point on the video determined based on the selected comment marker, includes automatically defining a segment corresponding to the selected comment marker utilizing the segment timeline bar by moving the initial handle of the segment timeline bar to a beginning point corresponding to the point on the video determined based on the selected comment marker, and automatically moving the final handle of the segment timeline bar to an end point corresponding to a second point on the video determined based on the selected comment.
  • the method further includes receiving, at the electronic device from the user, input corresponding to engagement of a play button of the video annotation interface; and repeatedly playing the video in the video pane in a loop from a frame corresponding to the position of the initial handle to a frame corresponding to the position of the final handle.
  • the received data corresponding to a plurality of comments associated with the video includes data corresponding to a particular point comment associated with the later-selected comment marker, and wherein the data corresponding to the particular point comment includes an identification of a point of the video associated with the comment.
  • the identification of a point of the video comprises a timestamp.
  • the identification of a point of the video comprises a frame identification.
  • the received data corresponding to a plurality of comments associated with the video includes data corresponding to a particular segment comment associated with the later-selected comment marker, and wherein the data corresponding to the particular segment comment includes an identification of a starting point of a segment of the video associated with the particular segment comment.
  • the data corresponding to the particular segment comment includes an identification of a length of the segment comment.
  • the data corresponding to the particular segment comment includes an identification of an end point of a segment of the video associated with the particular segment comment.
  • Another aspect of the present invention relates to a method comprising displaying, to a user via an electronic display associated with an electronic device, a login interface; receiving, at the electronic device from the user, input corresponding to login credentials of a user account of the user; communicating, from the electronic device, data corresponding to the login credentials; displaying, to the user via the electronic display, a video selection interface configured to allow the user to a select a video for viewing; receiving, at the electronic device from the user, input corresponding to selection of a video for viewing; communicating, from the electronic device, data corresponding to the selected video; and receiving, at the electronic device, data corresponding to the video, and data corresponding to a plurality of comments associated with the video, each of the plurality of comments being associated with a particular user account of a plurality of user accounts.
  • the method further includes displaying, to the user via the electronic display associated with the electronic device, a video annotation interface comprising a video pane configured to display the video, a video timeline bar including a video play-head indicating a current point of the video which is being played, a segment timeline bar including initial and final handles configured to define a segment of the video for playing, a plurality of comment markers displayed in connection with the video timeline bar, each of the plurality of comment markers corresponding to one of the plurality of comments associated with the video, a comment display pane displaying text corresponding to at least some of the plurality of comments associated with the video, and a comment button configured to allow the user to add a comment to the video.
  • a video annotation interface comprising a video pane configured to display the video, a video timeline bar including a video play-head indicating a current point of the video which is being played, a segment timeline bar including initial and final handles configured to define a segment of the video for playing, a plurality of comment markers displayed in connection with the video timeline bar,
  • the method still further includes receiving, at the electronic device from the user, input corresponding to selection of a particular portion of the video; receiving, at the electronic device from the user, input corresponding to engagement of the comment button; in response to receiving input corresponding to engagement of the comment button, displaying, to the user via the electronic display associated with the electronic device, a comment interface; receiving, at the electronic device from the user, input corresponding to one or more desired annotations; and in response to receiving input corresponding to one or more desired annotations, associating the input one or more annotations with the selected particular portion of the video, updating the video annotation interface so that the plurality of comment markers displayed in connection with the video timeline bar includes a new comment marker indicating a point on the video timeline bar corresponding to the particular portion of the video, and displaying an indication of the input one or more annotations overlaid over the video in the video pane.
  • the step of associating the input one or more annotations with the selected particular portion of the video comprises communicating, by the electronic device, data corresponding to the input one or more annotations and the selected particular portion of the video to a remote server for storage.
  • the step of associating the input one or more annotations with the selected particular portion of the video comprises storing, in a computer readable medium, an association of the input one or more annotations with the selected particular portion of the video.
  • the step of associating the input one or more annotations with the selected particular portion of the video comprises storing, in a computer readable medium at the electronic device, an association of the input one or more annotations with the selected particular portion of the video.
  • the particular portion of the video is a point.
  • the particular portion of the video is a frame.
  • the particular portion of the video is a segment.
  • receiving, at the electronic device from the user, input corresponding to one or more desired annotations includes receiving input corresponding to text for a text annotation.
  • receiving, at the electronic device from the user, input corresponding to one or more desired annotations includes receiving input corresponding to a drawing annotation.
  • receiving, at the electronic device from the user, input corresponding to one or more desired annotations includes receiving input corresponding to text for a text annotation.
  • receiving, at the electronic device from the user, input corresponding to one or more desired annotations includes receiving input corresponding to text for a text annotation and receiving input corresponding to selecting a location for the text annotation to be displayed.
  • receiving, at the electronic device from the user, input corresponding to one or more desired annotations includes receiving input corresponding to text for a text annotation and receiving input corresponding to moving the text annotation.
  • receiving, at the electronic device from the user, input corresponding to selection of a particular portion of the video comprises receiving input corresponding to selection of a segment that includes input corresponding to moving the initial handle of the segment timeline bar and moving the final handle of the segment timeline bar.
  • the method further includes loading, at the electronic device, a key code library representing an association of keys with one or more respective desired annotations, and wherein receiving, at the electronic device from the user, input corresponding to one or more desired annotations comprises receiving input corresponding to a particular key which is associated with one or more respective desired annotations in the key code library.
  • receiving, at the electronic device from the user, input corresponding to one or more desired annotations includes receiving input corresponding to selection of a color and input corresponding to a drawing annotation.
  • receiving, at the electronic device from the user, input corresponding to one or more desired annotations includes receiving input corresponding to selection of a brush size for a drawing annotation and input corresponding to a drawing annotation.
  • receiving, at the electronic device from the user, input corresponding to one or more desired annotations includes receiving input corresponding to a video annotation.
  • receiving, at the electronic device from the user, input corresponding to one or more desired annotations includes receiving input corresponding to a video annotation via a webcam.
  • receiving, at the electronic device from the user, input corresponding to one or more desired annotations includes receiving input corresponding to an audio annotation.
  • receiving, at the electronic device from the user, input corresponding to one or more desired annotations includes receiving input corresponding to an audio annotation via a microphone.
  • receiving, at the electronic device from the user, input corresponding to one or more desired annotations includes receiving input corresponding to a video or audio annotation via an indication of a file representing such an annotation.
  • the comment interface is overlaid over the video annotation interface.
  • the comment interface is overlaid over the video pane.
  • Another aspect of the present invention relates to a method comprising displaying, to a user via an electronic display associated with an electronic device, a login interface; receiving, at the electronic device from the user, input corresponding to login credentials of a user account of the user; communicating, from the electronic device, data corresponding to the login credentials; displaying, to the user via the electronic display, a video selection interface configured to allow the user to a select a video for viewing; receiving, at the electronic device from the user, input corresponding to selection of a video for viewing; communicating, from the electronic device, data corresponding to the selected video; receiving, at the electronic device, data corresponding to the video, and data corresponding to a plurality of comments associated with the video, each of the plurality of comments being associated with a particular user account of a plurality of user accounts; and displaying, to the user via the electronic display associated with the electronic device, a video annotation interface comprising a video pane configured to display the video, a video timeline bar including a video play-head indicating a current point of the video
  • the method further includes receiving, at the electronic device from the user, input corresponding to engagement of a play button of the video annotation interface; playing, in the video pane, the video; receiving, at the electronic device from the user, input corresponding to engagement of the comment button; in response to receiving input corresponding to engagement of the comment button, pausing playback of the video, and displaying, to the user via the electronic display associated with the electronic device, a comment interface; receiving, at the electronic device from the user, input corresponding to one or more desired annotations; and in response to receiving input corresponding to one or more desired annotations, associating the input one or more annotations with the selected particular portion of the video, updating the video annotation interface so that the plurality of comment markers displayed in connection with the video timeline bar includes a new comment marker indicating a point on the video timeline bar corresponding to the particular portion of the video, and displaying an indication of the input one or more annotations overlaid over the video in the video pane.
  • the method further includes loading, at the electronic device, a key code library representing an association of keys with one or more respective desired annotations, and wherein receiving, at the electronic device from the user, input corresponding to one or more desired annotations comprises receiving input corresponding to a particular key which is associated with one or more respective desired annotations in the key code library.
  • FIG. 1 illustrates the architecture of an implementation of an exemplary application
  • FIG. 2 illustrates a login interface of an exemplary application
  • FIG. 3 illustrates a video selection interface of the exemplary application of FIG. 2 ;
  • FIG. 4 illustrates a video upload interface of the exemplary application of FIG. 2 ;
  • FIG. 5 illustrates the display of properties of a video on the video selection interface of FIG. 3 ;
  • FIG. 6 illustrates a video annotation interface of the exemplary application of FIG. 2 ;
  • FIGS. 7A-C illustrate functionality of a video play-head of the video annotation interface of FIG. 6 ;
  • FIGS. 8A-H illustrate segment functionality of the video annotation interface of FIG. 6 ;
  • FIGS. 9A-I and 10 A-B illustrate commenting functionality of the video annotation interface of FIG. 6 ;
  • FIG. 11 illustrates a change log pane of the video annotation interface of FIG. 6 ;
  • FIG. 12 illustrates focus area functionality of a video annotation interface
  • FIG. 13 illustrates zoom functionality of a video annotation interface
  • FIG. 14 illustrates a journal interface of the exemplary application of FIG. 2 .
  • any embodiment may incorporate only one or a plurality of the above-disclosed aspects of the invention and may further incorporate only one or a plurality of the above-disclosed features.
  • any embodiment discussed and identified as being “preferred” is considered to be part of a best mode contemplated for carrying out the present invention.
  • Other embodiments also may be discussed for additional illustrative purposes in providing a full and enabling disclosure of the present invention.
  • any embodiment may incorporate only one or a plurality of the above-disclosed aspects of the invention and may further incorporate only one or a plurality of the above-disclosed features.
  • many embodiments, such as adaptations, variations, modifications, and equivalent arrangements, will be implicitly disclosed by the embodiments described herein and fall within the scope of the present invention.
  • any sequence(s) and/or temporal order of steps of various processes or methods that are described herein are illustrative and not restrictive. Accordingly, it should be understood that, although steps of various processes or methods may be shown and described as being in a sequence or temporal order, the steps of any such processes or methods are not limited to being carried out in any particular sequence or order, absent an indication otherwise. Indeed, the steps in such processes or methods generally may be carried out in various different sequences and orders while still falling within the scope of the present invention. Accordingly, it is intended that the scope of patent protection afforded the present invention is to be defined by the appended claims rather than the description set forth herein.
  • a picnic basket having an apple describes “a picnic basket having at least one apple” as well as “a picnic basket having apples.”
  • a picnic basket having a single apple describes “a picnic basket having only one apple.”
  • One or more preferred embodiments in accordance with the present invention represent an online collaboration system and tool.
  • a system and tool represents a web-based application that allows users to annotate a video using a variety of modalities including text, digital ink, audio, and video.
  • An interface which allows allow notes, comments or sketches to be added in the form of texts or images on a video can be characterized as an annotation interface.
  • annotation system utilized by such an application can be characterized as “multimodal” in that it provides support for multiple modes of input or output.
  • an annotation system supports comment input in the form of text, sketching, audio, and video and hence allows input in four different modes.
  • This can be contrasted with various existing video systems, such as, for example, Youtube which supports a text modality, viddler, which supports video and text modalities, bublePly, which supports clipart and text modalities, and videoANT, which supports a text modality.
  • the annotation system also supports visual indexing of annotations and a rich navigational control for inserting and editing video annotations.
  • an annotation system utilizes context based navigation, which provides a way to navigate to a certain point in time of a video by linking that point to a comment or annotation.
  • context based navigation provides a way to navigate to a certain point in time of a video by linking that point to a comment or annotation.
  • when a comment is added to a video or image it is added in context to a point or span on the timeline of video.
  • the comment registers the time of the video where the comment has been made. Later, a user can navigate through the comments utilizing either a comment-list or markers on a video timeline exposing the frame(s) of video in context of which that particular comment was made.
  • annotations are anchored to a point in the timeline of a video, but are stored externally in a separate file. It is believed that annotations can enhance the end-user experience when they are displayed in context (at a cue-point where they were made).
  • Users can upload videos and then post comments in the form of text, digital sketches, audio, and video comments. Users can thereafter log in, analyze the video and read (and listen to or view) comments posted by other users (such as, for example, choreographers) in context, and can respond with video, ink, audio, or text comments.
  • users such as, for example, choreographers
  • Such an application preferably enables synchronous collaboration as multiple users can login and work simultaneously, as well as asynchronous collaboration as different users can insert annotations across multiple sessions separated in time.
  • Exemplary synchronous collaboration might involve, for example, collaboration between two or more users located in the same or different spaces interacting with an annotation system concurrently.
  • Video-conferencing is an example of synchronous collaboration in another context.
  • Exemplary asynchronous collaboration might involve collaboration between two or more users located in the same or different spaces interacting with an annotation system at different, disjoint times.
  • Email is an example of asynchronous collaboration in another context.
  • Such video annotation systems, and related methodologies, are applicable for use in a wide variety of contexts.
  • one such context is a dance context.
  • a video annotation system might be utilized by an ornithologist to review and annotate video of a particular bird.
  • exemplary contexts include an educational context (such as online education), a physical therapy context (e.g. annotation exercises or movements), or a sports context (e.g. annotating game or practice tapes).
  • an educational context such as online education
  • a physical therapy context e.g. annotation exercises or movements
  • a sports context e.g. annotating game or practice tapes.
  • FIG. 1 illustrates the architecture of an implementation of an exemplary application, which might be configured for use in one or more of these contexts, or another context (or may be generic to two or more contexts).
  • Such implementation utilizes HTML5 and Javascript and works in all major web browsers, supporting mp4 and mov video files.
  • users can load videos, play them, and insert annotations.
  • the application sends the text, digital ink drawings, or video annotations to a MySQL database via PHP. Text is stored as a string while digital ink drawings are stored as a snapshot of the transparent canvas on which the drawing was made. Video comments are stored (e.g. as an .mov file) in a separate folder.
  • Stored annotations can then subsequently be retrieved from the MySQL database.
  • the annotations are first retrieved as a proxy XML feed, which is written dynamically using PHP, and then are rendered in the application.
  • Users can access the exemplary application via a web browser by navigating their web browser to an address associated with the application. Thereafter, if users are not already logged in (which the application might determine utilizing, for example, cookie information), users will be presented with a login interface which allows a user to login, as illustrated in FIG. 2 .
  • Each user is preferably assigned a username and password, which can be accomplished in various ways. For example, these may be pre-assigned and configured into the application, may be entered by a system administrator, may be requested by a user, or may be selected by a user.
  • Logging in both allows user comments to be associated with that user, and additionally might provide access to private videos that are not publicly available. Further, in one or more preferred implementations, each user has a color assigned to them so that each user's comments are color-coded.
  • the video selection interface includes a file display section which displays files and folders forming part of a file archival system of the application.
  • This file archival system represents a repository of files stored in different folders following a tree structure, and allows files to be stored in various named folders making it easier for users to access a required file in an efficient manner.
  • some files or folders of this archival system are private, and are only viewable and accessible by certain users who have the appropriate permissions. Such permissions may be sharable, and/or may only be grantable by a system administrator.
  • the application includes a navigation bar which allows users to effect various tasks, such as navigating between the video selection interface and various other interfaces.
  • the navigation bar allows a user to upload a video, access a journal, select a video (via the video selection interface), or log out.
  • a user can utilize the navigation bar to effect navigation to a video upload interface, which is illustrated in FIG. 4 .
  • the video upload interface includes a file display system, and allows a user to select a folder to upload a video to, as well as a video to upload.
  • the video upload interface further allows a user to add a description of the video. Preferably, this description can be edited later by the user, or, in at least some implementations, by another use with appropriate permissions.
  • a user Upon uploading a video, a user can return (and in at least some implementations is automatically taken back) to the video selection screen, where the user can then select that video (or another video) for annotation and/or review.
  • properties of that video are provided, as illustrated in FIG. 5 .
  • a user can then select that video for annotation and/or review.
  • a user Upon selecting a video, a user is taken to a video annotation interface, as illustrated in FIG. 6 , which shows the selected video and comments posted to that video by users (preferably including both comments posted by other users and comments posted by the logged in user, which may be color coded as described herein, such as, for example, by outlining each comment and/or the commenting user's name in a color associated with the commenting user).
  • the video annotation interface comprises a video pane 12 with controls therebelow, and a display pane 14 with tabs for alternately displaying a comment display pane and a change log pane.
  • the controls below the video pane 12 include a video timeline bar 20 , a segment timeline bar 30 , a play button 16 , a playback speed slider 40 , and a comment button 18 .
  • the video annotation interface allows a user to play the selected video with the play button.
  • the video timeline bar includes a video play-head 22 which indicates relatively what frame or portion of the video is currently displayed/being played, as illustrated in FIGS. 7A-7C .
  • the annotation interface displays (such as, for example, in seconds) where in playback the video play-head 22 currently is.
  • the segment timeline bar can be utilized by a user to limit playback to a particular segment of the video.
  • the segment timeline bar includes an initial position handle 32 and a final position handle 34 .
  • each position handle 32 , 34 preferably displays in association therewith a time of the video the handle is currently positioned at, as illustrated in FIG. 6 (such text is omitted for clarity in many illustrations).
  • Each of these position handles can be slid along the segment timeline bar by a user (e.g. using a mouse or a touchscreen), as illustrated by the arrows in FIGS. 8A-B , and in some implementations can even be slid to the same point, but in preferred implementations cannot be slid past one another (although in at least some implementations they may be).
  • the handles can be utilized by a user to define a segment to limit playback to (or associate a comment therewith, as described hereinbelow).
  • the handles are initially set to the start and end points of the video, and when the handles are so set the segment timeline bar appears generally similar in appearance to the video timeline bar (e.g. being displayed in the same color), but when one or both handles are moved so as to define a segment which is less than the entire video, the portion of the segment timeline bar between the handles is highlighted (e.g. is displayed in a brighter, darker, or different color, such as bright yellow).
  • a segment timeline bar is implemented utilizing HTML and a j Query UI slider library, currently available at jqueryui.com.
  • the video play-head can only be set as early as the time corresponding to the initial position handle (and will preferably be moved there if it was set to an earlier point in the video before the segment was defined, as illustrated by reference to FIGS. 8A-B ), and can only be set as late as the time corresponding to the final position handle (and will preferably be moved there if it was set to a later point in the video before the segment was defined).
  • the video will only play the portion of the video corresponding to the segment, and once it reaches the time corresponding to the final position handle, it will loop back to the time corresponding to the initial position handle.
  • the segment timeline bar can be characterized as providing a looping mechanism that works by selecting a defined segment of video and looping the playback repeatedly through that segment.
  • the mechanism involves the use of the timeline handles which can be dragged to indicate the beginning and end of a segment for looping. For example, if a dancer wants to see only a small part of video repeatedly so that he or she can master a given movement before proceeding to the next, he or she can select a segment of video he or she wants to see again and again.
  • a segment of a certain length may be a predefined or preconfigured length, such as, for example, four seconds, and/or may be calculated or determined based in whole or in part on a total length of the video. In at least some implementations, such a certain length may be wholly or partially user configurable.
  • a user can easily perform various operations to manipulate the segment. For example, a user can extend the segment by clicking on the segment timeline bar outside of the segment, in which case the handle closest to the point on the segment timeline bar the user clicked is moved to the point the user clicked, as illustrated by reference to FIG. 8C (illustrating where a user clicks with his or her mouse cursor), and FIG. 8D (which illustrates the change to the segment following such user click). Similarly, a user can shorten the segment by clicking on the segment timeline bar inside of the segment, in which case the handle closest to the point on the segment timeline bar the user clicked is moved to the point the user clicked, as illustrated by reference to FIG.
  • FIG. 8E illustrating where a user clicks with his or her mouse cursor
  • FIG. 8F which illustrates the change to the segment following such user click
  • a user can drag a defined segment across the segment timeline bar, in effect moving the start and end points of that segment, but keeping the same segment length, as illustrated by reference to FIGS. 8G-8H .
  • a user can deselect the segment, and select the entire video, by either moving the handles to the start and end points of the video, or can select the entire video (automatically moving the handles to the start and end of the video) by double clicking anywhere within the segment.
  • comments can be inserted on the video timeline bar in a plurality of different modalities by clicking the comment button.
  • a user chooses to insert a comment, he or she can choose between text, sketch, voice, video, or any combination of these modalities.
  • Sketch commenting preferably supports multi-color sketching with at least two different pen widths.
  • a user has the option of clearing the canvas if he or she is not satisfied with his or her sketch.
  • Video commenting can utilize a web-cam to capture a video comment of the user, or can utilize an existing file as a video comment.
  • a comment is added to the comment display pane and one or more markers (such as, for example, a color-coded circle, rectangle, or triangle) appears over the video timeline bar showing the position of the inserted comment in the video.
  • markers such as, for example, a color-coded circle, rectangle, or triangle
  • posted comments can be associated with either a single point in time (or frame) of a video, or with a longer segment of a video.
  • the former can be characterized as point comments, and might be represented by a first type of marker (such as, for example, a single colored rectangle) and stored in association with a single point in time, or frame, while the latter can be characterized as segment comments, and might be represented by a different type of marker (such as, for example, a colored triangle denoting a start of the segment, or even a pair of colored triangles denoting a start and an end of the segment), and stored in association with a start and end time, or start and end frames.
  • a user can add a point comment by clicking on the comment button while no segment is selected (that is, the entire video is selected for playback), as illustrated (starting from the situation in FIG. 7C ) in FIGS. 9A-H , and can add a segment comment by clicking on the comment button while a segment is selected, as illustrated (starting from the situation in FIG. 8H ) in FIGS. 10A-B .
  • a user can choose to move around the comment to overlay a different portion of the video, as illustrated via reference to FIGS. 9A-B ; the position of the comment is preferably saved such that the comment is presented to other users in that same position.
  • a user who originally posted a comment is preferably able to later edit, move, or delete such a comment, and other users may be able to do so as well if they enjoy sufficient permissions.
  • all users are preferably able to hide a displayed comment using an close button 72 , which is illustrated in FIG. 9H .
  • FIG. 9I illustrates similar display of a video comment (which video comment may be displayed with its own video timeline bar and segment timeline bar, as illustrated).
  • each user's comments are color-coded, which can help in visual identification of comments while navigating.
  • FIG. 9B illustrates an exemplary comment addition interface which is provided once a user presses the comment button.
  • the comment addition interface allows a user to input a text comment, move the position of that text comment relative to the video, select a color to use to annotate the video with using a plurality of displayed color boxes 74 (which can perhaps best be seen in FIG. 9C ), annotate the video by drawing directly on the video using the selected color, and clear the canvas of any drawn annotations (which may include multiple drawn annotations in multiple colors) using a clear button 76 (which can perhaps best be seen in FIG. 9C ).
  • the comment addition interface additionally allows a user to record a video or audio comment (e.g. utilizing a microphone and/or webcam), or upload such a video or audio comment (e.g. by selecting a file for upload).
  • video playback is paused when a point (and/or segment) associated with a comment is reached.
  • point comments may be displayed for a longer duration, such as, for example, for a period of time prior to and/or subsequent to the point in time they are associated with, which period of time may be preconfigured and/or determined/modified based on an overall length of the video.
  • Posted comments can preferably be explored several other ways as well.
  • a user can click on any comment in the comment display pane, or, alternatively, can click on a marker on the video timeline bar 20 .
  • the video play-head 22 is either set to the point in time associated with the comment (which can be characterized as a cue-point), or else to a point in time prior to the cue-point (which might be a predefined amount of time before the cue-point, and/or might be calculated based on the overall length of the video), the video is paused (although in at least some implementations it may not be), and any comments related to that cue-point are overlaid over the video.
  • the point in time associated with the comment which can be characterized as a cue-point
  • a point in time prior to the cue-point which might be a predefined amount of time before the cue-point, and/or might be calculated based on the overall length of the video
  • the video is paused (although in at least some implementations it may not be)
  • any comments related to that cue-point are overlaid over the video.
  • segment functionality of the segment timeline bar 30 is utilized when a point comment is selected to automatically define a segment around the point associated with the comment, the handles 32 , 34 of the segment timeline bar being automatically moved to define a segment having a length that might be predetermined or preconfigured (such as, for example, four seconds), or might be calculated or determined (such as, for example, based on a total length of the video).
  • different types of comments may trigger different treatment; for example, a video may be paused at a particular frame associated with a drawing if a drawn annotation is present, but may be paused and/or looped through a four second segment if a text comment is present.
  • the initial and final handles 32 , 34 are set to the positions associated with that segment comment, the play-head 22 is moved to the position corresponding to the initial handle 32 , the video is paused (although in at least some implementations it may not be), and the comment is displayed (other comments falling within that segment may or may not be displayed during playback).
  • Allowing a user to click on a comment and have a portion of the video associated therewith played back (together with the comment) delivers comments in context, which makes it easy for users to correlate comments with the portion of the video. For example, dancers watching a video can easily correlate comments with their movements.
  • a user can hover a cursor over a comment marker, and a preview of the comment (preferably with a frame of the video the comment is associated with) is displayed.
  • a preview may be displayed in the video pane in the form of the frame the comment is associated with and the comment in its position relative to the video, and/or a preview may be displayed in the form of the comment displayed above the comment marker.
  • the portion of the video corresponding to that segment is highlighted, for example on the segment timeline bar or the video timeline bar.
  • the video annotation interface includes a change log pane which is configured to track changes and/or user interaction with the application.
  • the application preferably includes self-documenting archival functionality which logs, for example, the creation, editing, and deletion of comments. Such changes are displayed in the change log pane, and in one or more preferred implementations, the application allows changes to be undone. Additionally, the application preferably logs other types of user interaction with the video annotation interface as well, as illustrated in FIG. 11 .
  • the video annotation interface includes focus area functionality which allows a user to define a focus area for a comment, such as, for example, a rectangular, circular, or irregularly shaped (and user defined) focus area which corresponds to a portion of the video the user wants other users to focus on when reviewing his or her comment.
  • a focus area may be defined by drawing a particular shape on the video pane, and/or by cropping out an area of the video pane.
  • the portions of the frames outside of the defined focus area are grayed out, as illustrated in FIG. 12 .
  • the video annotation interface allows a user to define a zoom area, and the frames of the video associated with the zoom area are displayed zoomed in on the zoom area when the comment is selected for playback, as illustrated in FIG. 13 (and, in at least some implementations, during normal play back too if all comments are selected for display, although in at least some implementations zooming may not be enabled in such situations).
  • the video annotation interface provides the ability to define a measurement scale and take measurements based on such scale.
  • An exemplary such methodology involves allowing a user to first establish a baseline measurement by drawing a line segment over the video pane that corresponds with a certain displayed element and enter a measurement corresponding to that displayed element and line segment. For example, if a user knows that the beak of a particular bird displayed in a video is 3 cm, then a user can draw a line segment over the beak and input 3 cm (preferably, the application alternatively allows for simple entry of a ratio, for example, if a user knows that 10 cm onscreen corresponds to 3 cm in the video, the user can simply enter that ratio).
  • the user can draw additional line segments over frames of the video and the application will calculate, based on the scale defined by the 3 cm beak, the relative length of whatever is displayed in the video based on the drawn line segment.
  • the video annotation interface includes a playback speed slider, which allows a user to control a playback speed of the video.
  • the playback speed slider preferably includes a selection handle which a user can slide along to vary the playback speed.
  • the playback speed slider enables slow motion playback, that is, play back of a video at a slower rate than its normal speed.
  • a normal playback speed is roughly 30 frames/second, so reducing a playback speed could, for example, effect playback in slow motion at a speed of 20 frames/second, 15 frames/second, 10 frames/second, 5 frames/second, etc.
  • the playback speed slider control can vary the playback speed from 0.1 ⁇ to 2 ⁇ , with a midpoint of the slider control corresponding to normal speed (1 ⁇ ), although in at least some other implementations these values may be different (and may even be user configurable).
  • a user can associate a point or segment comment with a particular playback speed so that the if a user later selects the comment, the portion of video corresponding to that comment will be played back at the associated speed. In one or more preferred implementations, this might be characterized as saving the state of a comment, such as the state of the playback speed slider.
  • the application includes functionality which allows a user to automatically generate a comment with a single key press.
  • a user For example, an ornithologist watching a video of a particular bird might want to annotate the video with bird behaviors that occur during the video, such as turning left, turning right, chirping, etc.
  • the application is preferably configured to allow the user to automatically generate certain configured comments with a single key press, such as automatically generating and saving a “left turn” comment by pressing “L”, or automatically generating and saving a “chirp” comment by pressing “C”.
  • these configured automatically generated comments are capable of being user configured.
  • the application is configured to utilize key code (or key press) libraries which configure what keys will automatically generate what comment.
  • a key press library might comprise a comma separated values file in which each line includes an indication of a key (such as “C”) and an indication of a comment to automatically generate (such as “chirp”).
  • Such a file might additionally include an (x,y) position on the video to anchor the comment, and/or an indication of a segment to create based on the key press (for example, the file might indicate to create a segment starting x seconds before the point at which the key was pressed and ending x seconds after the point at which the key was pressed, where either or both x's might be zero).
  • an application might provide key code (or key press) library functionality by allowing a user to upload and/or load one or more key press files.
  • the application includes a full screen mode which allows users to view a video in full screen mode.
  • this full screen mode includes comment markers displayed proximate a top or bottom of the screen.
  • the full screen mode includes one or both timeline bars displayed proximate a top or bottom of the screen.
  • the application is configured for use with a tablet or slate computer, such as an iPad, or a smart phone, such as an iPhone, Android, or Blackberry.
  • the application is configured to toggle between a normal mode and a full screen mode as an orientation of the device the application is running on is transitioned from a landscape orientation to a portrait orientation, or vice versa (e.g. based on accelerometer readings of the device).
  • commenting is configured to allow users to respond to comments made by other users, and responsive comments may be grouped in various ways to indicate that they are responsive to a previous comment.
  • comments may form a comment thread, and/or may all be grouped together in a manner that pictorially or numerically indicates that multiple comments are grouped or associated.
  • the application preferably includes a journal which can be utilized by users.
  • a journal which can be utilized by users.
  • choreographers and dancers often keep a journal documenting notes on their progress and ideas and images to be explored (in some production processes a journal is an explicit part of the development); the application includes journal functionality which can be used by choreographers or dancers to record and review private thoughts. So if a dancer got some instruction from a choreographer or if her or she made a mental note about something to remember, he or she can write it in his or her journal.
  • a user can access this journal functionality anytime while using the application, once logged in, and preferably journal entries are private to a user and cannot be seen by other users.
  • journal interface allows users to access a journal interface.
  • FIG. 14 illustrates such journal interface.
  • the journal interface allows a user to write and submit new journal entries, and review, edit, and delete older journal entries.

Abstract

A video annotation interface includes a video pane configured to display a video, a video timeline bar including a video play-head indicating a current point of the video which is being played, a segment timeline bar including initial and final handles configured to define a segment of the video for playing, and a plurality of color-coded comment markers displayed in connection with the video timeline bar. Each of the comment markers is associated with a frame or segment of the video and corresponds to one or more annotations for that frame or segment made by one of a plurality of users. Each of the users can make annotations and view annotations made by other users. The annotations can include annotations corresponding to a plurality of modalities, including text, drawing, video, and audio modalities.

Description

    CROSS-REFERENCE TO RELATED APPLICATIONS
  • The present application is a U.S. continuation patent application of, and claims priority under 35 U.S.C. § 120 to, U.S. nonprovisional patent application Ser. No. 15/144,976, filed May 3, 2016, which '976 application and any publications thereof and patents issuing therefrom are hereby incorporated herein by reference, and which '976 application is a U.S. continuation patent application of, and claims priority under 35 U.S.C. § 120 to, U.S. nonprovisional patent application Ser. No. 13/627,899, filed Sep. 26, 2012, which '899 application and any publications thereof and patents issuing therefrom are hereby incorporated herein by reference, and which '899 application is a U.S. nonprovisional patent application of, and claims priority under 35 U.S.C. § 119(e) to, U.S. provisional patent application Ser. No. 61/539,147, filed Sep. 26, 2011, which provisional patent application is hereby incorporated herein by reference. A copy of this provisional patent application is attached hereto as Appendix A, which Appendix is hereby incorporated herein by reference.
  • The present application also incorporates herein by reference the documents attached hereto as Appendices B, C, and D.
  • GOVERNMENT SUPPORT
  • One or more inventions disclosed herein was made with Government support under award number IIS-0855882 by the National Science Foundation. The Government has certain rights in such inventions.
  • COPYRIGHT STATEMENT
  • All of the material in this patent document is subject to copyright protection under the copyright laws of the United States and other countries. The copyright owner has no objection to the facsimile reproduction by anyone of the patent document or the patent disclosure, as it appears in official governmental records but, otherwise, all other copyright rights whatsoever are reserved.
  • BACKGROUND OF THE INVENTION
  • The present invention generally relates to collaborative video annotation.
  • Video annotation interfaces to facilitate communication between different stakeholders have been explored by different groups. Interfaces have been developed to support multimedia annotation as educational tools for the distance learning market, such as the Classroom 2000 project and the Microsoft Research Annotation System. The “WACTool” supports P2P capabilities that allow user collaboration and content sharing using voice and digital ink comments over individual frames of a video. “Project Pad” is a standalone web application for adding text notes to media clips and sharing those notes with others. “Video Image Annotation” tool is a Windows application that provides an interface to manually annotate a video with text. “VideoANT” is a web-based video annotation application developed for e-learning, to annotate videos with time point markers and text. “Advene” allows people to design and exchange videos with annotations. “Anvil” is another tool that allows annotation of audiovisual material by inserting time-anchored comments and gesture elements.
  • There are many video-based interfaces with text annotation that are commercially available. Youtube allows user to annotate videos they upload, and others can see the annotation. “BubblePLY” is a web application that allows users to annotate remote videos hosted on other websites such as Youtube. Users can add text, drawing, pre-defined clipart, subtitles etc. on a remote video to create bubbles (annotated videos) and can edit their own bubbles. “veotag” and “viddler” also support annotations of video content. Sports and News broadcast domains have also seen a significant amount of work dealing with video annotations.
  • The applications discussed typically support annotations in only one form (mostly text) and lack multimodal capabilities. None of these applications support fine-grained access and navigational control, multimodal annotation grouping, rich annotation positioning and user-based color coding. All of the applications discussed allow some form of video annotation but none of them comprehensively address the issues of collaboratively working on a creative temporal project, such as dance.
  • A need exists for improvement in collaborative video annotation. This, and other needs, are addressed by one or more aspects of the present invention.
  • SUMMARY OF THE INVENTION
  • The present invention includes many aspects and features. Moreover, while many aspects and features relate to, and are described in, the context of collaborative video annotation, the present invention is not limited to use only in this context, as will become apparent from the following summaries and detailed descriptions of aspects, features, and one or more embodiments of the present invention.
  • Accordingly, one aspect of the present invention relates to a method comprising displaying, to a user via an electronic display associated with an electronic device, a login interface; receiving, at the electronic device from the user, input corresponding to login credentials of a user account of the user; communicating, from the electronic device, data corresponding to the login credentials; displaying, to the user via the electronic display, a video selection interface configured to allow the user to a select a video for viewing; receiving, at the electronic device from the user, input corresponding to selection of a video for viewing; communicating, from the electronic device, data corresponding to the selected video; receiving, at the electronic device, data corresponding to the video, and data corresponding to a plurality of comments associated with the video, each of the plurality of comments being associated with a particular user account of a plurality of user accounts. The method further includes displaying, to the user via the electronic display associated with the electronic device, a video annotation interface comprising a video pane configured to display the video, a video timeline bar including a video play-head indicating a current point of the video which is being played, a segment timeline bar including initial and final handles configured to define a segment of the video for playing, a plurality of comment markers displayed in connection with the video timeline bar, each of the plurality of comment markers corresponding to one of the plurality of comments associated with the video, a comment display pane displaying text corresponding to at least some of the plurality of comments associated with the video, and a comment button configured to allow the user to add a comment to the video. The method still further includes receiving, at the electronic device from the user, input corresponding to selection of a displayed comment marker; and in response to receiving input corresponding to selection of a displayed comment marker, automatically moving the video play-head to a point on the video determined based on the selected comment marker, and displaying, in the video pane, a frame corresponding to the position of the video play-head, and displaying, in the video pane overlaid over the displayed frame of the video, one or more annotations associated with the comment marker.
  • In a feature of this aspect, automatically moving the video play-head to a point on the video determined based on the selected comment marker comprises automatically moving the video play-head to a point of the video a predetermined period of time prior to a point on the video associated with the comment marker. In at least some implementations, the predetermined period of time is two seconds.
  • In a feature of this aspect, in addition to automatically moving the video play-head to a point on the video determined based on the selected comment marker, the method includes automatically defining a segment corresponding to the selected comment marker utilizing the segment timeline bar by moving the initial handle of the segment timeline bar to a beginning point corresponding to the point on the video determined based on the selected comment marker, and automatically moving the final handle of the segment timeline bar to an end point corresponding to a second point on the video determined based on the selected comment. In at least some implementations, the method further includes receiving, at the electronic device from the user, input corresponding to engagement of a play button of the video annotation interface; and repeatedly playing the video in the video pane in a loop from a frame corresponding to the position of the initial handle to a frame corresponding to the position of the final handle. In at least some implementations, the received data corresponding to a plurality of comments associated with the video includes data corresponding to a particular point comment associated with the later-selected comment marker, and wherein the data corresponding to the particular point comment includes an identification of a point of the video associated with the comment. In at least some implementations, the identification of a point of the video comprises a timestamp. In at least some implementations, the identification of a point of the video comprises a frame identification. In at least some implementations, the received data corresponding to a plurality of comments associated with the video includes data corresponding to a particular segment comment associated with the later-selected comment marker, and wherein the data corresponding to the particular segment comment includes an identification of a starting point of a segment of the video associated with the particular segment comment. In at least some implementations, the data corresponding to the particular segment comment includes an identification of a length of the segment comment. In at least some implementations, the data corresponding to the particular segment comment includes an identification of an end point of a segment of the video associated with the particular segment comment.
  • In a feature of this aspect, the displayed video annotation interface further comprises a playback speed slider configured to allow the user to vary the speed of playback of the video in the video pane.
  • In a feature of this aspect, the step of receiving, at the electronic device, data corresponding to the video, and data corresponding to a plurality of comments associated with the video comprises receiving such data at the electronic device via a communication adapter.
  • In a feature of this aspect, the electronic display is part of the electronic device.
  • In a feature of this aspect, the electronic device is a tablet, slate computer, or smartphone.
  • In a feature of this aspect, the electronic display is connected to the electronic device.
  • In a feature of this aspect, the electronic display is a touchscreen display, and wherein at least one of the steps of receiving, at the electronic device from the user, input, comprises receiving input from the user input via the touchscreen display.
  • In a feature of this aspect, the method further includes, in response to receiving input corresponding to selection of a displayed comment marker, highlighting a portion of the video displayed in the video pane that corresponds to a focus area associated with a comment corresponding to the comment marker.
  • In a feature of this aspect, the method further includes, in response to receiving input corresponding to selection of a displayed comment marker, zooming in on a portion of the video displayed in the video pane that corresponds to a zoom area associated with a comment corresponding to the comment marker.
  • In a feature of this aspect, the method further includes receiving, at the electronic device from the user, input corresponding to selection of a baseline definition tool; receiving, at the electronic device from the user, input corresponding to drawing a baseline line segment on the video pane; receiving, at the electronic device from the user, input representing a baseline measurement for the user-drawn baseline line segment; receiving, at the electronic device from the user, input corresponding to drawing a measurement line segment on the video pane; automatically determining a measurement for the measurement line segment based on a length of the user-drawn baseline line segment, the user input representing a baseline measurement for the user-drawn baseline line segment, and a length of the user-drawn measurement line segment. In this feature, the method still further includes displaying, to the user via the electronic display, the automatically determined measurement for the measurement line segment.
  • In a feature of this aspect, each of the plurality of comment markers is color coded based on the particular user account associated with a comment the comment marker is associated with.
  • In a feature of this aspect, the one or more annotations includes a text annotation.
  • In a feature of this aspect, the one or more annotations includes a video annotation.
  • In a feature of this aspect, the one or more annotations includes an audio annotation.
  • In a feature of this aspect, the one or more annotations includes a drawing annotation.
  • In a feature of this aspect, the one or more annotations includes annotations in a plurality of different modalities.
  • In a feature of this aspect, wherein the one or more annotations includes annotations in two or more of the following modalities: text, drawing, video, audio.
  • Another aspect of the present invention relates to a method comprising receiving, at an electronic device, data corresponding to a video, and data corresponding to a plurality of comments associated with the video, each of the plurality of comments being associated with a particular user account of a plurality of user accounts; displaying, to a user via an electronic display associated with the electronic device, a video annotation interface comprising a video pane configured to display the video, a video timeline bar including a video play-head indicating a current point of the video which is being played, a segment timeline bar including initial and final handles configured to define a segment of the video for playing, and a plurality of comment markers displayed in connection with the video timeline bar, each of the plurality of comment markers corresponding to one of the plurality of comments associated with the video. The method further includes receiving, at the electronic device from the user, input corresponding to selection of a displayed comment marker; and in response to receiving input corresponding to selection of a displayed comment marker, automatically moving the video play-head to a point on the video determined based on the selected comment marker, and displaying, in the video pane, a frame corresponding to the position of the video play-head, and displaying, in the video pane overlaid over the displayed frame of the video, one or more comments associated with the comment marker.
  • In a feature of this aspect, automatically moving the video play-head to a point on the video determined based on the selected comment marker comprises automatically moving the video play-head to a point of the video a predetermined period of time prior to a point on the video associated with the comment marker. In at least some implementations, the predetermined period of time is two seconds.
  • In a feature of this aspect, in addition to automatically moving the video play-head to a point on the video determined based on the selected comment marker, the method includes automatically defining a segment corresponding to the selected comment marker utilizing the segment timeline bar by moving the initial handle of the segment timeline bar to a beginning point corresponding to the point on the video determined based on the selected comment marker, and automatically moving the final handle of the segment timeline bar to an end point corresponding to a second point on the video determined based on the selected comment. In at least some implementations, the method further includes receiving, at the electronic device from the user, input corresponding to engagement of a play button of the video annotation interface; and repeatedly playing the video in the video pane in a loop from a frame corresponding to the position of the initial handle to a frame corresponding to the position of the final handle. In at least some implementations, the received data corresponding to a plurality of comments associated with the video includes data corresponding to a particular point comment associated with the later-selected comment marker, and wherein the data corresponding to the particular point comment includes an identification of a point of the video associated with the comment. In at least some implementations, the identification of a point of the video comprises a timestamp. In at least some implementations, the identification of a point of the video comprises a frame identification. In at least some implementations, the received data corresponding to a plurality of comments associated with the video includes data corresponding to a particular segment comment associated with the later-selected comment marker, and wherein the data corresponding to the particular segment comment includes an identification of a starting point of a segment of the video associated with the particular segment comment. In at least some implementations, the data corresponding to the particular segment comment includes an identification of a length of the segment comment. In at least some implementations, the data corresponding to the particular segment comment includes an identification of an end point of a segment of the video associated with the particular segment comment.
  • Another aspect of the present invention relates to a method comprising displaying, to a user via an electronic display associated with an electronic device, a login interface; receiving, at the electronic device from the user, input corresponding to login credentials of a user account of the user; communicating, from the electronic device, data corresponding to the login credentials; displaying, to the user via the electronic display, a video selection interface configured to allow the user to a select a video for viewing; receiving, at the electronic device from the user, input corresponding to selection of a video for viewing; communicating, from the electronic device, data corresponding to the selected video; and receiving, at the electronic device, data corresponding to the video, and data corresponding to a plurality of comments associated with the video, each of the plurality of comments being associated with a particular user account of a plurality of user accounts. The method further includes displaying, to the user via the electronic display associated with the electronic device, a video annotation interface comprising a video pane configured to display the video, a video timeline bar including a video play-head indicating a current point of the video which is being played, a segment timeline bar including initial and final handles configured to define a segment of the video for playing, a plurality of comment markers displayed in connection with the video timeline bar, each of the plurality of comment markers corresponding to one of the plurality of comments associated with the video, a comment display pane displaying text corresponding to at least some of the plurality of comments associated with the video, and a comment button configured to allow the user to add a comment to the video. The method still further includes receiving, at the electronic device from the user, input corresponding to selection of a particular portion of the video; receiving, at the electronic device from the user, input corresponding to engagement of the comment button; in response to receiving input corresponding to engagement of the comment button, displaying, to the user via the electronic display associated with the electronic device, a comment interface; receiving, at the electronic device from the user, input corresponding to one or more desired annotations; and in response to receiving input corresponding to one or more desired annotations, associating the input one or more annotations with the selected particular portion of the video, updating the video annotation interface so that the plurality of comment markers displayed in connection with the video timeline bar includes a new comment marker indicating a point on the video timeline bar corresponding to the particular portion of the video, and displaying an indication of the input one or more annotations overlaid over the video in the video pane.
  • In a feature of this aspect, the step of associating the input one or more annotations with the selected particular portion of the video comprises communicating, by the electronic device, data corresponding to the input one or more annotations and the selected particular portion of the video to a remote server for storage.
  • In a feature of this aspect, the step of associating the input one or more annotations with the selected particular portion of the video comprises storing, in a computer readable medium, an association of the input one or more annotations with the selected particular portion of the video.
  • In a feature of this aspect, the step of associating the input one or more annotations with the selected particular portion of the video comprises storing, in a computer readable medium at the electronic device, an association of the input one or more annotations with the selected particular portion of the video.
  • In a feature of this aspect, the particular portion of the video is a point.
  • In a feature of this aspect, the particular portion of the video is a frame.
  • In a feature of this aspect, the particular portion of the video is a segment.
  • In a feature of this aspect, receiving, at the electronic device from the user, input corresponding to one or more desired annotations includes receiving input corresponding to text for a text annotation.
  • In a feature of this aspect, receiving, at the electronic device from the user, input corresponding to one or more desired annotations includes receiving input corresponding to a drawing annotation.
  • In a feature of this aspect, receiving, at the electronic device from the user, input corresponding to one or more desired annotations includes receiving input corresponding to text for a text annotation.
  • In a feature of this aspect, receiving, at the electronic device from the user, input corresponding to one or more desired annotations includes receiving input corresponding to text for a text annotation and receiving input corresponding to selecting a location for the text annotation to be displayed.
  • In a feature of this aspect, receiving, at the electronic device from the user, input corresponding to one or more desired annotations includes receiving input corresponding to text for a text annotation and receiving input corresponding to moving the text annotation.
  • In a feature of this aspect, receiving, at the electronic device from the user, input corresponding to selection of a particular portion of the video comprises receiving input corresponding to selection of a segment that includes input corresponding to moving the initial handle of the segment timeline bar and moving the final handle of the segment timeline bar.
  • In a feature of this aspect, the method further includes loading, at the electronic device, a key code library representing an association of keys with one or more respective desired annotations, and wherein receiving, at the electronic device from the user, input corresponding to one or more desired annotations comprises receiving input corresponding to a particular key which is associated with one or more respective desired annotations in the key code library.
  • In a feature of this aspect, receiving, at the electronic device from the user, input corresponding to one or more desired annotations includes receiving input corresponding to selection of a color and input corresponding to a drawing annotation.
  • In a feature of this aspect, receiving, at the electronic device from the user, input corresponding to one or more desired annotations includes receiving input corresponding to selection of a brush size for a drawing annotation and input corresponding to a drawing annotation.
  • In a feature of this aspect, receiving, at the electronic device from the user, input corresponding to one or more desired annotations includes receiving input corresponding to a video annotation.
  • In a feature of this aspect, receiving, at the electronic device from the user, input corresponding to one or more desired annotations includes receiving input corresponding to a video annotation via a webcam.
  • In a feature of this aspect, receiving, at the electronic device from the user, input corresponding to one or more desired annotations includes receiving input corresponding to an audio annotation.
  • In a feature of this aspect, receiving, at the electronic device from the user, input corresponding to one or more desired annotations includes receiving input corresponding to an audio annotation via a microphone.
  • In a feature of this aspect, receiving, at the electronic device from the user, input corresponding to one or more desired annotations includes receiving input corresponding to a video or audio annotation via an indication of a file representing such an annotation.
  • In a feature of this aspect, the comment interface is overlaid over the video annotation interface.
  • In a feature of this aspect, the comment interface is overlaid over the video pane.
  • Another aspect of the present invention relates to a method comprising displaying, to a user via an electronic display associated with an electronic device, a login interface; receiving, at the electronic device from the user, input corresponding to login credentials of a user account of the user; communicating, from the electronic device, data corresponding to the login credentials; displaying, to the user via the electronic display, a video selection interface configured to allow the user to a select a video for viewing; receiving, at the electronic device from the user, input corresponding to selection of a video for viewing; communicating, from the electronic device, data corresponding to the selected video; receiving, at the electronic device, data corresponding to the video, and data corresponding to a plurality of comments associated with the video, each of the plurality of comments being associated with a particular user account of a plurality of user accounts; and displaying, to the user via the electronic display associated with the electronic device, a video annotation interface comprising a video pane configured to display the video, a video timeline bar including a video play-head indicating a current point of the video which is being played, a segment timeline bar including initial and final handles configured to define a segment of the video for playing, a plurality of comment markers displayed in connection with the video timeline bar, each of the plurality of comment markers corresponding to one of the plurality of comments associated with the video, a comment display pane displaying text corresponding to at least some of the plurality of comments associated with the video, and a comment button configured to allow the user to add a comment to the video. The method further includes receiving, at the electronic device from the user, input corresponding to engagement of a play button of the video annotation interface; playing, in the video pane, the video; receiving, at the electronic device from the user, input corresponding to engagement of the comment button; in response to receiving input corresponding to engagement of the comment button, pausing playback of the video, and displaying, to the user via the electronic display associated with the electronic device, a comment interface; receiving, at the electronic device from the user, input corresponding to one or more desired annotations; and in response to receiving input corresponding to one or more desired annotations, associating the input one or more annotations with the selected particular portion of the video, updating the video annotation interface so that the plurality of comment markers displayed in connection with the video timeline bar includes a new comment marker indicating a point on the video timeline bar corresponding to the particular portion of the video, and displaying an indication of the input one or more annotations overlaid over the video in the video pane.
  • In a feature of this aspect, the method further includes loading, at the electronic device, a key code library representing an association of keys with one or more respective desired annotations, and wherein receiving, at the electronic device from the user, input corresponding to one or more desired annotations comprises receiving input corresponding to a particular key which is associated with one or more respective desired annotations in the key code library.
  • In addition to the aforementioned aspects and features of the present invention, it should be noted that the present invention further encompasses the various possible combinations and subcombinations of such aspects and features. Thus, for example, any aspect may be combined with an aforementioned feature in accordance with the present invention without requiring any other aspect or feature.
  • BRIEF DESCRIPTION OF THE DRAWINGS
  • One or more preferred embodiments of the present invention now will be described in detail with reference to the accompanying drawings, wherein the same elements are referred to with the same reference numerals, and wherein:
  • FIG. 1 illustrates the architecture of an implementation of an exemplary application;
  • FIG. 2 illustrates a login interface of an exemplary application;
  • FIG. 3 illustrates a video selection interface of the exemplary application of FIG. 2;
  • FIG. 4 illustrates a video upload interface of the exemplary application of FIG. 2;
  • FIG. 5 illustrates the display of properties of a video on the video selection interface of FIG. 3;
  • FIG. 6 illustrates a video annotation interface of the exemplary application of FIG. 2;
  • FIGS. 7A-C illustrate functionality of a video play-head of the video annotation interface of FIG. 6;
  • FIGS. 8A-H illustrate segment functionality of the video annotation interface of FIG. 6;
  • FIGS. 9A-I and 10A-B illustrate commenting functionality of the video annotation interface of FIG. 6;
  • FIG. 11 illustrates a change log pane of the video annotation interface of FIG. 6;
  • FIG. 12 illustrates focus area functionality of a video annotation interface;
  • FIG. 13 illustrates zoom functionality of a video annotation interface; and
  • FIG. 14 illustrates a journal interface of the exemplary application of FIG. 2.
  • DETAILED DESCRIPTION
  • As a preliminary matter, it will readily be understood by one having ordinary skill in the relevant art (“Ordinary Artisan”) that the present invention has broad utility and application. As should be understood, any embodiment may incorporate only one or a plurality of the above-disclosed aspects of the invention and may further incorporate only one or a plurality of the above-disclosed features. Furthermore, any embodiment discussed and identified as being “preferred” is considered to be part of a best mode contemplated for carrying out the present invention. Other embodiments also may be discussed for additional illustrative purposes in providing a full and enabling disclosure of the present invention. As should be understood, any embodiment may incorporate only one or a plurality of the above-disclosed aspects of the invention and may further incorporate only one or a plurality of the above-disclosed features. Moreover, many embodiments, such as adaptations, variations, modifications, and equivalent arrangements, will be implicitly disclosed by the embodiments described herein and fall within the scope of the present invention.
  • Accordingly, while the present invention is described herein in detail in relation to one or more embodiments, it is to be understood that this disclosure is illustrative and exemplary of the present invention, and is made merely for the purposes of providing a full and enabling disclosure of the present invention. The detailed disclosure herein of one or more embodiments is not intended, nor is to be construed, to limit the scope of patent protection afforded the present invention, which scope is to be defined by the claims and the equivalents thereof. It is not intended that the scope of patent protection afforded the present invention be defined by reading into any claim a limitation found herein that does not explicitly appear in the claim itself.
  • Thus, for example, any sequence(s) and/or temporal order of steps of various processes or methods that are described herein are illustrative and not restrictive. Accordingly, it should be understood that, although steps of various processes or methods may be shown and described as being in a sequence or temporal order, the steps of any such processes or methods are not limited to being carried out in any particular sequence or order, absent an indication otherwise. Indeed, the steps in such processes or methods generally may be carried out in various different sequences and orders while still falling within the scope of the present invention. Accordingly, it is intended that the scope of patent protection afforded the present invention is to be defined by the appended claims rather than the description set forth herein.
  • Additionally, it is important to note that each term used herein refers to that which the Ordinary Artisan would understand such term to mean based on the contextual use of such term herein. To the extent that the meaning of a term used herein—as understood by the Ordinary Artisan based on the contextual use of such term—differs in any way from any particular dictionary definition of such term, it is intended that the meaning of the term as understood by the Ordinary Artisan should prevail.
  • Regarding applicability of 35 U.S.C. § 112, ¶6, no claim element is intended to be read in accordance with this statutory provision unless the explicit phrase “means for” or “step for” is actually used in such claim element, whereupon this statutory provision is intended to apply in the interpretation of such claim element.
  • Furthermore, it is important to note that, as used herein, “a” and “an” each generally denotes “at least one,” but does not exclude a plurality unless the contextual use dictates otherwise. Thus, reference to “a picnic basket having an apple” describes “a picnic basket having at least one apple” as well as “a picnic basket having apples.” In contrast, reference to “a picnic basket having a single apple” describes “a picnic basket having only one apple.”
  • When used herein to join a list of items, “or” denotes “at least one of the items,” but does not exclude a plurality of items of the list. Thus, reference to “a picnic basket having cheese or crackers” describes “a picnic basket having cheese without crackers”, “a picnic basket having crackers without cheese”, and “a picnic basket having both cheese and crackers.” Finally, when used herein to join a list of items, “and” denotes “all of the items of the list.” Thus, reference to “a picnic basket having cheese and crackers” describes “a picnic basket having cheese, wherein the picnic basket further has crackers,” as well as describes “a picnic basket having crackers, wherein the picnic basket further has cheese.”
  • Referring now to the drawings, one or more preferred embodiments of the present invention are next described. The following description of one or more preferred embodiments is merely exemplary in nature and is in no way intended to limit the invention, its implementations, or uses.
  • One or more preferred embodiments in accordance with the present invention represent an online collaboration system and tool. In one or more preferred implementations, such a system and tool represents a web-based application that allows users to annotate a video using a variety of modalities including text, digital ink, audio, and video. An interface which allows allow notes, comments or sketches to be added in the form of texts or images on a video can be characterized as an annotation interface.
  • The annotation system utilized by such an application can be characterized as “multimodal” in that it provides support for multiple modes of input or output. For example, in one or more preferred implementations, an annotation system supports comment input in the form of text, sketching, audio, and video and hence allows input in four different modes. This can be contrasted with various existing video systems, such as, for example, Youtube which supports a text modality, viddler, which supports video and text modalities, bublePly, which supports clipart and text modalities, and videoANT, which supports a text modality.
  • The annotation system also supports visual indexing of annotations and a rich navigational control for inserting and editing video annotations. In one or more preferred implementations, an annotation system utilizes context based navigation, which provides a way to navigate to a certain point in time of a video by linking that point to a comment or annotation. In an exemplary such implementation, when a comment is added to a video or image, it is added in context to a point or span on the timeline of video. The comment registers the time of the video where the comment has been made. Later, a user can navigate through the comments utilizing either a comment-list or markers on a video timeline exposing the frame(s) of video in context of which that particular comment was made.
  • Preferably, annotations are anchored to a point in the timeline of a video, but are stored externally in a separate file. It is believed that annotations can enhance the end-user experience when they are displayed in context (at a cue-point where they were made).
  • Users can upload videos and then post comments in the form of text, digital sketches, audio, and video comments. Users can thereafter log in, analyze the video and read (and listen to or view) comments posted by other users (such as, for example, choreographers) in context, and can respond with video, ink, audio, or text comments.
  • Such an application preferably enables synchronous collaboration as multiple users can login and work simultaneously, as well as asynchronous collaboration as different users can insert annotations across multiple sessions separated in time. Exemplary synchronous collaboration might involve, for example, collaboration between two or more users located in the same or different spaces interacting with an annotation system concurrently. Video-conferencing is an example of synchronous collaboration in another context. Exemplary asynchronous collaboration might involve collaboration between two or more users located in the same or different spaces interacting with an annotation system at different, disjoint times. Email is an example of asynchronous collaboration in another context.
  • Such video annotation systems, and related methodologies, are applicable for use in a wide variety of contexts. For example, as noted elsewhere herein, one such context is a dance context.
  • In the dance production process, choreographers often use video cameras to record a rehearsal, and, then, all participants might sit on a studio floor and discuss and critique the rehearsal, for example by watching the video on a small TV. However, the nature of playback, such as, for example, as VCR playback, might restrict the ability to navigate the rehearsal video. It is believed that a digital video interface with annotation support would ease the rehearsal process, and a collaborative video annotation system is contemplated for use in a dance context.
  • Another exemplary context is a science or research context. For example, in an exemplary methodology of use, a video annotation system might be utilized by an ornithologist to review and annotate video of a particular bird.
  • Other exemplary contexts include an educational context (such as online education), a physical therapy context (e.g. annotation exercises or movements), or a sports context (e.g. annotating game or practice tapes).
  • It will be appreciated that these are merely exemplary contexts, and many other contexts are contemplated as well.
  • FIG. 1 illustrates the architecture of an implementation of an exemplary application, which might be configured for use in one or more of these contexts, or another context (or may be generic to two or more contexts). Such implementation utilizes HTML5 and Javascript and works in all major web browsers, supporting mp4 and mov video files. After logging in, users can load videos, play them, and insert annotations. The application sends the text, digital ink drawings, or video annotations to a MySQL database via PHP. Text is stored as a string while digital ink drawings are stored as a snapshot of the transparent canvas on which the drawing was made. Video comments are stored (e.g. as an .mov file) in a separate folder. Stored annotations can then subsequently be retrieved from the MySQL database. Preferably, the annotations are first retrieved as a proxy XML feed, which is written dynamically using PHP, and then are rendered in the application.
  • Users can access the exemplary application via a web browser by navigating their web browser to an address associated with the application. Thereafter, if users are not already logged in (which the application might determine utilizing, for example, cookie information), users will be presented with a login interface which allows a user to login, as illustrated in FIG. 2. Each user is preferably assigned a username and password, which can be accomplished in various ways. For example, these may be pre-assigned and configured into the application, may be entered by a system administrator, may be requested by a user, or may be selected by a user.
  • Logging in both allows user comments to be associated with that user, and additionally might provide access to private videos that are not publicly available. Further, in one or more preferred implementations, each user has a color assigned to them so that each user's comments are color-coded.
  • Upon logging in, users are presented with a video selection interface, as illustrated in FIG. 3. The video selection interface includes a file display section which displays files and folders forming part of a file archival system of the application. This file archival system represents a repository of files stored in different folders following a tree structure, and allows files to be stored in various named folders making it easier for users to access a required file in an efficient manner. In one or more preferred implementations, some files or folders of this archival system are private, and are only viewable and accessible by certain users who have the appropriate permissions. Such permissions may be sharable, and/or may only be grantable by a system administrator.
  • The application includes a navigation bar which allows users to effect various tasks, such as navigating between the video selection interface and various other interfaces. For example, the navigation bar allows a user to upload a video, access a journal, select a video (via the video selection interface), or log out.
  • A user can utilize the navigation bar to effect navigation to a video upload interface, which is illustrated in FIG. 4. The video upload interface includes a file display system, and allows a user to select a folder to upload a video to, as well as a video to upload. The video upload interface further allows a user to add a description of the video. Preferably, this description can be edited later by the user, or, in at least some implementations, by another use with appropriate permissions.
  • Upon uploading a video, a user can return (and in at least some implementations is automatically taken back) to the video selection screen, where the user can then select that video (or another video) for annotation and/or review.
  • Upon initially identifying a video using the video selection interface, properties of that video are provided, as illustrated in FIG. 5. A user can then select that video for annotation and/or review.
  • Upon selecting a video, a user is taken to a video annotation interface, as illustrated in FIG. 6, which shows the selected video and comments posted to that video by users (preferably including both comments posted by other users and comments posted by the logged in user, which may be color coded as described herein, such as, for example, by outlining each comment and/or the commenting user's name in a color associated with the commenting user). In the illustrated implementation, the video annotation interface comprises a video pane 12 with controls therebelow, and a display pane 14 with tabs for alternately displaying a comment display pane and a change log pane.
  • The controls below the video pane 12 include a video timeline bar 20, a segment timeline bar 30, a play button 16, a playback speed slider 40, and a comment button 18.
  • The video annotation interface allows a user to play the selected video with the play button. The video timeline bar includes a video play-head 22 which indicates relatively what frame or portion of the video is currently displayed/being played, as illustrated in FIGS. 7A-7C. Preferably, the annotation interface displays (such as, for example, in seconds) where in playback the video play-head 22 currently is.
  • The segment timeline bar can be utilized by a user to limit playback to a particular segment of the video. The segment timeline bar includes an initial position handle 32 and a final position handle 34. In at least some implementations, each position handle 32,34 preferably displays in association therewith a time of the video the handle is currently positioned at, as illustrated in FIG. 6 (such text is omitted for clarity in many illustrations). Each of these position handles can be slid along the segment timeline bar by a user (e.g. using a mouse or a touchscreen), as illustrated by the arrows in FIGS. 8A-B, and in some implementations can even be slid to the same point, but in preferred implementations cannot be slid past one another (although in at least some implementations they may be). The handles can be utilized by a user to define a segment to limit playback to (or associate a comment therewith, as described hereinbelow). In one or more preferred implementations, the handles are initially set to the start and end points of the video, and when the handles are so set the segment timeline bar appears generally similar in appearance to the video timeline bar (e.g. being displayed in the same color), but when one or both handles are moved so as to define a segment which is less than the entire video, the portion of the segment timeline bar between the handles is highlighted (e.g. is displayed in a brighter, darker, or different color, such as bright yellow).
  • In a preferred implementation, a segment timeline bar is implemented utilizing HTML and a j Query UI slider library, currently available at jqueryui.com.
  • If a particular segment is selected (as opposed to the entire video), then the video play-head can only be set as early as the time corresponding to the initial position handle (and will preferably be moved there if it was set to an earlier point in the video before the segment was defined, as illustrated by reference to FIGS. 8A-B), and can only be set as late as the time corresponding to the final position handle (and will preferably be moved there if it was set to a later point in the video before the segment was defined). Thereafter, when playback is commenced, the video will only play the portion of the video corresponding to the segment, and once it reaches the time corresponding to the final position handle, it will loop back to the time corresponding to the initial position handle.
  • In this way, the segment timeline bar can be characterized as providing a looping mechanism that works by selecting a defined segment of video and looping the playback repeatedly through that segment. The mechanism involves the use of the timeline handles which can be dragged to indicate the beginning and end of a segment for looping. For example, if a dancer wants to see only a small part of video repeatedly so that he or she can master a given movement before proceeding to the next, he or she can select a segment of video he or she wants to see again and again.
  • In one or more preferred implementations, when no segment is selected and the handles of the segment timeline bar are at the start and end of the video, double clicking anywhere on the segment timeline bar will automatically move the handles to correspond to a segment of a certain length centered around the point clicked on. Such certain length may be a predefined or preconfigured length, such as, for example, four seconds, and/or may be calculated or determined based in whole or in part on a total length of the video. In at least some implementations, such a certain length may be wholly or partially user configurable.
  • In one or more preferred implementations, once a segment is defined, a user can easily perform various operations to manipulate the segment. For example, a user can extend the segment by clicking on the segment timeline bar outside of the segment, in which case the handle closest to the point on the segment timeline bar the user clicked is moved to the point the user clicked, as illustrated by reference to FIG. 8C (illustrating where a user clicks with his or her mouse cursor), and FIG. 8D (which illustrates the change to the segment following such user click). Similarly, a user can shorten the segment by clicking on the segment timeline bar inside of the segment, in which case the handle closest to the point on the segment timeline bar the user clicked is moved to the point the user clicked, as illustrated by reference to FIG. 8E (illustrating where a user clicks with his or her mouse cursor) and FIG. 8F (which illustrates the change to the segment following such user click). Further, a user can drag a defined segment across the segment timeline bar, in effect moving the start and end points of that segment, but keeping the same segment length, as illustrated by reference to FIGS. 8G-8H. Further still, a user can deselect the segment, and select the entire video, by either moving the handles to the start and end points of the video, or can select the entire video (automatically moving the handles to the start and end of the video) by double clicking anywhere within the segment.
  • Using the video annotation interface, comments can be inserted on the video timeline bar in a plurality of different modalities by clicking the comment button. Once a user chooses to insert a comment, he or she can choose between text, sketch, voice, video, or any combination of these modalities. Sketch commenting preferably supports multi-color sketching with at least two different pen widths. Preferably, a user has the option of clearing the canvas if he or she is not satisfied with his or her sketch. Video commenting can utilize a web-cam to capture a video comment of the user, or can utilize an existing file as a video comment.
  • Once a comment is submitted, it is added to the comment display pane and one or more markers (such as, for example, a color-coded circle, rectangle, or triangle) appears over the video timeline bar showing the position of the inserted comment in the video.
  • In preferred implementations, posted comments can be associated with either a single point in time (or frame) of a video, or with a longer segment of a video. The former can be characterized as point comments, and might be represented by a first type of marker (such as, for example, a single colored rectangle) and stored in association with a single point in time, or frame, while the latter can be characterized as segment comments, and might be represented by a different type of marker (such as, for example, a colored triangle denoting a start of the segment, or even a pair of colored triangles denoting a start and an end of the segment), and stored in association with a start and end time, or start and end frames. Preferably, a user can add a point comment by clicking on the comment button while no segment is selected (that is, the entire video is selected for playback), as illustrated (starting from the situation in FIG. 7C) in FIGS. 9A-H, and can add a segment comment by clicking on the comment button while a segment is selected, as illustrated (starting from the situation in FIG. 8H) in FIGS. 10A-B.
  • In the case of text (and, in at least some implementations, video) comments, a user can choose to move around the comment to overlay a different portion of the video, as illustrated via reference to FIGS. 9A-B; the position of the comment is preferably saved such that the comment is presented to other users in that same position. A user who originally posted a comment is preferably able to later edit, move, or delete such a comment, and other users may be able to do so as well if they enjoy sufficient permissions. Additionally, all users are preferably able to hide a displayed comment using an close button 72, which is illustrated in FIG. 9H. FIG. 9I illustrates similar display of a video comment (which video comment may be displayed with its own video timeline bar and segment timeline bar, as illustrated).
  • Preferably, each user's comments are color-coded, which can help in visual identification of comments while navigating.
  • FIG. 9B illustrates an exemplary comment addition interface which is provided once a user presses the comment button. The comment addition interface allows a user to input a text comment, move the position of that text comment relative to the video, select a color to use to annotate the video with using a plurality of displayed color boxes 74 (which can perhaps best be seen in FIG. 9C), annotate the video by drawing directly on the video using the selected color, and clear the canvas of any drawn annotations (which may include multiple drawn annotations in multiple colors) using a clear button 76 (which can perhaps best be seen in FIG. 9C). In at least some preferred implementations, the comment addition interface additionally allows a user to record a video or audio comment (e.g. utilizing a microphone and/or webcam), or upload such a video or audio comment (e.g. by selecting a file for upload).
  • When a user watches a video, he or she can choose to display all comments (such as by utilizing a “Show all Comments” checkbox) during playback, in which case the comments associated with that video are displayed during playback. In at least some implementations, video playback is paused when a point (and/or segment) associated with a comment is reached. In at least some implementations, even though point comments are only associated with a single point in time, they may be displayed for a longer duration, such as, for example, for a period of time prior to and/or subsequent to the point in time they are associated with, which period of time may be preconfigured and/or determined/modified based on an overall length of the video.
  • Posted comments can preferably be explored several other ways as well. A user can click on any comment in the comment display pane, or, alternatively, can click on a marker on the video timeline bar 20.
  • In either of these cases, when a user has selected a point comment, the video play-head 22 is either set to the point in time associated with the comment (which can be characterized as a cue-point), or else to a point in time prior to the cue-point (which might be a predefined amount of time before the cue-point, and/or might be calculated based on the overall length of the video), the video is paused (although in at least some implementations it may not be), and any comments related to that cue-point are overlaid over the video. In one or more preferred implementations, segment functionality of the segment timeline bar 30 is utilized when a point comment is selected to automatically define a segment around the point associated with the comment, the handles 32,34 of the segment timeline bar being automatically moved to define a segment having a length that might be predetermined or preconfigured (such as, for example, four seconds), or might be calculated or determined (such as, for example, based on a total length of the video). In at least some implementations, different types of comments may trigger different treatment; for example, a video may be paused at a particular frame associated with a drawing if a drawn annotation is present, but may be paused and/or looped through a four second segment if a text comment is present.
  • Similarly to some implementations of point comments, if a user has selected a segment comment, the initial and final handles 32,34 are set to the positions associated with that segment comment, the play-head 22 is moved to the position corresponding to the initial handle 32, the video is paused (although in at least some implementations it may not be), and the comment is displayed (other comments falling within that segment may or may not be displayed during playback).
  • Allowing a user to click on a comment and have a portion of the video associated therewith played back (together with the comment) delivers comments in context, which makes it easy for users to correlate comments with the portion of the video. For example, dancers watching a video can easily correlate comments with their movements.
  • In one or more preferred implementations, a user can hover a cursor over a comment marker, and a preview of the comment (preferably with a frame of the video the comment is associated with) is displayed. For example, a preview may be displayed in the video pane in the form of the frame the comment is associated with and the comment in its position relative to the video, and/or a preview may be displayed in the form of the comment displayed above the comment marker.
  • Additionally, in one or more preferred implementations, when a user hovers over a segment comment marker, the portion of the video corresponding to that segment is highlighted, for example on the segment timeline bar or the video timeline bar.
  • As noted hereinabove, the video annotation interface includes a change log pane which is configured to track changes and/or user interaction with the application. The application preferably includes self-documenting archival functionality which logs, for example, the creation, editing, and deletion of comments. Such changes are displayed in the change log pane, and in one or more preferred implementations, the application allows changes to be undone. Additionally, the application preferably logs other types of user interaction with the video annotation interface as well, as illustrated in FIG. 11.
  • In one or more preferred implementations, the video annotation interface includes focus area functionality which allows a user to define a focus area for a comment, such as, for example, a rectangular, circular, or irregularly shaped (and user defined) focus area which corresponds to a portion of the video the user wants other users to focus on when reviewing his or her comment. Such a focus area may be defined by drawing a particular shape on the video pane, and/or by cropping out an area of the video pane. In some preferred implementations, during playback of frames of the video associated with the comment having a defined focus area, the portions of the frames outside of the defined focus area are grayed out, as illustrated in FIG. 12. Similarly, in at least some implementations, the video annotation interface allows a user to define a zoom area, and the frames of the video associated with the zoom area are displayed zoomed in on the zoom area when the comment is selected for playback, as illustrated in FIG. 13 (and, in at least some implementations, during normal play back too if all comments are selected for display, although in at least some implementations zooming may not be enabled in such situations).
  • In one or more preferred implementations, the video annotation interface provides the ability to define a measurement scale and take measurements based on such scale. An exemplary such methodology involves allowing a user to first establish a baseline measurement by drawing a line segment over the video pane that corresponds with a certain displayed element and enter a measurement corresponding to that displayed element and line segment. For example, if a user knows that the beak of a particular bird displayed in a video is 3 cm, then a user can draw a line segment over the beak and input 3 cm (preferably, the application alternatively allows for simple entry of a ratio, for example, if a user knows that 10 cm onscreen corresponds to 3 cm in the video, the user can simply enter that ratio). Thereafter, the user can draw additional line segments over frames of the video and the application will calculate, based on the scale defined by the 3 cm beak, the relative length of whatever is displayed in the video based on the drawn line segment. In a preferred implementation, this is calculated as a simple ratio; for example, a 10 cm line segment that is drawn by a user and indicated to correspond to a 3 cm beak could be used to determine that a 20 cm line segment drawn by the user corresponds to 6 cm in the video using the simple ratio: 10/3=20/x, or x=(20*3)/10.
  • The video annotation interface includes a playback speed slider, which allows a user to control a playback speed of the video. The playback speed slider preferably includes a selection handle which a user can slide along to vary the playback speed. Preferably, the playback speed slider enables slow motion playback, that is, play back of a video at a slower rate than its normal speed. In preferred implementations, a normal playback speed is roughly 30 frames/second, so reducing a playback speed could, for example, effect playback in slow motion at a speed of 20 frames/second, 15 frames/second, 10 frames/second, 5 frames/second, etc. In a preferred implementation, the playback speed slider control can vary the playback speed from 0.1× to 2×, with a midpoint of the slider control corresponding to normal speed (1×), although in at least some other implementations these values may be different (and may even be user configurable). In one or more preferred implementations, a user can associate a point or segment comment with a particular playback speed so that the if a user later selects the comment, the portion of video corresponding to that comment will be played back at the associated speed. In one or more preferred implementations, this might be characterized as saving the state of a comment, such as the state of the playback speed slider.
  • In one or more preferred implementations, the application includes functionality which allows a user to automatically generate a comment with a single key press. For example, an ornithologist watching a video of a particular bird might want to annotate the video with bird behaviors that occur during the video, such as turning left, turning right, chirping, etc. In order to facilitate rapid annotation of these behaviors, the application is preferably configured to allow the user to automatically generate certain configured comments with a single key press, such as automatically generating and saving a “left turn” comment by pressing “L”, or automatically generating and saving a “chirp” comment by pressing “C”. Preferably, these configured automatically generated comments are capable of being user configured. In one or more preferred implementations, the application is configured to utilize key code (or key press) libraries which configure what keys will automatically generate what comment. For example, a key press library might comprise a comma separated values file in which each line includes an indication of a key (such as “C”) and an indication of a comment to automatically generate (such as “chirp”). Such a file might additionally include an (x,y) position on the video to anchor the comment, and/or an indication of a segment to create based on the key press (for example, the file might indicate to create a segment starting x seconds before the point at which the key was pressed and ending x seconds after the point at which the key was pressed, where either or both x's might be zero). In one or more preferred implementations, an application might provide key code (or key press) library functionality by allowing a user to upload and/or load one or more key press files.
  • In one or more preferred implementations, the application includes a full screen mode which allows users to view a video in full screen mode. In at least some implementations, this full screen mode includes comment markers displayed proximate a top or bottom of the screen. Similarly, in at least some implementations, the full screen mode includes one or both timeline bars displayed proximate a top or bottom of the screen.
  • Although illustrated and described herein largely in the context of a computer application configured for use with a mouse and keyboard, in one or more preferred implementations the application is configured for use with a tablet or slate computer, such as an iPad, or a smart phone, such as an iPhone, Android, or Blackberry. In some preferred implementations, the application is configured to toggle between a normal mode and a full screen mode as an orientation of the device the application is running on is transitioned from a landscape orientation to a portrait orientation, or vice versa (e.g. based on accelerometer readings of the device).
  • In one or more preferred implementations, commenting is configured to allow users to respond to comments made by other users, and responsive comments may be grouped in various ways to indicate that they are responsive to a previous comment. For example, comments may form a comment thread, and/or may all be grouped together in a manner that pictorially or numerically indicates that multiple comments are grouped or associated.
  • In addition to allowing users to collaboratively comment on videos, the application preferably includes a journal which can be utilized by users. For example, choreographers and dancers often keep a journal documenting notes on their progress and ideas and images to be explored (in some production processes a journal is an explicit part of the development); the application includes journal functionality which can be used by choreographers or dancers to record and review private thoughts. So if a dancer got some instruction from a choreographer or if her or she made a mental note about something to remember, he or she can write it in his or her journal. A user can access this journal functionality anytime while using the application, once logged in, and preferably journal entries are private to a user and cannot be seen by other users.
  • Specifically, the application's navigation bar allows users to access a journal interface. FIG. 14 illustrates such journal interface. The journal interface allows a user to write and submit new journal entries, and review, edit, and delete older journal entries.
  • Based on the foregoing description, it will be readily understood by those persons skilled in the art that the present invention is susceptible of broad utility and application. Many embodiments and adaptations of the present invention other than those specifically described herein, as well as many variations, modifications, and equivalent arrangements, will be apparent from or reasonably suggested by the present invention and the foregoing descriptions thereof, without departing from the substance or scope of the present invention. Accordingly, while the present invention has been described herein in detail in relation to one or more preferred embodiments, it is to be understood that this disclosure is only illustrative and exemplary of the present invention and is made merely for the purpose of providing a full and enabling disclosure of the invention. The foregoing disclosure is not intended to be construed to limit the present invention or otherwise exclude any such other embodiments, adaptations, variations, modifications or equivalent arrangements, the present invention being limited only by the claims appended hereto and the equivalents thereof.

Claims (20)

What is claimed is:
1: A method comprising:
(a) authenticating a user account of a user;
(b) receiving an identification of a video;
(c) receiving, at an electronic device,
(i) data corresponding to the video, and
(ii) data corresponding to a plurality of point comments associated with the video, each of the plurality of point comments being associated with a single point in time and with a particular user account of a plurality of user accounts,
(iii) data corresponding to a plurality of segment comments associated with the video, each of the plurality of segment comments being associated with a video segment representing a span of time of the video and with a particular user account of the plurality of user accounts;
(d) displaying, to the user via an electronic display associated with the electronic device, a video annotation interface comprising
(i) a video pane configured to display the video,
(ii) a first video timeline bar including a video play-head indicating a current point of the video which is being played,
(iii) a second segment timeline bar disposed below the first video timeline bar, the second segment timeline bar including initial and final handles configured to define a segment of the video for playing,
(iv) a first plurality of point comment markers identifiable as point comment markers by the presence of a first geometric shape displayed in connection with the video timeline bar, each of the first plurality of comment markers corresponding to one of the plurality of point comments associated with the video,
(v) a second plurality of segment comment markers identifiable as segment comment markers by the presence of a second geometric shape displayed in connection with the video timeline bar, each of the second plurality of comment markers corresponding to one of the plurality of segment comments associated with the video,
(vi) a comment display pane displaying text corresponding to at least some of the plurality of comments associated with the video, and
(vii) a comment control configured to allow the user to add a comment to the video; and
(e) receiving, at the electronic device from the user, input corresponding to engagement at a first point on the segment timeline bar;
(f) automatically, in response to receiving the input corresponding to engagement at the first point on the segment timeline bar, moving the initial and final handles of the segment timeline bar to define a first segment of a first length centered around the first point, the first length being a length calculated based on a total length of the video;
(g) receiving, at the electronic device from the user, input corresponding to dragging of the final handle to change the length of the first segment to a second length;
(h) receiving, at the electronic device from the user, input corresponding to dragging of the first segment on the segment timeline bar, and simultaneously moving the initial and final handles while keeping the first segment its current length in response thereto;
(i) receiving, at the electronic device from the user, input corresponding to engagement of the comment control;
(j) in response to receiving input corresponding to engagement of the comment control, displaying, to the user via the electronic display associated with the electronic device, a comment interface;
(k) receiving, at the electronic device from the user, input corresponding to one or more desired annotations; and
(l) in response to receiving input corresponding to one or more desired annotations,
(i) associating the input one or more annotations with the selected first segment of the video,
(ii) updating the video annotation interface so that the plurality of comment markers displayed in connection with the video timeline bar includes a new segment comment marker corresponding to the first segment, and
(iii) displaying an indication of the input one or more annotations overlaid over the video in the video pane.
2: The method of claim 1, wherein the received data corresponding to a plurality of point comments associated with the video includes data corresponding to a particular point comment which includes an identification of a point of the video associated with the comment.
3: The method of claim 2, wherein the identification of a point of the video comprises a timestamp.
4: The method of claim 1, wherein the identification of a point of the video comprises a frame identification.
5: The method of claim 1, wherein the received data corresponding to a plurality of segment comments associated with the video includes data corresponding to a particular segment comment, and wherein the data corresponding to the particular segment comment includes an identification of a starting point of a segment of the video associated with the particular segment comment.
6: The method of claim 5, wherein the data corresponding to the particular segment comment includes an identification of a length of the segment comment.
7: The method of claim 5, wherein the data corresponding to the particular segment comment includes an identification of an end point of a segment of the video associated with the particular segment comment.
8: The method of claim 1, wherein the electronic display is part of the electronic device.
9: The method of claim 1, wherein the electronic device is a tablet, slate computer, or smartphone.
10: The method of claim 1, wherein the electronic display is connected to the electronic device.
11: The method of claim 1, wherein the electronic display is a touchscreen display, and wherein at least one of the steps of receiving, at the electronic device from the user, input, comprises receiving input from the user input via the touchscreen display.
12: The method of claim 1, wherein the method further includes
(a) receiving, at the electronic device from the user, input corresponding to selection of a baseline definition tool;
(b) receiving, at the electronic device from the user, input corresponding to drawing a baseline line segment on the video pane;
(c) receiving, at the electronic device from the user, input representing a baseline measurement for the user-drawn baseline line segment;
(d) receiving, at the electronic device from the user, input corresponding to drawing a measurement line segment on the video pane;
(e) automatically determining a measurement for the measurement line segment based on
(i) a length of the user-drawn baseline line segment,
(ii) the user input representing a baseline measurement for the user-drawn baseline line segment, and
(iii) a length of the user-drawn measurement line segment;
(f) displaying, to the user via the electronic display, the automatically determined measurement for the measurement line segment.
13: The method of claim 1, wherein each of the plurality of comment markers is color coded based on the particular user account associated with a comment the comment marker is associated with.
14: The method of claim 1, wherein the one or more annotations includes a video annotation.
15: The method of claim 1, wherein the one or more annotations includes an audio annotation.
16: The method of claim 1, wherein the one or more annotations includes a drawing annotation.
17: The method of claim 1, wherein the one or more annotations includes annotations in a plurality of different modalities.
18: The method of claim 1, wherein the method further includes loading, at the electronic device, a key code library representing an association of keys with one or more respective desired annotations, and wherein receiving, at the electronic device from the user, input corresponding to one or more desired annotations comprises receiving input corresponding to a particular key which is associated with one or more respective desired annotations in the key code library.
19: A method comprising:
(a) authenticating a user account of a user;
(b) receiving an identification of a video;
(c) receiving, at an electronic device,
(i) data corresponding to the video, and
(ii) data corresponding to a plurality of point comments associated with the video, each of the plurality of point comments being associated with a single point in time and with a particular user account of a plurality of user accounts,
(iii) data corresponding to a plurality of segment comments associated with the video, each of the plurality of segment comments being associated with a video segment representing a span of time of the video and with a particular user account of the plurality of user accounts;
(d) displaying, to the user via an electronic display associated with the electronic device, a video annotation interface comprising
(i) a video pane configured to display the video,
(ii) a first video timeline bar including a video play-head indicating a current point of the video which is being played,
(iii) a second segment timeline bar disposed below the first video timeline bar, the second segment timeline bar including initial and final handles configured to define a segment of the video for playing,
(iv) a first plurality of point comment markers identifiable as point comment markers by the presence of a first geometric shape displayed in connection with the video timeline bar, each of the first plurality of comment markers corresponding to one of the plurality of point comments associated with the video,
(v) a second plurality of segment comment markers identifiable as segment comment markers by the presence of a second geometric shape displayed in connection with the video timeline bar, each of the second plurality of comment markers corresponding to one of the plurality of segment comments associated with the video,
(vi) a comment display pane displaying text corresponding to at least some of the plurality of comments associated with the video, and
(vii) a comment control configured to allow the user to add a comment to the video; and
(e) receiving, at the electronic device from the user, input corresponding to engagement at a first point on the segment timeline bar;
(f) automatically, in response to receiving the input corresponding to engagement at the first point on the segment timeline bar, moving the initial and final handles of the segment timeline bar to define a first segment of a first length centered around the first point, the first length being a preconfigured length;
(g) receiving, at the electronic device from the user, input corresponding to dragging of the final handle to change the length of the first segment to a second length;
(h) receiving, at the electronic device from the user, input corresponding to dragging of the first segment on the segment timeline bar, and simultaneously moving the initial and final handles while keeping the first segment its current length in response thereto;
(i) receiving, at the electronic device from the user, input corresponding to engagement of the comment control;
(j) in response to receiving input corresponding to engagement of the comment control, displaying, to the user via the electronic display associated with the electronic device, a comment interface;
(k) receiving, at the electronic device from the user, input corresponding to one or more desired annotations; and
(l) in response to receiving input corresponding to one or more desired annotations,
(i) associating the input one or more annotations with the selected first segment of the video,
(ii) updating the video annotation interface so that the plurality of comment markers displayed in connection with the video timeline bar includes a new segment comment marker corresponding to the first segment, and
(iii) displaying an indication of the input one or more annotations overlaid over the video in the video pane.
20: A method comprising:
(a) receiving, at an electronic device,
(i) data corresponding to a video, and
(ii) data corresponding to a plurality of point comments associated with the video, each of the plurality of point comments being associated with a single point in time and with a particular user account of a plurality of user accounts,
(iii) data corresponding to a plurality of segment comments associated with the video, each of the plurality of segment comments being associated with a video segment representing a span of time of the video and with a particular user account of the plurality of user accounts;
(b) displaying, to a user via an electronic display associated with the electronic device, a video annotation interface comprising
(i) a video pane configured to display the video,
(ii) a first video timeline bar including a video play-head indicating a current point of the video which is being played,
(iii) a second segment timeline bar disposed below the first video timeline bar, the second segment timeline bar including initial and final handles configured to define a segment of the video for playing,
(iv) a first plurality of point comment markers identifiable as point comment markers by the presence of a first geometric shape displayed in connection with the video timeline bar, each of the first plurality of comment markers corresponding to one of the plurality of point comments associated with the video,
(v) a second plurality of segment comment markers identifiable as segment comment markers by the presence of a second geometric shape displayed in connection with the video timeline bar, each of the second plurality of comment markers corresponding to one of the plurality of segment comments associated with the video,
(vi) a comment display pane displaying text corresponding to at least some of the plurality of comments associated with the video, and
(vii) a comment control configured to allow the user to add a comment to the video; and
(c) receiving, at the electronic device from the user, input corresponding to engagement at a first point on the segment timeline bar;
(d) automatically, in response to receiving the input corresponding to engagement at the first point on the segment timeline bar, moving the initial and final handles of the segment timeline bar to define a first segment of a first length centered around the first point;
(e) receiving, at the electronic device from the user, input corresponding to dragging of the final handle to change the length of the first segment to a second length;
(f) receiving, at the electronic device from the user, input corresponding to dragging of the first segment on the segment timeline bar, and simultaneously moving the initial and final handles while keeping the first segment its current length in response thereto;
(g) receiving, at the electronic device from the user, input corresponding to engagement of the comment control;
(h) in response to receiving input corresponding to engagement of the comment control, displaying, to the user via the electronic display associated with the electronic device, a comment interface;
(i) receiving, at the electronic device from the user, input corresponding to one or more desired annotations; and
(j) in response to receiving input corresponding to one or more desired annotations,
(i) associating the input one or more annotations with the selected first segment of the video,
(ii) updating the video annotation interface so that the plurality of comment markers displayed in connection with the video timeline bar includes a new segment comment marker corresponding to the first segment, and
(iii) displaying an indication of the input one or more annotations overlaid over the video in the video pane.
US16/103,307 2011-09-26 2018-08-14 Multi-modal collaborative web-based video annotation system Abandoned US20180358049A1 (en)

Priority Applications (2)

Application Number Priority Date Filing Date Title
US16/103,307 US20180358049A1 (en) 2011-09-26 2018-08-14 Multi-modal collaborative web-based video annotation system
US16/850,176 US20200286519A1 (en) 2011-09-26 2020-04-16 Multi-modal collaborative web-based video annotation system

Applications Claiming Priority (4)

Application Number Priority Date Filing Date Title
US201161539147P 2011-09-26 2011-09-26
US13/627,899 US9354763B2 (en) 2011-09-26 2012-09-26 Multi-modal collaborative web-based video annotation system
US15/144,976 US10079039B2 (en) 2011-09-26 2016-05-03 Multi-modal collaborative web-based video annotation system
US16/103,307 US20180358049A1 (en) 2011-09-26 2018-08-14 Multi-modal collaborative web-based video annotation system

Related Parent Applications (2)

Application Number Title Priority Date Filing Date
US15/144,976 Continuation-In-Part US10079039B2 (en) 2011-09-26 2016-05-03 Multi-modal collaborative web-based video annotation system
US15/144,976 Continuation US10079039B2 (en) 2011-09-26 2016-05-03 Multi-modal collaborative web-based video annotation system

Related Child Applications (1)

Application Number Title Priority Date Filing Date
US16/850,176 Continuation US20200286519A1 (en) 2011-09-26 2020-04-16 Multi-modal collaborative web-based video annotation system

Publications (1)

Publication Number Publication Date
US20180358049A1 true US20180358049A1 (en) 2018-12-13

Family

ID=56693141

Family Applications (3)

Application Number Title Priority Date Filing Date
US15/144,976 Active 2033-10-27 US10079039B2 (en) 2011-09-26 2016-05-03 Multi-modal collaborative web-based video annotation system
US16/103,307 Abandoned US20180358049A1 (en) 2011-09-26 2018-08-14 Multi-modal collaborative web-based video annotation system
US16/850,176 Abandoned US20200286519A1 (en) 2011-09-26 2020-04-16 Multi-modal collaborative web-based video annotation system

Family Applications Before (1)

Application Number Title Priority Date Filing Date
US15/144,976 Active 2033-10-27 US10079039B2 (en) 2011-09-26 2016-05-03 Multi-modal collaborative web-based video annotation system

Family Applications After (1)

Application Number Title Priority Date Filing Date
US16/850,176 Abandoned US20200286519A1 (en) 2011-09-26 2020-04-16 Multi-modal collaborative web-based video annotation system

Country Status (1)

Country Link
US (3) US10079039B2 (en)

Cited By (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN112839258A (en) * 2021-04-22 2021-05-25 北京世纪好未来教育科技有限公司 Video note generation method, video note playing method, video note generation device, video note playing device and related equipment
CN113811948A (en) * 2019-04-01 2021-12-17 黑魔法设计私人有限公司 User interface for video editing system
WO2022039301A1 (en) * 2020-08-20 2022-02-24 주식회사 누날 Video curation service method
US11721365B2 (en) 2020-11-09 2023-08-08 Blackmagic Design Pty Ltd Video editing or media management system
US11942117B2 (en) 2019-04-01 2024-03-26 Blackmagic Design Pty Ltd Media management system

Families Citing this family (40)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US9098611B2 (en) * 2012-11-26 2015-08-04 Intouch Technologies, Inc. Enhanced video interaction for a user interface of a telepresence network
US9557885B2 (en) 2011-08-09 2017-01-31 Gopro, Inc. Digital media editing
US10079039B2 (en) * 2011-09-26 2018-09-18 The University Of North Carolina At Charlotte Multi-modal collaborative web-based video annotation system
US9245020B2 (en) * 2011-12-14 2016-01-26 Microsoft Technology Licensing, Llc Collaborative media sharing
US9135602B2 (en) 2012-07-25 2015-09-15 E-Plan, Inc. Management of building plan documents utilizing comments and a correction list
US9699488B2 (en) * 2014-06-02 2017-07-04 Google Inc. Smart snap to interesting points in media content
CN105468347B (en) * 2014-09-05 2018-07-27 富泰华工业(深圳)有限公司 Suspend the system and method for video playing
US10606941B2 (en) * 2015-08-10 2020-03-31 Open Text Holdings, Inc. Annotating documents on a mobile device
US10897490B2 (en) 2015-08-17 2021-01-19 E-Plan, Inc. Systems and methods for augmenting electronic content
US10015212B2 (en) * 2015-11-12 2018-07-03 Genband Us Llc Asynchronous collaboration session linked to a synchronous collaboration session
US10109319B2 (en) 2016-01-08 2018-10-23 Gopro, Inc. Digital media editing
US9812175B2 (en) 2016-02-04 2017-11-07 Gopro, Inc. Systems and methods for annotating a video
CN106792250A (en) * 2017-01-17 2017-05-31 百度在线网络技术(北京)有限公司 Barrage information interacting method and device
EP3364311A1 (en) 2017-02-15 2018-08-22 Televic Education NV Visualisation system
CN107404668A (en) * 2017-05-08 2017-11-28 武汉鲨鱼网络直播技术有限公司 A kind of colorful barrage rendering intent and device
US11226731B1 (en) * 2018-01-24 2022-01-18 Snap Inc. Simulated interactive panoramas
US10390097B1 (en) 2018-05-30 2019-08-20 Rovi Guides, Inc. Systems and methods for creating an asynchronous social watching experience among users
US11012403B1 (en) * 2018-09-04 2021-05-18 Facebook, Inc. Storylines: collaborative feedback system
US10929909B2 (en) * 2018-09-19 2021-02-23 International Business Machines Corporation Media enhancement with customized add-on content
US10617949B1 (en) 2018-10-08 2020-04-14 Facebook, Inc. Digital feedback prompt
USD904425S1 (en) 2018-10-08 2020-12-08 Facebook, Inc. Display screen with a graphical user interface
USD904426S1 (en) 2018-10-08 2020-12-08 Facebook, Inc. Display screen with a graphical user interface
US10623917B1 (en) * 2018-10-08 2020-04-14 Facebook, Inc. Collaborative digital story system
US10924446B1 (en) 2018-10-08 2021-02-16 Facebook, Inc. Digital story reply container
CN109725803B (en) * 2018-12-20 2020-02-04 网易(杭州)网络有限公司 Comment information processing method and device, storage medium and electronic equipment
CN115185445A (en) * 2019-04-17 2022-10-14 华为技术有限公司 Method for adding annotations and electronic equipment
US11678031B2 (en) * 2019-04-19 2023-06-13 Microsoft Technology Licensing, Llc Authoring comments including typed hyperlinks that reference video content
US11785194B2 (en) 2019-04-19 2023-10-10 Microsoft Technology Licensing, Llc Contextually-aware control of a user interface displaying a video and related user text
US11960562B1 (en) 2019-05-03 2024-04-16 Meta Platforms, Inc. Channels of content for display in an online system
US11100182B1 (en) 2019-05-03 2021-08-24 Facebook, Inc. Channels of content for display in a online system
US10990263B1 (en) * 2019-09-03 2021-04-27 Gopro, Inc. Interface for trimming videos
EP4147452A4 (en) * 2020-05-06 2023-12-20 ARRIS Enterprises LLC Interactive commenting in an on-demand video
KR20210114536A (en) * 2020-06-24 2021-09-23 바이두 온라인 네트웍 테크놀러지 (베이징) 캄파니 리미티드 Video data production method and apparatus, electronic device, computer readable medium
US11528533B2 (en) * 2020-10-06 2022-12-13 Rovi Guides, Inc. Dynamic media content runtime modification
US11550452B1 (en) * 2021-06-09 2023-01-10 Gopro, Inc. Presentation of a media sequence
US11424945B1 (en) 2021-06-30 2022-08-23 Dropbox, Inc. Techniques for avoiding conflicting user actions during a video collaboration session
US11863600B2 (en) 2021-06-30 2024-01-02 Dropbox, Inc. Techniques for efficient communication during a video collaboration session
US20230088315A1 (en) * 2021-09-22 2023-03-23 Motorola Solutions, Inc. System and method to support human-machine interactions for public safety annotations
GB2621403A (en) * 2022-08-12 2024-02-14 Sony Group Corp Data processing apparatuses and methods
CN116744048B (en) * 2023-08-14 2023-12-15 杭州面朝互动科技有限公司 Online video marking method, system and storage medium

Citations (52)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US5717879A (en) * 1995-11-03 1998-02-10 Xerox Corporation System for the capture and replay of temporal data representing collaborative activities
US5717869A (en) * 1995-11-03 1998-02-10 Xerox Corporation Computer controlled display system using a timeline to control playback of temporal data representing collaborative activities
US5786814A (en) * 1995-11-03 1998-07-28 Xerox Corporation Computer controlled display system activities using correlated graphical and timeline interfaces for controlling replay of temporal data representing collaborative activities
US6452615B1 (en) * 1999-03-24 2002-09-17 Fuji Xerox Co., Ltd. System and apparatus for notetaking with digital video and ink
US20030231198A1 (en) * 2002-06-18 2003-12-18 Koninklijke Philips Electronics N.V. System and method for providing videomarks for a video program
US20040047589A1 (en) * 1999-05-19 2004-03-11 Kim Kwang Su Method for creating caption-based search information of moving picture data, searching and repeating playback of moving picture data based on said search information, and reproduction apparatus using said method
US20060098941A1 (en) * 2003-04-04 2006-05-11 Sony Corporation 7-35 Kitashinagawa Video editor and editing method, recording medium, and program
US20060143560A1 (en) * 1998-09-15 2006-06-29 Microsoft Corporation Annotations for multiple versions of media content
US20060282776A1 (en) * 2005-06-10 2006-12-14 Farmer Larry C Multimedia and performance analysis tool
US20080095404A1 (en) * 2006-10-18 2008-04-24 Ut-Battelle Llc Method and system for determining a volume of an object from two-dimensional images
US20080313570A1 (en) * 2007-06-14 2008-12-18 Yahoo! Inc. Method and system for media landmark identification
US20080313541A1 (en) * 2007-06-14 2008-12-18 Yahoo! Inc. Method and system for personalized segmentation and indexing of media
US20090049400A1 (en) * 2007-08-15 2009-02-19 Sony Corporation Graphical user interface, display control device, display method, and program
US20090132952A1 (en) * 2007-11-16 2009-05-21 Microsoft Corporation Localized thumbnail preview of related content during spatial browsing
US20090187825A1 (en) * 2008-01-23 2009-07-23 Microsoft Corporation Annotating and Sharing Content
US20090210779A1 (en) * 2008-02-19 2009-08-20 Mihai Badoiu Annotating Video Intervals
US20090297118A1 (en) * 2008-06-03 2009-12-03 Google Inc. Web-based system for generation of interactive games based on digital videos
US20090327856A1 (en) * 2008-06-28 2009-12-31 Mouilleseaux Jean-Pierre M Annotation of movies
US20100070878A1 (en) * 2008-09-12 2010-03-18 At&T Intellectual Property I, L.P. Providing sketch annotations with multimedia programs
US20100180218A1 (en) * 2009-01-15 2010-07-15 International Business Machines Corporation Editing metadata in a social network
US7765491B1 (en) * 2005-11-16 2010-07-27 Apple Inc. User interface widget for selecting a point or range
US20100199182A1 (en) * 2006-03-28 2010-08-05 Cisco Media Solutions, Inc., a California corporation System allowing users to embed comments at specific points in time into media presentation
US20100223128A1 (en) * 2009-03-02 2010-09-02 John Nicholas Dukellis Software-based Method for Assisted Video Creation
US20100241962A1 (en) * 2009-03-23 2010-09-23 Peterson Troy A Multiple content delivery environment
US20100251120A1 (en) * 2009-03-26 2010-09-30 Google Inc. Time-Marked Hyperlinking to Video Content
US20110052144A1 (en) * 2009-09-01 2011-03-03 2Cimple, Inc. System and Method for Integrating Interactive Call-To-Action, Contextual Applications with Videos
US20110112665A1 (en) * 2009-11-10 2011-05-12 At&T Intellectual Property I, L.P. Method and apparatus for presenting media programs
US20110138326A1 (en) * 2009-12-04 2011-06-09 At&T Intellectual Property I, L.P. Apparatus and Method for Tagging Media Content and Managing Marketing
US20110158605A1 (en) * 2009-12-18 2011-06-30 Bliss John Stuart Method and system for associating an object to a moment in time in a digital video
US20110161820A1 (en) * 2009-12-31 2011-06-30 John Lee Management of multimedia segment data over a communications network
US20110176788A1 (en) * 2009-12-18 2011-07-21 Bliss John Stuart Method and System for Associating an Object to a Moment in Time in a Digital Video
US20110289401A1 (en) * 2010-05-20 2011-11-24 Salesforce.Com, Inc. Multiple graphical annotations of documents using overlays
US20120023407A1 (en) * 2010-06-15 2012-01-26 Robert Taylor Method, system and user interface for creating and displaying of presentations
US20120030182A1 (en) * 2010-07-27 2012-02-02 Timothy Claman Hierarchical multimedia program composition
US20120047119A1 (en) * 2009-07-21 2012-02-23 Porto Technology, Llc System and method for creating and navigating annotated hyperlinks between video segments
US20120110455A1 (en) * 2010-11-01 2012-05-03 Microsoft Corporation Video viewing and tagging system
US20120151346A1 (en) * 2010-12-10 2012-06-14 Mcclements Iv James Burns Parallel echo version of media content for comment creation and delivery
US20120236201A1 (en) * 2011-01-27 2012-09-20 In The Telling, Inc. Digital asset management, authoring, and presentation techniques
US20120282951A1 (en) * 2011-01-10 2012-11-08 Samsung Electronics Co., Ltd. Anchoring and sharing locations and enjoyment experience information on a presentation timeline for multimedia content streamed over a network
US20120308196A1 (en) * 2009-11-25 2012-12-06 Thomas Bowman System and method for uploading and downloading a video file and synchronizing videos with an audio file
US20130004138A1 (en) * 2011-06-30 2013-01-03 Hulu Llc Commenting Correlated To Temporal Point Of Video Data
US20130042282A1 (en) * 2011-08-12 2013-02-14 Sony Corporation Information processing apparatus, information processing method, program, and information processing system
US20130091469A1 (en) * 2009-11-20 2013-04-11 At&T Intellectual Property I, Lp Apparatus and Method for Collaborative Network in an Enterprise Setting
US20130104054A1 (en) * 2011-10-19 2013-04-25 Microsoft Corporation Techniques to Facilitate Asynchronous Communication
US20130132839A1 (en) * 2010-11-30 2013-05-23 Michael Berry Dynamic Positioning of Timeline Markers for Efficient Display
US20130173531A1 (en) * 2010-05-24 2013-07-04 Intersect Ptp, Inc. Systems and methods for collaborative storytelling in a virtual space
US8724960B2 (en) * 2009-07-09 2014-05-13 Sony Corporation Imaging device, image processing method, and program thereof
US8775922B2 (en) * 2006-12-22 2014-07-08 Google Inc. Annotation framework for video
US8826117B1 (en) * 2009-03-25 2014-09-02 Google Inc. Web-based system for video editing
US20140249931A1 (en) * 2009-11-06 2014-09-04 At&T Intellectual Property I, Lp Apparatus and Method for Managing Marketing
US8839306B2 (en) * 2009-11-20 2014-09-16 At&T Intellectual Property I, Lp Method and apparatus for presenting media programs
US10079039B2 (en) * 2011-09-26 2018-09-18 The University Of North Carolina At Charlotte Multi-modal collaborative web-based video annotation system

Family Cites Families (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US8392821B2 (en) 2006-03-17 2013-03-05 Viddler, Inc. Methods and systems for displaying videos with overlays and tags
US9354763B2 (en) 2011-09-26 2016-05-31 The University Of North Carolina At Charlotte Multi-modal collaborative web-based video annotation system

Patent Citations (53)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US5717869A (en) * 1995-11-03 1998-02-10 Xerox Corporation Computer controlled display system using a timeline to control playback of temporal data representing collaborative activities
US5786814A (en) * 1995-11-03 1998-07-28 Xerox Corporation Computer controlled display system activities using correlated graphical and timeline interfaces for controlling replay of temporal data representing collaborative activities
US5717879A (en) * 1995-11-03 1998-02-10 Xerox Corporation System for the capture and replay of temporal data representing collaborative activities
US20060143560A1 (en) * 1998-09-15 2006-06-29 Microsoft Corporation Annotations for multiple versions of media content
US6452615B1 (en) * 1999-03-24 2002-09-17 Fuji Xerox Co., Ltd. System and apparatus for notetaking with digital video and ink
US20040047589A1 (en) * 1999-05-19 2004-03-11 Kim Kwang Su Method for creating caption-based search information of moving picture data, searching and repeating playback of moving picture data based on said search information, and reproduction apparatus using said method
US20030231198A1 (en) * 2002-06-18 2003-12-18 Koninklijke Philips Electronics N.V. System and method for providing videomarks for a video program
US20060098941A1 (en) * 2003-04-04 2006-05-11 Sony Corporation 7-35 Kitashinagawa Video editor and editing method, recording medium, and program
US20060282776A1 (en) * 2005-06-10 2006-12-14 Farmer Larry C Multimedia and performance analysis tool
US7765491B1 (en) * 2005-11-16 2010-07-27 Apple Inc. User interface widget for selecting a point or range
US20100199182A1 (en) * 2006-03-28 2010-08-05 Cisco Media Solutions, Inc., a California corporation System allowing users to embed comments at specific points in time into media presentation
US20080095404A1 (en) * 2006-10-18 2008-04-24 Ut-Battelle Llc Method and system for determining a volume of an object from two-dimensional images
US8775922B2 (en) * 2006-12-22 2014-07-08 Google Inc. Annotation framework for video
US20080313570A1 (en) * 2007-06-14 2008-12-18 Yahoo! Inc. Method and system for media landmark identification
US20080313541A1 (en) * 2007-06-14 2008-12-18 Yahoo! Inc. Method and system for personalized segmentation and indexing of media
US20090049400A1 (en) * 2007-08-15 2009-02-19 Sony Corporation Graphical user interface, display control device, display method, and program
US20090132952A1 (en) * 2007-11-16 2009-05-21 Microsoft Corporation Localized thumbnail preview of related content during spatial browsing
US20090187825A1 (en) * 2008-01-23 2009-07-23 Microsoft Corporation Annotating and Sharing Content
US20090210779A1 (en) * 2008-02-19 2009-08-20 Mihai Badoiu Annotating Video Intervals
US20090297118A1 (en) * 2008-06-03 2009-12-03 Google Inc. Web-based system for generation of interactive games based on digital videos
US20090327856A1 (en) * 2008-06-28 2009-12-31 Mouilleseaux Jean-Pierre M Annotation of movies
US20100070878A1 (en) * 2008-09-12 2010-03-18 At&T Intellectual Property I, L.P. Providing sketch annotations with multimedia programs
US20100180218A1 (en) * 2009-01-15 2010-07-15 International Business Machines Corporation Editing metadata in a social network
US20100223128A1 (en) * 2009-03-02 2010-09-02 John Nicholas Dukellis Software-based Method for Assisted Video Creation
US20100241962A1 (en) * 2009-03-23 2010-09-23 Peterson Troy A Multiple content delivery environment
US8826117B1 (en) * 2009-03-25 2014-09-02 Google Inc. Web-based system for video editing
US20100251120A1 (en) * 2009-03-26 2010-09-30 Google Inc. Time-Marked Hyperlinking to Video Content
US8724960B2 (en) * 2009-07-09 2014-05-13 Sony Corporation Imaging device, image processing method, and program thereof
US20120047119A1 (en) * 2009-07-21 2012-02-23 Porto Technology, Llc System and method for creating and navigating annotated hyperlinks between video segments
US20110052144A1 (en) * 2009-09-01 2011-03-03 2Cimple, Inc. System and Method for Integrating Interactive Call-To-Action, Contextual Applications with Videos
US20140249931A1 (en) * 2009-11-06 2014-09-04 At&T Intellectual Property I, Lp Apparatus and Method for Managing Marketing
US20110112665A1 (en) * 2009-11-10 2011-05-12 At&T Intellectual Property I, L.P. Method and apparatus for presenting media programs
US8839306B2 (en) * 2009-11-20 2014-09-16 At&T Intellectual Property I, Lp Method and apparatus for presenting media programs
US20130091469A1 (en) * 2009-11-20 2013-04-11 At&T Intellectual Property I, Lp Apparatus and Method for Collaborative Network in an Enterprise Setting
US20120308196A1 (en) * 2009-11-25 2012-12-06 Thomas Bowman System and method for uploading and downloading a video file and synchronizing videos with an audio file
US20110138326A1 (en) * 2009-12-04 2011-06-09 At&T Intellectual Property I, L.P. Apparatus and Method for Tagging Media Content and Managing Marketing
US20110158605A1 (en) * 2009-12-18 2011-06-30 Bliss John Stuart Method and system for associating an object to a moment in time in a digital video
US20110176788A1 (en) * 2009-12-18 2011-07-21 Bliss John Stuart Method and System for Associating an Object to a Moment in Time in a Digital Video
US20110161820A1 (en) * 2009-12-31 2011-06-30 John Lee Management of multimedia segment data over a communications network
US20110289401A1 (en) * 2010-05-20 2011-11-24 Salesforce.Com, Inc. Multiple graphical annotations of documents using overlays
US20130173531A1 (en) * 2010-05-24 2013-07-04 Intersect Ptp, Inc. Systems and methods for collaborative storytelling in a virtual space
US20120023407A1 (en) * 2010-06-15 2012-01-26 Robert Taylor Method, system and user interface for creating and displaying of presentations
US20120030182A1 (en) * 2010-07-27 2012-02-02 Timothy Claman Hierarchical multimedia program composition
US20120110455A1 (en) * 2010-11-01 2012-05-03 Microsoft Corporation Video viewing and tagging system
US20130132839A1 (en) * 2010-11-30 2013-05-23 Michael Berry Dynamic Positioning of Timeline Markers for Efficient Display
US20120151346A1 (en) * 2010-12-10 2012-06-14 Mcclements Iv James Burns Parallel echo version of media content for comment creation and delivery
US9384512B2 (en) * 2010-12-10 2016-07-05 Quib, Inc. Media content clip identification and combination architecture
US20120282951A1 (en) * 2011-01-10 2012-11-08 Samsung Electronics Co., Ltd. Anchoring and sharing locations and enjoyment experience information on a presentation timeline for multimedia content streamed over a network
US20120236201A1 (en) * 2011-01-27 2012-09-20 In The Telling, Inc. Digital asset management, authoring, and presentation techniques
US20130004138A1 (en) * 2011-06-30 2013-01-03 Hulu Llc Commenting Correlated To Temporal Point Of Video Data
US20130042282A1 (en) * 2011-08-12 2013-02-14 Sony Corporation Information processing apparatus, information processing method, program, and information processing system
US10079039B2 (en) * 2011-09-26 2018-09-18 The University Of North Carolina At Charlotte Multi-modal collaborative web-based video annotation system
US20130104054A1 (en) * 2011-10-19 2013-04-25 Microsoft Corporation Techniques to Facilitate Asynchronous Communication

Cited By (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN113811948A (en) * 2019-04-01 2021-12-17 黑魔法设计私人有限公司 User interface for video editing system
US20220189511A1 (en) * 2019-04-01 2022-06-16 Blackmagic Design Pty Ltd User interface for video editing system
US11942117B2 (en) 2019-04-01 2024-03-26 Blackmagic Design Pty Ltd Media management system
WO2022039301A1 (en) * 2020-08-20 2022-02-24 주식회사 누날 Video curation service method
US11721365B2 (en) 2020-11-09 2023-08-08 Blackmagic Design Pty Ltd Video editing or media management system
CN112839258A (en) * 2021-04-22 2021-05-25 北京世纪好未来教育科技有限公司 Video note generation method, video note playing method, video note generation device, video note playing device and related equipment

Also Published As

Publication number Publication date
US10079039B2 (en) 2018-09-18
US20200286519A1 (en) 2020-09-10
US20160247535A1 (en) 2016-08-25

Similar Documents

Publication Publication Date Title
US20200286519A1 (en) Multi-modal collaborative web-based video annotation system
US9354763B2 (en) Multi-modal collaborative web-based video annotation system
US11669225B2 (en) Categorized and tagged video annotation
US11417367B2 (en) Systems and methods for reviewing video content
US10222946B2 (en) Video lesson builder system and method
US11249715B2 (en) Collaborative remote interactive platform
US20180011627A1 (en) Meeting collaboration systems, devices, and methods
US8943408B2 (en) Text image review process
US20150121189A1 (en) Systems and Methods for Creating and Displaying Multi-Slide Presentations
US20150121232A1 (en) Systems and Methods for Creating and Displaying Multi-Slide Presentations
US11190557B1 (en) Collaborative remote interactive platform
WO2012103267A2 (en) Digital asset management, authoring, and presentation techniques
US20160057500A1 (en) Method and system for producing a personalized project repository for content creators
US20160063087A1 (en) Method and system for providing location scouting information
US11349889B1 (en) Collaborative remote interactive platform
US9942297B2 (en) System and methods for facilitating the development and management of creative assets
Clark et al. Captivate and Camtasia
Almahr Re-imaging the UI for metadata entry & media production
Antjon SynergE-Learning: A PaperPoint-Based Lecture Capturing and Annotation Tool
Pongnumkul Facilitating Interactive Video Browsing through Content-Aware Task-Centric Interfaces

Legal Events

Date Code Title Description
STPP Information on status: patent application and granting procedure in general

Free format text: DOCKETED NEW CASE - READY FOR EXAMINATION

STPP Information on status: patent application and granting procedure in general

Free format text: NON FINAL ACTION MAILED

STCB Information on status: application discontinuation

Free format text: ABANDONED -- FAILURE TO RESPOND TO AN OFFICE ACTION