US20120189284A1 - Automatic highlight reel producer - Google Patents

Automatic highlight reel producer Download PDF

Info

Publication number
US20120189284A1
US20120189284A1 US13/041,370 US201113041370A US2012189284A1 US 20120189284 A1 US20120189284 A1 US 20120189284A1 US 201113041370 A US201113041370 A US 201113041370A US 2012189284 A1 US2012189284 A1 US 2012189284A1
Authority
US
United States
Prior art keywords
clips
video
user
digital video
clip
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Abandoned
Application number
US13/041,370
Inventor
Andrew Morrison
Chris Park
Kevin Lau
Parthu Kishen
Desmond Smith
Michael Bibik
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
T Mobile USA Inc
Original Assignee
Individual
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Individual filed Critical Individual
Priority to US13/041,370 priority Critical patent/US20120189284A1/en
Assigned to T-MOBILE USA, INC. reassignment T-MOBILE USA, INC. ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: BIBIK, MICHAEL, KISHEN, PARTHU, LAU, KEVIN, MORRISON, ANDREW, PARK, CHRIS, SMITH, DESMOND
Publication of US20120189284A1 publication Critical patent/US20120189284A1/en
Assigned to DEUTSCHE TELEKOM AG reassignment DEUTSCHE TELEKOM AG INTELLECTUAL PROPERTY SECURITY AGREEMENT Assignors: T-MOBILE USA, INC.
Assigned to T-MOBILE USA, INC., METROPCS WIRELESS, INC., T-MOBILE SUBSIDIARY IV CORPORATION, IBSV LLC, PushSpring, Inc., Layer3 TV, Inc., MetroPCS Communications, Inc. reassignment T-MOBILE USA, INC. RELEASE BY SECURED PARTY (SEE DOCUMENT FOR DETAILS). Assignors: DEUTSCHE BANK AG NEW YORK BRANCH
Assigned to T-MOBILE USA, INC., IBSV LLC reassignment T-MOBILE USA, INC. RELEASE BY SECURED PARTY (SEE DOCUMENT FOR DETAILS). Assignors: DEUTSCHE TELEKOM AG
Abandoned legal-status Critical Current

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/50Information retrieval; Database structures therefor; File system structures therefor of still image data
    • G06F16/58Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually
    • G06F16/583Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually using metadata automatically derived from the content
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/50Information retrieval; Database structures therefor; File system structures therefor of still image data
    • G06F16/58Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually

Definitions

  • An individual person may shoot several hours of digital video at a single event (such as a wedding or a party) or even hundreds of hours of video on a trip (such as to the Grand Canyon) or over a season (the summer). Later, sometimes long after the event, trip, or season is over, the individual videographer may attempt to edit the many hours of video spending hours watching them and editing the videos using the available conventional approaches for doing so. With these conventional approaches, a person may use a crude user interface to view, select, and piece together the clips of the digital video over a timeline and perhaps form a watchable summary of the event, trip, or season.
  • a typical mobile personal communication device While the power and memory capacities have increased, a typical mobile personal communication device still has a small display and, consequently, more constrained user-interface capabilities than a desktop computer. Accordingly, a user of a typical mobile personal communication device is forced to abandon his mobile environment and move to a desktop computer in order to view and manage her large media catalogue.
  • Described herein are techniques related to automatic selection of a subset of digital-video clips (i.e., “highlight reel”) from a set of digital-video clips.
  • the automatic selection is based, at least in part, upon various weighted criteria regarding properties (e.g., metadata, enhanced metadata, and/or content) of the clips.
  • a video-capturing device automatically produces a highlight reel by selecting the superlative clips (e.g., the best clips). If the user does not agree that the clips of the highlight reel are superlative, the device may produce a new highlight reel based upon one or more reweighted criteria.
  • FIG. 1 illustrates a video-capturing apparatus that is configured to implement the techniques described herein for automatic selection of one or more digital-video clips from a set of digital-video clips based, at least in part, upon weighted criteria.
  • FIGS. 2-4 are flowcharts of processes that are configured to implement the techniques described herein for automatic selection of one or more digital-video clips from a set of digital-video clips based, at least in part, upon weighted criteria.
  • FIG. 5 depicts a video-capturing telecommunications apparatus within a telecommunications environment.
  • the depicted apparatus is configured to implement the techniques described herein for automatic selection of one or more digital-video clips from a set of digital-video clips based, at least in part, upon weighted criteria.
  • Described herein are techniques related to an automatic production of a highlight reel by selecting of one or more digital-video clips from a larger set of such clips based, at least in part, upon weighted criteria.
  • a video-capturing device such as a digital camcorder or camera-equipped mobile phone
  • a user accumulates a large collection of digital-video clips by, for example, videoing over several days, weeks, or even months.
  • her video-capturing device automatically selects the superlative clips based upon various weighted criteria regarding properties (e.g., metadata, enhanced metadata and/or content) of the clips.
  • properties e.g., metadata, enhanced metadata and/or content
  • the user views the highlight reel.
  • the user may indicate approval or disapproval of the highlight reel. If approved, the weight values assigned to the various weighted criteria remain unchanged because those values seem to lead to a proper selection of clips. If the user disagrees with the selection, then the weight values are adjusted and the auto-selection is performed again but with reweighted values for the criteria.
  • Hope While generally happy with the auto-selected highlight reel, Hope is displeased that so few of the clips include her family members. So, when asked, she indicates that she does not agree that these clips are the “best” in her subjective opinion. Hope may indicate her displeasure by indicating that one or more particular clips do not belong amongst the “best” clips by removing them from the “best” group.
  • the smartphone In response to her dissatisfaction, the smartphone soon offers another auto-selection of seven minutes of clips.
  • the clips that she was displeased with e.g., removed from the “best” group
  • her contacts in her favorites list (which includes many friends and family members) receive a notification (via email, Short Message Service (SMS), social networking site, etc.) about Hope's new highlight reel for them to see. So, with little effort from Hope, the best seven minutes of over twelve hours of vacation videos are shared with so many of the people who are most important to her. This highlight reel is shared in a timely manner rather than weeks after her vacation and after manually editing her videos.
  • SMS Short Message Service
  • An example of an embodiment of a video-capturing apparatus that employs the new video clip auto-selection techniques described herein may be referred to as an “exemplary video-capturing device.” While one or more example embodiments are described herein, the reader should understand that the claimed invention may be practiced using different details than the exemplary ones described herein.
  • FIG. 1 shows a user 100 (such as Hope) holding a digital camcorder 110 with data-communication capability.
  • the device 110 is also shown in an expanded view.
  • the device 110 may be used to automatically select a small group of digital-video clips from amongst a large video collection.
  • the videos of the large collection may have been shot by the device 110 itself.
  • the device 110 has the capability to communicate (via cellular technology, WLAN, etc.) with a network infrastructure 120 .
  • This network infrastructure 120 may be a private or public network of computing or telecommunication resources. It may include the Internet. It may include the so-called “cloud.” As such, the network infrastructure 120 may be called the cloud-computing infrastructure.
  • the video-capturing device 110 includes a video-capturing unit 130 , a video-storage unit 140 , a highlight-production unit 150 , a highlight-presentation unit 160 , a user-interaction unit 170 , an actions unit 180 , and a communications unit 190 .
  • Each of these units may be implemented (at least in part) by a set of processor-executable instructions (e.g., software modules).
  • each of these units may include or employ one or more lower-level components of the device 110 .
  • these lower-level components include processors, memory, storage, video display, user-input device (e.g., keyboard), transceiver, and the like.
  • the video-capturing unit 130 is configured, designed, and/or programmed to capture digital-video clips. That is, a person using the device 110 may shoot a scene and/or person using the video-capturing unit 130 of the device.
  • the video-capturing unit 130 includes various lower-level camera components for capturing moving digital-video clips with audio (i.e., digital video) and perhaps still digital-video clips (i.e., photographs or clips).
  • the camera components may include (by way of example and not limitation): digital sensor chip (e.g., CCD (charge-coupled device) or CMOS (complementary metal-oxide-semiconductor)), lenses, display, view-finder, and the like.
  • the video-capturing unit 130 may be implemented, at least in part, by a software module resident, at least in part, in the device's memory and executed by one or more processors of the device.
  • the video-capturing unit 130 may include, or have associated therewith, a video-handling software module (e.g., application) that may be used to enhance the amount of information recorded in the digital-video file relating to the captured digital-video clips.
  • a video-handling software module e.g., application
  • the video-handling application may use information from other sources and/or applications to add data to the captured digital-video clips that are stored on the video-storage unit 140 . This added data may be called “metadata.”
  • the video-handling application may be configured to obtain information from other hardware, firmware, or software components to add data to the digital-video files. Examples of other components include (but are not limited to) a location application, a calendar application, and/or a contacts application.
  • the file formats for storing digital videos are sometimes referred to container formats from audio and video.
  • Common container formats for storing digital videos include (but are not limited to): 3GPP multimedia file (0.3 gp); anime music video file (.amv); advanced systems format file (.asf); Microsoft ASF redirector file (.asx); audio video interleave file (.avi); DivX-encoded movie (.divx); Nintendo DS movie file (.dpg); Apple QuickTime movie (.moov, .mov, .qt); MPEG-4 video file (.mp4); MPEG video file (.mpg); Real media file (.rm); Macromedia Flash movie (.swf); and Windows media video file (.wmv).
  • a “digital video” is one or more digital-video files.
  • a digital video includes one or more digital-video clips.
  • a digital-video file refers to the formatted container of a digital video as it is stored on the video-storage unit 140 or other storage device.
  • a digital-video clip (or simply a clip) refers to a short portion of a digital video that often contains an individual scene, which is a video sequence that is typically shot in one continuous take.
  • the video-handling application may determine or identify particular clips in the one or more stored files of the digital video.
  • the video-handling application identifies a particular clip by using existing metadata that specifies the boundaries of the particular clip (i.e., where an individual clip begins and ends).
  • the video-handling application may utilize existing or new video-editing technology to locate a clip based upon the content of the digital video. These video-editing technologies are commonly known as “automatic scene detection” to those of ordinary skill in the art.
  • the video-handling application may be designed to enhance user functionality once clips have been obtained.
  • the video-handling application may also be configured to display clips to the user in cooperation with or as part of the highlight-presentation unit 160 .
  • the video-handling application may include various filters or criteria used to limit the number of clips displayed to the user. As discussed below, these filters or criteria may be user selectable, may use the data in the digital-video file obtained from non-video sources or applications, or may be configured based on data in the digital-video files, etc. As another example, similar filters or criteria may also be used to cluster clips into folders (such as virtual albums, system file folders, etc.). As still another example, the video-handling application may use data stored in the digital-video files, contact information, calendar information, and/or upload information to increase the ease of sharing clips.
  • the clips operated on by the video-handling application may include clips captured by the video-capturing unit 130 , and/or may include clips obtained from sources other than the video-capturing unit 130 .
  • images may be transferred to device 110 using one or more wireless and/or wireline interfaces from the network 120 .
  • the video-handling application may be a stand-alone application, or may be integrated into other applications or other units. Moreover, the video-handling application may be formed by a combination of functions of separate, distinct programs or units of the device 110 .
  • the video-storage unit 140 is configured, designed, and/or programmed to store digital-video clips as digital-video files and possibly other forms of data and software. That is, the device 110 stores the digital-video files of clips shot by the person using the device 110 on the video-storage unit 140 of the device.
  • the video-storage unit 140 includes one or more lower-level memory or storage components for storing moving digital images with audio (i.e., digital video).
  • the memory or storage components may be volatile or non-volatile, dynamic or static, read/write-able or read only, random- or sequential-access, location- or file-addressable, and the like.
  • the memory or storage components may be magnetic, optical, holographic, and the like.
  • the memory or storage components may be internal to the device 110 , attached externally to the device 110 , or available via data communications (e.g., on the network or the cloud).
  • the clips stored by the video-storage unit 140 may include clips shot by the video-capturing unit 130 , and/or may include clips obtained from sources other than the video-capturing unit 130 .
  • clips may be transferred to the device 110 using one or more wireless and/or wireline interfaces from the network 120 .
  • the highlight-production unit 150 is configured, designed, and/or programmed to automatically select a group of superlative (e.g., top-ranked or bottom-ranked) clips of a collection of such clips stored by the video-storage unit 140 .
  • the highlight-production unit 150 may cluster the collection of clips before the auto-selection is performed.
  • the highlight-production unit 150 may cluster the auto-selected group of clips after the auto-selection is performed.
  • the highlight-production unit 150 shown in FIG. 1 is implemented as a software module which would reside, at least in part, in the device's memory and be executed by the device's one or more processors.
  • the highlight-production unit 150 may be implemented as a collection of or as part of dedicated hardware or firmware.
  • the highlight-production unit 150 may be implemented as a combination of hardware, firmware, or software.
  • the highlight-presentation unit 160 is configured, designed, and/or programmed to visually present digital-video clips to the user for her enjoyment and/or consideration. That is, a person using the device 110 may view a selection of clips using the highlight-presentation unit 160 of the device.
  • the highlight-presentation unit 160 includes various lower-level audio/visual presentation components for showing moving digital images with audio (i.e., digital video).
  • the audio/visual components may include (by way of example and not limitation): a liquid crystal display (LCD), a flat panel, organic light-emitting diode (OLED) displays, pico-projection displays, a solid state display or other visual display devices, speakers, and the like.
  • the highlight-presentation unit 160 may be implemented, at least in part, by a software module resident, at least in part, in the device's memory and executed by one or more processors of the device.
  • the user-interaction unit 170 is configured, designed, and/or programmed to attain feedback (e.g., obtain input) from the user and, in particular, feedback related to the clips presented in cooperation with the highlight-presentation unit 160 . That is, a person using the device 110 indicates approval of or disapproval with the auto-selected group of clips presented on-screen by using the user-interaction unit 170 of the device.
  • the user-interaction unit 170 includes various lower-level user-input components for receiving input from the user, such as (by way of example and not limitation): keyboard, touchscreen, mouse, touchpad, trackball, and the like.
  • the user-interaction unit 170 may use some of the same audio/visual components of the highlight-presentation unit 160 .
  • the user-interaction unit 170 is implemented as a software module which would reside, at least in part, in the device's memory and be executed by the device's one or more processors.
  • the highlight-production unit 150 may be implemented as a collection of or as part of dedicated hardware or firmware.
  • the user-interaction unit 170 may be implemented as a combination of hardware, firmware, or software.
  • the actions unit 180 is configured, designed, and/or programmed to automatically perform a defined action on the group of top-ranked (or alternatively bottom-ranked) clips of a collection of such clips stored by the video-storage unit 140 .
  • the defined actions include (by way of example and not limitation): archiving, sharing, burning, conversion/reformat, and the like.
  • Presuming the clips are bottom-ranked, the defined actions include (by way of example and not limitation): deleting, recycling, and the like.
  • Sharing clips involves storing clips in a different and perhaps more reliable location, such as onto the cloud-computing infrastructure.
  • Sharing clips includes sending copies of the clips to another person via one or more various ways of sending such data or notices of the same.
  • sharing clips includes sending a link to one or more of the clips via one or more various ways of sending such links or notices of the same. Examples of such ways to send clips, links, and/or notices thereof include (but are not limited to): email, posting on a social network, posting on a blog or website, text message, MMS (multimedia messaging service), and the like.
  • MMS multimedia messaging service
  • the clips When the clips are burned, they are packaged and written onto a removeable media (like a DVD disc) so that the clips can be viewed on a television with an attached DVD player. Also, the clips may be automatically converted or reformatted in a pre-defined manner.
  • Deleting clips involves permanently removing one or more clips from the video-storage unit 140 . Recycling clips involves placing one or more clips into a queue of files to be deleted later. This queue is sometimes called the “trash” or “recycle bin.”
  • the actions unit 180 shown in FIG. 1 is implemented as a software module which would reside, at least in part, in the device's memory and be executed by the device's one or more processors.
  • the actions unit 180 may be implemented as a collection of or as part of dedicated hardware or firmware.
  • the actions unit 180 may be implemented as a combination of hardware, firmware, or software.
  • the communications unit 190 is configured, designed, and/or programmed to transmit (and/or receive) notifications of, links to, and/or copies of digital-video clips through the network 120 and onto another user or device accessible from that network.
  • a user of the device 110 may have an auto-selected group of the clips that the person shot sent to their friends upon the person's approval of that auto-selected group.
  • the communications unit 190 includes various lower-level communications components for sending and/or receiving data communications from/by the device. Using transceiver, transmitter, receiver, network interface controller (NIC), and the like, the communications unit 190 utilizes wired (e.g., universal serial bus (USB) or Ethernet) or wireless communications. Examples of wireless communications are (by way of example and not limitation): cellular, satellite, conventional analog AM or FM radio, Wi-FiTM, wireless local area network (WLAN or IEEE 802.11), WiMAXTM (Worldwide Interoperability for Microwave Access), and other analog and digital wireless voice and data transmission technologies.
  • the communications unit 190 may be implemented, at least in part, by a software module resident, at least in part, in the device's memory and executed by one or more processors of the device.
  • FIGS. 2-4 are flowcharts illustrating exemplary processes 200 , 300 , and 400 that implement the techniques described herein for automatic highlight reel production.
  • the exemplary processes 200 , 300 , and 400 are performed, at least in part, by a video-capturing device, such as device 110 of FIG. 1 , a video-capturing device 204 of FIG. 2 , and/or a video-capturing telecommunications device 510 of FIG. 5 . Additionally and alternatively, the exemplary processes 200 , 300 , and 400 are performed, at least in part, by a cloud-based service.
  • the process 200 starts, at 210 , with an update of a set 212 of digital-video clips (i.e., clips) by, for example, a user 202 shooting new clips with a video-capturing device 204 , such as a digital camcorder.
  • the set 212 of digital-video clips may be updated via other sources such as from a nearby BluetoothTM device, via cellular connection, from the Internet (e.g., email, social networking site, website, etc.), or some cloud-computing infrastructure resource.
  • the set 212 of digital-video clips may be stored on the device 204 , on another data-communicative device, on a network server, on the Internet, in the cloud, or some combination thereof.
  • the box representing operations 210 is vertically long and horizontally narrow so as to indicate that this operation may occur while the other to-be-described operations are performed.
  • the set 212 of digital-video clips may represent all of the clips stored and/or accessible by the device 204 .
  • the set 212 includes a group of clips that are less than all of the clips stored and/or accessible by the device 204 .
  • the set 212 is a collection or cluster of some but not all of the stored and/or available clips. Clips may clustered based on one or more criteria, such as location, time, date, and calendar information from one or more events, such as those on private and/or publicly available calendars.
  • clustering criteria may also consider event specifics, such as, faces in the clusters, colors associated with clips, background and lighting of the clips, including overall brightness, color similarities or dissimilarities, and scene information.
  • Clusters of clips may be multilayered. For example, a cluster of wedding clips may have a number of secondary clusters. One secondary cluster may include clips of just the groom and bride, where another secondary cluster may include clips of just the bridesmaids.
  • the device 204 provides enhanced metadata for the digital video.
  • This enhanced-metadata provision includes generating the enhanced metadata and associating it with the digital video (and the clips that make up that associated video).
  • the enhanced metadata is derived from the existing metadata already associated with the digital video (e.g., location and time) and/or an analysis of the content of the video.
  • the content analysis detects and determines occurrence (or lack thereof) of temporal events in the video rather than something based on a single image or frame.
  • the content of the audio and/or the moving images portions of the digital video are analyzed.
  • This analysis may look for and note one or more of the following content conditions based upon temporal events (provided by way of illustration only and not limitation): blank segment, silent or quiet audio, lack of sync between audio and video, no audible voices, existence of faces, identification of faces, emotions expressed on faces, motion detection, human motion detected, presence and identification of particular objects, person or object tracking, music detected, language detected, and the like.
  • the device 204 auto-selects a subset of the set 212 of digital-video clips.
  • the subset of clips forms an effective “highlight reel” of the larger set 212 of digital-video clips.
  • the subset presumptively includes some, but not all of, the digital-video clips of the set 212 .
  • the selected subset includes the superlative clips from the set 212 of clips.
  • the term “superlative” refers to a subset of clips having the highest kind, quality, or order with respect to the other clips of a set. For example, clips that are top-ranked or bottom-ranked in some category or characteristic are superlative. Also, for example, clips considered the “best” or “worst” of the set are considered superlative.
  • the auto-selection is based upon one or more weighted selection criteria 218 of one or more properties of the clips.
  • FIG. 3 provides more information regarding both the auto-selection operation 216 and the selection criteria 218 .
  • the auto-selection operation 216 may be designed to find the “best” clips (i.e., the top-ranked clips). Alternatively, it may be designed to find the “worst” clips (i.e., the bottom-ranked clips).
  • the device 204 presents the highlight reel to the user 202 via a user interface (of a telecommunication device or a cloud-connected computer device).
  • a user interface of a telecommunication device or a cloud-connected computer device.
  • the user 202 indicates his approval of the highlight reel.
  • the user 202 may be, for example, agreeing that the highlight reel represents the “best” clips in his subjective opinion.
  • disapproving the user 202 may be, for example, disagreeing that the highlight reel represents the “best” clips in his subjective opinion.
  • the device 204 attains feedback from the user regarding the highlight reel.
  • the device 204 determines whether input from the user indicates approval of the highlight reel. If not approved, then the device 204 updates (i.e., adjusts or alters) the weight values assigned to one or more of the selection criteria 218 and the process returns to the auto-selection operation 216 to perform a reweighted selection of the clips based upon the now updated criteria.
  • the device 204 automatically reselects an updated highlight reel. If the user input indicates, at 222 , that the user disapproved, the process proceeds to operation 224 .
  • the user 202 may indicate his disapproval by him choosing to remove or add one or more clips from/to the highlight reel.
  • the device 204 may “learn” the types of images that a user prefers or does not prefer based on iterations of approval and disapproval by the user.
  • the device 204 reduces the weight values for the removed clip's strong properties and/or increases the weight values for the removed clip's weak properties. And vice versa for the clips that the user manually adds to the subset.
  • the multiple iterations hone the weight values assigned to various weighted selection criteria.
  • the device determines whether input from the user (most likely in addition to the user input regarding approval) indicates whether to take a defined action upon the subset of clips and/or which of several possible defined actions to take upon the subset. If the user does not want any defined action, then the process returns to the beginning operation at 210 . If the user wants a defined action to be performed upon the subset, then the process proceeds to action operation at 226 .
  • operation 224 is shown as a dashed box to indicate that it is optional and some implementations may not offer the user this choice.
  • the device 204 may perform one or more defined actions by default.
  • the user may have pre-set whether and/or which one or more of the defined actions will be performed upon the clips of the highlight reel.
  • the device 204 performs one or more defined actions on one or more of the clips of the highlight reel. Presuming the clips are top-ranked (i.e., “best”), the defined actions include (by way of example and not limitation): archiving, sharing, media burning, converting/reformatting, and the like. Presuming the clips are bottom-ranked (i.e., “worst”), the defined actions include (by way of example and not limitation): deleting, recycling, and the like.
  • the process returns to the video-set updating operation 210 .
  • the operations 214 through 226 may be performed in real-time with the video-set updating operation 210 .
  • FIG. 3 shows the process 300 and offers more details regarding the auto-selection operation 216 and the selection criteria 218 .
  • the auto-selection operation 218 includes a binary selection filter operation 302 followed by a weighted-criteria selection filter operation 304 .
  • one of more of the criteria 210 may be applied as a binary filter to, for example, remove clips that do not meet the one or more criteria.
  • one or more of the criteria are applied in a weighted manner.
  • the main purpose of the binary selection filter operation 302 is to remove outright “bad” clips. So, it could be called the bad-clip filter.
  • a bad clip is one where the content is difficult for a human to discern, incoherent, would certainly be uninteresting, and/or simple post-processing is not likely to alleviate the problem.
  • the binary filter may use one or more criteria related to one or more properties that might make a clip be considered “bad.” Examples of such “bad clip” criteria include (but are not limited to): shaky, blurry, over or under exposed, poor resolution, poor contrast, out-of-focus, too short, stillness and/or silent for too long, audio-video out of sync, red-eye, and the like. Out-of-sync and red-eye might not be part of the binary selection filter operation 302 in some implementations because they can be fixed with some post-processing.
  • Some implementations may not employ the binary selection filter operation 302 . Those implementations may rely upon the weighted-criteria selection filter operation 304 to remove the “bad” clips.
  • the binary selection filter operation 302 may be described as grouping the set 212 of digital-video clips into an intermediate grouping that includes two or more of the digital-video clips of the set 212 .
  • the weighted-criteria selection filter operation 304 may be alternatively described as ranking the digital-video clips of the intermediate grouping based upon one or more of the multiple weighted criteria 218 and designating an allotment of top-ranked (or bottom-ranked) digital-video clips of the intermediate grouping as the selected subset of digital-video clips that results from the auto-selection operation 216 .
  • the size (i.e., number, cardinality) of the allotment may be a fixed number (e.g., ten) or calculated. That allotment size may refer to number of clips, number of frames, length of highlight reel (e.g., ten minutes), or the like. If fixed, the allotment size may be a factory-set default number or a user-defined number. If calculated, the calculation may be, for example, based upon linear or non-linear proportions of the quantity or length of clips in the set 212 , in the intermediate grouping, and/or the like. These proportions may be factory-set or user-defined. A percentage is an example of such proportions.
  • the allotment size for the subset may be user-defined to be five percent of the number or length of clips in the set 212 . If, for this example, the set 212 included 100 clips totaling three hours of video, then the allotment size might be five clips or perhaps nine minutes. Consequently, the user 202 would be presented with fives clips or perhaps nine minutes of clips for his approval.
  • the criteria used for the binary filtering differ from those used for the weighted filtering. However, in some implementations, the criteria used for one may overlap the criteria used for the other.
  • the criteria are related to properties of a clip. Those properties may be derived from the content of the clip (based upon an image or video analysis) and/or from metadata associated with the clip.
  • clip metadata examples include (but are not limited to): technical metadata (such as encoding details, identification of clip boundaries, size, resolution, color profile, ISO speed and other camera settings), descriptive metadata (captions, headlines, titles, keywords, location of capture, etc.), and administrative metadata (such as licensing or rights usage terms, specific restrictions on using an image, model releases, provenance information, such as the identity of the creator, and contact information for the rights holder or licensor).
  • technical metadata such as encoding details, identification of clip boundaries, size, resolution, color profile, ISO speed and other camera settings
  • descriptive metadata examples, headlines, titles, keywords, location of capture, etc.
  • administrative metadata such as licensing or rights usage terms, specific restrictions on using an image, model releases, provenance information, such as the identity of the creator, and contact information for the rights holder or licensor).
  • the weighted selection criteria 218 include (by way of example and not limitation): presence or absence of audio from a corresponding clip, motion, smiling faces, laughter, black screens, identification of video-capture device, identification of video-capture device user, location and/or time of video-capture, focus, contrast, shake, red eye, person in clip is a favorite (e.g., in a social network or on the device), person in clip is a contact, clip is tagged (in general or with a particular tag), user-added criteria, clip was auto-corrected, flash used, social network ranking of a clip, etc.
  • FIG. 4 illustrates the process 400 for determining a best clip that may be used with embodiments described herein. This process 400 may be employed as part of process 200 and/or process 300 herein.
  • the process 400 starts with operation 402 where a video-capture device retrieves the next clip (or the first when initially executed) from a memory like that of video-storage unit 140 of device 110 (as shown in FIG. 1 ).
  • a determination is made if the retrieved clip is shaky or if redeye is detected. If so, a next clip is retrieved and the process 400 starts again with that next clip. If the retrieved clip is the last (as determined at operation 438 ), then the process 400 proceeds to a clip-ranking operation 440 , which is discussed later.
  • the process 400 continues onto operation 408 where it is determined if the resolution of the retrieved photo is greater than a defined threshold resolution. If not, then a next photo is retrieved and the process begins again. If so, based on the resolution of the image, a resolution index is determined or calculated at operation 410 . In one example, the resolution index may be determined or calculated based on the resolution (megapixels) associated with a photo undergoing consideration.
  • the process determines if the histogram associated with the photo is within a defined threshold. If not, then a next photo is retrieved and the process begins again. If so, based on the histogram of the photo, a histogram index is determined or calculated at 414 . In one example, the histogram index is determined or calculated based on a consideration of the number of “dark” pixels (underexposure), “bright” pixels (overexposure), and/or the number of pixels that fall between the underexposure and overexposure range.
  • the process continues, at 416 , to determine if a duplicate clip of the one currently being processed is stored in the device or otherwise known or available. If so, a next clip is retrieved and the process begins again. If not, the process determines, at 418 , if one or more faces are detected in the clip. If no faces are detected in the clip, the process calculates, at 420 , a NF (“no face”) clip score based on the calculated resolution index and histogram index, and associates the NF score with the clip and adds the clip to a “no face” clip queue at 422 . If one or more faces is detected, the process determines, at 424 , if a subject in the clip has their eyes shut or partially shut (i.e., “eyes closed detection”). If that is detected, a next clip is retrieved and the process begins again.
  • NF no face
  • the process calculates face, event, alignment and user like indexes at operations 426 , 428 , 430 , 432 , respectively.
  • the face index calculation considers if one or more expected or desired faces are in a photo
  • the event index calculation considers if an expected event (e.g., wedding or party) is part of a photo
  • the alignment calculation considers a degree of alignment (e.g., vertical or horizontal) of a photo
  • the user like index considers a relative user acceptance or approval (e.g., like or dislike) of one or more faces in a photo.
  • the queue is automatically resorted based on scores associated with an already processed and queued clip.
  • the process picks the highest scored clip as the best clip at 440 .
  • operation 440 may also select, based on clip score, a set of clips for presentation to a user of the portable device.
  • FIG. 5 shows an exemplary telecommunication environment 500 which includes a video-capturing telecommunications device 510 . Both the environment 500 and the device 510 may be included as part of one or more implementation of the techniques described herein.
  • the device 510 is a smartphone that is capable of communicating using analog and/or digital wireless voice and/or data transmission.
  • FIG. 5 shows the device 510 in communication with a cell phone 520 via a so-called cellular network 522 as represented by cell tower.
  • the device 510 is coupled to a cloud-computing infrastructure 524 .
  • the cloud-computing infrastructure 524 may include scalable and virtualized compute resources over a private or public network. Typically, a cloud-computing infrastructure uses the internet and central remote servers to maintain data and applications.
  • the cloud-computing infrastructure 524 includes a group of individual computers connected by high-speed communications and some form of management software guiding the collective actions and usage of the individual computers. This technology allows for much more efficient computing by centralizing storage, memory, processing and bandwidth. While depicted in FIG. 5 as the cloud-computing infrastructure, item 524 may alternatively be viewed as a network or the Internet.
  • the device 510 includes many components, such as a wireless transceiver 530 , a NIC (network interface controller) 532 , one or more processors 534 , one or more cameras 536 , one or more input subsystems 538 , one or more output subsystems 540 , a secondary storage 542 with at least one processor-readable medium 544 , one or more location subsystems 546 , one or more power systems 548 , one or more audio systems 550 , one or more video displays 552 , and a memory 554 having software 556 and data (e.g., digital photo or digital video) 558 stored thereon.
  • a wireless transceiver 530 a NIC (network interface controller) 532 , one or more processors 534 , one or more cameras 536 , one or more input subsystems 538 , one or more output subsystems 540 , a secondary storage 542 with at least one processor-readable medium 544 , one or more location subsystems 546 , one
  • the NIC 532 is hardware component that connects the device 510 to one or more computer networks and allow for communication to and from that network.
  • the NIC 532 operates as both an Open Systems Interconnection (OSI) layer 1 (i.e., physical layer) and layer 2 (i.e., data link layer) device, as it provides physical access to a networking medium and provides a low-level addressing system through the use of Media Access Control (MAC) addresses. It allows devices to connect to each other either by using cables or wirelessly.
  • OSI Open Systems Interconnection
  • layer 2 i.e., data link layer
  • MAC Media Access Control
  • the one or more processors 534 may include a single or multi-core central processing unit (CPU), a graphics processing unit (GPU), other processing unit or component, or some combination thereof.
  • CPU central processing unit
  • GPU graphics processing unit
  • the camera 536 may be configurable to capture still images (clips), moving images (video), or both still and moving images.
  • the camera components may include (by way of example and not limitation): digital sensor chip (e.g., CCD (charge-coupled device) or CMOS (complementary metal-oxide-semiconductor)), lenses, flash, and the like.
  • the input subsystems 538 are physical devices designed to allow for input from, for example, a human user.
  • the input subsystems 538 may include (by way of example and not limitation): keyboard, keypad, touchscreen, touchpad, mouse, trackball, paddle, light pen, scanner, stylus, and/or a micro-telecommunications device.
  • the output subsystems 540 are mechanisms for providing a tactile output to the user in a manner that is neither audio nor video based. Such tactile output is typically generated by an offset rotating motor that vibrates the device 510 to give a notification to the user.
  • the secondary storage 542 is typically a read-only memory (ROM), a programmable memory (such as EEPROM, EPROM, or PROM), a static read/writable memory (such as flash memory), and a mechanical read/writeable device (such as hard drive), which may be magnetic, optical, holographic, and the like.
  • the components of the secondary storage 542 may be internal to the device 510 , attached externally to the device 510 , or available via data communications (e.g., on the cloud).
  • At least one processor-readable medium 544 is stored on the secondary storage 542 .
  • the processor-readable medium 544 stores one or more sets of instructions (e.g., software) embodying any one or more of the methodologies or functions described herein.
  • the instructions may also reside, completely or at least partially, within the memory 554 as software 556 and within the processor(s) 534 during execution thereof by the device 510 .
  • the location subsystems 546 may be designed to determine the physical location of the device 510 .
  • the one or more location subsystems 546 may include (but is not limited to) an application and hardware for determining location via global positioning system (GPS) and/or by communicating with a cellular network.
  • GPS global positioning system
  • the power subsystems 548 may include (but is not limited to) one or more batteries and one or more external power supplies, such as an interface to an external power supply source.
  • the audio systems 550 are configured to generate sound, noise, voices, and music. In addition, the audio systems 550 are configured to capture and process sound, noise, voices, and music.
  • the audio systems 550 includes (but is not limited to) speakers and microphones.
  • the video display 552 is configured to display images, videos, text, user-interfaces, and the like.
  • the video display 552 may include (but is not limited to) a liquid crystal display (LCD), a flat panel, a solid state display or other device.
  • the video display 552 may operate as a digital viewfinder that allows a user to preview a scene before capturing an image and/or to view a movie as it is being captured.
  • the memory 554 is configured to store software 556 and data (e.g., digital photo or digital video) 558 thereon.
  • the memory 554 is a working space and may include (but is not limited to) random access memory (RAM).
  • RAM random access memory
  • the memory 554 may be one or more of the following (by way of example and not limitation): volatile or non-volatile, dynamic or static, read/write-able or read only, random- or sequential-access, location- or file-addressable, and the like.
  • the components of the memory 554 may be internal to the device 510 , attached externally to the device 510 , or available via data communications (e.g., on the cloud).
  • the exemplary devices and apparatuses discussed herein that implement at least some aspect of the techniques discussed herein include (by way of illustration and not limitation): the video-capturing device 110 from FIG. 1 , the video-capturing device 204 of FIG. 2 , and the video-capturing telecommunications device 510 of FIG. 5 .
  • Such exemplary devices may be what is commonly referred to as a “mobile phone,” “smartphone,” and/or “cellphone.”
  • the described techniques can be used in conjunction with non-cellular technologies such as conventional analog AM or FM radio, Wi-FiTM, wireless local area network (WLAN or IEEE 802.11), WiMAXTM (Worldwide Interoperability for Microwave Access), BluetoothTM, and other analog and digital wireless voice and data transmission technologies.
  • Alternative implementations of such devices might not have any telecommunications or wireless communication capability.
  • Still other alternative implementations may utilize a wired communication link instead of or in addition to wireless communication.
  • the exemplary devices are not limited to those having a camera. Any device with the capability to acquire, collect, and/or manipulate digital images or moving digital images (i.e., video) may implement some aspect of the techniques discussed herein. Examples of such alternative devices include (but are not limited to) the following: tablet-based computer, other handheld computer, netbook, computer, digital camera, digital camcorder, handheld multimedia device, digital single-lens reflex (DSLR) camera, GPS navigational system, vehicle-based computer, or other portable electronics.
  • DSLR digital single-lens reflex
  • the exemplary devices may operate on digital videos acquired from sources other than their own video-capturing components. These other video sources may be other video-capturing devices or from non-camera-equipped sources.
  • the non-camera-equipped sources include other devices or services where images and/or videos are stored, collected, accessed, handled, manipulated, and/or viewed.
  • image/video sources include (but are not limited to): a clip-processing kiosk, portable and removable storage media or device (e.g., CD-ROM, DVD-ROM, other optical media, USB flash drive, flash memory card, external hard drive, etc.), electronics with radio-frequency identification (RFID), a social-networking services (e.g., FacebookTM, MySpaceTM, TwitterTM, LinkedinTM, NingTM, and the like), and photo/video sharing services (e.g., FlickrTM, PhotobucketTM, PicasaTM, ShutterflyTM, and the like).
  • portable and removable storage media or device e.g., CD-ROM, DVD-ROM, other optical media, USB flash drive, flash memory card, external hard drive, etc.
  • RFID radio-frequency identification
  • social-networking services e.g., FacebookTM, MySpaceTM, TwitterTM, LinkedinTM, NingTM, and the like
  • photo/video sharing services e.g., FlickrTM, PhotobucketTM, PicasaTM
  • One or more implementations of the techniques described herein may include an initial training session for the auto-selecting device or service to learn what makes a photo and/or video clip superlative (e.g., best or worst) in the user's opinion. This training may be repeated from time to time. This training will set the weight values of one or more of the weighted criteria. There may be various profiles for specific conditions and situations where the weighted criteria have their own trained values. Each user of an auto-selecting device or service may have their own customized weighted criteria values and profiles of such values. An auto-selecting device or service may have a default set of values or profiles assigned. That default set or profiles may be extrapolated from a statistically evaluation (e.g., mean or median) of the trained values derived from multiple users.
  • a statistically evaluation e.g., mean or median
  • one or more implementations may take advantage of a system of user-ratings (e.g., thumbs up/down or 1 to 5 stars) for photos and/or video clips. These user-rating are effective data for training the exemplary video-capture device to learn what makes a photo and/or video clip superlative (e.g., best or worst) in the user's opinion.
  • the user may define values for one or more of the weighted criteria. This may be accomplished using, for example, slider-bar user interface for the user to adjust weights assigned to particular criteria or to categories of criteria.
  • exemplary is used herein to mean serving as an example, instance, or illustration. Any aspect or design described herein as “exemplary” is not necessarily to be construed as preferred or advantageous over other aspects or designs. Rather, use of the word exemplary is intended to present concepts and techniques in a concrete fashion.
  • techniques may refer to one or more devices, apparatuses, systems, methods, articles of manufacture, and/or computer-readable instructions as indicated by the context described herein.
  • the term “or” is intended to mean an inclusive “or” rather than an exclusive “or.” That is, unless specified otherwise or clear from context, “X employs A or B” is intended to mean any of the natural inclusive permutations. That is, if X employs A; X employs B; or X employs both A and B, then “X employs A or B” is satisfied under any of the foregoing instances.
  • the articles “a” and “an” as used in this application and the appended claims should generally be construed to mean “one or more,” unless specified otherwise or clear from context to be directed to a singular form.
  • the exemplary processes discussed herein are illustrated as a collection of blocks in a logical flow graph, which represents a sequence of operations that can be implemented with hardware, software, firmware, or some combination thereof.
  • the blocks represent instructions stored on one or more processor-readable storage media that, when executed by one or more processors, perform the recited operations.
  • the operations of the exemplary processes may be rendered in virtually any programming language or environment including (by way of example and not limitation): C/C++, Fortran, COBOL, PASCAL, assembly language, markup languages (e.g., HTML, SGML, XML, VoXML), and the like, as well as object-oriented environments such as the Common Object Request Broker Architecture (CORBA), JavaTM (including J2ME, Java Beans, etc.), Binary Runtime Environment (BREW), and the like.
  • CORBA Common Object Request Broker Architecture
  • JavaTM including J2ME, Java Beans, etc.
  • BREW Binary Runtime Environment
  • processor-readable media includes processor-storage media.
  • processor-storage media may include, but are not limited to, magnetic storage devices (e.g., hard disk, floppy disk, and magnetic strips), optical disks (e.g., compact disk (CD) and digital versatile disk (DVD)), smart cards, flash memory devices (e.g., thumb drive, stick, key drive, and SD cards), and volatile and non-volatile memory (e.g., random access memory (RAM), read-only memory (ROM)).
  • magnetic storage devices e.g., hard disk, floppy disk, and magnetic strips
  • optical disks e.g., compact disk (CD) and digital versatile disk (DVD)
  • smart cards e.g., compact disk (CD) and digital versatile disk (DVD)
  • smart cards e.g., compact disk (CD) and digital versatile disk (DVD)
  • flash memory devices e.g., thumb drive, stick, key drive, and SD cards
  • volatile and non-volatile memory e.g.,
  • Coupled and “connected” may have been used to describe how various elements interface. Such described interfacing of various elements may be either direct or indirect.

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Library & Information Science (AREA)
  • General Physics & Mathematics (AREA)
  • Physics & Mathematics (AREA)
  • General Engineering & Computer Science (AREA)
  • Databases & Information Systems (AREA)
  • Data Mining & Analysis (AREA)
  • Information Transfer Between Computers (AREA)
  • Studio Devices (AREA)
  • Television Signal Processing For Recording (AREA)
  • Information Retrieval, Db Structures And Fs Structures Therefor (AREA)
  • Management, Administration, Business Operations System, And Electronic Commerce (AREA)

Abstract

Described herein are techniques related to automatic selection of a subset of digital-video clips (i.e., “highlight reel”) from a set of digital-video clips. The automatic selection is based, at least in part, upon various weighted criteria regarding properties (e.g., metadata, enhanced metadata, and/or content) of the clips. A video-capturing device automatically produces a highlight reel by selecting the superlative clips (e.g., the best clips). This Abstract is submitted with the understanding that it will not be used to interpret or limit the scope or meaning of the claims.

Description

    RELATED APPLICATIONS
  • This application claims the benefit of priority to U.S. Provisional Patent Application Ser. No. 61/435,757, filed on Jan. 24, 2011, the disclosure of which is incorporated by reference herein.
  • BACKGROUND
  • With the widespread use of digital consumer electronic video-capturing devices such as digital camcorders and camera-equipped mobile phones, the size of each consumers' digital-video collections continue to increase very rapidly. As such collections grow ever larger and increasingly more unwieldy, a person is less able to handle the sheer volume of these collections.
  • An individual person may shoot several hours of digital video at a single event (such as a wedding or a party) or even hundreds of hours of video on a trip (such as to the Grand Canyon) or over a season (the summer). Later, sometimes long after the event, trip, or season is over, the individual videographer may attempt to edit the many hours of video spending hours watching them and editing the videos using the available conventional approaches for doing so. With these conventional approaches, a person may use a crude user interface to view, select, and piece together the clips of the digital video over a timeline and perhaps form a watchable summary of the event, trip, or season.
  • Software applications that manage media (e.g., digital photographs and videos) collections have become widely adopted as the amount of digital media, including digital video, has grown. Because of its large display, processing power, and memory capacity, most of these conventional approaches are concentrated on use with personal “desktop” computers. Using the ample room of the desktop computer's large displays, the desktop user interface offers a broad workspace for a user to view and manage a large catalogue of digital-video clips.
  • As evidenced by seeming ubiquity of mobile personal communication devices (e.g., the so-called “smartphones”), mobile communication and computing technology has rapidly developed. With the processing power and memory capacity of such mobile technology, it is possible to have large media collections on mobile personal communication devices or have access to such collections via high-speed data telecommunication networks.
  • While the power and memory capacities have increased, a typical mobile personal communication device still has a small display and, consequently, more constrained user-interface capabilities than a desktop computer. Accordingly, a user of a typical mobile personal communication device is forced to abandon his mobile environment and move to a desktop computer in order to view and manage her large media catalogue.
  • SUMMARY
  • Described herein are techniques related to automatic selection of a subset of digital-video clips (i.e., “highlight reel”) from a set of digital-video clips. The automatic selection is based, at least in part, upon various weighted criteria regarding properties (e.g., metadata, enhanced metadata, and/or content) of the clips. A video-capturing device automatically produces a highlight reel by selecting the superlative clips (e.g., the best clips). If the user does not agree that the clips of the highlight reel are superlative, the device may produce a new highlight reel based upon one or more reweighted criteria.
  • This Summary is submitted with the understanding that it will not be used to interpret or limit the scope or meaning of the claims. This Summary is not intended to identify key features or essential features of the claimed subject matter, nor is it intended to be used as an aid in determining the scope of the claimed subject matter.
  • BRIEF DESCRIPTION OF THE DRAWINGS
  • FIG. 1 illustrates a video-capturing apparatus that is configured to implement the techniques described herein for automatic selection of one or more digital-video clips from a set of digital-video clips based, at least in part, upon weighted criteria.
  • FIGS. 2-4 are flowcharts of processes that are configured to implement the techniques described herein for automatic selection of one or more digital-video clips from a set of digital-video clips based, at least in part, upon weighted criteria.
  • FIG. 5 depicts a video-capturing telecommunications apparatus within a telecommunications environment. The depicted apparatus is configured to implement the techniques described herein for automatic selection of one or more digital-video clips from a set of digital-video clips based, at least in part, upon weighted criteria.
  • The Detailed Description references the accompanying figures. In the figures, the left-most digit(s) of a reference number identifies the figure in which the reference number first appears. The same numbers are used throughout the drawings to reference like features and components
  • DETAILED DESCRIPTION
  • Described herein are techniques related to an automatic production of a highlight reel by selecting of one or more digital-video clips from a larger set of such clips based, at least in part, upon weighted criteria. Using a video-capturing device (such as a digital camcorder or camera-equipped mobile phone), a user accumulates a large collection of digital-video clips by, for example, videoing over several days, weeks, or even months. Rather than manually watching and editing the large video collection to find a group of superlative clips (e.g., the best clips), her video-capturing device automatically selects the superlative clips based upon various weighted criteria regarding properties (e.g., metadata, enhanced metadata and/or content) of the clips. When assembled together, the auto-selected collection of clips is called a “highlight reel” herein.
  • Via a user-interface on the device or on another device with a display, the user views the highlight reel. The user may indicate approval or disapproval of the highlight reel. If approved, the weight values assigned to the various weighted criteria remain unchanged because those values seem to lead to a proper selection of clips. If the user disagrees with the selection, then the weight values are adjusted and the auto-selection is performed again but with reweighted values for the criteria.
  • For instance, consider the following scenario to illustrate how the techniques described herein may be employed. Using her camera-equipped smartphone, Hope shoots about twelve hours of video during a summer vacation to the Oregon Coast with her family. At some point during her vacation, Hope notices a notification on her smartphone asking her if she would like to look at her “highlight reel.” Curious, she, of course, watches the approximately seven-minute long highlight reel. In this example, the highlight reel includes the top-rated one percent (i.e., about seven minutes) of the twelve hours that the smartphone had auto-selected as the “best” of the clips that she captured during her vacation.
  • While generally happy with the auto-selected highlight reel, Hope is displeased that so few of the clips include her family members. So, when asked, she indicates that she does not agree that these clips are the “best” in her subjective opinion. Hope may indicate her displeasure by indicating that one or more particular clips do not belong amongst the “best” clips by removing them from the “best” group.
  • In response to her dissatisfaction, the smartphone soon offers another auto-selection of seven minutes of clips. The clips that she was displeased with (e.g., removed from the “best” group) are not part of the new offering of “best” clips. Since this offering seems to include many more of her family members in the clips, she agrees that these clips are the “best.”
  • In response to her agreement and without any additional prompting from her, her contacts in her favorites list (which includes many friends and family members) receive a notification (via email, Short Message Service (SMS), social networking site, etc.) about Hope's new highlight reel for them to see. So, with little effort from Hope, the best seven minutes of over twelve hours of vacation videos are shared with so many of the people who are most important to her. This highlight reel is shared in a timely manner rather than weeks after her vacation and after manually editing her videos.
  • An example of an embodiment of a video-capturing apparatus that employs the new video clip auto-selection techniques described herein may be referred to as an “exemplary video-capturing device.” While one or more example embodiments are described herein, the reader should understand that the claimed invention may be practiced using different details than the exemplary ones described herein.
  • The following co-owned U.S. patent applications are incorporated by reference herein:
      • U.S. Provisional Patent Application Ser. No. 61/435,757, filed on Jan. 24, 2011;
      • U.S. patent application Ser. No. ______, titled “Automatic Selection Of Digital-Video Clips With An Apparatus,” filed on Mar. ___, 2011;
      • U.S. patent application Ser. No. ______, titled “Automatic Selection Of Digital Images From A Multi-Sourced Collection Of Digital Images,” filed on Mar. ___, 2011; and
      • U.S. patent application Ser. No. ______, titled “Automatic Sharing of Superlative Digital Images,” filed on Mar. ___, 2011.
    Exemplary Video-Capturing Device
  • FIG. 1 shows a user 100 (such as Hope) holding a digital camcorder 110 with data-communication capability. The device 110 is also shown in an expanded view. The device 110 may be used to automatically select a small group of digital-video clips from amongst a large video collection. The videos of the large collection may have been shot by the device 110 itself.
  • The device 110 has the capability to communicate (via cellular technology, WLAN, etc.) with a network infrastructure 120. This network infrastructure 120 may be a private or public network of computing or telecommunication resources. It may include the Internet. It may include the so-called “cloud.” As such, the network infrastructure 120 may be called the cloud-computing infrastructure.
  • The video-capturing device 110 includes a video-capturing unit 130, a video-storage unit 140, a highlight-production unit 150, a highlight-presentation unit 160, a user-interaction unit 170, an actions unit 180, and a communications unit 190. Each of these units may be implemented (at least in part) by a set of processor-executable instructions (e.g., software modules). Furthermore, each of these units may include or employ one or more lower-level components of the device 110. For example, these lower-level components include processors, memory, storage, video display, user-input device (e.g., keyboard), transceiver, and the like.
  • The video-capturing unit 130 is configured, designed, and/or programmed to capture digital-video clips. That is, a person using the device 110 may shoot a scene and/or person using the video-capturing unit 130 of the device.
  • The video-capturing unit 130 includes various lower-level camera components for capturing moving digital-video clips with audio (i.e., digital video) and perhaps still digital-video clips (i.e., photographs or clips). The camera components may include (by way of example and not limitation): digital sensor chip (e.g., CCD (charge-coupled device) or CMOS (complementary metal-oxide-semiconductor)), lenses, display, view-finder, and the like. The video-capturing unit 130 may be implemented, at least in part, by a software module resident, at least in part, in the device's memory and executed by one or more processors of the device.
  • In addition, the video-capturing unit 130 may include, or have associated therewith, a video-handling software module (e.g., application) that may be used to enhance the amount of information recorded in the digital-video file relating to the captured digital-video clips. For example, the video-handling application may use information from other sources and/or applications to add data to the captured digital-video clips that are stored on the video-storage unit 140. This added data may be called “metadata.” Specifically, the video-handling application may be configured to obtain information from other hardware, firmware, or software components to add data to the digital-video files. Examples of other components include (but are not limited to) a location application, a calendar application, and/or a contacts application.
  • The file formats for storing digital videos are sometimes referred to container formats from audio and video. Common container formats for storing digital videos include (but are not limited to): 3GPP multimedia file (0.3 gp); anime music video file (.amv); advanced systems format file (.asf); Microsoft ASF redirector file (.asx); audio video interleave file (.avi); DivX-encoded movie (.divx); Nintendo DS movie file (.dpg); Apple QuickTime movie (.moov, .mov, .qt); MPEG-4 video file (.mp4); MPEG video file (.mpg); Real media file (.rm); Macromedia Flash movie (.swf); and Windows media video file (.wmv).
  • As used herein, a “digital video” is one or more digital-video files. A digital video includes one or more digital-video clips. A digital-video file refers to the formatted container of a digital video as it is stored on the video-storage unit 140 or other storage device. A digital-video clip (or simply a clip) refers to a short portion of a digital video that often contains an individual scene, which is a video sequence that is typically shot in one continuous take.
  • Additionally, the video-handling application may determine or identify particular clips in the one or more stored files of the digital video. The video-handling application identifies a particular clip by using existing metadata that specifies the boundaries of the particular clip (i.e., where an individual clip begins and ends). Also, the video-handling application may utilize existing or new video-editing technology to locate a clip based upon the content of the digital video. These video-editing technologies are commonly known as “automatic scene detection” to those of ordinary skill in the art.
  • Furthermore, the video-handling application may be designed to enhance user functionality once clips have been obtained. For example, the video-handling application may also be configured to display clips to the user in cooperation with or as part of the highlight-presentation unit 160. The video-handling application may include various filters or criteria used to limit the number of clips displayed to the user. As discussed below, these filters or criteria may be user selectable, may use the data in the digital-video file obtained from non-video sources or applications, or may be configured based on data in the digital-video files, etc. As another example, similar filters or criteria may also be used to cluster clips into folders (such as virtual albums, system file folders, etc.). As still another example, the video-handling application may use data stored in the digital-video files, contact information, calendar information, and/or upload information to increase the ease of sharing clips.
  • The clips operated on by the video-handling application may include clips captured by the video-capturing unit 130, and/or may include clips obtained from sources other than the video-capturing unit 130. For example, images may be transferred to device 110 using one or more wireless and/or wireline interfaces from the network 120.
  • The video-handling application may be a stand-alone application, or may be integrated into other applications or other units. Moreover, the video-handling application may be formed by a combination of functions of separate, distinct programs or units of the device 110.
  • The video-storage unit 140 is configured, designed, and/or programmed to store digital-video clips as digital-video files and possibly other forms of data and software. That is, the device 110 stores the digital-video files of clips shot by the person using the device 110 on the video-storage unit 140 of the device.
  • The video-storage unit 140 includes one or more lower-level memory or storage components for storing moving digital images with audio (i.e., digital video). The memory or storage components may be volatile or non-volatile, dynamic or static, read/write-able or read only, random- or sequential-access, location- or file-addressable, and the like. The memory or storage components may be magnetic, optical, holographic, and the like. The memory or storage components may be internal to the device 110, attached externally to the device 110, or available via data communications (e.g., on the network or the cloud).
  • The clips stored by the video-storage unit 140 may include clips shot by the video-capturing unit 130, and/or may include clips obtained from sources other than the video-capturing unit 130. For example, clips may be transferred to the device 110 using one or more wireless and/or wireline interfaces from the network 120.
  • The highlight-production unit 150 is configured, designed, and/or programmed to automatically select a group of superlative (e.g., top-ranked or bottom-ranked) clips of a collection of such clips stored by the video-storage unit 140. Alternatively, the highlight-production unit 150 may cluster the collection of clips before the auto-selection is performed. Alternatively still, the highlight-production unit 150 may cluster the auto-selected group of clips after the auto-selection is performed.
  • The highlight-production unit 150 shown in FIG. 1 is implemented as a software module which would reside, at least in part, in the device's memory and be executed by the device's one or more processors. Alternatively, the highlight-production unit 150 may be implemented as a collection of or as part of dedicated hardware or firmware. Alternatively still, the highlight-production unit 150 may be implemented as a combination of hardware, firmware, or software.
  • The highlight-presentation unit 160 is configured, designed, and/or programmed to visually present digital-video clips to the user for her enjoyment and/or consideration. That is, a person using the device 110 may view a selection of clips using the highlight-presentation unit 160 of the device.
  • The highlight-presentation unit 160 includes various lower-level audio/visual presentation components for showing moving digital images with audio (i.e., digital video). The audio/visual components may include (by way of example and not limitation): a liquid crystal display (LCD), a flat panel, organic light-emitting diode (OLED) displays, pico-projection displays, a solid state display or other visual display devices, speakers, and the like. The highlight-presentation unit 160 may be implemented, at least in part, by a software module resident, at least in part, in the device's memory and executed by one or more processors of the device.
  • The user-interaction unit 170 is configured, designed, and/or programmed to attain feedback (e.g., obtain input) from the user and, in particular, feedback related to the clips presented in cooperation with the highlight-presentation unit 160. That is, a person using the device 110 indicates approval of or disapproval with the auto-selected group of clips presented on-screen by using the user-interaction unit 170 of the device.
  • The user-interaction unit 170 includes various lower-level user-input components for receiving input from the user, such as (by way of example and not limitation): keyboard, touchscreen, mouse, touchpad, trackball, and the like. The user-interaction unit 170 may use some of the same audio/visual components of the highlight-presentation unit 160. The user-interaction unit 170 is implemented as a software module which would reside, at least in part, in the device's memory and be executed by the device's one or more processors. Alternatively, the highlight-production unit 150 may be implemented as a collection of or as part of dedicated hardware or firmware. Alternatively still, the user-interaction unit 170 may be implemented as a combination of hardware, firmware, or software.
  • The actions unit 180 is configured, designed, and/or programmed to automatically perform a defined action on the group of top-ranked (or alternatively bottom-ranked) clips of a collection of such clips stored by the video-storage unit 140. Presuming the clips are top-ranked, the defined actions include (by way of example and not limitation): archiving, sharing, burning, conversion/reformat, and the like. Presuming the clips are bottom-ranked, the defined actions include (by way of example and not limitation): deleting, recycling, and the like.
  • Archiving clips involves storing clips in a different and perhaps more reliable location, such as onto the cloud-computing infrastructure. Sharing clips includes sending copies of the clips to another person via one or more various ways of sending such data or notices of the same. Alternatively, sharing clips includes sending a link to one or more of the clips via one or more various ways of sending such links or notices of the same. Examples of such ways to send clips, links, and/or notices thereof include (but are not limited to): email, posting on a social network, posting on a blog or website, text message, MMS (multimedia messaging service), and the like. When the clips are burned, they are packaged and written onto a removeable media (like a DVD disc) so that the clips can be viewed on a television with an attached DVD player. Also, the clips may be automatically converted or reformatted in a pre-defined manner.
  • Deleting clips involves permanently removing one or more clips from the video-storage unit 140. Recycling clips involves placing one or more clips into a queue of files to be deleted later. This queue is sometimes called the “trash” or “recycle bin.”
  • The actions unit 180 shown in FIG. 1 is implemented as a software module which would reside, at least in part, in the device's memory and be executed by the device's one or more processors. Alternatively, the actions unit 180 may be implemented as a collection of or as part of dedicated hardware or firmware. Alternatively still, the actions unit 180 may be implemented as a combination of hardware, firmware, or software.
  • The communications unit 190 is configured, designed, and/or programmed to transmit (and/or receive) notifications of, links to, and/or copies of digital-video clips through the network 120 and onto another user or device accessible from that network. For example, a user of the device 110 may have an auto-selected group of the clips that the person shot sent to their friends upon the person's approval of that auto-selected group.
  • The communications unit 190 includes various lower-level communications components for sending and/or receiving data communications from/by the device. Using transceiver, transmitter, receiver, network interface controller (NIC), and the like, the communications unit 190 utilizes wired (e.g., universal serial bus (USB) or Ethernet) or wireless communications. Examples of wireless communications are (by way of example and not limitation): cellular, satellite, conventional analog AM or FM radio, Wi-Fi™, wireless local area network (WLAN or IEEE 802.11), WiMAX™ (Worldwide Interoperability for Microwave Access), and other analog and digital wireless voice and data transmission technologies. The communications unit 190 may be implemented, at least in part, by a software module resident, at least in part, in the device's memory and executed by one or more processors of the device.
  • Exemplary Processes
  • FIGS. 2-4 are flowcharts illustrating exemplary processes 200, 300, and 400 that implement the techniques described herein for automatic highlight reel production. The exemplary processes 200, 300, and 400 are performed, at least in part, by a video-capturing device, such as device 110 of FIG. 1, a video-capturing device 204 of FIG. 2, and/or a video-capturing telecommunications device 510 of FIG. 5. Additionally and alternatively, the exemplary processes 200, 300, and 400 are performed, at least in part, by a cloud-based service.
  • The process 200 starts, at 210, with an update of a set 212 of digital-video clips (i.e., clips) by, for example, a user 202 shooting new clips with a video-capturing device 204, such as a digital camcorder. Of course, the set 212 of digital-video clips may be updated via other sources such as from a nearby Bluetooth™ device, via cellular connection, from the Internet (e.g., email, social networking site, website, etc.), or some cloud-computing infrastructure resource. The set 212 of digital-video clips may be stored on the device 204, on another data-communicative device, on a network server, on the Internet, in the cloud, or some combination thereof. The box representing operations 210 is vertically long and horizontally narrow so as to indicate that this operation may occur while the other to-be-described operations are performed.
  • The set 212 of digital-video clips may represent all of the clips stored and/or accessible by the device 204. Typically, the set 212 includes a group of clips that are less than all of the clips stored and/or accessible by the device 204. Accordingly, the set 212 is a collection or cluster of some but not all of the stored and/or available clips. Clips may clustered based on one or more criteria, such as location, time, date, and calendar information from one or more events, such as those on private and/or publicly available calendars.
  • Additionally, the clustering criteria may also consider event specifics, such as, faces in the clusters, colors associated with clips, background and lighting of the clips, including overall brightness, color similarities or dissimilarities, and scene information. Clusters of clips may be multilayered. For example, a cluster of wedding clips may have a number of secondary clusters. One secondary cluster may include clips of just the groom and bride, where another secondary cluster may include clips of just the bridesmaids.
  • Next, at operation 214, the device 204 provides enhanced metadata for the digital video. This enhanced-metadata provision includes generating the enhanced metadata and associating it with the digital video (and the clips that make up that associated video). The enhanced metadata is derived from the existing metadata already associated with the digital video (e.g., location and time) and/or an analysis of the content of the video. The content analysis detects and determines occurrence (or lack thereof) of temporal events in the video rather than something based on a single image or frame.
  • To produce enhanced metadata, the content of the audio and/or the moving images portions of the digital video are analyzed. This analysis may look for and note one or more of the following content conditions based upon temporal events (provided by way of illustration only and not limitation): blank segment, silent or quiet audio, lack of sync between audio and video, no audible voices, existence of faces, identification of faces, emotions expressed on faces, motion detection, human motion detected, presence and identification of particular objects, person or object tracking, music detected, language detected, and the like.
  • After that, at 216, the device 204 auto-selects a subset of the set 212 of digital-video clips. The subset of clips forms an effective “highlight reel” of the larger set 212 of digital-video clips. As used herein, the subset presumptively includes some, but not all of, the digital-video clips of the set 212.
  • Typically, the selected subset includes the superlative clips from the set 212 of clips. As used herein with regard to clips, the term “superlative” refers to a subset of clips having the highest kind, quality, or order with respect to the other clips of a set. For example, clips that are top-ranked or bottom-ranked in some category or characteristic are superlative. Also, for example, clips considered the “best” or “worst” of the set are considered superlative.
  • The auto-selection is based upon one or more weighted selection criteria 218 of one or more properties of the clips. FIG. 3 provides more information regarding both the auto-selection operation 216 and the selection criteria 218. The auto-selection operation 216 may be designed to find the “best” clips (i.e., the top-ranked clips). Alternatively, it may be designed to find the “worst” clips (i.e., the bottom-ranked clips).
  • At 220, the device 204 presents the highlight reel to the user 202 via a user interface (of a telecommunication device or a cloud-connected computer device). Via the user-interface (including a user-input device), the user 202 indicates his approval of the highlight reel. By approving, the user 202 may be, for example, agreeing that the highlight reel represents the “best” clips in his subjective opinion. Conversely, by disapproving, the user 202 may be, for example, disagreeing that the highlight reel represents the “best” clips in his subjective opinion.
  • At 222, the device 204 attains feedback from the user regarding the highlight reel. The device 204 determines whether input from the user indicates approval of the highlight reel. If not approved, then the device 204 updates (i.e., adjusts or alters) the weight values assigned to one or more of the selection criteria 218 and the process returns to the auto-selection operation 216 to perform a reweighted selection of the clips based upon the now updated criteria. The device 204 automatically reselects an updated highlight reel. If the user input indicates, at 222, that the user disapproved, the process proceeds to operation 224.
  • In one or more implementations, the user 202 may indicate his disapproval by him choosing to remove or add one or more clips from/to the highlight reel. In the same implementation or others, the device 204 may “learn” the types of images that a user prefers or does not prefer based on iterations of approval and disapproval by the user. When a user removes a clip from the subset, the device 204 reduces the weight values for the removed clip's strong properties and/or increases the weight values for the removed clip's weak properties. And vice versa for the clips that the user manually adds to the subset. Using machine learning techniques, the multiple iterations hone the weight values assigned to various weighted selection criteria.
  • At 224, the device determines whether input from the user (most likely in addition to the user input regarding approval) indicates whether to take a defined action upon the subset of clips and/or which of several possible defined actions to take upon the subset. If the user does not want any defined action, then the process returns to the beginning operation at 210. If the user wants a defined action to be performed upon the subset, then the process proceeds to action operation at 226.
  • In FIG. 2, operation 224 is shown as a dashed box to indicate that it is optional and some implementations may not offer the user this choice. Alternatively, the device 204 may perform one or more defined actions by default. Alternatively still, the user may have pre-set whether and/or which one or more of the defined actions will be performed upon the clips of the highlight reel.
  • At 226, the device 204 performs one or more defined actions on one or more of the clips of the highlight reel. Presuming the clips are top-ranked (i.e., “best”), the defined actions include (by way of example and not limitation): archiving, sharing, media burning, converting/reformatting, and the like. Presuming the clips are bottom-ranked (i.e., “worst”), the defined actions include (by way of example and not limitation): deleting, recycling, and the like.
  • After the one or more actions are performed on the clips of the highlight reel, the process returns to the video-set updating operation 210. The operations 214 through 226 may be performed in real-time with the video-set updating operation 210.
  • FIG. 3 shows the process 300 and offers more details regarding the auto-selection operation 216 and the selection criteria 218. The auto-selection operation 218 includes a binary selection filter operation 302 followed by a weighted-criteria selection filter operation 304. At 302, one of more of the criteria 210 may be applied as a binary filter to, for example, remove clips that do not meet the one or more criteria. After that, at 304, one or more of the criteria are applied in a weighted manner.
  • Generally, the main purpose of the binary selection filter operation 302 is to remove outright “bad” clips. So, it could be called the bad-clip filter. A bad clip is one where the content is difficult for a human to discern, incoherent, would certainly be uninteresting, and/or simple post-processing is not likely to alleviate the problem. The binary filter may use one or more criteria related to one or more properties that might make a clip be considered “bad.” Examples of such “bad clip” criteria include (but are not limited to): shaky, blurry, over or under exposed, poor resolution, poor contrast, out-of-focus, too short, stillness and/or silent for too long, audio-video out of sync, red-eye, and the like. Out-of-sync and red-eye might not be part of the binary selection filter operation 302 in some implementations because they can be fixed with some post-processing.
  • Some implementations may not employ the binary selection filter operation 302. Those implementations may rely upon the weighted-criteria selection filter operation 304 to remove the “bad” clips.
  • Alternatively, the binary selection filter operation 302 may be described as grouping the set 212 of digital-video clips into an intermediate grouping that includes two or more of the digital-video clips of the set 212. Similarly, the weighted-criteria selection filter operation 304 may be alternatively described as ranking the digital-video clips of the intermediate grouping based upon one or more of the multiple weighted criteria 218 and designating an allotment of top-ranked (or bottom-ranked) digital-video clips of the intermediate grouping as the selected subset of digital-video clips that results from the auto-selection operation 216.
  • The size (i.e., number, cardinality) of the allotment may be a fixed number (e.g., ten) or calculated. That allotment size may refer to number of clips, number of frames, length of highlight reel (e.g., ten minutes), or the like. If fixed, the allotment size may be a factory-set default number or a user-defined number. If calculated, the calculation may be, for example, based upon linear or non-linear proportions of the quantity or length of clips in the set 212, in the intermediate grouping, and/or the like. These proportions may be factory-set or user-defined. A percentage is an example of such proportions.
  • For example, the allotment size for the subset may be user-defined to be five percent of the number or length of clips in the set 212. If, for this example, the set 212 included 100 clips totaling three hours of video, then the allotment size might be five clips or perhaps nine minutes. Consequently, the user 202 would be presented with fives clips or perhaps nine minutes of clips for his approval.
  • Typically, the criteria used for the binary filtering differ from those used for the weighted filtering. However, in some implementations, the criteria used for one may overlap the criteria used for the other. The criteria are related to properties of a clip. Those properties may be derived from the content of the clip (based upon an image or video analysis) and/or from metadata associated with the clip. Examples of clip metadata include (but are not limited to): technical metadata (such as encoding details, identification of clip boundaries, size, resolution, color profile, ISO speed and other camera settings), descriptive metadata (captions, headlines, titles, keywords, location of capture, etc.), and administrative metadata (such as licensing or rights usage terms, specific restrictions on using an image, model releases, provenance information, such as the identity of the creator, and contact information for the rights holder or licensor).
  • The weighted selection criteria 218 include (by way of example and not limitation): presence or absence of audio from a corresponding clip, motion, smiling faces, laughter, black screens, identification of video-capture device, identification of video-capture device user, location and/or time of video-capture, focus, contrast, shake, red eye, person in clip is a favorite (e.g., in a social network or on the device), person in clip is a contact, clip is tagged (in general or with a particular tag), user-added criteria, clip was auto-corrected, flash used, social network ranking of a clip, etc.
  • FIG. 4 illustrates the process 400 for determining a best clip that may be used with embodiments described herein. This process 400 may be employed as part of process 200 and/or process 300 herein.
  • The process 400 starts with operation 402 where a video-capture device retrieves the next clip (or the first when initially executed) from a memory like that of video-storage unit 140 of device 110 (as shown in FIG. 1). Next, at 404 and 406, a determination is made if the retrieved clip is shaky or if redeye is detected. If so, a next clip is retrieved and the process 400 starts again with that next clip. If the retrieved clip is the last (as determined at operation 438), then the process 400 proceeds to a clip-ranking operation 440, which is discussed later.
  • If no shake or redeye is detected in a clip, then the process 400 continues onto operation 408 where it is determined if the resolution of the retrieved photo is greater than a defined threshold resolution. If not, then a next photo is retrieved and the process begins again. If so, based on the resolution of the image, a resolution index is determined or calculated at operation 410. In one example, the resolution index may be determined or calculated based on the resolution (megapixels) associated with a photo undergoing consideration. Next, at 412, the process determines if the histogram associated with the photo is within a defined threshold. If not, then a next photo is retrieved and the process begins again. If so, based on the histogram of the photo, a histogram index is determined or calculated at 414. In one example, the histogram index is determined or calculated based on a consideration of the number of “dark” pixels (underexposure), “bright” pixels (overexposure), and/or the number of pixels that fall between the underexposure and overexposure range.
  • Next, the process continues, at 416, to determine if a duplicate clip of the one currently being processed is stored in the device or otherwise known or available. If so, a next clip is retrieved and the process begins again. If not, the process determines, at 418, if one or more faces are detected in the clip. If no faces are detected in the clip, the process calculates, at 420, a NF (“no face”) clip score based on the calculated resolution index and histogram index, and associates the NF score with the clip and adds the clip to a “no face” clip queue at 422. If one or more faces is detected, the process determines, at 424, if a subject in the clip has their eyes shut or partially shut (i.e., “eyes closed detection”). If that is detected, a next clip is retrieved and the process begins again.
  • If no blink is detected, the process calculates face, event, alignment and user like indexes at operations 426, 428, 430, 432, respectively. In one implementation, the face index calculation considers if one or more expected or desired faces are in a photo, the event index calculation considers if an expected event (e.g., wedding or party) is part of a photo, the alignment calculation considers a degree of alignment (e.g., vertical or horizontal) of a photo and the user like index considers a relative user acceptance or approval (e.g., like or dislike) of one or more faces in a photo. These indexes and the previously calculated indexes associated with resolution and histogram are considered when calculating an F (“face”) photo score at 434. The face image score is associated with the photo and the photo is added to a face queue at 436.
  • Each time a clip is added to a queue, the queue is automatically resorted based on scores associated with an already processed and queued clip. Once the last clip is processed (as determined by operation 438), the process picks the highest scored clip as the best clip at 440. In addition, operation 440 may also select, based on clip score, a set of clips for presentation to a user of the portable device.
  • In alternative implementations, other criteria (like those of 218 in FIGS. 2 and 3) and/or different permutations of criteria may be employed at operations 404, 406, 408, 416, 418, and 424. For example, with operations 404, 406, and 408 other “bad clip” type criteria may be used and with operations 416, 418, and 424 other weighed selection criteria may be employed.
  • Exemplary Device within Exemplary Telecommunication Environment
  • FIG. 5 shows an exemplary telecommunication environment 500 which includes a video-capturing telecommunications device 510. Both the environment 500 and the device 510 may be included as part of one or more implementation of the techniques described herein.
  • As depicted, the device 510 is a smartphone that is capable of communicating using analog and/or digital wireless voice and/or data transmission. FIG. 5 shows the device 510 in communication with a cell phone 520 via a so-called cellular network 522 as represented by cell tower. Via that cellular network 522 or via some other wireless or wired communication technology (e.g., Wi-Fi™, WiMAX™, WLAN, Ethernet, etc.), the device 510 is coupled to a cloud-computing infrastructure 524.
  • The cloud-computing infrastructure 524 may include scalable and virtualized compute resources over a private or public network. Typically, a cloud-computing infrastructure uses the internet and central remote servers to maintain data and applications. The cloud-computing infrastructure 524 includes a group of individual computers connected by high-speed communications and some form of management software guiding the collective actions and usage of the individual computers. This technology allows for much more efficient computing by centralizing storage, memory, processing and bandwidth. While depicted in FIG. 5 as the cloud-computing infrastructure, item 524 may alternatively be viewed as a network or the Internet.
  • The device 510 includes many components, such as a wireless transceiver 530, a NIC (network interface controller) 532, one or more processors 534, one or more cameras 536, one or more input subsystems 538, one or more output subsystems 540, a secondary storage 542 with at least one processor-readable medium 544, one or more location subsystems 546, one or more power systems 548, one or more audio systems 550, one or more video displays 552, and a memory 554 having software 556 and data (e.g., digital photo or digital video) 558 stored thereon.
  • The NIC 532 is hardware component that connects the device 510 to one or more computer networks and allow for communication to and from that network. Typically, the NIC 532 operates as both an Open Systems Interconnection (OSI) layer 1 (i.e., physical layer) and layer 2 (i.e., data link layer) device, as it provides physical access to a networking medium and provides a low-level addressing system through the use of Media Access Control (MAC) addresses. It allows devices to connect to each other either by using cables or wirelessly.
  • The one or more processors 534 may include a single or multi-core central processing unit (CPU), a graphics processing unit (GPU), other processing unit or component, or some combination thereof.
  • The camera 536 may be configurable to capture still images (clips), moving images (video), or both still and moving images. The camera components may include (by way of example and not limitation): digital sensor chip (e.g., CCD (charge-coupled device) or CMOS (complementary metal-oxide-semiconductor)), lenses, flash, and the like.
  • The input subsystems 538 are physical devices designed to allow for input from, for example, a human user. The input subsystems 538 may include (by way of example and not limitation): keyboard, keypad, touchscreen, touchpad, mouse, trackball, paddle, light pen, scanner, stylus, and/or a micro-telecommunications device.
  • The output subsystems 540 are mechanisms for providing a tactile output to the user in a manner that is neither audio nor video based. Such tactile output is typically generated by an offset rotating motor that vibrates the device 510 to give a notification to the user.
  • The secondary storage 542 is typically a read-only memory (ROM), a programmable memory (such as EEPROM, EPROM, or PROM), a static read/writable memory (such as flash memory), and a mechanical read/writeable device (such as hard drive), which may be magnetic, optical, holographic, and the like. The components of the secondary storage 542 may be internal to the device 510, attached externally to the device 510, or available via data communications (e.g., on the cloud).
  • At least one processor-readable medium 544 is stored on the secondary storage 542. The processor-readable medium 544 stores one or more sets of instructions (e.g., software) embodying any one or more of the methodologies or functions described herein. The instructions may also reside, completely or at least partially, within the memory 554 as software 556 and within the processor(s) 534 during execution thereof by the device 510.
  • The location subsystems 546 may be designed to determine the physical location of the device 510. The one or more location subsystems 546 may include (but is not limited to) an application and hardware for determining location via global positioning system (GPS) and/or by communicating with a cellular network.
  • The power subsystems 548 may include (but is not limited to) one or more batteries and one or more external power supplies, such as an interface to an external power supply source.
  • The audio systems 550 are configured to generate sound, noise, voices, and music. In addition, the audio systems 550 are configured to capture and process sound, noise, voices, and music. The audio systems 550 includes (but is not limited to) speakers and microphones.
  • The video display 552 is configured to display images, videos, text, user-interfaces, and the like. The video display 552 may include (but is not limited to) a liquid crystal display (LCD), a flat panel, a solid state display or other device. The video display 552 may operate as a digital viewfinder that allows a user to preview a scene before capturing an image and/or to view a movie as it is being captured.
  • The memory 554 is configured to store software 556 and data (e.g., digital photo or digital video) 558 thereon. The memory 554 is a working space and may include (but is not limited to) random access memory (RAM). The memory 554 may be one or more of the following (by way of example and not limitation): volatile or non-volatile, dynamic or static, read/write-able or read only, random- or sequential-access, location- or file-addressable, and the like. The components of the memory 554 may be internal to the device 510, attached externally to the device 510, or available via data communications (e.g., on the cloud).
  • Additional and Alternative Implementation Notes
  • The exemplary devices and apparatuses discussed herein that implement at least some aspect of the techniques discussed herein include (by way of illustration and not limitation): the video-capturing device 110 from FIG. 1, the video-capturing device 204 of FIG. 2, and the video-capturing telecommunications device 510 of FIG. 5.
  • Such exemplary devices may be what is commonly referred to as a “mobile phone,” “smartphone,” and/or “cellphone.” However, the described techniques can be used in conjunction with non-cellular technologies such as conventional analog AM or FM radio, Wi-Fi™, wireless local area network (WLAN or IEEE 802.11), WiMAX™ (Worldwide Interoperability for Microwave Access), Bluetooth™, and other analog and digital wireless voice and data transmission technologies. Alternative implementations of such devices might not have any telecommunications or wireless communication capability. Still other alternative implementations may utilize a wired communication link instead of or in addition to wireless communication.
  • The exemplary devices are not limited to those having a camera. Any device with the capability to acquire, collect, and/or manipulate digital images or moving digital images (i.e., video) may implement some aspect of the techniques discussed herein. Examples of such alternative devices include (but are not limited to) the following: tablet-based computer, other handheld computer, netbook, computer, digital camera, digital camcorder, handheld multimedia device, digital single-lens reflex (DSLR) camera, GPS navigational system, vehicle-based computer, or other portable electronics.
  • The exemplary devices may operate on digital videos acquired from sources other than their own video-capturing components. These other video sources may be other video-capturing devices or from non-camera-equipped sources. The non-camera-equipped sources include other devices or services where images and/or videos are stored, collected, accessed, handled, manipulated, and/or viewed. Examples of such alternative image/video sources include (but are not limited to): a clip-processing kiosk, portable and removable storage media or device (e.g., CD-ROM, DVD-ROM, other optical media, USB flash drive, flash memory card, external hard drive, etc.), electronics with radio-frequency identification (RFID), a social-networking services (e.g., Facebook™, MySpace™, Twitter™, Linkedin™, Ning™, and the like), and photo/video sharing services (e.g., Flickr™, Photobucket™, Picasa™, Shutterfly™, and the like).
  • One or more implementations of the techniques described herein may include an initial training session for the auto-selecting device or service to learn what makes a photo and/or video clip superlative (e.g., best or worst) in the user's opinion. This training may be repeated from time to time. This training will set the weight values of one or more of the weighted criteria. There may be various profiles for specific conditions and situations where the weighted criteria have their own trained values. Each user of an auto-selecting device or service may have their own customized weighted criteria values and profiles of such values. An auto-selecting device or service may have a default set of values or profiles assigned. That default set or profiles may be extrapolated from a statistically evaluation (e.g., mean or median) of the trained values derived from multiple users.
  • Instead of a training session, one or more implementations may take advantage of a system of user-ratings (e.g., thumbs up/down or 1 to 5 stars) for photos and/or video clips. These user-rating are effective data for training the exemplary video-capture device to learn what makes a photo and/or video clip superlative (e.g., best or worst) in the user's opinion. In some implementations, the user may define values for one or more of the weighted criteria. This may be accomplished using, for example, slider-bar user interface for the user to adjust weights assigned to particular criteria or to categories of criteria.
  • In the above description of exemplary implementations, for purposes of explanation, specific numbers, materials configurations, and other details are set forth in order to better explain the invention, as claimed. However, it will be apparent to one skilled in the art that the claimed invention may be practiced using different details than the exemplary ones described herein. In other instances, well-known features are omitted or simplified to clarify the description of the exemplary implementations.
  • The inventors intend the described exemplary implementations to be primarily examples. The inventors do not intend these exemplary implementations to limit the scope of the appended claims. Rather, the inventors have contemplated that the claimed invention might also be embodied and implemented in other ways, in conjunction with other present or future technologies.
  • Moreover, the word “exemplary” is used herein to mean serving as an example, instance, or illustration. Any aspect or design described herein as “exemplary” is not necessarily to be construed as preferred or advantageous over other aspects or designs. Rather, use of the word exemplary is intended to present concepts and techniques in a concrete fashion. The term “techniques,” for instance, may refer to one or more devices, apparatuses, systems, methods, articles of manufacture, and/or computer-readable instructions as indicated by the context described herein.
  • As used in this application, the term “or” is intended to mean an inclusive “or” rather than an exclusive “or.” That is, unless specified otherwise or clear from context, “X employs A or B” is intended to mean any of the natural inclusive permutations. That is, if X employs A; X employs B; or X employs both A and B, then “X employs A or B” is satisfied under any of the foregoing instances. In addition, the articles “a” and “an” as used in this application and the appended claims should generally be construed to mean “one or more,” unless specified otherwise or clear from context to be directed to a singular form.
  • The exemplary processes discussed herein are illustrated as a collection of blocks in a logical flow graph, which represents a sequence of operations that can be implemented with hardware, software, firmware, or some combination thereof. In the context of software/firmware, the blocks represent instructions stored on one or more processor-readable storage media that, when executed by one or more processors, perform the recited operations. The operations of the exemplary processes may be rendered in virtually any programming language or environment including (by way of example and not limitation): C/C++, Fortran, COBOL, PASCAL, assembly language, markup languages (e.g., HTML, SGML, XML, VoXML), and the like, as well as object-oriented environments such as the Common Object Request Broker Architecture (CORBA), Java™ (including J2ME, Java Beans, etc.), Binary Runtime Environment (BREW), and the like.
  • Note that the order in which the processes are described is not intended to be construed as a limitation, and any number of the described process blocks can be combined in any order to implement the processes or an alternate process. Additionally, individual blocks may be deleted from the processes without departing from the spirit and scope of the subject matter described herein.
  • The term “processor-readable media” includes processor-storage media. For example, processor-storage media may include, but are not limited to, magnetic storage devices (e.g., hard disk, floppy disk, and magnetic strips), optical disks (e.g., compact disk (CD) and digital versatile disk (DVD)), smart cards, flash memory devices (e.g., thumb drive, stick, key drive, and SD cards), and volatile and non-volatile memory (e.g., random access memory (RAM), read-only memory (ROM)).
  • For the purposes of this disclosure and the claims that follow, the terms “coupled” and “connected” may have been used to describe how various elements interface. Such described interfacing of various elements may be either direct or indirect.

Claims (20)

1. A method comprising:
obtaining a digital video;
producing a highlight reel of clips from the digital video, the producing comprising:
analyzing metadata and content of the clips of the digital video;
selecting multiple clips from the digital video based upon the analyzing and weighted selection criteria;
presenting, via a user-interface of a computing apparatus, the highlight reel to a user for feedback;
attaining feedback from the user regarding user-acceptability of the produced highlight reel; and
in response to attained user-feedback indicating a lack of user-acceptability of the produced highlight reel, producing an updated highlight reel using a reweighted selection of clips from the digital video.
2. A method as recited in claim 1, further comprising providing enhanced metadata to the digital video, the providing comprising:
generating enhanced metadata based upon, at least in part, existing metadata associated with the digital video;
associating the generated enhanced metadata with the digital video.
3. A method as recited in claim 1, wherein the selecting includes:
segregating the digital video into an intermediate grouping of clips based upon whether the clips match one or more specified criteria;
ranking the clips of the segregated intermediate grouping based upon multiple weighted criteria;
designating an allotment of similarly ranked clips of the segregated intermediate grouping as the produced highlight reel.
4. A method as recited in claim 3, wherein the designated allotment includes clips that are top ranked amongst the segregated intermediate grouping.
5. A method as recited in claim 3, wherein the designated allotment includes clips that are bottom ranked amongst the segregated intermediate grouping.
6. A method as recited in claim 3, further comprising calculating a cardinality of the designated allotment based upon a percentage of the segregated intermediate grouping.
7. A method as recited in claim 1, wherein the production of the updated highlight reel includes:
adjusting one or more weight values of the weighted criteria;
assembling the updated highlight reel by reselecting a collection of clips from the digital video based upon the adjusted weighted criteria.
8. A method as recited in claim 1, wherein the weighted criteria regarding a particular clip are selected from a group of clip properties consisting of identification of the video-capture apparatus, clip-acquisition location, presence of audio, absence of audio, motion, smiling faces, laughter, black frames, location, focus, contrast, shake, red eye, person in clip is a favorite, person in clip is a contact, clip is tagged, user-added criteria, clip was auto-corrected, and social network ranking of the clip.
9. A method as recited in claim 1, further comprising clustering the digital video into multiple clusters of clips, wherein the produced highlight reel is selected from one or more of the multiple clusters.
10. A method as recited in claim 1, further comprising clustering the highlight reel into multiple clusters.
11. One or more processor-readable storage devices having processor-executable instructions embodied thereon, the processor-executable instructions, when executed by one or more processors, direct the one or more processors to perform operations comprising:
obtaining a digital video;
producing a highlight reel of clips from the digital video, the producing comprising:
analyzing metadata and content of the clips of the digital video;
selecting multiple clips from the digital video based upon the analyzing and weighted selection criteria;
presenting, via a user-interface of a computing apparatus, the highlight reel.
12. One or more processor-readable storage devices as recited in claim 11, further comprising:
providing enhanced metadata to the digital video, the providing comprising:
generating enhanced metadata based upon, at least in part, content of one or more clips of the digital video;
associating the generated enhanced metadata with the digital video,
wherein the analyzing includes analyzing the enhanced metadata.
13. One or more processor-readable storage devices as recited in claim 11, further comprising:
attaining feedback regarding user-acceptability of the produced highlight reel;
in response to the attained user-feedback indicating a lack of user-acceptability of the produced highlight reel, producing an updated highlight reel using a reweighted selection of clips from the digital video.
14. One or more processor-readable storage devices as recited in claim 11, wherein the selecting includes:
segregating the digital video into an intermediate grouping of clips based upon whether the clips match one or more specified criteria;
ranking the clips of the segregated intermediate grouping based upon multiple weighted criteria;
designating an allotment of similarly top or bottom ranked clips of the segregated intermediate grouping as the produced highlight reel.
15. One or more processor-readable storage devices as recited in claim 11, wherein the production of the updated highlight reel includes:
adjusting one or more weight values of the weighted criteria;
assembling the updated highlight reel by reselecting a collection of clips from the digital video based upon the adjusted weighted criteria.
16. One or more processor-readable storage devices as recited in claim 11, wherein the weighted criteria regarding a particular clip are selected from a group of clip properties consisting of identification of the video-capture apparatus, clip-acquisition location, presence of audio, absence of audio, motion, smiling faces, laughter, black frames, location, focus, contrast, shake, red eye, person in clip is a favorite, person in clip is a contact, clip is tagged, user-added criteria, clip was auto-corrected, and social network ranking of the clip.
17. One or more processor-readable storage devices as recited in claim 11, wherein the obtaining comprises retrieving the digital video via a telecommunication network.
18. An apparatus comprising:
a video-storage unit configured to store digital videos;
a highlight-production unit configured to produce a highlight reel of clips from one or more of the stored digital videos, the highlight-production unit being further configured to:
analyze metadata and content of the clips of the digital video;
select multiple clips from the digital video based upon the analyzing and weighted selection criteria;
a highlight-presentation unit configured to present the produced highlight reel.
19. An apparatus as recited in claim 18, further comprising a user-interaction unit configured to attain feedback regarding user-acceptability of the produced highlight reel presented to a user via the highlight-presentation unit.
20. An apparatus as recited in claim 18, wherein the highlight-production unit is further configured to produce, in response to attained user-feedback indicating a lack of user-acceptability of the produced highlight reel, an updated highlight reel using a reweighted selection of clips from the digital video.
US13/041,370 2011-01-24 2011-03-05 Automatic highlight reel producer Abandoned US20120189284A1 (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
US13/041,370 US20120189284A1 (en) 2011-01-24 2011-03-05 Automatic highlight reel producer

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
US201161435757P 2011-01-24 2011-01-24
US13/041,370 US20120189284A1 (en) 2011-01-24 2011-03-05 Automatic highlight reel producer

Publications (1)

Publication Number Publication Date
US20120189284A1 true US20120189284A1 (en) 2012-07-26

Family

ID=46543901

Family Applications (4)

Application Number Title Priority Date Filing Date
US13/041,361 Active 2031-11-24 US9268792B2 (en) 2011-01-24 2011-03-05 Automatic selection of digital images with an apparatus
US13/041,370 Abandoned US20120189284A1 (en) 2011-01-24 2011-03-05 Automatic highlight reel producer
US13/041,369 Active 2032-09-18 US9195678B2 (en) 2011-01-24 2011-03-05 Automatic selection of digital images from a multi-sourced collection of digital images
US13/041,372 Active 2031-06-26 US9223802B2 (en) 2011-01-24 2011-03-05 Automatic sharing of superlative digital images

Family Applications Before (1)

Application Number Title Priority Date Filing Date
US13/041,361 Active 2031-11-24 US9268792B2 (en) 2011-01-24 2011-03-05 Automatic selection of digital images with an apparatus

Family Applications After (2)

Application Number Title Priority Date Filing Date
US13/041,369 Active 2032-09-18 US9195678B2 (en) 2011-01-24 2011-03-05 Automatic selection of digital images from a multi-sourced collection of digital images
US13/041,372 Active 2031-06-26 US9223802B2 (en) 2011-01-24 2011-03-05 Automatic sharing of superlative digital images

Country Status (1)

Country Link
US (4) US9268792B2 (en)

Cited By (26)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20130051756A1 (en) * 2011-08-26 2013-02-28 Cyberlink Corp. Systems and Methods of Detecting Significant Faces in Video Streams
US20130160051A1 (en) * 2011-12-15 2013-06-20 Microsoft Corporation Dynamic Personalized Program Content
US20130282747A1 (en) * 2012-04-23 2013-10-24 Sri International Classification, search, and retrieval of complex video events
US20130336639A1 (en) * 2011-12-27 2013-12-19 William C. DeLeeuw Method and system to play linear video in variable time frames
US20140325568A1 (en) * 2013-04-26 2014-10-30 Microsoft Corporation Dynamic creation of highlight reel tv show
US20150026578A1 (en) * 2013-07-22 2015-01-22 Sightera Technologies Ltd. Method and system for integrating user generated media items with externally generated media items
US20150078726A1 (en) * 2013-09-17 2015-03-19 Babak Robert Shakib Sharing Highlight Reels
US20150147045A1 (en) * 2013-11-26 2015-05-28 Sony Corporation Computer ecosystem with automatically curated video montage
US20150160836A1 (en) * 2013-12-04 2015-06-11 Autodesk, Inc. Extracting demonstrations from in-situ video content
US9110929B2 (en) 2012-08-31 2015-08-18 Facebook, Inc. Sharing television and video programming through social networking
US9195678B2 (en) 2011-01-24 2015-11-24 T-Mobile Usa, Inc. Automatic selection of digital images from a multi-sourced collection of digital images
US9301016B2 (en) 2012-04-05 2016-03-29 Facebook, Inc. Sharing television and video programming through social networking
WO2016200692A1 (en) * 2015-06-11 2016-12-15 Vieu Labs, Inc. Editing, sharing, and viewing video
US20160365120A1 (en) * 2015-06-11 2016-12-15 Eran Steinberg Video editing system for generating multiple final cut clips
US20170092324A1 (en) * 2015-09-30 2017-03-30 Apple Inc. Automatic Video Compositing
US20170092331A1 (en) 2015-09-30 2017-03-30 Apple Inc. Synchronizing Audio and Video Components of an Automatically Generated Audio/Video Presentation
US20180063253A1 (en) * 2015-03-09 2018-03-01 Telefonaktiebolaget Lm Ericsson (Publ) Method, system and device for providing live data streams to content-rendering devices
US10269387B2 (en) 2015-09-30 2019-04-23 Apple Inc. Audio authoring and compositing
US20190182565A1 (en) * 2017-12-13 2019-06-13 Playable Pty Ltd System and Method for Algorithmic Editing of Video Content
US20190214054A1 (en) * 2016-02-16 2019-07-11 Shimmeo, Inc System and Method for Automated Video Editing
US20200053203A1 (en) * 2018-08-13 2020-02-13 Avaya Inc. Selective communication event extraction
US10726594B2 (en) 2015-09-30 2020-07-28 Apple Inc. Grouping media content for automatically generating a media presentation
US10971192B2 (en) 2018-11-07 2021-04-06 Genetec Inc. Methods and systems for detection of anomalous motion in a video stream and for creating a video summary
US11019252B2 (en) 2014-05-21 2021-05-25 Google Technology Holdings LLC Enhanced image capture
US11126317B2 (en) 2015-02-24 2021-09-21 Axon Enterprise, Inc. Systems and methods for bulk redaction of recorded data
US11315600B2 (en) * 2017-11-06 2022-04-26 International Business Machines Corporation Dynamic generation of videos based on emotion and sentiment recognition

Families Citing this family (65)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US10298781B2 (en) * 2009-03-18 2019-05-21 Shutterfly, Inc. Cloud storage for image data, image product designs, and image projects
US9485365B2 (en) * 2009-03-18 2016-11-01 Shutterfly, Inc. Cloud storage for image data, image product designs, and image projects
US8762826B2 (en) * 2010-10-18 2014-06-24 Blayne White Organization and design system and online community for images
KR101740059B1 (en) * 2011-04-01 2017-05-26 한국전자통신연구원 Method and apparatus for providing time machine service based on social network service
US9125022B2 (en) * 2011-12-02 2015-09-01 Microsoft Technology Licensing, Llc Inferring positions with content item matching
WO2013093186A2 (en) * 2011-12-22 2013-06-27 Nokia Corporation Method and apparatus for sharing media upon request via social networks
US9147131B2 (en) 2012-07-30 2015-09-29 Evernote Corporation Extracting multiple facial photos from a video clip
CN104583901B (en) * 2012-08-20 2019-10-01 爱维士软件私人有限公司 The packing of multi-medium data based on set and display methods and system
US20140058980A1 (en) * 2012-08-22 2014-02-27 Avant Garde Events, Inc. Method for facilitating business networking
JP6300295B2 (en) 2012-08-24 2018-03-28 三星電子株式会社Samsung Electronics Co.,Ltd. Friend recommendation method, server therefor, and terminal
CN104685861B (en) * 2012-10-05 2018-08-21 索尼公司 Content acquisition unit, portable unit, server, information processing unit and storage medium
US20140108405A1 (en) * 2012-10-16 2014-04-17 Realnetworks, Inc. User-specified image grouping systems and methods
KR102067642B1 (en) * 2012-12-17 2020-01-17 삼성전자주식회사 Apparataus and method for providing videotelephony in a portable terminal
US8925060B2 (en) * 2013-01-02 2014-12-30 International Business Machines Corporation Selecting image or video files for cloud storage
US20140280561A1 (en) * 2013-03-15 2014-09-18 Fujifilm North America Corporation System and method of distributed event based digital image collection, organization and sharing
US9232137B2 (en) * 2013-04-24 2016-01-05 Microsoft Technology Licensing, Llc Motion blur avoidance
US9398057B2 (en) * 2013-06-04 2016-07-19 Dropbox, Inc. System and method for group participation in a digital media presentation
WO2015017865A1 (en) 2013-08-02 2015-02-05 Shoto, Inc. Discovery and sharing of photos between devices
US9275066B2 (en) 2013-08-20 2016-03-01 International Business Machines Corporation Media file replacement
US10185841B2 (en) 2013-10-10 2019-01-22 Elwha Llc Devices, methods, and systems for managing representations of entities through use of privacy beacons
US10834290B2 (en) 2013-10-10 2020-11-10 Elwha Llc Methods, systems, and devices for delivering image data from captured images to devices
US10346624B2 (en) * 2013-10-10 2019-07-09 Elwha Llc Methods, systems, and devices for obscuring entities depicted in captured images
US20150106628A1 (en) * 2013-10-10 2015-04-16 Elwha Llc Devices, methods, and systems for analyzing captured image data and privacy data
JP6389249B2 (en) 2013-10-14 2018-09-12 ノキア テクノロジーズ オサケユイチア Method and apparatus for identifying media files based on contextual relationships
US9407728B2 (en) 2013-11-08 2016-08-02 Dropbox, Inc. Content item presentation system
WO2015103498A1 (en) * 2014-01-03 2015-07-09 Snapcious LLC Method and system for distributed collection and distribution of photographs
US10372747B1 (en) * 2014-02-25 2019-08-06 Google Llc Defining content presentation interfaces based on identified similarities between received and stored media content items
US9602679B2 (en) * 2014-02-27 2017-03-21 Lifeprint Llc Distributed printing social network
US20150286897A1 (en) * 2014-04-03 2015-10-08 Microsoft Corporation Automated techniques for photo upload and selection
US9614724B2 (en) 2014-04-21 2017-04-04 Microsoft Technology Licensing, Llc Session-based device configuration
US9639742B2 (en) 2014-04-28 2017-05-02 Microsoft Technology Licensing, Llc Creation of representative content based on facial analysis
US9773156B2 (en) 2014-04-29 2017-09-26 Microsoft Technology Licensing, Llc Grouping and ranking images based on facial recognition data
US9430667B2 (en) 2014-05-12 2016-08-30 Microsoft Technology Licensing, Llc Managed wireless distribution network
US10111099B2 (en) 2014-05-12 2018-10-23 Microsoft Technology Licensing, Llc Distributing content in managed wireless distribution networks
US9384335B2 (en) 2014-05-12 2016-07-05 Microsoft Technology Licensing, Llc Content delivery prioritization in managed wireless distribution networks
US9384334B2 (en) 2014-05-12 2016-07-05 Microsoft Technology Licensing, Llc Content discovery in managed wireless distribution networks
US9874914B2 (en) 2014-05-19 2018-01-23 Microsoft Technology Licensing, Llc Power management contracts for accessory devices
JP2015225445A (en) * 2014-05-27 2015-12-14 ソニー株式会社 Information processing apparatus and information processing method
US10037202B2 (en) 2014-06-03 2018-07-31 Microsoft Technology Licensing, Llc Techniques to isolating a portion of an online computing service
US9367490B2 (en) 2014-06-13 2016-06-14 Microsoft Technology Licensing, Llc Reversible connector for accessory devices
US9460493B2 (en) 2014-06-14 2016-10-04 Microsoft Technology Licensing, Llc Automatic video quality enhancement with temporal smoothing and user override
US9373179B2 (en) 2014-06-23 2016-06-21 Microsoft Technology Licensing, Llc Saliency-preserving distinctive low-footprint photograph aging effect
US10074003B2 (en) * 2014-07-11 2018-09-11 Intel Corporation Dynamic control for data capture
US10108626B2 (en) * 2014-09-03 2018-10-23 Verizon Patent And Licensing Inc. Proactively clearing digital storage
US9558553B2 (en) 2014-11-17 2017-01-31 Ricoh Company, Ltd. Image acquisition and management using a reference image
CN107408212B (en) * 2015-02-11 2021-09-21 爱维士软件有限责任公司 System and method for identifying unwanted photos stored on a device
US9594776B2 (en) 2015-05-05 2017-03-14 Microsoft Technology Licensing, Llc Dynamic, parameterized image resource selection
KR102336448B1 (en) * 2015-06-10 2021-12-07 삼성전자주식회사 Electronic apparatus and method for photograph extraction
JP6083546B1 (en) * 2015-11-26 2017-02-22 株式会社パブセン Information management apparatus, information management method and program
CN105631404B (en) * 2015-12-17 2018-11-30 小米科技有限责任公司 The method and device that photo is clustered
US10229324B2 (en) 2015-12-24 2019-03-12 Intel Corporation Video summarization using semantic information
US10664500B2 (en) 2015-12-29 2020-05-26 Futurewei Technologies, Inc. System and method for user-behavior based content recommendations
US20170192625A1 (en) * 2015-12-31 2017-07-06 Young Kim Data managing and providing method and system for the same
US20170329796A1 (en) * 2016-05-13 2017-11-16 Motorola Mobility Llc Systems and methods for context-based file sharing
US10305977B2 (en) 2016-08-05 2019-05-28 International Business Machines Corporation Social network image filtering
US10425498B2 (en) * 2017-01-25 2019-09-24 International Business Machines Corporation Targeted profile picture selection
US10572121B2 (en) * 2018-02-09 2020-02-25 International Business Machines Corporation Contextual user profile photo selection
WO2019229519A1 (en) * 2018-06-01 2019-12-05 Patung Investments Ltd. Capture and presentation of authenticated data
WO2020026238A1 (en) * 2018-07-31 2020-02-06 Pic2Go Ltd. Batch-based, multi-photograph sharing methods and system
US20200058079A1 (en) * 2018-08-14 2020-02-20 Mastercard International Incorporated Systems and methods for using machine learning and simulations for intelligent budgeting
CN110022451B (en) * 2019-04-18 2020-07-28 环爱网络科技(上海)有限公司 Method and system for generating sub-video and medium storing corresponding program
KR102114223B1 (en) * 2019-12-10 2020-05-22 셀렉트스타 주식회사 Method for filtering a similar image based on deep learning and apparatus using the same
US20240121293A1 (en) * 2020-02-12 2024-04-11 Rajeshkumar K. Mehta Methods and systems for implementing a digital photograph sharing mobile application
US11863860B2 (en) 2020-09-30 2024-01-02 Snap Inc. Image capture eyewear with context-based sending
US12051150B2 (en) 2021-12-21 2024-07-30 Samsung Eletronica Da Amazonia Ltda. Computer implemented method and system for classifying an input image for new view synthesis in a 3D visual effect, and non-transitory computer readable storage medium

Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20030187844A1 (en) * 2002-02-11 2003-10-02 Mingjing Li Statistical bigram correlation model for image retrieval
US20040025180A1 (en) * 2001-04-06 2004-02-05 Lee Begeja Method and apparatus for interactively retrieving content related to previous query results
US20060203261A1 (en) * 2006-05-12 2006-09-14 Dhiraj Kacker Image ranking for imaging products and services
US20080141307A1 (en) * 2006-12-06 2008-06-12 Verizon Services Organization Inc. Customized media on demand
US20090254539A1 (en) * 2008-04-03 2009-10-08 Microsoft Corporation User Intention Modeling For Interactive Image Retrieval
US20120158717A1 (en) * 2009-09-04 2012-06-21 Koninklijke Philips Electronics N.V. Visualization of relevance for content-based image retrieval

Family Cites Families (9)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US7779117B2 (en) * 2002-05-31 2010-08-17 Aol Inc. Monitoring digital images
US7003515B1 (en) 2001-05-16 2006-02-21 Pandora Media, Inc. Consumer item matching method and system
EP1783632B1 (en) 2005-11-08 2012-12-19 Intel Corporation Content recommendation method with user feedback
US7958130B2 (en) 2008-05-26 2011-06-07 Microsoft Corporation Similarity-based content sampling and relevance feedback
US10210179B2 (en) 2008-11-18 2019-02-19 Excalibur Ip, Llc Dynamic feature weighting
US8112428B2 (en) 2008-11-24 2012-02-07 Yahoo! Inc. Clustering image search results through voting: reciprocal election
US8676803B1 (en) 2009-11-04 2014-03-18 Google Inc. Clustering images
US8416997B2 (en) 2010-01-27 2013-04-09 Apple Inc. Method of person identification using social connections
US9268792B2 (en) * 2011-01-24 2016-02-23 T-Mobile Usa, Inc. Automatic selection of digital images with an apparatus

Patent Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20040025180A1 (en) * 2001-04-06 2004-02-05 Lee Begeja Method and apparatus for interactively retrieving content related to previous query results
US20030187844A1 (en) * 2002-02-11 2003-10-02 Mingjing Li Statistical bigram correlation model for image retrieval
US20060203261A1 (en) * 2006-05-12 2006-09-14 Dhiraj Kacker Image ranking for imaging products and services
US20080141307A1 (en) * 2006-12-06 2008-06-12 Verizon Services Organization Inc. Customized media on demand
US20090254539A1 (en) * 2008-04-03 2009-10-08 Microsoft Corporation User Intention Modeling For Interactive Image Retrieval
US20120158717A1 (en) * 2009-09-04 2012-06-21 Koninklijke Philips Electronics N.V. Visualization of relevance for content-based image retrieval

Cited By (82)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US9195678B2 (en) 2011-01-24 2015-11-24 T-Mobile Usa, Inc. Automatic selection of digital images from a multi-sourced collection of digital images
US9223802B2 (en) 2011-01-24 2015-12-29 T-Mobile Usa, Inc. Automatic sharing of superlative digital images
US9268792B2 (en) 2011-01-24 2016-02-23 T-Mobile Usa, Inc. Automatic selection of digital images with an apparatus
US9576610B2 (en) * 2011-08-26 2017-02-21 Cyberlink Corp. Systems and methods of detecting significant faces in video streams
US9179201B2 (en) * 2011-08-26 2015-11-03 Cyberlink Corp. Systems and methods of detecting significant faces in video streams
US20160027474A1 (en) * 2011-08-26 2016-01-28 Cyberlink Corp. Systems and Methods of Detecting Significant Faces in Video Streams
US20130051756A1 (en) * 2011-08-26 2013-02-28 Cyberlink Corp. Systems and Methods of Detecting Significant Faces in Video Streams
US9967621B2 (en) * 2011-12-15 2018-05-08 Rovi Technologies Corporation Dynamic personalized program content
US20130160051A1 (en) * 2011-12-15 2013-06-20 Microsoft Corporation Dynamic Personalized Program Content
US9454992B2 (en) * 2011-12-27 2016-09-27 Intel Corporation Method and system to play linear video in variable time frames
US9298718B2 (en) * 2011-12-27 2016-03-29 Intel Corporation Method and system to play linear video in variable time frames
US8983272B2 (en) * 2011-12-27 2015-03-17 Intel Corporation Method and system to play linear video in variable time frames
US20130336639A1 (en) * 2011-12-27 2013-12-19 William C. DeLeeuw Method and system to play linear video in variable time frames
US9301016B2 (en) 2012-04-05 2016-03-29 Facebook, Inc. Sharing television and video programming through social networking
US20130282747A1 (en) * 2012-04-23 2013-10-24 Sri International Classification, search, and retrieval of complex video events
US9244924B2 (en) * 2012-04-23 2016-01-26 Sri International Classification, search, and retrieval of complex video events
US9686337B2 (en) 2012-08-31 2017-06-20 Facebook, Inc. Sharing television and video programming through social networking
US9549227B2 (en) 2012-08-31 2017-01-17 Facebook, Inc. Sharing television and video programming through social networking
US9171017B2 (en) 2012-08-31 2015-10-27 Facebook, Inc. Sharing television and video programming through social networking
US9110929B2 (en) 2012-08-31 2015-08-18 Facebook, Inc. Sharing television and video programming through social networking
US10536738B2 (en) 2012-08-31 2020-01-14 Facebook, Inc. Sharing television and video programming through social networking
US10425671B2 (en) 2012-08-31 2019-09-24 Facebook, Inc. Sharing television and video programming through social networking
US9386354B2 (en) 2012-08-31 2016-07-05 Facebook, Inc. Sharing television and video programming through social networking
US20190289354A1 (en) 2012-08-31 2019-09-19 Facebook, Inc. Sharing Television and Video Programming through Social Networking
US10405020B2 (en) 2012-08-31 2019-09-03 Facebook, Inc. Sharing television and video programming through social networking
US9461954B2 (en) * 2012-08-31 2016-10-04 Facebook, Inc. Sharing television and video programming through social networking
US9491133B2 (en) 2012-08-31 2016-11-08 Facebook, Inc. Sharing television and video programming through social networking
US9497155B2 (en) 2012-08-31 2016-11-15 Facebook, Inc. Sharing television and video programming through social networking
US10257554B2 (en) 2012-08-31 2019-04-09 Facebook, Inc. Sharing television and video programming through social networking
US10158899B2 (en) 2012-08-31 2018-12-18 Facebook, Inc. Sharing television and video programming through social networking
US10154297B2 (en) 2012-08-31 2018-12-11 Facebook, Inc. Sharing television and video programming through social networking
US9912987B2 (en) 2012-08-31 2018-03-06 Facebook, Inc. Sharing television and video programming through social networking
US9578390B2 (en) 2012-08-31 2017-02-21 Facebook, Inc. Sharing television and video programming through social networking
US9201904B2 (en) 2012-08-31 2015-12-01 Facebook, Inc. Sharing television and video programming through social networking
US10142681B2 (en) 2012-08-31 2018-11-27 Facebook, Inc. Sharing television and video programming through social networking
US10028005B2 (en) 2012-08-31 2018-07-17 Facebook, Inc. Sharing television and video programming through social networking
US9992534B2 (en) 2012-08-31 2018-06-05 Facebook, Inc. Sharing television and video programming through social networking
US9660950B2 (en) 2012-08-31 2017-05-23 Facebook, Inc. Sharing television and video programming through social networking
US9667584B2 (en) 2012-08-31 2017-05-30 Facebook, Inc. Sharing television and video programming through social networking
US9674135B2 (en) 2012-08-31 2017-06-06 Facebook, Inc. Sharing television and video programming through social networking
US9854303B2 (en) 2012-08-31 2017-12-26 Facebook, Inc. Sharing television and video programming through social networking
US9699485B2 (en) 2012-08-31 2017-07-04 Facebook, Inc. Sharing television and video programming through social networking
US9723373B2 (en) 2012-08-31 2017-08-01 Facebook, Inc. Sharing television and video programming through social networking
US9743157B2 (en) 2012-08-31 2017-08-22 Facebook, Inc. Sharing television and video programming through social networking
US9807454B2 (en) 2012-08-31 2017-10-31 Facebook, Inc. Sharing television and video programming through social networking
US20140325568A1 (en) * 2013-04-26 2014-10-30 Microsoft Corporation Dynamic creation of highlight reel tv show
US20150026578A1 (en) * 2013-07-22 2015-01-22 Sightera Technologies Ltd. Method and system for integrating user generated media items with externally generated media items
US20150078726A1 (en) * 2013-09-17 2015-03-19 Babak Robert Shakib Sharing Highlight Reels
US11200916B2 (en) 2013-09-17 2021-12-14 Google Llc Highlighting media through weighting of people or contexts
US9652475B2 (en) 2013-09-17 2017-05-16 Google Technology Holdings LLC Highlight reels
US10811050B2 (en) * 2013-09-17 2020-10-20 Google Technology Holdings LLC Highlighting media through weighting of people or contexts
US9436705B2 (en) 2013-09-17 2016-09-06 Google Technology Holdings LLC Grading images and video clips
US20150082172A1 (en) * 2013-09-17 2015-03-19 Babak Robert Shakib Highlighting Media Through Weighting of People or Contexts
US20150147045A1 (en) * 2013-11-26 2015-05-28 Sony Corporation Computer ecosystem with automatically curated video montage
US11042274B2 (en) * 2013-12-04 2021-06-22 Autodesk, Inc. Extracting demonstrations from in-situ video content
US20150160836A1 (en) * 2013-12-04 2015-06-11 Autodesk, Inc. Extracting demonstrations from in-situ video content
US11575829B2 (en) 2014-05-21 2023-02-07 Google Llc Enhanced image capture
US11290639B2 (en) 2014-05-21 2022-03-29 Google Llc Enhanced image capture
US11943532B2 (en) 2014-05-21 2024-03-26 Google Technology Holdings LLC Enhanced image capture
US11019252B2 (en) 2014-05-21 2021-05-25 Google Technology Holdings LLC Enhanced image capture
US11397502B2 (en) 2015-02-24 2022-07-26 Axon Enterprise, Inc. Systems and methods for bulk redaction of recorded data
EP3262833B1 (en) * 2015-02-24 2023-02-22 Axon Enterprise, Inc. Systems and methods for bulk redaction of recorded data
US11126317B2 (en) 2015-02-24 2021-09-21 Axon Enterprise, Inc. Systems and methods for bulk redaction of recorded data
US20180063253A1 (en) * 2015-03-09 2018-03-01 Telefonaktiebolaget Lm Ericsson (Publ) Method, system and device for providing live data streams to content-rendering devices
WO2016200692A1 (en) * 2015-06-11 2016-12-15 Vieu Labs, Inc. Editing, sharing, and viewing video
US20160365120A1 (en) * 2015-06-11 2016-12-15 Eran Steinberg Video editing system for generating multiple final cut clips
US20160365118A1 (en) * 2015-06-11 2016-12-15 Eran Steinberg Video editing system with multi-stakeholder control
US10269387B2 (en) 2015-09-30 2019-04-23 Apple Inc. Audio authoring and compositing
US20170092324A1 (en) * 2015-09-30 2017-03-30 Apple Inc. Automatic Video Compositing
US20170092331A1 (en) 2015-09-30 2017-03-30 Apple Inc. Synchronizing Audio and Video Components of an Automatically Generated Audio/Video Presentation
US10726594B2 (en) 2015-09-30 2020-07-28 Apple Inc. Grouping media content for automatically generating a media presentation
US10692537B2 (en) 2015-09-30 2020-06-23 Apple Inc. Synchronizing audio and video components of an automatically generated audio/video presentation
US10062415B2 (en) 2015-09-30 2018-08-28 Apple Inc. Synchronizing audio and video components of an automatically generated audio/video presentation
US10381041B2 (en) * 2016-02-16 2019-08-13 Shimmeo, Inc. System and method for automated video editing
US20190214054A1 (en) * 2016-02-16 2019-07-11 Shimmeo, Inc System and Method for Automated Video Editing
US11315600B2 (en) * 2017-11-06 2022-04-26 International Business Machines Corporation Dynamic generation of videos based on emotion and sentiment recognition
US20190182565A1 (en) * 2017-12-13 2019-06-13 Playable Pty Ltd System and Method for Algorithmic Editing of Video Content
US11729478B2 (en) * 2017-12-13 2023-08-15 Playable Pty Ltd System and method for algorithmic editing of video content
US10880428B2 (en) * 2018-08-13 2020-12-29 Avaya Inc. Selective communication event extraction
US20200053203A1 (en) * 2018-08-13 2020-02-13 Avaya Inc. Selective communication event extraction
US11893796B2 (en) 2018-11-07 2024-02-06 Genetec Inc. Methods and systems for detection of anomalous motion in a video stream and for creating a video summary
US10971192B2 (en) 2018-11-07 2021-04-06 Genetec Inc. Methods and systems for detection of anomalous motion in a video stream and for creating a video summary

Also Published As

Publication number Publication date
US20120188382A1 (en) 2012-07-26
US9195678B2 (en) 2015-11-24
US20120188405A1 (en) 2012-07-26
US9223802B2 (en) 2015-12-29
US20120191709A1 (en) 2012-07-26
US9268792B2 (en) 2016-02-23

Similar Documents

Publication Publication Date Title
US20120189284A1 (en) Automatic highlight reel producer
US11995530B2 (en) Systems and methods for providing feedback for artificial intelligence-based image capture devices
US20220068312A1 (en) Highlighting Media Through Weighting of People or Contexts
US10885380B2 (en) Automatic suggestion to share images
US10628680B2 (en) Event-based image classification and scoring
US9569658B2 (en) Image sharing with facial recognition models
EP3210371B1 (en) Camera application
US10248867B2 (en) Systems and methods for providing playback of selected video segments
US10459968B2 (en) Image processing system and image processing method
US20120226663A1 (en) Preconfigured media file uploading and sharing
US9521211B2 (en) Content processing device, content processing method, computer-readable recording medium, and integrated circuit
US9692963B2 (en) Method and electronic apparatus for sharing photographing setting values, and sharing system
US20150242405A1 (en) Methods, devices and systems for context-sensitive organization of media files
US10070175B2 (en) Method and system for synchronizing usage information between device and server
JP5878523B2 (en) Content processing apparatus and integrated circuit, method and program thereof
US20150189118A1 (en) Photographing apparatus, photographing system, photographing method, and recording medium recording photographing control program
CN115696049A (en) Micro video system, format and generation method
EP3073743A1 (en) Systems and methods for providing playback of selected video segments
WO2016079609A1 (en) Generation apparatus and method for evaluation information, electronic device and server
CN111949618A (en) Processing method and processing device

Legal Events

Date Code Title Description
AS Assignment

Owner name: T-MOBILE USA, INC., WASHINGTON

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:MORRISON, ANDREW;PARK, CHRIS;LAU, KEVIN;AND OTHERS;REEL/FRAME:025906/0423

Effective date: 20110304

STCB Information on status: application discontinuation

Free format text: ABANDONED -- FAILURE TO RESPOND TO AN OFFICE ACTION

AS Assignment

Owner name: DEUTSCHE TELEKOM AG, GERMANY

Free format text: INTELLECTUAL PROPERTY SECURITY AGREEMENT;ASSIGNOR:T-MOBILE USA, INC.;REEL/FRAME:041225/0910

Effective date: 20161229

AS Assignment

Owner name: IBSV LLC, WASHINGTON

Free format text: RELEASE BY SECURED PARTY;ASSIGNOR:DEUTSCHE BANK AG NEW YORK BRANCH;REEL/FRAME:052969/0314

Effective date: 20200401

Owner name: T-MOBILE SUBSIDIARY IV CORPORATION, WASHINGTON

Free format text: RELEASE BY SECURED PARTY;ASSIGNOR:DEUTSCHE BANK AG NEW YORK BRANCH;REEL/FRAME:052969/0314

Effective date: 20200401

Owner name: IBSV LLC, WASHINGTON

Free format text: RELEASE BY SECURED PARTY;ASSIGNOR:DEUTSCHE TELEKOM AG;REEL/FRAME:052969/0381

Effective date: 20200401

Owner name: PUSHSPRING, INC., WASHINGTON

Free format text: RELEASE BY SECURED PARTY;ASSIGNOR:DEUTSCHE BANK AG NEW YORK BRANCH;REEL/FRAME:052969/0314

Effective date: 20200401

Owner name: METROPCS WIRELESS, INC., WASHINGTON

Free format text: RELEASE BY SECURED PARTY;ASSIGNOR:DEUTSCHE BANK AG NEW YORK BRANCH;REEL/FRAME:052969/0314

Effective date: 20200401

Owner name: LAYER3 TV, INC., WASHINGTON

Free format text: RELEASE BY SECURED PARTY;ASSIGNOR:DEUTSCHE BANK AG NEW YORK BRANCH;REEL/FRAME:052969/0314

Effective date: 20200401

Owner name: T-MOBILE USA, INC., WASHINGTON

Free format text: RELEASE BY SECURED PARTY;ASSIGNOR:DEUTSCHE TELEKOM AG;REEL/FRAME:052969/0381

Effective date: 20200401

Owner name: METROPCS COMMUNICATIONS, INC., WASHINGTON

Free format text: RELEASE BY SECURED PARTY;ASSIGNOR:DEUTSCHE BANK AG NEW YORK BRANCH;REEL/FRAME:052969/0314

Effective date: 20200401

Owner name: T-MOBILE USA, INC., WASHINGTON

Free format text: RELEASE BY SECURED PARTY;ASSIGNOR:DEUTSCHE BANK AG NEW YORK BRANCH;REEL/FRAME:052969/0314

Effective date: 20200401