US20190320213A1 - Media management based on derived quantitative data of quality - Google Patents

Media management based on derived quantitative data of quality Download PDF

Info

Publication number
US20190320213A1
US20190320213A1 US16/452,768 US201916452768A US2019320213A1 US 20190320213 A1 US20190320213 A1 US 20190320213A1 US 201916452768 A US201916452768 A US 201916452768A US 2019320213 A1 US2019320213 A1 US 2019320213A1
Authority
US
United States
Prior art keywords
video
video file
segment
files
file
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Abandoned
Application number
US16/452,768
Inventor
Robert Sayko
Sean Carolan
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
AT&T Corp
AT&T Intellectual Property II LP
Original Assignee
AT&T Corp
AT&T Intellectual Property II LP
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by AT&T Corp, AT&T Intellectual Property II LP filed Critical AT&T Corp
Priority to US16/452,768 priority Critical patent/US20190320213A1/en
Assigned to AT&T CORP. reassignment AT&T CORP. ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: CAROLAN, SEAN, SAYKO, ROBERT
Publication of US20190320213A1 publication Critical patent/US20190320213A1/en
Abandoned legal-status Critical Current

Links

Images

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/20Servers specifically adapted for the distribution of content, e.g. VOD servers; Operations thereof
    • H04N21/23Processing of content or additional data; Elementary server operations; Server middleware
    • H04N21/234Processing of video elementary streams, e.g. splicing of video streams or manipulating encoded video stream scene graphs
    • H04N21/23418Processing of video elementary streams, e.g. splicing of video streams or manipulating encoded video stream scene graphs involving operations for analysing video streams, e.g. detecting features or characteristics
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/40Information retrieval; Database structures therefor; File system structures therefor of multimedia data, e.g. slideshows comprising image and additional audio data
    • G06F16/43Querying
    • G06F16/438Presentation of query results
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/40Information retrieval; Database structures therefor; File system structures therefor of multimedia data, e.g. slideshows comprising image and additional audio data
    • G06F16/48Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/90Details of database functions independent of the retrieved data types
    • G06F16/95Retrieval from the web
    • G06F16/951Indexing; Web crawling techniques
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/10Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
    • H04N19/134Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the element, parameter or criterion affecting or controlling the adaptive coding
    • H04N19/146Data rate or code amount at the encoder output
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/10Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
    • H04N19/169Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding
    • H04N19/17Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/85Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using pre-processing or post-processing specially adapted for video compression
    • H04N19/86Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using pre-processing or post-processing specially adapted for video compression involving reduction of coding artifacts, e.g. of blockiness
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/20Servers specifically adapted for the distribution of content, e.g. VOD servers; Operations thereof
    • H04N21/23Processing of content or additional data; Elementary server operations; Server middleware
    • H04N21/231Content storage operation, e.g. caching movies for short term storage, replicating data over plural servers, prioritizing data for deletion
    • H04N21/23109Content storage operation, e.g. caching movies for short term storage, replicating data over plural servers, prioritizing data for deletion by placing content in organized collections, e.g. EPG data repository
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/20Servers specifically adapted for the distribution of content, e.g. VOD servers; Operations thereof
    • H04N21/23Processing of content or additional data; Elementary server operations; Server middleware
    • H04N21/235Processing of additional data, e.g. scrambling of additional data or processing content descriptors
    • H04N21/2353Processing of additional data, e.g. scrambling of additional data or processing content descriptors specifically adapted to content descriptors, e.g. coding, compressing or processing of metadata
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/20Servers specifically adapted for the distribution of content, e.g. VOD servers; Operations thereof
    • H04N21/23Processing of content or additional data; Elementary server operations; Server middleware
    • H04N21/239Interfacing the upstream path of the transmission network, e.g. prioritizing client content requests
    • H04N21/2393Interfacing the upstream path of the transmission network, e.g. prioritizing client content requests involving handling client requests
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/70Information retrieval; Database structures therefor; File system structures therefor of video data
    • G06F16/78Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually
    • G06F16/783Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually using metadata automatically derived from the content
    • G06F16/7844Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually using metadata automatically derived from the content using original textual content or text extracted from visual content or transcript of audio data
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/70Information retrieval; Database structures therefor; File system structures therefor of video data
    • G06F16/78Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually
    • G06F16/783Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually using metadata automatically derived from the content
    • G06F16/7847Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually using metadata automatically derived from the content using low-level visual features of the video content

Definitions

  • This invention generally pertains to qualitative analysis of data and more particularly to analyzing and using quality data to rank multiple copies of a media file.
  • Media files may have the same file type, associated bit rates and/or other similar static attributes which all imply a certain level of quality. However, quality of such media can still be significantly different as it may be affected by numerous conditions that can create random “noise” or artifacts. This in turn affects how a user perceives the media when viewing/hearing it (e.g. bad video or audio), or can reduce/interfere with compression algorithms, impede delivery protocols or subsequent copy and reproduction of the media.
  • quality of such media can still be significantly different as it may be affected by numerous conditions that can create random “noise” or artifacts. This in turn affects how a user perceives the media when viewing/hearing it (e.g. bad video or audio), or can reduce/interfere with compression algorithms, impede delivery protocols or subsequent copy and reproduction of the media.
  • multiple copies of a media file may be available but in different formats which may further affect the media file or play back. There may be multiple reasons why media files may have differing qualities.
  • An exemplary method embodiment of the invention comprises receiving a request for a media file, identifying one or more media files associated with the request, measuring at least one audio or visual quality associated with each media file by analyzing at least one of: compression artifacts or grading for each media file, generating quantitative data based on the measured at least one audio or visual quality for each media file, and returning the quantitative data associated with each media file.
  • FIG. 1 illustrates a basic system or computing device embodiment of the invention
  • FIG. 2 illustrates an image developing artifacts after compressiong
  • FIG. 3 illustrates examples of sorting similar media files
  • FIG. 4 illustrates an example method embodiment of the invention.
  • an exemplary system for implementing the invention includes a general-purpose computing device 100 , including a processing unit (CPU) 120 and a system bus 110 that couples various system components including the system memory such as read only memory (ROM) 140 and random access memory (RAM) 150 to the processing unit 120 .
  • system memory 130 may be available for use as well.
  • the system bus 110 may be any of several types of bus structures including a memory bus or memory controller, a peripheral bus, and a local bus using any of a variety of bus architectures.
  • the computing device 100 further includes storage means such as a hard disk drive 160 , a magnetic disk drive, an optical disk drive, tape drive or the like.
  • the storage device 160 is connected to the system bus 110 by a drive interface.
  • the drives and the associated computer readable media provide nonvolatile storage of computer readable instructions, data structures, program modules and other data for the computing device 100 .
  • the basic components are known to those of skill in the art and appropriate variations are contemplated depending on the type of device, such as whether the device is a small, handheld computing device, a desktop computer, or a computer server.
  • an input device 190 represents any number of input mechanisms, such as a microphone for speech, a touch sensitive screen for gesture or graphical input, keyboard, mouse, motion input, speech and so forth.
  • the input may be used by the presenter to indicate the beginning of a speech search query.
  • the device output 170 can also be one or more of a number of output means.
  • multimodal systems enable a user to provide multiple types of input to communicate with the computing device 100 .
  • the communications interface 180 generally governs and manages the user input and system output. There is no restriction on the invention operating on any particular hardware arrangement and therefore the basic features here may easily be substituted for improved hardware or firmware arrangements as they are developed.
  • the illustrative embodiment of the present invention is presented as comprising individual functional blocks (including functional blocks labeled as a “processor”).
  • the functions these blocks represent may be provided through the use of either shared or dedicated hardware, including, but not limited to, hardware capable of executing software.
  • the functions of one or more processors presented in FIG. 1 may be provided by a single shared processor or multiple processors.
  • Illustrative embodiments may comprise microprocessor and/or digital signal processor (DSP) hardware, read-only memory (ROM) for storing software performing the operations discussed below, and random access memory (RAM) for storing results.
  • DSP digital signal processor
  • ROM read-only memory
  • RAM random access memory
  • VLSI Very large scale integration
  • media file(s) is intended to encompass all forms of digital media, including, but not limited to, graphic images, sounds, animations, photographs, and video.
  • a media file may be stored in any format, or copies of the same or similar media file may be stored in multiple formats.
  • the present invention relates to media management based on derived quantitative quality of data.
  • Quality data of media files is usually tagged via the file type and parameters such as an associated bit rate (e.g. MP3 with 192 bitrate).
  • quality of such media can still be significantly different as it may be affected by numerous conditions that can create random “noise” or artifacts.
  • the present invention can identify the quality of a media file, much like a human might analyze the quality of an audio or video file, and then use that quality in a beneficial manner to aid the user in differentiating between media files.
  • Some encoding and transcoding applications can detect these qualitative issues and can include a measure of the quality into the meta-data for the media.
  • a specific real world example embodiment applies to the YouTube.com experience.
  • YouTube there are often multiple instances of the same video (e.g. multiple copies of the television segment of a Tiger Woods putt). While the files are all supposed to be of the same type, not all are of the same quality. This is perhaps due to the method used by the different submitters of the file to process/capture it, but there are of course other things that could also have affected quality.
  • This invention therefore includes not only the method of compiling and using quality data about a file for sorting, but also the concept of adding this as a service/feature to a application like YouTube. Alternatively, this capability can be completely client based, but in the specific example of YouTube, it would be more effective as a server/network based application. Similarly, YouTube itself could analyze files in this way to determine a quality rating and then associate that with a quantitative grade that would allow users to block/filter videos that they feel are not worth viewing.
  • FIG. 2 illustrates an example of an image developing artifacts during compression 200 .
  • the original image 202 is clear with a good color grade.
  • the image becomes fuzzy with a loss of edge clarity 204 .
  • the preferred embodiment can detect errors in the creation of the data which can come across as artifacts.
  • Example artifacts include, but are not limited to noise, pixilation, saturation, resolution, inconsistencies in the file such as “blips,” and any errors or losses during the compression, encoding, or processing of a media file.
  • the preferred embodiment derives a quantitative representation of the quality of the media file, which may be included in metadata associated with the media file.
  • the quantitative representation should not be limited to a number.
  • the representation can be a matrix, vector, or other representation of a quality rating in any space consisting of one or more numbers, characters, symbols, or other tokens.
  • the preferred embodiment also takes into account a quality grading component.
  • the quality grading looks at a variety of factors and determines quality on a case to case basis.
  • the grading may be based on comparative picture quality.
  • the grading can look at how often pixels in a given decompressed output frame smear together with ones adjacent to them, or if the color depth and resolution is what it should be in the produced output. For example, most people don't realize that color video tape does not have nearly as much resolution in its color component as it does in its detailing black-and-white component.
  • the system is aware of that, and factors this in to the grading determination, thus taking into account the quality of the source (as well as can be measured) as well as the quality of the current copy generation being analyzed.
  • multiple media files of the same event might exist. For example, there might be multiple videos of a music concert. One video was taken with a cell phone while another with a video camera from a different angle.
  • factors the quality grading component may take into account include, but are not limited to, lighting, clarity, noise, pixilation, resolution, and color.
  • an example embodiment may determine how well it “understands” the person in order to measure the quality of the file.
  • the grading of the media file is quantized and may be included in metadata associated with the media file.
  • the quality of the grading component and the artifacts can be measured not only during the compression, encoding, or processing of a file, but also in realtime.
  • An example embodiment uses a real time tool to “parse” or preview a file to get an instantaneous quality rating.
  • a media file that has not received quantitative data of quality during compressing, encoding, or processing can be streamed and given an instantaneous quality rating.
  • defining the rating scale is a separate issue that can be done in a variety of ways.
  • the scope of the present invention also includes a service that can be offered to users of content. It can save time for users who may want the best version of a media file and/or users engaged in archiving. For example, a news company may want to only save the best version of a particular event they recorded. The news company can use the present invention to automatically determine and archive only the best version of a media file. This alleviates a person having to view all the versions of a given media file.
  • the news agency can filter files that do not meet a certain threshold in quality.
  • An example embodiment may archive files above a threshold quality rating and discard those below.
  • Users of services such as websites that stream media, can also filter files that do not meet a threshold or that are not the best version of a file. For instance, if two of the same music video exist on a website, and one of the videos is inferior in visual or audio quality than the other, the user can filter out the inferior video so his search only results in the best music video being returned.
  • FIG. 3 illustrates what the results of a YouTube search 300 for “john Mayer daughters” 302 may look like. The music video appears twice 304 along with various live recordings 306 . YouTube allows the user to sort by relevance, date added, view count, or user rating 308 .
  • the present invention allows the option of sorting based on a quality rating 310 . This rating may be included in the metadata or can be acquired in realtime. Both an objective 314 and relative 316 quality rating 312 may be presented to the user.
  • the objective rating is based on a scale of ten.
  • the rating can be defined in numerous ways such as using a green symbol to indicate a good-quality file, a yellow for an average-quality file, and a red symbol for a bad-quality file. Defining the rating is not at issue, but rather the ability to measure the quality of the media file and manage media content based on the measure.
  • the quality rating may depend on a case by case basis. For example, the user might wish to select the video with the best sound quality. The user could also select the video based on the best picture quality or any other relevant factor or combination of factors.
  • Relative ratings are the quality of a media file compared to one or more other media files. The comparisons may be based on groupings determined as described below. In FIG. 3 , the first John Mayer video is ranked as the best of the two in its group, as indicated by the “1 ⁇ 2” 316 . Files may have the same objective rating yet still have different relative ratings.
  • the media files may be grouped in various ways when returned from a request for a media file. For example, a user looking for the best version of a song to download can search the availability of the song on various disparate databases. The song can be compared, grouped, and presented to the user; though not necessarily in that order. Media files may also be grouped according to their time. For instance, media files that are close in length (time), such as a one second difference, may be grouped together while media files that differ greatly in time. The name of the media file may also be used for grouping. In FIG.
  • Grouping may also be based on the content of a media file. To illustrate, the content of a media file may be compared with the content of another media file. If the content is deemed similar enough, then the media files may be grouped. Grouping may take into account one or more of the aforementioned factors.
  • Another aspect of this invention is the quality of media files can be compared.
  • various websites offer compressed music files to purchase for download.
  • the quality of the music, from service to service may differ based on the processing, compression, or encoding of the file.
  • the present invention can do a quality comparison of these files.
  • the comparison may be used in order to obtain quantitative data relating to the quality of the media file.
  • a comparison of quantitative data of quality may be acquired independently for each media file, and the quantitative data can be compared.
  • the comparison may be between files of the same type (e.g. MP3 and MP3) or files of different types (e.g. MP3 and MP4).
  • FIG. 4 illustrates an exemplary method embodiment of the invention.
  • the method comprises receiving a request for a media file 402 .
  • the request may be either user or machine (e.g. computing device) generated.
  • the method also comprises identifying one or more media files associated with the request measuring at least one audio or visual quality 404 .
  • the method further comprises measuring at least one of audio and/or visual quality associated with each media file by analyzing at least one of: compression artifacts or grading for each media file 406 . There may be different ratings for different audio or visual qualities for each media file. To illustrate, there may be separate ratings for both the bass and treble on an audio track.
  • the method further comprises generating quantitative data based on the measured at least one audio or visual quality for each media file 408 .
  • the quantitative data comprises any data that is representative of a quality rating of at least one audio or visual quality. There may be different data for different qualities or a single rating for multiple qualities.
  • the data can be included in metadata or be obtained instantaneously using a real-time parsing tool. As discussed above, the data can be used to sort, filter, compare, archive, or select media files.
  • the method also comprises returning the quantitative data associated with each media file 410 .
  • the information may also be presented to a user. Such information may or may not aide the user in selecting a media file. For instance, multiple files of a musician performing a song may exist, albeit the performance takes place at different locations.
  • the quality of the media files of the different performances can be presented for the user's information, although a show at a particular location might interest a viewer more even though the file quality may be worse.
  • the information may also be returned to a computing device. For instance, the information may assist in automatic archival of media files.
  • the system may analyze multiple media files and discover that portions of one media file may be degraded or have artifacts while other portions are of relative high quality.
  • an aspect of the invention may be to parse media files into separately rated segments and then concatenate the highest quality segments together. The end result in this scenario would be a media file that has a higher quality than any individual media file from which the best version was constructed.
  • the system may perform some automatic analysis of format and perform format conversion to insure that the final media file is of the best quality and of the appropriate format.
  • a dialog with the user may aid in some decisions (i.e., “Would you prefer the media file in format X or YT').
  • the user can data in a media file of the highest quality possible.
  • Embodiments within the scope of the present invention may also include computer-readable media for carrying or having computer-executable instructions or data structures stored thereon.
  • Such computer-readable media can be any available media that can be accessed by a general purpose or special purpose computer.
  • Such computer-readable media can comprise RAM, ROM, EEPROM, CD-ROM or other optical disk storage, magnetic disk storage or other magnetic storage devices, or any other medium which can be used to carry or store desired program code means in the form of computer-executable instructions or data structures.
  • a network or another communications connection either hardwired, wireless, or combination thereof
  • any such connection is properly termed a computer-readable medium. Combinations of the above should also be included within the scope of the computer-readable media.
  • Computer-executable instructions include, for example, instructions and data which cause a general purpose computer, special purpose computer, or special purpose processing device to perform a certain function or group of functions.
  • Computer-executable instructions also include program modules that are executed by computers in stand-alone or network environments.
  • program modules include routines, programs, objects, components, and data structures, etc. that perform particular tasks or implement particular abstract data types.
  • Computer-executable instructions, associated data structures, and program modules represent examples of the program code means for executing steps of the methods disclosed herein. The particular sequence of such executable instructions or associated data structures represents examples of corresponding acts for implementing the functions described in such steps.
  • Embodiments of the invention may be practiced in network computing environments with many types of computer system configurations, including personal computers, hand-held devices, multi-processor systems, microprocessor-based or programmable consumer electronics, network PCs, minicomputers, mainframe computers, and the like. Embodiments may also be practiced in distributed computing environments where tasks are performed by local and remote processing devices that are linked (either by hardwired links, wireless links, or by a combination thereof) through a communications network. In a distributed computing environment, program modules may be located in both local and remote memory storage devices.

Landscapes

  • Engineering & Computer Science (AREA)
  • Multimedia (AREA)
  • Theoretical Computer Science (AREA)
  • Signal Processing (AREA)
  • Databases & Information Systems (AREA)
  • Data Mining & Analysis (AREA)
  • Physics & Mathematics (AREA)
  • General Engineering & Computer Science (AREA)
  • General Physics & Mathematics (AREA)
  • Library & Information Science (AREA)
  • Two-Way Televisions, Distribution Of Moving Picture Or The Like (AREA)

Abstract

Systems, methods, and computer readable media for media management provide a derived quantitative data of quality for video media files. The method embodiment comprises receiving a request for a video media file, identifying one or more video media files associated with the request, measuring at least one audio or visual quality associated with each video media file by analyzing at least one of: compression artifacts or grading for the each video media file, generating quantitative data based on the measured at least one audio or visual quality for the each video media file, and returning the quantitative data associated with the each video media file. Other features include the ability to compare, archive, filter, sort and select video media files based on the quantitative data.

Description

    PRIORITY INFORMATION
  • The present application is a continuation of U.S. patent application Ser. No. 14/966,297, filed Dec. 11, 2015, which is a continuation of U.S. patent application Ser. No. 11/847,800, filed Aug. 30, 2007, the content of which is incorporated herein by reference in its entirety.
  • BACKGROUND OF THE INVENTION 1. Field of the Invention
  • This invention generally pertains to qualitative analysis of data and more particularly to analyzing and using quality data to rank multiple copies of a media file.
  • 2. Introduction
  • Digital cameras, portable media players, and web sites that allow visitors to stream media clips have contributed to a growing demand for media content. With so much media content available and easily accessible, the audio or visual quality of the media file can be an important factor in terms of sorting, filtering, and selection.
  • Media files may have the same file type, associated bit rates and/or other similar static attributes which all imply a certain level of quality. However, quality of such media can still be significantly different as it may be affected by numerous conditions that can create random “noise” or artifacts. This in turn affects how a user perceives the media when viewing/hearing it (e.g. bad video or audio), or can reduce/interfere with compression algorithms, impede delivery protocols or subsequent copy and reproduction of the media.
  • Furthermore, multiple copies of a media file may be available but in different formats which may further affect the media file or play back. There may be multiple reasons why media files may have differing qualities.
  • Accordingly, what is needed in the art is an improved way to provide derived quantitative data of quality for media files.
  • SUMMARY OF THE INVENTION
  • Additional features and advantages of the invention will be set forth in the description which follows, and in part will be obvious from the description, or may be learned by practice of the invention. The features and advantages of the invention may be realized and obtained by means of the instruments and combinations particularly pointed out in the appended claims. These and other features of the present invention will become more fully apparent from the following description and appended claims, or may be learned by the practice of the invention as set forth herein.
  • Disclosed herein are systems, methods, and computer readable media for providing a ranking of media files. An exemplary method embodiment of the invention comprises receiving a request for a media file, identifying one or more media files associated with the request, measuring at least one audio or visual quality associated with each media file by analyzing at least one of: compression artifacts or grading for each media file, generating quantitative data based on the measured at least one audio or visual quality for each media file, and returning the quantitative data associated with each media file.
  • BRIEF DESCRIPTION OF THE DRAWINGS
  • In order to describe the manner in which the above-recited and other advantages and features of the invention can be obtained, a more particular description of the invention briefly described above will be rendered by reference to specific embodiments thereof which are illustrated in the appended drawings. Understanding that these drawings depict only typical embodiments of the invention and are not therefore to be considered to be limiting of its scope, the invention will be described and explained with additional specificity and detail through the use of the accompanying drawings in which:
  • FIG. 1 illustrates a basic system or computing device embodiment of the invention;
  • FIG. 2 illustrates an image developing artifacts after compressiong;
  • FIG. 3 illustrates examples of sorting similar media files; and
  • FIG. 4 illustrates an example method embodiment of the invention.
  • DETAILED DESCRIPTION OF THE INVENTION
  • Various embodiments of the invention are discussed in detail below. While specific implementations are discussed, it should be understood that this is done for illustration purposes only. A person skilled in the relevant art will recognize that other components and configurations may be used without parting from the spirit and scope of the invention.
  • With reference to FIG. 1, an exemplary system for implementing the invention includes a general-purpose computing device 100, including a processing unit (CPU) 120 and a system bus 110 that couples various system components including the system memory such as read only memory (ROM) 140 and random access memory (RAM) 150 to the processing unit 120. Other system memory 130 may be available for use as well. It can be appreciated that the invention may operate on a computing device with more than one CPU 120 or on a group or cluster of computing devices networked together to provide greater processing capability. The system bus 110 may be any of several types of bus structures including a memory bus or memory controller, a peripheral bus, and a local bus using any of a variety of bus architectures. A basic input/output (BIOS), containing the basic routine that helps to transfer information between elements within the computing device 100, such as during start-up, is typically stored in ROM 140. The computing device 100 further includes storage means such as a hard disk drive 160, a magnetic disk drive, an optical disk drive, tape drive or the like. The storage device 160 is connected to the system bus 110 by a drive interface. The drives and the associated computer readable media provide nonvolatile storage of computer readable instructions, data structures, program modules and other data for the computing device 100. The basic components are known to those of skill in the art and appropriate variations are contemplated depending on the type of device, such as whether the device is a small, handheld computing device, a desktop computer, or a computer server.
  • Although the exemplary environment described herein employs the hard disk, it should be appreciated by those skilled in the art that other types of computer readable media which can store data that are accessible by a computer, such as magnetic cassettes, flash memory cards, digital versatile disks, cartridges, random access memories (RAMs), read only memory (ROM), a cable or wireless signal containing a bit stream and the like, may also be used in the exemplary operating environment.
  • To enable user interaction with the computing device 100, an input device 190 represents any number of input mechanisms, such as a microphone for speech, a touch sensitive screen for gesture or graphical input, keyboard, mouse, motion input, speech and so forth. The input may be used by the presenter to indicate the beginning of a speech search query. The device output 170 can also be one or more of a number of output means. In some instances, multimodal systems enable a user to provide multiple types of input to communicate with the computing device 100. The communications interface 180 generally governs and manages the user input and system output. There is no restriction on the invention operating on any particular hardware arrangement and therefore the basic features here may easily be substituted for improved hardware or firmware arrangements as they are developed.
  • For clarity of explanation, the illustrative embodiment of the present invention is presented as comprising individual functional blocks (including functional blocks labeled as a “processor”). The functions these blocks represent may be provided through the use of either shared or dedicated hardware, including, but not limited to, hardware capable of executing software. For example the functions of one or more processors presented in FIG. 1 may be provided by a single shared processor or multiple processors. (Use of the term “processor” should not be construed to refer exclusively to hardware capable of executing software.) Illustrative embodiments may comprise microprocessor and/or digital signal processor (DSP) hardware, read-only memory (ROM) for storing software performing the operations discussed below, and random access memory (RAM) for storing results. Very large scale integration (VLSI) hardware embodiments, as well as custom VLSI circuitry in combination with a general purpose DSP circuit, may also be provided.
  • As used herein, the term “media file(s)” is intended to encompass all forms of digital media, including, but not limited to, graphic images, sounds, animations, photographs, and video. A media file may be stored in any format, or copies of the same or similar media file may be stored in multiple formats.
  • The present invention relates to media management based on derived quantitative quality of data. Quality data of media files is usually tagged via the file type and parameters such as an associated bit rate (e.g. MP3 with 192 bitrate). However, as suggested above, quality of such media can still be significantly different as it may be affected by numerous conditions that can create random “noise” or artifacts. The present invention can identify the quality of a media file, much like a human might analyze the quality of an audio or video file, and then use that quality in a beneficial manner to aid the user in differentiating between media files. Some encoding and transcoding applications can detect these qualitative issues and can include a measure of the quality into the meta-data for the media. Alternatively, there are now new content management tools that can actually “view” or “listen” to media to attain a “user perspective” on quality. Information like this can also be associated with the media. In the latter case, the media can be “examined” in real time each time the file is to be used or is transferred—thus unlike meta-data created at time of file encoding, quality can be evaluated as desired. All this allows a user of media/content to select, sort and use the best quality file from a group of similar media/files.
  • A specific real world example embodiment applies to the YouTube.com experience. On YouTube, there are often multiple instances of the same video (e.g. multiple copies of the television segment of a Tiger Woods putt). While the files are all supposed to be of the same type, not all are of the same quality. This is perhaps due to the method used by the different submitters of the file to process/capture it, but there are of course other things that could also have affected quality. This invention therefore includes not only the method of compiling and using quality data about a file for sorting, but also the concept of adding this as a service/feature to a application like YouTube. Alternatively, this capability can be completely client based, but in the specific example of YouTube, it would be more effective as a server/network based application. Similarly, YouTube itself could analyze files in this way to determine a quality rating and then associate that with a quantitative grade that would allow users to block/filter videos that they feel are not worth viewing.
  • In the YouTube example, deriving a quantitative quality of data can take into account compression artifacts, qualitative grading, or any combination of the two. During the encoding or processing, a media file may develop compression artifacts. FIG. 2 illustrates an example of an image developing artifacts during compression 200. The original image 202 is clear with a good color grade. However, after being converted to JPEG format, the image becomes fuzzy with a loss of edge clarity 204. The preferred embodiment can detect errors in the creation of the data which can come across as artifacts. Example artifacts include, but are not limited to noise, pixilation, saturation, resolution, inconsistencies in the file such as “blips,” and any errors or losses during the compression, encoding, or processing of a media file. After detecting the artifacts, the preferred embodiment derives a quantitative representation of the quality of the media file, which may be included in metadata associated with the media file. The quantitative representation, however, should not be limited to a number. For example, the representation can be a matrix, vector, or other representation of a quality rating in any space consisting of one or more numbers, characters, symbols, or other tokens.
  • The preferred embodiment also takes into account a quality grading component. In the preferred embodiment the quality grading looks at a variety of factors and determines quality on a case to case basis. For visual media files, the grading may be based on comparative picture quality. The grading can look at how often pixels in a given decompressed output frame smear together with ones adjacent to them, or if the color depth and resolution is what it should be in the produced output. For example, most people don't realize that color video tape does not have nearly as much resolution in its color component as it does in its detailing black-and-white component. In the preferred embodiment, the system is aware of that, and factors this in to the grading determination, thus taking into account the quality of the source (as well as can be measured) as well as the quality of the current copy generation being analyzed. In many instances, multiple media files of the same event might exist. For example, there might be multiple videos of a music concert. One video was taken with a cell phone while another with a video camera from a different angle. In this example, factors the quality grading component may take into account include, but are not limited to, lighting, clarity, noise, pixilation, resolution, and color. In a recording of a person speaking, an example embodiment may determine how well it “understands” the person in order to measure the quality of the file. Once again, the grading of the media file is quantized and may be included in metadata associated with the media file.
  • The quality of the grading component and the artifacts can be measured not only during the compression, encoding, or processing of a file, but also in realtime. An example embodiment uses a real time tool to “parse” or preview a file to get an instantaneous quality rating. For example, a media file that has not received quantitative data of quality during compressing, encoding, or processing, can be streamed and given an instantaneous quality rating. As suggested above, defining the rating scale is a separate issue that can be done in a variety of ways.
  • The scope of the present invention also includes a service that can be offered to users of content. It can save time for users who may want the best version of a media file and/or users engaged in archiving. For example, a news company may want to only save the best version of a particular event they recorded. The news company can use the present invention to automatically determine and archive only the best version of a media file. This alleviates a person having to view all the versions of a given media file.
  • Likewise, the news agency can filter files that do not meet a certain threshold in quality. An example embodiment may archive files above a threshold quality rating and discard those below. Users of services, such as websites that stream media, can also filter files that do not meet a threshold or that are not the best version of a file. For instance, if two of the same music video exist on a website, and one of the videos is inferior in visual or audio quality than the other, the user can filter out the inferior video so his search only results in the best music video being returned.
  • Media files can also be sorted based on the derived quantitative data relating to the quality of the file. In many cases, a service has different recordings of the same event or different encodings of the same media clip. For example, FIG. 3 illustrates what the results of a YouTube search 300 for “john Mayer daughters” 302 may look like. The music video appears twice 304 along with various live recordings 306. YouTube allows the user to sort by relevance, date added, view count, or user rating 308. The present invention allows the option of sorting based on a quality rating 310. This rating may be included in the metadata or can be acquired in realtime. Both an objective 314 and relative 316 quality rating 312 may be presented to the user. In the example embodiment, the objective rating is based on a scale of ten. The first music video returned received an objective rating of “8314. However, the rating can be defined in numerous ways such as using a green symbol to indicate a good-quality file, a yellow for an average-quality file, and a red symbol for a bad-quality file. Defining the rating is not at issue, but rather the ability to measure the quality of the media file and manage media content based on the measure. The quality rating may depend on a case by case basis. For example, the user might wish to select the video with the best sound quality. The user could also select the video based on the best picture quality or any other relevant factor or combination of factors. Relative ratings are the quality of a media file compared to one or more other media files. The comparisons may be based on groupings determined as described below. In FIG. 3, the first John Mayer video is ranked as the best of the two in its group, as indicated by the “½” 316. Files may have the same objective rating yet still have different relative ratings.
  • The media files may be grouped in various ways when returned from a request for a media file. For example, a user looking for the best version of a song to download can search the availability of the song on various disparate databases. The song can be compared, grouped, and presented to the user; though not necessarily in that order. Media files may also be grouped according to their time. For instance, media files that are close in length (time), such as a one second difference, may be grouped together while media files that differ greatly in time. The name of the media file may also be used for grouping. In FIG. 3, out of the four concerts returned 306 by the search for “john mayer daughters” 302, three have the title “John Mayer—Daughters—Irvine” and are grouped and compared to each other 318. The user who posted the media file may also factor into grouping. Generally, a user would not post the same media file twice on the same site. Grouping may also be based on the content of a media file. To illustrate, the content of a media file may be compared with the content of another media file. If the content is deemed similar enough, then the media files may be grouped. Grouping may take into account one or more of the aforementioned factors.
  • Another aspect of this invention is the quality of media files can be compared. For example, various websites offer compressed music files to purchase for download. The quality of the music, from service to service, may differ based on the processing, compression, or encoding of the file. The present invention can do a quality comparison of these files. In one embodiment, the comparison may be used in order to obtain quantitative data relating to the quality of the media file. In another embodiment a comparison of quantitative data of quality may be acquired independently for each media file, and the quantitative data can be compared. The comparison may be between files of the same type (e.g. MP3 and MP3) or files of different types (e.g. MP3 and MP4).
  • FIG. 4 illustrates an exemplary method embodiment of the invention. The method comprises receiving a request for a media file 402. The request may be either user or machine (e.g. computing device) generated. The method also comprises identifying one or more media files associated with the request measuring at least one audio or visual quality 404. The method further comprises measuring at least one of audio and/or visual quality associated with each media file by analyzing at least one of: compression artifacts or grading for each media file 406. There may be different ratings for different audio or visual qualities for each media file. To illustrate, there may be separate ratings for both the bass and treble on an audio track. The method further comprises generating quantitative data based on the measured at least one audio or visual quality for each media file 408. The quantitative data comprises any data that is representative of a quality rating of at least one audio or visual quality. There may be different data for different qualities or a single rating for multiple qualities. The data can be included in metadata or be obtained instantaneously using a real-time parsing tool. As discussed above, the data can be used to sort, filter, compare, archive, or select media files. The method also comprises returning the quantitative data associated with each media file 410. The information may also be presented to a user. Such information may or may not aide the user in selecting a media file. For instance, multiple files of a musician performing a song may exist, albeit the performance takes place at different locations. The quality of the media files of the different performances can be presented for the user's information, although a show at a particular location might interest a viewer more even though the file quality may be worse. The information may also be returned to a computing device. For instance, the information may assist in automatic archival of media files.
  • In another aspect, the system may analyze multiple media files and discover that portions of one media file may be degraded or have artifacts while other portions are of relative high quality. With digital media files, an aspect of the invention may be to parse media files into separately rated segments and then concatenate the highest quality segments together. The end result in this scenario would be a media file that has a higher quality than any individual media file from which the best version was constructed.
  • The system may perform some automatic analysis of format and perform format conversion to insure that the final media file is of the best quality and of the appropriate format. A dialog with the user may aid in some decisions (i.e., “Would you prefer the media file in format X or YT'). Thus, using the quality analysis approach, in this aspect of the invention, the user can data in a media file of the highest quality possible.
  • Embodiments within the scope of the present invention may also include computer-readable media for carrying or having computer-executable instructions or data structures stored thereon. Such computer-readable media can be any available media that can be accessed by a general purpose or special purpose computer. By way of example, and not limitation, such computer-readable media can comprise RAM, ROM, EEPROM, CD-ROM or other optical disk storage, magnetic disk storage or other magnetic storage devices, or any other medium which can be used to carry or store desired program code means in the form of computer-executable instructions or data structures. When information is transferred or provided over a network or another communications connection (either hardwired, wireless, or combination thereof) to a computer, the computer properly views the connection as a computer-readable medium. Thus, any such connection is properly termed a computer-readable medium. Combinations of the above should also be included within the scope of the computer-readable media.
  • Computer-executable instructions include, for example, instructions and data which cause a general purpose computer, special purpose computer, or special purpose processing device to perform a certain function or group of functions. Computer-executable instructions also include program modules that are executed by computers in stand-alone or network environments. Generally, program modules include routines, programs, objects, components, and data structures, etc. that perform particular tasks or implement particular abstract data types. Computer-executable instructions, associated data structures, and program modules represent examples of the program code means for executing steps of the methods disclosed herein. The particular sequence of such executable instructions or associated data structures represents examples of corresponding acts for implementing the functions described in such steps.
  • Those of skill in the art will appreciate that other embodiments of the invention may be practiced in network computing environments with many types of computer system configurations, including personal computers, hand-held devices, multi-processor systems, microprocessor-based or programmable consumer electronics, network PCs, minicomputers, mainframe computers, and the like. Embodiments may also be practiced in distributed computing environments where tasks are performed by local and remote processing devices that are linked (either by hardwired links, wireless links, or by a combination thereof) through a communications network. In a distributed computing environment, program modules may be located in both local and remote memory storage devices.
  • Although the above description may contain specific details, they should not be construed as limiting the claims in any way. Other configurations of the described embodiments of the invention are part of the scope of this invention. For example, the present invention could be configured according to the type of media file being evaluated or according to the device on which it is implemented. To illustrate, a photographer could rapidly snap multiple photographs of the same scene with the camera having an automatic filter keeping only the best quality photograph. Accordingly, the appended claims and their legal equivalents should only define the invention, rather than any specific examples given.

Claims (20)

I claim:
1. A method comprising:
identifying a respective segment within each video file of a set of video files, to yield a respective identified segment of each video file;
rating a signal quality of the respective identified segment of each video file according to a number of compression artifacts found in the respective identified segment of each video file, to yield a respective rated video segment; and
concatenating a composite version of a video using the respective rated video segment from multiple instances of the set of video files.
2. The method of claim 1, further comprising:
identifying the set of video files based on a request, where each video file in the set of video files is an instance of the video.
3. The method of claim 2, further comprising:
returning, in response to the request, the composite version.
4. The method of claim 1, further comprising returning, with the composite version, metadata associated with the composite version.
5. The method of claim 4, wherein the metadata comprises quantitative data associated with the signal quality.
6. The method of claim 1, wherein rating the signal quality occurs concurrently with the identifying of the respective segment within each video file.
7. The method of claim 1, wherein rating of the signal quality is further based on a comparison between each video file and another video file in the set of video files.
8. The method of claim 1, further comprising archiving the set of video files according to the signal quality of the respective identified segment of each video file.
9. The method of claim 1, further comprising:
identifying a plurality of segments within each video file of the set of video files, to yield identified segments of each video file; and
rating the signal quality of the identified segments of each video file according to a number of compression artifacts found in the identified segments of each video file, to yield the respective rated video segment, wherein the composite version uses a highest ranked segment of the identified segments of each video file.
10. A system comprising:
a processor; and
a computer-readable storage medium having instructions stored which, when executed by the processor, cause the processor to perform operations comprising:
identifying a respective segment within each video file of a set of video files, to yield a respective identified segment of each video file;
rating a signal quality of the respective identified segment of each video file according to a number of compression artifacts found in the respective identified segment of each video file, to yield a respective rated video segment; and
concatenating a composite version of a video using the respective rated video segment from multiple instances of the set of video files.
11. The system of claim 10, wherein the computer-readable storage medium stores additional instructions stored which, when executed by the processor, cause the processor to perform operations further comprising:
identifying the set of video files based on a request, where each video file in the set of video files is an instance of the video.
12. The system of claim 11, wherein the computer-readable storage medium stores additional instructions stored which, when executed by the processor, cause the processor to perform operations further comprising:
returning, in response to the request, the composite version.
13. The system of claim 10, wherein the computer-readable storage medium stores additional instructions stored which, when executed by the processor, cause the processor to perform operations further comprising:
returning, with the composite version, metadata associated with the composite version.
14. The system of claim 13, wherein the metadata comprises quantitative data associated with the signal quality.
15. The system of claim 10, wherein rating the signal quality occurs concurrently with the identifying of the respective segment within each video file.
16. The system of claim 10, wherein rating of the signal quality is further based on a comparison between each video file and another video file in the set of video files.
17. The system of claim 10, wherein the computer-readable storage medium stores additional instructions stored which, when executed by the processor, cause the processor to perform operations further comprising:
archiving the set of video files according to the signal quality of the respective identified segment of each video file.
18. The system of claim 10, wherein the computer-readable storage medium stores additional instructions stored which, when executed by the processor, cause the processor to perform operations further comprising:
identifying a plurality of segments within each video file of the set of video files, to yield identified segments of each video file; and
rating the signal quality of the identified segments of each video file according to a number of compression artifacts found in the identified segments of each video file, to yield the respective rated video segment, wherein the composite version uses a highest ranked segment of the identified segments of each video file.
19. A non-transitory computer-readable storage device having instructions stored which, when executed by a computing device, cause the computing device to perform operations comprising:
identifying a respective segment within each video file of a set of video files, to yield a respective identified segment of each video file;
rating a signal quality of the respective identified segment of each video file according to a number of compression artifacts found in the respective identified segment of each video file, to yield a respective rated video segment; and
concatenating a composite version of a video using the respective rated video segment from multiple instances of the set of video files.
20. The transitory computer-readable storage device of claim 19, wherein the non-transitory computer-readable storage device stores additional instructions stored which, when executed by the computing device, cause the computing device to perform operations further comprising:
identifying the set of video files based on a request, where each video file in the set of video files is an instance of the video.
US16/452,768 2007-08-30 2019-06-26 Media management based on derived quantitative data of quality Abandoned US20190320213A1 (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
US16/452,768 US20190320213A1 (en) 2007-08-30 2019-06-26 Media management based on derived quantitative data of quality

Applications Claiming Priority (3)

Application Number Priority Date Filing Date Title
US11/847,800 US9304994B2 (en) 2007-08-30 2007-08-30 Media management based on derived quantitative data of quality
US14/966,297 US10341695B2 (en) 2007-08-30 2015-12-11 Media management based on derived quantitative data of quality
US16/452,768 US20190320213A1 (en) 2007-08-30 2019-06-26 Media management based on derived quantitative data of quality

Related Parent Applications (1)

Application Number Title Priority Date Filing Date
US14/966,297 Continuation US10341695B2 (en) 2007-08-30 2015-12-11 Media management based on derived quantitative data of quality

Publications (1)

Publication Number Publication Date
US20190320213A1 true US20190320213A1 (en) 2019-10-17

Family

ID=40409148

Family Applications (3)

Application Number Title Priority Date Filing Date
US11/847,800 Expired - Fee Related US9304994B2 (en) 2007-08-30 2007-08-30 Media management based on derived quantitative data of quality
US14/966,297 Active 2028-05-19 US10341695B2 (en) 2007-08-30 2015-12-11 Media management based on derived quantitative data of quality
US16/452,768 Abandoned US20190320213A1 (en) 2007-08-30 2019-06-26 Media management based on derived quantitative data of quality

Family Applications Before (2)

Application Number Title Priority Date Filing Date
US11/847,800 Expired - Fee Related US9304994B2 (en) 2007-08-30 2007-08-30 Media management based on derived quantitative data of quality
US14/966,297 Active 2028-05-19 US10341695B2 (en) 2007-08-30 2015-12-11 Media management based on derived quantitative data of quality

Country Status (1)

Country Link
US (3) US9304994B2 (en)

Families Citing this family (15)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US8005801B2 (en) * 2008-06-02 2011-08-23 Microsoft Corporation Aging and compressing multimedia content
US20100251291A1 (en) * 2009-03-24 2010-09-30 Pino Jr Angelo J System, Method and Computer Program Product for Processing Video Data
FR2947069A1 (en) * 2009-06-19 2010-12-24 Thomson Licensing METHOD OF SELECTING VERSIONS OF A DOCUMENT AMONG A PLURALITY OF VERSIONS RECEIVED FOLLOWING A SEARCH, AND ASSOCIATED RECEIVER
US8718145B1 (en) * 2009-08-24 2014-05-06 Google Inc. Relative quality score for video transcoding
US9465879B2 (en) * 2010-01-14 2016-10-11 Excalibur Ip, Llc Search engine recency using content preview
US8712930B1 (en) 2010-08-09 2014-04-29 Google Inc. Encoding digital content based on models for predicting similarity between exemplars
US8601334B2 (en) 2011-05-10 2013-12-03 At&T Intellectual Property I, L.P. System and method for delivering content over a multicast network
US9677886B2 (en) * 2013-02-10 2017-06-13 Qualcomm Incorporated Method and apparatus for navigation based on media density along possible routes
EP3540725A1 (en) * 2013-09-27 2019-09-18 Koninklijke Philips N.V. Simultaneously displaying video data of multiple video sources
US20150146012A1 (en) * 2013-11-27 2015-05-28 Sprint Communications Company L.P. Video presentation quality display in a wireless communication device
US9639275B2 (en) 2014-08-06 2017-05-02 Seagate Technology Llc Managing data within a storage device based on file system metadata
US9749686B2 (en) 2015-09-21 2017-08-29 Sling Media Pvt Ltd. Video analyzer
US9693063B2 (en) * 2015-09-21 2017-06-27 Sling Media Pvt Ltd. Video analyzer
WO2017096277A1 (en) * 2015-12-04 2017-06-08 Sling Media, Inc. Network-based event recording
CN106126526A (en) * 2016-06-13 2016-11-16 浪潮电子信息产业股份有限公司 A kind of data managing method and device

Family Cites Families (23)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US5249053A (en) * 1991-02-05 1993-09-28 Dycam Inc. Filmless digital camera with selective image compression
US5216503A (en) * 1991-12-24 1993-06-01 General Instrument Corporation Statistical multiplexer for a multichannel image compression system
US5966465A (en) * 1994-09-21 1999-10-12 Ricoh Corporation Compression/decompression using reversible embedded wavelets
US5867166A (en) * 1995-08-04 1999-02-02 Microsoft Corporation Method and system for generating images using Gsprites
US6125201A (en) * 1997-06-25 2000-09-26 Andrew Michael Zador Method, apparatus and system for compressing data
US5990955A (en) * 1997-10-03 1999-11-23 Innovacom Inc. Dual encoding/compression method and system for picture quality/data density enhancement
US6195780B1 (en) * 1997-12-10 2001-02-27 Lucent Technologies Inc. Method and apparatus for generating cyclical redundancy code
US6493023B1 (en) * 1999-03-12 2002-12-10 The United States Of America As Represented By The Administrator Of The National Aeronautics And Space Administration Method and apparatus for evaluating the visual quality of processed digital video sequences
US7281034B1 (en) * 2000-01-24 2007-10-09 Friskit, Inc. System and method for media playback over a network using links that contain control signals and commands
US6970859B1 (en) * 2000-03-23 2005-11-29 Microsoft Corporation Searching and sorting media clips having associated style and attributes
US7296231B2 (en) * 2001-08-09 2007-11-13 Eastman Kodak Company Video structuring by probabilistic merging of video segments
US6954543B2 (en) * 2002-02-28 2005-10-11 Ipac Acquisition Subsidiary I, Llc Automated discovery, assignment, and submission of image metadata to a network-based photosharing service
US20030185301A1 (en) * 2002-04-02 2003-10-02 Abrams Thomas Algie Video appliance
US7038710B2 (en) * 2002-07-17 2006-05-02 Koninklijke Philips Electronics, N.V. Method and apparatus for measuring the quality of video data
US7298930B1 (en) * 2002-11-29 2007-11-20 Ricoh Company, Ltd. Multimodal access of meeting recordings
US7933338B1 (en) * 2004-11-10 2011-04-26 Google Inc. Ranking video articles
JP4981026B2 (en) * 2005-03-31 2012-07-18 コーニンクレッカ フィリップス エレクトロニクス エヌ ヴィ Composite news story synthesis
US20070209025A1 (en) * 2006-01-25 2007-09-06 Microsoft Corporation User interface for viewing images
US7620716B2 (en) * 2006-01-31 2009-11-17 Dell Products L.P. System and method to predict the performance of streaming media over wireless links
US7683931B2 (en) * 2006-03-06 2010-03-23 Dell Products L.P. Image artifact detection in video quality benchmarks
US20070217761A1 (en) * 2006-03-07 2007-09-20 Coban Research And Technologies, Inc. Method for video/audio recording using unrestricted pre-event/post-event buffering with multiple bit and frame rates buffer files
WO2007112111A2 (en) * 2006-03-29 2007-10-04 Packetvideo Corp. System and method for securing content ratings
WO2008008150A2 (en) * 2006-07-10 2008-01-17 Thomson Licensing Methods and apparatus for enhanced performance in a multi-pass video encoder

Also Published As

Publication number Publication date
US20160100204A1 (en) 2016-04-07
US9304994B2 (en) 2016-04-05
US10341695B2 (en) 2019-07-02
US20090063561A1 (en) 2009-03-05

Similar Documents

Publication Publication Date Title
US20190320213A1 (en) Media management based on derived quantitative data of quality
US10698952B2 (en) Using digital fingerprints to associate data with a work
US5903892A (en) Indexing of media content on a network
US6370543B2 (en) Display of media previews
US5983176A (en) Evaluation of media content in media files
US9179200B2 (en) Method and system for determining content treatment
US8687941B2 (en) Automatic static video summarization
US8156132B1 (en) Systems for comparing image fingerprints
US20150154204A1 (en) Signature-based system and method for generation of personalized multimedia channels
US20240212706A1 (en) Audio data processing
WO2003088665A1 (en) Meta data edition device, meta data reproduction device, meta data distribution device, meta data search device, meta data reproduction condition setting device, and meta data distribution method
US20220107978A1 (en) Method for recommending video content
US7774385B1 (en) Techniques for providing a surrogate heuristic identification interface
US7302437B2 (en) Methods, systems, and computer-readable media for a global video format schema defining metadata relating to video media
US20080018503A1 (en) Method and apparatus for encoding/playing multimedia contents
CN110149529B (en) Media information processing method, server and storage medium
US8549022B1 (en) Fingerprint generation of multimedia content based on a trigger point with the multimedia content
US8320667B2 (en) Automatic and scalable image selection
US20190311746A1 (en) Indexing media content library using audio track fingerprinting
CN113591853A (en) Keyword extraction method and device and electronic equipment
EP1527453A1 (en) Method, system and program product for generating a content-based table of contents
WO2014103374A1 (en) Information management device, server and control method
WO2009044351A1 (en) Generation of image data summarizing a sequence of video frames
US9020964B1 (en) Generation of fingerprints for multimedia content based on vectors and histograms
US20120290437A1 (en) System and Method of Selecting and Acquiring Still Images from Video

Legal Events

Date Code Title Description
AS Assignment

Owner name: AT&T CORP., NEW YORK

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:SAYKO, ROBERT;CAROLAN, SEAN;REEL/FRAME:049591/0684

Effective date: 20070828

STPP Information on status: patent application and granting procedure in general

Free format text: APPLICATION DISPATCHED FROM PREEXAM, NOT YET DOCKETED

STPP Information on status: patent application and granting procedure in general

Free format text: NON FINAL ACTION MAILED

STCB Information on status: application discontinuation

Free format text: ABANDONED -- FAILURE TO RESPOND TO AN OFFICE ACTION