US20150237341A1 - Method and apparatus for managing audio visual, audio or visual content - Google Patents

Method and apparatus for managing audio visual, audio or visual content Download PDF

Info

Publication number
US20150237341A1
US20150237341A1 US14/623,354 US201514623354A US2015237341A1 US 20150237341 A1 US20150237341 A1 US 20150237341A1 US 201514623354 A US201514623354 A US 201514623354A US 2015237341 A1 US2015237341 A1 US 2015237341A1
Authority
US
United States
Prior art keywords
fingerprint
audio
stream
fingerprints
values
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Abandoned
Application number
US14/623,354
Inventor
Jonathan Diggins
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Snell Advanced Media Ltd
Original Assignee
Snell Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Snell Ltd filed Critical Snell Ltd
Assigned to SNELL LIMITED reassignment SNELL LIMITED ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: DIGGINS, JONATHAN
Publication of US20150237341A1 publication Critical patent/US20150237341A1/en
Priority to US15/459,860 priority Critical patent/US10219033B2/en
Priority to US16/283,491 priority patent/US10893323B2/en
Abandoned legal-status Critical Current

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V20/00Scenes; Scene-specific elements
    • G06V20/40Scenes; Scene-specific elements in video content
    • G06V20/46Extracting features or characteristics from the video content, e.g. video fingerprints, representative shots or key frames
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/60Information retrieval; Database structures therefor; File system structures therefor of audio data
    • G06F16/68Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually
    • G06F16/683Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually using metadata automatically derived from the content
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/70Information retrieval; Database structures therefor; File system structures therefor of video data
    • G06F16/78Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually
    • G06F16/783Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually using metadata automatically derived from the content
    • G06F17/30743
    • G06F17/30784
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/22Matching criteria, e.g. proximity measures
    • G06K9/00744
    • G06K9/00758
    • G06K9/6201
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/70Arrangements for image or video recognition or understanding using pattern recognition or machine learning
    • G06V10/74Image or video pattern matching; Proximity measures in feature spaces
    • G06V10/75Organisation of the matching processes, e.g. simultaneous or sequential comparisons of image or video features; Coarse-fine approaches, e.g. multi-scale approaches; using context analysis; Selection of dictionaries
    • G06V10/758Involving statistics of pixels or of feature values, e.g. histogram matching
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V20/00Scenes; Scene-specific elements
    • G06V20/40Scenes; Scene-specific elements in video content
    • G06V20/41Higher-level, semantic clustering, classification or understanding of video scenes, e.g. detection, labelling or Markovian modelling of sport events or news items
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V20/00Scenes; Scene-specific elements
    • G06V20/40Scenes; Scene-specific elements in video content
    • G06V20/48Matching video sequences
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N17/00Diagnosis, testing or measuring for television systems or their details
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/43Processing of content or additional data, e.g. demultiplexing additional data from a digital video stream; Elementary client operations, e.g. monitoring of home network or synchronising decoder's clock; Client middleware
    • H04N21/44Processing of video elementary streams, e.g. splicing a video clip retrieved from local storage with an incoming video stream, rendering scenes according to MPEG-4 scene graphs
    • H04N21/44008Processing of video elementary streams, e.g. splicing a video clip retrieved from local storage with an incoming video stream, rendering scenes according to MPEG-4 scene graphs involving operations for analysing video streams, e.g. detecting features or characteristics in the video stream
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/43Processing of content or additional data, e.g. demultiplexing additional data from a digital video stream; Elementary client operations, e.g. monitoring of home network or synchronising decoder's clock; Client middleware
    • H04N21/442Monitoring of processes or resources, e.g. detecting the failure of a recording device, monitoring the downstream bandwidth, the number of times a movie has been viewed, the storage space available from the internal hard disk
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/80Generation or processing of content or additional data by content creator independently of the distribution process; Content per se
    • H04N21/83Generation or processing of protective or descriptive data associated with content; Content structuring
    • H04N21/835Generation of protective data, e.g. certificates
    • H04N21/8358Generation of protective data, e.g. certificates involving watermark
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04HBROADCAST COMMUNICATION
    • H04H20/00Arrangements for broadcast or for distribution combined with broadcast
    • H04H20/12Arrangements for observation, testing or troubleshooting
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04HBROADCAST COMMUNICATION
    • H04H2201/00Aspects of broadcast communication
    • H04H2201/90Aspects of broadcast communication characterised by the use of signatures
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N17/00Diagnosis, testing or measuring for television systems or their details
    • H04N2017/006Diagnosis, testing or measuring for television systems or their details for television sound

Definitions

  • This invention concerns automatic monitoring or other managing of audio, video and audio visual content.
  • QC and monitoring equipment analyses audio visual data using a variety of different algorithms that identify specific characteristics of the content such as:
  • results of this analysis may be stored as ‘metadata’ that is associated with the audio visual content; or, it may be used in a monitoring system that detects defects in distributed content and alerts an operator, or automatically makes changes to signal routing etc. to correct the defect.
  • QC and monitoring processing is complex, and the resulting volume of metadata is large.
  • QC equipment is therefore usually placed at only a few points in a distribution or processing system, perhaps only at the system's input and output points.
  • This invention takes advantage of another area of development in the field of audio visual content production and distribution is the processing of audio and video content to form ‘signatures’ or ‘fingerprints’ that describe some characteristic of the content with a very small amount of data.
  • signatures or fingerprints are associated with some temporal position or segment within the content, such as a video frame, and enable the relative timing between content streams to be measured; and, the equivalence of content at different points in a distribution network to be confirmed.
  • fingerprint will be used to describe this type of data.
  • a compressed image may be produced by a lossy encoding process which is irreversible, the compressed image remains an image and can be converted to viewable form through a suitable decoding process.
  • a fingerprint cannot by any sensible process be converted to a viewable image.
  • Fingerprint generating equipment is typically simple, cheap and placed at many points within a distribution or processing system.
  • the invention consists in one aspect in a method and apparatus for inferring metadata from a plurality of fingerprints derived by an irreversible data reduction process from respective temporal regions within a particular audio visual, audio or visual content stream wherein the said metadata is not directly encoded in the fingerprints and the plurality of fingerprints is received via a communication network from a fingerprint generator that is physically separate from the inference process.
  • a stream of fingerprint values is windowed and the frequencies of occurrence of particular fingerprint values or ranges of fingerprint values are compared with expected frequencies of occurrence for particular types of audio visual content.
  • the sustained occurrence of particular values of a spatial video fingerprint are detected and compared with one or more expected values for one or more expected images so as to generate metadata indicating the presence of a particular expected image.
  • the sustained occurrence of low values of an audio fingerprint are detected and metadata indicating silence is generated.
  • the pattern of differences between succeeding values of a temporal video fingerprint is compared with expected patterns of film cadence and metadata indicating a film cadence is generated.
  • FIG. 1 shows an exemplary system according to an embodiment of the invention.
  • FIG. 2 shows a metadata processor according to an embodiment of the invention.
  • FIG. 3 shows a sequence of video temporal fingerprint values from which the positions of shot changes can be identified.
  • FIGS. 4 a - 4 c show three examples of sequences of video temporal fingerprint values from which film cadence can be identified.
  • FIG. 5 shows a metadata processor according to an alternative embodiment of the invention.
  • FIG. 6 shows a metadata processor according to a further alternative embodiment of the invention.
  • FIG. 1 A system according to an embodiment of the invention is shown in FIG. 1 .
  • An audio visual data stream ( 1 ) is input to a fingerprint generator ( 2 ) at a point in an audio visual content distribution system.
  • the fingerprint generator ( 2 ) outputs a fingerprint stream ( 3 ) that describes the audio visual data stream ( 1 ).
  • the fingerprint stream ( 3 ) may describe either the audio or the video elements of the audio visual data stream ( 1 ), but typically will contain information relating to both.
  • the fingerprint stream ( 3 ) comprises a sequence of fingerprints, where each member of the sequence relates to a different temporal position in the data stream ( 1 ).
  • the video element of each fingerprint is derived from a different frame of video data; and, the audio element of each fingerprint is derived from a different set of audio samples.
  • the data rate of fingerprint stream ( 3 ) is very much less than the data rate of the audio visual data stream ( 1 ).
  • the audio component of the fingerprint stream ( 3 ) has a data rate of around 150 byte/s
  • the video component of the fingerprint stream ( 3 ) has a data rate of around 500 byte/s.
  • the derivation of the fingerprint from the audio visual data is a non-reversible process; it is not possible to re-construct the audio visual data from the fingerprint.
  • the fingerprint can be considered a hash-function of the audio visual data such that it is highly unlikely that different audio visual data will give the same fingerprint.
  • the fingerprint stream ( 3 ) is input to a fingerprint processor ( 4 ) that derives metadata ( 5 ) from the fingerprint stream ( 3 ) and is further described below.
  • This second fingerprint stream is also routed to the fingerprint processor ( 4 ).
  • Other unrelated audio, video or audio visual streams from different points within the audio visual content production and distribution process can be fingerprinted and the results routed to the fingerprint processor ( 4 ).
  • the fingerprint stream ( 10 ) describing the audio visual data stream ( 9 ) is shown as a further input to the fingerprint processor ( 4 ).
  • the respective fingerprint streams can be conveyed to the fingerprint processor ( 4 ) over low bandwidth links; for example, narrow-band internet connections could be used.
  • the metadata ( 5 ) output from the metadata processor ( 4 ) comprises metadata describing the first and second audio visual streams ( 1 ) and ( 6 ) and any other audio visual streams whose respective fingerprint streams are input to it.
  • the fingerprint processor ( 4 ) would be situated at a central monitoring location, and its output metadata ( 5 ) would be input to a manual or automatic control system that seeks to maintain the correct operation of the audio visual content production and distribution system.
  • An input fingerprint stream ( 200 ) comprises spatial video fingerprint data, temporal video fingerprint data, and audio fingerprint data relating to a sequence of temporal positions in the audiovisual data stream from which it was derived. Typically this sequence of temporal positions corresponds to fields of an interlaced video stream, or frames of a progressive video stream. In the following description it is assumed that a fingerprint is input for every field of the audio visual sequence.
  • a separator ( 201 ) separates out the three components of each input fingerprint of the fingerprint stream ( 200 ).
  • the separated spatial video fingerprint stream ( 202 ) comprises respective pixel-value summations for a set of regions of each video field. This is input to a black detector ( 205 ) that compares the values with a threshold and detects the simultaneous occurrence of low values in all the regions for several consecutive fields. When this condition is detected, a Black metadata component ( 211 ) is output to a monitoring process.
  • the separated spatial video fingerprint stream ( 202 ) is also input to a test signal detector ( 206 ) that detects a sustained set of pixel-value summation values for a set of regions within each video field.
  • the test signal detector ( 206 ) compares the regional pixel-value summations contained within each fingerprint of the fingerprint sequence ( 202 ) with previously-derived regional pixel-value summations for known test signals.
  • the comparison results are compared with one or more thresholds to identify near equivalence of the values in the fingerprints with the respective values for known test signals. If a set of values closely corresponding to values for a particular known test signal, colour bars for example, is found in a consecutive sequence of fingerprints, a test signal metadata component ( 212 ) that identifies the presence of the particular test signal is output.
  • the separated temporal video fingerprint stream ( 203 ) is input to a still-image detector ( 207 ).
  • the separated temporal video fingerprint stream ( 203 ) typically comprises a measure of inter-field differences between pixel-value summations for a set of regions within each video field.
  • An example is a sum of the sums of inter-field differences for a set of regions within the frame, evaluated between a current field and a previous field. If the fingerprint contains an inter-frame difference value, or if an inter-frame difference can be derived from the fingerprint, then this is used. If a sustained low-value inter-field or inter-frame difference measure is found in a consecutive sequence of fingerprints, a still-image metadata component ( 213 ) that identifies lack of motion is output.
  • the separated temporal video fingerprint stream ( 203 ) is also input to a shot-change detector ( 208 ), which identifies isolated high values of the temporal video fingerprint by comparing the respective value differences between a fingerprint and its closely preceding and succeeding fingerprints with a threshold. If the temporal fingerprint for a field is significantly greater than the corresponding fingerprints for preceding and succeeding fields, then that field is identified as the first field of a new shot, and it is identified in a shot-change metadata output ( 214 ).
  • a graph of temporal fingerprint value versus time for a video sequence containing shot changes is shown in FIG. 3 .
  • the isolated peaks ( 31 ) to ( 36 ) correspond to shot-changes.
  • FIG. 4 shows examples of sequences of temporal video fingerprint values for three different film cadences.
  • the sequence of temporal fingerprints for succeeding fields is analysed in the film cadence detector ( 209 ), and the sequence of differences between the fingerprints is identified. If successive pairs of temporal fingerprints from adjacent fields have similar values (i.e. the differences are less than a threshold), as shown in FIG. 4 a , then it is inferred that each pair comes from a new film frame; this is commonly known a 2:2 film cadence. If two pairs of similar values are followed by a significantly different value in a continuing sequence, as shown in FIG.
  • 3:2 film cadence in which the ratio of the film frame rate to the video field rate is 2:5, is identified. And, if there is no pattern of similarity between the temporal fingerprints for succeeding fields, as shown in FIG. 4 c , then video cadence is identified.
  • the film cadence detector ( 209 ) detects the pattern of changes between the fingerprints for succeeding fields by a known method, such as correlation of sequences of inter-fingerprint difference values with candidate sequences of differences. Metadata indicating detected video cadence ( 215 ), detected 2:2 film cadence ( 216 ) or detected 3:2 film cadence ( 217 ) is output.
  • the separated audio fingerprint stream ( 204 ) is input to a silence detector ( 210 ).
  • Typical audio fingerprints are derived from the magnitudes of a sequence of adjacent audio samples. When the audio is silent the sample magnitudes are small and a sequence of low-value fingerprints results. When a sustained sequence of audio fingerprint values less than a low-value threshold is detected by the silence detector ( 210 ), it outputs silence metadata ( 218 ).
  • a further audio visual fingerprint analysis process is shown in FIG. 5 .
  • Each set of fingerprint values is converted, in a histogram generator ( 502 ), to a histogram giving the respective frequencies of occurrence of values, or ranges of values, within the set.
  • the sequence of histograms from the histogram generator ( 502 ), corresponding the sequence of adjacent fingerprint values from the window selector ( 501 ), is analysed statistically in a moment processor ( 503 ) and an entropy processor ( 504 ).
  • the moment processor ( 503 ) determines known statistical parameters of each histogram: The mean (or first moment); the variance (or second moment); the skew (or third moment); and the kurtosis (or fourth moment). The derivation of these known dimensionless parameters of the distribution of values within a set of values will not be described here as it is well-known to those skilled in the art.
  • the entropy processor ( 504 ) determines the entropy E, or ‘distinctiveness’ of each histogram.
  • a suitable measure is given by the following equation:
  • the stream of sets of dimensionless statistical parameters ( 505 ) from the moment processor ( 503 ), and the stream of entropy values ( 506 ) from the entropy processor ( 504 ) are input to a classifier ( 507 ) that compares each of its input data sets with reference data sets corresponding to known types of audiovisual content.
  • the output from the classifier ( 507 ) is metadata ( 508 ) that describes the type of audio visual content from which the fingerprint value sequence ( 500 ) was derived.
  • the output of the classifier ( 507 ) is a weighted sum of the outputs from a number of different, known comparison functions, where the weights and the functions have been previously selected in a known ‘training’ process.
  • candidate sets of comparison functions are applied iteratively to sets of statistical data ( 505 ) and entropy data ( 506 ) that have been derived from analysis (as shown in FIG. 5 ) of fingerprint data from known types of audio visual content.
  • the weights and comparison functions are selected during this training so as to obtain the best agreement between the result of the weighted sum of comparisons, and the known content type of the respective training data set.
  • the classifier ( 507 ) uses a set of comparison functions and respective weights determined in a prior training process to identify when its input corresponds to a particular member of a set of reference data sets that corresponds with a particular type of audio visual content.
  • the stream of sets of frequency components ( 603 ) from the transform processor ( 602 ) is input to a classifier ( 604 ) that operates in the same way as the above-described classifier ( 507 ) to recognise the spectral characteristics of known types of audio visual content.
  • Metadata ( 605 ) that describes the type of audio visual content from which the fingerprint value sequence ( 600 ) was derived is output from the classifier ( 604 ).
  • Some audio fingerprints for example the ‘bar code’ audio signature described in international patent application WO 2009/104022, comprise a sequence of one-bit binary values. These fingerprints can conveniently be described by run-length coding, in which a sequence of run-length values indicates counts of succeeding identical fingerprint values. This is a well-known method of data compression that represents a sequence of consecutive values by a single descriptor and run-length value. In the case of binary data, the descriptor is not required, as each run-length value represents a change of state of the binary data.
  • Run-length values for rolling windows of a fingerprint sequence can be histogrammed and the histograms of the frequencies of occurrence of run-length values, or ranges of run-length values used to identify characteristics of the material from which the fingerprints were derived.
  • the reliability of all the above-described methods of extracting metadata from fingerprint data can be improved by applying a temporal low-pass filter to the derived metadata.
  • Simple recursive filters a running average for example, are suitable.
  • speed of response is different for different types of metadata.
  • Some parameters describe a single frame, for example a black frame identifier.
  • Other parameters relate to a short sequence of frames, for example film cadence.
  • Yet others relate to hundreds, or even thousands, of frames, for example type of content.
  • the temporal filters applicable to these different types of metadata will have different bandwidths.
  • Changes in the values of metadata derived by the methods described in this specification contain useful information which can be used to derive higher level metadata. For example, the frequency of occurrence of shot changes can be used to infer content type.
  • a metadata inference process according to the invention can use one or more of these methods; not all elements of a particular fingerprint need be analysed.
  • the temporal position of an available audio fingerprint will have a fixed relationship to the temporal position of an associated available video fingerprint for the same content stream at the same point in an audio visual content production or distribution process.
  • combination of the results video fingerprint analysis according to the invention with the results of audio fingerprint analysis according to the invention will give a more reliable determination of metadata for the audio visual sequence than would be achieved by analysis of the audio or video fingerprints in isolation.

Abstract

To manage audio visual content, a stream of fingerprints is derived in a fingerprint generator and received at a fingerprint processor that is physically separate from the fingerprint generator. Metadata is generated by processing the fingerprints to detect the sustained occurrence of low values of an audio fingerprint to generate metadata indicating silence; comparing the pattern of differences between temporally succeeding values of a fingerprint with expected patterns of film cadence to generate metadata indicating a film cadence; and comparing differences between temporally succeeding values of a fingerprint with a threshold to generate metadata indicating a still image or freeze frame.

Description

    FIELD OF THE INVENTION
  • This invention concerns automatic monitoring or other managing of audio, video and audio visual content.
  • BACKGROUND OF THE INVENTION
  • The very large numbers of ‘channels’ output to terrestrial, satellite and cable distribution systems by typical broadcasters cannot be monitored economically by human viewers and listeners. And, audio visual content, such as films, television shows and commercials received from content providers cannot always be checked for conformance with technical standards by human operators when ‘ingested’ into a broadcaster's digital storage system. The historic practice of checking by a person who looks for defects and non-conformance with standards is no longer economic, or even feasible, for a modern digital broadcaster.
  • These developments have led to great advances in automated quality checking (QC) and monitoring systems for audio visual content. Typically QC and monitoring equipment analyses audio visual data using a variety of different algorithms that identify specific characteristics of the content such as:
      • Audio dynamic range
      • Duration of periods of silent audio or black video
      • Presence of subtitles
      • Presence of test signals
      • Video aspect ratio and presence or absence of ‘black bars’ at the edges of the video frame
      • Audio to video synchronisation
  • The results of this analysis may be stored as ‘metadata’ that is associated with the audio visual content; or, it may be used in a monitoring system that detects defects in distributed content and alerts an operator, or automatically makes changes to signal routing etc. to correct the defect.
  • Typical QC and monitoring processing is complex, and the resulting volume of metadata is large. QC equipment is therefore usually placed at only a few points in a distribution or processing system, perhaps only at the system's input and output points.
  • SUMMARY OF THE INVENTION
  • It is an object of certain embodiments of the present invention to provide improved method or apparatus for automatic monitoring or other managing of audio, video and audio visual content.
  • This invention takes advantage of another area of development in the field of audio visual content production and distribution is the processing of audio and video content to form ‘signatures’ or ‘fingerprints’ that describe some characteristic of the content with a very small amount of data. Typically these signatures or fingerprints are associated with some temporal position or segment within the content, such as a video frame, and enable the relative timing between content streams to be measured; and, the equivalence of content at different points in a distribution network to be confirmed. In the remainder of this specification the term fingerprint will be used to describe this type of data.
  • It is important to distinguish between fingerprints, which are primarily for content identification and audio to video synchronisation, and ancillary data associated with audio visual data. Ancillary data will often contain data derived from a QC process, and the ancillary data may be carried with the audio and video data in a similar way to the carriage of fingerprint data. However, ancillary data directly encodes metadata, and typically can be extracted by simple de-multiplexing and decoding.
  • It is also important to distinguish between fingerprints and compressed images. Whilst a compressed image may be produced by a lossy encoding process which is irreversible, the compressed image remains an image and can be converted to viewable form through a suitable decoding process. A fingerprint cannot by any sensible process be converted to a viewable image.
  • Fingerprint generating equipment is typically simple, cheap and placed at many points within a distribution or processing system.
  • The invention consists in one aspect in a method and apparatus for inferring metadata from a plurality of fingerprints derived by an irreversible data reduction process from respective temporal regions within a particular audio visual, audio or visual content stream wherein the said metadata is not directly encoded in the fingerprints and the plurality of fingerprints is received via a communication network from a fingerprint generator that is physically separate from the inference process.
  • In a first embodiment, characteristics of a stream of fingerprints are compared in a classifier with expected characteristics of particular types of audio visual content, and the inferred metadata identifies the content type from which the fingerprints were derived. Suitably, a stream of fingerprint values is converted to the frequency domain, and the resulting frequency components are compared with expected frequency components for particular types of audio visual content.
  • Alternatively, a stream of fingerprint values is windowed and the frequencies of occurrence of particular fingerprint values or ranges of fingerprint values are compared with expected frequencies of occurrence for particular types of audio visual content. In a second embodiment, the sustained occurrence of particular values of a spatial video fingerprint are detected and compared with one or more expected values for one or more expected images so as to generate metadata indicating the presence of a particular expected image.
  • In a third embodiment, the sustained occurrence of low values of an audio fingerprint are detected and metadata indicating silence is generated.
  • In a fourth embodiment, the pattern of differences between succeeding values of a temporal video fingerprint is compared with expected patterns of film cadence and metadata indicating a film cadence is generated.
  • BRIEF DESCRIPTION OF THE DRAWINGS
  • FIG. 1 shows an exemplary system according to an embodiment of the invention.
  • FIG. 2 shows a metadata processor according to an embodiment of the invention.
  • FIG. 3 shows a sequence of video temporal fingerprint values from which the positions of shot changes can be identified.
  • FIGS. 4 a-4 c show three examples of sequences of video temporal fingerprint values from which film cadence can be identified.
  • FIG. 5 shows a metadata processor according to an alternative embodiment of the invention.
  • FIG. 6 shows a metadata processor according to a further alternative embodiment of the invention.
  • DETAILED DESCRIPTION OF THE INVENTION
  • A system according to an embodiment of the invention is shown in FIG. 1. An audio visual data stream (1) is input to a fingerprint generator (2) at a point in an audio visual content distribution system. The fingerprint generator (2) outputs a fingerprint stream (3) that describes the audio visual data stream (1). The fingerprint stream (3) may describe either the audio or the video elements of the audio visual data stream (1), but typically will contain information relating to both.
  • The fingerprint stream (3) comprises a sequence of fingerprints, where each member of the sequence relates to a different temporal position in the data stream (1). Typically the video element of each fingerprint is derived from a different frame of video data; and, the audio element of each fingerprint is derived from a different set of audio samples. The data rate of fingerprint stream (3) is very much less than the data rate of the audio visual data stream (1). Typically the audio component of the fingerprint stream (3) has a data rate of around 150 byte/s, and the video component of the fingerprint stream (3) has a data rate of around 500 byte/s. The derivation of the fingerprint from the audio visual data is a non-reversible process; it is not possible to re-construct the audio visual data from the fingerprint. The fingerprint can be considered a hash-function of the audio visual data such that it is highly unlikely that different audio visual data will give the same fingerprint.
  • There are many known methods of deriving fingerprints from audio and video. International patent application WO 2009/104022 (which is hereby incorporated by reference) describes how an audio fingerprint can be derived from a stream of audio samples, and how spatial and temporal video fingerprints can be derived from video frames. Standards defining audio and video fingerprints for establishing temporal synchronization between audio and video streams are being developed.
  • Returning to FIG. 1, the fingerprint stream (3) is input to a fingerprint processor (4) that derives metadata (5) from the fingerprint stream (3) and is further described below.
  • At another place in the content distribution system a second audio visual data stream (6), that is not related to the first audio visual stream (1), is input to a second fingerprint processor (7) that generates a second fingerprint stream (8) from the second audio visual data stream (6). This second fingerprint stream is also routed to the fingerprint processor (4). Other unrelated audio, video or audio visual streams from different points within the audio visual content production and distribution process can be fingerprinted and the results routed to the fingerprint processor (4). For example, the fingerprint stream (10) describing the audio visual data stream (9) is shown as a further input to the fingerprint processor (4). As the fingerprints comprise small volumes of data, the respective fingerprint streams can be conveyed to the fingerprint processor (4) over low bandwidth links; for example, narrow-band internet connections could be used.
  • The metadata (5) output from the metadata processor (4) comprises metadata describing the first and second audio visual streams (1) and (6) and any other audio visual streams whose respective fingerprint streams are input to it. Typically the fingerprint processor (4) would be situated at a central monitoring location, and its output metadata (5) would be input to a manual or automatic control system that seeks to maintain the correct operation of the audio visual content production and distribution system.
  • The operations carried out by the metadata processor (4) on one of its input fingerprint streams are illustrated in FIG. 2. An input fingerprint stream (200) comprises spatial video fingerprint data, temporal video fingerprint data, and audio fingerprint data relating to a sequence of temporal positions in the audiovisual data stream from which it was derived. Typically this sequence of temporal positions corresponds to fields of an interlaced video stream, or frames of a progressive video stream. In the following description it is assumed that a fingerprint is input for every field of the audio visual sequence.
  • A separator (201) separates out the three components of each input fingerprint of the fingerprint stream (200). The separated spatial video fingerprint stream (202) comprises respective pixel-value summations for a set of regions of each video field. This is input to a black detector (205) that compares the values with a threshold and detects the simultaneous occurrence of low values in all the regions for several consecutive fields. When this condition is detected, a Black metadata component (211) is output to a monitoring process.
  • The separated spatial video fingerprint stream (202) is also input to a test signal detector (206) that detects a sustained set of pixel-value summation values for a set of regions within each video field. The test signal detector (206) compares the regional pixel-value summations contained within each fingerprint of the fingerprint sequence (202) with previously-derived regional pixel-value summations for known test signals. The comparison results are compared with one or more thresholds to identify near equivalence of the values in the fingerprints with the respective values for known test signals. If a set of values closely corresponding to values for a particular known test signal, colour bars for example, is found in a consecutive sequence of fingerprints, a test signal metadata component (212) that identifies the presence of the particular test signal is output.
  • The separated temporal video fingerprint stream (203) is input to a still-image detector (207). The separated temporal video fingerprint stream (203) typically comprises a measure of inter-field differences between pixel-value summations for a set of regions within each video field. An example is a sum of the sums of inter-field differences for a set of regions within the frame, evaluated between a current field and a previous field. If the fingerprint contains an inter-frame difference value, or if an inter-frame difference can be derived from the fingerprint, then this is used. If a sustained low-value inter-field or inter-frame difference measure is found in a consecutive sequence of fingerprints, a still-image metadata component (213) that identifies lack of motion is output.
  • The separated temporal video fingerprint stream (203) is also input to a shot-change detector (208), which identifies isolated high values of the temporal video fingerprint by comparing the respective value differences between a fingerprint and its closely preceding and succeeding fingerprints with a threshold. If the temporal fingerprint for a field is significantly greater than the corresponding fingerprints for preceding and succeeding fields, then that field is identified as the first field of a new shot, and it is identified in a shot-change metadata output (214). A graph of temporal fingerprint value versus time for a video sequence containing shot changes is shown in FIG. 3. The isolated peaks (31) to (36) correspond to shot-changes.
  • The separated temporal video fingerprint stream (203) is also analysed to detect ‘film cadence’ in a film cadence detector (209). FIG. 4 shows examples of sequences of temporal video fingerprint values for three different film cadences. The sequence of temporal fingerprints for succeeding fields is analysed in the film cadence detector (209), and the sequence of differences between the fingerprints is identified. If successive pairs of temporal fingerprints from adjacent fields have similar values (i.e. the differences are less than a threshold), as shown in FIG. 4 a, then it is inferred that each pair comes from a new film frame; this is commonly known a 2:2 film cadence. If two pairs of similar values are followed by a significantly different value in a continuing sequence, as shown in FIG. 4 b, then 3:2 film cadence, in which the ratio of the film frame rate to the video field rate is 2:5, is identified. And, if there is no pattern of similarity between the temporal fingerprints for succeeding fields, as shown in FIG. 4 c, then video cadence is identified.
  • The film cadence detector (209) detects the pattern of changes between the fingerprints for succeeding fields by a known method, such as correlation of sequences of inter-fingerprint difference values with candidate sequences of differences. Metadata indicating detected video cadence (215), detected 2:2 film cadence (216) or detected 3:2 film cadence (217) is output.
  • The separated audio fingerprint stream (204) is input to a silence detector (210). Typical audio fingerprints are derived from the magnitudes of a sequence of adjacent audio samples. When the audio is silent the sample magnitudes are small and a sequence of low-value fingerprints results. When a sustained sequence of audio fingerprint values less than a low-value threshold is detected by the silence detector (210), it outputs silence metadata (218).
  • A further audio visual fingerprint analysis process is shown in FIG. 5. A sequence of spatial or temporal video fingerprints (500), corresponding to fields or frames of a video or audio visual sequence, is input to a rolling window selector (501), which selects and outputs a stream of sets of adjacent fingerprint values. Typically each set corresponds to one or two seconds of video, and the sets overlap each other by a few hundred milliseconds.
  • Each set of fingerprint values is converted, in a histogram generator (502), to a histogram giving the respective frequencies of occurrence of values, or ranges of values, within the set. The sequence of histograms from the histogram generator (502), corresponding the sequence of adjacent fingerprint values from the window selector (501), is analysed statistically in a moment processor (503) and an entropy processor (504).
  • The moment processor (503) determines known statistical parameters of each histogram: The mean (or first moment); the variance (or second moment); the skew (or third moment); and the kurtosis (or fourth moment). The derivation of these known dimensionless parameters of the distribution of values within a set of values will not be described here as it is well-known to those skilled in the art.
  • The entropy processor (504) determines the entropy E, or ‘distinctiveness’ of each histogram. A suitable measure is given by the following equation:

  • E=−Σp i log(p i)
      • Where: pi is the number of occurrences of fingerprint value i divided by the number of fingerprint values in the set; and,
      • The summation is made over all values of i that occur in the set.
  • The stream of sets of dimensionless statistical parameters (505) from the moment processor (503), and the stream of entropy values (506) from the entropy processor (504) are input to a classifier (507) that compares each of its input data sets with reference data sets corresponding to known types of audiovisual content. The output from the classifier (507) is metadata (508) that describes the type of audio visual content from which the fingerprint value sequence (500) was derived.
  • Typically the output of the classifier (507) is a weighted sum of the outputs from a number of different, known comparison functions, where the weights and the functions have been previously selected in a known ‘training’ process. In such prior training, candidate sets of comparison functions are applied iteratively to sets of statistical data (505) and entropy data (506) that have been derived from analysis (as shown in FIG. 5) of fingerprint data from known types of audio visual content. The weights and comparison functions are selected during this training so as to obtain the best agreement between the result of the weighted sum of comparisons, and the known content type of the respective training data set. The classifier (507) uses a set of comparison functions and respective weights determined in a prior training process to identify when its input corresponds to a particular member of a set of reference data sets that corresponds with a particular type of audio visual content.
  • Typically the following types of audio visual stream are used as training data, and are identified by the classifier (507):
      • Specific sports
      • Studio news presentation
      • ‘Talking heads’
      • Episodic drama
      • Film/movie drama
      • Commercials
      • Cartoon animation
      • Credit sequences
      • Loss of signal conditions
      • Recorder ‘shuttle’ modes
  • Other content types may be more suitable for the control and monitoring of a particular audio visual production or distribution process.
  • Another embodiment of the invention is shown in FIG. 6. A sequence of audio or video fingerprint values (600) is separated into sets of rolling windows by a rolling window selector (601) that operates in the same way as the previously-described window selector (501). Temporally-ordered, windowed sets of adjacent fingerprint values are transformed from the time domain to the frequency domain in a transform processor (602), whose output comprises a stream of sets of spectral components, one set for each temporal position of the rolling window applied by the window selector (601). Typically the transform processor (602) uses the well-known Fourier transform, but other time-domain to frequency-domain conversions could be used.
  • The stream of sets of frequency components (603) from the transform processor (602) is input to a classifier (604) that operates in the same way as the above-described classifier (507) to recognise the spectral characteristics of known types of audio visual content. Metadata (605) that describes the type of audio visual content from which the fingerprint value sequence (600) was derived is output from the classifier (604).
  • Some audio fingerprints, for example the ‘bar code’ audio signature described in international patent application WO 2009/104022, comprise a sequence of one-bit binary values. These fingerprints can conveniently be described by run-length coding, in which a sequence of run-length values indicates counts of succeeding identical fingerprint values. This is a well-known method of data compression that represents a sequence of consecutive values by a single descriptor and run-length value. In the case of binary data, the descriptor is not required, as each run-length value represents a change of state of the binary data.
  • Run-length values for rolling windows of a fingerprint sequence can be histogrammed and the histograms of the frequencies of occurrence of run-length values, or ranges of run-length values used to identify characteristics of the material from which the fingerprints were derived.
  • The reliability of all the above-described methods of extracting metadata from fingerprint data can be improved by applying a temporal low-pass filter to the derived metadata. Simple recursive filters, a running average for example, are suitable. However, there is a trade-off between reliability and speed of response. The required speed of response is different for different types of metadata. Some parameters describe a single frame, for example a black frame identifier. Other parameters relate to a short sequence of frames, for example film cadence. Yet others relate to hundreds, or even thousands, of frames, for example type of content. The temporal filters applicable to these different types of metadata will have different bandwidths.
  • Changes in the values of metadata derived by the methods described in this specification contain useful information which can be used to derive higher level metadata. For example, the frequency of occurrence of shot changes can be used to infer content type.
  • Several different methods of analysing fingerprint data have been described. A metadata inference process according to the invention can use one or more of these methods; not all elements of a particular fingerprint need be analysed.
  • Processing of spatial video fingerprints, temporal video fingerprints and audio fingerprints has been described. These methods of obtaining metadata from fingerprint data are applicable to one type of fingerprint, or combinations of different types of fingerprint derived from the same temporal position within an audio visual content stream. The relationship between different fingerprint types derived from the same content can be used to determine metadata applicable to that content.
  • Typically the temporal position of an available audio fingerprint will have a fixed relationship to the temporal position of an associated available video fingerprint for the same content stream at the same point in an audio visual content production or distribution process. In this case combination of the results video fingerprint analysis according to the invention with the results of audio fingerprint analysis according to the invention will give a more reliable determination of metadata for the audio visual sequence than would be achieved by analysis of the audio or video fingerprints in isolation.
  • The principles of the invention can be applied to many different types of audio video or audio visual fingerprint. Audio and/or video data may be sub-sampled prior to generating the applicable fingerprint or fingerprints. Video fingerprints may be derived from fields or frames.

Claims (17)

1. A method of managing audio visual, audio or visual content, comprising the steps of:
receiving a stream of fingerprints, derived in a fingerprint generator by an irreversible data reduction process from respective temporal regions within a particular audio visual, audio or visual content stream, at a fingerprint processor that is physically separate from the fingerprint generator via a communication network; and
processing said fingerprints in the fingerprint processor to generate metadata which is not directly encoded in the fingerprints, with one or more processes selected from the group consisting of:
detecting the sustained occurrence of low values of an audio fingerprint to generate metadata indicating silence;
comparing the pattern of differences between temporally succeeding values of a fingerprint with expected patterns of film cadence to generate metadata indicating a film cadence; and
comparing differences between temporally succeeding values of a fingerprint with a threshold to generate metadata indicating a still image or freeze frame.
2. The method according to claim 1, wherein said communication network comprises the Internet.
3. The method according to claim 1, wherein an audio fingerprint stream has a data rate of less than about 500 byte/s per audio channel.
4. The method according to claim 1, wherein an audio fingerprint stream has a data rate of less than about 250 byte/s per audio channel.
5. The method according to claim 1, wherein a video fingerprint stream has a data rate of less than about 500 byte per field.
6. The method according to claim 1, wherein a video fingerprint stream has a data rate of less than about 200 byte per field.
7. The method according to claim 1, wherein said content comprises a video stream of video frames and wherein a fingerprint is generated for substantially every frame in the video stream.
8. A method of managing audio visual, audio or visual content, comprising the steps of:
receiving a stream of fingerprints, derived in a fingerprint generator by an irreversible data reduction process from respective temporal regions within a particular audio visual, audio or visual content stream, at a fingerprint processor that is physically separate from the fingerprint generator via a communication network; and
processing said fingerprints in the fingerprint processor to generate metadata which is not directly encoded in the fingerprints; wherein said processing includes
windowing the stream of fingerprints with a time window,
deriving frequencies of occurrence of particular fingerprint values or ranges of fingerprint values within each time window,
determining statistical moments or entropy values of said frequencies of occurrence,
comparing said statistical moments or entropy values with expected values for particular types of content, and
generating metadata representing the type of the audio visual, audio or visual content.
9. The method according to claim 8, wherein said statistical moment comprises one or more of the mean; variance; skew or kurtosis of said frequencies of occurrence.
10. The method according to claim 8, wherein said communication network comprises the Internet.
11. The method according to claim 8, wherein a video fingerprint stream has a data rate of less than about 500 byte per field.
12. The method according to claim 8, wherein a video fingerprint stream has a data rate of less than about 200 byte per field.
13. The method according to claim 8, wherein said content comprises a video stream of video frames and wherein a fingerprint is generated for substantially every frame in the video stream.
14. An apparatus for use in managing audio visual, audio or visual content, the apparatus comprising:
a fingerprint processor configured to receive via a communication network a stream of fingerprints derived in a fingerprint generator that is physically separate from the fingerprint processor by an irreversible data reduction process from respective temporal regions within a particular audio visual, audio or visual content stream, at a fingerprint processor generator; the fingerprint processor including
a window unit configure to receive said stream of fingerprints and apply a time window,
a frequency of occurrence histogram unit configured to derive the frequencies of occurrence of particular fingerprint values in each time window,
a statistical moment unit configured to derive statistical moments of said frequencies of occurrence, and
a classifier configured to generate from said statistical moments metadata representing the type of the audio visual, audio or visual content.
15. The apparatus according to claim 14, further comprising an entropy unit configured to derive entropy values for histograms of frequencies of occurrence and wherein said classifier is configured to generate said metadata representing the type of the audio visual, audio or visual content additionally from said entropy values.
16. A non-transitory computer program product adapted to cause programmable apparatus to implement a method of managing audio visual, audio or visual content, comprising the steps of:
receiving a stream of fingerprints, derived in a fingerprint generator by an irreversible data reduction process from respective temporal regions within a particular audio visual, audio or visual content stream, at a fingerprint processor that is physically separate from the fingerprint generator via a communication network; and
processing said fingerprints in the fingerprint processor to generate metadata which is not directly encoded in the fingerprints, with one or more processes selected from the group consisting of,
detecting the sustained occurrence of low values of an audio fingerprint to generate metadata indicating silence,
comparing the pattern of differences between temporally succeeding values of a fingerprint with expected patterns of film cadence to generate metadata indicating a film cadence, and
comparing differences between temporally succeeding values of a fingerprint with a threshold to generate metadata indicating a still image or freeze frame.
17. A non-transitory computer program product adapted to cause programmable apparatus to implement a method of managing audio visual, audio or visual content, comprising the steps of:
receiving a stream of fingerprints, derived in a fingerprint generator by an irreversible data reduction process from respective temporal regions within a particular audio visual, audio or visual content stream, at a fingerprint processor that is physically separate from the fingerprint generator via a communication network; and
processing said fingerprints in the fingerprint processor to generate metadata which is not directly encoded in the fingerprints; wherein said processing includes
windowing the stream of fingerprints with a time window,
deriving frequencies of occurrence of particular fingerprint values or ranges of fingerprint values within each time window,
determining statistical moments or entropy values of said frequencies of occurrence,
comparing said statistical moments or entropy values with expected values for particular types of content, and
generating metadata representing the type of the audio visual, audio or visual content.
US14/623,354 2014-02-17 2015-02-16 Method and apparatus for managing audio visual, audio or visual content Abandoned US20150237341A1 (en)

Priority Applications (2)

Application Number Priority Date Filing Date Title
US15/459,860 US10219033B2 (en) 2014-02-17 2017-03-15 Method and apparatus of managing visual content
US16/283,491 US10893323B2 (en) 2014-02-17 2019-02-22 Method and apparatus of managing visual content

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
GB1402775.9 2014-02-17
GB1402775.9A GB2523311B (en) 2014-02-17 2014-02-17 Method and apparatus for managing audio visual, audio or visual content

Related Child Applications (1)

Application Number Title Priority Date Filing Date
US15/459,860 Continuation US10219033B2 (en) 2014-02-17 2017-03-15 Method and apparatus of managing visual content

Publications (1)

Publication Number Publication Date
US20150237341A1 true US20150237341A1 (en) 2015-08-20

Family

ID=50440289

Family Applications (3)

Application Number Title Priority Date Filing Date
US14/623,354 Abandoned US20150237341A1 (en) 2014-02-17 2015-02-16 Method and apparatus for managing audio visual, audio or visual content
US15/459,860 Expired - Fee Related US10219033B2 (en) 2014-02-17 2017-03-15 Method and apparatus of managing visual content
US16/283,491 Active US10893323B2 (en) 2014-02-17 2019-02-22 Method and apparatus of managing visual content

Family Applications After (2)

Application Number Title Priority Date Filing Date
US15/459,860 Expired - Fee Related US10219033B2 (en) 2014-02-17 2017-03-15 Method and apparatus of managing visual content
US16/283,491 Active US10893323B2 (en) 2014-02-17 2019-02-22 Method and apparatus of managing visual content

Country Status (4)

Country Link
US (3) US20150237341A1 (en)
EP (1) EP2908269A3 (en)
CN (2) CN110443108A (en)
GB (1) GB2523311B (en)

Cited By (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20160353182A1 (en) * 2013-12-27 2016-12-01 Thomson Licensing Method for synchronising metadata with an audiovisual document by using parts of frames and a device for producing such metadata
US20170309298A1 (en) * 2016-04-20 2017-10-26 Gracenote, Inc. Digital fingerprint indexing
US20180184179A1 (en) * 2015-06-30 2018-06-28 British Broadcasting Corporation Audio-Video Content Control
CN111104552A (en) * 2019-12-24 2020-05-05 浙江大学 Method for predicting movie scoring category based on movie structural information and brief introduction
US11417099B1 (en) 2021-11-08 2022-08-16 9219-1568 Quebec Inc. System and method for digital fingerprinting of media content

Families Citing this family (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
KR102414783B1 (en) 2017-11-01 2022-06-30 삼성전자주식회사 Electronic apparatus and controlling method thereof
CN109640169B (en) 2018-11-27 2020-09-22 Oppo广东移动通信有限公司 Video enhancement control method and device and electronic equipment
CN112887678A (en) * 2021-01-28 2021-06-01 泰豪信息技术有限公司 Supervision method, system, storage medium and equipment for supervision personnel on duty

Citations (10)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US6055018A (en) * 1997-11-04 2000-04-25 Ati Technologies, Inc. System and method for reconstructing noninterlaced captured content for display on a progressive screen
US20020120925A1 (en) * 2000-03-28 2002-08-29 Logan James D. Audio and video program recording, editing and playback systems using metadata
GB2444094A (en) * 2006-11-22 2008-05-28 Half Minute Media Ltd Identifying repeating video sections by comparing video fingerprints from detected candidate video sequences
US20110222787A1 (en) * 2008-02-28 2011-09-15 Stefan Thiemert Frame sequence comparison in multimedia streams
US20110289114A1 (en) * 2011-08-08 2011-11-24 Lei Yu System and method for auto content recognition
US20130254793A1 (en) * 2011-08-30 2013-09-26 Clear Channel Management Services, Inc. Broadcast Source Identification Based on Matching Via Bit Count
US20140259041A1 (en) * 2013-03-05 2014-09-11 Google Inc. Associating audio tracks of an album with video content
US20140277641A1 (en) * 2013-03-15 2014-09-18 Facebook, Inc. Managing Silence In Audio Signal Identification
US9159327B1 (en) * 2012-12-20 2015-10-13 Google Inc. System and method for adding pitch shift resistance to an audio fingerprint
US9275427B1 (en) * 2013-09-05 2016-03-01 Google Inc. Multi-channel audio video fingerprinting

Family Cites Families (20)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US6469749B1 (en) * 1999-10-13 2002-10-22 Koninklijke Philips Electronics N.V. Automatic signature-based spotting, learning and extracting of commercials and other video content
CA2310769C (en) * 1999-10-27 2013-05-28 Nielsen Media Research, Inc. Audio signature extraction and correlation
WO2005041109A2 (en) * 2003-10-17 2005-05-06 Nielsen Media Research, Inc. Methods and apparatus for identifiying audio/video content using temporal signal characteristics
US10032192B2 (en) * 2003-12-23 2018-07-24 Roku, Inc. Automatic localization of advertisements
GB2425431A (en) * 2005-04-14 2006-10-25 Half Minute Media Ltd Video entity recognition in compressed digital video streams
US20060271947A1 (en) * 2005-05-23 2006-11-30 Lienhart Rainer W Creating fingerprints
CN101278568B (en) * 2005-08-16 2010-12-15 尼尔森(美国)有限公司 Display device on/off detection methods and apparatus
US20100039517A1 (en) * 2007-01-03 2010-02-18 Koninklijke Philips Electronics N.V. Film cadence detection
EP2168061A1 (en) * 2007-06-06 2010-03-31 Dolby Laboratories Licensing Corporation Improving audio/video fingerprint search accuracy using multiple search combining
US8238669B2 (en) * 2007-08-22 2012-08-07 Google Inc. Detection and classification of matches between time-based media
GB2457694B (en) * 2008-02-21 2012-09-26 Snell Ltd Method of Deriving an Audio-Visual Signature
US9510044B1 (en) * 2008-06-18 2016-11-29 Gracenote, Inc. TV content segmentation, categorization and identification and time-aligned applications
US8635211B2 (en) * 2009-06-11 2014-01-21 Dolby Laboratories Licensing Corporation Trend analysis in content identification based on fingerprinting
US9047371B2 (en) * 2010-07-29 2015-06-02 Soundhound, Inc. System and method for matching a query against a broadcast stream
US8515990B2 (en) * 2010-11-19 2013-08-20 Lg Electronics Inc. Mobile terminal and method of managing video using metadata therein
US9093120B2 (en) * 2011-02-10 2015-07-28 Yahoo! Inc. Audio fingerprint extraction by scaling in time and resampling
US20130318114A1 (en) * 2012-05-13 2013-11-28 Harry E. Emerson, III Discovery of music artist and title by broadcast radio receivers
EP2868109B1 (en) * 2012-06-28 2016-08-17 Google, Inc. Generating a sequence of audio fingerprints at a set top box
US9460204B2 (en) * 2012-10-19 2016-10-04 Sony Corporation Apparatus and method for scene change detection-based trigger for audio fingerprinting analysis
KR102587635B1 (en) * 2017-11-22 2023-10-10 피피지 인더스트리즈 오하이오 인코포레이티드 patterning paste

Patent Citations (10)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US6055018A (en) * 1997-11-04 2000-04-25 Ati Technologies, Inc. System and method for reconstructing noninterlaced captured content for display on a progressive screen
US20020120925A1 (en) * 2000-03-28 2002-08-29 Logan James D. Audio and video program recording, editing and playback systems using metadata
GB2444094A (en) * 2006-11-22 2008-05-28 Half Minute Media Ltd Identifying repeating video sections by comparing video fingerprints from detected candidate video sequences
US20110222787A1 (en) * 2008-02-28 2011-09-15 Stefan Thiemert Frame sequence comparison in multimedia streams
US20110289114A1 (en) * 2011-08-08 2011-11-24 Lei Yu System and method for auto content recognition
US20130254793A1 (en) * 2011-08-30 2013-09-26 Clear Channel Management Services, Inc. Broadcast Source Identification Based on Matching Via Bit Count
US9159327B1 (en) * 2012-12-20 2015-10-13 Google Inc. System and method for adding pitch shift resistance to an audio fingerprint
US20140259041A1 (en) * 2013-03-05 2014-09-11 Google Inc. Associating audio tracks of an album with video content
US20140277641A1 (en) * 2013-03-15 2014-09-18 Facebook, Inc. Managing Silence In Audio Signal Identification
US9275427B1 (en) * 2013-09-05 2016-03-01 Google Inc. Multi-channel audio video fingerprinting

Cited By (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20160353182A1 (en) * 2013-12-27 2016-12-01 Thomson Licensing Method for synchronising metadata with an audiovisual document by using parts of frames and a device for producing such metadata
US20180184179A1 (en) * 2015-06-30 2018-06-28 British Broadcasting Corporation Audio-Video Content Control
US10701459B2 (en) * 2015-06-30 2020-06-30 British Broadcasting Corporation Audio-video content control
US20170309298A1 (en) * 2016-04-20 2017-10-26 Gracenote, Inc. Digital fingerprint indexing
CN111104552A (en) * 2019-12-24 2020-05-05 浙江大学 Method for predicting movie scoring category based on movie structural information and brief introduction
US11417099B1 (en) 2021-11-08 2022-08-16 9219-1568 Quebec Inc. System and method for digital fingerprinting of media content
US11783583B2 (en) 2021-11-08 2023-10-10 9219-1568 Quebec Inc. System and method for digital fingerprinting of media content

Also Published As

Publication number Publication date
CN104853244B (en) 2019-09-17
EP2908269A3 (en) 2015-12-16
US20170188091A1 (en) 2017-06-29
US10219033B2 (en) 2019-02-26
CN110443108A (en) 2019-11-12
CN104853244A (en) 2015-08-19
EP2908269A2 (en) 2015-08-19
US10893323B2 (en) 2021-01-12
GB2523311B (en) 2021-07-14
US20190191213A1 (en) 2019-06-20
GB2523311A (en) 2015-08-26
GB201402775D0 (en) 2014-04-02

Similar Documents

Publication Publication Date Title
US10893323B2 (en) Method and apparatus of managing visual content
US9536545B2 (en) Audio visual signature, method of deriving a signature, and method of comparing audio-visual data background
US8860883B2 (en) Method and apparatus for providing signatures of audio/video signals and for making use thereof
US8717499B2 (en) Audio video offset detector
US20160337691A1 (en) System and method for detecting streaming of advertisements that occur while streaming a media program
KR101595875B1 (en) Method and apparatus for evaluating multimedia quality
US8212939B2 (en) Non-intrusive determination of an objective mean opinion score of a video sequence
US11706500B2 (en) Computing system with content-characteristic-based trigger feature
US20130002887A1 (en) System And Method For Automated Set-Top Box Testing Via Configurable Event Time Measurements
Staelens et al. Viqid: A no-reference bit stream-based visual quality impairment detector
RU2642402C1 (en) Method of detecting duplicate video frames
CA3168392A1 (en) Real-time latency measurement of video streams
US7233348B2 (en) Test method
Grbić et al. Real-time video freezing detection for 4K UHD videos
GB2487499A (en) Audio-Visual Signature, Method of Deriving a Signature, and Method of Comparing Audio-Visual Data
Pribula et al. Real-time video sequences matching using the spatio-temporal fingerprint
KR101361442B1 (en) Video transcoding optimizations using high frequency frame analysis
KR20170095039A (en) Apparatus for editing contents for seperating shot and method thereof
Kyperountas et al. Scene change detection using audiovisual clues

Legal Events

Date Code Title Description
AS Assignment

Owner name: SNELL LIMITED, UNITED KINGDOM

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:DIGGINS, JONATHAN;REEL/FRAME:035408/0416

Effective date: 20150407

STCB Information on status: application discontinuation

Free format text: ABANDONED -- FAILURE TO RESPOND TO AN OFFICE ACTION