US5724100A - Method and apparatus for detecting scene-cuts in a block-based video coding system - Google Patents

Method and apparatus for detecting scene-cuts in a block-based video coding system Download PDF

Info

Publication number
US5724100A
US5724100A US08/606,613 US60661396A US5724100A US 5724100 A US5724100 A US 5724100A US 60661396 A US60661396 A US 60661396A US 5724100 A US5724100 A US 5724100A
Authority
US
United States
Prior art keywords
macroblock
input
variance
macroblocks
image
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Expired - Lifetime
Application number
US08/606,613
Inventor
Prashanth Kuchibhotla
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
MediaTek Inc
Original Assignee
David Sarnoff Research Center Inc
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by David Sarnoff Research Center Inc filed Critical David Sarnoff Research Center Inc
Priority to US08/606,613 priority Critical patent/US5724100A/en
Assigned to DAVID SARNOFF RESEARCH CENTER, INC. reassignment DAVID SARNOFF RESEARCH CENTER, INC. ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: KUCHIBHOTLA, PRASHANTH
Priority to KR10-1998-0706667A priority patent/KR100382676B1/en
Priority to JP9530283A priority patent/JP2000505614A/en
Priority to PCT/US1997/002563 priority patent/WO1997031485A1/en
Priority to DE69738901T priority patent/DE69738901D1/en
Priority to EP97907683A priority patent/EP0885526B1/en
Publication of US5724100A publication Critical patent/US5724100A/en
Application granted granted Critical
Assigned to SARNOFF CORPORATION reassignment SARNOFF CORPORATION MERGER (SEE DOCUMENT FOR DETAILS). Assignors: DAVID SARNOFF RESEARCH CENTER, INC.
Assigned to MEDIATEK, INC. reassignment MEDIATEK, INC. ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: SARNOFF CORPORATION
Anticipated expiration legal-status Critical
Expired - Lifetime legal-status Critical Current

Links

Images

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/10Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
    • H04N19/134Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the element, parameter or criterion affecting or controlling the adaptive coding
    • H04N19/142Detection of scene cut or scene change
    • GPHYSICS
    • G11INFORMATION STORAGE
    • G11BINFORMATION STORAGE BASED ON RELATIVE MOVEMENT BETWEEN RECORD CARRIER AND TRANSDUCER
    • G11B27/00Editing; Indexing; Addressing; Timing or synchronising; Monitoring; Measuring tape travel
    • G11B27/10Indexing; Addressing; Timing or synchronising; Measuring tape travel
    • G11B27/19Indexing; Addressing; Timing or synchronising; Measuring tape travel by using information detectable on the record carrier
    • G11B27/28Indexing; Addressing; Timing or synchronising; Measuring tape travel by using information detectable on the record carrier by using information signals recorded by the same method as the main recording
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/10Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
    • H04N19/169Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding
    • H04N19/179Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being a scene or a shot
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/50Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding
    • H04N19/503Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding involving temporal prediction
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/85Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using pre-processing or post-processing specially adapted for video compression
    • H04N19/87Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using pre-processing or post-processing specially adapted for video compression involving scene cut or scene change detection in combination with video compression
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N5/00Details of television systems
    • H04N5/14Picture signal circuitry for video frequency region
    • H04N5/147Scene change detection

Definitions

  • the invention relates to a block-based video coding technique and, more particularly, the invention relates to a method and apparatus for detecting scene cuts in a video sequence within a block-based video coding system.
  • Block-based video coding systems typically use a coding technique that takes advantage of both spatial and temporal redundancy within an image (intra-picture) and between images (inter-picture) within a sequence of images.
  • Such block-based image coding systems including those that utilize the well-known Moving Picture Experts Group (MPEG) standard of video coding); namely, ISO/IEC international standards 11172-2 (1994) (generally referred to as MPEG-1) and 13818-2 (Jan. 20, 1995 draft) (generally referred to as MPEG-2).
  • MPEG Moving Picture Experts Group
  • block-based coding techniques assume that the sequential pictures within an input video sequence contain substantially similar information, i.e., the imaged scene changes very little from picture to picture. A scene cut occurring in the picture sequence violates the underlying assumption for efficient coding. Consequently, after a scene change (scene cut), a block-based coding technique must use a substantial number of bits to code the first picture following the scene change. Because the number of bits available to code any one image is typically limited, a scene cut can cause substantial errors in the coding and lead to substantial distortion of the decoded picture.
  • the disadvantages heretofore associated with the prior art are overcome by the present invention of a scene cut detector for incorporation into a conventional block-based video coding system.
  • the scene cut detector compares predicted macroblocks from a predicted image to input macroblocks from an input image on a macroblock-by-macroblock basis to generate a residual macroblock representing the difference between each predicted macroblock and each input macroblock.
  • a variance for each residual macroblock and a variance for each input macroblock is computed after each comparison.
  • the variances of the residual macroblock and the input macroblock are compared to a decision function. Depending upon the outcome of the decision function comparison, a counter is incremented.
  • the scene cut detector repeats this process until each macroblock in the predicted image is compared to each input macroblock. If the count value ever exceeds a threshold level while a input image is being processed, the scene cut detector identifies the input image as being within a new scene and, in response, sets a scene cut indicator flag.
  • FIG. 1 depicts a block diagram of a block-based coding system incorporating a scene cut detector of the present invention
  • FIG. 2 depicts a detailed block diagram of a scene cut detector of the present invention.
  • FIG. 3 depicts a graph of a decision function used by the IID estimator of the present invention.
  • FIG. 1 depicts a block diagram of a block-based coding system 100 (specifically, an MPEG encoder) incorporating the present invention.
  • the input signal, at port 102, to the system is a preprocessed image that has been partitioned into a plurality of blocks, where the blocks are sequentially provided as an input to the system.
  • these blocks of pixels are commonly known as macroblocks, e.g., a 16 ⁇ 16 pixel block.
  • macroblock is intended to describe a block of pixels of any size that is used for the basis of motion compensation.
  • the system computes, from the system output signal, a series of predicted macroblocks (P). Each predicted macroblock is illustratively produced by decoding the output signal, at port 104, just as the receiver of the transmitted output signal would decode the received signal.
  • Subtractor 106 generates, on path 107, a residual signal (also referred to in the art as simply the residual or the residual macroblock) by subtracting the predicted macroblock from the input macroblock.
  • the input macroblock is said to be motion compensated.
  • the difference between the predicted macroblock and the input macroblock is substantial, the residuals are difficult to code. Consequently, the system is better off directly coding the input macroblock in lieu of coding the motion compensated residual macroblock.
  • This selection is known as a selection of the cooling mode. Coding the input macroblock I is referred to as intra-coding, while coding the residuals is referred to as inter-coding. The selection between these two modes is known as the Intra-Inter-Decision (IID).
  • the IID is made by the IID circuit 110, which sets the coding mode switch 108.
  • the IID is computed by first computing the variance of the residual macroblock (Var R) and the variance of the input macroblock (Var I).
  • the coding decision is based on these values. There are several functions that can be used to make this decision. For example, if Var R is less than Var I, the IID selects the Inter-mode. Conversely, if Var I is less than Var R, the IID selects the Intra-mode.
  • the selected block is processed in a discrete cosine transform (DCT) block 112.
  • the DCT produces coefficients representing the input signal to the DCT.
  • the quantizer 114 quantizes the coefficients to produce the output block at port 104.
  • the rate control block 116 controls the quantization scale (step size) used to quantize the coefficients.
  • the encoder needs access to the decoded images.
  • the quantizer 114 output is passed through both the inverse quantizer 118 and inverse DCT 120.
  • the output of the inverse DCT is ideally identical to the input to the DCT 112.
  • the decoded macroblock is produced by summing the output of the inverse DCT and the predicted macroblock.
  • the decoded macroblock is simply the output of the inverse DCT.
  • the decoded macroblocks are then stored in the frame store 124.
  • the frame store accumulates a plurality of these "reconstructed" macroblocks that constitute an entire reconstructed frame of image information.
  • the reconstructed frame is used by the motion vector predictor 126 to produce motion vectors that are used in generating predicted macroblocks for forthcoming input images.
  • the motion vector predictor 126 comprises three components: a full-pel motion estimator 128, a half-pel motion estimator 130, and a motion mode block 132.
  • the full-pel motion estimator 128 is a "coarse" motion vector generator that searches for a coarse match between a macroblock in a previous image and the present input macroblock.
  • the previous image is referred to as an anchor image.
  • the anchor image is what is known as an I or P frame within an image sequence known as a Group Of Pictures (GOP).
  • the motion vector is a vector representing the relative position where a coarse match was found between the two macroblocks.
  • the coarse motion vector generator produces a motion vector that is accurate to one picture element (pel).
  • the accuracy of the full-pel motion estimator is improved in the half-pel motion estimator.
  • the half-pel estimator uses the full-pel motion vectors and the reconstructed macroblocks from the frame store 124 to compute motion vectors to half-pel accuracy.
  • the half-pel motion vectors are then sent to the motion modes block 132. Typically, there are multiple motion vectors related to each macroblock.
  • the modes block 132 selects the best motion vector for representing motion for each input macroblock.
  • the full-pel estimator is a computationally intensive task compared to the half-pel estimator. For this reason, in several implementations, it is computed separately on dedicated hardware. Often, all the full-pel motion vectors are computed before half-pel processing begins.
  • the MPEG encoder system discussed above is a conventional system that is available as a set of integrated circuits as model L64120 from LSI Logic, Inc. of Milpitas, Calif. Importantly, this MPEG encoder stores an entire frame of full-pel motion vectors before the half-pel estimator begins operation.
  • the concepts of Motion Estimation and Motion Compensation are based on an underlying assumption that the current picture is not very different from a previously occurring picture (the anchor image).
  • the anchor pictures are substantially different from the current picture.
  • the predicted macroblocks are very inaccurate and the residuals are large.
  • the IID selects the input macroblock (intra-mode) for coding in lieu of coding the residuals (inter-mode). It should be noted that this coding decision occurs even in the case where there is no scene change, and that a normally coded picture may contain a mixture of intra- and inter-coded macroblocks.
  • the scene cut detector of the present invention analyzes all the macroblocks in a picture and then determines whether a scene cut has occurred. This is accomplished by counting the number of intra-coded macroblocks and comparing the count to a threshold level. Specifically, if the percentage of I coded macroblocks in any given frame exceeds the threshold level, that frame is deemed to follow a scene cut.
  • the actual IID decision is made after the half-pel motion vectors are generated and best motion vector is chosen. Since the full-pel estimator 128 generates motion vectors for the entire frame before the first macroblock is coded by the encoder, the inventive scene cut detector apparatus 134 monitoring these full-pel results is able to produce an IID estimate for all macroblocks, i.e., an estimate that the IID would make when it does analyze the residuals.
  • the scene cut detector comprises an IID estimator 136 connected in series to an intra-code macroblock counter 138. The counter 138 generates a scene cut decision (a flag) indicating that the scene cut detector has determined that a scene cut has occurred.
  • FIG. 2 depicts a detailed block diagram of the inventive scene cut detector 134 of the MPEG encoder 100.
  • the full-pel motion estimator 128 is provided the I macroblocks as well as the appropriate I or P anchor image in which the predicted macroblocks are found.
  • the anchor images are stored in the frame memory 140.
  • the full-pel motion vector generator 141 generates the motion vectors for each input macroblock using one of many well-known methods including those disclosed in U.S. Pat. No. 5,351,095, issued Sep. 27, 1994, incorporated herein by reference, and U.S. patent application Ser. No. 08/300,023, filed Sep. 2, 1994, incorporated herein by reference.
  • the motion compensator 145 uses the full-pel motion vectors and the anchor images from the anchor image store 140, the motion compensator 145 generates the predicted macroblock (P) for each input macroblock (I).
  • the input macroblocks (path 120) and the predicted image macroblocks (P) form inputs to the scene cut detector 134.
  • the IID estimator 136 computes full-pel residual by subtracting (subtractor 142) the predicted macroblock from the input macroblock.
  • the IID estimator then computes, using the variance blocks 146 and 148, the variance of the input macroblocks (Var I) and the variance of the full-pel residuals (Var R).
  • the IID circuit 150 then makes its IID estimate based on these variances.
  • the counter 152 counts the number of intra-mode decisions and compares, in block 154, the count number to a threshold count value. If, while processing a given picture, the count number ever exceeds the threshold, the intra MB counter generates a scene cut decision flag.
  • FIG. 3 depicts a graph 300 of a typical decision function used by the IID estimator in comparing Var R and Var I.
  • the simplest decision function is linear function 302. This function is implemented using a comparator such that when Var I is less than Var R, the IID estimates is in intra-mode (region 310). Conversely, when Var R is less than Vat I, the IID estimate is inter-mode (region 312).
  • non-linear function 304 represents a more typical function. This function is vertical at a particular value 306 of Var R and then is linear. In operation, macroblocks with relatively small values of Var R are coded using the inter-mode. Any value of Var R greater than value 306 is compared using the direct comparison of function 302.
  • a macroblock having Var R value that is only slightly greater than the Var I value e.g., a point lying near, but just beneath the curve 302
  • the IID of the MPEG encoder would use inter-mode coding, yet the IID estimation in the scene cut detector would estimate that intra-mode coding would be used.
  • function 308 is generally used. Function 308 is similar in form to function 304; however, the function 308 is shifted slightly below function 304. Thus, incorrect IID estimations are avoided, i.e., those valves that would lie near the curve, now lie in the inter-mode region.
  • the threshold is generally set to be exceeded when 33% (one-third) of the estimated decisions are intra-mode decisions.
  • the threshold can be set at any percentage that is required by the expected image sequence content.
  • the invention is also useful in detecting pictures that are not going to code very well. For example, if an imaged scene is changing unpredictably, as occurs when an object moves in and out from behind another object within the scene, two consecutive pictures in a sequence representing the scene will be substantially different. For such images, the encoder could exceed a coding bit budget for the picture sequence, i.e., utilize more bits to code the images than can be transmitted through the communications channel.
  • the inventive detector is used to detect, before coding begins, the picture that will be difficult to predict and code. Consequently, the encoder can alter the coding strategy to avoid exceeding the bit budget.

Landscapes

  • Engineering & Computer Science (AREA)
  • Multimedia (AREA)
  • Signal Processing (AREA)
  • Compression Or Coding Systems Of Tv Signals (AREA)

Abstract

The scene cut detector compares predicted macroblocks from an anchor image to input macroblocks from an input image on a macroblock-by-macroblock basis to generate a residual macroblock representing the difference between each predicted macroblock and each input macroblock. A variance for each residual macroblock and a variance for each input macroblock is computed after each comparison. The residual variance is compared to the input macroblock variance. Whenever the variance of the residuals macroblock exceeds the variance of the input macroblock, a counter is incremented. The scene cut detector repeats this process until each macroblock in the predicted image is compared to each input macroblock. If the count value ever exceeds a threshold level while a input image is being processed, the scene cut detector sets a scene cut indicator flag.

Description

The invention relates to a block-based video coding technique and, more particularly, the invention relates to a method and apparatus for detecting scene cuts in a video sequence within a block-based video coding system.
BACKGROUND OF THE DISCLOSURE
Block-based video coding systems typically use a coding technique that takes advantage of both spatial and temporal redundancy within an image (intra-picture) and between images (inter-picture) within a sequence of images. Such block-based image coding systems including those that utilize the well-known Moving Picture Experts Group (MPEG) standard of video coding); namely, ISO/IEC international standards 11172-2 (1994) (generally referred to as MPEG-1) and 13818-2 (Jan. 20, 1995 draft) (generally referred to as MPEG-2). To take advantage of the redundancies in an input video sequence and efficiently code a video sequence into a transmittable bitstream, block-based coding techniques assume that the sequential pictures within an input video sequence contain substantially similar information, i.e., the imaged scene changes very little from picture to picture. A scene cut occurring in the picture sequence violates the underlying assumption for efficient coding. Consequently, after a scene change (scene cut), a block-based coding technique must use a substantial number of bits to code the first picture following the scene change. Because the number of bits available to code any one image is typically limited, a scene cut can cause substantial errors in the coding and lead to substantial distortion of the decoded picture.
Therefore, a need exists in the art for a method and apparatus that detects the occurrence of a scene cut prior to coding the pictures such that the coding system can take certain actions to avoid substantial coding errors.
SUMMARY OF THE INVENTION
The disadvantages heretofore associated with the prior art are overcome by the present invention of a scene cut detector for incorporation into a conventional block-based video coding system. The scene cut detector compares predicted macroblocks from a predicted image to input macroblocks from an input image on a macroblock-by-macroblock basis to generate a residual macroblock representing the difference between each predicted macroblock and each input macroblock. A variance for each residual macroblock and a variance for each input macroblock is computed after each comparison. The variances of the residual macroblock and the input macroblock are compared to a decision function. Depending upon the outcome of the decision function comparison, a counter is incremented. The scene cut detector repeats this process until each macroblock in the predicted image is compared to each input macroblock. If the count value ever exceeds a threshold level while a input image is being processed, the scene cut detector identifies the input image as being within a new scene and, in response, sets a scene cut indicator flag.
BRIEF DESCRIPTION OF THE DRAWINGS
The teachings of the present invention can be readily understood by considering the following detailed description in conjunction with the accompanying drawings, in which:
FIG. 1 depicts a block diagram of a block-based coding system incorporating a scene cut detector of the present invention;
FIG. 2 depicts a detailed block diagram of a scene cut detector of the present invention; and
FIG. 3 depicts a graph of a decision function used by the IID estimator of the present invention.
To facilitate understanding, identical reference numerals have been used, where possible, to designate identical elements that are common to the figures.
DETAILED DESCRIPTION
FIG. 1 depicts a block diagram of a block-based coding system 100 (specifically, an MPEG encoder) incorporating the present invention. The input signal, at port 102, to the system is a preprocessed image that has been partitioned into a plurality of blocks, where the blocks are sequentially provided as an input to the system. Under the MPEG standard, these blocks of pixels are commonly known as macroblocks, e.g., a 16×16 pixel block. The following disclosure uses the MPEG standard terminology; however, it should be understood that the term macroblock is intended to describe a block of pixels of any size that is used for the basis of motion compensation.
The system computes, from the system output signal, a series of predicted macroblocks (P). Each predicted macroblock is illustratively produced by decoding the output signal, at port 104, just as the receiver of the transmitted output signal would decode the received signal. Subtractor 106 generates, on path 107, a residual signal (also referred to in the art as simply the residual or the residual macroblock) by subtracting the predicted macroblock from the input macroblock.
If the predicted macroblock is substantially similar to the input macroblock, the residuals are relatively small and are easily coded using very few bits. In such a scenario, the input macroblock is said to be motion compensated. However, if the difference between the predicted macroblock and the input macroblock is substantial, the residuals are difficult to code. Consequently, the system is better off directly coding the input macroblock in lieu of coding the motion compensated residual macroblock. This selection is known as a selection of the cooling mode. Coding the input macroblock I is referred to as intra-coding, while coding the residuals is referred to as inter-coding. The selection between these two modes is known as the Intra-Inter-Decision (IID).
The IID is made by the IID circuit 110, which sets the coding mode switch 108. The IID is computed by first computing the variance of the residual macroblock (Var R) and the variance of the input macroblock (Var I). The coding decision is based on these values. There are several functions that can be used to make this decision. For example, if Var R is less than Var I, the IID selects the Inter-mode. Conversely, if Var I is less than Var R, the IID selects the Intra-mode.
The selected block is processed in a discrete cosine transform (DCT) block 112. The DCT produces coefficients representing the input signal to the DCT. The quantizer 114 quantizes the coefficients to produce the output block at port 104. The rate control block 116 controls the quantization scale (step size) used to quantize the coefficients.
To produce correct predicted blocks and to accomplish efficient half-pel motion vector generation, the encoder needs access to the decoded images. In order to accomplish such access, the quantizer 114 output is passed through both the inverse quantizer 118 and inverse DCT 120. The output of the inverse DCT is ideally identical to the input to the DCT 112. In the inter-mode, the decoded macroblock is produced by summing the output of the inverse DCT and the predicted macroblock. During the intra-mode, the decoded macroblock is simply the output of the inverse DCT. The decoded macroblocks are then stored in the frame store 124. The frame store accumulates a plurality of these "reconstructed" macroblocks that constitute an entire reconstructed frame of image information. The reconstructed frame is used by the motion vector predictor 126 to produce motion vectors that are used in generating predicted macroblocks for forthcoming input images.
To generate motion vectors, the motion vector predictor 126 comprises three components: a full-pel motion estimator 128, a half-pel motion estimator 130, and a motion mode block 132. The full-pel motion estimator 128 is a "coarse" motion vector generator that searches for a coarse match between a macroblock in a previous image and the present input macroblock. The previous image is referred to as an anchor image. Under the MPEG standards, the anchor image is what is known as an I or P frame within an image sequence known as a Group Of Pictures (GOP). The motion vector is a vector representing the relative position where a coarse match was found between the two macroblocks. The coarse motion vector generator produces a motion vector that is accurate to one picture element (pel).
The accuracy of the full-pel motion estimator is improved in the half-pel motion estimator. The half-pel estimator uses the full-pel motion vectors and the reconstructed macroblocks from the frame store 124 to compute motion vectors to half-pel accuracy. The half-pel motion vectors are then sent to the motion modes block 132. Typically, there are multiple motion vectors related to each macroblock. The modes block 132 selects the best motion vector for representing motion for each input macroblock.
The full-pel estimator is a computationally intensive task compared to the half-pel estimator. For this reason, in several implementations, it is computed separately on dedicated hardware. Often, all the full-pel motion vectors are computed before half-pel processing begins.
The MPEG encoder system discussed above is a conventional system that is available as a set of integrated circuits as model L64120 from LSI Logic, Inc. of Milpitas, Calif. Importantly, this MPEG encoder stores an entire frame of full-pel motion vectors before the half-pel estimator begins operation.
The concepts of Motion Estimation and Motion Compensation are based on an underlying assumption that the current picture is not very different from a previously occurring picture (the anchor image). However, when a scene change (also called a scene cut) occurs, the anchor pictures are substantially different from the current picture. Hence, the predicted macroblocks are very inaccurate and the residuals are large. As such, for most input macroblocks of a picture, the IID selects the input macroblock (intra-mode) for coding in lieu of coding the residuals (inter-mode). It should be noted that this coding decision occurs even in the case where there is no scene change, and that a normally coded picture may contain a mixture of intra- and inter-coded macroblocks. However, the percentage of intra-coded macroblocks significantly increases when a scene cut occurs. The scene cut detector of the present invention analyzes all the macroblocks in a picture and then determines whether a scene cut has occurred. This is accomplished by counting the number of intra-coded macroblocks and comparing the count to a threshold level. Specifically, if the percentage of I coded macroblocks in any given frame exceeds the threshold level, that frame is deemed to follow a scene cut.
In a typical MPEG encoder, the actual IID decision is made after the half-pel motion vectors are generated and best motion vector is chosen. Since the full-pel estimator 128 generates motion vectors for the entire frame before the first macroblock is coded by the encoder, the inventive scene cut detector apparatus 134 monitoring these full-pel results is able to produce an IID estimate for all macroblocks, i.e., an estimate that the IID would make when it does analyze the residuals. The scene cut detector comprises an IID estimator 136 connected in series to an intra-code macroblock counter 138. The counter 138 generates a scene cut decision (a flag) indicating that the scene cut detector has determined that a scene cut has occurred.
FIG. 2 depicts a detailed block diagram of the inventive scene cut detector 134 of the MPEG encoder 100. The full-pel motion estimator 128 is provided the I macroblocks as well as the appropriate I or P anchor image in which the predicted macroblocks are found. The anchor images are stored in the frame memory 140. The full-pel motion vector generator 141 generates the motion vectors for each input macroblock using one of many well-known methods including those disclosed in U.S. Pat. No. 5,351,095, issued Sep. 27, 1994, incorporated herein by reference, and U.S. patent application Ser. No. 08/300,023, filed Sep. 2, 1994, incorporated herein by reference. Using the full-pel motion vectors and the anchor images from the anchor image store 140, the motion compensator 145 generates the predicted macroblock (P) for each input macroblock (I).
The input macroblocks (path 120) and the predicted image macroblocks (P) form inputs to the scene cut detector 134. The IID estimator 136 computes full-pel residual by subtracting (subtractor 142) the predicted macroblock from the input macroblock. The IID estimator then computes, using the variance blocks 146 and 148, the variance of the input macroblocks (Var I) and the variance of the full-pel residuals (Var R). The IID circuit 150 then makes its IID estimate based on these variances. The counter 152 counts the number of intra-mode decisions and compares, in block 154, the count number to a threshold count value. If, while processing a given picture, the count number ever exceeds the threshold, the intra MB counter generates a scene cut decision flag.
FIG. 3 depicts a graph 300 of a typical decision function used by the IID estimator in comparing Var R and Var I. The simplest decision function is linear function 302. This function is implemented using a comparator such that when Var I is less than Var R, the IID estimates is in intra-mode (region 310). Conversely, when Var R is less than Vat I, the IID estimate is inter-mode (region 312).
However, though simple, a linear function does not tend to provide the best results. Thus, non-linear function 304 represents a more typical function. This function is vertical at a particular value 306 of Var R and then is linear. In operation, macroblocks with relatively small values of Var R are coded using the inter-mode. Any value of Var R greater than value 306 is compared using the direct comparison of function 302.
Since the half-pel motion estimator more accurately defines the motion vectors, a macroblock having Var R value that is only slightly greater than the Var I value, e.g., a point lying near, but just beneath the curve 302, can be shifted above the curve when more accurate estimation is accomplished using the half-pel estimator. As such, the IID of the MPEG encoder would use inter-mode coding, yet the IID estimation in the scene cut detector would estimate that intra-mode coding would be used. To compensate for this anomaly, function 308 is generally used. Function 308 is similar in form to function 304; however, the function 308 is shifted slightly below function 304. Thus, incorrect IID estimations are avoided, i.e., those valves that would lie near the curve, now lie in the inter-mode region.
Of course, the functions depicted in FIG. 3 are merely illustrative. Other linear and non-linear functions are used to achieve particular results. The invention is intended to include any function within the IID estimator.
During a typical sequence of macroblocks (no scene cut), the predicted macroblocks, even at full-pel accuracy, will yield mostly inter-mode decisions. Thus, the scene cut flag will not be set. Note that t percentage of intra-mode decisions made after half-pel refinement will be lower because of the additional accuracy in the motion vectors.
When a scene cut occurs, it does not matter whether the system generates motion vectors to half-pel or full-pel accuracy. The motion estimation will be inaccurate in either case. A substantial number of intra-mode macroblocks are selected at both the full-pel and the half-pel stages. Therefore, the rough count of intra-mode decisions using the IID estimator at full-pel accuracy is sufficient for scene cut detection.
The threshold is generally set to be exceeded when 33% (one-third) of the estimated decisions are intra-mode decisions. Of course, the threshold can be set at any percentage that is required by the expected image sequence content.
In addition to detecting scene cuts, the invention is also useful in detecting pictures that are not going to code very well. For example, if an imaged scene is changing unpredictably, as occurs when an object moves in and out from behind another object within the scene, two consecutive pictures in a sequence representing the scene will be substantially different. For such images, the encoder could exceed a coding bit budget for the picture sequence, i.e., utilize more bits to code the images than can be transmitted through the communications channel. The inventive detector is used to detect, before coding begins, the picture that will be difficult to predict and code. Consequently, the encoder can alter the coding strategy to avoid exceeding the bit budget.
One particular apparatus that uses the scene cut detector flag is described in U.S. patent application Ser. No. 08/606,622 filed simultaneously herewith (Attorney Docket 11817) and incorporated herein by reference. In response to the flag, this apparatus alters the quantization scale such that the encoder coarsely codes the picture and the bit budget is preserved.
Although a single embodiment which incorporates the teachings of the present invention has been shown and described in detail herein, those skilled in the art can readily devise many other varied embodiments that still incorporate these teachings.

Claims (14)

What is claimed is:
1. In a block-based video encoder that partitions each input image of a sequence of input images into a plurality of macroblocks, computes at least one motion vector for each of said macroblocks, and generates a sequence of predicted images, where each predicted image in said sequence comprises a plurality of predicted macroblocks derived from said input images and said motion vectors, a method of detecting that a first image is substantially different from a second image within said sequence of input images comprising the steps of:
(a) estimating a coding mode that is to be used by the block-based video encoder to code a macroblock within said second image;
(b) incrementing a count value whenever the coding mode is estimated to be a particular type of coding mode;
(c) repeating steps a and b until the coding mode is estimated for each macroblock in the second image; and
(d) when the count value exceeds a threshold level, setting an indicator flag.
2. The method of claim 1 wherein said estimating step further comprises the steps of:
comparing a predicted macroblock to an input macroblock from said first image to generate a residuals macroblock;
computing a variance of the input macroblock and the residuals macroblock;
comparing the variance of the input macroblock and the variance of the residuals macroblock to a decision function; and
in response to a particular outcome of said comparing step, incrementing said count value.
3. The method of claim 1 wherein said particular type of coding mode is an intra-block coding mode.
4. The method of claim 1 wherein the count value exceeds the threshold level whenever a scene cut has occurred in said sequence of images.
5. The method of claim 1 wherein said threshold level is set to thirty percent of the total number of macroblocks in the input image.
6. In a block-based video encoder that partitions each input image of a sequence of input images into a plurality of macroblocks, computes at least one motion vector for each of said macroblocks, and generates a sequence of predicted images, where each predicted image in said sequence comprises a plurality of predicted macroblocks derived from said input images and said motion vectors, a method of detecting that a first image is substantially different from a second image within said sequence of input images comprising the steps of:
(a) comparing a predicted macroblock to an input macroblock from said first image to generate a residuals macroblock;
(b) computing a variance of the input macroblock and the residuals macroblock;
(c) comparing the variance of the input macroblock to the variance of the residuals macroblock;
(d) if the variance of the residuals macroblock exceeds the variance of the input macroblock, incrementing a count value;
(e) repeating steps a, b, c, d, and e, until each macroblock in the predicted image is compared to each input macroblock; and
(f) when the count value exceeds a threshold level, setting an indicator flag.
7. The method of claim 6 wherein said variance of the residuals macroblock exceeds the variance of the input macroblock to produce a count value that exceeds the threshold level whenever a scene cut has occurred in said sequence of images.
8. The method of claim 6 wherein said threshold level is set to approximately thirty-three percent of the total number of macroblocks in the input image.
9. In a block-based video encoder that partitions each input image of a sequence of input images into a plurality of macroblocks, computes at least one motion vector for each of said macroblocks, and generates a sequence of predicted images, where each predicted image in said sequence comprises a plurality of predicted macroblocks derived from said input images and said motion vectors, apparatus of detecting that a first image is substantially different from a second image within said sequence of input images comprising:
a coding mode estimator, connected to a motion estimator within said block-based video encoder and a source of input images, for estimating the coding mode that is to be used by said block-based video encoder to code each macro-block of said second image;
a counter, connected to said coding mode estimator, for incrementing a count value whenever a number of macroblocks coded using a first coding mode and a number of macroblocks coded using a second coding mode compare favorably with a decision function; and
thresholding means, connected to said counter, for setting an indicator flag when the count value exceeds a threshold level.
10. The apparatus of claim 9 wherein said coding mode estimator comprises:
subtractor, connected to said motion estimator and a source of said input image, for comparing each of the predicted macroblocks to an input macroblock to generate a residuals macroblock;
a residuals variance generator, connected to said subtractor, for computing a variance of the residuals macroblocks;
a input variance generator, connected to said source of the input image, for computing a variance of the input macroblocks;
a decision circuit, connected to the residuals variance generator and the input variance generator, for comparing the variance of the input macroblocks and the variance of the residuals macroblocks to a decision function, wherein when the variance of said residuals macroblock and the variance of the input macroblock compares with the decision function to produce a first output, the block-based video encoder is estimated to be using the second coding mode and when the variance of said residuals macroblock and the variance of the input macroblock compares with the decision function to produce a second output, the block-based video encoder is estimated to be using the first coding mode.
11. The apparatus of claim 10 wherein said first coding mode is an inter-block coding mode and the second coding mode is an intra-block coding mode.
12. The apparatus of claim 9 wherein the count value exceeds the threshold level whenever a scene cut has occurred in said sequence of images.
13. The apparatus of claim 9 wherein said threshold level is set to approximately thirty-three percent of the total number of macroblocks in the input image.
14. The apparatus of claim 10 wherein said decision circuit is a comparator, connected to the residuals variance generator and the input variance generator, for comparing the variance of the input macroblocks to the variance of the residuals macroblocks, wherein when the variance of said residuals macroblock exceeds the variance of the input macroblock, the block-based video encoder is estimated to be using the second coding mode and when the variance of said input macroblock exceeds the variance of the residuals macroblock, the block-based video encoder is estimated to be using the first coding mode.
US08/606,613 1996-02-26 1996-02-26 Method and apparatus for detecting scene-cuts in a block-based video coding system Expired - Lifetime US5724100A (en)

Priority Applications (6)

Application Number Priority Date Filing Date Title
US08/606,613 US5724100A (en) 1996-02-26 1996-02-26 Method and apparatus for detecting scene-cuts in a block-based video coding system
DE69738901T DE69738901D1 (en) 1996-02-26 1997-02-26 METHOD AND DEVICE FOR DETERMINING SCENARIOUS POINTS IN A BLOCK-BASED VIDEO SIGNALING CODING SYSTEM
JP9530283A JP2000505614A (en) 1996-02-26 1997-02-26 Method and apparatus for detecting scene cuts in block-based video coding systems
PCT/US1997/002563 WO1997031485A1 (en) 1996-02-26 1997-02-26 Method and apparatus for detecting scene cuts in a block-based video coding system
KR10-1998-0706667A KR100382676B1 (en) 1996-02-26 1997-02-26 Method and apparatus for detecting scene cuts in a block-based video coding system
EP97907683A EP0885526B1 (en) 1996-02-26 1997-02-26 Method and apparatus for detecting scene cuts in a block-based video coding system

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
US08/606,613 US5724100A (en) 1996-02-26 1996-02-26 Method and apparatus for detecting scene-cuts in a block-based video coding system

Publications (1)

Publication Number Publication Date
US5724100A true US5724100A (en) 1998-03-03

Family

ID=24428691

Family Applications (1)

Application Number Title Priority Date Filing Date
US08/606,613 Expired - Lifetime US5724100A (en) 1996-02-26 1996-02-26 Method and apparatus for detecting scene-cuts in a block-based video coding system

Country Status (6)

Country Link
US (1) US5724100A (en)
EP (1) EP0885526B1 (en)
JP (1) JP2000505614A (en)
KR (1) KR100382676B1 (en)
DE (1) DE69738901D1 (en)
WO (1) WO1997031485A1 (en)

Cited By (30)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO1999038314A2 (en) * 1998-01-21 1999-07-29 Sarnoff Corporation Apparatus and method for using side information to improve a coding system
EP0979011A1 (en) * 1998-08-06 2000-02-09 STMicroelectronics S.r.l. Detection of a change of scene in a motion estimator of a video encoder
US6137544A (en) * 1997-06-02 2000-10-24 Philips Electronics North America Corporation Significant scene detection and frame filtering for a visual indexing system
US6167155A (en) * 1997-07-28 2000-12-26 Physical Optics Corporation Method of isomorphic singular manifold projection and still/video imagery compression
US6339617B1 (en) * 1997-12-19 2002-01-15 Nec Corporation Moving picture compressing apparatus and moving picture compressing method
US6351493B1 (en) 1998-06-30 2002-02-26 Compaq Computer Corporation Coding an intra-frame upon detecting a scene change in a video sequence
US20020078438A1 (en) * 2000-03-31 2002-06-20 U.S. Philps Corporation Video signal analysis and storage
EP1244291A2 (en) * 2001-03-16 2002-09-25 Kabushiki Kaisha Toshiba Moving image compression and cut detection
US6493390B2 (en) * 1997-10-02 2002-12-10 Kokusai Denshin Denwa Kabushiki Kaisha Motion compensation image encoding device
US20030058947A1 (en) * 2001-09-26 2003-03-27 Peterson Richard Mateer Scene cut detection in a video bitstream
US20030215012A1 (en) * 1999-02-25 2003-11-20 Matsushita Electric Industrial Co., Ltd. Method and apparatus for transforming moving picture coding system
WO2004002150A1 (en) * 2002-06-25 2003-12-31 Motorola, Inc., A Corporation Of The State Of Delaware Video encoder and method for encoding video frames
US20040156549A1 (en) * 1998-10-01 2004-08-12 Cirrus Logic, Inc. Feedback scheme for video compression system
US20050089232A1 (en) * 2003-10-23 2005-04-28 Chun-Ming Hsu Method of video compression that accommodates scene changes
US20050105615A1 (en) * 2003-11-13 2005-05-19 Khaled El-Maleh Selective and/or scalable complexity control for video codecs
EP1575300A1 (en) * 2004-03-10 2005-09-14 LG Electronics Inc. System and method for controlling bit rate of an image
US20050201460A1 (en) * 2004-03-10 2005-09-15 Lg Electronics Inc. System and method for controlling bit rate of an image
US20050224902A1 (en) * 2002-02-06 2005-10-13 Ramsey Craig C Wireless substrate-like sensor
US20050243917A1 (en) * 2004-04-30 2005-11-03 Samsung Electronics Co., Ltd. H.263/MPEG video encoder using average histogram difference and method for controlling the same
US20060008152A1 (en) * 1999-10-08 2006-01-12 Rakesh Kumar Method and apparatus for enhancing and indexing video and audio signals
US7020351B1 (en) 1999-10-08 2006-03-28 Sarnoff Corporation Method and apparatus for enhancing and indexing video and audio signals
US20060150102A1 (en) * 2005-01-06 2006-07-06 Thomson Licensing Method of reproducing documents comprising impaired sequences and, associated reproduction device
US20060165162A1 (en) * 2005-01-24 2006-07-27 Ren-Wei Chiang Method and system for reducing the bandwidth access in video encoding
US20060239348A1 (en) * 2005-04-25 2006-10-26 Bo Zhang Method and system for encoding video data
US20070147512A1 (en) * 2000-04-18 2007-06-28 Ati International Srl Method and apparatus for rate control for constant-bit-rate-finite-buffer-size video encoder
US20070248160A1 (en) * 1996-09-20 2007-10-25 At&T Corp. System and method of providing directional information for direct prediction
US20090016438A1 (en) * 1998-12-08 2009-01-15 Mcdade Darryn Method and apparatus for a motion compensation instruction generator
US20100309984A1 (en) * 2009-06-09 2010-12-09 Sony Corporation Dual-mode compression of images and videos for reliable real-time transmission
US20110051809A1 (en) * 2009-09-02 2011-03-03 Sony Computer Entertainment Inc. Scene change detection
US20110075739A1 (en) * 1996-09-20 2011-03-31 At&T Intellectual Property Ii, L.P. Video Coder Providing Implicit Coefficient Prediction and Scan Adaptation for Image Coding and Intra Coding of Video

Families Citing this family (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US6125229A (en) * 1997-06-02 2000-09-26 Philips Electronics North America Corporation Visual indexing system
GB2351406A (en) * 1999-06-21 2000-12-27 Sony Uk Ltd Video data compression with scene change detection
GB2406737B (en) * 2001-07-20 2005-10-05 Hewlett Packard Co System and method for detecting the border of recorded video data

Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US5377051A (en) * 1993-01-13 1994-12-27 Hitachi America, Ltd. Digital video recorder compatible receiver with trick play image enhancement
US5404174A (en) * 1992-06-29 1995-04-04 Victor Company Of Japan, Ltd. Scene change detector for detecting a scene change of a moving picture
US5459517A (en) * 1992-12-15 1995-10-17 Fuji Xerox Co., Ltd. Moving picture scene detection system
US5493345A (en) * 1993-03-08 1996-02-20 Nec Corporation Method for detecting a scene change and image editing apparatus
US5532746A (en) * 1993-11-09 1996-07-02 Korea Telecommunication Authority Bit allocation method for controlling transmission rate of video encoder

Family Cites Families (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP3338460B2 (en) * 1991-05-09 2002-10-28 ソニー株式会社 Video signal encoding device and video signal encoding method
EP1304884B1 (en) * 1992-01-29 2005-11-02 Mitsubishi Denki Kabushiki Kaisha Video information recording medium and reproducing apparatus

Patent Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US5404174A (en) * 1992-06-29 1995-04-04 Victor Company Of Japan, Ltd. Scene change detector for detecting a scene change of a moving picture
US5459517A (en) * 1992-12-15 1995-10-17 Fuji Xerox Co., Ltd. Moving picture scene detection system
US5377051A (en) * 1993-01-13 1994-12-27 Hitachi America, Ltd. Digital video recorder compatible receiver with trick play image enhancement
US5493345A (en) * 1993-03-08 1996-02-20 Nec Corporation Method for detecting a scene change and image editing apparatus
US5532746A (en) * 1993-11-09 1996-07-02 Korea Telecommunication Authority Bit allocation method for controlling transmission rate of video encoder

Cited By (56)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20070248162A1 (en) * 1996-09-20 2007-10-25 Haskell Barin Geoffry Video coder providing implicit coefficient prediction and scan adaptation for image coding and intra coding of video
US9693051B2 (en) 1996-09-20 2017-06-27 At&T Intellectual Property Ii, L.P. Video coder providing implicit coefficient prediction and scan adaptation for image coding and intra coding of video
US20070248160A1 (en) * 1996-09-20 2007-10-25 At&T Corp. System and method of providing directional information for direct prediction
US20070248161A1 (en) * 1996-09-20 2007-10-25 At&T Corp. System and method for generating video data for implicit coefficient prediction decoding
US7646809B2 (en) 1996-09-20 2010-01-12 At&T Intellectual Property Ii, Lp. System and method of providing directional information for direct prediction
US7869502B2 (en) 1996-09-20 2011-01-11 At&T Intellectual Property Ii, L.P. Video coder providing implicit coefficient prediction and scan adaptation for image coding and intra coding of video
US8625665B2 (en) 1996-09-20 2014-01-07 At&T Intellectual Property Ii, L.P. Video coder providing implicit coefficient prediction and scan adaptation for image coding and intra coding of video
US20110075739A1 (en) * 1996-09-20 2011-03-31 At&T Intellectual Property Ii, L.P. Video Coder Providing Implicit Coefficient Prediction and Scan Adaptation for Image Coding and Intra Coding of Video
US7974346B2 (en) 1996-09-20 2011-07-05 AT&T Intellectual II, L.P. System and method for generating video data for implicit coefficient prediction decoding
US6137544A (en) * 1997-06-02 2000-10-24 Philips Electronics North America Corporation Significant scene detection and frame filtering for a visual indexing system
US6167155A (en) * 1997-07-28 2000-12-26 Physical Optics Corporation Method of isomorphic singular manifold projection and still/video imagery compression
US6493390B2 (en) * 1997-10-02 2002-12-10 Kokusai Denshin Denwa Kabushiki Kaisha Motion compensation image encoding device
US6339617B1 (en) * 1997-12-19 2002-01-15 Nec Corporation Moving picture compressing apparatus and moving picture compressing method
JP2002502158A (en) * 1998-01-21 2002-01-22 サーノフ コーポレイション Apparatus and method for using side information to improve coding systems
US6233278B1 (en) * 1998-01-21 2001-05-15 Sarnoff Corporation Apparatus and method for using side information to improve a coding system
WO1999038314A3 (en) * 1998-01-21 2000-01-06 Sarnoff Corp Apparatus and method for using side information to improve a coding system
WO1999038314A2 (en) * 1998-01-21 1999-07-29 Sarnoff Corporation Apparatus and method for using side information to improve a coding system
US6100940A (en) * 1998-01-21 2000-08-08 Sarnoff Corporation Apparatus and method for using side information to improve a coding system
US6351493B1 (en) 1998-06-30 2002-02-26 Compaq Computer Corporation Coding an intra-frame upon detecting a scene change in a video sequence
US6480543B1 (en) 1998-08-06 2002-11-12 Stmicroelectronics S.R.L. Detection of a change of scene in a motion estimator of a video encoder
EP0979011A1 (en) * 1998-08-06 2000-02-09 STMicroelectronics S.r.l. Detection of a change of scene in a motion estimator of a video encoder
US20040156549A1 (en) * 1998-10-01 2004-08-12 Cirrus Logic, Inc. Feedback scheme for video compression system
US7158681B2 (en) 1998-10-01 2007-01-02 Cirrus Logic, Inc. Feedback scheme for video compression system
US20090016438A1 (en) * 1998-12-08 2009-01-15 Mcdade Darryn Method and apparatus for a motion compensation instruction generator
US7292632B2 (en) * 1999-02-25 2007-11-06 Matsushita Electric Industrial Co., Ltd. Method and apparatus for transforming moving picture coding system
US20030215012A1 (en) * 1999-02-25 2003-11-20 Matsushita Electric Industrial Co., Ltd. Method and apparatus for transforming moving picture coding system
US20060008152A1 (en) * 1999-10-08 2006-01-12 Rakesh Kumar Method and apparatus for enhancing and indexing video and audio signals
US7020351B1 (en) 1999-10-08 2006-03-28 Sarnoff Corporation Method and apparatus for enhancing and indexing video and audio signals
US20020078438A1 (en) * 2000-03-31 2002-06-20 U.S. Philps Corporation Video signal analysis and storage
US7277483B1 (en) * 2000-04-18 2007-10-02 Ati International Srl Method and apparatus for rate control for constant-bit-rate finite-buffer-size video encoder
US9414078B2 (en) 2000-04-18 2016-08-09 Ati Technologies Ulc Method for rate control for constant-bit-rate-finite-buffer-size video encoder
US10462473B2 (en) 2000-04-18 2019-10-29 Ati Technologies Ulc Method and apparatus for rate control for constant-bit-rate-finite-buffer-size video encoder
US20070147512A1 (en) * 2000-04-18 2007-06-28 Ati International Srl Method and apparatus for rate control for constant-bit-rate-finite-buffer-size video encoder
EP1244291A2 (en) * 2001-03-16 2002-09-25 Kabushiki Kaisha Toshiba Moving image compression and cut detection
EP1244291A3 (en) * 2001-03-16 2003-06-25 Kabushiki Kaisha Toshiba Moving image compression and cut detection
US6996183B2 (en) * 2001-09-26 2006-02-07 Thomson Licensing Scene cut detection in a video bitstream
US20030058947A1 (en) * 2001-09-26 2003-03-27 Peterson Richard Mateer Scene cut detection in a video bitstream
US20050224902A1 (en) * 2002-02-06 2005-10-13 Ramsey Craig C Wireless substrate-like sensor
WO2004002150A1 (en) * 2002-06-25 2003-12-31 Motorola, Inc., A Corporation Of The State Of Delaware Video encoder and method for encoding video frames
US20050089232A1 (en) * 2003-10-23 2005-04-28 Chun-Ming Hsu Method of video compression that accommodates scene changes
US20050105615A1 (en) * 2003-11-13 2005-05-19 Khaled El-Maleh Selective and/or scalable complexity control for video codecs
US9351013B2 (en) * 2003-11-13 2016-05-24 Qualcomm Incorporated Selective and/or scalable complexity control for video codecs
US20050201460A1 (en) * 2004-03-10 2005-09-15 Lg Electronics Inc. System and method for controlling bit rate of an image
EP1575300A1 (en) * 2004-03-10 2005-09-14 LG Electronics Inc. System and method for controlling bit rate of an image
US20050243917A1 (en) * 2004-04-30 2005-11-03 Samsung Electronics Co., Ltd. H.263/MPEG video encoder using average histogram difference and method for controlling the same
US20060150102A1 (en) * 2005-01-06 2006-07-06 Thomson Licensing Method of reproducing documents comprising impaired sequences and, associated reproduction device
US9043701B2 (en) * 2005-01-06 2015-05-26 Thomson Licensing Method and apparatus for indicating the impaired sequences of an audiovisual document
US20060165162A1 (en) * 2005-01-24 2006-07-27 Ren-Wei Chiang Method and system for reducing the bandwidth access in video encoding
US9667999B2 (en) * 2005-04-25 2017-05-30 Avago Technologies General Ip (Singapore) Pte. Ltd. Method and system for encoding video data
US20060239348A1 (en) * 2005-04-25 2006-10-26 Bo Zhang Method and system for encoding video data
WO2010144488A3 (en) * 2009-06-09 2011-02-10 Sony Corporation Dual-mode compression of images and videos for reliable real-time transmission
US8964851B2 (en) 2009-06-09 2015-02-24 Sony Corporation Dual-mode compression of images and videos for reliable real-time transmission
WO2010144488A2 (en) * 2009-06-09 2010-12-16 Sony Corporation Dual-mode compression of images and videos for reliable real-time transmission
US20100309984A1 (en) * 2009-06-09 2010-12-09 Sony Corporation Dual-mode compression of images and videos for reliable real-time transmission
US20110051809A1 (en) * 2009-09-02 2011-03-03 Sony Computer Entertainment Inc. Scene change detection
US8345750B2 (en) * 2009-09-02 2013-01-01 Sony Computer Entertainment Inc. Scene change detection

Also Published As

Publication number Publication date
EP0885526A4 (en) 2003-03-05
KR100382676B1 (en) 2005-09-20
WO1997031485A1 (en) 1997-08-28
EP0885526A1 (en) 1998-12-23
DE69738901D1 (en) 2008-09-25
KR19990087266A (en) 1999-12-15
EP0885526B1 (en) 2008-08-13
JP2000505614A (en) 2000-05-09

Similar Documents

Publication Publication Date Title
US5724100A (en) Method and apparatus for detecting scene-cuts in a block-based video coding system
US5731835A (en) Dynamic coding rate control in a block-based video coding system
EP0710031B1 (en) System for coding a video signal in the presence of an image intensity gradient
KR100626994B1 (en) Variable bitrate video coding method and corresponding videocoder
EP2250813B1 (en) Method and apparatus for predictive frame selection supporting enhanced efficiency and subjective quality
US6591015B1 (en) Video coding method and apparatus with motion compensation and motion vector estimator
US5500689A (en) System for detecting a video signal image intensity gradient
US6167088A (en) Method and apparatus for performing adaptive encoding rate control of a video information stream including 3:2 pull-down video information
US6907072B2 (en) Moving image encoding apparatus
US20050254577A1 (en) Image picture encoding apparatus and method, computer program and computer-readable storage medium
US20060239347A1 (en) Method and system for scene change detection in a video encoder
US20120320979A1 (en) Method and digital video encoder system for encoding digital video data
US7092443B2 (en) Process and device for video coding using the MPEG4 standard
US7991048B2 (en) Device and method for double-pass encoding of a video data stream
US9131236B2 (en) Code amount control method and apparatus
JP2008005536A (en) Method and apparatus for detecting scene cuts in block-based video coding system
KR950006777B1 (en) Error correcting system of decoder
KR20070029109A (en) Video encoding method and device
Milicevic et al. B Frame Processing Modeling in H. 264/AVC
KR20000025291A (en) Motion compensation and mode determination apparatus of mpeg-2 image encoder

Legal Events

Date Code Title Description
AS Assignment

Owner name: DAVID SARNOFF RESEARCH CENTER, INC., NEW JERSEY

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:KUCHIBHOTLA, PRASHANTH;REEL/FRAME:007982/0714

Effective date: 19960509

STCF Information on status: patent grant

Free format text: PATENTED CASE

AS Assignment

Owner name: SARNOFF CORPORATION, NEW JERSEY

Free format text: MERGER;ASSIGNOR:DAVID SARNOFF RESEARCH CENTER, INC.;REEL/FRAME:009912/0460

Effective date: 19970404

FEPP Fee payment procedure

Free format text: PAYOR NUMBER ASSIGNED (ORIGINAL EVENT CODE: ASPN); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY

FPAY Fee payment

Year of fee payment: 4

AS Assignment

Owner name: MEDIATEK, INC., TAIWAN

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:SARNOFF CORPORATION;REEL/FRAME:016283/0355

Effective date: 20041022

FPAY Fee payment

Year of fee payment: 8

FPAY Fee payment

Year of fee payment: 12