EP1012778A1 - Apparatus and method for macroblock based rate control in a coding system - Google Patents

Apparatus and method for macroblock based rate control in a coding system

Info

Publication number
EP1012778A1
EP1012778A1 EP98935565A EP98935565A EP1012778A1 EP 1012778 A1 EP1012778 A1 EP 1012778A1 EP 98935565 A EP98935565 A EP 98935565A EP 98935565 A EP98935565 A EP 98935565A EP 1012778 A1 EP1012778 A1 EP 1012778A1
Authority
EP
European Patent Office
Prior art keywords
frame
block
bit rate
macroblock
target
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
EP98935565A
Other languages
German (de)
French (fr)
Other versions
EP1012778B1 (en
EP1012778A4 (en
Inventor
Tihao Chiang
Hung-Ju Lee
Ya-Qin Zhang
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
MediaTek Inc
Original Assignee
Sarnoff Corp
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Sarnoff Corp filed Critical Sarnoff Corp
Publication of EP1012778A1 publication Critical patent/EP1012778A1/en
Publication of EP1012778A4 publication Critical patent/EP1012778A4/en
Application granted granted Critical
Publication of EP1012778B1 publication Critical patent/EP1012778B1/en
Anticipated expiration legal-status Critical
Expired - Lifetime legal-status Critical Current

Links

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/10Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
    • H04N19/189Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the adaptation method, adaptation tool or adaptation type used for the adaptive coding
    • H04N19/196Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the adaptation method, adaptation tool or adaptation type used for the adaptive coding being specially adapted for the computation of encoding parameters, e.g. by averaging previously computed encoding parameters
    • H04N19/198Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the adaptation method, adaptation tool or adaptation type used for the adaptive coding being specially adapted for the computation of encoding parameters, e.g. by averaging previously computed encoding parameters including smoothing of a sequence of encoding parameters, e.g. by averaging, by choice of the maximum, minimum or median value
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/10Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
    • H04N19/102Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the element, parameter or selection affected or controlled by the adaptive coding
    • H04N19/115Selection of the code volume for a coding unit prior to coding
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/10Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
    • H04N19/102Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the element, parameter or selection affected or controlled by the adaptive coding
    • H04N19/124Quantisation
    • H04N19/126Details of normalisation or weighting functions, e.g. normalisation matrices or variable uniform quantisers
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/10Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
    • H04N19/134Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the element, parameter or criterion affecting or controlling the adaptive coding
    • H04N19/136Incoming video signal characteristics or properties
    • H04N19/137Motion inside a coding unit, e.g. average field, frame or block difference
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/10Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
    • H04N19/134Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the element, parameter or criterion affecting or controlling the adaptive coding
    • H04N19/146Data rate or code amount at the encoder output
    • H04N19/147Data rate or code amount at the encoder output according to rate distortion criteria
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/10Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
    • H04N19/169Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding
    • H04N19/17Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object
    • H04N19/172Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object the region being a picture, frame or field
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/10Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
    • H04N19/169Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding
    • H04N19/17Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object
    • H04N19/176Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object the region being a block, e.g. a macroblock
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/10Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
    • H04N19/189Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the adaptation method, adaptation tool or adaptation type used for the adaptive coding
    • H04N19/196Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the adaptation method, adaptation tool or adaptation type used for the adaptive coding being specially adapted for the computation of encoding parameters, e.g. by averaging previously computed encoding parameters
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/60Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using transform coding
    • H04N19/61Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using transform coding in combination with predictive coding
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/10Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
    • H04N19/134Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the element, parameter or criterion affecting or controlling the adaptive coding
    • H04N19/146Data rate or code amount at the encoder output
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/10Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
    • H04N19/134Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the element, parameter or criterion affecting or controlling the adaptive coding
    • H04N19/146Data rate or code amount at the encoder output
    • H04N19/152Data rate or code amount at the encoder output by measuring the fullness of the transmission buffer

Definitions

  • the present invention relates to an apparatus and concomitant method for optimizing the coding of motion video. More particularly, this invention relates to a method and apparatus that adaptively adjusts a quantizer scale for each macroblock within a frame to maintain the overall quality of the motion video while optimizing the coding rate.
  • MPEG Moving Picture Experts Group
  • ISO/IEC ISO/IEC international Standards 11172 and 13818
  • MPEG-1 and MPEG-2 format respectively
  • MPEG-1 and MPEG-2 format respectively
  • MPEG-1 and MPEG-2 format respectively
  • these MPEG standards specify a general coding methodology and syntax for generating an MPEG compliant bitstream, many variations are permitted to accommodate a plurality of different applications and services such as desktop video publishing, video conferencing, digital storage media and television broadcast.
  • MPEG does not define a specific method for controlling the bit rate of an encoder. It is the task of the encoder designer to devise a rate control process for controlling the bit rate such that the decoder input buffer neither overflows nor underflows.
  • bit rate is to alter the quantization process, which will affect the distortion of the input video image.
  • the quantizer scale step size
  • the quantizer scale for each frame is selected by assuming that all the pictures of the same type have identical complexity within a group of pictures.
  • the quantizer scale selected by this criterion may not achieve optimal coding performance, since the complexity of each picture will vary with time.
  • encoders that utilize global-type transforms have similar problems.
  • wavelet transforms are applied to an important aspect of low bit rate image coding: the coding of a binary map (a wavelet tree) indicating the locations of the non-zero values, otherwise known as the significance map of the transform coefficients.
  • Quantization and entropy coding are then used to achieve very low bit rates. It follows that a significant improvement in the proper selection of a quantizer scale for encoding the significance map (the wavelet tree) will translate into a significant improvement in compression efficiency and coding rate.
  • rate control can be implemented at lower levels within a frame e.g., at the macroblock or block levels.
  • An embodiment of the present invention is a method and apparatus for selecting a quantizer scale for each block, e.g., a macroblock, within each frame to maintain the overall quality of the video image while optimizing the coding rate. Namely, a quantizer scale is selected for each macroblock within each frame (picture) such that the target bit rate for the picture is achieved while maintaining a uniform visual quality over the entire frame.
  • FIG. 1 illustrates a block diagram of the apparatus of the present invention
  • FIG. 2 illustrates a block diagram of a flowchart of a method for deriving and allocating the target bit rate for an image based on blocks within the image;
  • FIG. 3 illustrates a flowchart of a method for determining a target frame bit rate
  • FIG. 4 illustrates a flowchart of a method for determining one or more target macroblock bit rates for the macroblocks within the current image
  • FIG. 5 illustrates a block diagram of a second embodiment of the apparatus of the present invention
  • FIG. 6 is a graphical representation of a wavelet tree
  • FIG. 7 illustrates an encoding system of the present invention. To facilitate understanding, identical reference numerals have been used, where possible, to designate identical elements that are common to the figures.
  • FIG. 1 depicts a block diagram of the apparatus 100 of the present invention for deriving a quantizer scale for each macroblock within each frame to maintain the overall quality of the video image while controlling the coding rate.
  • the present invention is described below with reference to a MPEG compliant encoder, those skilled in the art will realize that the present invention can be adapted to other encoders that are compliant with other coding/decoding standards.
  • the apparatus 100 is an encoder or a portion of a more complex block-based motion compensation coding system.
  • the apparatus 100 comprises a motion estimation module 140, a motion compensation module 150, a rate control module 130, a DCT module 160, a quantization (Q) module 170, a variable length coding (VLC) module 180, a buffer 190, an inverse quantization (Q 1 ) module 175, an inverse DCT (DCT 1 ) transform module 165, a subtractor 115 and a summer 155.
  • the apparatus 100 comprises a plurality of modules, those skilled in the art will realize that the functions performed by the various modules are not required to be isolated into separate modules as shown in FIG. 1.
  • FIG. 1 illustrates an input image (image sequence) on signal path 110 which is digitized and represented as a luminance and two color difference signals (Y, C r , C b ) in accordance with the MPEG standards. These signals are further divided into a plurality of layers (sequence, group of pictures, picture, slice, macroblock and block) such that each picture (frame) is represented by a plurality of macroblocks.
  • Each macroblock comprises four (4) luminance blocks, one C r block and one C b block where a block is defined as an eight (8) by eight (8) sample array.
  • the division of a picture into block units improves the ability to discern changes between two successive pictures and improves image compression through the elimination of low amplitude transformed coefficients (discussed below).
  • the digitized signal may optionally undergo preprocessing such as format conversion for selecting an appropriate window, resolution and input format.
  • macroblock or block is intended to describe a block of pixels of any size or shape. Broadly speaking, a "macroblock" could be as small as a single pixel, or as large as an entire video frame.
  • the input image on path 110 is received into motion estimation module 140 for estimating motion vectors.
  • a motion vector is a two- dimensional vector which is used by motion compensation to provide an offset from the coordinate position of a block in the current picture to the coordinates in a reference frame.
  • the reference frames can be a previous frame (P-frame), or previous and/or future frames (B-frames).
  • the use of motion vectors greatly enhances image compression by reducing the amount of information that is transmitted on a channel because only the changes between the current and reference frames are coded and transmitted.
  • the motion vectors from the motion estimation module 140 are received by the motion compensation module 150 for improving the efficiency of the prediction of sample values.
  • Motion compensation involves a prediction that uses motion vectors to provide offsets into the past and/or future reference frames containing previously decoded sample values that are used to form the prediction error.
  • the motion compensation module 150 uses the previously decoded frame and the motion vectors to construct an estimate of the current frame.
  • a coding mode Prior to performing motion compensation prediction for a given macroblock, a coding mode must be selected.
  • MPEG provides a plurality of different macroblock coding modes. Specifically, MPEG-2 provides macroblock coding modes which include intra mode, no motion compensation mode (No MC), frame/field/dual-prime motion compensation inter mode, forward/backward/average inter mode and field/frame DCT mode.
  • motion compensation module 150 generates a motion compensated prediction (predicted image) on path 152 of the contents of the block based on past and/or future reference pictures.
  • This motion compensated prediction on path 152 is subtracted via sub tractor 115 from the video image on path 110 in the current macroblock to form an error signal or predictive residual signal on path 153.
  • the formation of the predictive residual signal effectively removes redundant information in the input video image. It should be noted that if a current frame is encoded as an I-frame, then the signal on path 153 is simply the original picture and not a predictive residual signal.
  • the DCT module 160 then applies a forward discrete cosine transform process to each block of the predictive residual signal to produce a set of eight (8) by eight (8) block of DCT coefficients.
  • the DCT basis function or subband decomposition permits effective use of psychovisual criteria which is important for the next step of quantization.
  • the resulting 8 x 8 block of DCT coefficients is received by quantization module 170 where the DCT coefficients are quantized.
  • the process of quantization reduces the accuracy with which the DCT coefficients are represented by dividing the DCT coefficients by a set of quantization values with appropriate rounding to form integer values.
  • the quantization values can be set individually for each DCT coefficient, using criteria based on the visibility of the basis functions (known as visually weighted quantization). Namely, the quantization value corresponds to the threshold for visibility of a given basis function, i.e., the coefficient amplitude that is just detectable by the human eye. By quantizing the DCT coefficients with this value, many of the DCT coefficients are converted to the value "zero", thereby improving image compression efficiency.
  • the process of quantization is a key operation and is an important tool to achieve visual quality and to control the encoder to match its output to a given bit rate (rate control). Since a different quantization value can be applied to each DCT coefficient, a "quantization matrix" is generally established as a reference table, e.g., a luminance quantization table or a chrominance quantization table. Thus, the encoder chooses a quantization matrix that determines how each frequency coefficient in the transformed block is quantized.
  • variable length coding (VLC) module 180 receives the resulting 8 x 8 block of quantized DCT coefficients via signal connection 171, where the two-dimensional block of quantized coefficients is scanned in a "zig-zag" order to convert it into a one-dimensional string of quantized DCT coefficients.
  • This zig-zag scanning order is an approximate sequential ordering of the DCT coefficients from the lowest spatial frequency to the highest.
  • Variable length coding (VLC) module 180 then encodes the string of quantized DCT coefficients and all side-information for the macroblock using variable length coding and run-length coding.
  • the data stream is received into a "First In-First Out" (FIFO) buffer 190.
  • FIFO First In-First Out
  • the output signal of FIFO buffer 190 on path 195 is a compressed representation of the input video image on path 110 (or a compressed difference signal between the input image and a predicted image), where it is sent to a storage medium or telecommunication channel via path 195.
  • the rate control module 130 serves to monitor and adjust the bit rate of the data stream entering the FIFO buffer 190 to prevent overflow and underflow on the decoder side (within a receiver or target storage device, not shown) after transmission of the data stream. Thus, it is the task of the rate control module 130 to monitor the status of buffer 190 to control the number of bits generated by the encoder.
  • rate control module 130 selects a quantizer scale for each block, e.g., a macroblock within each frame to maintain the overall quality of the video image while controlling the coding rate. Namely, a frame can be evaluated to determine if certain blocks within the frame require more or less bit rate allocation. It has been observed that for different applications, various blocks are of more interest than other blocks, e.g., the face of a person in a video phone application is more important to a human viewer than the background in general.
  • a quantizer scale is selected for each macroblock within each -J7- frame such that target bit rate for the frame is achieved while maintaining a uniform visual quality over the entire frame.
  • the rate control module 130 comprises a frame rate allocator 131 and a macroblock rate allocator 132.
  • the frame rate allocator 131 allocates a bit budget (target frame bit rate) for a current frame
  • the macroblock rate allocator 132 allocates a bit budget (target macroblock bit rate or target block bit rate) for each macroblock within the current frame.
  • the frame rate allocator 131 initially obtains a rough estimate of the complexity of a specific type of picture (I, P, B) from previously encoded pictures or by implementing various MPEG test models. This estimated complexity is used to derive a predicted number of bits necessary to code each frame.
  • a quantizer scale is calculated for the frame in accordance with a complexity measure having a polynomial form. This complexity measure is derived to meet the constraint that the selected quantizer scale for the frame should approach the target bit rate for the picture.
  • the rate control module recursively adjusts the complexity measure through the use of a polynomial regression process. That is, the actual number of bits necessary to code the macroblock is used to refine the complexity measure so as to improve the prediction of a quantizer scale for the next frame.
  • the "target frame bit rate" is also recursively updated.
  • This frame rate allocating method was disclosed in patent application entitled “Apparatus And Method For Optimizing The Rate Control In A Coding System", filed on February 11, 1998, with serial number 09/022,349 (attorney docket SAR12459), which is incorporated herein by reference. It should be understood that the present invention can be implemented using other frame bit rate allocating methods, e.g., frame bit rate allocating methods that are based on a distortion measure and the like.
  • the macroblock rate allocator 132 then applies the calculated target frame bit rate to determine one or more target macroblock bit rates, where the bits of the target frame bit rate are distributed proportional to the mean of the absolute differences (MAD) and the weighting for a macroblock.
  • MAD mean of the absolute differences
  • the importance of a macroblock is determined by an optional macroblock classifying module 120.
  • the macroblock classifying module 120 contains the necessary criteria to define the importance of the macroblocks within each frame.
  • Various macroblock classifying methods are available, e.g., as disclosed in patent application entitled “Apparatus And Method For Employing M-Ary Pyramids To Enhance Feature-Based Classification And Motion Estimation", filed on December 31, 1997, with serial number 09/002258 (attorney docket SAR12626), which is incorporated herein by reference.
  • the "importance" of a macroblock is accounted through the use of weighting as described below. In brief, if a macroblock is very important, then more bits are allocated to the macroblock, whereas if a macroblock is not very important, then less bits are allocated to the macroblock.
  • the "macroblock based" information e.g., which macroblocks are more important, which macroblocks carry what type of information, e.g., foreground, background, or objects in a frame and the like, can be obtained directly from the image sequence on path 112. Namely, if the image sequence was previously processed and stored on a storage medium, e.g., a stored video sequence or program on a server, then it is possible that the encoder that generated the stored video sequence may pass along "macroblock based" information. In other words, "macroblock based" information can be transmitted to the encoder 100 along with the image sequence. In such implementation, the macroblock classifying module 120 can be omitted, since the macroblock based information is readily available.
  • a storage medium e.g., a stored video sequence or program on a server
  • the resulting 8 x 8 block of quantized DCT coefficients from the quantization module 170 is also received by the inverse quantization module 175 via signal connection 172.
  • the encoder regenerates I-frames and P-frames of the input video image by decoding the data so that they are used as reference frames for subsequent encoding.
  • the resulting dequantized 8 x 8 block of DCT coefficients are passed to the inverse DCT module 165 where inverse DCT is applied to each macroblock to produce the decoded error signal.
  • This error signal is added back to the prediction signal from the motion compensation module via summer 155 to produce a decoded reference picture (reconstructed image).
  • FIG. 2 depicts a block diagram of a flowchart of a method 200 for deriving and allocating bits for an image based on macroblocks within the image. More specifically, method 200 starts in step 205 and proceeds to step 210 where a target frame bit rate is determined for a current frame. In the preferred embodiment, the target frame bit rate is determined using a complexity measure that is recursively adjusted through the use of a polynomial regression process (as illustrated in FIG. 3).
  • FIG. 3 illustrates a flowchart of a method 300 for determining a target frame bit rate.
  • the method begins at step 305 and proceeds to step 310, where the method determines the target bits (target frame bit rate) for a frame, T frame as:
  • T fmme — x (1 - past _ percent) + T previousframe x past _ percent (1)
  • R is the remaining number of bits for a sequence of frame
  • Nf is the number of remaining frames in the sequence
  • T pre ⁇ iousfra ⁇ ne is the number of bits used for encoding the previous frame
  • the past_percent is a constant.
  • the constant past_percent is selected to be 0.05.
  • the present invention is not so limited. Other values can be employed, that depend on the specific applications or the context of the images. In fact, these values can be adjusted temporally.
  • equation (1) allows the target frame bit rate to be computed based on the bits available and the last encoded frame bits. If the last frame is complex and uses many bits, it leads to the premise that more bits should be assigned to the current frame. However, this increased allocation will diminish the available number of bits for encoding the remaining frames, thereby limiting the increased allocation to this frame.
  • a weighted average reflects a compromise of these two factors, as illustrated in the second term in equation (1).
  • step 320 method 300 then adjusts the calculated target frame bit rate, T fr by the current buffer fullness as:
  • T frame is the adjusted target bit rate
  • "a" is the current buffer fullness (the portion of the buffer that contains bits to be sent to the decoder)
  • b is (the physical buffer size - buffer fullness (a))
  • c is a constant selected to be a value of 2 (other values can be used).
  • equation (3) allows T fram ⁇ to take the greater (max) of two possible values, where R s is a bit rate for the sequence (or segment), e.g., 24000 bits/sec). Namely, a lower bound of target rate (R/30) is used to maintain or guarantee a minimal quality, e.g., 800 bits/frame can be set as a minimum. If the minimal quality cannot be maintained, the encoder has the option to skip the current frame altogether. Method 300 then ends in step 340.
  • frame bit rate allocation methods can be used, e.g., the MPEG TM4 and TM5, with the present invention.
  • the computational overhead and the accuracy of the target macroblock bit rates are affected by the frame bit rate allocation method that is employed. For example, if it is desirable to minimize computational complexity at the expense of performance, then it may be appropriate to employ the frame bit rate allocation methods of MPEG TM4 and TM5. In contrast, if it is desirable to maximize performance at the expense of increasing computational complexity, then it may be appropriate to employ the frame bit rate allocation method of the patent application (attorney docket SAR12459), or other more complex frame bit rate allocation methods.
  • method 200 determines one or more target macroblock bit rates for the macroblocks within the current image as illustrated in FIG. 4 below. Method 200 then ends in step 230.
  • FIG. 4 illustrates a flowchart of a method 400 for determining one or more target macroblock bit rates for the macroblocks within the current image.
  • the method starts in step 405 and proceeds to step 410, where a sum of absolute difference (SAD) S j is performed for each macroblock i. Namely, the absolute difference between each pixel value (in the original image) and the corresponding pixel value (in the predicted image) is performed for pixels defined within the macroblock. Next, the sum of all the absolute differences of the pixels for the macroblock is performed to generate the SAD for the macroblock i.
  • method 400 queries whether S ⁇ is greater than the threshold H .
  • the threshold H q is selected to remove various macroblocks from the present macroblock bit allocation method.
  • the threshold H q serves to eliminate various macroblocks from consideration.
  • the threshold H q is selected as the average of all the mean absolute difference (MADs) that have been skipped in the previous frame, where MAD j is defined as the S, divided by the number of pixels in the macroblock i. However, if the current frame is the first frame in the image sequence, then H q is set to be half of the average of all the existing MADs in this current frame.
  • MADs mean absolute difference
  • Mad w K, «- «- for Mad k having SAD k > H (5)
  • Mad ⁇ , k , where Mad ; is the mean absolute difference (MAD) of an macroblock i, "n" is the number of macroblocks in a frame, "w” is a weighting factor and R j is the estimated target macroblock bit rate for macroblock i.
  • the weighting factor w allows R ; for a given macroblock to be adjusted in accordance with other criteria. Namely, it has been observed that some macroblocks can be viewed as being more important than other macroblocks, where "importance" is not based solely on the MAD of an macroblock.
  • Various applications as illustrated above, may place emphasis on certain macroblocks.
  • By incorporating an optional weighting factor w more or less bits can be allocated to a particular macroblock based upon application specific criteria. In the preferred embodiment, w is set to a value of "1", but other values can be used as necessary for a particular application.
  • the MADs of macroblocks a and b are not used in the determination of R c and R d .
  • the above example illustrates the need to update T frame after each R ; is allocated to a macroblock.
  • method 400 in step 440 calculates a quantization scale Q j for each macroblock i.
  • the quantization scale Q is calculated in accordance with a distortion measure as described in US patent application with serial number 09/022,349 (attorney docket SAR12459).
  • E represents a distortion measure.
  • E ] represents a sum of all the mean absolute differences (MADs) for those macroblocks having R, for the current frame.
  • the measure E t provides a method of adjusting the macroblock bit budget to account for the differences in the macroblock between successive frames in a sequence.
  • the greater the differences between an macroblock in the current frame and the same macroblock in a previous frame the greater the number of bits that will be required to code the macroblock in the current frame.
  • other distortion measures can be used, such that E, may represent mean square error or just-noticeable difference (jnd).
  • R a represents the bit rate for the sequence (or segment), e.g., 24000 bits per second.
  • N s represents the distance between encoded frames. Namely, due to low bit rate applications, certain frames within a sequence may not be encoded (skipped), e.g., the encoder may only encode every fourth frame. It should be understood that the number of skipped frames can be tailored to the requirement of a particular application.
  • the quantizer scale Q t can be determined in accordance with equations (6) and (7).
  • the calculated Q is limited by the condition that it should not be varied too significantly from macroblock to macroblock. Namely Q, is limited by the conditions:
  • Q last is the quantizer scale for an immediate previous macroblock having an R
  • "e" is a constant selected to be a value "2" (other values can be used depending on the application).
  • the calculated Q t is limited by a change to be no greater than a value of "2" from macroblock to macroblock. This limitation maintains uniform visual quality, i.e., minimizing significant changes in visual quality from macroblock to macroblock. It should be noted that if an immediate macroblock is skipped (a macroblock without an R,), then Q last is based on the next immediate "non- skipped" macroblock.
  • step 450 the Q, is used to encode the macroblock to generate an "actual R t " (the actual number of bits used to encode the macroblock).
  • the actual R t and the Q ⁇ are used to update the parameters X and X j , using a polynomial regression model or a quadratic regression model to refine the parameters X t and X.
  • the constants X x and X are updated to account for the discrepancy between the bits allocated to an macroblock and the actual number of bits needed to the code the macroblock for a particular quantizer level or scale.
  • Regression models are well known in the art. For a detailed discussion of various regression models, see e.g., Bowerman and O'Connell, Forecasting and Time Series. 3rd Edition, Duxbury Press, (1993 , chapter 4).
  • T frame is then updated by subtracting the actual R,:
  • the actual T frame (the actual number of bits used to encode the frame) can then be used to update the method that is -lo- tasked with generating a target frame bit rate, e.g. as illustrated in FIG. 3 above.
  • step 470 method 400 queries whether there is a next macroblock in the frame. If the query is negatively answered, then method 400 ends in step 480. If the query is positively answered, then method 400 returns to step 430, where steps 430-460 are repeated until all the macroblocks are evaluated in the current frame.
  • the encoder contains a block motion compensator (BMC) and motion vector coder 504, subtractor 502, discrete wavelet transform (DWT) coder 506, bit rate controller 510, DWT decoder 512 and output buffer 514.
  • BMC block motion compensator
  • DWT discrete wavelet transform
  • the encoder has three functions: first, it produces, using the BMC and its coder 504, a plurality of motion vectors that represent motion that occurs between frames; second, it predicts the present frame using a reconstructed version of the previous frame combined with the motion vectors; and third, the predicted frame is subtracted from the present frame to produce a frame of residuals that are coded and transmitted along with the motion vectors to a receiver.
  • the discrete wavelet transform performs a wavelet hierarchical subband decomposition to produce a conventional wavelet tree representation of the input image.
  • the image is decomposed using times two subsampling into high horizontal-high vertical (HH), high horizontal-low vertical (HL), low horizontal-high vertical (LH), and low horizontal-low vertical (LL), frequency subbands.
  • the LL subband is then further subsampled times two to produce a set of HH, HL, LH and LL subbands.
  • This subsampling is accomplished recursively to produce an array of subbands such as that illustrated in FIG. 6 where three subsamplings have been used. Preferably six subsamplings are used in practice.
  • the parent-child dependencies between subbands are illustrated as arrows pointing from the subband of the parent nodes to the subbands of the child nodes.
  • the lowest frequency subband is the top left LL 1? and the highest frequency subband is at the bottom right HH 3 . In this example, all child nodes have one parent.
  • a detailed discussion of subband decomposition is presented in J.M. Shapiro, "Embedded Image Coding Using Zerotrees of Wavelet Coefficients", IEEE Trans, on Signal Processing, Vol. 41, No. 12, pp. 3445-62, December 1993.
  • the DWT coder of FIG. 5 codes the coefficients of the wavelet tree in either a "breadth first" or "depth first” pattern.
  • a breadth first pattern traverse the wavelet tree in a bit-plane by bit-plane pattern, i.e., quantize all parent nodes, then all children, then all grandchildren and so on.
  • a depth first pattern traverses each tree from the root in the low-low subband (LL X ) through the children (top down) or children through the low-low subband (bottom up).
  • the selection of the proper quantization level by the rate controller 510 is as discussed above to control the bit rate for each macroblock within each frame of a sequence.
  • the present invention can be adapted to various types of encoders that use different transforms.
  • FIG. 7 illustrates an encoding system 700 of the present invention.
  • the encoding system comprises a general purpose computer 710 and various input/output devices 720.
  • the general purpose computer comprises a central processing unit (CPU) 712, a memory 714 and an encoder 716 for receiving and encoding a sequence of images.
  • CPU central processing unit
  • the encoder 716 is simply the encoder 100 and/or encoder 500 as discussed above.
  • the encoder 716 can be a physical device which is coupled to the CPU 712 through a communication channel.
  • the encoder 716 can be represented by a software application (or a combination of software and hardware, e.g., application specific integrated circuits (ASIC)) which is loaded from a storage device and resides in the memory 712 of the computer.
  • ASIC application specific integrated circuits
  • the encoder 100 and 500 of the present invention can be stored on a computer readable medium, e.g., a memory or storage device.
  • the computer 710 can be coupled to a plurality of input and output devices 720, such as a keyboard, a mouse, a camera, a camcorder, a video monitor, any number of imaging devices or storage devices, including but not limited to, a tape drive, a floppy drive, a hard disk drive or a compact disk drive.
  • the input devices serve to provide inputs to the computer for producing the encoded video bitstreams or to receive the sequence of video images from a storage device or an imaging device.
  • a communication channel 730 is shown where the encoded signal from the encoding system is forwarded to a decoding system (not shown).

Landscapes

  • Engineering & Computer Science (AREA)
  • Multimedia (AREA)
  • Signal Processing (AREA)
  • Theoretical Computer Science (AREA)
  • Computing Systems (AREA)
  • Bioinformatics & Computational Biology (AREA)
  • Artificial Intelligence (AREA)
  • Bioinformatics & Cheminformatics (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Data Mining & Analysis (AREA)
  • Evolutionary Biology (AREA)
  • Evolutionary Computation (AREA)
  • Physics & Mathematics (AREA)
  • General Engineering & Computer Science (AREA)
  • General Physics & Mathematics (AREA)
  • Compression Or Coding Systems Of Tv Signals (AREA)
  • Compression, Expansion, Code Conversion, And Decoders (AREA)

Abstract

A method and apparatus for selecting a quantizer scale for each macroblock within a frame to optimize the coding rate is presented (130). A quantizer scale is selected for each macroblock (120) within each frame such that the target bit rate for the frame is achieved while maintaining a uniform visual quality over the entire frame.

Description

APPARATUS AND METHOD FOR MACROBLOCK BASED RATE CONTROL IN A CODING SYSTEM
This application claims the benefit of U.S. Provisional Application No. 60/052,437 filed July 14, 1997, which is herein incorporated by reference.
The present invention relates to an apparatus and concomitant method for optimizing the coding of motion video. More particularly, this invention relates to a method and apparatus that adaptively adjusts a quantizer scale for each macroblock within a frame to maintain the overall quality of the motion video while optimizing the coding rate.
BACKGROUND OF THE INVENTION The Moving Picture Experts Group (MPEG) created the ISO/IEC international Standards 11172 and 13818 (generally referred to as MPEG-1 and MPEG-2 format respectively) to establish a standard for coding/decoding strategies. Although these MPEG standards specify a general coding methodology and syntax for generating an MPEG compliant bitstream, many variations are permitted to accommodate a plurality of different applications and services such as desktop video publishing, video conferencing, digital storage media and television broadcast.
In the area of rate control, MPEG does not define a specific method for controlling the bit rate of an encoder. It is the task of the encoder designer to devise a rate control process for controlling the bit rate such that the decoder input buffer neither overflows nor underflows.
Currently, one way of controlling the bit rate is to alter the quantization process, which will affect the distortion of the input video image. By altering the quantizer scale (step size), the bit rate can be changed and controlled.
Although changing the quantizer scale is an effective method of implementing the rate control of an encoder, it has been shown that a poor rate control process will actually degrade the visual quality of the video image, i.e., failing to alter the quantizer scale in an efficient manner such that it is necessary to drastically alter the quantizer scale toward the end of a picture to avoid overflow and underflow conditions. Since altering the quantizer scale affects both image quality and compression efficiency, it is important for a rate control process to control the bit rate without sacrificing image quality.
In the current MPEG coding strategies (e.g., various MPEG test models), the quantizer scale for each frame is selected by assuming that all the pictures of the same type have identical complexity within a group of pictures. However, the quantizer scale selected by this criterion may not achieve optimal coding performance, since the complexity of each picture will vary with time.
Furthermore, encoders that utilize global-type transforms, e.g., wavelet transform (otherwise known as hierarchical subband decomposition), have similar problems. For example, wavelet transforms are applied to an important aspect of low bit rate image coding: the coding of a binary map (a wavelet tree) indicating the locations of the non-zero values, otherwise known as the significance map of the transform coefficients. Quantization and entropy coding are then used to achieve very low bit rates. It follows that a significant improvement in the proper selection of a quantizer scale for encoding the significance map (the wavelet tree) will translate into a significant improvement in compression efficiency and coding rate. Furthermore, rate control can be implemented at lower levels within a frame e.g., at the macroblock or block levels. However, macroblock level rate control is generally more costly, since there is an additional overhead if quantization parameter (quantizer scale) is changed within a frame. Namely, more bits are needed to communicate to the decoder of the different quantizer scales for different macroblocks within each frame. This criticality is exacerbated in low bit rate applications, where proper bit management is very important. -o-
Therefore, a need exists in the art for an apparatus and method that adaptively adjusts a quantizer scale for each macroblock within a frame to maintain the overall quality of the motion video while optimizing the coding rate.
SUMMARY OF THE INVENTION An embodiment of the present invention is a method and apparatus for selecting a quantizer scale for each block, e.g., a macroblock, within each frame to maintain the overall quality of the video image while optimizing the coding rate. Namely, a quantizer scale is selected for each macroblock within each frame (picture) such that the target bit rate for the picture is achieved while maintaining a uniform visual quality over the entire frame.
BRIEF DESCRIPTION OF THE DRAWINGS The teachings of the present invention can be readily understood by considering the following detailed description in conjunction with the accompanying drawings, in which:
FIG. 1 illustrates a block diagram of the apparatus of the present invention; FIG. 2 illustrates a block diagram of a flowchart of a method for deriving and allocating the target bit rate for an image based on blocks within the image;
FIG. 3 illustrates a flowchart of a method for determining a target frame bit rate; FIG. 4 illustrates a flowchart of a method for determining one or more target macroblock bit rates for the macroblocks within the current image;
FIG. 5 illustrates a block diagram of a second embodiment of the apparatus of the present invention; FIG. 6 is a graphical representation of a wavelet tree; and
FIG. 7 illustrates an encoding system of the present invention. To facilitate understanding, identical reference numerals have been used, where possible, to designate identical elements that are common to the figures.
DETAILED DESCRIPTION
FIG. 1 depicts a block diagram of the apparatus 100 of the present invention for deriving a quantizer scale for each macroblock within each frame to maintain the overall quality of the video image while controlling the coding rate. Although the present invention is described below with reference to a MPEG compliant encoder, those skilled in the art will realize that the present invention can be adapted to other encoders that are compliant with other coding/decoding standards.
In the preferred embodiment of the present invention, the apparatus 100 is an encoder or a portion of a more complex block-based motion compensation coding system. The apparatus 100 comprises a motion estimation module 140, a motion compensation module 150, a rate control module 130, a DCT module 160, a quantization (Q) module 170, a variable length coding (VLC) module 180, a buffer 190, an inverse quantization (Q 1) module 175, an inverse DCT (DCT1) transform module 165, a subtractor 115 and a summer 155. Although the apparatus 100 comprises a plurality of modules, those skilled in the art will realize that the functions performed by the various modules are not required to be isolated into separate modules as shown in FIG. 1. For example, the set of modules comprising the motion compensation module 150, inverse quantization module 175 and inverse DCT module 165 is generally known as an "embedded decoder". FIG. 1 illustrates an input image (image sequence) on signal path 110 which is digitized and represented as a luminance and two color difference signals (Y, Cr, Cb) in accordance with the MPEG standards. These signals are further divided into a plurality of layers (sequence, group of pictures, picture, slice, macroblock and block) such that each picture (frame) is represented by a plurality of macroblocks. Each macroblock comprises four (4) luminance blocks, one Cr block and one Cb block where a block is defined as an eight (8) by eight (8) sample array. The division of a picture into block units improves the ability to discern changes between two successive pictures and improves image compression through the elimination of low amplitude transformed coefficients (discussed below). The digitized signal may optionally undergo preprocessing such as format conversion for selecting an appropriate window, resolution and input format.
The following disclosure uses the MPEG standard terminology; however, it should be understood that the term macroblock or block is intended to describe a block of pixels of any size or shape. Broadly speaking, a "macroblock" could be as small as a single pixel, or as large as an entire video frame.
The input image on path 110 is received into motion estimation module 140 for estimating motion vectors. A motion vector is a two- dimensional vector which is used by motion compensation to provide an offset from the coordinate position of a block in the current picture to the coordinates in a reference frame. The reference frames can be a previous frame (P-frame), or previous and/or future frames (B-frames). The use of motion vectors greatly enhances image compression by reducing the amount of information that is transmitted on a channel because only the changes between the current and reference frames are coded and transmitted. The motion vectors from the motion estimation module 140 are received by the motion compensation module 150 for improving the efficiency of the prediction of sample values. Motion compensation involves a prediction that uses motion vectors to provide offsets into the past and/or future reference frames containing previously decoded sample values that are used to form the prediction error. Namely, the motion compensation module 150 uses the previously decoded frame and the motion vectors to construct an estimate of the current frame. Furthermore, those skilled in the art will realize that the functions performed by the motion estimation module and the motion compensation module can be implemented in a combined module, e.g., a single block motion compensator. Furthermore, prior to performing motion compensation prediction for a given macroblock, a coding mode must be selected. In the area of coding mode decision, MPEG provides a plurality of different macroblock coding modes. Specifically, MPEG-2 provides macroblock coding modes which include intra mode, no motion compensation mode (No MC), frame/field/dual-prime motion compensation inter mode, forward/backward/average inter mode and field/frame DCT mode.
Once a coding mode is selected, motion compensation module 150 generates a motion compensated prediction (predicted image) on path 152 of the contents of the block based on past and/or future reference pictures. This motion compensated prediction on path 152 is subtracted via sub tractor 115 from the video image on path 110 in the current macroblock to form an error signal or predictive residual signal on path 153. The formation of the predictive residual signal effectively removes redundant information in the input video image. It should be noted that if a current frame is encoded as an I-frame, then the signal on path 153 is simply the original picture and not a predictive residual signal.
The DCT module 160 then applies a forward discrete cosine transform process to each block of the predictive residual signal to produce a set of eight (8) by eight (8) block of DCT coefficients. The DCT basis function or subband decomposition permits effective use of psychovisual criteria which is important for the next step of quantization.
The resulting 8 x 8 block of DCT coefficients is received by quantization module 170 where the DCT coefficients are quantized. The process of quantization reduces the accuracy with which the DCT coefficients are represented by dividing the DCT coefficients by a set of quantization values with appropriate rounding to form integer values. The quantization values can be set individually for each DCT coefficient, using criteria based on the visibility of the basis functions (known as visually weighted quantization). Namely, the quantization value corresponds to the threshold for visibility of a given basis function, i.e., the coefficient amplitude that is just detectable by the human eye. By quantizing the DCT coefficients with this value, many of the DCT coefficients are converted to the value "zero", thereby improving image compression efficiency. The process of quantization is a key operation and is an important tool to achieve visual quality and to control the encoder to match its output to a given bit rate (rate control). Since a different quantization value can be applied to each DCT coefficient, a "quantization matrix" is generally established as a reference table, e.g., a luminance quantization table or a chrominance quantization table. Thus, the encoder chooses a quantization matrix that determines how each frequency coefficient in the transformed block is quantized.
However, subjective perception of quantization error greatly varies with the frequency and it is advantageous to use coarser quantization values for the higher frequencies. Namely, human perceptual sensitivity of quantization errors are lower for the higher spatial frequencies. As a result, high frequencies are quantized more coarsely with fewer allowed values than low frequencies. Furthermore, an exact quantization matrix depends on many external parameters such as the characteristics of the intended display, the viewing distance and the amount of noise in the source. Thus, it is possible to tailor a particular quantization matrix for an application or even for an individual sequence of frames. Generally, a customized quantization matrix can be stored as context together with the compressed video image. The proper selection of a quantizer scale is performed by the rate control module 130.
Next, the resulting 8 x 8 block of quantized DCT coefficients is received by variable length coding (VLC) module 180 via signal connection 171, where the two-dimensional block of quantized coefficients is scanned in a "zig-zag" order to convert it into a one-dimensional string of quantized DCT coefficients. This zig-zag scanning order is an approximate sequential ordering of the DCT coefficients from the lowest spatial frequency to the highest. Variable length coding (VLC) module 180 then encodes the string of quantized DCT coefficients and all side-information for the macroblock using variable length coding and run-length coding. The data stream is received into a "First In-First Out" (FIFO) buffer 190. A consequence of using different picture types and variable length coding is that the overall bit rate into the FIFO is variable. Namely, the number of bits used to code each frame can be different. In applications that involve a fixed-rate channel, a FIFO buffer is used to match the encoder output to the channel for smoothing the bit rate. Thus, the output signal of FIFO buffer 190 on path 195 is a compressed representation of the input video image on path 110 (or a compressed difference signal between the input image and a predicted image), where it is sent to a storage medium or telecommunication channel via path 195.
The rate control module 130 serves to monitor and adjust the bit rate of the data stream entering the FIFO buffer 190 to prevent overflow and underflow on the decoder side (within a receiver or target storage device, not shown) after transmission of the data stream. Thus, it is the task of the rate control module 130 to monitor the status of buffer 190 to control the number of bits generated by the encoder.
In the preferred embodiment of the present invention, rate control module 130 selects a quantizer scale for each block, e.g., a macroblock within each frame to maintain the overall quality of the video image while controlling the coding rate. Namely, a frame can be evaluated to determine if certain blocks within the frame require more or less bit rate allocation. It has been observed that for different applications, various blocks are of more interest than other blocks, e.g., the face of a person in a video phone application is more important to a human viewer than the background in general. Other examples include medical applications, where certain blocks of an image, i.e., a potential tumor is more important than the surrounding tissues or in surveillance applications, where certain blocks of an image, i.e., a military assess is more important than the surrounding camouflage, and so on. Thus, the particular application will dictate the criteria that define the importance of relevant blocks within a frame. In the present invention, a quantizer scale is selected for each macroblock within each -J7- frame such that target bit rate for the frame is achieved while maintaining a uniform visual quality over the entire frame.
It should be understood that although the present invention is described with an encoder implementing temporal (e.g., motion estimation/compensation) and spatial encoding (e.g., discreet cosine transform), the present invention is not so limited. Other temporal and spatial encoding methods can be used, including no use of any temporal and spatial encoding.
Specifically, the rate control module 130 comprises a frame rate allocator 131 and a macroblock rate allocator 132. The frame rate allocator 131 allocates a bit budget (target frame bit rate) for a current frame, whereas the macroblock rate allocator 132 allocates a bit budget (target macroblock bit rate or target block bit rate) for each macroblock within the current frame. In brief, the frame rate allocator 131 initially obtains a rough estimate of the complexity of a specific type of picture (I, P, B) from previously encoded pictures or by implementing various MPEG test models. This estimated complexity is used to derive a predicted number of bits necessary to code each frame. With this knowledge, a quantizer scale is calculated for the frame in accordance with a complexity measure having a polynomial form. This complexity measure is derived to meet the constraint that the selected quantizer scale for the frame should approach the target bit rate for the picture. Once the frame is encoded, the rate control module recursively adjusts the complexity measure through the use of a polynomial regression process. That is, the actual number of bits necessary to code the macroblock is used to refine the complexity measure so as to improve the prediction of a quantizer scale for the next frame. In the course of computing the quantizer scale, the "target frame bit rate" is also recursively updated. This frame rate allocating method was disclosed in patent application entitled "Apparatus And Method For Optimizing The Rate Control In A Coding System", filed on February 11, 1998, with serial number 09/022,349 (attorney docket SAR12459), which is incorporated herein by reference. It should be understood that the present invention can be implemented using other frame bit rate allocating methods, e.g., frame bit rate allocating methods that are based on a distortion measure and the like. In brief, the macroblock rate allocator 132 then applies the calculated target frame bit rate to determine one or more target macroblock bit rates, where the bits of the target frame bit rate are distributed proportional to the mean of the absolute differences (MAD) and the weighting for a macroblock. A detailed description of the target macroblock bit rate and corresponding quantizer scale selection method is discussed below with reference to FIG. 4.
However, due to human visual responses, some macroblocks may be deemed to be more important than other macroblock to a human viewer. The importance of a macroblock is determined by an optional macroblock classifying module 120. The macroblock classifying module 120 contains the necessary criteria to define the importance of the macroblocks within each frame. Various macroblock classifying methods are available, e.g., as disclosed in patent application entitled "Apparatus And Method For Employing M-Ary Pyramids To Enhance Feature-Based Classification And Motion Estimation", filed on December 31, 1997, with serial number 09/002258 (attorney docket SAR12626), which is incorporated herein by reference. The "importance" of a macroblock is accounted through the use of weighting as described below. In brief, if a macroblock is very important, then more bits are allocated to the macroblock, whereas if a macroblock is not very important, then less bits are allocated to the macroblock.
Alternatively, the "macroblock based" information, e.g., which macroblocks are more important, which macroblocks carry what type of information, e.g., foreground, background, or objects in a frame and the like, can be obtained directly from the image sequence on path 112. Namely, if the image sequence was previously processed and stored on a storage medium, e.g., a stored video sequence or program on a server, then it is possible that the encoder that generated the stored video sequence may pass along "macroblock based" information. In other words, "macroblock based" information can be transmitted to the encoder 100 along with the image sequence. In such implementation, the macroblock classifying module 120 can be omitted, since the macroblock based information is readily available.
Returning to FIG. 1, the resulting 8 x 8 block of quantized DCT coefficients from the quantization module 170 is also received by the inverse quantization module 175 via signal connection 172. At this stage, the encoder regenerates I-frames and P-frames of the input video image by decoding the data so that they are used as reference frames for subsequent encoding.
The resulting dequantized 8 x 8 block of DCT coefficients are passed to the inverse DCT module 165 where inverse DCT is applied to each macroblock to produce the decoded error signal. This error signal is added back to the prediction signal from the motion compensation module via summer 155 to produce a decoded reference picture (reconstructed image).
FIG. 2 depicts a block diagram of a flowchart of a method 200 for deriving and allocating bits for an image based on macroblocks within the image. More specifically, method 200 starts in step 205 and proceeds to step 210 where a target frame bit rate is determined for a current frame. In the preferred embodiment, the target frame bit rate is determined using a complexity measure that is recursively adjusted through the use of a polynomial regression process (as illustrated in FIG. 3).
FIG. 3 illustrates a flowchart of a method 300 for determining a target frame bit rate. Referring to FIG. 3, the method begins at step 305 and proceeds to step 310, where the method determines the target bits (target frame bit rate) for a frame, Tframe as:
Tfmme = — x (1 - past _ percent) + Tpreviousframe x past _ percent (1) where R is the remaining number of bits for a sequence of frame, Nf is the number of remaining frames in the sequence, Tpreϊiousfraιne is the number of bits used for encoding the previous frame, and the past_percent is a constant. In the preferred embodiment, the constant past_percent is selected to be 0.05. However, the present invention is not so limited. Other values can be employed, that depend on the specific applications or the context of the images. In fact, these values can be adjusted temporally.
In sum, equation (1) allows the target frame bit rate to be computed based on the bits available and the last encoded frame bits. If the last frame is complex and uses many bits, it leads to the premise that more bits should be assigned to the current frame. However, this increased allocation will diminish the available number of bits for encoding the remaining frames, thereby limiting the increased allocation to this frame. A weighted average reflects a compromise of these two factors, as illustrated in the second term in equation (1).
In step 320, method 300 then adjusts the calculated target frame bit rate, Tfr by the current buffer fullness as:
. (a + c x b) „,
T = - - X. T (2) fmme (c a + b) fmme
where Tframe is the adjusted target bit rate, "a" is the current buffer fullness (the portion of the buffer that contains bits to be sent to the decoder), and b is (the physical buffer size - buffer fullness (a)), and c is a constant selected to be a value of 2 (other values can be used). As such "b" represents the remaining space in the buffer. Equation (2) indicates that if the buffer is more than half full, the adjusted target bit rate Tframe is decreased. Conversely, if the buffer is less than half full, the adjusted target bit rate Tframe is increased. If the buffer is exactly at half, no adjustment is necessary, since equation (2) reduces to T' framβ = Tframe. In step 330, method 300 then optionally verifies that a lower bound of target frame bit rate (R/30) is maintained as: Tframe = Max(Rs / 30, Tfmme) (3)
It should be noted that equation (3) allows Tframβ to take the greater (max) of two possible values, where Rs is a bit rate for the sequence (or segment), e.g., 24000 bits/sec). Namely, a lower bound of target rate (R/30) is used to maintain or guarantee a minimal quality, e.g., 800 bits/frame can be set as a minimum. If the minimal quality cannot be maintained, the encoder has the option to skip the current frame altogether. Method 300 then ends in step 340.
It should be understood that other frame bit rate allocation methods can be used, e.g., the MPEG TM4 and TM5, with the present invention. However, since the target macroblock bit rates are derived using the target frame bit rate, the computational overhead and the accuracy of the target macroblock bit rates are affected by the frame bit rate allocation method that is employed. For example, if it is desirable to minimize computational complexity at the expense of performance, then it may be appropriate to employ the frame bit rate allocation methods of MPEG TM4 and TM5. In contrast, if it is desirable to maximize performance at the expense of increasing computational complexity, then it may be appropriate to employ the frame bit rate allocation method of the patent application (attorney docket SAR12459), or other more complex frame bit rate allocation methods.
Returning to FIG. 2, once Tfrmne is determined, method 200 then determines one or more target macroblock bit rates for the macroblocks within the current image as illustrated in FIG. 4 below. Method 200 then ends in step 230.
FIG. 4 illustrates a flowchart of a method 400 for determining one or more target macroblock bit rates for the macroblocks within the current image. The method starts in step 405 and proceeds to step 410, where a sum of absolute difference (SAD) Sj is performed for each macroblock i. Namely, the absolute difference between each pixel value (in the original image) and the corresponding pixel value (in the predicted image) is performed for pixels defined within the macroblock. Next, the sum of all the absolute differences of the pixels for the macroblock is performed to generate the SAD for the macroblock i. In step 420, method 400 queries whether Sα is greater than the threshold H . The threshold Hq is selected to remove various macroblocks from the present macroblock bit allocation method. Namely, it is anticipated that information in certain macroblocks will be eliminated or reduced to zero through either spatial filtering or quantization. Since these macroblocks will be deemed to carry no information, bits will not be allocated to these macroblocks. Thus, the threshold Hq serves to eliminate various macroblocks from consideration.
In operation, the threshold Hq is selected as the average of all the mean absolute difference (MADs) that have been skipped in the previous frame, where MADj is defined as the S, divided by the number of pixels in the macroblock i. However, if the current frame is the first frame in the image sequence, then Hq is set to be half of the average of all the existing MADs in this current frame.
Thus, if the query at step 420 is answered negatively, then method 400 proceeds to step 425, where the present macroblock i is removed from consideration for receiving bit allocation. If the query at step 420 is answered positively, then method 400 proceeds to step 423, where the present macroblock i will be considered for receiving bit allocation, as discussed below. In step 430, target macroblock bit rate, R,, is determined for each macroblock i (i= 1, 2, 3...) as follows:
R, = K, x Tframe (4)
Mad w K, = «-«- for Madk having SADk > H (5) Mad Λ, k=, where Mad; is the mean absolute difference (MAD) of an macroblock i, "n" is the number of macroblocks in a frame, "w" is a weighting factor and Rjis the estimated target macroblock bit rate for macroblock i. The weighting factor w allows R; for a given macroblock to be adjusted in accordance with other criteria. Namely, it has been observed that some macroblocks can be viewed as being more important than other macroblocks, where "importance" is not based solely on the MAD of an macroblock. Various applications as illustrated above, may place emphasis on certain macroblocks. By incorporating an optional weighting factor w, more or less bits can be allocated to a particular macroblock based upon application specific criteria. In the preferred embodiment, w is set to a value of "1", but other values can be used as necessary for a particular application.
Thus, the target frame bits Tframe are distributed proportional to the mean of the absolute differences (MAD) of a macroblock. For example, if an image has only four (4) macroblocks "a", "b", "c" and "d", having Mada=l , Madb =1, Madc=3 and Madd=4, respectively, Tframe = 100 and Hq is such that macroblocks a and b are removed from consideration for receiving bit allocation, then, Rc and Rd are given respectively as:
3w„ 4vv,
R = x 100 and R, =- x(100- Λ).
2>wc + 4wd 3w„ + Aw,
Thus, the MADs of macroblocks a and b are not used in the determination of Rc and Rd. Furthermore, the above example illustrates the need to update Tframe after each R; is allocated to a macroblock. Once R; for each macroblock in the image is determined, method 400, in step 440 calculates a quantization scale Qj for each macroblock i. In the preferred embodiment, the quantization scale Q; is calculated in accordance with a distortion measure as described in US patent application with serial number 09/022,349 (attorney docket SAR12459). In brief, the quantizer scale Q; is derived from a quadratic rate-distortion method as expressed below: R^ X^Q. '+ X.E.Q;2 (6) where X R, = r/rflme
where R, is the actual bits used for macroblock i, Qx represents a quantization level or scale selected for the macroblock i, E: represents a distortion measure. In the preferred embodiment, E] represents a sum of all the mean absolute differences (MADs) for those macroblocks having R, for the current frame. Namely, the measure Et provides a method of adjusting the macroblock bit budget to account for the differences in the macroblock between successive frames in a sequence. In other words, the greater the differences between an macroblock in the current frame and the same macroblock in a previous frame, the greater the number of bits that will be required to code the macroblock in the current frame. Furthermore, other distortion measures can be used, such that E, may represent mean square error or just-noticeable difference (jnd).
During initialization, R, in equation (6) is substituted with the calculated R, to generate Q,. The parameters X1 and X, are initialized as follows:
X, = (Rs*Ns)/2 , = 0 (7)
Ra represents the bit rate for the sequence (or segment), e.g., 24000 bits per second. Ns represents the distance between encoded frames. Namely, due to low bit rate applications, certain frames within a sequence may not be encoded (skipped), e.g., the encoder may only encode every fourth frame. It should be understood that the number of skipped frames can be tailored to the requirement of a particular application. Thus, in step 440, the quantizer scale Qt can be determined in accordance with equations (6) and (7). However, the calculated Q, is limited by the condition that it should not be varied too significantly from macroblock to macroblock. Namely Q, is limited by the conditions:
if Q, < (Qlast -e), then Q, = (Qlast -e), (8) f Qj > (Q,a t +e), then Q, = (Qlast +e),
where Qlast is the quantizer scale for an immediate previous macroblock having an R, and "e" is a constant selected to be a value "2" (other values can be used depending on the application). Namely, the calculated Qt is limited by a change to be no greater than a value of "2" from macroblock to macroblock. This limitation maintains uniform visual quality, i.e., minimizing significant changes in visual quality from macroblock to macroblock. It should be noted that if an immediate macroblock is skipped (a macroblock without an R,), then Qlast is based on the next immediate "non- skipped" macroblock.
In step 450, the Q, is used to encode the macroblock to generate an "actual Rt" (the actual number of bits used to encode the macroblock). After actual encoding, the actual Rt and the Qι are used to update the parameters X and Xj, using a polynomial regression model or a quadratic regression model to refine the parameters Xt and X. Namely, the constants Xx and X, are updated to account for the discrepancy between the bits allocated to an macroblock and the actual number of bits needed to the code the macroblock for a particular quantizer level or scale. Regression models are well known in the art. For a detailed discussion of various regression models, see e.g., Bowerman and O'Connell, Forecasting and Time Series. 3rd Edition, Duxbury Press, (1993 , chapter 4).
Furthermore, Tframe is then updated by subtracting the actual R,:
Tframe = Tframe - actual R, (9)
When the entire frame is encoded, the actual Tframe (the actual number of bits used to encode the frame) can then be used to update the method that is -lo- tasked with generating a target frame bit rate, e.g. as illustrated in FIG. 3 above.
Although the above embodiment employs R; to compute a quantizer scale for a macroblock, other coding parameters such as allocation of computing resources can be implemented. Namely, if it is determined that a particular macroblock or a series of macroblocks has a high Rt, then it is possible to allocate more processing power, e.g., dedicating more processors in a multiple processors coding system in coding a particular macroblock or a series of macroblocks. In step 470, method 400 queries whether there is a next macroblock in the frame. If the query is negatively answered, then method 400 ends in step 480. If the query is positively answered, then method 400 returns to step 430, where steps 430-460 are repeated until all the macroblocks are evaluated in the current frame. FIG. 5 depicts a wavelet-based encoder 500 that incorporates the present invention. The encoder contains a block motion compensator (BMC) and motion vector coder 504, subtractor 502, discrete wavelet transform (DWT) coder 506, bit rate controller 510, DWT decoder 512 and output buffer 514. In general, as discussed above the input signal is a video image (a two-dimensional array of pixels (pels) defining a frame in a video sequence). To accurately transmit the image through a low bit rate channel, the spatial and temporal redundancy in the video frame sequence must be substantially reduced. This is generally accomplished by coding and transmitting only the differences between successive frames. The encoder has three functions: first, it produces, using the BMC and its coder 504, a plurality of motion vectors that represent motion that occurs between frames; second, it predicts the present frame using a reconstructed version of the previous frame combined with the motion vectors; and third, the predicted frame is subtracted from the present frame to produce a frame of residuals that are coded and transmitted along with the motion vectors to a receiver. -ιy-
The discrete wavelet transform performs a wavelet hierarchical subband decomposition to produce a conventional wavelet tree representation of the input image. To accomplish such image decomposition, the image is decomposed using times two subsampling into high horizontal-high vertical (HH), high horizontal-low vertical (HL), low horizontal-high vertical (LH), and low horizontal-low vertical (LL), frequency subbands. The LL subband is then further subsampled times two to produce a set of HH, HL, LH and LL subbands. This subsampling is accomplished recursively to produce an array of subbands such as that illustrated in FIG. 6 where three subsamplings have been used. Preferably six subsamplings are used in practice. The parent-child dependencies between subbands are illustrated as arrows pointing from the subband of the parent nodes to the subbands of the child nodes. The lowest frequency subband is the top left LL1? and the highest frequency subband is at the bottom right HH3. In this example, all child nodes have one parent. A detailed discussion of subband decomposition is presented in J.M. Shapiro, "Embedded Image Coding Using Zerotrees of Wavelet Coefficients", IEEE Trans, on Signal Processing, Vol. 41, No. 12, pp. 3445-62, December 1993. The DWT coder of FIG. 5 codes the coefficients of the wavelet tree in either a "breadth first" or "depth first" pattern. A breadth first pattern traverse the wavelet tree in a bit-plane by bit-plane pattern, i.e., quantize all parent nodes, then all children, then all grandchildren and so on. In contrast, a depth first pattern traverses each tree from the root in the low-low subband (LLX) through the children (top down) or children through the low-low subband (bottom up). The selection of the proper quantization level by the rate controller 510 is as discussed above to control the bit rate for each macroblock within each frame of a sequence. As such, the present invention can be adapted to various types of encoders that use different transforms. FIG. 7 illustrates an encoding system 700 of the present invention.
The encoding system comprises a general purpose computer 710 and various input/output devices 720. The general purpose computer comprises a central processing unit (CPU) 712, a memory 714 and an encoder 716 for receiving and encoding a sequence of images.
In the preferred embodiment, the encoder 716 is simply the encoder 100 and/or encoder 500 as discussed above. The encoder 716 can be a physical device which is coupled to the CPU 712 through a communication channel. Alternatively, the encoder 716 can be represented by a software application (or a combination of software and hardware, e.g., application specific integrated circuits (ASIC)) which is loaded from a storage device and resides in the memory 712 of the computer. As such, the encoder 100 and 500 of the present invention can be stored on a computer readable medium, e.g., a memory or storage device.
The computer 710 can be coupled to a plurality of input and output devices 720, such as a keyboard, a mouse, a camera, a camcorder, a video monitor, any number of imaging devices or storage devices, including but not limited to, a tape drive, a floppy drive, a hard disk drive or a compact disk drive. The input devices serve to provide inputs to the computer for producing the encoded video bitstreams or to receive the sequence of video images from a storage device or an imaging device. Finally, a communication channel 730 is shown where the encoded signal from the encoding system is forwarded to a decoding system (not shown).
There has thus been shown and described a novel apparatus and method that selects a quantizer scale for each macroblock within each frame to maintain the overall quality of the video image while optimizing the coding rate. Many changes, modifications, variations and other uses and applications of the subject invention will, however, become apparent to those skilled in the art after considering this specification and the accompanying drawings which disclose the embodiments thereof. All such changes, modifications, variations and other uses and applications which do not depart from the spirit and scope of the invention are deemed to be covered by the invention.

Claims

What is claimed is:
1. A method for allocating bits to encode each frame of an image sequence, each of said frame having at least one block, said method comprising the steps of:
(a) determining a target frame bit rate for the frame; and
(b) allocating said target frame bit rate among the at least one block.
2. The method of claim 1, wherein said allocating step (b) comprises the step of allocating said target frame bit rate in accordance with a target block bit rate for the at least one block.
3. The method of claim 2, wherein said target block bit rate for the at least one block is selected in accordance with a mean absolute differences (Mad) of said block.
4) The method of claim 3, wherein said target block bit rate is determined in accordance with:
R, = K, x Tfmme
K, = for Madk having SADk > Hq
where Mad is the mean absolute difference (MAD) of a block, n is a number of blocks in the frame, w is a weighting factor and Rjis said target block bit rate, Sad is a sum of absolute difference (SAD) of a block, and H is a constant.
5. The method of claim 3, wherein said target block bit rate is adjusted in accordance with a threshold H .
6. The method of claim 5, wherein said target block bit rate is adjusted by removing each block within the frame having a sum of absolute difference (SAD) that is less than said threshold Hq from said step (b) of allocating said target frame bit rate among the at least one block.
7. The method of claim 1, wherein said target frame bit rate, Tftame, is derived in accordance with:
n
Tframe = f x (* ~ past _ percent) + Tpreviousframe x past _ percent
where R is a remaining number of bits for the image sequence, Nf is a number of remaining frames in the image sequence, Tprevi0U8frame is a number of bits used for encoding a previous frame, and past_percent is a constant.
8. The method of claim 2, further comprising the step of:
(c) generating a quantizer scale for said at least one block in accordance with said target block bit rate.
9. Apparatus for encoding each frame of an image sequence, said frame having at least one block, said apparatus comprising: a motion compensator for generating a predicted image of a current frame; a transform module for applying a transformation to a difference signal between the current frame and said predicted image, where said transformation produces a plurality of coefficients; a quantizer for quantizing said plurality of coefficients with at least one quantizer scale; and a controller for selectively adjusting said at least one quantizer scale for a current frame in response to a target block bit rate for the at least one block.
10. The apparatus of claim 9, wherein said target block bit rate for the at least one block is selected in accordance with a mean absolute differences (Mad) of said block.
EP98935565A 1997-07-14 1998-07-09 Apparatus and method for macroblock based rate control in a coding system Expired - Lifetime EP1012778B1 (en)

Applications Claiming Priority (5)

Application Number Priority Date Filing Date Title
US5243797P 1997-07-14 1997-07-14
US52437 1998-03-31
US09/099,617 US6690833B1 (en) 1997-07-14 1998-06-18 Apparatus and method for macroblock based rate control in a coding system
US99617 1998-06-18
PCT/US1998/014132 WO1999004359A1 (en) 1997-07-14 1998-07-09 Apparatus and method for macroblock based rate control in a coding system

Publications (3)

Publication Number Publication Date
EP1012778A1 true EP1012778A1 (en) 2000-06-28
EP1012778A4 EP1012778A4 (en) 2006-04-26
EP1012778B1 EP1012778B1 (en) 2011-06-01

Family

ID=26730598

Family Applications (1)

Application Number Title Priority Date Filing Date
EP98935565A Expired - Lifetime EP1012778B1 (en) 1997-07-14 1998-07-09 Apparatus and method for macroblock based rate control in a coding system

Country Status (8)

Country Link
US (1) US6690833B1 (en)
EP (1) EP1012778B1 (en)
KR (1) KR100559924B1 (en)
CN (1) CN1223959C (en)
AU (1) AU8478298A (en)
BR (1) BR9810588A (en)
CA (1) CA2294225A1 (en)
WO (1) WO1999004359A1 (en)

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN113282553A (en) * 2021-05-24 2021-08-20 北京航空航天大学 Rapid lossless compression storage method for far-field data in electromagnetic environment

Families Citing this family (42)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US6539124B2 (en) 1999-02-03 2003-03-25 Sarnoff Corporation Quantizer selection based on region complexities derived using a rate distortion model
US7277483B1 (en) * 2000-04-18 2007-10-02 Ati International Srl Method and apparatus for rate control for constant-bit-rate finite-buffer-size video encoder
US6904094B1 (en) * 2000-09-20 2005-06-07 General Instrument Corporation Processing mode selection for channels in a video multi-processor system
KR100355831B1 (en) * 2000-12-06 2002-10-19 엘지전자 주식회사 Motion vector coding method based on 2-demension least bits prediction
JP4534106B2 (en) * 2000-12-26 2010-09-01 日本電気株式会社 Video encoding system and method
FR2821998A1 (en) * 2001-03-06 2002-09-13 Koninkl Philips Electronics Nv Method for coding digital images in macroblocks with exclusion based on reconstruction capacity estimated on the basis of concealment of errors
GB0119708D0 (en) 2001-08-13 2001-10-03 Unilever Plc Process for the production of detergent granules
JP4769392B2 (en) * 2001-09-25 2011-09-07 キヤノン株式会社 Signal processing device
WO2003061294A2 (en) * 2001-12-28 2003-07-24 Koninklijke Philips Electronics N.V. Video encoding method
US20030123539A1 (en) * 2001-12-28 2003-07-03 Hyung-Suk Kim Method and apparatus for video bit-rate control
KR100466596B1 (en) * 2002-01-07 2005-01-24 주식회사 팬택앤큐리텔 Method for Improvement of Adaptive Rate Control in Based Block Video Coder using Intra Refresh
US7388912B1 (en) * 2002-05-30 2008-06-17 Intervideo, Inc. Systems and methods for adjusting targeted bit allocation based on an occupancy level of a VBV buffer model
US20040066849A1 (en) * 2002-10-04 2004-04-08 Koninklijke Philips Electronics N.V. Method and system for significance-based embedded motion-compensation wavelet video coding and transmission
ATE435567T1 (en) * 2003-08-29 2009-07-15 Koninkl Philips Electronics Nv SYSTEM AND METHOD FOR ENCODING AND DECODING IMPROVEMENT LEVEL DATA USING DESCRIPTIVE MODEL PARAMETERS
KR100987765B1 (en) * 2003-09-30 2010-10-13 삼성전자주식회사 Prediction method and apparatus in video encoder
FR2862168B1 (en) * 2003-11-06 2006-05-19 Canon Kk METHOD AND DEVICE FOR CONTROLLING THE FLOW OF AN IMAGE SEQUENCE
US7483583B2 (en) * 2003-12-16 2009-01-27 Infocus Corporation System and method for processing image data
CN100385453C (en) * 2003-12-16 2008-04-30 富可视公司 System and method for processing image data
EP1551185A1 (en) * 2004-01-05 2005-07-06 Thomson Licensing S.A. Encoding method, decoding method, and encoding apparatus for a digital picture sequence
US20050207501A1 (en) * 2004-03-18 2005-09-22 Sony Corporation Method of and system for video bit allocation for scene cuts and scene changes
CN101069432B (en) * 2004-12-02 2015-10-21 汤姆逊许可公司 For the determination method and apparatus of the quantization parameter that video encoder rate controls
US20060120612A1 (en) * 2004-12-08 2006-06-08 Sharath Manjunath Motion estimation techniques for video encoding
US20060146929A1 (en) * 2005-01-06 2006-07-06 Lg Electronics Inc. Method and system for acceleration of lossy video encoding owing to adaptive discarding poor-informative macroblocks
CA2595583A1 (en) * 2005-01-28 2006-08-10 Thomson Licensing Method and apparatus for real-time frame encoding
US7956905B2 (en) * 2005-02-28 2011-06-07 Fujifilm Corporation Titling apparatus, a titling method, and a machine readable medium storing thereon a computer program for titling
US7403562B2 (en) * 2005-03-09 2008-07-22 Eg Technology, Inc. Model based rate control for predictive video encoder
CN100442848C (en) * 2005-04-11 2008-12-10 华为技术有限公司 Method for controlling code rate in H263 coding
US7848409B2 (en) * 2005-05-02 2010-12-07 Qualcomm Incorporated Macroblock level bit allocation
CN101223790B (en) * 2005-07-14 2013-03-27 英特尔公司 Rate control method and apparatus
US8761259B2 (en) * 2005-09-22 2014-06-24 Qualcomm Incorporated Multi-dimensional neighboring block prediction for video encoding
JP4795223B2 (en) * 2006-01-31 2011-10-19 キヤノン株式会社 Image processing device
KR101172430B1 (en) * 2007-08-17 2012-08-08 삼성전자주식회사 Method and apparatus for bit rate control
US20090161766A1 (en) * 2007-12-21 2009-06-25 Novafora, Inc. System and Method for Processing Video Content Having Redundant Pixel Values
TWI428022B (en) * 2008-10-24 2014-02-21 Nat Univ Chung Cheng Video control method for instant video encoding chips
JP5950157B2 (en) * 2012-05-18 2016-07-13 ソニー株式会社 Image processing apparatus and method, and program
KR102249819B1 (en) * 2014-05-02 2021-05-10 삼성전자주식회사 System on chip and data processing system including the same
US10200716B2 (en) 2015-06-25 2019-02-05 Sony Interactive Entertainment Inc. Parallel intra-prediction encoding/decoding process utilizing PIPCM and/or PIDC for selected sections
US10142049B2 (en) * 2015-10-10 2018-11-27 Dolby Laboratories Licensing Corporation Near optimal forward error correction system and method
US11750811B2 (en) * 2018-04-06 2023-09-05 Comcast Cable Communications, Llc Systems, methods, and apparatuses for processing video
US12088821B1 (en) * 2020-06-30 2024-09-10 Amazon Technologies, Inc. Dynamic encoder-time scaling service for live and on-demand adaptive streaming
CN112738515B (en) * 2020-12-28 2023-03-24 北京百度网讯科技有限公司 Quantization parameter adjustment method and apparatus for adaptive quantization
CN113596451B (en) * 2021-06-28 2024-01-26 无锡唐古半导体有限公司 Video encoding method, video decoding method and related devices

Family Cites Families (13)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US5134476A (en) 1990-03-30 1992-07-28 At&T Bell Laboratories Video signal encoding with bit rate control
US5231484A (en) 1991-11-08 1993-07-27 International Business Machines Corporation Motion video compression system with adaptive bit allocation and quantization
US5253059A (en) 1992-05-15 1993-10-12 Bell Communications Research, Inc. Method and circuit for adjusting the size of a video frame
US5440345A (en) 1992-07-17 1995-08-08 Kabushiki Kaisha Toshiba High efficient encoding/decoding system
US5367629A (en) * 1992-12-18 1994-11-22 Sharevision Technology, Inc. Digital video compression system utilizing vector adaptive transform
AU6099594A (en) 1993-02-03 1994-08-29 Qualcomm Incorporated Interframe video encoding and decoding system
US5440346A (en) * 1993-06-16 1995-08-08 Intel Corporation Mode selection for method and system for encoding images
US5509089A (en) * 1993-09-09 1996-04-16 Intel Corporation Method and system for encoding images using temporal filtering
US5493513A (en) * 1993-11-24 1996-02-20 Intel Corporation Process, apparatus and system for encoding video signals using motion estimation
US5493514A (en) * 1993-11-24 1996-02-20 Intel Corporation Process, apparatus, and system for encoding and decoding video signals
US5528238A (en) * 1993-11-24 1996-06-18 Intel Corporation Process, apparatus and system for decoding variable-length encoded signals
US5532940A (en) * 1993-11-24 1996-07-02 Intel Corporation Process, apparatus and system for selecting quantization levels for encoding video signals
US5594504A (en) 1994-07-06 1997-01-14 Lucent Technologies Inc. Predictive video coding using a motion vector updating routine

Non-Patent Citations (6)

* Cited by examiner, † Cited by third party
Title
A. VETRO, H. SUN: "MPEG97/M1631: joint rate control for multiple video objects based on quadratic rate-distortion" ISO/IEC/JTC1/SC29/WG11, February 1997 (1997-02), XP002369677 *
CUBISS C. ET AL: 'VARIABLE BLOCK SIZE MOTION COMPENSATED CONDITIONAL REPLENISHMENT IMAGE CODER' PROCEEDINGS OF THE PICTURE CODING SYMPOSIUM (PCS). LAUSANNE, MAR. 17 - 19, 1993; [PROCEEDINGS OF THE PICTURE CODING SYMPOSIUM (PCS)], LAUSANNE, SFIT, CH vol. -, 17 March 1993, pages 13.6/A - 13.6/B, XP000346453 *
DAVID N. HEIN ET AL: 'Video Compression Using Conditional Replenishment and Motion Prediction' IEEE TRANSACTIONS ON ELECTROMAGNETIC COMPATIBILITY, IEEE SERVICE CENTER, NEW YORK, NY, US vol. EMC-10, no. 3, 01 August 1984, pages 134 - 142, XP011165174 ISSN: 0018-9375 *
See also references of WO9904359A1 *
T. CHIANG, H.J. LEE, Y.Q. ZHANG: "Q2 core experiment: macroblock rate control" ISO/IEC/JTC1/SC29/WG11, 31 March 1997 (1997-03-31), XP002369676 *
TIHAO CHIANG ET AL: "A new rate control scheme using quadratic rate distortion model" PROCEEDINGS OF THE INTERNATIONAL CONFERENCE ON IMAGE PROCESSING (ICIP) LAUSANNE, SEPT. 16 - 19, 1996, NEW YORK, IEEE, US, vol. VOL. 1, 16 September 1996 (1996-09-16), pages 73-76, XP010202597 ISBN: 0-7803-3259-8 *

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN113282553A (en) * 2021-05-24 2021-08-20 北京航空航天大学 Rapid lossless compression storage method for far-field data in electromagnetic environment
CN113282553B (en) * 2021-05-24 2021-11-02 北京航空航天大学 Rapid lossless compression storage method for far-field data in electromagnetic environment

Also Published As

Publication number Publication date
KR100559924B1 (en) 2006-03-13
AU8478298A (en) 1999-02-10
CN1223959C (en) 2005-10-19
US6690833B1 (en) 2004-02-10
CA2294225A1 (en) 1999-01-28
EP1012778B1 (en) 2011-06-01
WO1999004359A1 (en) 1999-01-28
EP1012778A4 (en) 2006-04-26
CN1274446A (en) 2000-11-22
KR20010021879A (en) 2001-03-15
BR9810588A (en) 2000-09-05

Similar Documents

Publication Publication Date Title
US6690833B1 (en) Apparatus and method for macroblock based rate control in a coding system
US6023296A (en) Apparatus and method for object based rate control in a coding system
US6243497B1 (en) Apparatus and method for optimizing the rate control in a coding system
US6160846A (en) Apparatus and method for optimizing the rate control in a coding system
US6084908A (en) Apparatus and method for quadtree based variable block size motion estimation
US6895050B2 (en) Apparatus and method for allocating bits temporaly between frames in a coding system
US6192081B1 (en) Apparatus and method for selecting a coding mode in a block-based coding system
US7653129B2 (en) Method and apparatus for providing intra coding frame bit budget
WO2000079802A1 (en) Apparatus and method for optimizing the rate control for multiscale entropy encoding
JP2000511366A6 (en) Apparatus and method for variable block size motion estimation based on quadrant tree
US20050036549A1 (en) Method and apparatus for selection of scanning mode in dual pass encoding
US6363113B1 (en) Methods and apparatus for context-based perceptual quantization
US20060146928A1 (en) Method and apparatus for providing a rate control
JP4532607B2 (en) Apparatus and method for selecting a coding mode in a block-based coding system

Legal Events

Date Code Title Description
PUAI Public reference made under article 153(3) epc to a published international application that has entered the european phase

Free format text: ORIGINAL CODE: 0009012

17P Request for examination filed

Effective date: 20000124

AK Designated contracting states

Kind code of ref document: A1

Designated state(s): DE ES FR GB IT NL

RAP1 Party data changed (applicant data changed or rights of an application transferred)

Owner name: MEDIATEK INC.

A4 Supplementary search report drawn up and despatched

Effective date: 20060313

17Q First examination report despatched

Effective date: 20071213

GRAP Despatch of communication of intention to grant a patent

Free format text: ORIGINAL CODE: EPIDOSNIGR1

GRAJ Information related to disapproval of communication of intention to grant by the applicant or resumption of examination proceedings by the epo deleted

Free format text: ORIGINAL CODE: EPIDOSDIGR1

GRAP Despatch of communication of intention to grant a patent

Free format text: ORIGINAL CODE: EPIDOSNIGR1

GRAS Grant fee paid

Free format text: ORIGINAL CODE: EPIDOSNIGR3

GRAA (expected) grant

Free format text: ORIGINAL CODE: 0009210

AK Designated contracting states

Kind code of ref document: B1

Designated state(s): DE ES FR GB IT NL

REG Reference to a national code

Ref country code: GB

Ref legal event code: FG4D

REG Reference to a national code

Ref country code: DE

Ref legal event code: R096

Ref document number: 69842296

Country of ref document: DE

Effective date: 20110714

REG Reference to a national code

Ref country code: NL

Ref legal event code: VDEP

Effective date: 20110601

PG25 Lapsed in a contracting state [announced via postgrant information from national office to epo]

Ref country code: ES

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20110912

PG25 Lapsed in a contracting state [announced via postgrant information from national office to epo]

Ref country code: NL

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20110601

PLBE No opposition filed within time limit

Free format text: ORIGINAL CODE: 0009261

STAA Information on the status of an ep patent application or granted ep patent

Free format text: STATUS: NO OPPOSITION FILED WITHIN TIME LIMIT

26N No opposition filed

Effective date: 20120302

PG25 Lapsed in a contracting state [announced via postgrant information from national office to epo]

Ref country code: IT

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20110601

REG Reference to a national code

Ref country code: DE

Ref legal event code: R097

Ref document number: 69842296

Country of ref document: DE

Effective date: 20120302

REG Reference to a national code

Ref country code: FR

Ref legal event code: PLFP

Year of fee payment: 18

REG Reference to a national code

Ref country code: FR

Ref legal event code: PLFP

Year of fee payment: 19

REG Reference to a national code

Ref country code: FR

Ref legal event code: PLFP

Year of fee payment: 20

PGFP Annual fee paid to national office [announced via postgrant information from national office to epo]

Ref country code: GB

Payment date: 20170727

Year of fee payment: 20

Ref country code: DE

Payment date: 20170727

Year of fee payment: 20

Ref country code: FR

Payment date: 20170726

Year of fee payment: 20

REG Reference to a national code

Ref country code: DE

Ref legal event code: R071

Ref document number: 69842296

Country of ref document: DE

REG Reference to a national code

Ref country code: GB

Ref legal event code: PE20

Expiry date: 20180708

PG25 Lapsed in a contracting state [announced via postgrant information from national office to epo]

Ref country code: GB

Free format text: LAPSE BECAUSE OF EXPIRATION OF PROTECTION

Effective date: 20180708