US8442114B2 - Moving picture encoding apparatus and distribution system - Google Patents

Moving picture encoding apparatus and distribution system Download PDF

Info

Publication number
US8442114B2
US8442114B2 US12/923,421 US92342110A US8442114B2 US 8442114 B2 US8442114 B2 US 8442114B2 US 92342110 A US92342110 A US 92342110A US 8442114 B2 US8442114 B2 US 8442114B2
Authority
US
United States
Prior art keywords
encoding rate
key frame
encoding
entropy
correction
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Expired - Fee Related, expires
Application number
US12/923,421
Other versions
US20110075733A1 (en
Inventor
Takahiro Yamasaki
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Oki Electric Industry Co Ltd
Original Assignee
Oki Electric Industry Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Oki Electric Industry Co Ltd filed Critical Oki Electric Industry Co Ltd
Assigned to OKI ELECTRIC INDUSTRY CO., LTD. reassignment OKI ELECTRIC INDUSTRY CO., LTD. ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: YAMASAKI, TAKAHIRO
Publication of US20110075733A1 publication Critical patent/US20110075733A1/en
Application granted granted Critical
Publication of US8442114B2 publication Critical patent/US8442114B2/en
Expired - Fee Related legal-status Critical Current
Adjusted expiration legal-status Critical

Links

Images

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/10Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
    • H04N19/102Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the element, parameter or selection affected or controlled by the adaptive coding
    • H04N19/115Selection of the code volume for a coding unit prior to coding
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/30Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using hierarchical techniques, e.g. scalability
    • H04N19/395Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using hierarchical techniques, e.g. scalability involving distributed video coding [DVC], e.g. Wyner-Ziv video coding or Slepian-Wolf video coding

Definitions

  • the present invention relates to the encoding and decoding of moving picture sequences and is applicable in, for example, a system that uses distributed video coding techniques to distribute moving picture data.
  • DVC Distributed video coding
  • Slepian-Wolf and Wyner-Ziv theorems A basic DVC coding method is described by Aaron et al. in ‘Transform-Domain Wyner-Ziv Codec for Video’, Proc. SPIE Visual Communications and Image Processing, 2004.
  • the encoder treats some frames in a received video sequence as key frames and the rest as Wyner-Ziv frames.
  • the key frames are coded as intraframes.
  • a discrete cosine transform is used to transform each Wyner-Ziv frame to the coefficient domain, the coefficients are grouped into bands, the coefficients in the k-th band are quantized by a 2 Mk -level quantizer, the quantized coefficients (q k ) are expressed in fixed numbers of bits, and the bit planes are extracted and supplied to a Slepian-Wolf encoder that uses a turbo code to produce data bits and error-correcting code bits, generally referred to as parity bits. The data bits are discarded.
  • DCT discrete cosine transform
  • the decoder decodes the key frames, uses the decoded key frames to generate a predicted image for each Wyner-Ziv frame, applies a DCT to convert the predicted image to the coefficient domain, groups the coefficients into bands, and inputs the coefficients in each band as side information to a Slepian-Wolf decoder.
  • the Slepian-Wolf decoder uses parity bits received from the encoder to correct prediction errors in the side information by an iterative process, in which the decoder originally receives a subset of the parity bits and may request further parity bits as required.
  • an inverse discrete cosine transform IDCT
  • a problem with this method is that feedback from the decoder to the encoder is necessary in order to request additional parity bits. As a result, the encoder and decoder cannot operate independently, and there are inevitable delays involved with requesting and obtaining additional parity bits.
  • Brites et al. use a combination of two methods to estimate the necessary encoding rate.
  • One method assumes that the distribution of the differences between the DCT coefficients of the original image and the predicted image can be approximately modeled by a Laplacian distribution. This model is used to estimate the decoder's prediction error probability. A conditional entropy is then calculated from the estimated error probability, and the necessary encoding rate is estimated from the conditional entropy.
  • a problem is that the supplementary rate is always added, even when it is not needed, so the encoder sometimes outputs more than the necessary number of parity bits. In that case, the supplementary parity bits do not improve the quality of the decoded image; the decoding result is the same as if the supplementary parity bits had not been sent.
  • the only effect of the supplementary parity bits is to waste computational resources in the encoder and decoder and communication resources on the link between them.
  • a general object of the present invention is to encode moving pictures more efficiently.
  • a more specific object is to determine when an encoder needs to supplement an estimated necessary encoding rate.
  • the invention provides a moving picture encoding apparatus for encoding a moving picture signal including a sequence of frames.
  • a frame classifier designates each frame as a key frame or a non-key frame.
  • a key frame encoder encodes the key frames.
  • a side information generator uses the encoded key frame data to generate a predicted image for each non-key frame.
  • An encoding rate estimator uses the predicted image to determine a final encoding rate for each non-key frame.
  • a non-key frame encoder then encodes each non-key frame to obtain an error-correcting code at the final encoding rate determined by the encoding rate estimator, and outputs the error-correcting code as encoded data of the non-key frame.
  • the encoding rate estimator determines a first encoding rate by using a first method to estimate the number of error-correcting bits needed to correct prediction errors in the predicted image generated for the non-key frame, and estimates at least one second encoding rate by using a second method or methods different from the first method to re-estimate the necessary number of error-correcting bits. From the second encoding rate(s), the encoding rate estimator decides whether to supplement the first encoding rate. If the decision is not to supplement the first encoding rate, the first encoding rate is used as the final encoding rate. If the decision is to supplement the first encoding rate, the encoding rate estimator adds a calculated encoding rate correction to the first encoding rate.
  • the first encoding rate may be determined by a distribution model entropy calculation or a bit error entropy calculation.
  • the second encoding rate may be determined by a prediction entropy calculation based on a model of the relation of the predicted image to the non-key frame, or on a direct comparison of the predicted image with the non-key frame.
  • Calculating the second encoding rate enables the encoding efficiency to be improved by supplementing the first encoding rate only when necessary.
  • FIG. 1 is a block diagram illustrating the basic structure of a moving picture distribution system embodying the invention
  • FIG. 2 is a functional block diagram illustrating the structure of a moving picture encoding apparatus in a first embodiment
  • FIGS. 3 to 5 constitute a flowchart illustrating the operation of the moving image encoding apparatus in FIG. 2 ;
  • FIG. 6 is a functional block diagram illustrating the structure of a moving picture encoding apparatus in a second embodiment.
  • FIGS. 7 to 10 constitute a flowchart illustrating the operation of the moving picture encoding apparatus in FIG. 6 .
  • Both embodiments concern a moving picture distribution system 1 ( 1 A) with the basic structure shown in FIG. 1 , including a moving picture distribution apparatus 10 ( 10 A) and a moving picture receiving apparatus 20 .
  • the moving picture distribution apparatus 10 includes a moving picture encoding apparatus 100 ( 100 A) that encodes a video signal, and a transmitter (not shown) that transmits the encoded video signal to the moving picture receiving apparatus 20 .
  • the moving picture receiving apparatus 20 includes a moving picture decoding apparatus 200 that decodes the video signal, and other apparatus (not shown) for output of the decoded signal.
  • the type of output is not limited. Exemplary forms of output include display, writing onto a storage medium, and transmission over a communication link.
  • Reference characters 1 A, 10 A, and 100 A are used in the second embodiment.
  • the video signal input to the moving picture encoding apparatus 100 may be any type of video signal, digital or analog, but it will be assumed in the following description that the signal is a digital signal organized as a sequence of frames. This is the type of signal that would be received from a video camera equipped with a charge coupled device (CCD) imager. If an analog video signal is input, the moving picture encoding apparatus 100 may include a frame grabber that captures the signal by converting it to a digitized frame sequence.
  • CCD charge coupled device
  • the moving picture encoding apparatus 100 may be a special-purpose device or it may be a general-purpose computer having, for example, a central processing unit (CPU), read only memory (ROM), random access memory (RAM), electrically programmable and erasable read only memory (EEPROM), a hard disk, and other well known facilities for storing data and storing and executing programs, including a program that implements the functions of the moving picture distribution apparatus 10 (or 10 A) in the embodiments described below.
  • the invention can be practiced by installing such a program in a computer. Other programs for communication processing, data processing, and so on may also be installed as necessary.
  • the moving picture decoding apparatus 200 may also be either a specialized device or a general-purpose computing device equipped with specialized software, and may include a display unit for displaying the decoded video signal output by the moving picture receiving apparatus 20 , or a data storage unit for storing the decoded video signal output by the moving picture receiving apparatus 20 .
  • the moving picture decoding apparatus 200 may have, for example, the structure shown by Brites et al.
  • the moving picture encoding apparatus 100 in the first embodiment comprises a frame classifier 105 , a Wyner-Ziv frame encoder 110 and a key frame encoder 160 .
  • the frame classifier 105 shown schematically as a switch, receives the incoming frame sequence, designates each frame as a key frame or a non-key frame, sends the key frames to the key frame encoder 160 , and sends the non-key frames to the Wyner-Ziv frame encoder 110 .
  • the Wyner-Ziv frame encoder 110 and key frame encoder 160 both output encoded data to the moving picture receiving apparatus 20 in FIG. 1 .
  • Non-key frames will be referred to below as Wyner-Ziv frames.
  • the frame classifier 105 may designate frames as key frames or Wyner-Ziv frames according to their position in the frame sequence. For example, odd-numbered frames may be key frames and even-numbered frames may be Wyner-Ziv frames. Alternatively, a scheme in which Wyner-Ziv frames occur consecutively, or key frames occur consecutively, may be used. The scheme may be selected for compatibility with existing DVC equipment.
  • the key frame encoder 160 receives and encodes the key frames and sends the encoded data to the moving picture receiving apparatus 20 .
  • Known intraframe coding methods such as the methods standardized by the Joint Photographic Experts Group (JPEG) or the Moving Picture Experts Group (MPEG) may be used, as taught by Aaron et al.
  • Distributed video coding of Wyner-Ziv frames may be carried out either in the pixel domain, by processing the pixel data directly, or in the transform domain, by first performing a transform such as a DCT, as also taught by Aaron et al.
  • the moving picture encoding apparatus 100 will be described as operating in the transform domain, but the invention is not limited to either domain.
  • the Wyner-Ziv frame encoder 110 includes a quantizer 120 , a Slepian-Wolf encoder 130 , an encoding rate estimator 140 , and a side information generator 150 .
  • the quantizer 120 receives, transforms, and quantizes Wyner-Ziv frames, divides the quantized data into bit planes, and gives the bit planes to the Slepian-Wolf encoder 130 and encoding rate estimator 140 .
  • the side information generator 150 generates a predicted image of a Wyner-Ziv frame from either or both of the key-frames preceding and following the Wyner-Ziv frame, transforms and quantizes the predicted image, and divides the quantized data into bit planes.
  • the encoding rate estimator 140 estimates an encoding rate that enables prediction errors in the input bit planes to be corrected, by using the original image of the input frame and the predicted image generated by the side information generator 150 . Details will be given later.
  • the Slepian-Wolf encoder 130 receives quantized bit planes from the quantizer 120 , performs Slepian-Wolf encoding of each bit plane by using a turbo code or a low-density parity-check (LDPC) code, for example, and sends resulting error-correcting code bits or parity bits to the moving picture receiving apparatus 20 at the encoding rate calculated by the encoding rate estimator 140 .
  • LDPC low-density parity-check
  • the encoding rate estimator 140 includes a distribution model entropy calculator 141 , an encoding rate correction calculator 142 , a correction decision unit 143 , and a correction unit 144 .
  • the basic operation of each component of the encoding rate estimator 140 is described below. Further details will be given later.
  • the distribution model entropy calculator 141 obtains the bit planes of the transformed original image in the Wyner-Ziv frame from the quantizer 120 and the coefficients of the corresponding predicted image from the side information generator 150 , estimates the error probability of each bit in the bit planes by using a distribution model, and then calculates an entropy.
  • the encoding rate correction calculator 142 obtains the bit planes of the transformed original image in the Wyner-Ziv frame from the quantizer 120 and the bit planes of the corresponding predicted image from the side information generator 150 and calculates an encoding rate correction.
  • the correction decision unit 143 compares the entropy obtained from the distribution model entropy calculator 141 and an entropy obtained by measuring bit differences (prediction errors) between the bit planes of the original image and the corresponding bit planes of the predicted image, and decides whether the encoding rate correction is needed.
  • the correction unit 144 corrects the calculation made by the distribution model entropy calculator 141 in accordance with the decision made by the correction decision unit 143 and supplies the corrected information to the Slepian-Wolf encoder 130 as the final encoding rate.
  • the frame is input to the quantizer 120 (step S 301 ) and transformed to the coefficient domain by a DCT, and the coefficients are grouped into spatial frequency bands, quantized, and divided into bit planes (step S 302 ).
  • the bit planes may cover the entire area of the frame, or the frame area may be divided into sub-areas and the coefficient data corresponding to each sub-area may be divided into separate bit planes.
  • the image of either the preceding or following key frame or both is input from the key frame encoder 160 to the side information generator 150 (step S 303 ) and a predicted image is generated for the Wyner-Ziv frame (step S 304 ).
  • the predicted image may be generated by using the preceding key frame image as the predicted image, by averaging the preceding and following key frame images, by carrying out a limited motion search, or by various other methods.
  • the predicted image generated by the side information generator 150 is transformed and quantized, and the quantized data are divided into bit planes (step S 305 ).
  • the length of the bit planes in step S 305 is the same as the length of the bit planes obtained by the quantizer 120 .
  • the coefficients of the original image of the frame are input from the quantizer 120 to the distribution model entropy calculator 141 (step S 306 ).
  • the distribution model entropy calculator 141 estimates the number of prediction errors on the assumption that the distribution of differences in coefficients between the original image and the predicted image can be approximately modeled by a Laplacian distribution, and calculates an entropy value.
  • the coefficients of the predicted image of the current frame are input from the side information generator 150 to the distribution model entropy calculator 141 (step S 307 ).
  • the distribution model entropy calculator 141 calculates the parameter ⁇ of the Laplacian distribution for each band (step S 308 ).
  • the parameter ⁇ of the Laplacian distribution is related to the variance ⁇ 2 of the distribution as in equation (1) below. If the mean value of the distribution is 0, the variance ⁇ 2 can be calculated by equation (2) below, where X and Y represent coefficients derived from the original image and the predicted image, respectively, and N is the length of the bit plane.
  • a bit plane of the original image is now input from the quantizer 120 to the distribution model entropy calculator 141 (step S 309 ).
  • the corresponding bit plane of the predicted image is input from the side information generator 150 to the distribution model entropy calculator 141 (step S 310 ).
  • the distribution model entropy calculator 141 calculates an error probability by using the distribution model as follows (step S 311 ).
  • Equation (3) the conditional probability distribution of X for a given value of Y can be represented by equation (3).
  • the probability P n that the nth bit x n j in the jth bit plane in the original image has the value ‘1’ can be expressed in terms of conditional probabilities as in equation (4), where x n j-1 is the corresponding bit in the preceding (j ⁇ 1)th bit plane of the original image and y n j is the corresponding bit in the jth bit plane of the predicted image.
  • the (j ⁇ 1)th bit plane is more significant than the jth bit plane, and will normally have been processed before the jth bit plane is processed.
  • the distribution model entropy calculator 141 calculates the entropy H(P n ) given by the equation (5) below (step S 312 in FIG. 4 ).
  • H ( P n ) ⁇ P n log( P n ) ⁇ (1 ⁇ P n )log(1 ⁇ P n ) (5)
  • the entropy expressed by equation (5) represents the amount of information needed to correct an error in the nth bit in the jth bit plane. If the same calculation is carried out for each bit in the jth bit plane and the results are averaged, the conditional entropy H X
  • This entropy value will be referred to as the distribution model entropy.
  • This value also represents the first encoding rate calculated by the encoding rate estimator 140 .
  • the distribution model entropy calculated by the distribution model entropy calculator 141 is input to the correction unit 144 (step S 313 ).
  • an encoding rate correction is prepared for use if the first encoding rate is not adequate for decoding.
  • the relative error entropy described by Brites et al. will be used as an exemplary encoding rate correction.
  • the bit plane of the original image is input from the quantizer 120 to the encoding rate correction calculator 142 (step S 314 ); the corresponding bit plane of the predicted image is input from the side information generator 150 to the encoding rate correction calculator 142 (step S 315 ); the encoding rate correction calculator 142 compares the two input bit planes and calculates a relative error probability (step S 316 ).
  • the relative error probability can be obtained by dividing the number of bits that differ from the corresponding predicted bits in the current bit plane but not in higher bit planes by the length of the bit plane.
  • the relative error entropy is then calculated from the relative error probability as shown by Brites et al. to obtain the encoding rate correction.
  • the encoding rate correction obtained by the encoding rate correction calculator 142 is input to the correction unit 144 (step S 317 ), where it is treated as a supplementary encoding rate.
  • the correction decision unit 143 compares the distribution model entropy with, for example, a bit error entropy calculated as follows.
  • Y is input from the distribution model entropy calculator 141 to the correction decision unit 143 (step S 318 ).
  • the corresponding bit plane of the original image is input from the quantizer 120 to the correction decision unit 143 (step S 319 ).
  • the corresponding bit plane of the predicted image is input from the side information generator 150 to the correction decision unit 143 (step S 320 ).
  • the bit error entropy H(err) is calculated from the bit error rate as in equation (7) below.
  • the encoding rate represented by the bit error entropy will be referred to as the second encoding rate.
  • H (err) ⁇ err log(err) ⁇ (1 ⁇ err)log(1 ⁇ err) (7)
  • the correction decision unit 143 now compares the first and second encoding rates, that is, the bit error entropy H(err) and the distribution model entropy H X
  • the first encoding rate or distribution model entropy is based on an assumed general model, while the second encoding rate or bit error entropy is based on an actual comparison of the original and predicted images. If the first encoding rate is equal to or greater than the second encoding rate, it can be inferred that the first encoding rate is sufficient to enable the decoding apparatus to correct the errors that will actually occur in the predicted image generated at the decoding apparatus, and the correction decision unit 143 decides to use the first encoding rate as the final coding rate. If the second encoding rate exceeds the first encoding rate, it can be inferred that the first encoding rate may be inadequate, so the correction decision unit 143 decides to supplement the first coding rate.
  • the decision made by the correction decision unit 143 is input to the correction unit 144 (step S 323 in FIG. 5 ) and tested (step S 324 ). If the correction decision unit 143 has decided not to supplement the first coding rate, the first encoding rate becomes the final coding rate output by the encoding rate estimator 140 (step S 325 ). Otherwise, the correction unit 144 supplements the first coding rate by adding the encoding rate correction, that is, by adding the relative error entropy to the distribution model entropy, to obtain the final encoding rate (step S 326 ).
  • the correction unit 144 supplies the final encoding rate to the Wyner-Ziv frame encoder 110 (step S 327 ).
  • the bit plane to be encoded is now input from the quantizer 120 to the Slepian-Wolf encoder 130 (step S 328 ) and encoded (step S 329 ), and the Slepian-Wolf encoder 130 outputs an error-correcting code at the final encoding rate determined by the encoding rate estimator 140 (step S 330 ).
  • the Wyner-Ziv frame encoder 110 now decides whether all bit planes of the current frame have been processed (step S 331 ). If so, the processing of the frame ends. Otherwise, the Wyner-Ziv frame encoder 110 returns to step S 309 to process the next bit plane.
  • the Wyner-Ziv frame encoder 110 in the first embodiment does not always add a correction to the first encoding rate obtained from the distribution model.
  • a second coding rate is calculated by a different method (e.g., bit error entropy), the first and second encoding rates are compared, and the first encoding rate is corrected, by adding a supplementary encoding rate, only if the comparison indicates a correction to be necessary.
  • the Wyner-Ziv frame encoder is thereby able to avoid sending the moving picture decoding apparatus 200 an unnecessarily large amount of encoded data.
  • the first embodiment accordingly saves time, and conserves computational resources and communication resources.
  • the moving picture distribution system 1 A in the second embodiment has the same moving picture receiving apparatus 20 as in the first embodiment, but the moving picture distribution apparatus 10 A includes a different moving picture encoding apparatus 100 A.
  • the moving picture encoding apparatus 100 A includes the same frame classifier 105 and key frame encoder 160 as in the first embodiment, but has a Wyner-Ziv frame encoder 110 A including an encoding rate estimator 140 A with an encoding rate correction calculator 142 A that operates only when the correction decision unit 143 decides to supplement the encoding rate.
  • the other components of the encoding rate estimator 140 A operate as in the first embodiment.
  • the frame sequence consists of odd-numbered key frames, which are encoded by conventional methods, and even-numbered Wyner-Ziv frames, which are encoded as follows.
  • the encoding rate correction calculated by the encoding rate correction calculator 142 A in the second embodiment is the relative error entropy indicated by Brites et al., as in the first embodiment.
  • the quantizer 120 transforms and quantizes each input Wyner-Ziv frame and divides the quantized data into bit planes; the side information generator 150 generates a predicted image; and the distribution model entropy calculator 141 calculates a distribution model entropy (steps S 501 to S 513 ).
  • the same operations are performed in steps S 501 to S 513 as in steps S 301 to S 313 in the first embodiment, so descriptions will be omitted.
  • the correction decision unit 143 decides whether the first encoding rate obtained by the distribution model entropy calculator 141 needs to be supplemented. This decision is made by, for example, comparing the distribution model entropy with a bit error entropy, as in the first embodiment: the distribution model entropy is input from the distribution model entropy calculator 141 to the correction decision unit 143 (step S 514 in FIG.
  • one bit plane of the original image is input from the quantizer 120 to the correction decision unit 143 (step S 515 ); the corresponding bit plane of the predicted image is input from the side information generator 150 to the correction decision unit 143 (step S 516 ); the bit planes of the original image and the predicted image are compared, the number of bit errors is counted, and a bit error entropy is calculated (step S 517 ); the correction decision unit 143 compares the bit error entropy and the distribution model entropy and decides whether to supplement the encoding rate obtained from the distribution model entropy (step S 518 ); the decision made by the correction decision unit 143 is input to the correction unit 144 (step S 519 ) and tested (step S 520 ).
  • the correction decision unit 143 decides to supplement the first encoding rate, it also passes this decision to the encoding rate correction calculator 142 A (step S 521 in FIG. 9 ).
  • the encoding rate correction calculator 142 A then obtains the bit plane of the original image from the quantizer 120 (step S 522 ) and the corresponding bit plane of the predicted image from the side information generator 150 (step S 523 ), compares these two bit planes, calculates a relative error probability, and derives an encoding rate correction (step S 524 ), which is provided to the correction unit 144 (step S 525 ).
  • the correction unit 144 adds the encoding rate correction to the encoding rate calculated from the distribution model entropy (step S 526 ). The resulting sum is the final encoding rate output by the encoding rate estimator 140 A.
  • the correction unit 144 finds in step S 520 in FIG. 8 that the correction decision unit 143 has decided not to supplement the first encoding rate, the correction unit 144 uses the first encoding rate, i.e., the distribution model entropy calculated by the distribution model entropy calculator 141 , as the final encoding rate (step S 527 in FIG. 9 ).
  • the final encoding rate is input to the Slepian-Wolf encoder 130 (step S 528 in FIG. 10 ), the bit plane to be encoded is input from the quantizer 120 to the Slepian-Wolf encoder 130 (step S 529 ) and encoded (step S 530 ), and the resulting error-correcting code is output at the final encoding rate obtained by the encoding rate estimator 140 A and transmitted to the moving picture decoding apparatus 200 (step S 531 ).
  • the Wyner-Ziv frame encoder 110 A determines whether all bit planes of the current frame have been processed (step S 532 ). If so, the processing of the frame ends. Otherwise, the processing returns to step S 509 .
  • the effect of the second embodiment is to reduce the amount of encoding processing, because the encoding rate correction calculator 142 A calculates the encoding rate correction only when the correction decision unit 143 decides that the encoding rate correction is necessary.
  • the moving picture encoding apparatus is shown as part of a moving picture distribution apparatus.
  • the moving picture encoding apparatus may, however, be a stand-alone apparatus for the output of encoded data generated from an input sequence of frames.
  • the output may take the form of storing the data in a disk drive or other storage medium, or outputting the data to another apparatus.
  • the second encoding rate estimate obtained by the encoding rate estimator need not be calculated by the bit error entropy method; it may be calculated by any method differing from the distribution model entropy method used to calculate the first encoding rate estimate.
  • the encoding rate estimator may calculate a plurality of second encoding rates by different methods, compare them with the first encoding rate calculated by the distribution model entropy method, and then decide whether to supplement the first encoding rate.
  • the encoding rate estimator 140 may be decide to supplement the first encoding rate if at least one of the second encoding rates exceeds the first encoding rate.
  • bit error entropy method is used to calculate the first encoding rate
  • distribution model entropy method is used to calculate the second encoding rate

Abstract

A moving picture encoding apparatus divides a moving picture signal into key frames and non-key frames, encodes the key frames, predicts the non-key frames from the encoded key frames, determines a final encoding rate for each non-key frame, encodes each non-key frame, and outputs an error-correcting code at the final encoding rate as encoded data of the non-key frame. To determine the final encoding rate, the number of error-correcting bits needed to correct prediction errors is estimated by a first method and one or more second methods. The second methods are used to decide whether to use the number of bits estimated by the first method as the final encoding rate, or whether to supplement the estimated number of bits by adding an encoding rate correction. As a result, the encoding rate is not supplemented unnecessarily and the encoding efficiency is improved.

Description

BACKGROUND OF THE INVENTION
1. Field of the Invention
The present invention relates to the encoding and decoding of moving picture sequences and is applicable in, for example, a system that uses distributed video coding techniques to distribute moving picture data.
2. Description of the Related Art
Distributed video coding (DVC) is a new coding method, based on the Slepian-Wolf and Wyner-Ziv theorems, that has attracted much recent attention. A basic DVC coding method is described by Aaron et al. in ‘Transform-Domain Wyner-Ziv Codec for Video’, Proc. SPIE Visual Communications and Image Processing, 2004. The encoder treats some frames in a received video sequence as key frames and the rest as Wyner-Ziv frames. The key frames are coded as intraframes. A discrete cosine transform (DCT) is used to transform each Wyner-Ziv frame to the coefficient domain, the coefficients are grouped into bands, the coefficients in the k-th band are quantized by a 2Mk-level quantizer, the quantized coefficients (qk) are expressed in fixed numbers of bits, and the bit planes are extracted and supplied to a Slepian-Wolf encoder that uses a turbo code to produce data bits and error-correcting code bits, generally referred to as parity bits. The data bits are discarded.
The decoder decodes the key frames, uses the decoded key frames to generate a predicted image for each Wyner-Ziv frame, applies a DCT to convert the predicted image to the coefficient domain, groups the coefficients into bands, and inputs the coefficients in each band as side information to a Slepian-Wolf decoder. The Slepian-Wolf decoder uses parity bits received from the encoder to correct prediction errors in the side information by an iterative process, in which the decoder originally receives a subset of the parity bits and may request further parity bits as required. When a satisfactory decoded result is obtained, an inverse discrete cosine transform (IDCT) is applied to reconstruct the image of the Wyner-Ziv frame.
A problem with this method is that feedback from the decoder to the encoder is necessary in order to request additional parity bits. As a result, the encoder and decoder cannot operate independently, and there are inevitable delays involved with requesting and obtaining additional parity bits.
In ‘Encoder Rate Control for Transform Domain Wyner-Ziv Video Coding’, ICIP 2007, Brites et al. describe a DVC system that eliminates feedback. Instead, the encoder estimates the number of parity bits that the decoder will need for adequate decoding of each Wyner-Ziv frame by generating its own predicted image and comparing this predicted image with the original image. The encoder then sends the decoder the estimated number of parity bits without having to be asked for them. That is, the encoder controls its own encoding rate.
Brites et al. use a combination of two methods to estimate the necessary encoding rate. One method assumes that the distribution of the differences between the DCT coefficients of the original image and the predicted image can be approximately modeled by a Laplacian distribution. This model is used to estimate the decoder's prediction error probability. A conditional entropy is then calculated from the estimated error probability, and the necessary encoding rate is estimated from the conditional entropy.
Since the Laplacian model is only approximate, and since the encoder and decoder may generate somewhat different predicted images, the estimated necessary encoding rate will occasionally provide fewer parity bits than the decoder actually needs, causing the decoded image to be visibly distorted. Brites et al. therefore add a supplementary rate calculated from a relative error probability that indicates the probability of errors in new locations in bit planes. By adding the supplementary rate to the encoding rate estimated by the distribution model entropy method, they avoid sending the decoder data that it cannot decode.
A problem is that the supplementary rate is always added, even when it is not needed, so the encoder sometimes outputs more than the necessary number of parity bits. In that case, the supplementary parity bits do not improve the quality of the decoded image; the decoding result is the same as if the supplementary parity bits had not been sent. The only effect of the supplementary parity bits is to waste computational resources in the encoder and decoder and communication resources on the link between them.
This is a serious shortcoming in a video distribution system. It would be preferable if the encoder could tell when the supplementary parity bits were needed and send them only when necessary.
SUMMARY OF THE INVENTION
A general object of the present invention is to encode moving pictures more efficiently.
A more specific object is to determine when an encoder needs to supplement an estimated necessary encoding rate.
The invention provides a moving picture encoding apparatus for encoding a moving picture signal including a sequence of frames. In the moving picture encoding apparatus, a frame classifier designates each frame as a key frame or a non-key frame. A key frame encoder encodes the key frames. A side information generator uses the encoded key frame data to generate a predicted image for each non-key frame. An encoding rate estimator uses the predicted image to determine a final encoding rate for each non-key frame. A non-key frame encoder then encodes each non-key frame to obtain an error-correcting code at the final encoding rate determined by the encoding rate estimator, and outputs the error-correcting code as encoded data of the non-key frame.
The encoding rate estimator determines a first encoding rate by using a first method to estimate the number of error-correcting bits needed to correct prediction errors in the predicted image generated for the non-key frame, and estimates at least one second encoding rate by using a second method or methods different from the first method to re-estimate the necessary number of error-correcting bits. From the second encoding rate(s), the encoding rate estimator decides whether to supplement the first encoding rate. If the decision is not to supplement the first encoding rate, the first encoding rate is used as the final encoding rate. If the decision is to supplement the first encoding rate, the encoding rate estimator adds a calculated encoding rate correction to the first encoding rate.
The first encoding rate may be determined by a distribution model entropy calculation or a bit error entropy calculation.
The second encoding rate may be determined by a prediction entropy calculation based on a model of the relation of the predicted image to the non-key frame, or on a direct comparison of the predicted image with the non-key frame.
Calculating the second encoding rate enables the encoding efficiency to be improved by supplementing the first encoding rate only when necessary.
BRIEF DESCRIPTION OF THE DRAWINGS
In the attached drawings:
FIG. 1 is a block diagram illustrating the basic structure of a moving picture distribution system embodying the invention;
FIG. 2 is a functional block diagram illustrating the structure of a moving picture encoding apparatus in a first embodiment;
FIGS. 3 to 5 constitute a flowchart illustrating the operation of the moving image encoding apparatus in FIG. 2;
FIG. 6 is a functional block diagram illustrating the structure of a moving picture encoding apparatus in a second embodiment; and
FIGS. 7 to 10 constitute a flowchart illustrating the operation of the moving picture encoding apparatus in FIG. 6.
DETAILED DESCRIPTION OF THE INVENTION
Two embodiments of the invention will now be described with reference to the attached drawings, in which like elements are indicated by like reference characters.
Both embodiments concern a moving picture distribution system 1 (1A) with the basic structure shown in FIG. 1, including a moving picture distribution apparatus 10 (10A) and a moving picture receiving apparatus 20. The moving picture distribution apparatus 10 includes a moving picture encoding apparatus 100 (100A) that encodes a video signal, and a transmitter (not shown) that transmits the encoded video signal to the moving picture receiving apparatus 20. The moving picture receiving apparatus 20 includes a moving picture decoding apparatus 200 that decodes the video signal, and other apparatus (not shown) for output of the decoded signal. The type of output is not limited. Exemplary forms of output include display, writing onto a storage medium, and transmission over a communication link. Reference characters 1A, 10A, and 100A are used in the second embodiment.
The video signal input to the moving picture encoding apparatus 100 may be any type of video signal, digital or analog, but it will be assumed in the following description that the signal is a digital signal organized as a sequence of frames. This is the type of signal that would be received from a video camera equipped with a charge coupled device (CCD) imager. If an analog video signal is input, the moving picture encoding apparatus 100 may include a frame grabber that captures the signal by converting it to a digitized frame sequence.
The moving picture encoding apparatus 100 may be a special-purpose device or it may be a general-purpose computer having, for example, a central processing unit (CPU), read only memory (ROM), random access memory (RAM), electrically programmable and erasable read only memory (EEPROM), a hard disk, and other well known facilities for storing data and storing and executing programs, including a program that implements the functions of the moving picture distribution apparatus 10 (or 10A) in the embodiments described below. The invention can be practiced by installing such a program in a computer. Other programs for communication processing, data processing, and so on may also be installed as necessary.
The moving picture decoding apparatus 200 may also be either a specialized device or a general-purpose computing device equipped with specialized software, and may include a display unit for displaying the decoded video signal output by the moving picture receiving apparatus 20, or a data storage unit for storing the decoded video signal output by the moving picture receiving apparatus 20.
A detailed description of the moving picture receiving apparatus 20 will be omitted. The moving picture decoding apparatus 200 may have, for example, the structure shown by Brites et al.
First Embodiment
Referring to FIG. 2, the moving picture encoding apparatus 100 in the first embodiment comprises a frame classifier 105, a Wyner-Ziv frame encoder 110 and a key frame encoder 160. The frame classifier 105, shown schematically as a switch, receives the incoming frame sequence, designates each frame as a key frame or a non-key frame, sends the key frames to the key frame encoder 160, and sends the non-key frames to the Wyner-Ziv frame encoder 110. The Wyner-Ziv frame encoder 110 and key frame encoder 160 both output encoded data to the moving picture receiving apparatus 20 in FIG. 1. Non-key frames will be referred to below as Wyner-Ziv frames.
The frame classifier 105 may designate frames as key frames or Wyner-Ziv frames according to their position in the frame sequence. For example, odd-numbered frames may be key frames and even-numbered frames may be Wyner-Ziv frames. Alternatively, a scheme in which Wyner-Ziv frames occur consecutively, or key frames occur consecutively, may be used. The scheme may be selected for compatibility with existing DVC equipment.
The key frame encoder 160 receives and encodes the key frames and sends the encoded data to the moving picture receiving apparatus 20. Known intraframe coding methods such as the methods standardized by the Joint Photographic Experts Group (JPEG) or the Moving Picture Experts Group (MPEG) may be used, as taught by Aaron et al.
Distributed video coding of Wyner-Ziv frames may be carried out either in the pixel domain, by processing the pixel data directly, or in the transform domain, by first performing a transform such as a DCT, as also taught by Aaron et al. The moving picture encoding apparatus 100 will be described as operating in the transform domain, but the invention is not limited to either domain.
The Wyner-Ziv frame encoder 110 includes a quantizer 120, a Slepian-Wolf encoder 130, an encoding rate estimator 140, and a side information generator 150.
The quantizer 120 receives, transforms, and quantizes Wyner-Ziv frames, divides the quantized data into bit planes, and gives the bit planes to the Slepian-Wolf encoder 130 and encoding rate estimator 140.
The side information generator 150 generates a predicted image of a Wyner-Ziv frame from either or both of the key-frames preceding and following the Wyner-Ziv frame, transforms and quantizes the predicted image, and divides the quantized data into bit planes.
The encoding rate estimator 140 estimates an encoding rate that enables prediction errors in the input bit planes to be corrected, by using the original image of the input frame and the predicted image generated by the side information generator 150. Details will be given later.
The Slepian-Wolf encoder 130 receives quantized bit planes from the quantizer 120, performs Slepian-Wolf encoding of each bit plane by using a turbo code or a low-density parity-check (LDPC) code, for example, and sends resulting error-correcting code bits or parity bits to the moving picture receiving apparatus 20 at the encoding rate calculated by the encoding rate estimator 140. Instead of the term ‘parity bits’, the general term ‘error-correcting code bits’ will be used for uniformity below.
The encoding rate estimator 140 includes a distribution model entropy calculator 141, an encoding rate correction calculator 142, a correction decision unit 143, and a correction unit 144. The basic operation of each component of the encoding rate estimator 140 is described below. Further details will be given later.
The distribution model entropy calculator 141 obtains the bit planes of the transformed original image in the Wyner-Ziv frame from the quantizer 120 and the coefficients of the corresponding predicted image from the side information generator 150, estimates the error probability of each bit in the bit planes by using a distribution model, and then calculates an entropy.
The encoding rate correction calculator 142 obtains the bit planes of the transformed original image in the Wyner-Ziv frame from the quantizer 120 and the bit planes of the corresponding predicted image from the side information generator 150 and calculates an encoding rate correction.
The correction decision unit 143 compares the entropy obtained from the distribution model entropy calculator 141 and an entropy obtained by measuring bit differences (prediction errors) between the bit planes of the original image and the corresponding bit planes of the predicted image, and decides whether the encoding rate correction is needed.
The correction unit 144 corrects the calculation made by the distribution model entropy calculator 141 in accordance with the decision made by the correction decision unit 143 and supplies the corrected information to the Slepian-Wolf encoder 130 as the final encoding rate.
The encoding and decoding of Wyner-Ziv frames will now be described with reference to the flowchart in FIGS. 3 to 5. It will be assumed that the frame sequence consists of odd-numbered key frames, which are encoded by conventional methods, and even-numbered Wyner-Ziv frames, which are encoded as follows.
Referring to FIG. 3, when the encoding of a Wyner-Ziv frame such as the second frame begins, the frame is input to the quantizer 120 (step S301) and transformed to the coefficient domain by a DCT, and the coefficients are grouped into spatial frequency bands, quantized, and divided into bit planes (step S302). In each band, the bit planes may cover the entire area of the frame, or the frame area may be divided into sub-areas and the coefficient data corresponding to each sub-area may be divided into separate bit planes.
The image of either the preceding or following key frame or both is input from the key frame encoder 160 to the side information generator 150 (step S303) and a predicted image is generated for the Wyner-Ziv frame (step S304). The predicted image may be generated by using the preceding key frame image as the predicted image, by averaging the preceding and following key frame images, by carrying out a limited motion search, or by various other methods. The predicted image generated by the side information generator 150 is transformed and quantized, and the quantized data are divided into bit planes (step S305). The length of the bit planes in step S305 is the same as the length of the bit planes obtained by the quantizer 120.
The coefficients of the original image of the frame are input from the quantizer 120 to the distribution model entropy calculator 141 (step S306). In the next few steps, the distribution model entropy calculator 141 estimates the number of prediction errors on the assumption that the distribution of differences in coefficients between the original image and the predicted image can be approximately modeled by a Laplacian distribution, and calculates an entropy value.
First, the coefficients of the predicted image of the current frame are input from the side information generator 150 to the distribution model entropy calculator 141 (step S307).
The distribution model entropy calculator 141 calculates the parameter α of the Laplacian distribution for each band (step S308). The parameter α of the Laplacian distribution is related to the variance σ2 of the distribution as in equation (1) below. If the mean value of the distribution is 0, the variance σ2 can be calculated by equation (2) below, where X and Y represent coefficients derived from the original image and the predicted image, respectively, and N is the length of the bit plane.
α 2 = 2 σ 2 ( 1 ) σ 2 = 1 N ( X - Y ) 2 - ( X - Y _ ) 2 = 1 N ( X - Y ) 2 ( 2 )
A bit plane of the original image is now input from the quantizer 120 to the distribution model entropy calculator 141 (step S309). The corresponding bit plane of the predicted image is input from the side information generator 150 to the distribution model entropy calculator 141 (step S310). The distribution model entropy calculator 141 calculates an error probability by using the distribution model as follows (step S311).
If X is a random variable representing a coefficient of the original image and Y is a random variable representing the corresponding coefficient of the predicted image, the conditional probability distribution of X for a given value of Y can be represented by equation (3). The probability Pn that the nth bit xn j in the jth bit plane in the original image has the value ‘1’ can be expressed in terms of conditional probabilities as in equation (4), where xn j-1 is the corresponding bit in the preceding (j−1)th bit plane of the original image and yn j is the corresponding bit in the jth bit plane of the predicted image. The (j−1)th bit plane is more significant than the jth bit plane, and will normally have been processed before the jth bit plane is processed.
P ( X Y ) = α 2 exp ( - α Y - X ) ( 3 ) P n = P ( X j = 1 , X j - 1 = x j - 1 n Y = y n ) p ( X j = x j n Y j = y j n ) ( 4 )
Since the entropy calculation that follows is symmetrical with respect to Pn and 1−Pn and gives the amount of information needed to correct errors, the value of Pn can be considered as representing an error probability.
From the error probability Pn, the distribution model entropy calculator 141 calculates the entropy H(Pn) given by the equation (5) below (step S312 in FIG. 4).
H(P n)=−P n log(P n)−(1−P n)log(1−P n)  (5)
The entropy expressed by equation (5) represents the amount of information needed to correct an error in the nth bit in the jth bit plane. If the same calculation is carried out for each bit in the jth bit plane and the results are averaged, the conditional entropy HX|Y given by equation (6) below is obtained, representing the amount of information needed to correct errors in the entire jth bit plane. This entropy value will be referred to as the distribution model entropy. This value also represents the first encoding rate calculated by the encoding rate estimator 140.
H X Y = 1 N H ( P n ) ( 6 )
The distribution model entropy calculated by the distribution model entropy calculator 141 is input to the correction unit 144 (step S313).
In the next few steps an encoding rate correction is prepared for use if the first encoding rate is not adequate for decoding. The relative error entropy described by Brites et al. will be used as an exemplary encoding rate correction.
To calculate the relative error entropy, the bit plane of the original image is input from the quantizer 120 to the encoding rate correction calculator 142 (step S314); the corresponding bit plane of the predicted image is input from the side information generator 150 to the encoding rate correction calculator 142 (step S315); the encoding rate correction calculator 142 compares the two input bit planes and calculates a relative error probability (step S316). The relative error probability can be obtained by dividing the number of bits that differ from the corresponding predicted bits in the current bit plane but not in higher bit planes by the length of the bit plane. The relative error entropy is then calculated from the relative error probability as shown by Brites et al. to obtain the encoding rate correction.
The encoding rate correction obtained by the encoding rate correction calculator 142 is input to the correction unit 144 (step S317), where it is treated as a supplementary encoding rate. To decide whether to apply the encoding rate correction, the correction decision unit 143 compares the distribution model entropy with, for example, a bit error entropy calculated as follows.
The distribution model entropy HX|Y is input from the distribution model entropy calculator 141 to the correction decision unit 143 (step S318). The corresponding bit plane of the original image is input from the quantizer 120 to the correction decision unit 143 (step S319). The corresponding bit plane of the predicted image is input from the side information generator 150 to the correction decision unit 143 (step S320).
By comparing the corresponding bit planes the original and predicted images, the correction decision unit 143 obtains an error count e. Dividing the error count e by the bit plane length N gives a bit error rate err (err=e/N) (step S321). The bit error entropy H(err) is calculated from the bit error rate as in equation (7) below. The encoding rate represented by the bit error entropy will be referred to as the second encoding rate.
H(err)=−err log(err)−(1−err)log(1−err)  (7)
The correction decision unit 143 now compares the first and second encoding rates, that is, the bit error entropy H(err) and the distribution model entropy HX|Y, to decide whether or not to supplement the first encoding rate (step S322).
The first encoding rate or distribution model entropy is based on an assumed general model, while the second encoding rate or bit error entropy is based on an actual comparison of the original and predicted images. If the first encoding rate is equal to or greater than the second encoding rate, it can be inferred that the first encoding rate is sufficient to enable the decoding apparatus to correct the errors that will actually occur in the predicted image generated at the decoding apparatus, and the correction decision unit 143 decides to use the first encoding rate as the final coding rate. If the second encoding rate exceeds the first encoding rate, it can be inferred that the first encoding rate may be inadequate, so the correction decision unit 143 decides to supplement the first coding rate.
The decision made by the correction decision unit 143 is input to the correction unit 144 (step S323 in FIG. 5) and tested (step S324). If the correction decision unit 143 has decided not to supplement the first coding rate, the first encoding rate becomes the final coding rate output by the encoding rate estimator 140 (step S325). Otherwise, the correction unit 144 supplements the first coding rate by adding the encoding rate correction, that is, by adding the relative error entropy to the distribution model entropy, to obtain the final encoding rate (step S326).
The correction unit 144 supplies the final encoding rate to the Wyner-Ziv frame encoder 110 (step S327). The bit plane to be encoded is now input from the quantizer 120 to the Slepian-Wolf encoder 130 (step S328) and encoded (step S329), and the Slepian-Wolf encoder 130 outputs an error-correcting code at the final encoding rate determined by the encoding rate estimator 140 (step S330).
The Wyner-Ziv frame encoder 110 now decides whether all bit planes of the current frame have been processed (step S331). If so, the processing of the frame ends. Otherwise, the Wyner-Ziv frame encoder 110 returns to step S309 to process the next bit plane.
In contrast to the prior art, the Wyner-Ziv frame encoder 110 in the first embodiment does not always add a correction to the first encoding rate obtained from the distribution model. A second coding rate is calculated by a different method (e.g., bit error entropy), the first and second encoding rates are compared, and the first encoding rate is corrected, by adding a supplementary encoding rate, only if the comparison indicates a correction to be necessary. The Wyner-Ziv frame encoder is thereby able to avoid sending the moving picture decoding apparatus 200 an unnecessarily large amount of encoded data. The first embodiment accordingly saves time, and conserves computational resources and communication resources.
Second Embodiment
Referring again to FIG. 1, the moving picture distribution system 1A in the second embodiment has the same moving picture receiving apparatus 20 as in the first embodiment, but the moving picture distribution apparatus 10A includes a different moving picture encoding apparatus 100A. Referring to FIG. 6, the moving picture encoding apparatus 100A includes the same frame classifier 105 and key frame encoder 160 as in the first embodiment, but has a Wyner-Ziv frame encoder 110A including an encoding rate estimator 140A with an encoding rate correction calculator 142A that operates only when the correction decision unit 143 decides to supplement the encoding rate. The other components of the encoding rate estimator 140A operate as in the first embodiment.
The operation of the second embodiment will now be described with reference to the flowcharts in FIGS. 7 to 10. It will again be assumed that the frame sequence consists of odd-numbered key frames, which are encoded by conventional methods, and even-numbered Wyner-Ziv frames, which are encoded as follows.
The encoding rate correction calculated by the encoding rate correction calculator 142A in the second embodiment is the relative error entropy indicated by Brites et al., as in the first embodiment.
The quantizer 120 transforms and quantizes each input Wyner-Ziv frame and divides the quantized data into bit planes; the side information generator 150 generates a predicted image; and the distribution model entropy calculator 141 calculates a distribution model entropy (steps S501 to S513). The same operations are performed in steps S501 to S513 as in steps S301 to S313 in the first embodiment, so descriptions will be omitted.
In the next few steps, the correction decision unit 143 decides whether the first encoding rate obtained by the distribution model entropy calculator 141 needs to be supplemented. This decision is made by, for example, comparing the distribution model entropy with a bit error entropy, as in the first embodiment: the distribution model entropy is input from the distribution model entropy calculator 141 to the correction decision unit 143 (step S514 in FIG. 8); one bit plane of the original image is input from the quantizer 120 to the correction decision unit 143 (step S515); the corresponding bit plane of the predicted image is input from the side information generator 150 to the correction decision unit 143 (step S516); the bit planes of the original image and the predicted image are compared, the number of bit errors is counted, and a bit error entropy is calculated (step S517); the correction decision unit 143 compares the bit error entropy and the distribution model entropy and decides whether to supplement the encoding rate obtained from the distribution model entropy (step S518); the decision made by the correction decision unit 143 is input to the correction unit 144 (step S519) and tested (step S520).
If the correction decision unit 143 decides to supplement the first encoding rate, it also passes this decision to the encoding rate correction calculator 142A (step S521 in FIG. 9). The encoding rate correction calculator 142A then obtains the bit plane of the original image from the quantizer 120 (step S522) and the corresponding bit plane of the predicted image from the side information generator 150 (step S523), compares these two bit planes, calculates a relative error probability, and derives an encoding rate correction (step S524), which is provided to the correction unit 144 (step S525). The correction unit 144 adds the encoding rate correction to the encoding rate calculated from the distribution model entropy (step S526). The resulting sum is the final encoding rate output by the encoding rate estimator 140A.
If the correction unit 144 finds in step S520 in FIG. 8 that the correction decision unit 143 has decided not to supplement the first encoding rate, the correction unit 144 uses the first encoding rate, i.e., the distribution model entropy calculated by the distribution model entropy calculator 141, as the final encoding rate (step S527 in FIG. 9).
Following step S526 or S527, the final encoding rate is input to the Slepian-Wolf encoder 130 (step S528 in FIG. 10), the bit plane to be encoded is input from the quantizer 120 to the Slepian-Wolf encoder 130 (step S529) and encoded (step S530), and the resulting error-correcting code is output at the final encoding rate obtained by the encoding rate estimator 140A and transmitted to the moving picture decoding apparatus 200 (step S531).
The Wyner-Ziv frame encoder 110A then determines whether all bit planes of the current frame have been processed (step S532). If so, the processing of the frame ends. Otherwise, the processing returns to step S509.
The effect of the second embodiment is to reduce the amount of encoding processing, because the encoding rate correction calculator 142A calculates the encoding rate correction only when the correction decision unit 143 decides that the encoding rate correction is necessary.
The present invention is not limited to the embodiments described above. The following are some of the possible modifications.
In FIG. 1, the moving picture encoding apparatus is shown as part of a moving picture distribution apparatus. The moving picture encoding apparatus may, however, be a stand-alone apparatus for the output of encoded data generated from an input sequence of frames. The output may take the form of storing the data in a disk drive or other storage medium, or outputting the data to another apparatus.
The second encoding rate estimate obtained by the encoding rate estimator need not be calculated by the bit error entropy method; it may be calculated by any method differing from the distribution model entropy method used to calculate the first encoding rate estimate. Alternatively, the encoding rate estimator may calculate a plurality of second encoding rates by different methods, compare them with the first encoding rate calculated by the distribution model entropy method, and then decide whether to supplement the first encoding rate. For example, the encoding rate estimator 140 may be decide to supplement the first encoding rate if at least one of the second encoding rates exceeds the first encoding rate.
In another variation, the bit error entropy method is used to calculate the first encoding rate, and the distribution model entropy method is used to calculate the second encoding rate.
Those skilled in the art will recognize that further variations are possible within the scope of the invention, which is defined in the appended claims.

Claims (4)

What is claimed is:
1. A moving picture encoding apparatus for encoding a moving picture signal including a sequence of frames, the moving picture encoding apparatus comprising:
a frame classifier for designating each frame in the sequence as a key frame or a non-key frame;
a key frame encoder for encoding the designated key frames in the sequence to generate encoded key frame data;
a side information generator for using the encoded key frame data to generate a predicted image for each designated non-key frame in the sequence;
an encoding rate estimator for determining a final encoding rate for said each designated non-key frame by estimating a number of error-correcting bits needed to correct prediction errors in the predicted image generated for said each designated non-key frame; and
a non-key frame encoder for encoding said each designated non-key frame to obtain an error-correcting code at the final encoding rate determined by the encoding rate estimator, and outputting the error-correcting code as encoded data of said each designated non-key frame;
wherein the encoding rate estimator includes
a first encoding rate calculator configured to use a model that approximates a relation between said each designated non-key frame and the predicted image to calculate a distribution model entropy, and to calculate a first encoding rate using the distribution model entropy, the distribution model entropy corresponding to a first prediction error rate of the predicted image with respect to said each designated non-key frame,
an encoding rate correction calculator configured to compare a bit plane in said each designated non-key frame with a corresponding bit plane in the predicted image, and to calculate a relative error probability, to thereby determine an encoding rate correction,
a second encoding rate calculator configured to compare said each designated non-key frame with the predicted image to determine a difference therebetween, to calculate a bit error entropy using the difference, and to calculate a second encoding rate using the bit error entropy, the bit error entropy corresponding to a second prediction error rate of the predicted image with respect to said each designated non-key frame,
a correction decision unit configured to compare the distribution model entropy and the bit error entropy, and to decide whether or not to supplement the first encoding rate, the correction decision unit deciding to supplement the first encoding rate when the bit error entropy is larger than the distribution model entropy, and
a correction unit configured to use the first encoding rate as the final encoding rate when the correction decision unit decides not to supplement the first encoding rate, and to use a supplemented first encoding rate, obtained by adding the encoding rate correction to the first encoding rate, as the final coding rate when the correction decision unit decides to supplement the first encoding rate.
2. The moving picture encoding apparatus of claim 1, wherein the encoding rate correction calculator determines the encoding rate correction only when the correction decision unit decides to supplement the first encoding rate.
3. A machine-readable non-transitory medium storing a program executable by a computing device to encode a moving picture signal including a sequence of frames, the program including instructions for:
designating each frame in the sequence as a key frame or a non-key frame;
encoding the designated key frames in the sequence to generate encoded key frame data;
using the encoded key frame data to generate a predicted image for each designated non-key frame in the sequence;
calculating a distribution model entropy using a model that approximates a relation between said each designated non-key frame and the predicted image, and calculating a first encoding rate using the distribution model entropy, the distribution model entropy corresponding to a first prediction error rate of the predicted image with respect to said each designated non-key frame;
comparing said each designated non-key frame with the predicted image to determine a difference therebetween, calculating a bit error entropy using the difference, and calculating a second encoding rate using the bit error entropy, the bit error entropy corresponding to a second prediction error rate of the predicted image with respect to said each designated non-key frame;
comparing the distribution model entropy and the bit error entropy, and making a decision as to whether to supplement the first encoding rate, said making a decision including deciding to supplement the first encoding rate when the bit error entropy is larger than the distribution model entropy;
comparing a bit plane in said each designated non-key frame with a corresponding bit plane in the predicted image, and calculating a relative error probability, to thereby determine an encoding rate correction;
using the first encoding rate as a final encoding rate when the decision is not to supplement the first encoding rate, and using a supplemented first encoding rate, obtained by adding the encoding rate correction to the first encoding rate, as the final coding rate when the decision is to supplement the first encoding rate; and
encoding said each designated non-key frame to obtain an error-correcting code at the final encoding rate, and outputting the error-correcting code as encoded data of said each designated non-key frame.
4. A moving picture distribution system comprising:
a moving picture distribution apparatus including the moving picture encoding apparatus of claim 1; and
a moving picture receiving apparatus including a moving picture decoding apparatus for receiving the encoded key frame data and the encoded data of said each designated non-key frame and decoding the encoded key frame data and the encoded data of said each designated non-key frame.
US12/923,421 2009-09-30 2010-09-21 Moving picture encoding apparatus and distribution system Expired - Fee Related US8442114B2 (en)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
JP2009227919A JP5233939B2 (en) 2009-09-30 2009-09-30 Moving picture encoding apparatus and program, and moving picture distribution system
JP2009-227919 2009-09-30

Publications (2)

Publication Number Publication Date
US20110075733A1 US20110075733A1 (en) 2011-03-31
US8442114B2 true US8442114B2 (en) 2013-05-14

Family

ID=43780362

Family Applications (1)

Application Number Title Priority Date Filing Date
US12/923,421 Expired - Fee Related US8442114B2 (en) 2009-09-30 2010-09-21 Moving picture encoding apparatus and distribution system

Country Status (2)

Country Link
US (1) US8442114B2 (en)
JP (1) JP5233939B2 (en)

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US10924625B2 (en) * 2019-03-20 2021-02-16 Xerox Corporation Dynamic compression acceleration using real-time image data entropy analysis

Families Citing this family (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US8111755B2 (en) * 2008-06-25 2012-02-07 International Business Machines Corporation Method and system for low-complexity Slepian-Wolf rate estimation in Wyner-Ziv video encoding
JP5845750B2 (en) * 2011-09-12 2016-01-20 沖電気工業株式会社 Image coding system, image coding apparatus, and image coding program
JP5811823B2 (en) * 2011-12-12 2015-11-11 沖電気工業株式会社 Video encoding apparatus, system, and program
JP5900024B2 (en) * 2012-03-02 2016-04-06 沖電気工業株式会社 Moving picture encoding apparatus and program, moving picture decoding apparatus and program, and moving picture distribution system
JP6127487B2 (en) * 2012-12-10 2017-05-17 沖電気工業株式会社 Moving picture coding apparatus and program
JP6398149B2 (en) * 2013-04-12 2018-10-03 沖電気工業株式会社 Moving picture coding apparatus and program, and moving picture coding system
WO2023050431A1 (en) * 2021-09-30 2023-04-06 浙江大学 Encoding method, decoding method, decoder, encoder and computer-readable storage medium

Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20040076332A1 (en) * 2002-10-18 2004-04-22 Shunan Lin Adaptive error-resilient video encoding using multiple description motion compensation
US20060140269A1 (en) * 2003-01-30 2006-06-29 Bruls Wilhelmus Hendrikus A Video coding
US20080189073A1 (en) * 2007-01-16 2008-08-07 Ashish Jagmohan Method and Apparatus for Multi-Hypothesis Decoder Side-Information Coding

Family Cites Families (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP4559126B2 (en) * 2004-06-01 2010-10-06 日本電信電話株式会社 Video transmission method, video transmission apparatus, video transmission program, and computer-readable recording medium recording the program
JP2006340205A (en) * 2005-06-03 2006-12-14 Sharp Corp Error correction device
JP4819024B2 (en) * 2007-12-06 2011-11-16 日本電信電話株式会社 Image encoding / decoding method and system

Patent Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20040076332A1 (en) * 2002-10-18 2004-04-22 Shunan Lin Adaptive error-resilient video encoding using multiple description motion compensation
US20060140269A1 (en) * 2003-01-30 2006-06-29 Bruls Wilhelmus Hendrikus A Video coding
US20080189073A1 (en) * 2007-01-16 2008-08-07 Ashish Jagmohan Method and Apparatus for Multi-Hypothesis Decoder Side-Information Coding

Non-Patent Citations (2)

* Cited by examiner, † Cited by third party
Title
Anne Aaron, Shantanu Rane, Eric Setton, and Bernd Girod, "Transform-domain Wyner-Ziv Codec for Video." In: Proc, SPIE Visual Communications and Image Processing, San Jose, CA (2004).
C. Brites, F. Pereira, "Encoder Rate Control for Transform Domain Wyner-Ziv Video Coding", ICIP 2007, USA (Sep. 2007).

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US10924625B2 (en) * 2019-03-20 2021-02-16 Xerox Corporation Dynamic compression acceleration using real-time image data entropy analysis

Also Published As

Publication number Publication date
JP2011077881A (en) 2011-04-14
US20110075733A1 (en) 2011-03-31
JP5233939B2 (en) 2013-07-10

Similar Documents

Publication Publication Date Title
US8442114B2 (en) Moving picture encoding apparatus and distribution system
US8374255B2 (en) Moving picture encoder, decoder, and coding system
US8189682B2 (en) Decoding system and method for error correction with side information and correlation updater
US9667961B2 (en) Video encoding and decoding apparatus, method, and system
US8428124B2 (en) Moving picture encoding apparatus and distribution system
US8363733B2 (en) Video encoder and decoder apparatus deciding error in transform coefficients between an original image and a predictive image
US20130266078A1 (en) Method and device for correlation channel estimation
US20050281479A1 (en) Method of and apparatus for estimating noise of input image based on motion compensation, method of eliminating noise of input image and encoding video using the method for estimating noise of input image, and recording media having recorded thereon program for implementing those methods
JP5133290B2 (en) Video encoding apparatus and decoding apparatus
EP2263382A2 (en) Method and apparatus for encoding and decoding image
JP2003259372A (en) Method and apparatus to encode moving image with fixed computation complexity
US8243821B2 (en) For spatial Wyner Ziv coding
JP2012124890A (en) Method and decoder for decoding encoded signal based on statistical dependencies between encoded signal and side information
US8446949B2 (en) Distributed coded video decoding apparatus and method capable of successively improving side information on the basis of reliability of reconstructed data
Sheng et al. Feedback-free rate-allocation scheme for transform domain Wyner–Ziv video coding
Slowack et al. Distributed video coding with feedback channel constraints
US8218652B2 (en) Image encoder, image decoder, and image encoding system
US20040008898A1 (en) Method and apparatus for controlling amount of DCT computation performed to encode motion image
KR101035746B1 (en) Method of distributed motion estimation for video encoder and video decoder
KR101639434B1 (en) Wyner-Ziv coding and decoding system and method
KR101152482B1 (en) Methods of encoding and decoding using fast ldpca code and apparatuses using the same
JP2843024B2 (en) Method and apparatus for selecting transform coefficient of transform coding system
US8670480B2 (en) Transcoder
Min et al. Adaptive distributed video coding with motion vectors through a back channel
KR101038205B1 (en) Method and apparatus of modeling virtual channel used for distributed video coding scheme

Legal Events

Date Code Title Description
AS Assignment

Owner name: OKI ELECTRIC INDUSTRY CO., LTD., JAPAN

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:YAMASAKI, TAKAHIRO;REEL/FRAME:025059/0144

Effective date: 20100611

STCF Information on status: patent grant

Free format text: PATENTED CASE

FEPP Fee payment procedure

Free format text: PAYOR NUMBER ASSIGNED (ORIGINAL EVENT CODE: ASPN); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY

FPAY Fee payment

Year of fee payment: 4

FEPP Fee payment procedure

Free format text: MAINTENANCE FEE REMINDER MAILED (ORIGINAL EVENT CODE: REM.); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY

LAPS Lapse for failure to pay maintenance fees

Free format text: PATENT EXPIRED FOR FAILURE TO PAY MAINTENANCE FEES (ORIGINAL EVENT CODE: EXP.); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY

STCH Information on status: patent discontinuation

Free format text: PATENT EXPIRED DUE TO NONPAYMENT OF MAINTENANCE FEES UNDER 37 CFR 1.362

FP Lapsed due to failure to pay maintenance fee

Effective date: 20210514