WO2009091387A1 - Scintillement vidéo réduit - Google Patents

Scintillement vidéo réduit Download PDF

Info

Publication number
WO2009091387A1
WO2009091387A1 PCT/US2008/013962 US2008013962W WO2009091387A1 WO 2009091387 A1 WO2009091387 A1 WO 2009091387A1 US 2008013962 W US2008013962 W US 2008013962W WO 2009091387 A1 WO2009091387 A1 WO 2009091387A1
Authority
WO
WIPO (PCT)
Prior art keywords
source
coding
coded
modified
inter
Prior art date
Application number
PCT/US2008/013962
Other languages
English (en)
Inventor
Hua Yang
Jill Macdonald Boyce
Gad Moshe Berger
Original Assignee
Thomson Licensing
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Thomson Licensing filed Critical Thomson Licensing
Priority to US12/735,342 priority Critical patent/US20100278236A1/en
Publication of WO2009091387A1 publication Critical patent/WO2009091387A1/fr

Links

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/10Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
    • H04N19/102Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the element, parameter or selection affected or controlled by the adaptive coding
    • H04N19/103Selection of coding mode or of prediction mode
    • H04N19/107Selection of coding mode or of prediction mode between spatial and temporal predictive coding, e.g. picture refresh
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/10Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
    • H04N19/102Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the element, parameter or selection affected or controlled by the adaptive coding
    • H04N19/103Selection of coding mode or of prediction mode
    • H04N19/109Selection of coding mode or of prediction mode among a plurality of temporal predictive coding modes
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/50Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding
    • H04N19/503Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding involving temporal prediction
    • H04N19/51Motion estimation or motion compensation
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/60Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using transform coding
    • H04N19/61Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using transform coding in combination with predictive coding
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/85Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using pre-processing or post-processing specially adapted for video compression
    • H04N19/86Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using pre-processing or post-processing specially adapted for video compression involving reduction of coding artifacts, e.g. of blockiness

Definitions

  • At least one disclosed implementation generally relates to video processing and, more particularly, to compressed video.
  • GOPs The first frame of each GOP is coded as an I-frame.
  • the last frame of each GOP is coded as a P-frame.
  • the remaining frames of the GOP are coded as either P-frames or B-frames.
  • P- or B-frames involve inter- frame prediction, also called inter-prediction.
  • I-frames involve either intra-frame prediction, also called intra-prediction, or no prediction at all.
  • P-frames involve inter- prediction from prior frames only.
  • B-frames involve inter-prediction from both prior and subsequent frames.
  • Original video signals have naturally smooth optical flows. However, after lossy video encoding, the natural optical flow will be distorted in the coded video signals. The resultant temporal inconsistency across coded frames will then be perceived as the flickering artifact. In practice, flickering is more often perceived in static or low motion areas of a coded video. For example, several consecutive frames may share the same static background. Hence, all the collocated pixels in the static background across these frames bear the same or similar pixel values in the original input video. However, in video encoding, the collocated pixels of these frames may be predicted from different reference pixels in different frames, and hence after quantizing the residue, may yield different reconstruction values. Visually, the increased inter- frame differences across these frames will be perceived as flickering when the coded video is playing out.
  • a flickering artifact is typically more intensive for low or medium bit rate coding due to coarse quantization. Also, it is typically more obviously observed on I-frames than on P- or B-frames. This may be because for the same static areas, the prediction residue resultant from inter-frame prediction in P- or B-frames is usually much smaller than the resultant from intra-frame prediction or no-prediction in I-frames. Thus, with coarse quantization, the reconstructed static areas in an I-frame may demonstrate more noticeable difference from the collocated areas in previous P- or B-frames, and hence, a more noticeable flickering artifact.
  • a source image is inter-coded to produce coded source data.
  • the coded source data is decoded to produce a modified source.
  • the modified source is intra-coded to produce coded modified-source data.
  • the coded modified-source data is decoded to produce a reference image.
  • implementations may be configured or embodied in various manners. For example, an implementation may be performed as a method, or embodied as an apparatus configured to perform a set of operations, or embodied as an apparatus storing instructions for performing a set of operations, or embodied in a signal. Other aspects and features will become apparent from the following detailed description considered in conjunction with the accompanying drawings and the claims.
  • FIG. 1 is a schematic diagram of a relationship among frames at the end of one group of pictures and the beginning of a next group of pictures.
  • FIG. 2 is a block diagram of an implementation of an apparatus for encoding and transmitting or storing picture data.
  • FIG. 3 is a block diagram of an implementation of an apparatus for processing and transmitting or storing picture data.
  • FIG. 4 is a schematic diagram of another relationship among frames at the end of one group of pictures and the beginning of a next group of pictures in an implementation.
  • FIG. 5 is a process flow diagram of an implementation of a method for reducing an artifact in a portion of a reference frame.
  • FIG. 6 is a process flow diagram of another implementation of a method for reducing an artifact in a portion of a reference frame.
  • FIG. 7 is a block diagram of an implementation of an encoder adapted to reduce an artifact in a portion of a reference frame.
  • FIG. 8 is a process flow diagram of another implementation of a method for reducing an artifact in a portion of a reference frame.
  • FIG. 9 is a block diagram of an implementation of a video transmission system.
  • FIG. 10 is a block diagram of an implementation of a video receiving system.
  • FIG. 1 a digital video sequence is shown schematically.
  • the final three P-frames 104, 106, 108 of a group of pictures are shown.
  • frames is used in referring to I-frames, P-frames, and B-frames for convenience and due to the familiarity of the terms.
  • the more general term "pictures” is understood to be applicable and implementations may use I-pictures, P- pictures, and/or B-pictures.
  • the P-frames 104, 106, 108, 122, 124, and the I-frame 120 all represent actual frames having pixel data.
  • Each of these frames is a reconstruction (that is, decoded version of) a source frame.
  • source frame 110 is coded to produce coded data (for example, a coded bitstream) that represents the encoding.
  • coded data is then decoded to produce a reconstructed frame which is I-frame 120.
  • P-frames 104, 106, 108 use an I-frame of the first group of pictures as a reference. That I-frame was in turn encoded from a source frame.
  • a source frame 1 10 is encoded as the I-frame 120 of a next group of pictures.
  • P-frames 122, 124 use I-frame 120 as a prediction reference frame.
  • Flicker results from a discontinuity between P-frame 108 and I-frame 120.
  • One challenge is to reduce flicker in a manner that is transparent to the decoder.
  • a variety of different decoders may be used for any video. Accordingly, it is advantageous for any method and system of de-flickering to be implemented at an encoder and prior to transmission.
  • a system 200 includes an encoder 210 coupled to a transmit/store device 220.
  • the encoder 210 and the transmit/store device 220 may be implemented, for example, on a computer or a communications encoder.
  • the encoder 210 accesses unencoded picture data 205, and encodes the picture data according to one or more of a variety of coding algorithms, and provides an encoded data output 215 to the transmit/store device 220.
  • the transmit/store device 220 may include one or more of a storage device or a transmission device. Accordingly, the transmit/store device 220 accesses the encoded data 215 and either transmits the data 215 or stores the data 215.
  • a system 300 includes a processing device 310 coupled to a local storage device 320 and coupled to a transmit/store device 330.
  • the processing device 310 accesses unencoded picture data 305.
  • the processing device 310 encodes the picture data according to one or more of a variety of coding algorithms, and provides an encoded, data output 315 to the transmit/store device 330.
  • the processing device 310 may include, for example, the encoder 210.
  • the processing device 310 may cause data, including unencoded picture data, encoded picture data, and elements thereof to be stored in the local storage device 320, and may retrieve such data from the local storage device 320.
  • the transmit/store device 330 may include one or more of a storage device or a transmission device. Transmit/store device 330 may transmit the encoded picture data, including intra- coded reconstructed portions and inter-coded subsequent portions, as discussed below, to one or more decoders for decoding.
  • An I-frame in a video is an example of an I-picture
  • a P- frame in a video is an example of a P-picture.
  • P-frames 402, 404, 406 are the final frames of a group of pictures.
  • a no- flicker reference frame 412 is formed by encoding the unencoded current frame 410 to form a P-frame 412, predicted from previously coded P-frame 406. That is, the source frame 410 is inter-coded, using P-frame 406 as a prediction reference.
  • the inter-coding in a typical implementation, produces coded data that is decoded to produce the P-frame 412.
  • the P- frame 412 is then intra-coded (I-frame encoded) to produce coded data.
  • the coded data is then decoded to produce I-frame 414.
  • I-frame 414 is based on a "source" of P-frame 412. So the original source frame 410 has effectively been coded twice.
  • the first coding produces the no-tlicker reference 412, and the second coding provides an intra-coded frame 414 to being a new GOP.
  • the current frame 410 is coded on a macroblock by macroblock basis in two passes.
  • a macroblock of frame 410 is coded as a P- frame macroblock, and the resultant reconstruction is taken as a macroblock of the no-flicker reference 412. This is done for each macroblock of frame 410 separately.
  • a pseudo-P-frame coding may be referred to as, for example, a partial inter- frame coding according to the ITU-T H.264 standard, which means that the normal inter- frame coding process of the standard is only partially performed.
  • a macroblock from frame 410 is pseudo-P-frame coded, and then reconstructed, to produce a macroblock of frame 412.
  • the pseudo-P-frame coding process may include, for example, not performing an entropy coding step.
  • the P-frame 412 macroblock is reconstructed from quantized data without performing an entropy decoding step.
  • the pseudo-P-frame coding process may also, or alternatively, include, for example, not checking all available coding modes, as explained further below.
  • the macroblock of P-frame 412 is generated so that the distortion of this macroblock is close to the distortion of the corresponding macroblock in the prediction reference P-frame 406. Recall that the distortion of the macroblock of P-frame 412 is with respect to the corresponding macroblock in source frame 410. Recall also that the corresponding macroblock of P-frame 406 is determined when the macroblock of source frame 410 is P-frame coded (or pseudo-P-frame-coded). Recall that the distortion of the corresponding macroblock of P-frame 406 is with respect to the corresponding macroblock in its source frame (not shown).
  • the macroblock of the no-flicker reference 412 is intra-coded to produce a macroblock of I-frame 414 of the next group of pictures.
  • This second pass uses the no-flicker reference 412 macroblock, instead of the original 410 macroblock, as the target macroblock.
  • a small quantization parameter (“QP") is applied in quantization to ensure that the reconstructed macroblock of I-frame 414 closely approaches the target macroblock of P- frame (no-flicker reference frame) 412.
  • the two passes in the method are: (i) deriving a no-flicker reference as the target frame; and (ii) conducting actual I-frame coding with a small enough QP (or with enough coding bits) to closely approach the target frame.
  • the annoying I-frame flickering artifact can be effectively removed, or at least reduced, from the coded video, as observed in extensive experiments.
  • frames 416, 418, and/or succeeding frames in the GOP may be B-frames.
  • Portions (such as macroblocks) of frames 416, 418 may be inter-coded using the corresponding portions of I-frame 414 as a reference. Such inter-coding will produce an inter- coded portion, such as an inter-coded macroblock.
  • the second reconstruction is I-frame 414, which is used as the reference for encoding frames 416, 418.
  • the above 2-pass implementation is performed on a macroblock-by- macroblock basis. In this way, the entire frame 412 does not need to be stored or generated at one time. After a macroblock of frame 412 is generated and I-frame coded, and after the macroblock is no longer needed, then the macroblock may be discarded (that is, removed from memory and no longer stored or saved). As a result, this implementation requires less memory than an implementation that generates and stores the entire frame 412. Further, the use of a pseudo-P-frame coding process allows savings, for example, in terms of processing requirements. Referring now to FIG. 5, a process flow of a method according to an implementation will be explained. The method may be implemented by encoder 210 of FIG.
  • the process flow may commence with determining 505, for a portion of a reference picture, to replace the portion.
  • the term "reference image” refers to all or part of a reference picture, such as, for example, a macroblock. The determination may be made, for example, in order to reduce an artifact when an encoding of the reference picture is decoded and displayed. If it is determined to replace the portion, the process flow may continue with replacing the portion.
  • Replacing the portion includes (1) inter-coding 510 the portion of the reference picture to produce an inter-coded portion, and (2) intra-coding 515 a reconstruction of the inter-coded portion to produce an intra-coded reconstructed portion.
  • the process flow continues with inter-coding 520 a portion of a subsequent picture using the intra-coded reconstructed portion as a reference to produce an inter-coded subsequent portion.
  • the determining 505 may include performing an optimal rate- distortion evaluation of multiple options for replacing the portion. Based on the evaluation, an optimal option may be selected.
  • the multiple options may include at least one inter-coding mode from an applicable video compression standard, and one intra-coding mode from the applicable standard.
  • the standard may be, for example, the ITU-T H.264 standard, which is equivalent to the 1SO/1EC MPEG-4 Part 10 standard (also called MPEG-4 AVC and ISO/IEC 14496-10), MPEG-2, H.263, or MPEG-4 Part 2, by way of example.
  • the multiple options may include inter-coding with a 16x16 block size, intra-coding with a 16x16 block size, and intra- coding with a 4x4 block size.
  • the process flow may be carried out for multiple portions of a picture, or each portion of a picture.
  • the portions may be macroblocks, by way of example.
  • the picture may be an I-picture, and may be the first picture in a group of pictures.
  • the I-picture may be a reference frame.
  • the process flow commences with determining whether to pseudo-P-code a portion of the reference frame.
  • a quantization parameter is selected 605 for a pseudo P-frame coding equal to the average quantization parameter of all of the macroblocks of the previously coded P- frame, i.e., the last P-frame of the prior group of pictures.
  • the quantization parameter may be selected for the entire picture.
  • the process flow proceeds to performing 610 motion estimation for pseudo P-frame coding.
  • checking only one coding mode may be employed.
  • only the Interl6xl6 mode may be checked, although a different mode may be the only mode checked.
  • More than one coding mode may also be checked.
  • Other modes include, for example, Interl6x8 and Inter8xl6.
  • the motion vector is estimated, for example, using the motion vectors of one or more neighboring macroblocks.
  • the search may be performed over a limited motion vector search range. The search attempts, for example, to minimize a measure of the residue resulting from various motion vectors.
  • a measure may be, for example, the sum of absolute differences ("SAD") or the sum of square differences ("SSD").
  • the motion vector search range may be [-5, +5]. However, other limited search ranges, e.g., 3, 6, or 10 pixels, may be employed. As the goal of identifying macroblocks with a high probability of flicker is identifying macroblocks that change not at all or very little, checking only the Inter 16x16 mode and searching over a limited motion vector search range is sufficient to find the accurate motion for low or medium motion macroblocks.
  • either integer pixel level or subpixel e.g., at the 1 A pixel or % pixel levels, level motion vector search may be performed. In experiments, superior deflickering has been obtained through subpixel level motion vector search. After motion estimation, a mode is selected 615 for pseudo P-frame coding.
  • the process of mode selection is also the final step in determining whether to perform pseudo P- frame coding. If the mode selection process returns a result of intracoding, then pseudo P- frame coding is not performed.
  • the mode selection process may be performed by using an optimal rate-distortion evaluation, such as a rate-distortion optimized encoding strategy, using the mode that results in a minimum of a Lagrangian rate-distortion cost analysis.
  • an optimal rate-distortion evaluation such as a rate-distortion optimized encoding strategy
  • four coding modes may be checked, which may be: Skip (in which a motion vector based on a previous macroblock is used), Interl ⁇ xl ⁇ , Intral ⁇ xl ⁇ , and Intra4x4.
  • the Lagrangian rate-distortion costs of at least one mode may be multiplied by a factor in order to compensate for the absence of other modes.
  • the Lagrangian rate-distortion costs of the Skip and Interl ⁇ xl ⁇ modes are multiplied by a factor, which may be 0.7. This attempts to compensate for the absence of the other Inter-prediction modes that may have produced better results. If additional coding modes are checked in an implementation, then the Lagrangian rate-distortion costs of certain modes may need to be multiplied by a different factor.
  • the selected mode is an inter-coding mode 620, then a relatively low motion macroblock is indicated.
  • the process flow then proceeds to indicate whether the macroblock is characterized by very low motion or moderately low motion. If the selected mode is Skip mode, or the prediction residue is below a threshold, then a very low motion macroblock is indicated. The implementation is therefore satisfied with the QP value. If either of those is true 625, then the macroblock is reconstructed 630 with the selected mode (that is, the selected mode is used to code the macroblock, and then the coded data is decoded) and the reconstructed macroblock is used as the no-flicker reference.
  • the prediction residue threshold may be represented as the mean-absolute-difference being less than 10.
  • an updated quantization parameter for the no-flicker reference is determined 635.
  • the QP is implicitly assumed to be too high, as evidenced by the residue being above a threshold.
  • An updated quantization parameter may be selected which minimizes, or at least reduces, the mean-square-error (MSE) distortion difference between the current macroblock coded using Inter 16x16 mode and the prediction macroblock from the previous P-frame. That is, we attempt to make the distortion of the current macroblock (measured with respect to the original source) the same as the distortion of the prediction macroblock (measured with respect to its source).
  • the macroblock may be encoded 640 using Inter 16x16 mode and using the updated quantization parameter to obtain, after reconstruction, the no-flicker reference for the current macroblock.
  • a second pass may include applying de-flickering I-frame coding to the no-flicker macroblock produced in either of operations 630 or 640.
  • the applying of de-flickering I- frame coding may include checking all the Intra-prediction modes; using the no-flicker reference derived from the first pass pseudo P-frame coding as the target macroblock; and using a small quantization parameter, such that the reconstruction closely approaches the no- flicker reference.
  • the process flow proceeds to encode 650 the macroblock of the original source employing standard I-frame coding.
  • at least one modification is used to try and provide consistent picture quality within the resulting I frame.
  • the modification may be, for example, to use the macroblock average quantization parameter of the last frame of the prior group of pictures, as determined in operation 605.
  • the quantization parameter may be a fixed value for all the macroblocks in the frame. The fixed value is determined in the earlier frame-level bit allocation for the current I-frame.
  • the total bit budget is allocated to each frame such that every frame achieves the same coding quality or similar coding quality. For example, in an implementation, the allocated bits of each frame may be determined by assuming every frame will be coded with the same quantization parameter, QP, resulting in approximately the same quality, while consuming all the available bits.
  • a negative offset (denoted by is additionally applied to the I-frame quantization parameter, QP 1 .
  • the ratio of Intra-coded macroblocks to total macroblocks in the previous P-frame is denoted as prevFrmlntraRatio .
  • the following values may be employed for the negative offset: - 9, if prevFrmlntraRatio ⁇ 0.1
  • the value of prevFrmlntraRatio the lower the motion of the previous frame because intra-coding is typically associated with motion. Additionally, the current frame is assumed to have similar motion levels as the previous frame. Recall that low motion areas are more susceptible frame to flickering. Accordingly, larger values of the negative offset AQP 1 are applied for lower motion. As a result, more bits are used to code the I-frame if the motion is lower, thereby improving de-flickering.
  • the value of the quantization parameter actually employed in coding the I-frame QP actual co ⁇ ng is calculated using Equations (2) and (3) as follows. . .
  • Equation (2) it is assumed that for a group of pictures (GOP), only the first frame is an I- frame.
  • N denotes the total number of frames in the GOP
  • R GOP denotes the total bit budget of the GOP.
  • R 1 denotes the number bits of the I-frame
  • R 1 denotes the number of bits of frame i.
  • Equation 2 provides the QP (denoted QP*) that results in the target bit allocation. As can be seen, the same QP is assumed for the initial I-frame and all subsequent frames in the GOP.
  • Equation 3 The value of the quantization parameter actually employed in coding the I-frame is given in Equation 3 :
  • the use of the negative offset value results in a higher weight being assigned to the I-frame relative to conventional bit allocation strategies. As a result, a larger proportion of the bit budget of the GOP is assigned to the I-frame than in conventional bit allocation strategies.
  • the above process results in a desirable quantization parameter to accommodate allocation of a higher than conventional proportion of bits to the I-frame.
  • a different formula may be employed to determine the number of bits to assign to the I-frame. For example, in an implementation, the number of bits allocated to the I-frame may be increased by a percentage, which may be heuristically determined.
  • Encoder 700 receives picture data 705 and provides the data to flicker evaluator 710 for determining whether or not a no-flicker reference is to be generated for a given portion of picture data. This decision can also be characterized as determining whether or not an inter-coding operation, such as a pseudo-P-frame coding, is to be performed for the given portion of picture data.
  • Flicker evaluator 710 may perform, for example, operations 605-620 of Figure 6.
  • Flicker evaluator 710 is coupled to a no-flicker reference unit 720.
  • No-flicker reference unit 720 If a no-flicker reference is to be generated, then flicker evaluator 710 provides an indication of that decision, as well as any other appropriate information, to no-flicker reference unit 720.
  • No-flicker reference unit 720 generates a no-flicker reference for the given portion of picture data by, for example, inter-coding the given portion of picture data to produce coded data, and decoding the coded data to produce a modified portion of picture data.
  • No-flicker reference unit 720 accesses the picture data using input 705.
  • No-flicker reference unit 720 may perform, for example, operations 625-640 of Figure 6.
  • No-flicker reference unit 720 is coupled to an intra-coding unit 730 and provides the no-flicker reference to intra-coding unit 730.
  • Intra-coding unit 730 generates a reference image for the no flicker reference by intra- coding the modified portion of picture data to produce coded data, and decoding the coded data to produce the reference image.
  • Intra-coding unit 730 may perform, for example, I-frame coding, or modified I-frame coding, of the no flicker reference produced in operations 630 and 640 of Figure 6.
  • Intra-coding unit 730 is coupled to an inter-coding unit 740 and provides the reference image to inter-coding unit 740.
  • Inter-coding unit 740 inter-codes a subsequent image using the reference image as a reference.
  • Inter-coding unit 740 may, for example, use a reference image such as I-frame 414 to code subsequent P-frames 416 and 418, as shown in Figure 4.
  • Inter-coding unit 740 accesses the subsequent images from the picture data input 705.
  • Inter-coding unit 740 provides the reconstructions of the inter-coded images, for example, P-frames 416 and 418, as output.
  • FIG. 8 shows a process 800 for producing a reference image.
  • Process 800 includes inter-coding a source image to produce coded source data (810).
  • Process 800 further includes decoding the coded source data to produce a modified source (820).
  • Operations 810 and 820 may be performed, for example, by no-flicker reference unit 720.
  • Process 800 includes intra-coding the modified source to produce coded modified- source data (830).
  • Process 800 further includes decoding the coded modified-source data to produce a reference image (840).
  • Operations 830 and 840 may be performed, for example, by intra-coding unit 730.
  • FIG. 9 shows a video transmission system 900, to which the present principles may be applied, in accordance with an implementation of the present principles.
  • the video transmission system 900 may be, for example, a head-end or transmission system for transmitting a signal using any of a variety of media, such as, for example, satellite, cable, telephone-line, or terrestrial broadcast.
  • the transmission may be provided over the Internet or some other network.
  • the video transmission system 900 is capable of generating and delivering video, content. This is achieved by generating an encoded signal(s) including video information.
  • the video transmission system 900 includes an encoder 910 and a transmitter 920 capable of transmitting the encoded signal.
  • the encoder 910 receives video information and generates an encoded signal(s) therefrom.
  • the encoder 910 may be, for example, the encoder 700 described in detail above.
  • the transmitter 920 may be, for example, adapted to transmit a program signal having one or more bitstreams representing encoded pictures and/or information related thereto.
  • Typical transmitters perform functions such as, for example, one or more of providing error- correction coding, interleaving the data in the signal, randomizing the energy in the signal, and modulating the signal onto one or more carriers.
  • the transmitter may include, or interface with, an antenna (not shown).
  • Figure 10 shows a diagram of an implementation of a video receiving system 1000.
  • the video receiving system 1000 may be configured to receive signals over a variety of media, such as, for example, satellite, cable, telephone-line, or terrestrial broadcast.
  • the signals may be received over the Internet or some other network.
  • the video receiving system 1000 may be, for example, a cell-phone, a computer, a set- top box, a television, or other device that receives encoded video and provides, for example, decoded video for display to a user or for storage.
  • the video receiving system 1000 may provide its output to, for example, a screen of a television, a computer monitor, a computer (for storage, processing, or display), or some other storage, processing, or display device.
  • the video receiving system 1000 is capable of receiving and processing video content. This is achieved by receiving an encoded signal(s) including video information.
  • the video receiving system 1000 includes a receiver 1010 capable of receiving an encoded signal, such as for example the signals described in the implementations of this application, and a decoder 1020 capable of decoding the received signal.
  • the receiver 1010 may be, for example, adapted to receive a program signal having a plurality of bitstreams representing encoded pictures. Typical receivers perform functions such as, for example, one or more of receiving a modulated and encoded data signal, demodulating the data signal from one or more carriers, de-randomizing the energy in the signal, de-interleaving the data in the signal, and error-correction decoding the signal.
  • the receiver 1010 may include, or interface with, an antenna (not shown).
  • the decoder 1020 outputs video signals including video information.
  • the decoder 1020 may be, for example, a decoder operable to decode signals conforming with the ITU-T H.264 standard. Tests of performance of implementations have been conducted, and show that implementations result in elimination or near elimination of the annoying I-frame flickering artifact, and thus, greatly improve the overall perceptual video coding quality. It is also worthwhile mentioning that with I-frame de-flickering, lower PSNR will be yielded. This is because in our scheme a large amount of I-frame coding bits are spent to closely approach a lower quality no-flicker reference, rather than the original frame, which typically compromises the coding efficiency. In experiment, we generally observed a drop in PSNR of greater than 0.3 dB. This is, however, an example of the deficiency of PSNR as an objective video quality metric, because quality was improved.
  • picture includes, without limitation, a frame in a digital video, a field in a digital video, or a limited portion of a frame such as for example a macroblock or a partition of a macroblock. Additionally, the term “no flicker” and other similar expressions that are used throughout this application do not require the complete removal of flicker, but are, instead, intended to mean “reduced flicker”.
  • a picture level implementation may, for example, generate an entire no-flicker reference picture, rather than simply a macroblock, and intercode (I-frame encode, for example) the entire no-flicker reference picture in a pass.
  • This picture level implementation may also decide to generate no-flicker substitutes for the entire picture by, for example, pseudo-P-frame encoding the entire picture rather than selected macroblocks.
  • ⁇ sub-macroblock level implementation may, for example, make decisions regarding whether or not to generate a no-flicker substitute (replacement) on a partition basis.
  • a 16x16 macroblock may be partitioned into 4x4 partitions, and a separate and independent decision may be made for each partition regarding whether or not to do a pseudo-P- frame encoding and to generate a no-flicker substitute for that partition.
  • the implementations described herein may be implemented in, for example, a method or a process, an apparatus, or a software program. Even if only discussed in the context of a single form of implementation (for example, discussed only as a method), the implementation of features discussed may also be implemented in other forms (for example, an apparatus or program).
  • An apparatus may be implemented in, for example, appropriate hardware, software, and firmware.
  • the methods may be implemented in, for example, an apparatus such as, for example, a processor, which refers to processing devices in general, including, for example, a computer, a microprocessor, an integrated circuit, or a programmable logic device. Processors also include communication devices, such as, for example, computers, cell phones, portable/personal digital assistants ("PDAs”), and other devices that facilitate communication of information between end-users.
  • PDAs portable/personal digital assistants
  • Implementations of the various processes and features described herein may be embodied in a variety of different equipment or applications, particularly, for example, equipment or applications associated with data encoding and decoding.
  • equipment include video coders, video decoders, video codecs, web servers, set-top boxes, laptops, personal computers, cell phones, PDAs, and other communication devices.
  • the equipment may be mobile and even installed in a mobile vehicle.
  • the methods may be implemented by instructions being performed by a processor, and such instructions (and/or data values produced by an implementation) may be stored on a processor-readable medium such as, for example, an integrated circuit, a software carrier or other storage device such as, for example, a hard disk, a compact diskette, a random access memory ("RAM"), or a read-only memory (“ROM").
  • the instructions may form an application program tangibly embodied on a processor-readable medium. Instructions may be, for example, in hardware, firmware, software, or a combination. Instructions may be found in, for example, an operating system, a separate application, or a combination of the two.
  • a processor may be characterized, therefore, as, for example, both a device configured to carry out a process and a device that includes a processor-readable medium having instructions for carrying out a process.
  • implementations may produce a variety of signals formatted to carry information that may be, for example, stored or transmitted.
  • the information may include, for example, instructions for performing a method, or data produced by one of the described implementations.
  • a signal may be formatted to carry as data either coded data generated by coding of a no-flicker reference (such as, for example, no- flicker reference 412), or a reconstruction (such as, for example, reconstruction 414) of a no- flicker reference.
  • Such a signal may be formatted, for example, as an electromagnetic wave (for example, using a radio frequency portion of spectrum) or as a baseband signal.
  • the formatting may include, for example, encoding a data stream and modulating a carrier with the encoded data stream.
  • the information that the signal carries may be, for example, analog or digital information.
  • the signal may be transmitted over a variety of different wired or wireless links, as is known.

Landscapes

  • Engineering & Computer Science (AREA)
  • Multimedia (AREA)
  • Signal Processing (AREA)
  • Compression Or Coding Systems Of Tv Signals (AREA)

Abstract

Cette invention se rapporte à diverses mises en œuvre destinées à réduire des artefacts comme, par exemple, un scintillement de trame I. Des mises en œuvre particulières produisent une référence sans scintillement dans laquelle une image destinée à être codée comme une trame I est d'abord codée comme une trame P à l'aide d'une image de référence qui provient du GOP précédent. Ainsi, une continuité avec le GOP précédent est obtenue. Selon un aspect général, une image source est intercodée de manière à produire des données source codées. Les données source codées sont décodées de manière à produire une source modifiée. La source modifiée est intra-codée de manière à produire des données source modifiées codées. Les données source modifiées codées sont décodées de manière à produire une image de référence.
PCT/US2008/013962 2008-01-17 2008-12-22 Scintillement vidéo réduit WO2009091387A1 (fr)

Priority Applications (1)

Application Number Priority Date Filing Date Title
US12/735,342 US20100278236A1 (en) 2008-01-17 2008-12-22 Reduced video flicker

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
US1148508P 2008-01-17 2008-01-17
US61/011,485 2008-01-17

Publications (1)

Publication Number Publication Date
WO2009091387A1 true WO2009091387A1 (fr) 2009-07-23

Family

ID=40457034

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/US2008/013962 WO2009091387A1 (fr) 2008-01-17 2008-12-22 Scintillement vidéo réduit

Country Status (2)

Country Link
US (1) US20100278236A1 (fr)
WO (1) WO2009091387A1 (fr)

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2019017327A1 (fr) * 2017-07-19 2019-01-24 日本電気株式会社 Dispositif de codage d'image animée, procédé de décodage d'image animée, et support d'enregistrement sur lequel est stocké un programme de codage d'image animée

Families Citing this family (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2008187694A (ja) * 2007-01-31 2008-08-14 Matsushita Electric Ind Co Ltd 画像符号化装置およびその方法
US8654844B1 (en) * 2008-02-04 2014-02-18 Zenverge, Inc. Intra frame beating effect reduction
KR101522850B1 (ko) * 2010-01-14 2015-05-26 삼성전자주식회사 움직임 벡터를 부호화, 복호화하는 방법 및 장치
WO2012117744A1 (fr) * 2011-03-03 2012-09-07 Panasonic Corporation Procédé de codage d'une image dans une image codée, procédé de décodage d'une image codée et appareils associés
JP6034010B2 (ja) * 2011-10-24 2016-11-30 ソニー株式会社 符号化装置、符号化方法、およびプログラム
US10735773B2 (en) 2015-06-04 2020-08-04 Apple Inc. Video coding techniques for high quality coding of low motion content

Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20070036213A1 (en) * 2005-08-12 2007-02-15 Atsushi Matsumura Video encoding apparatus and video encoding method
US20070081591A1 (en) * 2005-10-06 2007-04-12 Samsung Electronics Co., Ltd. Method and apparatus for coding moving picture frame to reduce flickering
EP1845735A1 (fr) * 2004-12-28 2007-10-17 NEC Corporation Procede de codage d'images animees, dispositif faisant appel audit procede, et programme informatique

Family Cites Families (33)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US5623311A (en) * 1994-10-28 1997-04-22 Matsushita Electric Corporation Of America MPEG video decoder having a high bandwidth memory
US5682204A (en) * 1995-12-26 1997-10-28 C Cube Microsystems, Inc. Video encoder which uses intra-coding when an activity level of a current macro-block is smaller than a threshold level
US5786866A (en) * 1996-10-15 1998-07-28 Fairchild Semiconductor Corporation Video color subcarrier signal generator
US5870146A (en) * 1997-01-21 1999-02-09 Multilink, Incorporated Device and method for digital video transcoding
US6476779B1 (en) * 1998-03-31 2002-11-05 Sony Corporation Video display device
AU717480B2 (en) * 1998-08-01 2000-03-30 Korea Advanced Institute Of Science And Technology Loop-filtering method for image data and apparatus therefor
US6434197B1 (en) * 1999-01-07 2002-08-13 General Instrument Corporation Multi-functional transcoder for compressed bit streams
US6442206B1 (en) * 1999-01-25 2002-08-27 International Business Machines Corporation Anti-flicker logic for MPEG video decoder with integrated scaling and display functions
US6724825B1 (en) * 2000-09-22 2004-04-20 General Instrument Corporation Regeneration of program clock reference data for MPEG transport streams
US7088780B2 (en) * 2001-05-11 2006-08-08 Mitsubishi Electric Research Labs, Inc. Video transcoder with drift compensation
US6907071B2 (en) * 2001-05-24 2005-06-14 Telefonaktiebolaget Lm Ericsson (Publ) Selective prediction for intra-coding video data block
US20030090498A1 (en) * 2001-11-13 2003-05-15 Photela, Inc. Method and apparatus for the creation of digital photo albums
US7050097B2 (en) * 2001-11-13 2006-05-23 Microsoft Corporation Method and apparatus for the display of still images from image files
US7054367B2 (en) * 2001-12-31 2006-05-30 Emc Corporation Edge detection based on variable-length codes of block coded video
DE10300048B4 (de) * 2002-01-05 2005-05-12 Samsung Electronics Co., Ltd., Suwon Verfahren und Vorrichtung zur Bildcodierung und -decodierung
US7236521B2 (en) * 2002-03-27 2007-06-26 Scientific-Atlanta, Inc. Digital stream transcoder
US6807317B2 (en) * 2002-10-25 2004-10-19 Motorola, Inc. Method and decoder system for reducing quantization effects of a decoded image
JP2004350263A (ja) * 2003-04-28 2004-12-09 Canon Inc 画像処理装置及び画像処理方法
US7352909B2 (en) * 2003-06-02 2008-04-01 Seiko Epson Corporation Weighted overcomplete de-noising
JP4031455B2 (ja) * 2004-03-29 2008-01-09 株式会社東芝 画像符号化装置
US8085846B2 (en) * 2004-08-24 2011-12-27 Thomson Licensing Method and apparatus for decoding hybrid intra-inter coded blocks
US7760805B2 (en) * 2005-05-31 2010-07-20 Hewlett-Packard Development Company, L.P. Method of enhancing images extracted from video
GB0600141D0 (en) * 2006-01-05 2006-02-15 British Broadcasting Corp Scalable coding of video signals
US7956930B2 (en) * 2006-01-06 2011-06-07 Microsoft Corporation Resampling and picture resizing operations for multi-resolution video coding and decoding
US8036270B2 (en) * 2006-07-27 2011-10-11 Sharp Laboratories Of America, Inc. Intra-frame flicker reduction in video coding
US8467448B2 (en) * 2006-11-15 2013-06-18 Motorola Mobility Llc Apparatus and method for fast intra/inter macro-block mode decision for video encoding
MX2009006405A (es) * 2006-12-19 2009-06-23 Koninkl Philips Electronics Nv Metodo y sistema para codificar una señal de imagen, señal de imagen codificada, metodo y sistema para decodificar una señal de imagen.
JP2008252176A (ja) * 2007-03-29 2008-10-16 Toshiba Corp 動画像符号化装置及び方法
AU2007202789B9 (en) * 2007-06-15 2011-08-18 Canon Kabushiki Kaisha High-fidelity motion summarisation method
US8139883B2 (en) * 2008-07-29 2012-03-20 Sony Corporation System and method for image and video encoding artifacts reduction and quality improvement
JP5579731B2 (ja) * 2008-11-12 2014-08-27 トムソン ライセンシング 映像符号化方法、映像符号化装置及びコンピュータプログラム
EP2351374A1 (fr) * 2008-11-25 2011-08-03 Thomson Licensing Procédés et appareil de filtrage d élimination d artefacts parcimonieux pour codage et décodage vidéo
US20100265904A1 (en) * 2009-04-21 2010-10-21 Industrial Technology Research Institute Method, apparatus and computer program product for interference avoidance in uplink coordinated multi-point reception

Patent Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
EP1845735A1 (fr) * 2004-12-28 2007-10-17 NEC Corporation Procede de codage d'images animees, dispositif faisant appel audit procede, et programme informatique
US20070036213A1 (en) * 2005-08-12 2007-02-15 Atsushi Matsumura Video encoding apparatus and video encoding method
US20070081591A1 (en) * 2005-10-06 2007-04-12 Samsung Electronics Co., Ltd. Method and apparatus for coding moving picture frame to reduce flickering

Non-Patent Citations (3)

* Cited by examiner, † Cited by third party
Title
HUA YANG ET AL: "Effective flicker removal from periodic intra frames and accurate flicker measurement", PROCEEDINGS OF 15TH IEEE INTERNATIONAL CONFERENCE ON IMAGE PROCESSING, 12 October 2008 (2008-10-12), pages 2868 - 2871, XP031374640, ISBN: 978-1-4244-1765-0 *
KEIICHI CHONO ET AL: "Detented Quantization to Suppress Flicker Artifacts in Periodically Inserted Intra-Coded Pictures in H.264 Video Coding", PROCEEDINGS OF THE IEEE INTERNATIONAL CONFERENCE ON IMAGE PROCESSING, 1 October 2006 (2006-10-01), pages 1713 - 1716, XP031048986, ISBN: 978-1-4244-0480-3 *
WIEGAND T ET AL: "Rate-constrained coder control and comparison of video coding standards", IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY, vol. 13, no. 7, 1 July 2003 (2003-07-01), pages 688 - 703, XP011099260, ISSN: 1051-8215 *

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2019017327A1 (fr) * 2017-07-19 2019-01-24 日本電気株式会社 Dispositif de codage d'image animée, procédé de décodage d'image animée, et support d'enregistrement sur lequel est stocké un programme de codage d'image animée
US11336889B2 (en) 2017-07-19 2022-05-17 Nec Corporation Moving image encoding device and method for reducing flicker in a moving image

Also Published As

Publication number Publication date
US20100278236A1 (en) 2010-11-04

Similar Documents

Publication Publication Date Title
JP7077385B2 (ja) サンプル適応オフセットの処理を実行するための情報を符号化する符号化方法
US11758194B2 (en) Device and method for video decoding video blocks
US8036270B2 (en) Intra-frame flicker reduction in video coding
TWI399097B (zh) 用於編碼視訊之系統及方法,以及電腦可讀取媒體
Wang et al. Rate-distortion optimization of rate control for H. 264 with adaptive initial quantization parameter determination
US8619856B2 (en) Video coding with large macroblocks
US20070274396A1 (en) Complexity adaptive skip mode estimation for video encoding
US8184702B2 (en) Method for encoding/decoding a video sequence based on hierarchical B-picture using adaptively-adjusted GOP structure
US20100296579A1 (en) Adaptive picture type decision for video coding
US20100086029A1 (en) Video coding with large macroblocks
TWI392372B (zh) Image processing apparatus and image processing method
US20090161759A1 (en) Method and apparatus for video coding on pixel-wise prediction
US20100086031A1 (en) Video coding with large macroblocks
US20090097546A1 (en) System and method for enhanced video communication using real-time scene-change detection for control of moving-picture encoding data rate
US20100278236A1 (en) Reduced video flicker
JP2006519565A (ja) ビデオ符号化
US11212536B2 (en) Negative region-of-interest video coding
US20090080519A1 (en) Method for encoding/decoding video sequence based on mctf using adaptively-adjusted gop structure
JPH08251597A (ja) 動画像符号化/復号化装置
Goudemand et al. A low complexity image quality metric for real-time open-loop transcoding architectures
An et al. High Speed and Low-Complexity Mode Decision for Advanced Video Coding

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 08870686

Country of ref document: EP

Kind code of ref document: A1

WWE Wipo information: entry into national phase

Ref document number: 12735342

Country of ref document: US

NENP Non-entry into the national phase

Ref country code: DE

122 Ep: pct application non-entry in european phase

Ref document number: 08870686

Country of ref document: EP

Kind code of ref document: A1