US20100166069A1 - System and method for video compression using predictive coding - Google Patents

System and method for video compression using predictive coding Download PDF

Info

Publication number
US20100166069A1
US20100166069A1 US12/347,841 US34784108A US2010166069A1 US 20100166069 A1 US20100166069 A1 US 20100166069A1 US 34784108 A US34784108 A US 34784108A US 2010166069 A1 US2010166069 A1 US 2010166069A1
Authority
US
United States
Prior art keywords
pixels
block
macroblock
predictive coding
prediction mode
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
US12/347,841
Other versions
US8311112B2 (en
Inventor
Anurag Goel
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Dynamic Data Technologies LLC
Morgan Stanley Senior Funding Inc
Original Assignee
NXP BV
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Family has litigation
First worldwide family litigation filed litigation Critical https://patents.darts-ip.com/?family=42174218&utm_source=google_patent&utm_medium=platform_link&utm_campaign=public_patent_search&patent=US20100166069(A1) "Global patent litigation dataset” by Darts-ip is licensed under a Creative Commons Attribution 4.0 International License.
Assigned to NXP B.V. reassignment NXP B.V. ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: GOEL, ANURAG
Application filed by NXP BV filed Critical NXP BV
Priority to US12/347,841 priority Critical patent/US8311112B2/en
Priority to PCT/IB2009/055919 priority patent/WO2010076748A1/en
Assigned to TRIDENT MICROSYSTEMS (FAR EAST) LTD. reassignment TRIDENT MICROSYSTEMS (FAR EAST) LTD. ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: NXP HOLDING 1 B.V., TRIDENT MICROSYSTEMS (EUROPE) B.V.
Assigned to NXP HOLDING 1 B.V. reassignment NXP HOLDING 1 B.V. ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: NXP
Publication of US20100166069A1 publication Critical patent/US20100166069A1/en
Assigned to ENTROPIC COMMUNICATIONS, INC. reassignment ENTROPIC COMMUNICATIONS, INC. ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: TRIDENT MICROSYSTEMS (FAR EAST) LTD., TRIDENT MICROSYSTEMS, INC.
Publication of US8311112B2 publication Critical patent/US8311112B2/en
Application granted granted Critical
Assigned to ENTROPIC COMMUNICATIONS, INC. reassignment ENTROPIC COMMUNICATIONS, INC. MERGER AND CHANGE OF NAME (SEE DOCUMENT FOR DETAILS). Assignors: ENTROPIC COMMUNICATIONS, INC., EXCALIBUR ACQUISITION CORPORATION
Assigned to ENTROPIC COMMUNICATIONS, LLC reassignment ENTROPIC COMMUNICATIONS, LLC MERGER AND CHANGE OF NAME (SEE DOCUMENT FOR DETAILS). Assignors: ENTROPIC COMMUNICATIONS, INC., ENTROPIC COMMUNICATIONS, LLC, EXCALIBUR SUBSIDIARY, LLC
Assigned to MORGAN STANLEY SENIOR FUNDING, INC. reassignment MORGAN STANLEY SENIOR FUNDING, INC. SECURITY AGREEMENT SUPPLEMENT Assignors: NXP B.V.
Assigned to MORGAN STANLEY SENIOR FUNDING, INC. reassignment MORGAN STANLEY SENIOR FUNDING, INC. CORRECTIVE ASSIGNMENT TO CORRECT THE REMOVE APPLICATION 12092129 PREVIOUSLY RECORDED ON REEL 038017 FRAME 0058. ASSIGNOR(S) HEREBY CONFIRMS THE SECURITY AGREEMENT SUPPLEMENT. Assignors: NXP B.V.
Assigned to NXP B.V. reassignment NXP B.V. PATENT RELEASE Assignors: MORGAN STANLEY SENIOR FUNDING, INC.
Assigned to MORGAN STANLEY SENIOR FUNDING, INC. reassignment MORGAN STANLEY SENIOR FUNDING, INC. CORRECTIVE ASSIGNMENT TO CORRECT THE REMOVE APPLICATION 12681366 PREVIOUSLY RECORDED ON REEL 038017 FRAME 0058. ASSIGNOR(S) HEREBY CONFIRMS THE SECURITY AGREEMENT SUPPLEMENT. Assignors: NXP B.V.
Assigned to MORGAN STANLEY SENIOR FUNDING, INC. reassignment MORGAN STANLEY SENIOR FUNDING, INC. CORRECTIVE ASSIGNMENT TO CORRECT THE REMOVE APPLICATION 12681366 PREVIOUSLY RECORDED ON REEL 039361 FRAME 0212. ASSIGNOR(S) HEREBY CONFIRMS THE SECURITY AGREEMENT SUPPLEMENT. Assignors: NXP B.V.
Assigned to JPMORGAN CHASE BANK, N.A., AS COLLATERAL AGENT reassignment JPMORGAN CHASE BANK, N.A., AS COLLATERAL AGENT SECURITY AGREEMENT Assignors: ENTROPIC COMMUNICATIONS, LLC (F/K/A ENTROPIC COMMUNICATIONS, INC.), EXAR CORPORATION, MAXLINEAR, INC.
Assigned to DYNAMIC DATA TECHNOLOGIES LLC reassignment DYNAMIC DATA TECHNOLOGIES LLC ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: ENTROPIC COMMUNICATIONS LLC, MAXLINEAR INC.
Assigned to NXP B.V. reassignment NXP B.V. RELEASE BY SECURED PARTY (SEE DOCUMENT FOR DETAILS). Assignors: MORGAN STANLEY SENIOR FUNDING, INC.
Assigned to MORGAN STANLEY SENIOR FUNDING, INC. reassignment MORGAN STANLEY SENIOR FUNDING, INC. CORRECTIVE ASSIGNMENT TO CORRECT THE REMOVE APPLICATION 12298143 PREVIOUSLY RECORDED ON REEL 042985 FRAME 0001. ASSIGNOR(S) HEREBY CONFIRMS THE SECURITY AGREEMENT SUPPLEMENT. Assignors: NXP B.V.
Assigned to MORGAN STANLEY SENIOR FUNDING, INC. reassignment MORGAN STANLEY SENIOR FUNDING, INC. CORRECTIVE ASSIGNMENT TO CORRECT THE REMOVE APPLICATION 12298143 PREVIOUSLY RECORDED ON REEL 039361 FRAME 0212. ASSIGNOR(S) HEREBY CONFIRMS THE SECURITY AGREEMENT SUPPLEMENT. Assignors: NXP B.V.
Assigned to MORGAN STANLEY SENIOR FUNDING, INC. reassignment MORGAN STANLEY SENIOR FUNDING, INC. CORRECTIVE ASSIGNMENT TO CORRECT THE REMOVE APPLICATION 12298143 PREVIOUSLY RECORDED ON REEL 038017 FRAME 0058. ASSIGNOR(S) HEREBY CONFIRMS THE SECURITY AGREEMENT SUPPLEMENT. Assignors: NXP B.V.
Assigned to MORGAN STANLEY SENIOR FUNDING, INC. reassignment MORGAN STANLEY SENIOR FUNDING, INC. CORRECTIVE ASSIGNMENT TO CORRECT THE REMOVE APPLICATION 12298143 PREVIOUSLY RECORDED ON REEL 042762 FRAME 0145. ASSIGNOR(S) HEREBY CONFIRMS THE SECURITY AGREEMENT SUPPLEMENT. Assignors: NXP B.V.
Assigned to MUFG UNION BANK, N.A. reassignment MUFG UNION BANK, N.A. SUCCESSION OF AGENCY (REEL 042453 / FRAME 0001) Assignors: JPMORGAN CHASE BANK, N.A.
Assigned to MAXLINEAR, INC., MAXLINEAR COMMUNICATIONS LLC, EXAR CORPORATION reassignment MAXLINEAR, INC. RELEASE BY SECURED PARTY (SEE DOCUMENT FOR DETAILS). Assignors: MUFG UNION BANK, N.A.
Active legal-status Critical Current
Adjusted expiration legal-status Critical

Links

Images

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/50Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding
    • H04N19/503Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding involving temporal prediction
    • H04N19/507Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding involving temporal prediction using conditional replenishment
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/10Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
    • H04N19/102Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the element, parameter or selection affected or controlled by the adaptive coding
    • H04N19/103Selection of coding mode or of prediction mode
    • H04N19/107Selection of coding mode or of prediction mode between spatial and temporal predictive coding, e.g. picture refresh
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/10Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
    • H04N19/102Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the element, parameter or selection affected or controlled by the adaptive coding
    • H04N19/103Selection of coding mode or of prediction mode
    • H04N19/11Selection of coding mode or of prediction mode among a plurality of spatial predictive coding modes
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/10Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
    • H04N19/134Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the element, parameter or criterion affecting or controlling the adaptive coding
    • H04N19/146Data rate or code amount at the encoder output
    • H04N19/147Data rate or code amount at the encoder output according to rate distortion criteria
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/10Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
    • H04N19/169Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/10Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
    • H04N19/169Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding
    • H04N19/17Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object
    • H04N19/176Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object the region being a block, e.g. a macroblock
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/70Methods or arrangements for coding, decoding, compressing or decompressing digital video signals characterised by syntax aspects related to video coding, e.g. related to compression standards

Definitions

  • Embodiments of the invention relate generally to video coding systems and methods and, more particularly, to a system and method for video compression.
  • Video coding systems that utilize predictive coding to perform video compression, such as video coding systems according to Moving Picture Experts Group (MPEG)-4 and H.264 video compression standards, generate a prediction for a macroblock of a current frame by identifying a region or regions of maximum similarity with the macroblock in previously decoded frames.
  • a region of previous decoded frames that is used for generating the prediction is referred to as a reference region and the pixels of the reference region are referred to as reference pixels.
  • a concern with such video coding systems is the availability of the reference pixels. If new video content appears in the current frame, reference pixels of the new video content may not be available in the previous decoded frames. If the reference pixels of the new video content are not available in the previous decoded frames, the video coding systems search for the reference pixels of the new video content in neighboring frames. However, finding reference pixels in multiple reference frames increases memory requirement, reference frame management load, and motion estimation and motion compensation computation load of the video coding systems. Also, there is usually a practical limit on the number of reference frames that can be used. Additionally, because the reference pixels may not be present in the reference frames, the new video content may not be compressed efficiently. Thus, there is a need for a system and method for video compression that can achieve compression efficiency when new video content appears in the current frame.
  • a system and method for video compression performs predictive coding on a macroblock of a video frame such that a set of pixels of the macroblock is coded using some of the pixels from the same video frame as reference pixels and the rest of the macroblock is coded using reference pixels from at least one other video frame.
  • a method for video compression includes performing predictive coding on a set of pixels of a macroblock of pixels using a first group of reference pixels, where the macroblock of pixels and the first group of reference pixels are from a video frame, and performing predictive coding on the rest of the macroblock of pixels using a second group of reference pixels, where the second group of reference pixels is from at least one other video frame.
  • a system for video compression includes an intra-frame coding unit and an inter-frame coding unit.
  • the intra-frame coding unit is configured to perform predictive coding on a set of pixels of a macroblock of pixels using a first group of reference pixels, where the macroblock of pixels and the first group of reference pixels are from a video frame.
  • the inter-frame coding unit is configured to perform predictive coding on the rest of the macroblock of pixels using a second group of reference pixels, where the second group of reference pixels is from at least one other video frame.
  • a method for video compression includes performing predictive coding on a rectangular sub-block of pixels of a macroblock of pixels using a first group of reference pixels, where the macroblock of pixels and the first group of reference pixels are from a video frame and the first group of reference pixels borders the left edge and the top edge of the rectangular sub-block of pixels, and performing predictive coding on the rest of the macroblock of pixels using a second group of reference pixels, where the second group of reference pixels is from at least one other video frame.
  • FIG. 1 is a schematic block diagram of a system for video compression in accordance with an embodiment of the invention.
  • FIG. 2 depicts four intra-frame prediction modes for a 16 ⁇ 16 block of luminance pixels or an 8 ⁇ 8 block of chrominance pixels according to H.264 video compression standard.
  • FIG. 3 depicts nine intra-frame prediction modes for a 4 ⁇ 4 block of luminance pixels or an 8 ⁇ 8 block of luminance pixels according to H.264 video compression standard.
  • FIG. 4 depicts a 16 ⁇ 8 block of luminance pixels of new video content and corresponding reference pixels of the 16 ⁇ 8 block of luminance pixels in accordance with an embodiment of the invention.
  • FIG. 5 depicts an 8 ⁇ 16 block of luminance pixels of new video content and corresponding reference pixels of the 8 ⁇ 16 block of luminance pixels in accordance with an embodiment of the invention.
  • FIG. 6 depicts an 8 ⁇ 8 block of luminance pixels of new video content and corresponding reference pixels of the 8 ⁇ 8 block of luminance pixels in accordance with an embodiment of the invention.
  • FIG. 7 depicts an 8 ⁇ 4 block of chrominance pixels of new video content and corresponding reference pixels of the 8 ⁇ 4 block of chrominance pixels in accordance with an embodiment of the invention.
  • FIG. 8 depicts a 4 ⁇ 8 block of chrominance pixels of new video content and corresponding reference pixels of the 4 ⁇ 8 block of chrominance pixels in accordance with an embodiment of the invention.
  • FIG. 9 depicts a 4 ⁇ 4 block of chrominance pixels of new video content and corresponding reference pixels of the 4 ⁇ 4 block of chrominance pixels in accordance with an embodiment of the invention.
  • FIG. 10 depicts an exemplary 4 ⁇ 4 array that includes DC coefficients (amplitude of zero frequencies) of 4 ⁇ 4 pixel blocks of new video content under a first scenario in accordance with an embodiment of the invention.
  • FIG. 11 depicts an exemplary 4 ⁇ 4 array that includes DC coefficients of 4 ⁇ 4 pixel blocks of new video content under a second scenario in accordance with an embodiment of the invention.
  • FIG. 12 depicts four different patterns in which a single 8 ⁇ 8 block of luminance pixels can be present in a 16 ⁇ 16 block of luminance pixels of a video frame in accordance with an embodiment of the invention.
  • FIG. 13 depicts an exemplary 4 ⁇ 4 array that includes DC coefficients of 4 ⁇ 4 pixel blocks of new video content under a third scenario in accordance with an embodiment of the invention.
  • FIG. 14 depicts four different patterns in which three 8 ⁇ 8 blocks of luminance pixels can be present in a 16 ⁇ 16 block of luminance pixels of a video frame in accordance with an embodiment of the invention.
  • FIG. 15 depicts an exemplary 4 ⁇ 4 array that includes DC coefficients of 4 ⁇ 4 pixel blocks of new video content under a fourth scenario in accordance with an embodiment of the invention.
  • FIG. 16 depicts an exemplary 4 ⁇ 4 array that includes DC coefficients of 4 ⁇ 4 pixel blocks of new video content under a fifth scenario in accordance with an embodiment of the invention.
  • FIG. 17 depicts an exemplary 4 ⁇ 4 array that includes DC coefficients of 4 ⁇ 4 pixel blocks of new video content under a sixth scenario in accordance with an embodiment of the invention.
  • FIG. 18 depicts four different patterns that three 4 ⁇ 4 blocks of chrominance pixels can be present in an 8 ⁇ 8 block of chrominance pixels of a video frame in accordance with an embodiment of the invention.
  • FIG. 19 is a process flow diagram of a method for video compression in accordance with an embodiment of the invention.
  • FIG. 1 is a schematic block diagram of a system 100 for video compression in accordance with an embodiment of the invention.
  • the system of FIG. 1 can be used to achieve video compression efficiency when new video content appears in a macroblock of a video frame to be compressed.
  • the system can be used to improve H.264 video compression systems and scalable video codec (SVC) systems.
  • SVC scalable video codec
  • the system can also be used for data archiving and medical imaging.
  • the system of FIG. 1 can be implemented in video encoders and decoders, digital televisions, set top boxes and handheld multimedia devices.
  • the system 100 includes a predictive coding module 102 , a transforming module 104 , a quantization module 106 , and an entropy coding module 108 .
  • these modules of the system are shown as being separate components, some or all of these modules may be integrated into fewer components in other embodiments.
  • These modules of the system may be implemented in software, hardware, firmware, or a combination of software, hardware, and firmware.
  • the predictive coding module 102 includes a determining unit 110 , an intra-frame coding unit 112 , a decoding unit 113 , and an inter-frame coding unit 114 .
  • the predictive coding module is configured to perform predictive coding on a macroblock of a video frame.
  • the macroblock may be a rectangular block of pixels.
  • the macroblock may be a rectangular block of pixels of an YCbCr 4:2:0 color format, which includes a 16 ⁇ 16 block of luminance (luma) pixels “Y” and two corresponding 8 ⁇ 8 blocks of chrominance (chroma) pixels “Cb” and “Cr.” Predictive coding may be performed serially on the luminance pixels and the chrominance pixels, for example, first on the luminance pixels and then on the chrominance pixels.
  • the macroblock may have a color format different from the YCbCr 4:2:0 color format such as a RGB (red, green, and blue) color format, which has different chrominance pixel block sizes.
  • the determining unit 110 is configured to determine whether new video content appear in the macroblock of the video frame. Under certain circumstances, pixels of new video content may appear in the macroblock. For example, new objects may be introduced into the video frame and hence into the macroblock. New video content may be introduced into the macroblock because video content hidden behind a moving object may be revealed in the macroblock. New video content may also be introduced into the macroblock because of complex rigid body motion, such as rotation, zoom in, and zoom out, of an object or because of complex non-rigid body motion of an object. New video content may also be introduced into the macroblock because of irregular background texture.
  • the intra-frame coding unit 112 is configured to perform predictive coding on the new video content using reference pixels from the same video frame. That is, the reference pixels are pixels from the same video frame as the macroblock being coded. For example, the intra-frame coding unit may generate a prediction data block for pixels of the new video content using the reference pixels from the video frame and subtract the prediction data block from the pixels of the new video content to produce a residual data block. Details of the intra-frame coding unit are described below with reference to FIGS. 2-9 .
  • the reference pixels for the intra-frame coding unit 112 may be compressed.
  • the decoding unit 113 is configured to decompress and reconstruct the reference pixels.
  • the decoding unit may be configured to perform entropy decoding, inverse quantization, and inverse transformation on the reference pixels. In some embodiments, previously reconstructed reference pixels are input into the decoding unit.
  • the decoding unit is shown as a part of the predictive coding module 102 , the decoding unit may be separate from the predictive coding module in other embodiments.
  • the inter-frame coding unit 114 is configured to perform predictive coding on the rest of the macroblock using reference pixels from other video frames. For example, the inter-frame coding unit may generate a prediction data block for pixels of the rest of the macroblock using a reference group of pixels from previous decoded frames and subtract the prediction data block from the pixels of the rest of the macroblock to produce a residual data block.
  • the transforming module 104 , the quantization module 106 , and the entropy coding module 108 process the residual data from the predictive coding module 102 .
  • the transforming module is configured to transform the residual data to generate transformed residual data.
  • the transforming module may perform a discrete cosine transform and a Hadamard transform on the residual data.
  • the quantization module is configured to quantize the transformed residual data to generate quantized transformed residual data.
  • the entropy coding module is configured to entropy code the quantized transformed residual data.
  • the entropy coding module may process the quantized transformed residual data to generate run-length symbols and then entropy code the run-length symbols.
  • Pixels of new video content of a macroblock may have different pixel block sizes and different video data types.
  • the new video content may include a single block of pixels.
  • the new video content may include a 4 ⁇ 4 block of pixels, an 8 ⁇ 8 block of pixels, a 4 ⁇ 8 block of pixels, an 8 ⁇ 4 block of pixels, a 16 ⁇ 8 block of pixels, an 8 ⁇ 16 block of pixels, or a 16 ⁇ 16 block of pixels.
  • the new video content may include multiple blocks of pixels.
  • the new video content may include three 8 ⁇ 8 blocks of pixels or three 4 ⁇ 4 blocks of pixels.
  • the new video content may include pixels of a single video data type.
  • the new video content may includes luminance pixels or chrominance pixels.
  • the new video content may include pixels of multiple video data types.
  • the new video content may includes luminance pixels and chrominance pixels.
  • the intra-frame coding unit 112 performs predictive coding on the new video content of a macroblock of a video frame using reference pixels from the video frame.
  • the intra-frame coding unit 112 of FIG. 1 may perform predictive coding on the new video content in intra-frame prediction modes according to H.264 video compression standard.
  • H.264 video compression standard Four intra-frame prediction modes for a 16 ⁇ 16 block of luminance pixels or an 8 ⁇ 8 block of chrominance pixels according to H.264 video compression standard are described below with reference to FIG. 2 .
  • Nine intra-frame prediction modes for a 4 ⁇ 4 block of luminance pixels or an 8 ⁇ 8 block of luminance pixels according to H.264 video compression standard are described with reference to FIG. 3 .
  • FIG. 2 depicts four intra-frame prediction modes for a 16 ⁇ 16 block of luminance pixels or an 8 ⁇ 8 block of chrominance pixels according to H.264 video compression standard. Shaded region represents pixels of the new video content and white region labeled with “H” and “V” represents reference pixels. The pixels of the new video content and the corresponding reference pixels are from the same video frame. As shown in FIG. 2 , the four intra-frame prediction modes include a vertical prediction mode, a horizontal prediction mode, a DC prediction mode, and a plane prediction mode.
  • top horizontal (H) reference luminance pixel blocks and left vertical (V) reference luminance pixel blocks are used to perform predictive coding on the 16 ⁇ 16 block of luminance pixels or on the 8 ⁇ 8 block of chrominance pixels.
  • the vertical prediction mode predictive coding is performed on the 16 ⁇ 16 block of luminance pixels or on the 8 ⁇ 8 block of chrominance pixels along the vertical direction using the top reference pixel blocks.
  • the horizontal prediction mode predictive coding is performed on the 16 ⁇ 16 block of luminance pixels or on the 8 ⁇ 8 block of chrominance pixels along the horizontal direction using the left reference pixel blocks.
  • predictive coding is performed on the 16 ⁇ 16 block of luminance pixels or on the 8 ⁇ 8 block of chrominance pixels using a mean value calculated from the top reference pixel blocks and the left reference pixel blocks.
  • predictive coding is performed on the 16 ⁇ 16 block of luminance pixels or on the 8 ⁇ 8 block of chrominance pixels using a linear function of the top reference pixel blocks and the left reference pixel blocks.
  • FIG. 3 depicts nine intra-frame prediction modes for a 4 ⁇ 4 block of luminance pixels or an 8 ⁇ 8 block of luminance pixels according to H.264 video compression standard.
  • Shaded blocks represent pixels of the new video content and white blocks represent reference pixels. The pixels of the new video content and the corresponding reference pixels are from the same video frame.
  • each of the shaded block contains a pixel of the new video content and each of the reference blocks A-M contains a reference luminance pixel.
  • each of the shaded block contains a 2 ⁇ 2 block of pixels of the new video content and each of the reference blocks A-M contains two reference luminance pixels. As shown in FIG.
  • the nine intra-frame prediction modes include a vertical prediction mode, a horizontal prediction mode, a DC prediction mode, a diagonal down-left prediction mode, a diagonal down-right prediction mode, a vertical-right prediction mode, a horizontal-down prediction mode, a vertical-left prediction mode, and a horizontal-up prediction mode.
  • the vertical prediction mode predictive coding is performed on the 4 ⁇ 4 block of luminance pixels or on the 8 ⁇ 8 block of luminance pixels along the vertical direction using top reference blocks A-D.
  • the horizontal prediction mode predictive coding is performed on the 4 ⁇ 4 block of luminance pixels or on the 8 ⁇ 8 block of luminance pixels along the horizontal direction using left reference blocks I-L.
  • predictive coding is performed on the 4 ⁇ 4 block of luminance pixels or on the 8 ⁇ 8 block of luminance pixels using the mean of the top reference blocks A-D and the left reference blocks I-L.
  • predictive coding is performed on the 4 ⁇ 4 block of luminance pixels or on the 8 ⁇ 8 block of luminance pixels along the diagonal down-left direction using the reference blocks A-M.
  • predictive coding is performed on the 4 ⁇ 4 block of luminance pixels or on the 8 ⁇ 8 block of luminance pixels along the diagonal down-right direction using the reference blocks A-M.
  • the vertical-right prediction mode predictive coding is performed on the 4 ⁇ 4 block of luminance pixels or on the 8 ⁇ 8 block of luminance pixels along the vertical-right direction using the reference blocks A-M.
  • the horizontal-down prediction mode predictive coding is performed on the 4 ⁇ 4 block of luminance pixels or on the 8 ⁇ 8 block of luminance pixels along the horizontal-down direction using the reference blocks A-M.
  • the vertical-left prediction mode predictive coding is performed on the 4 ⁇ 4 block of luminance pixels or on the 8 ⁇ 8 block of luminance pixels along the vertical-left direction using the reference blocks A-M.
  • the horizontal-up prediction mode predictive coding is performed on the 4 ⁇ 4 block of luminance pixels or on the 8 ⁇ 8 block of luminance pixels along the horizontal-up direction using the reference blocks A-M.
  • the intra-frame coding unit 112 of FIG. 1 may perform predictive coding on the new video content in intra-frame prediction modes different from the intra-frame prediction modes shown in FIGS. 2-3 .
  • Four exemplary intra-frame prediction modes for a 16 ⁇ 8 block of luminance pixels of FIG. 4 four exemplary intra-frame prediction modes for an 8 ⁇ 16 block of luminance pixels of FIG. 5 , and four exemplary intra-frame prediction modes for an 8 ⁇ 8 block of luminance pixels of FIG. 6 are described below.
  • FIG. 4 depicts a 16 ⁇ 8 block of luminance pixels of new video content and corresponding reference pixels of the 16 ⁇ 8 block of luminance pixels.
  • the 16 ⁇ 8 block of luminance pixels and the corresponding reference pixels are from the same video frame.
  • each shaded box represents a pixel of the new video content and each white box represents a reference pixel.
  • the 16 ⁇ 8 block of luminance pixels includes sixteen columns C 1 -C 16 of pixels, where each column of pixels includes eight pixels.
  • the 16 ⁇ 8 block of luminance pixels includes eight rows R 1 -R 8 of pixels, where each row of pixels includes sixteen pixels.
  • the reference pixels include left vertical reference pixels V 1 -V 8 and top horizontal reference pixels H 0 -H 16 .
  • the four exemplary intra-frame prediction modes used by the intra-frame coding unit 112 for the 16 ⁇ 8 block of luminance pixels of FIG. 4 include a vertical prediction mode, a horizontal prediction mode, a DC prediction mode, and a plane prediction mode.
  • each top reference pixel is used to perform predictive coding on one of the sixteen columns C 1 -C 16 of pixels bordering the top reference pixel along the vertical direction.
  • the top reference pixel H 1 is used to perform predictive coding on the far left column C 1 of pixels.
  • the intra-frame coding unit 112 of FIG. 1 uses the top reference pixel H 1 to generate prediction data for each pixel of the far left column C 1 of pixels and subtracts the prediction data from the corresponding pixel of the far left column C 1 of pixels to produce residual data.
  • the top reference pixel H 16 is used to perform predictive coding on the far right column C 16 of pixels.
  • the intra-frame coding unit uses the top reference pixel H 16 to generate prediction data for each pixel of the far right column C 16 of pixels and subtracts the prediction data from the corresponding pixel of the far right column C 16 of pixels to produce residual data.
  • each left reference pixel is used to perform predictive coding on one of the eight rows R 1 -R 8 of pixels bordering the left reference pixel along the horizontal direction.
  • the left reference pixel V 1 is used to perform predictive coding on the top row R 1 of pixels.
  • the intra-frame coding unit 112 uses the left reference pixel V 1 to generate prediction data for each pixel of the top row R 1 of pixels and subtracts the prediction data from the corresponding pixel of the top row R 1 of pixels to produce residual data.
  • the left reference pixel V 8 is used to perform predictive coding on the bottom row R 8 of pixels.
  • the intra-frame coding unit uses the left reference pixel V 8 , the intra-frame coding unit generates prediction data for each pixel of the bottom row R 8 of pixels and subtracts the prediction data from corresponding pixel of the bottom row R 8 of pixels to produce residual data.
  • the intra-frame coding unit 112 may generate a DC prediction value for each pixel of the 16 ⁇ 8 block of luminance pixels and subtract the DC prediction value from the pixel to produce residual data.
  • the mean value of the top reference pixels H 1 -H 16 , Mean_H can be expressed as:
  • DC_value The DC prediction value of each pixel of the 16 ⁇ 8 block of luminance pixels
  • the intra-frame coding unit 112 may generate a plane prediction value for each pixel of the 16 ⁇ 8 block of luminance pixels and subtract the plane prediction value of the pixel from the pixel to produce residual data.
  • the plane prediction value of the pixel is referred to as P[x, y].
  • the plane prediction value of a pixel at position (x,y), P[x, y] can be expressed as:
  • Clip1 Y ( x ) Clip3(0,(1 ⁇ BitDepth Y ) ⁇ 1 , x )
  • Clip3( x, y, z ) x if z ⁇ x
  • FIG. 5 depicts an 8 ⁇ 16 block of luminance pixels of new video content and corresponding reference pixels of the 8 ⁇ 16 block of luminance pixels.
  • the 8 ⁇ 16 block of luminance pixels and the corresponding reference pixels are from the same video frame.
  • each shaded box represents a pixel of the new video content and each white box represents a reference pixel.
  • the 16 ⁇ 8 block of luminance pixels includes eight columns C 1 -C 8 of pixels, where each column of pixels includes sixteen pixels.
  • the 8 ⁇ 16 block of luminance pixels includes sixteen rows R 1 -R 16 of pixels, where each row of pixels includes eight pixels.
  • the reference pixels include left vertical reference pixels V 1 -V 16 and top horizontal reference pixels H 0 -H 8 .
  • the four exemplary intra-frame prediction modes used by the intra-frame coding unit 112 for the 8 ⁇ 16 block of luminance pixels of FIG. 5 include a vertical prediction mode, a horizontal prediction mode, a DC prediction mode, and a plane prediction mode.
  • each top reference pixel is used to perform predictive coding on one of the eight columns C 1 -C 8 of pixels bordering the top reference pixel along the vertical direction.
  • the top reference pixel H 1 is used to perform predictive coding on the far left column C 1 of pixels.
  • the top reference pixel H 8 is used to perform predictive coding on the far right column C 8 of pixels.
  • each left reference pixel is used to perform predictive coding on one of the sixteen rows R 1 -R 16 of pixels bordering the left reference pixel along the horizontal direction.
  • the left reference pixel V 1 is used to perform predictive coding on the top row R 1 of pixels.
  • the left reference pixel V 16 is used to perform predictive coding on the bottom row R 16 of pixels.
  • the intra-frame coding unit 112 may generate a DC prediction value for each pixel of the 8 ⁇ 16 block of luminance pixels and subtract the DC prediction value from the pixel to produce residual data.
  • the mean value of the top reference pixels H 1 -H 8 , Mean_H can be expressed as:
  • Mean_V The mean value of the left reference pixels V 1 -V 16 , Mean_V, can be expressed as:
  • DC_value The DC prediction value of each pixel of the 8 ⁇ 16 block of luminance pixels
  • the intra-frame coding unit 112 may generate a plane prediction value for each pixel of the 8 ⁇ 16 block of luminance pixels and subtract the plane prediction value of the pixel from the pixel to produce residual data.
  • the plane prediction value of the pixel is referred to as P[x, y].
  • the plane prediction 15 value of a pixel at position (x,y), P[x, y] can be expressed as:
  • FIG. 6 depicts an 8 ⁇ 8 block of luminance pixels of new video content and corresponding reference pixels of the 8 ⁇ 8 block of luminance pixels.
  • the 8 ⁇ 8 block of luminance pixels and the corresponding reference pixels are from the same video frame.
  • each shaded box represents a pixel of the new video content and each white box represents a reference pixel.
  • the 8 ⁇ 8 block of luminance pixels includes eight columns C 1 -C 8 of pixels, where each column of pixels includes eight pixels.
  • the 8 ⁇ 8 block of luminance pixels includes eight rows R 1 -R 8 of pixels, where each row of pixels includes eight pixels.
  • the reference pixels include left vertical reference pixels V 1 -V 8 and top horizontal reference pixels H 0 -H 8 .
  • the four exemplary intra-frame prediction modes used by the intra-frame coding unit 112 for the 8 ⁇ 8 block of luminance pixels of FIG. 6 include a vertical prediction mode, a horizontal prediction mode, a DC prediction mode, and a plane prediction mode.
  • each top reference pixel is used to perform predictive coding on one of the eight columns C 1 -C 8 of pixels bordering the top reference pixel along the vertical direction.
  • the top reference pixel H 1 is used to perform predictive coding on the far left column C 1 of pixels.
  • the intra-frame coding unit 112 uses the top reference pixel H 1 to generate prediction data for each pixel of the far left column C 1 of pixels and subtracts the prediction data from the corresponding pixel of the far left column C 1 of pixels to produce residual data.
  • the top reference pixel H 8 is used to perform predictive coding on the far right column C 8 of pixels.
  • the intra-frame coding unit uses the top reference pixel H 8 to generate prediction data for each pixel of the far right column C 8 of pixels and subtracts the prediction data from the corresponding pixel of the far right column C 8 of pixels to produce residual data.
  • each left reference pixel is used to perform predictive coding on one of the eight rows R 1 -R 8 of pixels bordering the left reference pixel along the horizontal direction.
  • the left reference pixel V 1 is used to perform predictive coding on the top row R 1 of pixels.
  • the intra-frame coding unit 112 uses the left reference pixel V 1 to generate prediction data for each pixel of the top row R 1 of pixels and subtracts the prediction data from the corresponding pixel of the top row R 1 of pixels to produce residual data.
  • the left reference pixel V 8 is used to perform predictive coding on the bottom row R 8 of pixels.
  • the intra-frame coding unit uses the left reference pixel V 8 , the intra-frame coding unit generates prediction data for each pixel of the bottom row R 8 of pixels and subtracts the prediction data from corresponding pixel of the bottom row R 8 of pixels to produce residual data.
  • the intra-frame coding unit 112 may generate a DC prediction value for each pixel of the 8 ⁇ 8 block of luminance pixels and subtract the DC prediction value from the pixel to produce residual data.
  • the intra-frame coding unit 112 may generate a plane prediction value for each pixel of the 8 ⁇ 8 block of luminance pixels according to the plane prediction mode of 8 ⁇ 8 chrominance pixels as specified in H.264 video coding standard and subtract the plane prediction value of the pixel from the pixel to produce residual data.
  • FIG. 7 depicts an 8 ⁇ 4 block of chrominance pixels of new video content and corresponding reference pixels of the 8 ⁇ 4 block of chrominance pixels.
  • the 8 ⁇ 4 block of chrominance pixels and the corresponding reference pixels are from the same video frame.
  • each shaded box represents a pixel of the new video content and each white box represents a reference pixel.
  • the 8 ⁇ 4 block of chrominance pixels includes eight columns C 1 -C 8 of pixels, where each column of pixels includes four pixels.
  • the 8 ⁇ 4 block of chrominance pixels includes four rows R 1 -R 4 of pixels, where each row of pixels includes eight pixels.
  • the reference pixels include left vertical reference pixels V 1 -V 4 and top horizontal reference pixels H 0 -H 8 .
  • the four exemplary intra-frame prediction modes used by the intra-frame coding unit 112 for the 8 ⁇ 4 block of chrominance pixels of FIG. 7 include a vertical prediction mode, a horizontal prediction mode, a DC prediction mode, and a plane prediction mode.
  • each top reference pixel is used to perform predictive coding on one of the eight columns C 1 -C 8 of pixels bordering the top reference pixel along the vertical direction.
  • the top reference pixel H 1 is used to perform predictive coding on the far left column C 1 of pixels.
  • the top reference pixel H 8 is used to perform predictive coding on the far right column C 8 of pixels.
  • the left reference pixel V 4 is used to perform predictive coding on the bottom row R 4 of pixels.
  • predictive coding is performed on the left 4 ⁇ 4 block of 8 ⁇ 4 block of chrominance pixels using a mean value of the top reference pixels H 1 -H 4 and the left reference pixels V 1 -V 4 .
  • Predictive coding on right 4 ⁇ 4 block of 8 ⁇ 4 block is performed using a mean value of the top reference pixels H 4 -H 8 and left reference pixels V 1 -V 4 .
  • the intra-frame coding unit 112 may generate a DC prediction value for each pixel of the 8 ⁇ 4 block of chrominance pixels and subtract the DC prediction value from the pixel to produce residual data.
  • the intra-frame coding unit 112 may generate a plane prediction value for each pixel of the 8 ⁇ 4 block of chrominance pixels and subtract the plane prediction value of the pixel from the pixel to produce residual data.
  • the plane prediction value of the pixel is referred to as P[x, y].
  • the plane prediction value of a pixel at position (x,y), P[x, y] can be expressed as:
  • FIG. 8 depicts a 4 ⁇ 8 block of chrominance pixels of new video content and corresponding reference pixels of the 4 ⁇ 8 block of chrominance pixels.
  • the 4 ⁇ 8 block of chrominance pixels and the corresponding reference pixels are from the same video frame.
  • each shaded box represents a pixel of the new video content and each white box represents a reference pixel.
  • the 4 ⁇ 8 block of chrominance pixels includes four columns C 1 -C 4 of pixels, where each column of pixels includes eight pixels.
  • the 4 ⁇ 8 block of chrominance pixels includes eight rows R 1 -R 8 of pixels, where each row of pixels includes four pixels.
  • the reference pixels include left vertical reference pixels V 1 -V 8 and top horizontal reference pixels H 0 -H 4 .
  • the four exemplary intra-frame prediction modes used by the intra-frame coding unit 112 for the 4 ⁇ 8 block of chrominance pixels of FIG. 8 include a vertical prediction mode, a horizontal prediction mode, a DC prediction mode, and a plane prediction mode.
  • each top reference pixel is used to perform predictive coding on one of the four columns C 1 -C 4 of pixels bordering the top reference pixel along the vertical direction.
  • the top reference pixel H 1 is used to perform predictive coding on the far left column C 1 of pixels.
  • the top reference pixel H 4 is used to perform predictive coding on the far right column C 4 of pixels.
  • each left reference pixel is used to perform predictive coding on one of the eight rows R 1 -R 8 of pixels bordering the left reference pixel along the horizontal direction.
  • the left reference pixel V 1 is used to perform predictive coding on the top row of R 1 pixels.
  • the left reference pixel V 8 is used to perform predictive coding on the bottom row R 8 of pixels.
  • predictive coding is performed on the top 4 ⁇ 4 block of 4 ⁇ 8 block of chrominance pixels using a mean value of the top reference pixels H 1 -H 4 and the left reference pixels V 1 -V 4 .
  • Predictive coding on bottom 4 ⁇ 4 block of 4 ⁇ 8 block is performed using a mean value of the top reference pixels H 1 -H 4 and the left reference pixels V 4 -V 8 .
  • the intra-frame coding unit 112 may generate a DC prediction value for each pixel of the 4 ⁇ 8 block of chrominance pixels and subtract the DC prediction value from the pixel to produce residual data.
  • the intra-frame coding unit 112 may generate a plane prediction value for each pixel of the 4 ⁇ 8 block of chrominance pixels and subtract the plane prediction value of the pixel from the pixel to produce residual data.
  • the plane prediction value of the pixel is referred to as P[x, y].
  • the plane prediction value of a pixel at position (x,y), P[x, y] can be expressed as:
  • FIG. 9 depicts a 4 ⁇ 4 block of chrominance pixels of new video content and corresponding reference pixels of the 4 ⁇ 4 block of chrominance pixels.
  • the 4 ⁇ 4 block of chrominance pixels and the corresponding reference pixels are from the same video frame.
  • each shaded box represents a pixel of the new video content and each white box represents a reference pixel.
  • the 4 ⁇ 4 block of chrominance pixels includes four columns C 1 -C 4 of pixels, where each column of pixels includes four pixels.
  • the 4 ⁇ 4 block of chrominance pixels includes four rows R 1 -R 4 of pixels, where each row of pixels includes four pixels.
  • the reference pixels include left vertical reference pixels V 1 -V 4 and top horizontal reference pixels H 0 -H 4 .
  • the four exemplary intra-frame prediction modes used by the intra-frame coding unit 112 for the 4 ⁇ 4 block of chrominance pixels of FIG. 9 include a vertical prediction mode, a horizontal prediction mode, a DC prediction mode, and a plane prediction mode.
  • each top reference pixel is used to perform predictive coding on one of the four columns C 1 -C 4 of pixels bordering the top reference pixel along the vertical direction.
  • the top reference pixel H 1 is used to perform predictive coding on the far left column C 1 of pixels.
  • the top reference pixel H 4 is used to perform predictive coding on the far right column C 4 of pixels.
  • each left reference pixel is used to perform predictive coding on one of the four rows R 1 -R 4 of pixels bordering the left reference pixel along the horizontal direction.
  • the left reference pixel V 1 is used to perform predictive coding on the top row R 1 of pixels.
  • the left reference pixel V 4 is used to perform predictive coding on the bottom row R 4 of pixels.
  • the intra-frame coding unit 112 may generate a DC prediction value for each pixel of the 4 ⁇ 4 block of chrominance pixels and subtract the DC prediction value from the pixel to produce residual data.
  • the new video content is a part of a 16 ⁇ 16 block of pixels or a part of an 8 ⁇ 8 block of pixels.
  • the 16 ⁇ 16 block of pixels and the 8 ⁇ 8 block of pixels are divided into 4 ⁇ 4 blocks of pixels.
  • the first scenario is that the new video content includes a 16 ⁇ 8 block of luminance pixels.
  • the second scenario is that the new video content includes an 8 ⁇ 16 block of luminance pixels.
  • the third scenario is that the new video content includes a single 8 ⁇ 8 block of luminance pixels.
  • the fourth scenario is that the new video content includes three 8 ⁇ 8 blocks of luminance pixels.
  • the fifth scenario is that the new video content includes an 8 ⁇ 4 block of chrominance pixels.
  • the sixth scenario is that the new video content includes a 4 ⁇ 8 block of chrominance pixels.
  • the seventh scenario is that the new video content includes a 4 ⁇ 4 block of chrominance pixels.
  • the eighth scenario is that the new video content includes three 4 ⁇ 4 blocks of chrominance pixels.
  • the determining unit 110 informs the intra-frame coding unit 112 that the new video content includes an 8 ⁇ 16 block of luminance pixels.
  • the new video content is a part of a 16 ⁇ 16 block of pixels.
  • the intra-frame coding unit performs predictive coding on the 8 ⁇ 16 block of luminance pixels in one of the intra-frame prediction modes shown in FIGS. 2-3 or one of the intra-frame prediction modes described above with respect to FIG. 5 .
  • the inter-frame coding unit 114 performs predictive coding on the rest of the 16 ⁇ 16 block of pixels using reference pixels from other video frames. If the 8 ⁇ 16 block is coded according to one of the prediction modes described above with respect to FIG. 2 or FIG.
  • FIG. 11 depicts an exemplary 4 ⁇ 4 array that includes DC coefficients of 4 ⁇ 4 pixel blocks of new video content under the second scenario.
  • DC 1 -DC 4 and DC 8 -DC 11 are DC coefficients of 4 ⁇ 4 pixel blocks of the new video content, which is located in the left 8 ⁇ 16 block of pixels of the 16 ⁇ 16 block of pixels.
  • the new video content includes a single 8 ⁇ 8 block of luminance pixels.
  • the new video content is a part of a 16 ⁇ 16 block of pixels.
  • FIG. 12 depicts four different patterns in which a single 8 ⁇ 8 block of luminance pixels can be present in a 16 ⁇ 16 block of luminance pixels of a video frame.
  • the blocks I 1 , I 2 , I 3 , and I 4 are 8 ⁇ 8 blocks of intra-frame luminance pixels.
  • the intra-frame coding unit 112 of FIG. 1 performs predictive coding on the 8 ⁇ 8 blocks I 1 -I 4 of luminance pixels using reference pixels from the video frame.
  • the blocks B 1 , B 2 , B 3 , and B 4 are 8 ⁇ 8 blocks of inter-frame luminance pixels.
  • the inter-frame coding unit 114 of FIG. 1 performs predictive coding on the 8 ⁇ 8 blocks B 1 -B 4 of luminance pixels using reference pixels from other video frames. From the top of FIG. 12 to the bottom of FIG. 12 , an 8 ⁇ 8 intra-frame luminance pixel block appears in the lower right corner, the upper left corner, the upper right corner, and the lower left corner of the 16 ⁇ 16 block of luminance pixels respectively.
  • the determining unit 110 informs the intra-frame coding unit that the new video content includes an 8 ⁇ 8 block of luminance pixels. Then, the intra-frame coding unit performs predictive coding on the 8 ⁇ 8 block of luminance pixels in one of the intra-frame prediction modes shown in FIGS. 2-3 or one of the intra-frame prediction modes described above with respect to FIG.
  • the inter-frame coding unit performs predictive coding on the rest of the 16 ⁇ 16 block of pixels using reference pixels from other video frames. If the 8 ⁇ 8 block is coded according to one of the prediction modes described above with respect to FIG. 2 or FIG. 6 then the transforming module 104 collects DC coefficients of 4 ⁇ 4 blocks of luminance pixels of the new video content.
  • FIG. 13 depicts an exemplary 4 ⁇ 4 array that includes DC coefficients 4 ⁇ 4 pixel blocks of new video content under the third scenario. As shown in FIG. 13 , DC 8 -DC 11 are DC coefficients of 4 ⁇ 4 pixel blocks of the new video content, which is located in the lower left 8 ⁇ 8 block of pixels of the 16 ⁇ 16 block of pixels.
  • the new video content includes three 8 ⁇ 8 blocks of luminance pixels.
  • the new video content is a part of a 16 ⁇ 16 block of pixels.
  • FIG. 14 depicts four different patterns in which three 8 ⁇ 8 blocks of luminance pixels can be presented in a 16 ⁇ 16 block of luminance pixels of a video frame.
  • the blocks I 1 , I 2 , I 3 , and I 4 are 8 ⁇ 8 blocks of intra-frame luminance pixels.
  • the intra-frame coding unit 112 of FIG. 1 performs predictive coding on the 8 ⁇ 8 blocks I 1 -I 4 of luminance pixels using reference pixels from the video frame.
  • the blocks B 1 , B 2 , B 3 , and B 4 are 8 ⁇ 8 blocks of inter-frame luminance pixels.
  • the determining unit 110 informs the intra-frame coding unit that the new video content includes three 8 ⁇ 8 blocks of luminance pixels. Then, the intra-frame coding unit performs predictive coding on the three 8 ⁇ 8 blocks of luminance pixels in one of the intra-frame prediction modes shown in FIGS. 2-3 .
  • the inter-frame coding unit performs predictive coding on the rest of the 16 ⁇ 16 block of pixels using reference pixels from other video frames.
  • the transforming module 104 collects DC coefficients of 4 ⁇ 4 blocks of luminance pixels of the new video content.
  • FIG. 15 depicts an exemplary 4 ⁇ 4 array that includes DC coefficients of 4 ⁇ 4 pixel blocks of new video content under the fourth scenario.
  • DC 1 -DC 4 and DC 8 -DC 15 are DC coefficients of 4 ⁇ 4 pixel blocks of the new video content, which is located in the three 8 ⁇ 8 blocks (the upper left, lower left and lower right 8 ⁇ 8 blocks) of pixels of the 16 ⁇ 16 block of pixels.
  • the determining unit 110 informs the intra-frame coding unit 112 that the new video content includes an 8 ⁇ 4 block of chrominance pixels.
  • the new video content is a part of an 8 ⁇ 8 block of pixels.
  • the intra-frame coding unit performs predictive coding on the 8 ⁇ 4 block of chrominance pixels in one of the intra-frame prediction modes shown in FIGS. 2-3 or one of the intra-frame prediction modes described above with respect to FIG. 7 .
  • the inter-frame coding unit 114 performs predictive coding on the rest of the 8 ⁇ 8 block of pixels using reference pixels from other video frames. If the 8 ⁇ 4 block is coded according to one of the prediction modes described above with respect to FIG. 2 or FIG.
  • FIG. 16 depicts an exemplary 4 ⁇ 4 array that includes DC coefficients 4 ⁇ 4 pixel blocks of new video content under the fifth scenario.
  • DC 3 -DC 4 are DC coefficients of 4 ⁇ 4 pixel blocks of the new video content, which is located in the lower 8 ⁇ 4 block of chrominance pixels of the 8 ⁇ 8 block of pixels.
  • the determining unit 110 informs the intra-frame coding unit 112 that the new video content includes a 4 ⁇ 8 block of chrominance pixels.
  • the new video content is a part of an 8 ⁇ 8 block of pixels.
  • the intra-frame coding unit performs predictive coding on the 4 ⁇ 8 block of chrominance pixels in one of the intra-frame prediction modes shown in FIGS. 2-3 or one of the intra-frame prediction modes described above with respect to FIG. 8 .
  • the inter-frame coding unit 114 performs predictive coding on the rest of the 8 ⁇ 8 block of pixels using reference pixels from other video frames. If the 4 ⁇ 8 block is coded according to one of the prediction modes described above with respect to FIG. 2 or FIG.
  • FIG. 17 depicts an exemplary 4 ⁇ 4 array that includes DC coefficients of 4 ⁇ 4 pixel blocks of new video content under the sixth scenario.
  • DC 1 and DC 3 are DC coefficients of 4 ⁇ 4 pixel blocks of the new video content, which is located in the left 8 ⁇ 4 block of chrominance pixels of the 8 ⁇ 8 block of pixels.
  • the transforming module 104 of FIG. 1 performs a Hadamard transform on the collected DC coefficients.
  • the transforming module may perform a four point Hadamard transform along the horizontal direction and then a two point Hadamard transform along the vertical direction.
  • the transforming module may perform a two point Hadamard transform along the horizontal direction and then a four point Hadamard transform along the vertical direction.
  • the transforming module may perform a two point Hadamard transform along the horizontal direction and then a two point Hadamard transform along the vertical direction.
  • the transforming module may perform a four point Hadamard transform along the horizontal direction and then a four point Hadamard transform along the vertical direction.
  • the transforming module may perform a two point Hadamard transform along the horizontal direction.
  • the transforming module may perform a two point Hadamard transform along the vertical direction. Order of applying Hadamard transform can be changed, i.e., vertical transform can be done first and horizontal transform can be done later, and vice versa.
  • the determining unit 110 informs the intra-frame coding unit 112 that the new video content includes a 4 ⁇ 4 block of chrominance pixels.
  • the new video content is a part of an 8 ⁇ 8 block of pixels.
  • the intra-frame coding unit performs predictive coding on the 4 ⁇ 4 block of chrominance pixels in one of the intra-frame prediction modes shown in FIGS. 2-3 or one of the intra-frame prediction modes described above with respect to FIG. 9 .
  • the inter-frame coding unit 114 performs predictive coding on the rest of the pixels using reference pixels from other video frames.
  • the new video content includes three 4 ⁇ 4 blocks of chrominance pixels.
  • the new video content is a part of an 8 ⁇ 8 block of pixels.
  • FIG. 18 depicts four different patterns in which three 4 ⁇ 4 blocks of chrominance pixels can be present in an 8 ⁇ 8 block of chrominance pixels of a video frame.
  • the blocks C 1 , C 2 , C 3 , and C 4 are 4 ⁇ 4 blocks of intra-frame chrominance pixels.
  • the intra-frame coding unit 112 of FIG. 1 performs predictive coding on the 4 ⁇ 4 blocks C 1 -C 4 of chrominance pixels using reference pixels from the same video frame.
  • the blocks B 1 , B 2 , B 3 , and B 4 are 4 ⁇ 4 blocks of inter-frame chrominance pixels.
  • the inter-frame coding unit 114 of FIG. 1 performs predictive coding on the 4 ⁇ 4 blocks B 1 -B 4 of luminance pixels using reference pixels from other video frames. From the top of FIG. 18 to the bottom of FIG. 18 , a 4 ⁇ 4 inter-frame chrominance pixel block is located in the lower right corner, the lower left corner, the upper left corner, and the upper right corner of the 16 ⁇ 16 block of luminance pixels respectively.
  • the determining unit 110 informs the intra-frame coding unit that the new video content includes three 4 ⁇ 4 blocks of chrominance pixels. Then, the intra-frame coding unit performs predictive coding on the three 4 ⁇ 4 blocks of chrominance pixels in one of the intra-frame prediction modes shown in FIGS. 2-3 .
  • the inter-frame coding unit performs predictive coding on the rest of the 8 ⁇ 8 block of pixels using reference pixels from other video frames.
  • a macroblock coded by the intra-frame coding unit 112 of FIG. 1 and the inter-frame coding unit 114 of FIG. 1 is encoded in a bitstream as a macroblock type “Intra_Inter_Macroblock.”
  • the macroblock type “Intra_Inter_Macroblock” is encoded according to encoding methods of such video coding standard for coding macroblock type related syntax elements.
  • macroblock type syntax elements are coded using exponential-Golomb codes or arithmetic codes.
  • the macroblock type “Intra_Inter_Macroblock” and syntax elements related to the macroblock type “Intra_Inter_Macroblock” are encoded in the bitstream at the level of macroblock layer syntax in formatted bitstream syntax.
  • the encoded video bitstream syntax may be hierarchically present.
  • the general structure of the encoded bitstream syntax may be as follows: (1) first, a sequence level syntax is present; (2) then, a group of picture syntax may be present; (3) then, a frame or picture level syntax is present; (4) then, a slice level syntax is present; and (5) after that, a macroblock layer syntax is present. After macroblock layer syntax is present, lower level syntax may be present.
  • An example of macroblock level syntax is described by the following pseudo code excerpt.
  • Intra_Top_Left_Bit // If Intra_Top_Left_Bit is 1, Top or Left block is intra-frame coded block.
  • the macroblock type “Intra_Inter_Macroblock” is present.
  • macroblock partition information “Macroblock_partition_bits” is present.
  • macroblock partition predictive coding type information is present.
  • the system 100 for video compression of FIG. 1 may be configured to compare predictive coding costs of multiple prediction modes for a macroblock to find a prediction mode with a minimum encoding cost.
  • the system for video compression of FIG. 1 may compare predictive coding costs of some or all of the prediction modes described above with respect to FIGS. 2-9 and inter-frame prediction modes according to H.264 video compression standard for a macroblock to find a prediction mode with a minimum encoding cost.
  • the system for video compression of FIG. 1 may compare predictive coding costs of some or all of the prediction modes described above with respect to FIGS. 2-9 and inter-frame prediction modes according to H.264 video compression standard for a macroblock, where the reference pixels for the prediction modes of FIGS.
  • Encoding cost of a prediction mode may include the number of bits that are required to encode the prediction modes, the number of bits that are required to encode residual data of the macroblock under the prediction modes, and the product of a weighting factor and a video quality indicator.
  • the weighting factor may be a result of trade off between compressed video file size and visual quality of compressed video.
  • FIG. 19 is a process flow diagram of a method for video compression in accordance with an embodiment of the invention.
  • predictive coding is performed on a set of pixels of a macroblock of pixels using a first group of reference pixels, where the macroblock of pixels and the first group of reference pixels are from a first video frame.
  • predictive coding is performed on the rest of the macroblock of pixels using a second group of reference pixels, where the second group of reference pixels is from at least one other video frame.

Abstract

A system and method for video compression performs predictive coding on a macroblock of a video frame such that a set of pixels of the macroblock is coded using some of the pixels from the same video frame as reference pixels and the rest of the macroblock is coded using reference pixels from at least one other video frame.

Description

  • Embodiments of the invention relate generally to video coding systems and methods and, more particularly, to a system and method for video compression.
  • Video coding systems that utilize predictive coding to perform video compression, such as video coding systems according to Moving Picture Experts Group (MPEG)-4 and H.264 video compression standards, generate a prediction for a macroblock of a current frame by identifying a region or regions of maximum similarity with the macroblock in previously decoded frames. A region of previous decoded frames that is used for generating the prediction is referred to as a reference region and the pixels of the reference region are referred to as reference pixels.
  • A concern with such video coding systems is the availability of the reference pixels. If new video content appears in the current frame, reference pixels of the new video content may not be available in the previous decoded frames. If the reference pixels of the new video content are not available in the previous decoded frames, the video coding systems search for the reference pixels of the new video content in neighboring frames. However, finding reference pixels in multiple reference frames increases memory requirement, reference frame management load, and motion estimation and motion compensation computation load of the video coding systems. Also, there is usually a practical limit on the number of reference frames that can be used. Additionally, because the reference pixels may not be present in the reference frames, the new video content may not be compressed efficiently. Thus, there is a need for a system and method for video compression that can achieve compression efficiency when new video content appears in the current frame.
  • A system and method for video compression performs predictive coding on a macroblock of a video frame such that a set of pixels of the macroblock is coded using some of the pixels from the same video frame as reference pixels and the rest of the macroblock is coded using reference pixels from at least one other video frame.
  • In an embodiment, a method for video compression includes performing predictive coding on a set of pixels of a macroblock of pixels using a first group of reference pixels, where the macroblock of pixels and the first group of reference pixels are from a video frame, and performing predictive coding on the rest of the macroblock of pixels using a second group of reference pixels, where the second group of reference pixels is from at least one other video frame.
  • In an embodiment, a system for video compression includes an intra-frame coding unit and an inter-frame coding unit. The intra-frame coding unit is configured to perform predictive coding on a set of pixels of a macroblock of pixels using a first group of reference pixels, where the macroblock of pixels and the first group of reference pixels are from a video frame. The inter-frame coding unit is configured to perform predictive coding on the rest of the macroblock of pixels using a second group of reference pixels, where the second group of reference pixels is from at least one other video frame.
  • In an embodiment, a method for video compression includes performing predictive coding on a rectangular sub-block of pixels of a macroblock of pixels using a first group of reference pixels, where the macroblock of pixels and the first group of reference pixels are from a video frame and the first group of reference pixels borders the left edge and the top edge of the rectangular sub-block of pixels, and performing predictive coding on the rest of the macroblock of pixels using a second group of reference pixels, where the second group of reference pixels is from at least one other video frame.
  • Other aspects and advantages of embodiments of the present invention will become apparent from the following detailed description, taken in conjunction with the accompanying drawings, depicted by way of example of the principles of the invention.
  • FIG. 1 is a schematic block diagram of a system for video compression in accordance with an embodiment of the invention.
  • FIG. 2 depicts four intra-frame prediction modes for a 16×16 block of luminance pixels or an 8×8 block of chrominance pixels according to H.264 video compression standard.
  • FIG. 3 depicts nine intra-frame prediction modes for a 4×4 block of luminance pixels or an 8×8 block of luminance pixels according to H.264 video compression standard.
  • FIG. 4 depicts a 16×8 block of luminance pixels of new video content and corresponding reference pixels of the 16×8 block of luminance pixels in accordance with an embodiment of the invention.
  • FIG. 5 depicts an 8×16 block of luminance pixels of new video content and corresponding reference pixels of the 8×16 block of luminance pixels in accordance with an embodiment of the invention.
  • FIG. 6 depicts an 8×8 block of luminance pixels of new video content and corresponding reference pixels of the 8×8 block of luminance pixels in accordance with an embodiment of the invention.
  • FIG. 7 depicts an 8×4 block of chrominance pixels of new video content and corresponding reference pixels of the 8×4 block of chrominance pixels in accordance with an embodiment of the invention.
  • FIG. 8 depicts a 4×8 block of chrominance pixels of new video content and corresponding reference pixels of the 4×8 block of chrominance pixels in accordance with an embodiment of the invention.
  • FIG. 9 depicts a 4×4 block of chrominance pixels of new video content and corresponding reference pixels of the 4×4 block of chrominance pixels in accordance with an embodiment of the invention.
  • FIG. 10 depicts an exemplary 4×4 array that includes DC coefficients (amplitude of zero frequencies) of 4×4 pixel blocks of new video content under a first scenario in accordance with an embodiment of the invention.
  • FIG. 11 depicts an exemplary 4×4 array that includes DC coefficients of 4×4 pixel blocks of new video content under a second scenario in accordance with an embodiment of the invention.
  • FIG. 12 depicts four different patterns in which a single 8×8 block of luminance pixels can be present in a 16×16 block of luminance pixels of a video frame in accordance with an embodiment of the invention.
  • FIG. 13 depicts an exemplary 4×4 array that includes DC coefficients of 4×4 pixel blocks of new video content under a third scenario in accordance with an embodiment of the invention.
  • FIG. 14 depicts four different patterns in which three 8×8 blocks of luminance pixels can be present in a 16×16 block of luminance pixels of a video frame in accordance with an embodiment of the invention.
  • FIG. 15 depicts an exemplary 4×4 array that includes DC coefficients of 4×4 pixel blocks of new video content under a fourth scenario in accordance with an embodiment of the invention.
  • FIG. 16 depicts an exemplary 4×4 array that includes DC coefficients of 4×4 pixel blocks of new video content under a fifth scenario in accordance with an embodiment of the invention.
  • FIG. 17 depicts an exemplary 4×4 array that includes DC coefficients of 4×4 pixel blocks of new video content under a sixth scenario in accordance with an embodiment of the invention.
  • FIG. 18 depicts four different patterns that three 4×4 blocks of chrominance pixels can be present in an 8×8 block of chrominance pixels of a video frame in accordance with an embodiment of the invention.
  • FIG. 19 is a process flow diagram of a method for video compression in accordance with an embodiment of the invention.
  • Throughout the description, similar reference numbers may be used to identify similar elements.
  • FIG. 1 is a schematic block diagram of a system 100 for video compression in accordance with an embodiment of the invention. The system of FIG. 1 can be used to achieve video compression efficiency when new video content appears in a macroblock of a video frame to be compressed. For example, the system can be used to improve H.264 video compression systems and scalable video codec (SVC) systems. The system can also be used for data archiving and medical imaging. The system of FIG. 1 can be implemented in video encoders and decoders, digital televisions, set top boxes and handheld multimedia devices.
  • As shown in FIG. 1, the system 100 includes a predictive coding module 102, a transforming module 104, a quantization module 106, and an entropy coding module 108. Although these modules of the system are shown as being separate components, some or all of these modules may be integrated into fewer components in other embodiments. These modules of the system may be implemented in software, hardware, firmware, or a combination of software, hardware, and firmware.
  • In the embodiment of FIG. 1, the predictive coding module 102 includes a determining unit 110, an intra-frame coding unit 112, a decoding unit 113, and an inter-frame coding unit 114. The predictive coding module is configured to perform predictive coding on a macroblock of a video frame. The macroblock may be a rectangular block of pixels. For example, the macroblock may be a rectangular block of pixels of an YCbCr 4:2:0 color format, which includes a 16×16 block of luminance (luma) pixels “Y” and two corresponding 8×8 blocks of chrominance (chroma) pixels “Cb” and “Cr.” Predictive coding may be performed serially on the luminance pixels and the chrominance pixels, for example, first on the luminance pixels and then on the chrominance pixels. In some embodiments, the macroblock may have a color format different from the YCbCr 4:2:0 color format such as a RGB (red, green, and blue) color format, which has different chrominance pixel block sizes.
  • The determining unit 110 is configured to determine whether new video content appear in the macroblock of the video frame. Under certain circumstances, pixels of new video content may appear in the macroblock. For example, new objects may be introduced into the video frame and hence into the macroblock. New video content may be introduced into the macroblock because video content hidden behind a moving object may be revealed in the macroblock. New video content may also be introduced into the macroblock because of complex rigid body motion, such as rotation, zoom in, and zoom out, of an object or because of complex non-rigid body motion of an object. New video content may also be introduced into the macroblock because of irregular background texture.
  • The intra-frame coding unit 112 is configured to perform predictive coding on the new video content using reference pixels from the same video frame. That is, the reference pixels are pixels from the same video frame as the macroblock being coded. For example, the intra-frame coding unit may generate a prediction data block for pixels of the new video content using the reference pixels from the video frame and subtract the prediction data block from the pixels of the new video content to produce a residual data block. Details of the intra-frame coding unit are described below with reference to FIGS. 2-9. The reference pixels for the intra-frame coding unit 112 may be compressed.
  • The decoding unit 113 is configured to decompress and reconstruct the reference pixels. The decoding unit may be configured to perform entropy decoding, inverse quantization, and inverse transformation on the reference pixels. In some embodiments, previously reconstructed reference pixels are input into the decoding unit. Although the decoding unit is shown as a part of the predictive coding module 102, the decoding unit may be separate from the predictive coding module in other embodiments.
  • The inter-frame coding unit 114 is configured to perform predictive coding on the rest of the macroblock using reference pixels from other video frames. For example, the inter-frame coding unit may generate a prediction data block for pixels of the rest of the macroblock using a reference group of pixels from previous decoded frames and subtract the prediction data block from the pixels of the rest of the macroblock to produce a residual data block.
  • The transforming module 104, the quantization module 106, and the entropy coding module 108 process the residual data from the predictive coding module 102. The transforming module is configured to transform the residual data to generate transformed residual data. For example, the transforming module may perform a discrete cosine transform and a Hadamard transform on the residual data. The quantization module is configured to quantize the transformed residual data to generate quantized transformed residual data. The entropy coding module is configured to entropy code the quantized transformed residual data. For example, the entropy coding module may process the quantized transformed residual data to generate run-length symbols and then entropy code the run-length symbols.
  • Pixels of new video content of a macroblock may have different pixel block sizes and different video data types. The new video content may include a single block of pixels. For example, the new video content may include a 4×4 block of pixels, an 8×8 block of pixels, a 4×8 block of pixels, an 8×4 block of pixels, a 16×8 block of pixels, an 8×16 block of pixels, or a 16×16 block of pixels. The new video content may include multiple blocks of pixels. For example, the new video content may include three 8×8 blocks of pixels or three 4×4 blocks of pixels. The new video content may include pixels of a single video data type. For example, the new video content may includes luminance pixels or chrominance pixels. The new video content may include pixels of multiple video data types. For example, the new video content may includes luminance pixels and chrominance pixels. According to the pixel block size and the video data type of the pixels of new video content, the intra-frame coding unit 112 performs predictive coding on the new video content of a macroblock of a video frame using reference pixels from the video frame.
  • For certain pixel block sizes and video data types of the pixels of the new video content, such as a 16×16 block of luminance pixels, an 8×8 block of chrominance pixels, a 4×4 block of luminance pixels, and an 8×8 block of luminance pixels, the intra-frame coding unit 112 of FIG. 1 may perform predictive coding on the new video content in intra-frame prediction modes according to H.264 video compression standard. Four intra-frame prediction modes for a 16×16 block of luminance pixels or an 8×8 block of chrominance pixels according to H.264 video compression standard are described below with reference to FIG. 2. Nine intra-frame prediction modes for a 4×4 block of luminance pixels or an 8×8 block of luminance pixels according to H.264 video compression standard are described with reference to FIG. 3.
  • FIG. 2 depicts four intra-frame prediction modes for a 16×16 block of luminance pixels or an 8×8 block of chrominance pixels according to H.264 video compression standard. Shaded region represents pixels of the new video content and white region labeled with “H” and “V” represents reference pixels. The pixels of the new video content and the corresponding reference pixels are from the same video frame. As shown in FIG. 2, the four intra-frame prediction modes include a vertical prediction mode, a horizontal prediction mode, a DC prediction mode, and a plane prediction mode. In the four intra-frame prediction modes, top horizontal (H) reference luminance pixel blocks and left vertical (V) reference luminance pixel blocks are used to perform predictive coding on the 16×16 block of luminance pixels or on the 8×8 block of chrominance pixels. In the vertical prediction mode, predictive coding is performed on the 16×16 block of luminance pixels or on the 8×8 block of chrominance pixels along the vertical direction using the top reference pixel blocks. In the horizontal prediction mode, predictive coding is performed on the 16×16 block of luminance pixels or on the 8×8 block of chrominance pixels along the horizontal direction using the left reference pixel blocks. In the DC prediction mode, predictive coding is performed on the 16×16 block of luminance pixels or on the 8×8 block of chrominance pixels using a mean value calculated from the top reference pixel blocks and the left reference pixel blocks. In the plane prediction mode, predictive coding is performed on the 16×16 block of luminance pixels or on the 8×8 block of chrominance pixels using a linear function of the top reference pixel blocks and the left reference pixel blocks.
  • FIG. 3 depicts nine intra-frame prediction modes for a 4×4 block of luminance pixels or an 8×8 block of luminance pixels according to H.264 video compression standard. Shaded blocks represent pixels of the new video content and white blocks represent reference pixels. The pixels of the new video content and the corresponding reference pixels are from the same video frame. In the case of performing predictive coding on the 4×4 block of luminance pixels, each of the shaded block contains a pixel of the new video content and each of the reference blocks A-M contains a reference luminance pixel. In the case of performing predictive coding on the 8×8 block of luminance pixels, each of the shaded block contains a 2×2 block of pixels of the new video content and each of the reference blocks A-M contains two reference luminance pixels. As shown in FIG. 3, the nine intra-frame prediction modes include a vertical prediction mode, a horizontal prediction mode, a DC prediction mode, a diagonal down-left prediction mode, a diagonal down-right prediction mode, a vertical-right prediction mode, a horizontal-down prediction mode, a vertical-left prediction mode, and a horizontal-up prediction mode. In the vertical prediction mode, predictive coding is performed on the 4×4 block of luminance pixels or on the 8×8 block of luminance pixels along the vertical direction using top reference blocks A-D. In the horizontal prediction mode, predictive coding is performed on the 4×4 block of luminance pixels or on the 8×8 block of luminance pixels along the horizontal direction using left reference blocks I-L. In the DC prediction mode, predictive coding is performed on the 4×4 block of luminance pixels or on the 8×8 block of luminance pixels using the mean of the top reference blocks A-D and the left reference blocks I-L. In the diagonal down-left prediction mode, predictive coding is performed on the 4×4 block of luminance pixels or on the 8×8 block of luminance pixels along the diagonal down-left direction using the reference blocks A-M. In the diagonal down-right prediction mode, predictive coding is performed on the 4×4 block of luminance pixels or on the 8×8 block of luminance pixels along the diagonal down-right direction using the reference blocks A-M. In the vertical-right prediction mode, predictive coding is performed on the 4×4 block of luminance pixels or on the 8×8 block of luminance pixels along the vertical-right direction using the reference blocks A-M. In the horizontal-down prediction mode, predictive coding is performed on the 4×4 block of luminance pixels or on the 8×8 block of luminance pixels along the horizontal-down direction using the reference blocks A-M. In the vertical-left prediction mode, predictive coding is performed on the 4×4 block of luminance pixels or on the 8×8 block of luminance pixels along the vertical-left direction using the reference blocks A-M. In the horizontal-up prediction mode, predictive coding is performed on the 4×4 block of luminance pixels or on the 8×8 block of luminance pixels along the horizontal-up direction using the reference blocks A-M.
  • For certain pixel block sizes and video data types of the pixels of the new video content, such as a 16×8 block, 8×16 block, or 8×8 block of luminance pixels, or an 8×4 block, 4×8 block, or 4×4 block of chrominance pixels, the intra-frame coding unit 112 of FIG. 1 may perform predictive coding on the new video content in intra-frame prediction modes different from the intra-frame prediction modes shown in FIGS. 2-3. Four exemplary intra-frame prediction modes for a 16×8 block of luminance pixels of FIG. 4, four exemplary intra-frame prediction modes for an 8×16 block of luminance pixels of FIG. 5, and four exemplary intra-frame prediction modes for an 8×8 block of luminance pixels of FIG. 6 are described below. Four exemplary intra-frame prediction modes for an 8×4 block of chrominance pixels of FIG. 7, four exemplary intra-frame prediction modes for a 4×8 block of chrominance pixels of FIG. 8, and four exemplary intra-frame prediction modes for a 4×4 block of chrominance pixels of FIG. 9 are described below as well.
  • FIG. 4 depicts a 16×8 block of luminance pixels of new video content and corresponding reference pixels of the 16×8 block of luminance pixels. The 16×8 block of luminance pixels and the corresponding reference pixels are from the same video frame. As shown in FIG. 4, each shaded box represents a pixel of the new video content and each white box represents a reference pixel. The 16×8 block of luminance pixels includes sixteen columns C1-C16 of pixels, where each column of pixels includes eight pixels. The 16×8 block of luminance pixels includes eight rows R1-R8 of pixels, where each row of pixels includes sixteen pixels. The reference pixels include left vertical reference pixels V1-V8 and top horizontal reference pixels H0-H16. The four exemplary intra-frame prediction modes used by the intra-frame coding unit 112 for the 16×8 block of luminance pixels of FIG. 4 include a vertical prediction mode, a horizontal prediction mode, a DC prediction mode, and a plane prediction mode.
  • In the vertical prediction mode, predictive coding is performed on the 16×8 block of luminance pixels along the vertical direction using the top reference pixels H1-H16. Each top reference pixel is used to perform predictive coding on one of the sixteen columns C1-C16 of pixels bordering the top reference pixel along the vertical direction. For example, the top reference pixel H1 is used to perform predictive coding on the far left column C1 of pixels. Using the top reference pixel H1, the intra-frame coding unit 112 of FIG. 1 generates prediction data for each pixel of the far left column C1 of pixels and subtracts the prediction data from the corresponding pixel of the far left column C1 of pixels to produce residual data. As another example, the top reference pixel H16 is used to perform predictive coding on the far right column C16 of pixels. Using the top reference pixel H16, the intra-frame coding unit generates prediction data for each pixel of the far right column C16 of pixels and subtracts the prediction data from the corresponding pixel of the far right column C16 of pixels to produce residual data.
  • In the horizontal prediction mode, predictive coding is performed on the 16×8 block of luminance pixels along the horizontal direction using the left reference pixels V1-V8. Each left reference pixel is used to perform predictive coding on one of the eight rows R1-R8 of pixels bordering the left reference pixel along the horizontal direction. For example, the left reference pixel V1 is used to perform predictive coding on the top row R1 of pixels. Using the left reference pixel V1, the intra-frame coding unit 112 generates prediction data for each pixel of the top row R1 of pixels and subtracts the prediction data from the corresponding pixel of the top row R1 of pixels to produce residual data. As another example, the left reference pixel V8 is used to perform predictive coding on the bottom row R8 of pixels. Using the left reference pixel V8, the intra-frame coding unit generates prediction data for each pixel of the bottom row R8 of pixels and subtracts the prediction data from corresponding pixel of the bottom row R8 of pixels to produce residual data.
  • In the DC prediction mode, predictive coding is performed on the 16×8 block of luminance pixels using a mean value of the top reference pixels H1-H16 and a mean value of the left reference pixels V1-V8 according to equations described below. For example, the intra-frame coding unit 112 may generate a DC prediction value for each pixel of the 16×8 block of luminance pixels and subtract the DC prediction value from the pixel to produce residual data. The mean value of the top reference pixels H1-H16, Mean_H, can be expressed as:

  • Mean H=(Sum(H1 to H16)+8)>>4,   (1)
  • where “>>” indicates division by two (2) raised to shift value.
    • The mean value of the left reference pixels V1-V8, Mean_V, can be expressed as:

  • Mean V=(Sum(V1 to V8)+4)>>3.   (2)
  • The DC prediction value of each pixel of the 16×8 block of luminance pixels, DC_value, can be expressed as:

  • DC_value=(Mean V+Mean H+1)>>1.   (3)
  • In the plane prediction mode, predictive coding is performed on the 16×8 block of luminance pixels using a linear function of the top reference pixels H0-H16 and the left reference pixels V1-V8. For example, the intra-frame coding unit 112 may generate a plane prediction value for each pixel of the 16×8 block of luminance pixels and subtract the plane prediction value of the pixel from the pixel to produce residual data. For a pixel at position (x,y), where x is an integer ranging from zero to fifteen and y is an integer ranging from zero to seven, the plane prediction value of the pixel is referred to as P[x, y]. The plane prediction value of a pixel at position (x,y), P[x, y], can be expressed as:

  • P[x,y]=Clip1Y((a+b*(x−7)+c*(y−3)+16)>>5)   (4)
    • a=16*(p[15, −1]+p[−1, 7])
    • b=(5*H+32)>>6
    • c=(17*V+16)>>5
    • H=Sum((x′+1)*(p[8+x′, −1]−p[6−x′, −1])), x′=0 to 7
    • V=Sum((y′+1)*(p[−1, 4+y′]−p[−1, 2−y′])), y′=0 to 3

  • Clip1Y(x)=Clip3(0,(1<<BitDepthY)−1, x)

  • Clip3(x, y, z)=x if z<x
      • y if Z>y
      • z otherwise
        BitDepthY=8, Luminance pixel Y is represented in 8 bits for YCbCr 4:2:0, where each of the top reference pixels H0-H16 is referred to as p [m,−1], where m is an integer ranging from minus one to fifteen, and each of the left reference pixels V1-V8 is referred to as p [−1,n], where n is an integer ranging from zero to seven.
  • FIG. 5 depicts an 8×16 block of luminance pixels of new video content and corresponding reference pixels of the 8×16 block of luminance pixels. The 8×16 block of luminance pixels and the corresponding reference pixels are from the same video frame. As shown in FIG. 5, each shaded box represents a pixel of the new video content and each white box represents a reference pixel. The 16×8 block of luminance pixels includes eight columns C1-C8 of pixels, where each column of pixels includes sixteen pixels. The 8×16 block of luminance pixels includes sixteen rows R1-R16 of pixels, where each row of pixels includes eight pixels. The reference pixels include left vertical reference pixels V1-V16 and top horizontal reference pixels H0-H8. The four exemplary intra-frame prediction modes used by the intra-frame coding unit 112 for the 8×16 block of luminance pixels of FIG. 5 include a vertical prediction mode, a horizontal prediction mode, a DC prediction mode, and a plane prediction mode.
  • In the vertical prediction mode, predictive coding is performed on the 8×16 block of luminance pixels along the vertical direction using the top reference pixels H1-H8. Each top reference pixel is used to perform predictive coding on one of the eight columns C1-C8 of pixels bordering the top reference pixel along the vertical direction. For example, the top reference pixel H1 is used to perform predictive coding on the far left column C1 of pixels. Using the top reference pixel H1, the intra-frame coding unit 112 generates prediction data, where prediction data=top reference pixel H1, for each pixel of the far left column C1 of pixels and subtracts the prediction data from the corresponding pixel of the far left column C1 of pixels to produce residual data. As another example, the top reference pixel H8 is used to perform predictive coding on the far right column C8 of pixels. Using the top reference pixel H8, the intra-frame coding unit generates prediction data, where prediction data=top reference pixel H8, for each pixel of the far right column C8 of pixels and subtracts the prediction data from the corresponding pixel of the far right column C8 of pixels to produce residual data.
  • In the horizontal prediction mode, predictive coding is performed on the 8×16 block of luminance pixels along the horizontal direction using the left reference pixels V1-V16. Each left reference pixel is used to perform predictive coding on one of the sixteen rows R1-R16 of pixels bordering the left reference pixel along the horizontal direction. For example, the left reference pixel V1 is used to perform predictive coding on the top row R1 of pixels. Using the left reference pixel V1, the intra-frame coding unit 112 generates prediction data, where prediction data=left reference pixel V1, for each pixel of the top row R1 of pixels and subtracts the prediction data from the corresponding pixel of the top row R1 of pixels to produce residual data. As another example, the left reference pixel V16 is used to perform predictive coding on the bottom row R16 of pixels. Using the left reference pixel V16, the intra-frame coding unit generates prediction data, where prediction data=left reference pixel V16, for each pixel of the bottom row R16 of pixels and subtracts the prediction data from corresponding pixel of the bottom row R16 of pixels to produce residual data.
  • In the DC prediction mode, predictive coding is performed on the 8×16 block of luminance pixels using a mean value of the top reference pixels H1-H8 and a mean value of the left reference pixels V1-V16 according to equations described below. For example, the intra-frame coding unit 112 may generate a DC prediction value for each pixel of the 8×16 block of luminance pixels and subtract the DC prediction value from the pixel to produce residual data. The mean value of the top reference pixels H1-H8, Mean_H, can be expressed as:

  • Mean H=(Sum(H1 to H8)+4)>>3.   (5)
  • The mean value of the left reference pixels V1-V16, Mean_V, can be expressed as:

  • Mean V=(Sum(V1 to V16)+8)>>4.   (6)
  • The DC prediction value of each pixel of the 8×16 block of luminance pixels, DC_value, can be expressed as:

  • DC_value=(Mean V+Mean H+1)>>1.   (7)
  • In the plane prediction mode, predictive coding is performed on the 8×16 block of luminance pixels using a linear function of the top reference pixels H0-H8 and the left reference pixels V1-V16. For example, the intra-frame coding unit 112 may generate a plane prediction value for each pixel of the 8×16 block of luminance pixels and subtract the plane prediction value of the pixel from the pixel to produce residual data. For a pixel at position (x,y), where x is an integer ranging from zero to seven and y is an integer ranging from zero to fifteen, the plane prediction value of the pixel is referred to as P[x, y]. The plane prediction 15 value of a pixel at position (x,y), P[x, y], can be expressed as:

  • P[x,y]=Clip1Y((a+b*(x−3)+c*(y−7)+16)>>5)   (8)
    • a=16*(p[−1, 15]+p[7, −1])
    • b=(17*H+16)>>5
    • c=(5*V+32)>>6
    • H=Sum((x′+1)*(p[4+x′, −1]−p[2−x′, −1])), x′=0 to 3
    • V=Sum((y′+1)*(p[−1, 8+y′]−p[−1, 6−y′])), y′=0 to 7,
      where each of the top reference pixels H0-H8 is referred to as p [m,−1], where m is an integer ranging from minus one to seven, and each of the left reference pixels V1-V16 is referred to as p [−1,n], where n is an integer ranging from zero to fifteen.
  • FIG. 6 depicts an 8×8 block of luminance pixels of new video content and corresponding reference pixels of the 8×8 block of luminance pixels. The 8×8 block of luminance pixels and the corresponding reference pixels are from the same video frame. As shown in FIG. 6, each shaded box represents a pixel of the new video content and each white box represents a reference pixel. The 8×8 block of luminance pixels includes eight columns C1-C8 of pixels, where each column of pixels includes eight pixels. The 8×8 block of luminance pixels includes eight rows R1-R8 of pixels, where each row of pixels includes eight pixels. The reference pixels include left vertical reference pixels V1-V8 and top horizontal reference pixels H0-H8. The four exemplary intra-frame prediction modes used by the intra-frame coding unit 112 for the 8×8 block of luminance pixels of FIG. 6 include a vertical prediction mode, a horizontal prediction mode, a DC prediction mode, and a plane prediction mode.
  • In the vertical prediction mode, predictive coding is performed on the 8×8 block of luminance pixels along the vertical direction using the top reference pixels H1-H8. Each top reference pixel is used to perform predictive coding on one of the eight columns C1-C8 of pixels bordering the top reference pixel along the vertical direction. For example, the top reference pixel H1 is used to perform predictive coding on the far left column C1 of pixels. Using the top reference pixel H1, the intra-frame coding unit 112 generates prediction data for each pixel of the far left column C1 of pixels and subtracts the prediction data from the corresponding pixel of the far left column C1 of pixels to produce residual data. As another example, the top reference pixel H8 is used to perform predictive coding on the far right column C8 of pixels. Using the top reference pixel H8, the intra-frame coding unit generates prediction data for each pixel of the far right column C8 of pixels and subtracts the prediction data from the corresponding pixel of the far right column C8 of pixels to produce residual data.
  • In the horizontal prediction mode, predictive coding is performed on the 8×8 block of luminance pixels along the horizontal direction using the left reference pixels V1-V8. Each left reference pixel is used to perform predictive coding on one of the eight rows R1-R8 of pixels bordering the left reference pixel along the horizontal direction. For example, the left reference pixel V1 is used to perform predictive coding on the top row R1 of pixels. Using the left reference pixel V1, the intra-frame coding unit 112 generates prediction data for each pixel of the top row R1 of pixels and subtracts the prediction data from the corresponding pixel of the top row R1 of pixels to produce residual data. As another example, the left reference pixel V8 is used to perform predictive coding on the bottom row R8 of pixels. Using the left reference pixel V8, the intra-frame coding unit generates prediction data for each pixel of the bottom row R8 of pixels and subtracts the prediction data from corresponding pixel of the bottom row R8 of pixels to produce residual data.
  • In the DC prediction mode, predictive coding is performed on the 8×8 block of luminance pixels using a mean value of the top reference pixels H1-H8 and the left reference pixels V1-V8. For example, the intra-frame coding unit 112 may generate a DC prediction value for each pixel of the 8×8 block of luminance pixels and subtract the DC prediction value from the pixel to produce residual data.
  • In the plane prediction mode, predictive coding is performed on the 8×8 block of luminance pixels using a linear function of the top reference pixels H0-H8 and the left reference pixels V1-V8. For example, the intra-frame coding unit 112 may generate a plane prediction value for each pixel of the 8×8 block of luminance pixels according to the plane prediction mode of 8×8 chrominance pixels as specified in H.264 video coding standard and subtract the plane prediction value of the pixel from the pixel to produce residual data.
  • FIG. 7 depicts an 8×4 block of chrominance pixels of new video content and corresponding reference pixels of the 8×4 block of chrominance pixels. The 8×4 block of chrominance pixels and the corresponding reference pixels are from the same video frame. As shown in FIG. 7, each shaded box represents a pixel of the new video content and each white box represents a reference pixel. The 8×4 block of chrominance pixels includes eight columns C1-C8 of pixels, where each column of pixels includes four pixels. The 8×4 block of chrominance pixels includes four rows R1-R4 of pixels, where each row of pixels includes eight pixels. The reference pixels include left vertical reference pixels V1-V4 and top horizontal reference pixels H0-H8. The four exemplary intra-frame prediction modes used by the intra-frame coding unit 112 for the 8×4 block of chrominance pixels of FIG. 7 include a vertical prediction mode, a horizontal prediction mode, a DC prediction mode, and a plane prediction mode.
  • In the vertical prediction mode, predictive coding is performed on the 8×4 block of chrominance pixels along the vertical direction using the top reference pixels H1-H8. Each top reference pixel is used to perform predictive coding on one of the eight columns C1-C8 of pixels bordering the top reference pixel along the vertical direction. For example, the top reference pixel H1 is used to perform predictive coding on the far left column C1 of pixels. Using the top reference pixel H1, the intra-frame coding unit 112 generates prediction data, where prediction data=top reference pixel H1, for each pixel of the far left column C1 of pixels and subtracts the prediction data from the corresponding pixel of the far left column C1 of pixels to produce residual data. As another example, the top reference pixel H8 is used to perform predictive coding on the far right column C8 of pixels. Using the top reference pixel H8, the intra-frame coding unit generates prediction data, where prediction data=top reference pixel H8, for each pixel of the far right column C8 of pixels and subtracts the prediction data from the corresponding pixel of the far right column C8 of pixels to produce residual data.
  • In the horizontal prediction mode, predictive coding is performed on the 8×4 block of chrominance pixels along the horizontal direction using the left reference pixels V1-V4. Each left reference pixel is used to perform predictive coding on one of the four rows R1-R4 of pixels bordering the left reference pixel along the horizontal direction. For example, the left reference pixel V1 is used to perform predictive coding on the top row R1 of pixels. Using the left reference pixel V1, the intra-frame coding unit 112 generates prediction data, where prediction data=left reference pixel V1, for each pixel of the top row R1 of pixels and subtracts the prediction data from the corresponding pixel of the top row R1 of pixels to produce residual data. As another example, the left reference pixel V4 is used to perform predictive coding on the bottom row R4 of pixels. Using the left reference pixel V4, the intra-frame coding unit generates prediction data, where prediction data=left reference pixel V4, for each pixel of the bottom row R4 of pixels and subtracts the prediction data from corresponding pixel of the bottom row R4 of pixels to produce residual data.
  • In the DC prediction mode, predictive coding is performed on the left 4×4 block of 8×4 block of chrominance pixels using a mean value of the top reference pixels H1-H4 and the left reference pixels V1-V4. Predictive coding on right 4×4 block of 8×4 block is performed using a mean value of the top reference pixels H4-H8 and left reference pixels V1-V4. For example, the intra-frame coding unit 112 may generate a DC prediction value for each pixel of the 8×4 block of chrominance pixels and subtract the DC prediction value from the pixel to produce residual data.
  • In the plane prediction mode, predictive coding is performed on the 8×4 block of chrominance pixels using a linear function of the top reference pixels H0-H8 and the left reference pixels V1-V4. For example, the intra-frame coding unit 112 may generate a plane prediction value for each pixel of the 8×4 block of chrominance pixels and subtract the plane prediction value of the pixel from the pixel to produce residual data. For a pixel at position (x,y), where x is an integer ranging from zero to seven and y is an integer ranging from zero to three, the plane prediction value of the pixel is referred to as P[x, y]. The plane prediction value of a pixel at position (x,y), P[x, y], can be expressed as:

  • P[x, y]=Clip1C((a+b*(x−3)+c*(y−1)+16)>>5)   (9)
    • a=16*(p[7, −1]+p[−1, 3])
    • b=(17*H+16)>>5
    • c=(54*V+8)>>4
    • H=Sum((x′+1)*(p[4+x′, −1]−p[2−x′, −1])), x′=0 to 3
    • V=Sum((y′+1)*(p[−1, 2+y′]−p[−1, 0−y′])), y′=0 to 1,
      Clip1C is same as Clip1Y, except that BitDepthC is used instead of BitDepthY, where each of the top reference pixels H0-H8 is referred to as p [m,−1], where m is an integer ranging from minus one to seven, and each of the left reference pixels V1-V4 is referred to as p [−1,n], where n is an integer ranging from zero to three.
  • FIG. 8 depicts a 4×8 block of chrominance pixels of new video content and corresponding reference pixels of the 4×8 block of chrominance pixels. The 4×8 block of chrominance pixels and the corresponding reference pixels are from the same video frame. As shown in FIG. 8, each shaded box represents a pixel of the new video content and each white box represents a reference pixel. The 4×8 block of chrominance pixels includes four columns C1-C4 of pixels, where each column of pixels includes eight pixels. The 4×8 block of chrominance pixels includes eight rows R1-R8 of pixels, where each row of pixels includes four pixels. The reference pixels include left vertical reference pixels V1-V8 and top horizontal reference pixels H0-H4. The four exemplary intra-frame prediction modes used by the intra-frame coding unit 112 for the 4×8 block of chrominance pixels of FIG. 8 include a vertical prediction mode, a horizontal prediction mode, a DC prediction mode, and a plane prediction mode.
  • In the vertical prediction mode, predictive coding is performed on the 4×8 block of chrominance pixels along the vertical direction using the top reference pixels H1-H4. Each top reference pixel is used to perform predictive coding on one of the four columns C1-C4 of pixels bordering the top reference pixel along the vertical direction. For example, the top reference pixel H1 is used to perform predictive coding on the far left column C1 of pixels. Using the top reference pixel H1, the intra-frame coding unit 112 generates prediction data, where prediction data=top reference pixel H1, for each pixel of the far left column C1 of pixels and subtracts the prediction data from the corresponding pixel of the far left column C1 of pixels to produce residual data. As another example, the top reference pixel H4 is used to perform predictive coding on the far right column C4 of pixels. Using the top reference pixel H4, the intra-frame coding unit generates prediction data, where prediction data=top reference pixel H4, for each pixel of the far right column C4 of pixels and subtracts the prediction data from the corresponding pixel of the far right column C4 of pixels to produce residual data.
  • In the horizontal prediction mode, predictive coding is performed on the 4×8 block of chrominance pixels along the horizontal direction using the left reference pixels V1-V8. Each left reference pixel is used to perform predictive coding on one of the eight rows R1-R8 of pixels bordering the left reference pixel along the horizontal direction. For example, the left reference pixel V1 is used to perform predictive coding on the top row of R1 pixels. Using the left reference pixel V1, the intra-frame coding unit 112 generates prediction data, where prediction data=left reference pixel V1, for each pixel of the top row R1 of pixels and subtracts the prediction data from the corresponding pixel of the top row R1 of pixels to produce residual data. As another example, the left reference pixel V8 is used to perform predictive coding on the bottom row R8 of pixels. Using the left reference pixel V8, the intra-frame coding unit generates prediction data, where prediction data=left reference pixel V8, for each pixel of the bottom row R8 of pixels and subtracts the prediction data from corresponding pixel of the bottom row R8 of pixels to produce residual data.
  • In the DC prediction mode, predictive coding is performed on the top 4×4 block of 4×8 block of chrominance pixels using a mean value of the top reference pixels H1-H4 and the left reference pixels V1-V4. Predictive coding on bottom 4×4 block of 4×8 block is performed using a mean value of the top reference pixels H1-H4 and the left reference pixels V4-V8. For example, the intra-frame coding unit 112 may generate a DC prediction value for each pixel of the 4×8 block of chrominance pixels and subtract the DC prediction value from the pixel to produce residual data.
  • In the plane prediction mode, predictive coding is performed on the 4×8 block of chrominance pixels using a linear function of the top reference pixels H0-H4 and the left reference pixels V1-V8. For example, the intra-frame coding unit 112 may generate a plane prediction value for each pixel of the 4×8 block of chrominance pixels and subtract the plane prediction value of the pixel from the pixel to produce residual data. For a pixel at position (x,y), where x is an integer ranging from zero to three and y is an integer ranging from zero to seven, the plane prediction value of the pixel is referred to as P[x, y]. The plane prediction value of a pixel at position (x,y), P[x, y], can be expressed as:

  • P[x,y]=Clip1C((a+b*(x−1)+c*(y−3)+16)>>5)   (10)
    • a=16*(p[−1, 7]+p[3, −1])
    • b=(54*H+8)>>4
    • c=(17*V+16)>>5
    • H=Sum((x′+1)*(p[2+x′, −1]−p[0−x′, −1])), x′=0 to 1
    • V=Sum((y′+1)*(p[−1, 4+y′]−p[−1, 2−y′])), y′=0 to 3,
      where each of the top reference pixels H0-H4 is referred to as p [m,−1], where m is an integer ranging from minus one to three, and each of the left reference pixels V1-V8 is referred to as p [−1,n], where n is an integer ranging from zero to seven.
  • FIG. 9 depicts a 4×4 block of chrominance pixels of new video content and corresponding reference pixels of the 4×4 block of chrominance pixels. The 4×4 block of chrominance pixels and the corresponding reference pixels are from the same video frame. As shown in FIG. 9, each shaded box represents a pixel of the new video content and each white box represents a reference pixel. The 4×4 block of chrominance pixels includes four columns C1-C4 of pixels, where each column of pixels includes four pixels. The 4×4 block of chrominance pixels includes four rows R1-R4 of pixels, where each row of pixels includes four pixels. The reference pixels include left vertical reference pixels V1-V4 and top horizontal reference pixels H0-H4. The four exemplary intra-frame prediction modes used by the intra-frame coding unit 112 for the 4×4 block of chrominance pixels of FIG. 9 include a vertical prediction mode, a horizontal prediction mode, a DC prediction mode, and a plane prediction mode.
  • In the vertical prediction mode, predictive coding is performed on the 4×4 block of chrominance pixels along the vertical direction using the top reference pixels H1-H4. Each top reference pixel is used to perform predictive coding on one of the four columns C1-C4 of pixels bordering the top reference pixel along the vertical direction. For example, the top reference pixel H1 is used to perform predictive coding on the far left column C1 of pixels. Using the top reference pixel H1, the intra-frame coding unit 112 generates prediction data, where prediction data=top reference pixel H1, for each pixel of the far left column C1 of pixels and subtracts the prediction data from the corresponding pixel of the far left column C1 of pixels to produce residual data. As another example, the top reference pixel H4 is used to perform predictive coding on the far right column C4 of pixels. Using the top reference pixel H4, the intra-frame coding unit generates prediction data, where prediction data=top reference pixel H4, for each pixel of the far right column C4 of pixels and subtracts the prediction data from the corresponding pixel of the far right column C4 of pixels to produce residual data.
  • In the horizontal prediction mode, predictive coding is performed on the 4×4 block of chrominance pixels along the horizontal direction using the left reference pixels V1-V4. Each left reference pixel is used to perform predictive coding on one of the four rows R1-R4 of pixels bordering the left reference pixel along the horizontal direction. For example, the left reference pixel V1 is used to perform predictive coding on the top row R1 of pixels. Using the top reference pixel V1, the intra-frame coding unit 112 generates prediction data, where prediction data=left reference pixel V1, for each pixel of the top row R1 of pixels and subtracts the prediction data from the corresponding pixel of the top row R1 of pixels to produce residual data. As another example, the left reference pixel V4 is used to perform predictive coding on the bottom row R4 of pixels. Using the left reference pixel V4, the intra-frame coding unit generates prediction data, where prediction data=left reference pixel V4, for each pixel of the bottom row R4 of pixels and subtracts the prediction data from corresponding pixel of the bottom row R4 of pixels to produce residual data.
  • In the DC prediction mode, predictive coding is performed on the 4×4 block of chrominance pixels using a mean value of the top reference pixels H1-H4 and the left reference pixels V1-V4. For example, the intra-frame coding unit 112 may generate a DC prediction value for each pixel of the 4×4 block of chrominance pixels and subtract the DC prediction value from the pixel to produce residual data.
  • In the plane prediction mode, predictive coding is performed on the 4×4 block of chrominance pixels using a linear function of the top reference pixels H0-H4 and the left reference pixels V1-V4. For example, the intra-frame coding unit 112 may generate a plane prediction value for each pixel of the 4×4 block of chrominance pixels and subtract the plane prediction value of the pixel from the pixel to produce residual data. For a pixel at position (x,y), where x is an integer ranging from zero to three and y is an integer ranging from zero to three, the plane prediction value of the pixel is referred to as P[x, y]. The plane prediction value of a pixel at position (x,y), P[x, y], can be expressed as:

  • P[x, y]=Clip1C((a+b*(x−1)+c*(y−1)+16)>>5)   (11)
    • a=16*(p[−1, 3]+p[3, −1])
    • b=(54*H+8)>>4
    • c=(54*V+8)>>4
    • H=Sum((x′+1)*(p[2+x′, −1]−p[0−x′, −1])), x′=0 to 1
    • V=Sum((y′+1)*(p[−1, 2+y′]−p[−1, 0−y′])), y′=0 to 1,
      where each of the top reference pixels H0-H4 is referred to as p [m,−1], where m is an integer ranging from minus one to three, and each of the left reference pixels V1-V4 is referred to as p [−1,n], where n is an integer ranging from zero to three.
  • Examples of performing predictive coding on new video content of different pixel block sizes and different pixel video data types are described below. Eight scenarios of pixel block sizes and pixel video data types of the new video content are considered for the operation of the system 100 of FIG. 1. In the eight scenarios, the new video content is a part of a 16×16 block of pixels or a part of an 8×8 block of pixels. The 16×16 block of pixels and the 8×8 block of pixels are divided into 4×4 blocks of pixels. The first scenario is that the new video content includes a 16×8 block of luminance pixels. The second scenario is that the new video content includes an 8×16 block of luminance pixels. The third scenario is that the new video content includes a single 8×8 block of luminance pixels. The fourth scenario is that the new video content includes three 8×8 blocks of luminance pixels. The fifth scenario is that the new video content includes an 8×4 block of chrominance pixels. The sixth scenario is that the new video content includes a 4×8 block of chrominance pixels. The seventh scenario is that the new video content includes a 4×4 block of chrominance pixels. The eighth scenario is that the new video content includes three 4×4 blocks of chrominance pixels.
  • Under the first scenario, the determining unit 110 informs the intra-frame coding unit 112 that the new video content is a 16×8 block of luminance pixels. The new video content is a part of a 16×16 block of pixels. Then, the intra-frame coding unit performs predictive coding on the 16×8 block of luminance pixels in one of the intra-frame prediction modes shown in FIGS. 2-3 or one of the intra-frame prediction modes described above with respect to FIG. 4. When one of the four intra-frame prediction modes for a 16×16 block of luminance pixels according to H.264 video compression standard is used, reference pixels for the 16×8 block of luminance are pixels bordering the left edge and the top edge of the 16×16 block of luminance pixels. The inter-frame coding unit 114 performs predictive coding on the rest of the 16×16 block of pixels using reference pixels from other video frames. If the 16×8 block is coded according to one of the prediction modes described above with respect to FIG. 2 or FIG. 4 then the transforming module 104 collects DC coefficients (amplitude of zero frequencies) of 4×4 blocks of luminance pixels of the new video content. FIG. 10 depicts an exemplary 4×4 array that includes DC coefficients (DCs) of 4×4 pixel blocks of new video content under the first scenario. As shown in FIG. 10, DC8-DC15 are DC coefficients of 4×4 pixel blocks of the new video content, which is located in the bottom 16×8 block of pixels of the 16×16 block of pixels.
  • Under the second scenario, the determining unit 110 informs the intra-frame coding unit 112 that the new video content includes an 8×16 block of luminance pixels. The new video content is a part of a 16×16 block of pixels. Then, the intra-frame coding unit performs predictive coding on the 8×16 block of luminance pixels in one of the intra-frame prediction modes shown in FIGS. 2-3 or one of the intra-frame prediction modes described above with respect to FIG. 5. The inter-frame coding unit 114 performs predictive coding on the rest of the 16×16 block of pixels using reference pixels from other video frames. If the 8×16 block is coded according to one of the prediction modes described above with respect to FIG. 2 or FIG. 5 then the transforming module 104 collects DC coefficients of 4×4 blocks of luminance pixels of the new video content. FIG. 11 depicts an exemplary 4×4 array that includes DC coefficients of 4×4 pixel blocks of new video content under the second scenario. As shown in FIG. 11, DC1-DC4 and DC8-DC11 are DC coefficients of 4×4 pixel blocks of the new video content, which is located in the left 8×16 block of pixels of the 16×16 block of pixels.
  • Under the third scenario, the new video content includes a single 8×8 block of luminance pixels. The new video content is a part of a 16×16 block of pixels. FIG. 12 depicts four different patterns in which a single 8×8 block of luminance pixels can be present in a 16×16 block of luminance pixels of a video frame. The blocks I1, I2, I3, and I4 are 8×8 blocks of intra-frame luminance pixels. The intra-frame coding unit 112 of FIG. 1 performs predictive coding on the 8×8 blocks I1-I4 of luminance pixels using reference pixels from the video frame. The blocks B1, B2, B3, and B4 are 8×8 blocks of inter-frame luminance pixels. The inter-frame coding unit 114 of FIG. 1 performs predictive coding on the 8×8 blocks B1-B4 of luminance pixels using reference pixels from other video frames. From the top of FIG. 12 to the bottom of FIG. 12, an 8×8 intra-frame luminance pixel block appears in the lower right corner, the upper left corner, the upper right corner, and the lower left corner of the 16×16 block of luminance pixels respectively. The determining unit 110 informs the intra-frame coding unit that the new video content includes an 8×8 block of luminance pixels. Then, the intra-frame coding unit performs predictive coding on the 8×8 block of luminance pixels in one of the intra-frame prediction modes shown in FIGS. 2-3 or one of the intra-frame prediction modes described above with respect to FIG. 6. The inter-frame coding unit performs predictive coding on the rest of the 16×16 block of pixels using reference pixels from other video frames. If the 8×8 block is coded according to one of the prediction modes described above with respect to FIG. 2 or FIG. 6 then the transforming module 104 collects DC coefficients of 4×4 blocks of luminance pixels of the new video content. FIG. 13 depicts an exemplary 4×4 array that includes DC coefficients 4×4 pixel blocks of new video content under the third scenario. As shown in FIG. 13, DC8-DC11 are DC coefficients of 4×4 pixel blocks of the new video content, which is located in the lower left 8×8 block of pixels of the 16×16 block of pixels.
  • Under the fourth scenario, the new video content includes three 8×8 blocks of luminance pixels. The new video content is a part of a 16×16 block of pixels. FIG. 14 depicts four different patterns in which three 8×8 blocks of luminance pixels can be presented in a 16×16 block of luminance pixels of a video frame. The blocks I1, I2, I3, and I4 are 8×8 blocks of intra-frame luminance pixels. The intra-frame coding unit 112 of FIG. 1 performs predictive coding on the 8×8 blocks I1-I4 of luminance pixels using reference pixels from the video frame. The blocks B1, B2, B3, and B4 are 8×8 blocks of inter-frame luminance pixels. The inter-frame coding unit 114 of FIG. 1 performs predictive coding on the 8×8 blocks B1-B4 of luminance pixels using reference pixels from other video frames. From the top of FIG. 14 to the bottom of FIG. 14, an 8×8 inter-frame luminance pixel block is located in the lower right corner, the lower left corner, the upper left corner, and the upper right corner of the 16×16 block of luminance pixels respectively. The determining unit 110 informs the intra-frame coding unit that the new video content includes three 8×8 blocks of luminance pixels. Then, the intra-frame coding unit performs predictive coding on the three 8×8 blocks of luminance pixels in one of the intra-frame prediction modes shown in FIGS. 2-3. The inter-frame coding unit performs predictive coding on the rest of the 16×16 block of pixels using reference pixels from other video frames. The transforming module 104 collects DC coefficients of 4×4 blocks of luminance pixels of the new video content. FIG. 15 depicts an exemplary 4×4 array that includes DC coefficients of 4×4 pixel blocks of new video content under the fourth scenario. As shown in FIG. 15, DC1-DC4 and DC8-DC15 are DC coefficients of 4×4 pixel blocks of the new video content, which is located in the three 8×8 blocks (the upper left, lower left and lower right 8×8 blocks) of pixels of the 16×16 block of pixels.
  • Under the fifth scenario, the determining unit 110 informs the intra-frame coding unit 112 that the new video content includes an 8×4 block of chrominance pixels. The new video content is a part of an 8×8 block of pixels. Then, the intra-frame coding unit performs predictive coding on the 8×4 block of chrominance pixels in one of the intra-frame prediction modes shown in FIGS. 2-3 or one of the intra-frame prediction modes described above with respect to FIG. 7. The inter-frame coding unit 114 performs predictive coding on the rest of the 8×8 block of pixels using reference pixels from other video frames. If the 8×4 block is coded according to one of the prediction modes described above with respect to FIG. 2 or FIG. 7 then the transforming module 104 collects DC coefficients of 4×4 blocks of luminance pixels of the new video content. FIG. 16 depicts an exemplary 4×4 array that includes DC coefficients 4×4 pixel blocks of new video content under the fifth scenario. As shown in FIG. 16, DC3-DC4 are DC coefficients of 4×4 pixel blocks of the new video content, which is located in the lower 8×4 block of chrominance pixels of the 8×8 block of pixels.
  • Under the sixth scenario, the determining unit 110 informs the intra-frame coding unit 112 that the new video content includes a 4×8 block of chrominance pixels. The new video content is a part of an 8×8 block of pixels. Then, the intra-frame coding unit performs predictive coding on the 4×8 block of chrominance pixels in one of the intra-frame prediction modes shown in FIGS. 2-3 or one of the intra-frame prediction modes described above with respect to FIG. 8. The inter-frame coding unit 114 performs predictive coding on the rest of the 8×8 block of pixels using reference pixels from other video frames. If the 4×8 block is coded according to one of the prediction modes described above with respect to FIG. 2 or FIG. 8, then the transforming module 104 collects DC coefficients of 4×4 blocks of luminance pixels of the new video content. FIG. 17 depicts an exemplary 4×4 array that includes DC coefficients of 4×4 pixel blocks of new video content under the sixth scenario. As shown in FIG. 17, DC1 and DC3 are DC coefficients of 4×4 pixel blocks of the new video content, which is located in the left 8×4 block of chrominance pixels of the 8×8 block of pixels.
  • Under the first, second, third, fourth, fifth and sixth scenarios, the transforming module 104 of FIG. 1 performs a Hadamard transform on the collected DC coefficients. For example, under the first scenario, the transforming module may perform a four point Hadamard transform along the horizontal direction and then a two point Hadamard transform along the vertical direction. Under the second scenario, the transforming module may perform a two point Hadamard transform along the horizontal direction and then a four point Hadamard transform along the vertical direction. Under the third scenario, the transforming module may perform a two point Hadamard transform along the horizontal direction and then a two point Hadamard transform along the vertical direction. Under the fourth scenario, the transforming module may perform a four point Hadamard transform along the horizontal direction and then a four point Hadamard transform along the vertical direction. Under the fifth scenario, the transforming module may perform a two point Hadamard transform along the horizontal direction. Under the sixth scenario, the transforming module may perform a two point Hadamard transform along the vertical direction. Order of applying Hadamard transform can be changed, i.e., vertical transform can be done first and horizontal transform can be done later, and vice versa.
  • Under the seventh scenario, the determining unit 110 informs the intra-frame coding unit 112 that the new video content includes a 4×4 block of chrominance pixels. The new video content is a part of an 8×8 block of pixels. Then, the intra-frame coding unit performs predictive coding on the 4×4 block of chrominance pixels in one of the intra-frame prediction modes shown in FIGS. 2-3 or one of the intra-frame prediction modes described above with respect to FIG. 9. The inter-frame coding unit 114 performs predictive coding on the rest of the pixels using reference pixels from other video frames.
  • Under the eighth scenario, the new video content includes three 4×4 blocks of chrominance pixels. The new video content is a part of an 8×8 block of pixels. FIG. 18 depicts four different patterns in which three 4×4 blocks of chrominance pixels can be present in an 8×8 block of chrominance pixels of a video frame. The blocks C1, C2, C3, and C4 are 4×4 blocks of intra-frame chrominance pixels. The intra-frame coding unit 112 of FIG. 1 performs predictive coding on the 4×4 blocks C1-C4 of chrominance pixels using reference pixels from the same video frame. The blocks B1, B2, B3, and B4 are 4×4 blocks of inter-frame chrominance pixels. The inter-frame coding unit 114 of FIG. 1 performs predictive coding on the 4×4 blocks B1-B4 of luminance pixels using reference pixels from other video frames. From the top of FIG. 18 to the bottom of FIG. 18, a 4×4 inter-frame chrominance pixel block is located in the lower right corner, the lower left corner, the upper left corner, and the upper right corner of the 16×16 block of luminance pixels respectively. Under the eighth scenario, the determining unit 110 informs the intra-frame coding unit that the new video content includes three 4×4 blocks of chrominance pixels. Then, the intra-frame coding unit performs predictive coding on the three 4×4 blocks of chrominance pixels in one of the intra-frame prediction modes shown in FIGS. 2-3. The inter-frame coding unit performs predictive coding on the rest of the 8×8 block of pixels using reference pixels from other video frames.
  • In an embodiment, a macroblock coded by the intra-frame coding unit 112 of FIG. 1 and the inter-frame coding unit 114 of FIG. 1 is encoded in a bitstream as a macroblock type “Intra_Inter_Macroblock.” When the system 100 for video compression of FIG. 1 is used in video coding systems according to a video coding standard, the macroblock type “Intra_Inter_Macroblock” is encoded according to encoding methods of such video coding standard for coding macroblock type related syntax elements. For example, when the system for video compression of FIG. 1 is used in video coding systems according to H.264 video compression standard, macroblock type syntax elements are coded using exponential-Golomb codes or arithmetic codes. The macroblock type “Intra_Inter_Macroblock” and syntax elements related to the macroblock type “Intra_Inter_Macroblock” are encoded in the bitstream at the level of macroblock layer syntax in formatted bitstream syntax. The encoded video bitstream syntax may be hierarchically present. As example, the general structure of the encoded bitstream syntax may be as follows: (1) first, a sequence level syntax is present; (2) then, a group of picture syntax may be present; (3) then, a frame or picture level syntax is present; (4) then, a slice level syntax is present; and (5) after that, a macroblock layer syntax is present. After macroblock layer syntax is present, lower level syntax may be present. An example of macroblock level syntax is described by the following pseudo code excerpt.
  • If (mb_type = = Intra_Inter_Macroblock)
    {
      If (macroblock_partition_type = = 16x8) Place Top_16x8_Inter_Block_bit.
    // When Top_16x8_Block_bit is 1, the top 16x8 block is inter-frame coded. When
    Top_16x8_Block_bit is not 1, the top 16x8 block is intra-frame coded.
      If (macroblock_partition_type = = 8x16) Place Left_8x16_Inter_Block_bit.
    // When Left_8x16_Inter_Block_bit is 1, the left 16x8 block is inter-frame coded.
    When Left_8x16_Inter_Block_bit is not 1, the left 16x8 block is intra-frame
    coded.
      If (macroblock_partition_type = = 8x8) Place Inter_8x8_block_bit's
    // When Inter_8x8_block_bit is 1, the 8x8 block is inter-frame coded. When
    Inter_8x8_block_bit is not 1, the 8x8 block is intra-framed coded.
      If (macroblock_partition type = = Inter_8x8_Intra_16x8_8x16)
      {
        Place Intra_16x8_8x16 bit
        // If Intra_16x8_8x16 bit is 1, mode Intra_16x8 is used. If
    Intra_16x8_8x16 bit is not 1, mode Intra_8x16 is used.
        Place Intra_Top_Left_Bit
        // If Intra_Top_Left_Bit is 1, Top or Left block is intra-frame coded
    block.
      }
    }

    The macroblock type “Intra_Inter_Macroblock” is present. Then macroblock partition information “Macroblock_partition_bits” is present. Then macroblock partition predictive coding type information is present.
  • The system 100 for video compression of FIG. 1 may be configured to compare predictive coding costs of multiple prediction modes for a macroblock to find a prediction mode with a minimum encoding cost. For example, the system for video compression of FIG. 1 may compare predictive coding costs of some or all of the prediction modes described above with respect to FIGS. 2-9 and inter-frame prediction modes according to H.264 video compression standard for a macroblock to find a prediction mode with a minimum encoding cost. Alternatively, for computation efficiency, the system for video compression of FIG. 1 may compare predictive coding costs of some or all of the prediction modes described above with respect to FIGS. 2-9 and inter-frame prediction modes according to H.264 video compression standard for a macroblock, where the reference pixels for the prediction modes of FIGS. 2-9 are the reference pixels bordering the top and left edge of the macroblock, to find a prediction mode with a minimum encoding cost. Encoding cost of a prediction mode may include the number of bits that are required to encode the prediction modes, the number of bits that are required to encode residual data of the macroblock under the prediction modes, and the product of a weighting factor and a video quality indicator. The weighting factor may be a result of trade off between compressed video file size and visual quality of compressed video. When an “Intra_Inter_Macroblock” type macroblock is used, the system for video compression of FIG. 1 may encode and place macroblock syntax information with residual data of the macroblock in the bit stream. For example, the system for video compression of FIG. 1 may first encode and place syntax information in the encoded bit stream then place the encoded residual data of the macroblock. As a result, video decoders can decode the “Intra_Inter_Macroblock” type macroblock using information from the encoded bit stream.
  • FIG. 19 is a process flow diagram of a method for video compression in accordance with an embodiment of the invention. At block 1902, predictive coding is performed on a set of pixels of a macroblock of pixels using a first group of reference pixels, where the macroblock of pixels and the first group of reference pixels are from a first video frame. At block 1904, predictive coding is performed on the rest of the macroblock of pixels using a second group of reference pixels, where the second group of reference pixels is from at least one other video frame.
  • Although the operations of the method herein are shown and described in a particular order, the order of the operations of the method may be altered so that certain operations may be performed in an inverse order or so that certain operations may be performed, at least in part, concurrently with other operations. In another embodiment, instructions or sub-operations of distinct operations may be implemented in an intermittent and/or alternating manner.
  • Although specific embodiments of the invention that have been described or depicted include several components described or depicted herein, other embodiments of the invention may include fewer or more components to implement less or more functionality.
  • Although specific embodiments of the invention have been described and depicted, the invention is not to be limited to the specific forms or arrangements of parts so described and depicted. The scope of the invention is to be defined by the claims appended hereto and their equivalents.

Claims (20)

1. A method for video compression, the method comprising:
performing predictive coding on a set of pixels of a macroblock of pixels using a first group of reference pixels, the macroblock of pixels and the first group of reference pixels being from a video frame; and
performing predictive coding on the rest of the macroblock of pixels using a second group of reference pixels, the second group of reference pixels being from at least one other video frame.
2. The method of claim 1 further comprising determining that new video content appears in the macroblock of pixels before performing the predictive coding on the macroblock of pixels, wherein the new video content appears in the set of pixels.
3. The method of claim 1 further comprising comparing predictive coding costs of a plurality of prediction modes, reference pixels for the predictive coding being top and left edge reference pixels bordering the macroblock or bordering a sub-block of the macroblock, before performing the predictive coding on the macroblock of pixels.
4. The method of claim 1, wherein the performing the predictive coding on the set of pixels includes performing the predictive coding on the set of pixels in at least one of a sub-block horizontal prediction mode, a sub-block vertical prediction mode, a sub-block DC prediction mode, a sub-block plane prediction mode, a horizontal prediction mode, a vertical prediction mode, a DC prediction mode, a plane prediction mode and directional prediction modes.
5. The method of claim 1, wherein the performing the predictive coding on the set of pixels of the macroblock of pixels using the first group of reference pixels includes generating a prediction data block for the set of pixels using the first group of reference pixels and subtracting the prediction data block from the set of pixels to produce a residual data block.
6. The method of claim 5 further comprising performing a Hadamard transform on at least a part of the transformed residual data block.
7. The method of claim 1, wherein the macroblock of pixels is a rectangular block of pixels.
8. The method of claim 7, wherein the first group of reference pixels borders the left edge and the top edge of the macroblock of pixels.
9. The method of claim 1, wherein the set of pixels is a rectangular block of pixels.
10. The method of claim 9, wherein the first group of reference pixels borders the left edge and the top edge of the set of pixels.
11. A system for video compression, the system comprising:
an intra-frame coding unit configured to perform predictive coding on a set of pixels of a macroblock of pixels using a first group of reference pixels, the macroblock of pixels and the first group of reference pixels being from a video frame; and
an inter-frame coding unit configured to perform predictive coding on the rest of the macroblock of pixels using a second group of reference pixels, the second group of reference pixels being from at least one other video frame.
12. The system of claim 11 further comprising a determining unit configured to determine that new video content appears in the macroblock of pixels before performing predictive coding on the macroblock of pixels, wherein the new video content appears in the set of pixels.
13. The system of claim 11 is configured to compare predictive coding costs of a plurality of prediction modes, reference pixels for the predictive coding being top and left edge reference pixels bordering the macroblock or bordering a sub-block of the macroblock, before performing the predictive coding on the macroblock of pixels.
14. The system of claim 11, wherein the intra-frame coding unit is configured to perform the predictive coding on the set of pixels in at least one of a sub-block horizontal prediction mode, a sub-block vertical prediction mode, a sub-block DC prediction mode, a sub-block plane prediction mode, a horizontal prediction mode, a vertical prediction mode, a DC prediction mode, a plane prediction mode and directional prediction modes.
15. The system of claim 11, wherein the intra-frame coding unit is configured to generate a prediction data block for the set of pixels using the first group of reference pixels and to subtract the prediction data block from the set of pixels to produce a residual data block.
16. The system of claim 15 further comprising a transforming module configured to perform a Hadamard transform on at least a part of the transformed residual data block.
17. A method for video compression, the method comprising:
performing predictive coding on a rectangular sub-block of pixels of a macroblock of pixels using a first group of reference pixels, the macroblock of pixels and the first group of reference pixels being from a video frame, the first group of reference pixels bordering the left edge and the top edge of the rectangular sub-block of pixels; and
performing predictive coding on the rest of the macroblock of pixels using a second group of reference pixels, the second group of reference pixels being from at least one other video frame.
18. The method of claim 17 further comprising determining that new video content appears in the macroblock of pixels before performing predictive coding on the macroblock of pixels, wherein the new video content appears in the rectangular sub-block of pixels.
19. The method of claim 17 further comprising comparing predictive coding costs of a plurality of prediction modes, reference pixels for the predictive coding being top and left edge reference pixels bordering the macroblock or bordering a sub-block of the macroblock, before performing the predictive coding on the macroblock of pixels.
20. The method of claim 17, wherein the performing the predictive coding on the rectangular sub-block of pixels includes performing predictive coding on the rectangular sub-block of pixels in at least one of a sub-block horizontal prediction mode, a sub-block vertical prediction mode, a sub-block DC prediction mode, a sub-block plane prediction mode, a horizontal prediction mode, a vertical prediction mode, a DC prediction mode, a plane prediction mode and directional prediction modes.
US12/347,841 2008-12-31 2008-12-31 System and method for video compression using predictive coding Active 2031-04-27 US8311112B2 (en)

Priority Applications (2)

Application Number Priority Date Filing Date Title
US12/347,841 US8311112B2 (en) 2008-12-31 2008-12-31 System and method for video compression using predictive coding
PCT/IB2009/055919 WO2010076748A1 (en) 2008-12-31 2009-12-23 System and method for video compression using predictive coding

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
US12/347,841 US8311112B2 (en) 2008-12-31 2008-12-31 System and method for video compression using predictive coding

Publications (2)

Publication Number Publication Date
US20100166069A1 true US20100166069A1 (en) 2010-07-01
US8311112B2 US8311112B2 (en) 2012-11-13

Family

ID=42174218

Family Applications (1)

Application Number Title Priority Date Filing Date
US12/347,841 Active 2031-04-27 US8311112B2 (en) 2008-12-31 2008-12-31 System and method for video compression using predictive coding

Country Status (2)

Country Link
US (1) US8311112B2 (en)
WO (1) WO2010076748A1 (en)

Cited By (11)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20120321207A1 (en) * 2010-03-04 2012-12-20 Megachips Corporation Image coding apparatus
US20130114714A1 (en) * 2010-07-20 2013-05-09 Kazushi Sato Image processing device and image processing method
US8787454B1 (en) * 2011-07-13 2014-07-22 Google Inc. Method and apparatus for data compression using content-based features
US20140233634A1 (en) * 2011-09-14 2014-08-21 Samsung Electronics Co., Ltd. Method and device for encoding and decoding video
US20150016515A1 (en) * 2013-07-12 2015-01-15 British Broadcasting Corporation Video encoding and decoding with prediction at higher precision
RU2597484C2 (en) * 2010-07-16 2016-09-10 Самсунг Электроникс Ко., Лтд. Method and apparatus for encoding and decoding image using intra-frame prediction
US9641337B2 (en) * 2014-04-28 2017-05-02 Nxp B.V. Interface compatible approach for gluing white-box implementation to surrounding program
US10003793B2 (en) 2012-10-01 2018-06-19 Google Technology Holdings LLC Processing of pulse code modulation (PCM) parameters
US10462494B2 (en) * 2009-02-23 2019-10-29 Korea Advanced Institute Of Science And Technology Video encoding method for encoding division block, video decoding method for decoding division block, and recording medium for implementing the same
US10750200B2 (en) * 2016-12-29 2020-08-18 Peking University Shenzhen Graduate School Encoding method, decoding method, encoder, and decoder
US11039138B1 (en) * 2012-03-08 2021-06-15 Google Llc Adaptive coding of prediction modes using probability distributions

Families Citing this family (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP4833309B2 (en) * 2009-03-06 2011-12-07 株式会社東芝 Video compression encoding device
CN104380166B (en) * 2012-07-12 2016-06-08 奥林巴斯株式会社 Camera head

Citations (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20020191698A1 (en) * 2001-06-15 2002-12-19 Solidstreaming, Inc. Video data CODEC system with low computational complexity
US20060171680A1 (en) * 2005-02-02 2006-08-03 Jun Makino Image processing apparatus and method

Family Cites Families (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP4163618B2 (en) * 2001-08-28 2008-10-08 株式会社エヌ・ティ・ティ・ドコモ Video encoding / transmission system, video encoding / transmission method, encoding apparatus, decoding apparatus, encoding method, decoding method, and program suitable for use in the same
US7466774B2 (en) * 2003-01-07 2008-12-16 Thomson Licensing Mixed inter/intra video coding of macroblock partitions
KR101088375B1 (en) * 2005-07-21 2011-12-01 삼성전자주식회사 apparatus and method for adaptive variable block-size transform and the appartus and the method for video encoding/decoding thereof.

Patent Citations (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20020191698A1 (en) * 2001-06-15 2002-12-19 Solidstreaming, Inc. Video data CODEC system with low computational complexity
US20060171680A1 (en) * 2005-02-02 2006-08-03 Jun Makino Image processing apparatus and method

Cited By (31)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US11659210B2 (en) 2009-02-23 2023-05-23 Korea Advanced Institute Of Science And Technology Video encoding method for encoding division block, video decoding method for decoding division block, and recording medium for implementing the same
US11076175B2 (en) 2009-02-23 2021-07-27 Korea Advanced Institute Of Science And Technology Video encoding method for encoding division block, video decoding method for decoding division block, and recording medium for implementing the same
US10462494B2 (en) * 2009-02-23 2019-10-29 Korea Advanced Institute Of Science And Technology Video encoding method for encoding division block, video decoding method for decoding division block, and recording medium for implementing the same
US20120321207A1 (en) * 2010-03-04 2012-12-20 Megachips Corporation Image coding apparatus
US8731312B2 (en) * 2010-03-04 2014-05-20 Megachips Corporation Image coding apparatus
RU2621008C2 (en) * 2010-07-16 2017-05-30 Самсунг Электроникс Ко., Лтд. Image coding and decoding method and device with intraframe prediction
US10212452B2 (en) 2010-07-16 2019-02-19 Samsung Electronics Co., Ltd. Method and apparatus for encoding and decoding image through intra prediction
US9661344B2 (en) 2010-07-16 2017-05-23 Samsung Electronics Co., Ltd. Method and apparatus for encoding and decoding image through intra prediction
US9654795B2 (en) 2010-07-16 2017-05-16 Samsung Electronics Co., Ltd. Method and apparatus for encoding and decoding image through intra prediction
US9654797B2 (en) 2010-07-16 2017-05-16 Samsung Electronics Co., Ltd. Method and apparatus for encoding and decoding image through intra prediction
RU2597484C2 (en) * 2010-07-16 2016-09-10 Самсунг Электроникс Ко., Лтд. Method and apparatus for encoding and decoding image using intra-frame prediction
US9654796B2 (en) 2010-07-16 2017-05-16 Samsung Electronics Co., Ltd. Method and apparatus for encoding and decoding image through intra prediction
RU2607233C2 (en) * 2010-07-16 2017-01-10 Самсунг Электроникс Ко., Лтд. Method and apparatus for encoding and decoding image using intra-frame prediction
US20130114714A1 (en) * 2010-07-20 2013-05-09 Kazushi Sato Image processing device and image processing method
US9282330B1 (en) 2011-07-13 2016-03-08 Google Inc. Method and apparatus for data compression using content-based features
US8787454B1 (en) * 2011-07-13 2014-07-22 Google Inc. Method and apparatus for data compression using content-based features
US20140233634A1 (en) * 2011-09-14 2014-08-21 Samsung Electronics Co., Ltd. Method and device for encoding and decoding video
US9538188B2 (en) * 2011-09-14 2017-01-03 Samsung Electronics Co., Ltd. Method and device for encoding and decoding video
US20150172656A1 (en) * 2011-09-14 2015-06-18 Samsung Electronics Co., Ltd. Method and device for encoding and decoding video
US20150172699A1 (en) * 2011-09-14 2015-06-18 Samsung Electronics Co., Ltd. Method and device for encoding and decoding video
US9578332B2 (en) * 2011-09-14 2017-02-21 Samsung Electronics Co., Ltd. Method and device for encoding and decoding video
US9538187B2 (en) * 2011-09-14 2017-01-03 Samsung Electronics Co., Ltd. Method and device for encoding and decoding video
US20150172696A1 (en) * 2011-09-14 2015-06-18 Samsung Electronics Co., Ltd. Method and device for encoding and decoding video
US9538184B2 (en) * 2011-09-14 2017-01-03 Samsung Electronics Co., Ltd. Method and device for encoding and decoding video
US11039138B1 (en) * 2012-03-08 2021-06-15 Google Llc Adaptive coding of prediction modes using probability distributions
US11627321B2 (en) 2012-03-08 2023-04-11 Google Llc Adaptive coding of prediction modes using probability distributions
US10003793B2 (en) 2012-10-01 2018-06-19 Google Technology Holdings LLC Processing of pulse code modulation (PCM) parameters
US9877025B2 (en) * 2013-07-12 2018-01-23 British Broadcasting Corporation Video encoding and decoding with prediction at higher precision
US20150016515A1 (en) * 2013-07-12 2015-01-15 British Broadcasting Corporation Video encoding and decoding with prediction at higher precision
US9641337B2 (en) * 2014-04-28 2017-05-02 Nxp B.V. Interface compatible approach for gluing white-box implementation to surrounding program
US10750200B2 (en) * 2016-12-29 2020-08-18 Peking University Shenzhen Graduate School Encoding method, decoding method, encoder, and decoder

Also Published As

Publication number Publication date
WO2010076748A1 (en) 2010-07-08
US8311112B2 (en) 2012-11-13

Similar Documents

Publication Publication Date Title
US8311112B2 (en) System and method for video compression using predictive coding
USRE49565E1 (en) Apparatus for encoding an image
RU2400941C1 (en) Device for coding of dynamic images, device for decoding of dynamic images, method for coding of dynamic images and method for decoding of dynamic images
CN107105241B (en) Image encoding device
JP5982612B2 (en) Intra prediction mode decoding method
KR101246294B1 (en) Method of and apparatus for video intraprediction encoding/decoding
US10887587B2 (en) Distance weighted bi-directional intra prediction
US8902980B2 (en) Apparatus and method for encoding and decoding high fidelity video, and computer-readable storage medium
US9083947B2 (en) Video encoder, video decoder, method for video encoding and method for video decoding, separately for each colour plane
US11102474B2 (en) Devices and methods for intra prediction video coding based on a plurality of reference pixel values
US8542938B2 (en) System and method for intra-frame compression using predictive coding
US20230276058A1 (en) Image encoder and decoder using unidirectional prediction
CN110741642B (en) Apparatus and method for directional intra prediction using fitted planes and reference samples

Legal Events

Date Code Title Description
AS Assignment

Owner name: NXP B.V.,NETHERLANDS

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:GOEL, ANURAG;REEL/FRAME:022047/0410

Effective date: 20081231

Owner name: NXP B.V., NETHERLANDS

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:GOEL, ANURAG;REEL/FRAME:022047/0410

Effective date: 20081231

AS Assignment

Owner name: NXP HOLDING 1 B.V.,NETHERLANDS

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:NXP;REEL/FRAME:023928/0489

Effective date: 20100207

Owner name: TRIDENT MICROSYSTEMS (FAR EAST) LTD.,CAYMAN ISLAND

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:TRIDENT MICROSYSTEMS (EUROPE) B.V.;NXP HOLDING 1 B.V.;REEL/FRAME:023928/0552

Effective date: 20100208

Owner name: NXP HOLDING 1 B.V., NETHERLANDS

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:NXP;REEL/FRAME:023928/0489

Effective date: 20100207

Owner name: TRIDENT MICROSYSTEMS (FAR EAST) LTD., CAYMAN ISLAN

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:TRIDENT MICROSYSTEMS (EUROPE) B.V.;NXP HOLDING 1 B.V.;REEL/FRAME:023928/0552

Effective date: 20100208

AS Assignment

Owner name: ENTROPIC COMMUNICATIONS, INC., CALIFORNIA

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:TRIDENT MICROSYSTEMS, INC.;TRIDENT MICROSYSTEMS (FAR EAST) LTD.;REEL/FRAME:028153/0440

Effective date: 20120411

STCF Information on status: patent grant

Free format text: PATENTED CASE

AS Assignment

Owner name: ENTROPIC COMMUNICATIONS, INC., CALIFORNIA

Free format text: MERGER AND CHANGE OF NAME;ASSIGNORS:EXCALIBUR ACQUISITION CORPORATION;ENTROPIC COMMUNICATIONS, INC.;ENTROPIC COMMUNICATIONS, INC.;REEL/FRAME:035706/0267

Effective date: 20150430

AS Assignment

Owner name: ENTROPIC COMMUNICATIONS, LLC, CALIFORNIA

Free format text: MERGER AND CHANGE OF NAME;ASSIGNORS:ENTROPIC COMMUNICATIONS, INC.;EXCALIBUR SUBSIDIARY, LLC;ENTROPIC COMMUNICATIONS, LLC;REEL/FRAME:035717/0628

Effective date: 20150430

AS Assignment

Owner name: MORGAN STANLEY SENIOR FUNDING, INC., MARYLAND

Free format text: SECURITY AGREEMENT SUPPLEMENT;ASSIGNOR:NXP B.V.;REEL/FRAME:038017/0058

Effective date: 20160218

FPAY Fee payment

Year of fee payment: 4

AS Assignment

Owner name: MORGAN STANLEY SENIOR FUNDING, INC., MARYLAND

Free format text: CORRECTIVE ASSIGNMENT TO CORRECT THE REMOVE APPLICATION 12092129 PREVIOUSLY RECORDED ON REEL 038017 FRAME 0058. ASSIGNOR(S) HEREBY CONFIRMS THE SECURITY AGREEMENT SUPPLEMENT;ASSIGNOR:NXP B.V.;REEL/FRAME:039361/0212

Effective date: 20160218

AS Assignment

Owner name: NXP B.V., NETHERLANDS

Free format text: PATENT RELEASE;ASSIGNOR:MORGAN STANLEY SENIOR FUNDING, INC.;REEL/FRAME:039707/0471

Effective date: 20160805

AS Assignment

Owner name: MORGAN STANLEY SENIOR FUNDING, INC., MARYLAND

Free format text: CORRECTIVE ASSIGNMENT TO CORRECT THE REMOVE APPLICATION 12681366 PREVIOUSLY RECORDED ON REEL 039361 FRAME 0212. ASSIGNOR(S) HEREBY CONFIRMS THE SECURITY AGREEMENT SUPPLEMENT;ASSIGNOR:NXP B.V.;REEL/FRAME:042762/0145

Effective date: 20160218

Owner name: MORGAN STANLEY SENIOR FUNDING, INC., MARYLAND

Free format text: CORRECTIVE ASSIGNMENT TO CORRECT THE REMOVE APPLICATION 12681366 PREVIOUSLY RECORDED ON REEL 038017 FRAME 0058. ASSIGNOR(S) HEREBY CONFIRMS THE SECURITY AGREEMENT SUPPLEMENT;ASSIGNOR:NXP B.V.;REEL/FRAME:042985/0001

Effective date: 20160218

AS Assignment

Owner name: JPMORGAN CHASE BANK, N.A., AS COLLATERAL AGENT, IL

Free format text: SECURITY AGREEMENT;ASSIGNORS:MAXLINEAR, INC.;ENTROPIC COMMUNICATIONS, LLC (F/K/A ENTROPIC COMMUNICATIONS, INC.);EXAR CORPORATION;REEL/FRAME:042453/0001

Effective date: 20170512

Owner name: JPMORGAN CHASE BANK, N.A., AS COLLATERAL AGENT, ILLINOIS

Free format text: SECURITY AGREEMENT;ASSIGNORS:MAXLINEAR, INC.;ENTROPIC COMMUNICATIONS, LLC (F/K/A ENTROPIC COMMUNICATIONS, INC.);EXAR CORPORATION;REEL/FRAME:042453/0001

Effective date: 20170512

AS Assignment

Owner name: DYNAMIC DATA TECHNOLOGIES LLC, MINNESOTA

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:MAXLINEAR INC.;ENTROPIC COMMUNICATIONS LLC;REEL/FRAME:047914/0274

Effective date: 20180418

AS Assignment

Owner name: NXP B.V., NETHERLANDS

Free format text: RELEASE BY SECURED PARTY;ASSIGNOR:MORGAN STANLEY SENIOR FUNDING, INC.;REEL/FRAME:050745/0001

Effective date: 20190903

AS Assignment

Owner name: MORGAN STANLEY SENIOR FUNDING, INC., MARYLAND

Free format text: CORRECTIVE ASSIGNMENT TO CORRECT THE REMOVE APPLICATION 12298143 PREVIOUSLY RECORDED ON REEL 042762 FRAME 0145. ASSIGNOR(S) HEREBY CONFIRMS THE SECURITY AGREEMENT SUPPLEMENT;ASSIGNOR:NXP B.V.;REEL/FRAME:051145/0184

Effective date: 20160218

Owner name: MORGAN STANLEY SENIOR FUNDING, INC., MARYLAND

Free format text: CORRECTIVE ASSIGNMENT TO CORRECT THE REMOVE APPLICATION 12298143 PREVIOUSLY RECORDED ON REEL 039361 FRAME 0212. ASSIGNOR(S) HEREBY CONFIRMS THE SECURITY AGREEMENT SUPPLEMENT;ASSIGNOR:NXP B.V.;REEL/FRAME:051029/0387

Effective date: 20160218

Owner name: MORGAN STANLEY SENIOR FUNDING, INC., MARYLAND

Free format text: CORRECTIVE ASSIGNMENT TO CORRECT THE REMOVE APPLICATION 12298143 PREVIOUSLY RECORDED ON REEL 042985 FRAME 0001. ASSIGNOR(S) HEREBY CONFIRMS THE SECURITY AGREEMENT SUPPLEMENT;ASSIGNOR:NXP B.V.;REEL/FRAME:051029/0001

Effective date: 20160218

Owner name: MORGAN STANLEY SENIOR FUNDING, INC., MARYLAND

Free format text: CORRECTIVE ASSIGNMENT TO CORRECT THE REMOVE APPLICATION12298143 PREVIOUSLY RECORDED ON REEL 039361 FRAME 0212. ASSIGNOR(S) HEREBY CONFIRMS THE SECURITY AGREEMENT SUPPLEMENT;ASSIGNOR:NXP B.V.;REEL/FRAME:051029/0387

Effective date: 20160218

Owner name: MORGAN STANLEY SENIOR FUNDING, INC., MARYLAND

Free format text: CORRECTIVE ASSIGNMENT TO CORRECT THE REMOVE APPLICATION12298143 PREVIOUSLY RECORDED ON REEL 042985 FRAME 0001. ASSIGNOR(S) HEREBY CONFIRMS THE SECURITY AGREEMENT SUPPLEMENT;ASSIGNOR:NXP B.V.;REEL/FRAME:051029/0001

Effective date: 20160218

Owner name: MORGAN STANLEY SENIOR FUNDING, INC., MARYLAND

Free format text: CORRECTIVE ASSIGNMENT TO CORRECT THE REMOVE APPLICATION 12298143 PREVIOUSLY RECORDED ON REEL 038017 FRAME 0058. ASSIGNOR(S) HEREBY CONFIRMS THE SECURITY AGREEMENT SUPPLEMENT;ASSIGNOR:NXP B.V.;REEL/FRAME:051030/0001

Effective date: 20160218

Owner name: MORGAN STANLEY SENIOR FUNDING, INC., MARYLAND

Free format text: CORRECTIVE ASSIGNMENT TO CORRECT THE REMOVE APPLICATION12298143 PREVIOUSLY RECORDED ON REEL 042762 FRAME 0145. ASSIGNOR(S) HEREBY CONFIRMS THE SECURITY AGREEMENT SUPPLEMENT;ASSIGNOR:NXP B.V.;REEL/FRAME:051145/0184

Effective date: 20160218

MAFP Maintenance fee payment

Free format text: PAYMENT OF MAINTENANCE FEE, 8TH YEAR, LARGE ENTITY (ORIGINAL EVENT CODE: M1552); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY

Year of fee payment: 8

AS Assignment

Owner name: MUFG UNION BANK, N.A., CALIFORNIA

Free format text: SUCCESSION OF AGENCY (REEL 042453 / FRAME 0001);ASSIGNOR:JPMORGAN CHASE BANK, N.A.;REEL/FRAME:053115/0842

Effective date: 20200701

AS Assignment

Owner name: MAXLINEAR, INC., CALIFORNIA

Free format text: RELEASE BY SECURED PARTY;ASSIGNOR:MUFG UNION BANK, N.A.;REEL/FRAME:056656/0204

Effective date: 20210623

Owner name: EXAR CORPORATION, CALIFORNIA

Free format text: RELEASE BY SECURED PARTY;ASSIGNOR:MUFG UNION BANK, N.A.;REEL/FRAME:056656/0204

Effective date: 20210623

Owner name: MAXLINEAR COMMUNICATIONS LLC, CALIFORNIA

Free format text: RELEASE BY SECURED PARTY;ASSIGNOR:MUFG UNION BANK, N.A.;REEL/FRAME:056656/0204

Effective date: 20210623