US20080159393A1 - Motion compensation method and apparatus that sequentially use global motion compensation and local motion compensation, decoding method, video encoder, and video decoder - Google Patents

Motion compensation method and apparatus that sequentially use global motion compensation and local motion compensation, decoding method, video encoder, and video decoder Download PDF

Info

Publication number
US20080159393A1
US20080159393A1 US11/748,537 US74853707A US2008159393A1 US 20080159393 A1 US20080159393 A1 US 20080159393A1 US 74853707 A US74853707 A US 74853707A US 2008159393 A1 US2008159393 A1 US 2008159393A1
Authority
US
United States
Prior art keywords
motion compensation
motion information
model
local
global motion
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
US11/748,537
Other versions
US8374244B2 (en
Inventor
Tammy Lee
Woo-jin Han
Mathew Manu
Kyo-hyuk Lee
Sang-Rae Lee
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Samsung Electronics Co Ltd
Original Assignee
Samsung Electronics Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Samsung Electronics Co Ltd filed Critical Samsung Electronics Co Ltd
Assigned to SAMSUNG ELECTRONICS CO., LTD reassignment SAMSUNG ELECTRONICS CO., LTD ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: HAN, WOO-JIN, LEE, KYO-HYUK, LEE, SANG-RAE, LEE, TAMMY, MANU, MATHEW
Publication of US20080159393A1 publication Critical patent/US20080159393A1/en
Application granted granted Critical
Publication of US8374244B2 publication Critical patent/US8374244B2/en
Expired - Fee Related legal-status Critical Current
Adjusted expiration legal-status Critical

Links

Images

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/50Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding
    • H04N19/503Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding involving temporal prediction
    • H04N19/51Motion estimation or motion compensation
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/50Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding
    • H04N19/503Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding involving temporal prediction
    • H04N19/51Motion estimation or motion compensation
    • H04N19/527Global motion vector estimation
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/60Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using transform coding
    • H04N19/61Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using transform coding in combination with predictive coding

Definitions

  • Methods and apparatuses consistent with the present invention generally relate to video coding and decoding, and more particularly, to motion compensation sequentially using global motion compensation and local motion compensation for video coding and decoding.
  • Motion estimation and compensation play an important role in video data compression and use high temporal redundancy between consecutive frames in a video sequence for high compression efficiency.
  • a motion in a video sequence is caused by movement of an object or panning, zooming, or rotation of a camera.
  • a motion caused by movement of an object is called a local motion and a motion caused by panning, zooming, or rotation of a camera is called a global motion.
  • Most objects have complicated motions, i.e., a global motion and a local motion, and the local motion may be very little. Therefore, it is necessary to perform motion compensation separately for the global motion and the local motion.
  • the present invention provides a motion compensation method and apparatus that sequentially use global motion compensation and local motion compensation, a video decoding method, a video encoder, and a video decoder.
  • a motion compensation method including: extracting global motion information of a reference block; performing global motion compensation by applying the extracted global motion information to the reference block; extracting local motion information of the global motion-compensated reference block; and performing local motion compensation by applying the local motion information to the global motion-compensated reference block.
  • a computer-readable recording medium having recorded thereon a program for implementing the motion compensation method.
  • a motion compensation apparatus including: a global motion estimation unit which extracts global motion information of a reference block; a global motion compensation unit which performs global motion compensation by applying the extracted global motion information to the reference block; a local motion estimation unit which extracts local motion information of the global motion-compensated reference block; and a local motion compensation unit which performs local motion compensation by applying the local motion information to the global motion-compensated reference block.
  • a video encoder including: a motion compensation apparatus which extracts motion information of a reference block and generates a motion-compensated reference block by sequentially applying global motion compensation and local motion compensation to the reference block; an addition unit which subtracts the motion-compensated reference block from a current block, thereby generating a residual block; a frequency transformation unit which transforms spatial-domain coefficients of the residual block output from the addition unit into frequency-domain coefficients; a quantization unit which quantizes the transformed frequency-domain coefficients of the residual block, and an entropy-coding unit which entropy-codes the quantized frequency-domain coefficients of the residual block.
  • a video decoder including: an entropy-decoding unit which entropy-decodes global motion information, local motion information, and a residual block from an encoded bitstream; an inverse quantization unit which inversely quantizes the entropy-decoded frequency-domain coefficients of the residual block; an inverse frequency transformation unit which inversely transforms the inversely quantized frequency-domain coefficients of the residual block into spatial-domain coefficients; a global motion compensation unit which performs global motion compensation by applying the entropy-decoded global motion information to a reference block; a local motion compensation unit which performs local motion compensation by applying the entropy-decoded local motion information to the global motion-compensated reference block; and an addition unit which adds the residual block having the inversely transformed spatial-domain coefficients to the local motion-compensated reference block, thereby generating a current block.
  • a video decoding method including: entropy-decoding global motion information, local motion information, and a residual block from an encoded bitstream; inversely quantizing the entropy-decoded frequency-domain coefficients of the residual block; inversely transforming the inversely quantized frequency-domain coefficients of the residual block into spatial-domain coefficients; performing global motion compensation by applying the entropy-decoded global motion information to a reference block, performing local motion compensation by applying the entropy-decoded local motion information to the global motion-compensated reference block, and adding the residual block having the inversely transformed spatial-domain coefficients to the local motion-compensated reference block, thereby generating a current block.
  • a computer-readable recording medium having recorded thereon a program for implementing the video decoding method.
  • the global motion information may comprise at least one parameter of one of an Affine model, a Translation model, a Perspective model, and a Projective model.
  • the local motion information may comprise at least one parameter of one of an Affine model, a Translation model, a Perspective model, and a Projective model, or comprises a local motion vector using a block matching algorithm.
  • FIG. 1 is a block diagram of a video encoder including a motion compensation apparatus according to an exemplary embodiment of the present invention
  • FIGS. 2A through 2D are views for explaining global motion compensation and local motion compensation according to an exemplary embodiment of the present invention
  • FIGS. 3A and 3B illustrate a global motion vector and a local motion vector according to an exemplary embodiment of the present invention
  • FIG. 4 is a flowchart of a video encoding method including a motion compensation process sequentially using global motion compensation and local motion compensation according to an exemplary embodiment of the present invention
  • FIG. 5 is a block diagram of a video decoder according to an exemplary embodiment of the present invention.
  • FIG. 6 is a flowchart of a video decoding method according to an exemplary embodiment of the present invention.
  • FIG. 1 is a block diagram of a video encoder 100 according to an exemplary embodiment of the present invention
  • FIGS. 2A through 2D are views for explaining global motion compensation and local motion compensation according to an exemplary embodiment of the present invention
  • FIGS. 3A and 3B illustrate a global motion vector G and a local motion vector L.
  • a motion compensation apparatus 120 includes a global motion estimation unit 101 , a global motion compensation unit 102 , a local motion estimation unit 103 , and a local motion compensation unit 104 .
  • the global motion estimation unit 101 extracts global motion information from a reference frame, i.e., a previous frame, stored in a frame storing unit 105 , and a current frame.
  • the extracted global motion information is output to the global motion compensation unit 102 and an entropy-coding unit 109 .
  • a global motion is caused by movement of a camera, i.e., panning, zooming, or rotation of the camera, and is interpreted as including an object-based motion in a broad sense.
  • various motion models may be selected and used. For example, one of a Translation model, an Affine model, and a Projective model may be used. Two parameters may be used as global motion information in the translation model, six parameters may be used as global motion information in the Affine model, and eight parameters may be used as global motion information in the Projective model.
  • a Translation model an Affine model
  • a Projective model Two parameters may be used as global motion information in the translation model
  • six parameters may be used as global motion information in the Affine model
  • eight parameters may be used as global motion information in the Projective model.
  • those motion models are only examples and other various motion models may be used to generate global motion information, as will be obvious to those skilled in the art.
  • the global motion compensation unit 102 performs global motion compensation on a reference block of a reference frame stored in the frame storing unit 105 using the global motion information received from the global motion estimation unit 101 .
  • a global motion vector is generated using global motion parameters and global motion compensation is performed pixel-by-pixel.
  • the global motion-compensated reference block is delivered to the local motion compensation unit 104 .
  • FIGS. 2A and 2D are views for explaining global motion compensation and local motion compensation according to an exemplary embodiment of the present invention. More specifically, an object of a reference frame before motion compensation is illustrated in FIG. 2A , an object of the current frame is illustrated in FIG. 2B , a dashed portion indicated by a bold-dotted line indicates an object before motion compensation and a portion indicated by a fine-dotted line indicates an object after global motion compensation and local motion compensation in FIG. 2C , and a dashed portion indicated by a bold-dotted line indicates an object after global motion compensation and local motion compensation and a portion indicated by a fine-dotted line indicates an object after global motion compensation in FIG. 2D .
  • FIGS. 2A through 2D illustrate only examples for facilitating understanding of the present invention and it is obvious to those skilled in the art that global and local motions of an object are more complicated than shown in FIGS. 2A through 2D .
  • the local motion estimation unit 103 extracts local motion information from the global motion-compensated reference block and the current block of the current frame.
  • the extracted local motion information is output to the local motion compensation unit 104 and the entropy-coding unit 109 .
  • a variety of motion models such as the translation model, the Affine model, or the Projective model may be selected and used for local motion estimation and in this case, motion parameters according to each of the motion models may be used as the local motion information.
  • the local motion information may also be obtained by a blocking matching algorithm and in this case, the local motion information is a motion vector.
  • FIGS. 3A and 3B illustrate a global motion vector G and a local motion vector L, in which a reference frame is illustrated in FIG. 3A and a current frame is illustrated in FIG. 3B .
  • a reference numeral 300 indicates a current block of a current frame
  • a reference numeral 301 indicates a reference block of a reference frame, which is located in the spatially same position as the current block 300
  • a reference numeral 302 indicates a block for explaining the global motion vector G
  • a reference numeral 303 indicates a block for explaining the local motion vector L.
  • motion compensation can be performed block-by-block or partition-by-partition.
  • local motion compensation is performed based on a point in which global motion compensation is performed in relation to FIGS. 3A and 3B .
  • the local motion compensation unit 104 applies the local motion information received from the local motion estimation unit 103 to the global motion-compensated reference block received from the global motion compensation unit 102 , thereby generating a local motion-compensated reference block.
  • the generated local motion-compensated reference block is output to an addition unit 111 a and an addition unit 106 .
  • the addition unit 106 subtracts the reference block received from the local motion compensation unit 104 from the current block of the current frame, thereby generating a residual block.
  • the frame storing unit 105 stores a reference frame for motion estimation and compensation of the current block.
  • the reference frame may temporally precede or follow the current frame.
  • a frequency transformation unit 107 transforms spatial-domain values of the residual block into frequency-domain values using a discrete cosine transformation (DCT) and outputs the residual block having the transformed frequency-domain values to a quantization unit 108 .
  • DCT discrete cosine transformation
  • the quantization unit 108 quantizes the residual block having the transformed frequency-domain values received from the frequency transformation unit 107 and outputs the quantized residual block to the entropy-coding unit 109 .
  • the entropy-coding unit 109 performs entropy-coding on the quantized residual block, and the local motion information and the global motion information that are received from the motion compensation apparatus 120 .
  • a data flow for reconstruction of a reference frame is as follows.
  • An inverse quantization unit 110 inversely quantizes the quantized residual block and outputs the inversely quantized residual block to an inverse frequency transformation unit 111 .
  • the inverse frequency transformation unit 111 inversely transforms the frequency-domain values of the inversely quantized residual block into spatial-domain values and outputs the residual block having the inversely transformed spatial-domain values to the addition unit 111 a.
  • the addition unit 111 a adds the local motion-compensated reference block received from the local motion compensation unit 104 to the residual block having the spatial-domain values in order to generate a reference block, and the generated reference block is stored in the frame storing unit 105 .
  • FIG. 4 is a flowchart of a video encoding method including a motion compensation process sequentially using global motion compensation and local motion compensation according to an exemplary embodiment of the present invention.
  • video frames are received in operation S 400 .
  • global motion information is extracted from a current block of a current frame and a reference block from among the received video frames.
  • Global motion compensation is performed by applying the global motion information to the reference block, thereby generating a global motion-compensated reference block.
  • local motion estimation is performed on the global motion-compensated reference block and the current block, thereby extracting local motion information.
  • the local motion information is applied to the global motion-compensated reference block, thereby generating a local motion-compensated reference block.
  • the reference block is subtracted from the current block in order to generate a residual block.
  • a DCT, quantization, and entropy-coding are sequentially performed on the residual block.
  • FIG. 5 is a block diagram of a video decoder 500 according to an exemplary embodiment of the present invention.
  • the video decoder 500 includes an entropy-decoding unit 501 , an inverse quantization unit 502 , an inverse frequency transformation unit 503 , a global motion compensation unit 504 , a local motion compensation unit 505 , an addition unit 506 , and a frame storing unit 507 .
  • the entropy-decoding unit 501 performs entropy-decoding on an encoded bitstream input from a video encoder and outputs the entropy-decoded bitstream to the inverse quantization unit 502 .
  • the entropy-decoding unit 501 performs entropy-decoding on global motion information, local motion information, and frequency-domain coefficients of a residual block.
  • the entropy-decoded global motion information is output to the global motion compensation unit 504
  • the entropy-decoded local motion information is output to the local motion compensation unit 505
  • the residual block including the entropy-decoded frequency-domain coefficients is output to the inverse quantization unit 502 .
  • the inverse quantization unit 502 inversely quantizes the entropy-decoded frequency-domain coefficients of the residual block received from the entropy-decoding unit 501 and outputs the inversely quantized frequency-domain coefficients to the inverse frequency transformation unit 503 .
  • the inverse frequency transformation unit 503 inversely transforms the inversely quantized frequency-domain coefficients of the residual block received from the inverse quantization unit 502 into spatial-domain coefficients and outputs the inversely transformed spatial-domain coefficients to the addition unit 506 .
  • the addition unit 506 adds the local motion-compensated reference block received from the local motion compensation unit 504 to the residual block received from the inverse frequency transformation unit 503 , thereby generating decoded data.
  • the output of the addition unit 506 is decoded video data of the current frame.
  • the global motion compensation unit 504 receives the decoded global motion information from the entropy-decoding unit 501 and performs global motion compensation by applying the received global motion information to the reference block received from the frame storing unit 507 .
  • the global motion-compensated reference block is output to the local motion compensation unit 505 .
  • the local motion compensation unit 505 performs local motion compensation by applying the local motion information received from the entropy-decoding unit 501 to the global motion-compensated reference block received from the global motion compensation unit 504 .
  • the local motion-compensated reference block is output to the addition unit 506 .
  • the frame storing unit 507 stores reference blocks referred to for the generation of the current block.
  • FIG. 6 is a flowchart of a video decoding method according to an exemplary embodiment of the present invention.
  • global motion information, local motion information, and a residual block are entropy-decoded from a bitstream in operation S 601 .
  • frequency-domain coefficients of the entropy-decoded residual block are inversely quantized.
  • the local motion-compensated reference block is added to the residual block having the transformed spatial-domain coefficients, thereby generating a current block.
  • the global motion information or the local motion information may be parameters of one of an Affine model, a Translation model, a Perspective model and a Projective model.
  • the local motion information is a local motion vector using a block matching algorithm.
  • the amount of information transmitted to a video decoder can be reduced by sequentially using global motion compensation and local motion compensation.
  • the exemplary embodiments of the present invention can also be embodied as computer-readable code on a computer-readable recording medium.
  • the computer-readable recording medium is any data storage device that can store data which can be thereafter read by a computer system. Examples of computer-readable recording media include read-only memory (ROM), random-access memory (RAM), CD-ROMs, magnetic tapes, floppy disks, and optical data storage devices.
  • ROM read-only memory
  • RAM random-access memory
  • CD-ROMs compact discs, digital versatile discs, and Blu-rays, and Blu-rays, and Blu-rays, and Blu-rays, and Blu-rays, etc.
  • the computer-readable recording medium can also be distributed over a network of coupled computer systems so that the computer-readable code is stored and executed in a decentralized fashion.

Abstract

A motion compensation method and apparatus that sequentially use global motion compensation and local motion compensation, a video decoding method, a video encoder, and a video decoder are provided. The motion compensation method includes extracting global motion information of a reference block, performing global motion compensation by applying the extracted global motion information to the reference block, extracting local motion information of the global motion-compensated reference block, and performing local motion compensation by applying the local motion information to the global motion-compensated reference block.

Description

    CROSS-REFERENCE TO RELATED PATENT APPLICATIONS
  • This application claims priority from Korean Patent Application No. 10-2007-0000709, filed on Jan. 3, 2007, in the Korean Intellectual Property Office, the disclosure of which is incorporated herein in its entirety by reference.
  • BACKGROUND OF THE INVENTION
  • 1. Field of the Invention
  • Methods and apparatuses consistent with the present invention generally relate to video coding and decoding, and more particularly, to motion compensation sequentially using global motion compensation and local motion compensation for video coding and decoding.
  • 2. Description of the Related Art
  • Motion estimation and compensation play an important role in video data compression and use high temporal redundancy between consecutive frames in a video sequence for high compression efficiency. A motion in a video sequence is caused by movement of an object or panning, zooming, or rotation of a camera. Generally, a motion caused by movement of an object is called a local motion and a motion caused by panning, zooming, or rotation of a camera is called a global motion. Most objects have complicated motions, i.e., a global motion and a local motion, and the local motion may be very little. Therefore, it is necessary to perform motion compensation separately for the global motion and the local motion.
  • SUMMARY OF THE INVENTION
  • The present invention provides a motion compensation method and apparatus that sequentially use global motion compensation and local motion compensation, a video decoding method, a video encoder, and a video decoder.
  • According to one aspect of the present invention, there is provided a motion compensation method including: extracting global motion information of a reference block; performing global motion compensation by applying the extracted global motion information to the reference block; extracting local motion information of the global motion-compensated reference block; and performing local motion compensation by applying the local motion information to the global motion-compensated reference block.
  • According to another aspect of the present invention, there is provided a computer-readable recording medium having recorded thereon a program for implementing the motion compensation method.
  • According to another aspect of the present invention, there is provided a motion compensation apparatus including: a global motion estimation unit which extracts global motion information of a reference block; a global motion compensation unit which performs global motion compensation by applying the extracted global motion information to the reference block; a local motion estimation unit which extracts local motion information of the global motion-compensated reference block; and a local motion compensation unit which performs local motion compensation by applying the local motion information to the global motion-compensated reference block.
  • According to another aspect of the present invention, there is provided a video encoder including: a motion compensation apparatus which extracts motion information of a reference block and generates a motion-compensated reference block by sequentially applying global motion compensation and local motion compensation to the reference block; an addition unit which subtracts the motion-compensated reference block from a current block, thereby generating a residual block; a frequency transformation unit which transforms spatial-domain coefficients of the residual block output from the addition unit into frequency-domain coefficients; a quantization unit which quantizes the transformed frequency-domain coefficients of the residual block, and an entropy-coding unit which entropy-codes the quantized frequency-domain coefficients of the residual block.
  • According to another aspect of the present invention, there is provided a video decoder including: an entropy-decoding unit which entropy-decodes global motion information, local motion information, and a residual block from an encoded bitstream; an inverse quantization unit which inversely quantizes the entropy-decoded frequency-domain coefficients of the residual block; an inverse frequency transformation unit which inversely transforms the inversely quantized frequency-domain coefficients of the residual block into spatial-domain coefficients; a global motion compensation unit which performs global motion compensation by applying the entropy-decoded global motion information to a reference block; a local motion compensation unit which performs local motion compensation by applying the entropy-decoded local motion information to the global motion-compensated reference block; and an addition unit which adds the residual block having the inversely transformed spatial-domain coefficients to the local motion-compensated reference block, thereby generating a current block.
  • According to another aspect of the present invention, there is provided a video decoding method including: entropy-decoding global motion information, local motion information, and a residual block from an encoded bitstream; inversely quantizing the entropy-decoded frequency-domain coefficients of the residual block; inversely transforming the inversely quantized frequency-domain coefficients of the residual block into spatial-domain coefficients; performing global motion compensation by applying the entropy-decoded global motion information to a reference block, performing local motion compensation by applying the entropy-decoded local motion information to the global motion-compensated reference block, and adding the residual block having the inversely transformed spatial-domain coefficients to the local motion-compensated reference block, thereby generating a current block.
  • According to another aspect of the present invention, there is provided a computer-readable recording medium having recorded thereon a program for implementing the video decoding method.
  • The global motion information may comprise at least one parameter of one of an Affine model, a Translation model, a Perspective model, and a Projective model.
  • The local motion information may comprise at least one parameter of one of an Affine model, a Translation model, a Perspective model, and a Projective model, or comprises a local motion vector using a block matching algorithm.
  • BRIEF DESCRIPTION OF THE DRAWINGS
  • The above and other aspects of the present invention will become more apparent by describing in detail an exemplary embodiment thereof with reference to the attached drawings, in which:
  • FIG. 1 is a block diagram of a video encoder including a motion compensation apparatus according to an exemplary embodiment of the present invention;
  • FIGS. 2A through 2D are views for explaining global motion compensation and local motion compensation according to an exemplary embodiment of the present invention;
  • FIGS. 3A and 3B illustrate a global motion vector and a local motion vector according to an exemplary embodiment of the present invention;
  • FIG. 4 is a flowchart of a video encoding method including a motion compensation process sequentially using global motion compensation and local motion compensation according to an exemplary embodiment of the present invention;
  • FIG. 5 is a block diagram of a video decoder according to an exemplary embodiment of the present invention; and
  • FIG. 6 is a flowchart of a video decoding method according to an exemplary embodiment of the present invention.
  • DETAILED DESCRIPTION OF EXEMPLARY EMBODIMENTS
  • Hereinafter, an exemplary embodiment of the present invention will be described in detail with reference to the accompanying drawings. It should be noticed that like reference numerals refer to like elements illustrated in one or more of the drawings. It would be obvious to those skilled in the art that many specifics like elements of a circuit are provided only to facilitate understanding of the present invention and the present invention can be implemented without those specifics. In the following description of the present invention, a detailed description of known functions and configurations incorporated herein will be omitted for conciseness and clarity.
  • FIG. 1 is a block diagram of a video encoder 100 according to an exemplary embodiment of the present invention, FIGS. 2A through 2D are views for explaining global motion compensation and local motion compensation according to an exemplary embodiment of the present invention, and FIGS. 3A and 3B illustrate a global motion vector G and a local motion vector L. In FIG. 1, a motion compensation apparatus 120 includes a global motion estimation unit 101, a global motion compensation unit 102, a local motion estimation unit 103, and a local motion compensation unit 104.
  • Referring to FIG. 1, the global motion estimation unit 101 extracts global motion information from a reference frame, i.e., a previous frame, stored in a frame storing unit 105, and a current frame. The extracted global motion information is output to the global motion compensation unit 102 and an entropy-coding unit 109. In the present invention, a global motion is caused by movement of a camera, i.e., panning, zooming, or rotation of the camera, and is interpreted as including an object-based motion in a broad sense.
  • To extract the global motion information, various motion models may be selected and used. For example, one of a Translation model, an Affine model, and a Projective model may be used. Two parameters may be used as global motion information in the translation model, six parameters may be used as global motion information in the Affine model, and eight parameters may be used as global motion information in the Projective model. However, those motion models are only examples and other various motion models may be used to generate global motion information, as will be obvious to those skilled in the art.
  • The global motion compensation unit 102 performs global motion compensation on a reference block of a reference frame stored in the frame storing unit 105 using the global motion information received from the global motion estimation unit 101. In other words, a global motion vector is generated using global motion parameters and global motion compensation is performed pixel-by-pixel. The global motion-compensated reference block is delivered to the local motion compensation unit 104.
  • FIGS. 2A and 2D are views for explaining global motion compensation and local motion compensation according to an exemplary embodiment of the present invention. More specifically, an object of a reference frame before motion compensation is illustrated in FIG. 2A, an object of the current frame is illustrated in FIG. 2B, a dashed portion indicated by a bold-dotted line indicates an object before motion compensation and a portion indicated by a fine-dotted line indicates an object after global motion compensation and local motion compensation in FIG. 2C, and a dashed portion indicated by a bold-dotted line indicates an object after global motion compensation and local motion compensation and a portion indicated by a fine-dotted line indicates an object after global motion compensation in FIG. 2D. However, FIGS. 2A through 2D illustrate only examples for facilitating understanding of the present invention and it is obvious to those skilled in the art that global and local motions of an object are more complicated than shown in FIGS. 2A through 2D.
  • Referring back to FIG. 1, the local motion estimation unit 103 extracts local motion information from the global motion-compensated reference block and the current block of the current frame. The extracted local motion information is output to the local motion compensation unit 104 and the entropy-coding unit 109. Similarly with global motion estimation, a variety of motion models such as the translation model, the Affine model, or the Projective model may be selected and used for local motion estimation and in this case, motion parameters according to each of the motion models may be used as the local motion information. The local motion information may also be obtained by a blocking matching algorithm and in this case, the local motion information is a motion vector.
  • FIGS. 3A and 3B illustrate a global motion vector G and a local motion vector L, in which a reference frame is illustrated in FIG. 3A and a current frame is illustrated in FIG. 3B. A reference numeral 300 indicates a current block of a current frame, a reference numeral 301 indicates a reference block of a reference frame, which is located in the spatially same position as the current block 300, a reference numeral 302 indicates a block for explaining the global motion vector G, and a reference numeral 303 indicates a block for explaining the local motion vector L. As illustrated in FIGS. 3A and 3B, motion compensation can be performed block-by-block or partition-by-partition. In particular, it should be noted that local motion compensation is performed based on a point in which global motion compensation is performed in relation to FIGS. 3A and 3B.
  • Referring back to FIG. 1, the local motion compensation unit 104 applies the local motion information received from the local motion estimation unit 103 to the global motion-compensated reference block received from the global motion compensation unit 102, thereby generating a local motion-compensated reference block. The generated local motion-compensated reference block is output to an addition unit 111 a and an addition unit 106.
  • The addition unit 106 subtracts the reference block received from the local motion compensation unit 104 from the current block of the current frame, thereby generating a residual block.
  • The frame storing unit 105 stores a reference frame for motion estimation and compensation of the current block. The reference frame may temporally precede or follow the current frame.
  • A frequency transformation unit 107 transforms spatial-domain values of the residual block into frequency-domain values using a discrete cosine transformation (DCT) and outputs the residual block having the transformed frequency-domain values to a quantization unit 108.
  • The quantization unit 108 quantizes the residual block having the transformed frequency-domain values received from the frequency transformation unit 107 and outputs the quantized residual block to the entropy-coding unit 109.
  • The entropy-coding unit 109 performs entropy-coding on the quantized residual block, and the local motion information and the global motion information that are received from the motion compensation apparatus 120.
  • A data flow for reconstruction of a reference frame is as follows.
  • An inverse quantization unit 110 inversely quantizes the quantized residual block and outputs the inversely quantized residual block to an inverse frequency transformation unit 111.
  • The inverse frequency transformation unit 111 inversely transforms the frequency-domain values of the inversely quantized residual block into spatial-domain values and outputs the residual block having the inversely transformed spatial-domain values to the addition unit 111 a.
  • The addition unit 111 a adds the local motion-compensated reference block received from the local motion compensation unit 104 to the residual block having the spatial-domain values in order to generate a reference block, and the generated reference block is stored in the frame storing unit 105.
  • FIG. 4 is a flowchart of a video encoding method including a motion compensation process sequentially using global motion compensation and local motion compensation according to an exemplary embodiment of the present invention.
  • Referring to FIG. 4, video frames are received in operation S400.
  • In operation S401, through global motion estimation, global motion information is extracted from a current block of a current frame and a reference block from among the received video frames. Global motion compensation is performed by applying the global motion information to the reference block, thereby generating a global motion-compensated reference block.
  • In operation S402, local motion estimation is performed on the global motion-compensated reference block and the current block, thereby extracting local motion information. The local motion information is applied to the global motion-compensated reference block, thereby generating a local motion-compensated reference block.
  • In operation S403, the reference block is subtracted from the current block in order to generate a residual block. In operation S404, a DCT, quantization, and entropy-coding are sequentially performed on the residual block.
  • FIG. 5 is a block diagram of a video decoder 500 according to an exemplary embodiment of the present invention. Referring to FIG. 5, the video decoder 500 includes an entropy-decoding unit 501, an inverse quantization unit 502, an inverse frequency transformation unit 503, a global motion compensation unit 504, a local motion compensation unit 505, an addition unit 506, and a frame storing unit 507.
  • The entropy-decoding unit 501 performs entropy-decoding on an encoded bitstream input from a video encoder and outputs the entropy-decoded bitstream to the inverse quantization unit 502. In particular, in the case of inter-mode coding, the entropy-decoding unit 501 performs entropy-decoding on global motion information, local motion information, and frequency-domain coefficients of a residual block. The entropy-decoded global motion information is output to the global motion compensation unit 504, the entropy-decoded local motion information is output to the local motion compensation unit 505, and the residual block including the entropy-decoded frequency-domain coefficients is output to the inverse quantization unit 502.
  • The inverse quantization unit 502 inversely quantizes the entropy-decoded frequency-domain coefficients of the residual block received from the entropy-decoding unit 501 and outputs the inversely quantized frequency-domain coefficients to the inverse frequency transformation unit 503.
  • The inverse frequency transformation unit 503 inversely transforms the inversely quantized frequency-domain coefficients of the residual block received from the inverse quantization unit 502 into spatial-domain coefficients and outputs the inversely transformed spatial-domain coefficients to the addition unit 506.
  • The addition unit 506 adds the local motion-compensated reference block received from the local motion compensation unit 504 to the residual block received from the inverse frequency transformation unit 503, thereby generating decoded data. The output of the addition unit 506 is decoded video data of the current frame.
  • More specifically, the global motion compensation unit 504 receives the decoded global motion information from the entropy-decoding unit 501 and performs global motion compensation by applying the received global motion information to the reference block received from the frame storing unit 507. The global motion-compensated reference block is output to the local motion compensation unit 505.
  • The local motion compensation unit 505 performs local motion compensation by applying the local motion information received from the entropy-decoding unit 501 to the global motion-compensated reference block received from the global motion compensation unit 504. The local motion-compensated reference block is output to the addition unit 506.
  • The frame storing unit 507 stores reference blocks referred to for the generation of the current block.
  • FIG. 6 is a flowchart of a video decoding method according to an exemplary embodiment of the present invention.
  • Referring to FIG. 6, global motion information, local motion information, and a residual block are entropy-decoded from a bitstream in operation S601.
  • In operation S602, frequency-domain coefficients of the entropy-decoded residual block are inversely quantized.
  • In operation S603, the inversely quantized frequency-domain coefficients of the residual block are transformed into spatial-domain coefficients.
  • In operation S604, global motion compensation is performed by applying the entropy-decoded local motion information to the reference block.
  • In operation S605, local motion compensation is performed by applying the entropy-decoded local motion information to the global motion-compensated reference block.
  • In operation S606, the local motion-compensated reference block is added to the residual block having the transformed spatial-domain coefficients, thereby generating a current block.
  • The global motion information or the local motion information may be parameters of one of an Affine model, a Translation model, a Perspective model and a Projective model. The local motion information is a local motion vector using a block matching algorithm.
  • As described above, according to the exemplary embodiments of the present invention, the amount of information transmitted to a video decoder can be reduced by sequentially using global motion compensation and local motion compensation.
  • Meanwhile, the exemplary embodiments of the present invention can also be embodied as computer-readable code on a computer-readable recording medium. The computer-readable recording medium is any data storage device that can store data which can be thereafter read by a computer system. Examples of computer-readable recording media include read-only memory (ROM), random-access memory (RAM), CD-ROMs, magnetic tapes, floppy disks, and optical data storage devices. The computer-readable recording medium can also be distributed over a network of coupled computer systems so that the computer-readable code is stored and executed in a decentralized fashion.
  • While the present invention has been particularly shown and described with reference to exemplary embodiments thereof, it will be understood by those of ordinary skill in the art that various changes in form and detail may be made therein without departing from the spirit and scope of the present invention as defined by the following claims.

Claims (24)

1. A motion compensation method comprising:
extracting global motion information of a reference block;
performing global motion compensation by applying the extracted global motion information to the reference block;
extracting local motion information of the global motion-compensated reference block; and
performing local motion compensation by applying the local motion information to the global motion-compensated reference block.
2. The motion compensation method of claim 1, wherein the global motion information comprises at least one parameter of one of an Affine model, a Translation model, a Perspective model, and a Projective model.
3. The motion compensation method of claim 1, wherein the local motion information comprises at least one parameter of one of an Affine model, a Translation model, a Perspective model, and a Projective model.
4. The motion compensation method of claim 1, wherein the local motion information comprises a local motion vector using a block matching algorithm.
5. The motion compensation method of claim 1, further comprising transmitting the global motion information and the local motion information.
6. A computer-readable recording medium having recorded thereon a program for implementing the motion compensation method of one of claim 1.
7. A motion compensation apparatus comprising:
a global motion estimation unit which extracts global motion information of a reference block;
a global motion compensation unit which performs global motion compensation by applying the extracted global motion information to the reference block;
a local motion estimation unit which extracts local motion information of the global motion-compensated reference block; and
a local motion compensation unit which performs local motion compensation by applying the local motion information to the global motion-compensated reference block.
8. The motion compensation apparatus of claim 7, wherein the global motion information comprises at least one parameter of one of an Affine model, a Translation model, a Perspective model, and a Projective model.
9. The motion compensation apparatus of claim 7, wherein the local motion information comprises at least one parameter of one of an Affine model, a Translation model, a Perspective model, and a Projective model.
10. The motion compensation apparatus of claim 7, wherein the local motion information comprises a local motion vector using a block matching algorithm.
11. A video encoder comprising:
a motion compensation apparatus which extracts motion information of a reference block and generates a motion-compensated reference block by sequentially applying global motion compensation and local motion compensation to the reference block;
an addition unit which subtracts the motion-compensated reference block from a current block, thereby generating a residual block;
a frequency transformation unit which transforms spatial-domain coefficients of the residual block output from the addition unit into frequency-domain coefficients;
a quantization unit which quantizes the transformed frequency-domain coefficients of the residual block; and
an entropy-coding unit which entropy-codes the quantized frequency-domain coefficients of the residual block.
12. The video encoder of claim 11, wherein the motion compensation apparatus further comprises:
a global motion estimation unit which extracts global motion information of the reference block;
a global motion compensation unit which performs global motion compensation by applying the extracted global motion information to the reference block;
a local motion estimation unit which extracts local motion information of the global motion-compensated reference block; and
a local motion compensation unit which performs local motion compensation by applying the local motion information to the global motion-compensated reference block.
13. The video encoder of claim 12, wherein the global motion information comprises at least one parameter of one of an Affine model, a Translation model, a Perspective model, and a Projective model.
14. The video encoder of claim 12, wherein the local motion information comprises at least one parameter of one of an Affine model, a Translation model, a Perspective model, and a Projective model.
15. The video encoder of claim 12, wherein the local motion information comprises a local motion vector using a block matching algorithm.
16. A video decoder comprising:
an entropy-decoding unit which entropy-decodes global motion information, local motion information, and a residual block from an encoded bitstream;
an inverse quantization unit which inversely quantizes the entropy-decoded frequency-domain coefficients of the residual block;
an inverse frequency transformation unit which inversely transforms the inversely quantized frequency-domain coefficients of the residual block into spatial-domain coefficients;
a global motion compensation unit which performs global motion compensation by applying the entropy-decoded global motion information to a reference block;
a local motion compensation unit which performs local motion compensation by applying the entropy-decoded local motion information to the global motion-compensated reference block; and
an addition unit which adds the residual block having the inversely transformed spatial-domain coefficients to the local motion-compensated reference block, thereby generating a current block.
17. The video decoder of claim 16, wherein the global motion information comprises at least one parameter of one of an Affine model, a Translation model, a Perspective model, and a Projective model.
18. The video decoder of claim 16, wherein the local motion information comprises at least one parameter of one of an Affine model, a Translation model, a Perspective model, and a Projective model.
19. The video decoder of claim 16, wherein the local motion information comprises a local motion vector using a block matching algorithm.
20. A video decoding method comprising:
entropy-decoding global motion information, local motion information, and a residual block from an encoded bitstream;
inversely quantizing the entropy-decoded frequency-domain coefficients of the residual block;
inversely transforming the inversely quantized frequency-domain coefficients of the residual block into spatial-domain coefficients;
performing global motion compensation by applying the entropy-decoded global motion information to a reference block;
performing local motion compensation by applying the entropy-decoded local motion information to the global motion-compensated reference block; and
adding the residual block having the inversely transformed spatial-domain coefficients to the local motion-compensated reference block, thereby generating a current block.
21. The video decoding method of claim 20, wherein the global motion information comprises at least one parameter of one of an Affine model, a Translation model, a Perspective model, and a Projective model.
22. The video decoding method of claim 20, wherein the local motion information comprises at least one parameter of one of an Affine model, a Translation model, a Perspective model, and a Projective model.
23. The video decoding method of claim 20, wherein the local motion information comprises a local motion vector using a block matching algorithm.
24. A computer-readable recording medium having recorded thereon a program for implementing the video decoding method of one of claim 20.
US11/748,537 2007-01-03 2007-05-15 Motion compensation method and apparatus that sequentially use global motion compensation and local motion compensation, decoding method, video encoder, and video decoder Expired - Fee Related US8374244B2 (en)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
KR1020070000709A KR101365445B1 (en) 2007-01-03 2007-01-03 Method of using global and local motion compensation sequentially, decoding method, apparatus, encoder and decoder
KR10-2007-0000709 2007-01-03

Publications (2)

Publication Number Publication Date
US20080159393A1 true US20080159393A1 (en) 2008-07-03
US8374244B2 US8374244B2 (en) 2013-02-12

Family

ID=39583954

Family Applications (1)

Application Number Title Priority Date Filing Date
US11/748,537 Expired - Fee Related US8374244B2 (en) 2007-01-03 2007-05-15 Motion compensation method and apparatus that sequentially use global motion compensation and local motion compensation, decoding method, video encoder, and video decoder

Country Status (3)

Country Link
US (1) US8374244B2 (en)
KR (1) KR101365445B1 (en)
WO (1) WO2008082062A1 (en)

Cited By (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20130243336A1 (en) * 2011-09-13 2013-09-19 Dominique Thoreau Method for coding and reconstructing a pixel block and corresponding devices
US20160088314A1 (en) * 2010-12-06 2016-03-24 Sk Telecom Co., Ltd. Method and apparatus for encoding/decoding image in spatial domain for noise component
US9426464B2 (en) 2012-07-04 2016-08-23 Thomson Licensing Method for coding and decoding a block of pixels from a motion model
WO2017087751A1 (en) * 2015-11-20 2017-05-26 Mediatek Inc. Method and apparatus for global motion compensation in video coding system
US20200227058A1 (en) * 2015-03-09 2020-07-16 Fraunhofer-Gesellschaft Zur Foerderung Der Angewandten Forschung E.V. Audio encoder, audio decoder, method for encoding an audio signal and method for decoding an encoded audio signal

Families Citing this family (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN102263958A (en) * 2011-07-26 2011-11-30 中兴通讯股份有限公司 method and device for obtaining initial point based on H264 motion estimation algorithm
CN108141594B (en) 2015-10-13 2021-02-26 三星电子株式会社 Method and apparatus for encoding or decoding image
EP3301928A1 (en) * 2016-09-30 2018-04-04 Thomson Licensing Methods, devices and stream to encode global rotation motion compensated images

Citations (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20020034251A1 (en) * 1996-10-30 2002-03-21 Yoshinori Suzuki Method and apparatus for image coding
US20060114997A1 (en) * 2004-11-29 2006-06-01 Dan Lelescu Temporal prediction in video coding

Family Cites Families (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US6205178B1 (en) 1996-09-20 2001-03-20 Hitachi, Ltd. Method and synthesizing a predicted image, video coding device and video coding method
CN100466747C (en) 1997-02-14 2009-03-04 日本电信电话株式会社 Predictive coding device and decoding device for dynamic image
KR100386583B1 (en) 2000-11-30 2003-06-02 엘지전자 주식회사 Apparatus and method for transcoding video
US20030123738A1 (en) 2001-11-30 2003-07-03 Per Frojdh Global motion compensation for video pictures
KR100610892B1 (en) 2005-01-18 2006-08-09 엘지전자 주식회사 Image Processing Device and Method for Plasma Display Panel

Patent Citations (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20020034251A1 (en) * 1996-10-30 2002-03-21 Yoshinori Suzuki Method and apparatus for image coding
US20060114997A1 (en) * 2004-11-29 2006-06-01 Dan Lelescu Temporal prediction in video coding

Cited By (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20160088314A1 (en) * 2010-12-06 2016-03-24 Sk Telecom Co., Ltd. Method and apparatus for encoding/decoding image in spatial domain for noise component
US9510017B2 (en) * 2010-12-06 2016-11-29 Sk Telecom Co., Ltd. Method and apparatus for encoding/decoding image in spatial domain for noise component
US20130243336A1 (en) * 2011-09-13 2013-09-19 Dominique Thoreau Method for coding and reconstructing a pixel block and corresponding devices
US9135721B2 (en) * 2011-09-13 2015-09-15 Thomson Licensing Method for coding and reconstructing a pixel block and corresponding devices
US9426464B2 (en) 2012-07-04 2016-08-23 Thomson Licensing Method for coding and decoding a block of pixels from a motion model
US20200227058A1 (en) * 2015-03-09 2020-07-16 Fraunhofer-Gesellschaft Zur Foerderung Der Angewandten Forschung E.V. Audio encoder, audio decoder, method for encoding an audio signal and method for decoding an encoded audio signal
WO2017087751A1 (en) * 2015-11-20 2017-05-26 Mediatek Inc. Method and apparatus for global motion compensation in video coding system
US11082713B2 (en) 2015-11-20 2021-08-03 Mediatek Inc. Method and apparatus for global motion compensation in video coding system

Also Published As

Publication number Publication date
KR20080064010A (en) 2008-07-08
KR101365445B1 (en) 2014-02-19
US8374244B2 (en) 2013-02-12
WO2008082062A1 (en) 2008-07-10

Similar Documents

Publication Publication Date Title
US8228989B2 (en) Method and apparatus for encoding and decoding based on inter prediction
US8374244B2 (en) Motion compensation method and apparatus that sequentially use global motion compensation and local motion compensation, decoding method, video encoder, and video decoder
US8649431B2 (en) Method and apparatus for encoding and decoding image by using filtered prediction block
US8503532B2 (en) Method and apparatus for inter prediction encoding/decoding an image using sub-pixel motion estimation
US8194748B2 (en) Apparatus for scalable encoding/decoding of moving image and method thereof
US9667961B2 (en) Video encoding and decoding apparatus, method, and system
US20060193382A1 (en) Method and apparatus for control of rate-distortion tradeoff by mode selection in video encoders
US20110002390A1 (en) Methods and systems for motion vector derivation at a video decoder
KR100694137B1 (en) Apparatus for encoding or decoding motion image, method therefor, and recording medium storing a program to implement thereof
KR101456491B1 (en) Method and apparatus for encoding and decoding image based on plurality of reference pictures
US20130128973A1 (en) Method and apparatus for encoding and decoding an image using a reference picture
Liu et al. Nonlinear motion-compensated interpolation for low-bit-rate video
US8699576B2 (en) Method of and apparatus for estimating motion vector based on sizes of neighboring partitions, encoder, decoding, and decoding method
JP2004363931A (en) Method and apparatus for re-encoding hierarchically encoded bit stream
WO1997004402A1 (en) Method and apparatus for regenerating a dense motion vector field
KR20090038278A (en) Method and apparatus for encoding and decoding image
US20090279610A1 (en) Method and apparatus for encoding/decoding with interlace scanning based motion vector transformation
KR20110024574A (en) Integrated video encoding method and apparatus
JPH08307860A (en) Scene re-encoder
KR20060069227A (en) Method and apparatus for deriving motion vectors of macro blocks from motion vectors of pictures of base layer when encoding/decoding video signal
US20040013200A1 (en) Advanced method of coding and decoding motion vector and apparatus therefor
US11647228B2 (en) Method and apparatus for encoding and decoding video signal using transform domain prediction for prediction unit partition
KR20090078114A (en) Multi-view image coding method and apparatus using variable gop prediction structure, multi-view image decoding apparatus and recording medium storing program for performing the method thereof
JP3691725B2 (en) Encoded image editing apparatus, encoded image editing method, and encoded image editing program recording medium
AU681324C (en) Method and apparatus for regenerating a dense motion vector field

Legal Events

Date Code Title Description
AS Assignment

Owner name: SAMSUNG ELECTRONICS CO., LTD, KOREA, REPUBLIC OF

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:LEE, TAMMY;HAN, WOO-JIN;MANU, MATHEW;AND OTHERS;REEL/FRAME:019293/0149

Effective date: 20070426

REMI Maintenance fee reminder mailed
LAPS Lapse for failure to pay maintenance fees
STCH Information on status: patent discontinuation

Free format text: PATENT EXPIRED DUE TO NONPAYMENT OF MAINTENANCE FEES UNDER 37 CFR 1.362

FP Lapsed due to failure to pay maintenance fee

Effective date: 20170212