WO2012020800A1 - 映像符号化方法,映像復号方法,映像符号化装置,映像復号装置およびそれらのプログラム - Google Patents
映像符号化方法,映像復号方法,映像符号化装置,映像復号装置およびそれらのプログラム Download PDFInfo
- Publication number
- WO2012020800A1 WO2012020800A1 PCT/JP2011/068286 JP2011068286W WO2012020800A1 WO 2012020800 A1 WO2012020800 A1 WO 2012020800A1 JP 2011068286 W JP2011068286 W JP 2011068286W WO 2012020800 A1 WO2012020800 A1 WO 2012020800A1
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- edge
- interpolation filter
- reference image
- unit
- region
- Prior art date
Links
Images
Classifications
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/50—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/10—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
- H04N19/102—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the element, parameter or selection affected or controlled by the adaptive coding
- H04N19/117—Filters, e.g. for pre-processing or post-processing
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/10—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
- H04N19/102—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the element, parameter or selection affected or controlled by the adaptive coding
- H04N19/119—Adaptive subdivision aspects, e.g. subdivision of a picture into rectangular or non-rectangular coding blocks
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/10—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
- H04N19/134—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the element, parameter or criterion affecting or controlling the adaptive coding
- H04N19/136—Incoming video signal characteristics or properties
- H04N19/14—Coding unit complexity, e.g. amount of activity or edge presence estimation
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/10—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
- H04N19/169—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding
- H04N19/17—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object
- H04N19/176—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object the region being a block, e.g. a macroblock
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/50—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding
- H04N19/503—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding involving temporal prediction
- H04N19/51—Motion estimation or motion compensation
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/50—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding
- H04N19/503—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding involving temporal prediction
- H04N19/51—Motion estimation or motion compensation
- H04N19/523—Motion estimation or motion compensation with sub-pixel accuracy
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/50—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding
- H04N19/59—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding involving spatial sub-sampling or interpolation, e.g. alteration of picture size or resolution
Definitions
- the present invention relates to a video encoding / decoding technique for improving performance of an adaptive interpolation filter in video encoding using motion compensation.
- inter-frame prediction (motion compensation) coding in which prediction is performed between different screens, a motion vector is obtained so as to minimize prediction error energy etc. by referring to already decoded frames.
- the residual signal is orthogonally transformed, quantized, entropy coded, and converted into binary data.
- a prediction method with higher prediction accuracy is required, and reduction of prediction error energy is indispensable.
- 16 ⁇ 16 and 8 ⁇ 8 16 ⁇ 8, 8 ⁇ 16, 8 ⁇ 4, 4 ⁇ 8, 4 ⁇ 4, etc.
- the block size can be divided finely. This tool is called variable block size prediction.
- a 1 ⁇ 2 precision pixel is interpolated from the integer precision pixel of the reference frame using a 6-tap filter, and a 1 ⁇ 4 precision pixel is generated by linear interpolation using the pixel. To do.
- prediction is applied to non-integer precision motion. This tool is called 1/4 pixel accuracy prediction.
- H. ITU / IEC “MPEG” International Organization / for Standardization / International Electrotechnical Commission “Moving Picture Experts Group”
- T “VCEG” International Telecommunication Union-Telecommunication Standardization Sector “Video Coding Experts Group”.
- KTA Key Technical Area
- an adaptive interpolation filter a tool that adaptively changes the interpolation filter coefficient of decimal precision pixels
- an adaptive interpolation filter which is effective for almost all images and was first adopted in the KTA software.
- JCT-VC Joint Collaborative Team On Video Coding
- MPEG-4 Part 2 generates pixels with 1 ⁇ 2 pixel accuracy using an average filter, but Advanced Simple Profile (ASP) also supports motion compensation with 1 ⁇ 4 pixel accuracy.
- the position of 1/2 pixel is calculated using a one-dimensional 8-tap filter as shown in FIG. 15B. Thereafter, the 1/4 pixel position is derived by an average value filter.
- the filter coefficients are [(1, -5, 20, 20, -5, 1) / 32], respectively.
- the 1 ⁇ 4 pixel position is interpolated using an average value filter of [1/2, 1/2]. Since it is necessary to interpolate all 1 ⁇ 2 pixel positions once, calculation complexity is high, but high-performance interpolation is possible, leading to improved coding efficiency. Details of the above-described fixed interpolation filter are shown in Non-Patent Document 1, Non-Patent Document 2, and Non-Patent Document 3.
- the filter coefficient value is constant regardless of input image conditions (sequence type / image size / frame rate) and encoding conditions (block size / GOP (Group of Pictures) structure / QP (Quantization Parameter)). is there.
- time-varying effects such as aliasing, quantization error, error due to motion estimation, and camera noise are not considered. Therefore, it is considered that there is a limit to the performance improvement in terms of coding efficiency. Therefore, a method of adaptively changing the interpolation filter coefficient has been proposed in Non-Patent Document 4, and is called a non-separable adaptive interpolation filter.
- Non-Patent Document 4 considers a two-dimensional interpolation filter (6 ⁇ 6 total 36 filter coefficients), and the filter coefficient is determined so as to minimize the prediction error energy. H. The encoding efficiency was higher than that of the one-dimensional 6-tap fixed interpolation filter used in H.264 / AVC, but the calculation complexity for obtaining the filter coefficient is very high.
- Non-Patent Document 5 introduces a proposal for reducing the above.
- Non-Patent Document 5 The method introduced in Non-Patent Document 5 is called a separable adaptive interpolation filter (SAIF), and uses a one-dimensional six-tap interpolation filter instead of using a two-dimensional interpolation filter. .
- SAIF separable adaptive interpolation filter
- 16A to 16C are diagrams illustrating a non-integer precision pixel interpolation method in the separable adaptive interpolation filter (SAIF).
- SAIF separable adaptive interpolation filter
- S is an original image
- P is a decoded reference image
- x and y are horizontal and vertical positions in the image, respectively.
- ⁇ y y + MV y
- MV y represents the vertical component of the motion vector.
- w ci represents a horizontal filter coefficient group c i (0 ⁇ c i ⁇ 6) to be obtained.
- the interpolation process in the vertical direction is performed as shown in Step 2 of FIG. 16C.
- the filter coefficient in the vertical direction is determined by solving a linear problem similar to that in the horizontal direction. Specifically, the vertical filter coefficient that minimizes the prediction error energy function E v 2 of Expression (2) is analytically determined.
- S is an original image
- ⁇ P ( ⁇ is a symbol on P) is an image subjected to interpolation processing in the horizontal direction after decoding
- x and y indicate horizontal and vertical positions in the image, respectively.
- ⁇ x 4 ⁇ (x + MV x ) is expressed
- MV x indicates a horizontal component of the rounded motion vector.
- w cj represents a vertical filter coefficient group c j (0 ⁇ c j ⁇ 6) to be obtained.
- the minimization process is performed independently for each decimal precision pixel, and 12 types of 6-tap filter coefficient groups are obtained. The remaining decimal precision pixels are interpolated using this filter coefficient.
- the filter coefficient to be transmitted for each frame is 51 (15 in the horizontal direction and 36 in the vertical direction).
- the unit of prediction error energy minimization processing is fixed in a frame.
- 51 filter coefficients are determined for one frame. If the encoding target frame is divided into two types (or a plurality of types) of texture areas, the optimum filter coefficient is a coefficient group that takes into account both (all textures). In the situation where a characteristic filter coefficient originally appears only in the vertical direction in the area A, but in the situation where the filter coefficient can be obtained only in the horizontal direction in the area B, the filter coefficient is derived in the form of averaging both.
- the prediction error energy can be reduced by preparing and switching a plurality of filter coefficient groups according to the local properties of the image without being limited to one filter coefficient group (51 coefficients) per frame.
- a method for improving the coding efficiency is proposed in Non-Patent Document 6.
- Non-Patent Document 6 considers a method using a plurality of optimized filter coefficient groups by region division for one frame.
- Non-Patent Document 6 employs motion vectors (horizontal and vertical components, orientation) and spatial coordinates (macroblock position, x-coordinate and y-coordinate of the block) as a method of area division, and considers various image properties. Area division.
- FIG. 18 shows a configuration example of a video encoding apparatus using a conventional area division type adaptive interpolation filter as shown in Non-Patent Document 6.
- the area dividing unit 101 divides the encoding target frame of the input video signal into a plurality of areas including a plurality of blocks which are units for switching interpolation filter coefficients adaptively.
- the interpolation filter coefficient switching unit 102 switches the interpolation filter coefficient of the decimal accuracy pixel used for the reference image in the predictive coding for each region divided by the region dividing unit 101.
- the filter coefficient optimized by the filter coefficient optimization unit 1021 is used as the interpolation filter coefficient to be switched.
- the filter coefficient optimization unit 1021 calculates an interpolation filter coefficient that minimizes the prediction error energy between the original image and the interpolated reference image for each region.
- the prediction signal generation unit 103 includes a reference image interpolation unit 1031 and a motion detection unit 1032.
- the reference image interpolation unit 1031 applies an interpolation filter based on the interpolation filter coefficient selected by the interpolation filter coefficient switching unit 102 to the decoded reference image stored in the reference image memory 107.
- the motion detection unit 1032 calculates a motion vector by performing motion search on the interpolated reference image.
- the prediction signal generation unit 103 generates a prediction signal by motion compensation using a motion vector with decimal precision calculated by the motion detection unit 1032.
- the prediction encoding unit 104 calculates a residual signal between the input video signal and the prediction signal, orthogonally transforms the residual signal, and performs prediction encoding by quantization of a transform coefficient.
- the decoding unit 106 decodes the result of predictive coding, and stores the decoded signal in the reference image memory 107 for later predictive coding.
- variable length coding unit 105 performs variable length coding on the quantized transform coefficients and motion vectors, and variable length codes the interpolation filter coefficients selected by the interpolation filter coefficient switching unit 102 for each region, and codes them. Output as a bitstream.
- FIG. 19 shows a configuration example of a video decoding apparatus using a conventional region division type adaptive interpolation filter.
- the stream encoded by the video encoding device 100 shown in FIG. 18 is decoded by the video decoding device 200 shown in FIG.
- the variable length decoding unit 201 receives an encoded bit stream and decodes a quantized transform coefficient, a motion vector, an interpolation filter coefficient group, and the like.
- the region determination unit 202 determines a region serving as a unit for switching interpolation filter coefficients adaptively for a decoding target frame.
- the interpolation filter coefficient switching unit 203 switches the interpolation filter coefficient decoded by the variable length decoding unit 201 for each region determined by the region determination unit 202.
- the reference image interpolation unit 2041 in the prediction signal generation unit 204 applies the interpolation filter based on the interpolation filter coefficient received from the interpolation filter coefficient switching unit 203 to the decoded reference image stored in the reference image memory 206, and calculates the decimal number of the reference image. Restore precision pixels.
- the prediction signal generation unit 204 generates a prediction signal of the decoding target block from the reference image in which the decimal precision pixels are restored.
- the predictive decoding unit 205 performs inverse quantization, inverse orthogonal transform, and the like of the quantized coefficient decoded by the variable length decoding unit 201, and the prediction residual signal calculated thereby and the prediction signal generated by the prediction signal generating unit 204 Are added together to generate a decoded signal and output as a decoded image.
- the decoded signal decoded by the predictive decoding unit 205 is stored in the reference image memory 206 for later predictive decoding.
- Hiroshi Harashima, Yoshinori Sakai, Toshiyuki Yoshida “Video coding”, Ohmsha, pp.135-136, 2001 Satoshi Okubo, Takuya Kakuno, Yoshihiro Kikuchi, Teruhiko Suzuki: Satoshi “H.264 / AVC textbook revised third edition”, Impress, pp.119-123, 2009 I. E. G. Richardson, G. J. Sullivan: “H.264 and MPEG-4 VIDEO COMPRESSION”, WILEY, pp.172-175, 2003 Y. Vatis, B. Edler, D. T. Nguyen, J.
- the region division type adaptive interpolation filter (Non-Patent Document 6) used by the video encoding device 100 as shown in FIG. 18 considers local properties of an image and switches a plurality of filter coefficient groups within a frame. It aims to improve the coding efficiency by reducing the prediction error energy. However, attention is paid to motion vectors and spatial coordinates in performing region division, and image texture components are not taken into consideration.
- the edge information of the image is considered to be one of the important indexes indicating the locality of the image. Therefore, it is considered that the encoding efficiency can be further improved by adopting it in the region segmentation method.
- the present invention improves coding efficiency by optimizing filter coefficients and further reducing residual energy by segmenting the edge of the reference image as a feature quantity. For the purpose.
- the edge component is calculated from the reference image and the region division is performed.
- the texture information of the image can be used for region division, and the encoding efficiency can be improved by reducing the prediction error energy.
- the present invention is a video coding method using motion compensation, which acquires a reference image indicated by a motion vector and calculates edge information from the obtained reference image. A region is divided in the frame from the calculated edge information, and the interpolation filter coefficient of the decimal precision pixel is optimized in each region. Thereafter, encoding is performed by interpolating decimal precision pixels using the optimized interpolation filter.
- the present invention is also a video decoding method for decoding an encoded stream encoded by the video encoding method, wherein a reference image obtained from a decoded motion vector by decoding interpolation filter coefficients of decimal precision pixels is decoded. Edge information is calculated from The region is determined using the calculated edge information, and the interpolation filter for the decimal precision pixel is switched for each region, the decimal precision pixel is interpolated, and decoding is performed.
- the operation of the present invention is as follows.
- the conventional area segmentation type adaptive interpolation filter employs an area segmentation method that uses motion vectors and spatial coordinates, and cannot perform segmentation using image texture information, and can fully consider the locality of the image. There wasn't.
- the influence of the texture of the image can be absorbed, the region segmentation function can be made flexible, and the encoding efficiency can be improved.
- the present invention it is possible to perform region division using texture (edge) information, which cannot be taken into consideration by the conventional separable adaptive interpolation filter, and to provide a wide range of region division types.
- the coding efficiency can be improved by extending the flexibility of locality and reducing the prediction error energy.
- SAIF separation-type adaptive interpolation filter
- region division type adaptive interpolation filter It is a figure which shows the comparison of the conventional adaptive interpolation filter and an area
- FIG. 1 is a diagram illustrating a configuration example of a video encoding device according to the present invention.
- the video encoding device 10 is a device that performs interpolation by fractional precision motion compensation by interpolating decimal precision pixels using an area division type adaptive interpolation filter that uses edge information, and is a unit for switching adaptive interpolation filters. This is different from the conventional video encoding apparatus 100 shown in FIG. 18 in that the region to be divided is performed using edge information detected from the reference image.
- the area dividing unit 11 uses a plurality of blocks as units for adaptively switching interpolation filter coefficients using the edge information calculated by the edge calculating unit 12 for the encoding target frame of the input video signal. Is divided into a plurality of regions.
- the edge calculation unit 12 performs processing for extracting edge information from the reference image at the time of motion compensation. Details will be described later.
- the interpolation filter coefficient switching unit 13 switches the interpolation filter coefficient of the decimal precision pixel used for the reference image in the predictive coding for each region divided by the region dividing unit 11.
- a filter coefficient optimized by the filter coefficient optimization unit 131 is used as the interpolation filter coefficient to be switched.
- the filter coefficient optimization unit 131 calculates an interpolation filter coefficient that minimizes the prediction error energy between the original image and the interpolated reference image for each region.
- the prediction signal generation unit 14 includes a reference image interpolation unit 141 and a motion detection unit 142.
- the reference image interpolation unit 141 applies an interpolation filter based on the interpolation filter coefficient selected by the interpolation filter coefficient switching unit 13 to the decoded reference image stored in the reference image memory 18.
- the motion detection unit 142 calculates a motion vector MV by performing a motion search on the interpolated reference image.
- the prediction signal generation unit 14 generates a prediction signal by motion compensation using a decimal precision motion vector calculated by the motion detection unit 142.
- the predictive encoding unit 15 calculates a residual signal between the input video signal and the predictive signal, orthogonally transforms it, and performs predictive encoding by quantizing the transform coefficient.
- the decoding unit 17 decodes the result of predictive encoding and stores the decoded signal in the reference image memory 18 for later predictive encoding.
- variable length coding unit 16 performs variable length coding on the quantized transform coefficients and motion vectors, and variable length codes the interpolation filter coefficients selected by the interpolation filter coefficient switching unit 13 for each region, and codes them. Output as a bitstream.
- FIG. 2 is a flowchart of video encoding processing executed by the video encoding device 10.
- processing is described assuming luminance signal processing.
- region segmentation is performed using the edge information of the image described in this example, and interpolation filter coefficients are optimized for each region.
- the encoding function can be applied not only to the luminance signal but also to the color difference signal.
- step S101 a frame to be encoded is input.
- step S102 the input frame is divided into blocks (for example, a conventional motion prediction block size such as 16 ⁇ 16 or 8 ⁇ 8), and the motion detector 142 calculates an optimal motion vector for each block.
- blocks for example, a conventional motion prediction block size such as 16 ⁇ 16 or 8 ⁇ 8
- the motion detector 142 calculates an optimal motion vector for each block.
- the conventional H.264 standard is used for interpolation of decimal precision pixels in the reference image in step S102.
- a fixed 6-tap filter based on H.264 / AVC is used.
- the edge calculation unit 12 acquires a reference image indicated by the motion vector using the motion vector obtained in step S102, and calculates edge information.
- a generally known edge detection method for example, a Sobel filter
- the region dividing unit 11 performs region determination for each block using the edge information obtained in step S103, and performs region division. For example, if the horizontal component of the edge is large, the region is divided into two types, such as R a1 , and if the vertical component is large, R a2 , or the edge size is also used as a judgment material. R b1 , R b2 if the edge size is greater than or equal to the threshold T and the horizontal component is large, R b2 if the edge size is greater than or equal to the threshold T and the vertical component is large, and R b3. You can also The number of area divisions can be set to any number depending on the edge definition.
- step S105 based on the result of region division in step S104, the filter coefficient optimizing unit 131 performs interpolation filter coefficient optimization processing for each region.
- the optimization process of the interpolation filter coefficient is performed for each decimal precision pixel in the horizontal direction using Expression (3) which is a prediction error energy function.
- ⁇ n indicates each region, n is a region number, S is an original image, P is a decoded reference image, and x and y are horizontal and vertical positions in the image, respectively.
- ⁇ y y + MV y , where MV y represents the vertical component of the motion vector.
- w ci represents a horizontal filter coefficient group c i (0 ⁇ c i ⁇ 6) to be obtained.
- step S106 using the horizontal direction interpolation filter coefficient obtained in step S105, the horizontal fractional pixel interpolation (a, b, in FIGS. 16A to 16C is performed independently for each region in the frame. c)).
- step S107 optimization processing of the interpolation filter coefficient in the vertical direction is performed.
- the optimization process of the interpolation filter coefficient for each decimal pixel in the vertical direction is performed using Expression (4) which is a prediction error energy function in the vertical direction.
- w cj represents a vertical filter coefficient group c j (0 ⁇ c j ⁇ 6) to be obtained.
- step S108 by using the vertical direction interpolation filter coefficient obtained in step S107, the vertical fractional pixel interpolation (d to o in FIG. 16A or FIG. 16C) is performed independently for each region in the frame. Interpolation).
- step S109 the motion detection unit 142 calculates the motion vector again using the image interpolated in the vertical direction obtained in step S108 as a reference image.
- step S110 a rate distortion cost (RD cost) is calculated and stored.
- step S111 the optimum interpolation filter coefficient obtained in steps S105 and S107 is sent to the variable length encoding unit 16, and the variable length encoding unit 16 encodes the interpolation filter coefficient.
- step S112 the variable-length encoding unit 16 stores the quantized coefficient value output from the predictive encoding unit 15, the motion vector (MV) output from the motion detecting unit 142, and other information to be encoded. Encode and output a bitstream.
- step S113 the processing is ended.
- the procedure for interpolating the vertical direction after interpolating the horizontal direction has been described.
- the interpolation of the horizontal direction after interpolating the vertical direction can also be processed by the same procedure.
- FIG. 3 is a detailed process flowchart of step S103 (edge calculation process) described in FIG.
- FIG. 4 is a diagram illustrating a specific example of the edge calculation process.
- step S201 it performs the entire target block B i of the horizontal filter f x, to compute the matrix M x.
- the block B i by applying a horizontal filter f x for calculating the difference of pixel values adjacent to the left and right, the matrix M x is derived.
- step S202 computing the sum S x of the respective components of the matrix M x.
- S x 39 is obtained.
- step S203 it performs the entire target block B i of the vertical filter f y, calculate the matrix M y.
- the block B i by applying a vertical filter f y for calculating the difference of pixel values adjacent above and below, the matrix M y is derived.
- step S204 computing the sum S y of the respective components of the matrix M y.
- step S206 the edge angle calculated in step S205, the block B i, allocate an area number indicating belongs to which region.
- FIG. 5 is a detailed flowchart of step S206 (area number allocation processing) described in FIG.
- step S301 the edge angle E is input and the area number table is inquired.
- step S302 the area number is determined from the inquiry result of the area number table.
- FIG. 6B shows an example of an area number table when the number of areas is two.
- the number of areas is 2
- the edge is horizontal, it is assigned to “area 1”
- the edge is vertical, it is assigned to “area 2”. That is, as shown in FIG. 6B, in the region number table, the region number is 1 (region 1) when ⁇ 1 ⁇ E ⁇ 1, and the region number is other than E ⁇ 1 or 1 ⁇ E.
- Definition information 2 (area 2) is stored.
- the block B i is determined to be “region 2”.
- FIG. 7B shows an example of an area number table when the number of areas is four.
- the number of areas is 4, for example, as shown in FIG. 7A, the areas are classified into four areas “area 1”, “area 2”, “area 3”, and “area 4” according to the edge angle. That is, as shown in FIG. 7B, the region number table shows that the region number is 1 (region 1) when ⁇ 0.414 ⁇ E ⁇ 0.414, and the region number when 0.414 ⁇ E ⁇ 2.414. When the number is 2 (region 2), 2.414 ⁇ E or E ⁇ 2.414, the region number is 3 (region 3), and when ⁇ 2.414 ⁇ E ⁇ 0.414, the region number is 4 Definition information (area 4) is stored. In the example of FIG. 4, since E ⁇ 1.128, the block B i is determined to be “region 2”.
- FIG. 8 is a detailed process flowchart of the second example of step S103 (edge calculation process) described in FIG.
- FIG. 9 is a diagram illustrating a specific example of the edge calculation process in consideration of the edge strength.
- step S403 by using the edge horizontal component S x and an edge vertical component S y calculated in step S401, it calculates an edge strength Z.
- the edge strength Z is calculated by the following equation, for example.
- step S404 allocates an edge angle E calculated in step S402, from the edge strength Z calculated in step S403, the block B i, an area number indicating belongs to which region.
- FIG. 10 is a detailed flowchart of step S404 (area number allocation processing) described in FIG.
- step S501 the edge angle E and the edge strength Z are input, and the area number table is inquired.
- step S502 the area number is determined from the inquiry result of the area number table.
- FIG. 11B shows an example of an area number table when the number of areas is three.
- the region number is “region 1” if the edge is strong and horizontal, “region 2” if the edge is strong and vertical, and “region 3” if the edge is weak. Is allocated.
- the threshold value Z t can be set arbitrarily.
- the region number table shows that when ⁇ 1 ⁇ E ⁇ 1 and Z t ⁇ Z, the region number is 1 (region 1), E ⁇ 1 or 1 ⁇ E, and Z t ⁇ Z. region number is 2 (region 2), the definition information that area number 3 (region 3) when Z ⁇ Z t are stored at.
- FIG. 12B shows an example of an area number table when the number of areas is five.
- the region number table shows that the region number is 1 (region 1) and 0.414 ⁇ E ⁇ 2.
- the region number is 2 (region 2), 2.414 ⁇ E or E ⁇ 2.414 and when Z t ⁇ Z, the region number is 3 (region 3), ⁇ 2.
- the threshold value Z t of the edge strength has been described as being arbitrarily set here. However, when the video encoding device 10 determines according to the feature amount or the encoding state of the image, the threshold value Z t is encoded and encoded. It may be included in the bit stream and transmitted to the video decoding device. Also may already determine the threshold Z t from the decoded image decoded by encoding, in this case, it is possible to calculate the threshold value Z t from even the decoded image in the video decoding apparatus, and coding the threshold Z t There is no need to transmit.
- the threshold value Z t used in the encoding of the 10-th frame as already mean value of the edge strength of all blocks in the ninth frame image from the first frame to be encoded, decoded.
- the threshold value Z t used in the encoding of the 10-th frame may be set to 2/5.
- FIG. 13 is a diagram showing a configuration example of a video decoding apparatus according to the present invention.
- the video decoding apparatus 20 receives the bit stream encoded by the video encoding apparatus 10 shown in FIG. 1, performs interpolation of decimal precision pixels using a domain division type adaptive interpolation filter using edge information, and performs decimal precision. Decoding by motion compensation. It differs from the conventional video decoding apparatus 200 shown in FIG. 19 in that a region is determined using edge information detected from a reference image, and an adaptive interpolation filter is switched to interpolate decimal precision pixels.
- the variable length decoding unit 21 receives the encoded bit stream and decodes a quantized transform coefficient, a motion vector, a group of interpolation filter coefficients, and the like.
- the edge calculation unit 22 performs processing for extracting edge information such as an edge angle or edge strength from a reference image at the time of motion compensation.
- the region determination unit 23 uses the edge information calculated by the edge calculation unit 22 to determine a region serving as a unit for switching the interpolation filter coefficient adaptively for the decoding target frame.
- the interpolation filter coefficient switching unit 24 switches the interpolation filter coefficient decoded by the variable length decoding unit 21 for each region determined by the region determination unit 23.
- the reference image interpolation unit 251 in the prediction signal generation unit 25 applies the interpolation filter based on the interpolation filter coefficient received from the interpolation filter coefficient switching unit 24 to the decoded reference image stored in the reference image memory 27, and calculates the decimal number of the reference image. Restore precision pixels.
- the prediction signal generation unit 25 generates a prediction signal of the decoding target block from the reference image in which the decimal precision pixels are restored.
- the prediction decoding unit 26 performs inverse quantization, inverse orthogonal transformation, and the like on the quantization coefficient decoded by the variable length decoding unit 21, and the prediction error signal calculated thereby, the prediction signal generated by the prediction signal generation unit 25, and Are added together to generate a decoded signal and output as a decoded image.
- the decoded signal decoded by the predictive decoding unit 26 is stored in the reference image memory 27 for subsequent predictive decoding.
- FIG. 14 is a flowchart of video decoding processing executed by the video decoding device 20.
- processing with a luminance signal is assumed unless otherwise specified.
- the present invention can be applied not only to a luminance signal but also to a color difference signal.
- step S601 the variable length decoding unit 21 acquires frame header information from the input bit stream.
- step S602 the interpolation filter coefficient used in each region is decoded.
- step S603 various information such as a motion vector (MV) is decoded.
- MV motion vector
- step S604 the edge calculation unit 22 acquires a reference image using the motion vector decoded in step S603, and calculates edge information from the reference image.
- the calculation of the edge information is the same as the edge calculation process described with reference to FIG. 3 or FIG.
- step S605 the region determination unit 23 determines which region the decoding target block belongs to in block units from the edge information obtained in step S604, and acquires region information (region number). To do.
- step S606 the interpolation filter coefficient switching unit 24 selects the optimum interpolation filter coefficient decoded in step S602 from the region number obtained in step S605, and the reference image interpolation unit 251 determines the decimal number of the reference image.
- the accuracy pixel is restored, and after the decimal accuracy pixel is restored, the prediction signal generation unit 25 generates a prediction signal of the decoding target block using the motion vector decoded in step S603.
- step S607 the variable length decoding unit 21 decodes the prediction residual signal of the decoding target block from the input bit stream.
- step S608 the prediction decoding unit 26 adds the prediction signal obtained in step S606 and the prediction residual signal obtained in step S607 to generate a decoded signal.
- the generated decoded signal is output as a decoded image and stored in the reference image memory 27.
- steps S601 to S608 are repeated until the decoding of all the frames is completed.
- the process is terminated (step S609).
- the procedure for interpolating the vertical direction after interpolating the horizontal direction has been described.
- the interpolation of the horizontal direction after interpolating the vertical direction can also be processed by the same procedure.
- the above video encoding and video decoding processes can be realized by a computer and a software program, and the program can be recorded on a computer-readable recording medium or provided through a network.
- the present invention can be used for, for example, video encoding and video decoding using motion compensation. According to the present invention, it is possible to perform region division using texture (edge) information, which cannot be taken into consideration with a conventional separable adaptive interpolation filter, to give a wide range of types of region division, and to be flexible with respect to locality within the screen. The encoding efficiency can be improved by reducing the prediction error energy.
Landscapes
- Engineering & Computer Science (AREA)
- Multimedia (AREA)
- Signal Processing (AREA)
- Compression Or Coding Systems Of Tv Signals (AREA)
Abstract
Description
本願は,2010年8月12日に日本へ出願された特願2010-180813号に基づき優先権を主張し,その内容をここに援用する。
過去の映像符号化標準方式MPEG-1/2では,図15Aに示されるように,1/2精度の画素を補間するために,両隣の2点の整数精度画素(単に整数画素ともいう)から加算平均を用いて補間画素を生成していた。すなわち,2点の整数画素に対して,[1/2,1/2]の平均値フィルタを施していることになる。非常に単純な処理のため,計算複雑度の観点からみると効果的であるが,1/4精度の画素を求める上ではフィルタの性能としては高くはない。
H.264/AVCでは,入力画像条件(シーケンス種類/画像サイズ/フレームレート)や符号化条件(ブロックサイズ/GOP(Group of Pictures)構造/QP(Quantization Parameter))に関わらず,フィルタ係数値は一定である。フィルタ係数値が固定である場合,例えば,エイリアシング,量子化誤差,動き推定による誤差,カメラノイズといった時間的に変化する効果が考慮されていない。したがって,符号化効率の点で性能向上に限界があると考えられる。そこで,補間フィルタ係数を適応的に変化させる方式が,非特許文献4では提案されており,非分離型の適応補間フィルタと呼ばれている。
図1は,本発明に係る映像符号化装置の構成例を示す図である。映像符号化装置10は,エッジ情報を利用する領域分割型適応補間フィルタを用いて,小数精度画素の補間を行い,小数精度の動き補償によって符号化を行う装置であり,適応補間フィルタを切り替える単位となる領域の分割を,参照画像から検出したエッジ情報を用いて行う点が,図18に示す従来の映像符号化装置100と異なる。
図2は,映像符号化装置10が実行する映像符号化処理のフローチャートである。以下では,特に断りがない限り,輝度信号での処理を仮定して説明するが,本例で述べている画像のエッジ情報を用いて領域分割を行い,領域単位で補間フィルタ係数を最適化して符号化する機能は,輝度信号だけでなく,色差信号にも適用可能である。
次に,図3および図4に従って,エッジ算出処理の詳細を説明する。図3は,図2で説明したステップS103(エッジ算出処理)の詳細な処理フローチャートである。図4は,エッジ算出処理の具体例を示す図である。
図5から図7Bに従って,領域番号の割り振り処理の詳細を説明する。図5は,図3で説明したステップS206(領域番号の割り振り処理)の詳細なフローチャートである。
エッジ算出処理の他の例について説明する。この例では,エッジ角度の他に,エッジ強度を算出し,エッジ角度とエッジ強度とから領域番号の割り振りを行う。図8は,図2で説明したステップS103(エッジ算出処理)の第2の例の詳細な処理フローチャートである。図9は,エッジ強度を考慮したエッジ算出処理の具体例を示す図である。
Sx =39およびSy =44の場合,図9の例では,Z≒58.80となる。なお,エッジ強度は,上の式に限らず,他の式によって定義してもよい。例えば,演算コストを重視する場合には,
Z=|Sx|+|Sy |
というように,エッジ強度をエッジ水平成分Sx とエッジ垂直成分Sy の絶対値の和として定義してもよい。
図10から図12Bに従って,第2の例における領域番号の割り振り処理の詳細を説明する。図10は,図8で説明したステップS404(領域番号の割り振り処理)の詳細なフローチャートである。
図13は,本発明に係る映像復号装置の構成例を示す図である。映像復号装置20は,図1に示す映像符号化装置10が符号化したビットストリームを入力し,エッジ情報を利用する領域分割型適応補間フィルタを用いて,小数精度画素の補間を行い,小数精度の動き補償によって復号する。参照画像から検出したエッジ情報を用いて領域を判定し,適応補間フィルタを切り替えて小数精度画素の補間を行う点が,図19に示す従来の映像復号装置200と異なる。
図14は,映像復号装置20が実行する映像復号処理のフローチャートである。以下では,特に断りがない限り,輝度信号での処理を仮定して説明するが,輝度信号だけでなく,色差信号にも適用可能である。
11 領域分割部
12 エッジ計算部
13 補間フィルタ係数切り替え部
131 フィルタ係数最適化部
14 予測信号生成部
141 参照画像補間部
142 動き検出部
15 予測符号化部
16 可変長符号化部
17 復号部
18 参照画像メモリ
20 映像復号装置
21 可変長復号部
22 エッジ算出部
23 領域判定部
24 補間フィルタ係数切り替え部
25 予測信号生成部
251 参照画像補間部
26 予測復号部
27 参照画像メモリ
Claims (8)
- 小数精度の動き補償を用いる映像符号化方法において,
動きベクトルの指し示す参照画像データを取得し,得られた参照画像データからエッジ情報を算出するステップと,
前記エッジ情報をもとに符号化対象フレームを,適応的に補間フィルタを適用する単位となる複数の領域に分割するステップと,
前記領域単位で小数精度画素の補間フィルタを最適化するステップと,
最適化された補間フィルタを用いて参照画像に対する小数精度画素の補間を行い,小数精度の動き補償により予測符号化を行うステップとを有する
映像符号化方法。 - 請求項1記載の映像符号化方法において,
前記エッジ情報は,前記動きベクトルの指し示す前記参照画像データから抽出されたエッジ角度,または前記エッジ角度とエッジ強度であり,前記エッジ角度の範囲,または前記エッジ角度の範囲と前記エッジ強度の強弱によって,前記領域の分割を行う
映像符号化方法。 - 小数精度の動き補償を用いる映像復号方法において,
小数精度画素の補間フィルタ係数を復号するステップと,
復号された動きベクトルから得られる参照画像データからエッジ情報を算出するステップと,
適応的に補間フィルタを適用する単位となる複数の領域のうち,復号対象データがどの領域に属するかを前記エッジ情報を用いて判定するステップと,
前記領域ごとに小数精度画素の補間フィルタを切り替えて,参照画像に対する小数精度画素の補間を行い,小数精度の動き補償により予測復号を行うステップとを有する
映像復号方法。 - 請求項3記載の映像復号方法において,
前記エッジ情報は,前記動きベクトルの指し示す前記参照画像データから抽出されたエッジ角度,または前記エッジ角度とエッジ強度であり,前記エッジ角度の範囲,または前記エッジ角度の範囲と前記エッジ強度の強弱によって,前記領域の判定を行う
映像復号方法。 - 小数精度の動き補償を用いる映像符号化装置において,
動きベクトルの指し示す参照画像データを取得し,得られた参照画像データからエッジ情報を算出するエッジ計算部と,
前記エッジ情報をもとに符号化対象フレームを,適応的に補間フィルタを適用する単位となる複数の領域に分割する領域分割部と,
前記領域単位で小数精度画素の補間フィルタを最適化するフィルタ係数最適化部と,
最適化された補間フィルタを用いて参照画像に対する小数精度画素の補間を行う参照画像補間部と,
小数精度の動き補償により予測符号化を行う予測符号化部とを備える
映像符号化装置。 - 小数精度の動き補償を用いる映像復号装置において,
小数精度画素の補間フィルタ係数を復号する補間フィルタ係数復号部と,
復号された動きベクトルから得られる参照画像データからエッジ情報を算出するエッジ算出部と,
適応的に補間フィルタを適用する単位となる複数の領域のうち,復号対象データがどの領域に属するかを前記エッジ情報を用いて判定する領域判定部と,
前記領域ごとに小数精度画素の補間フィルタを切り替えて,参照画像に対する小数精度画素の補間を行う参照画像補間部と,
小数精度の動き補償により予測復号を行う予測復号部とを備える
映像復号装置。 - 請求項1または請求項2に記載の映像符号化方法を,コンピュータに実行させるための映像符号化プログラム。
- 請求項3または請求項4に記載の映像復号方法を,コンピュータに実行させるための映像復号プログラム。
Priority Applications (8)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
BR112013002733A BR112013002733A2 (pt) | 2010-08-12 | 2011-08-10 | método de codificação de vídeo, método de decodoficação de vídeo, aparelho de codificação de vídeo, aparelho de decodificação de vídeo e programas dos mesmos |
KR1020157035181A KR20160003869A (ko) | 2010-08-12 | 2011-08-10 | 영상 부호화 방법, 영상 복호 방법, 영상 부호화 장치, 영상 복호 장치 및 이들의 프로그램 |
PL11816464T PL2590410T3 (pl) | 2010-08-12 | 2011-08-10 | Sposób kodowania wideo, sposób dekodowania wideo, urządzenie do kodowania wideo, urządzenie do dekodowania wideo i programy do tego celu |
CA2807327A CA2807327C (en) | 2010-08-12 | 2011-08-10 | Video encoding method, video decoding method, video encoding apparatus, video decoding apparatus, and programs thereof |
US13/814,144 US9609318B2 (en) | 2010-08-12 | 2011-08-10 | Video encoding method, video decoding method, video encoding apparatus, video decoding apparatus, and programs thereof |
KR1020137001194A KR20130055631A (ko) | 2010-08-12 | 2011-08-10 | 영상 부호화 방법, 영상 복호 방법, 영상 부호화 장치, 영상 복호 장치 및 이들의 프로그램 |
CN201180035072.4A CN103069803B (zh) | 2010-08-12 | 2011-08-10 | 视频编码方法、视频解码方法、视频编码装置、视频解码装置 |
EP11816464.9A EP2590410B1 (en) | 2010-08-12 | 2011-08-10 | Video encoding method, video decoding method, video encoding device, video decoding device, and programs for same |
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
JP2010-180813 | 2010-08-12 | ||
JP2010180813A JP5524762B2 (ja) | 2010-08-12 | 2010-08-12 | 映像符号化方法,映像復号方法,映像符号化装置,映像復号装置およびそれらのプログラム |
Publications (1)
Publication Number | Publication Date |
---|---|
WO2012020800A1 true WO2012020800A1 (ja) | 2012-02-16 |
Family
ID=45567762
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
PCT/JP2011/068286 WO2012020800A1 (ja) | 2010-08-12 | 2011-08-10 | 映像符号化方法,映像復号方法,映像符号化装置,映像復号装置およびそれらのプログラム |
Country Status (10)
Country | Link |
---|---|
US (1) | US9609318B2 (ja) |
EP (1) | EP2590410B1 (ja) |
JP (1) | JP5524762B2 (ja) |
KR (2) | KR20130055631A (ja) |
CN (1) | CN103069803B (ja) |
BR (1) | BR112013002733A2 (ja) |
CA (1) | CA2807327C (ja) |
PL (1) | PL2590410T3 (ja) |
TW (1) | TWI504236B (ja) |
WO (1) | WO2012020800A1 (ja) |
Families Citing this family (8)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
TWI508569B (zh) * | 2012-09-14 | 2015-11-11 | Realtek Semiconductor Corp | 行動高畫質連結資料轉換器以及行動高畫質連結資料轉換方法 |
WO2014078068A1 (en) * | 2012-11-13 | 2014-05-22 | Intel Corporation | Content adaptive transform coding for next generation video |
JP6230748B2 (ja) * | 2015-02-26 | 2017-11-15 | 三菱電機株式会社 | 画像処理装置 |
KR102437109B1 (ko) | 2015-11-11 | 2022-08-26 | 삼성전자주식회사 | 비디오 복호화 방법 및 그 장치 및 비디오 부호화 방법 및 그 장치 |
US11044490B2 (en) * | 2017-09-28 | 2021-06-22 | Sharp Kabushiki Kaisha | Motion compensation filter apparatus, image decoding apparatus, and video coding apparatus |
JP2019213096A (ja) * | 2018-06-06 | 2019-12-12 | Kddi株式会社 | 画像復号装置、画像符号化装置、画像処理システム、画像復号方法及びプログラム |
US11350108B2 (en) * | 2019-03-18 | 2022-05-31 | Tencent America LLC | Affine inter prediction refinement with optical flow |
KR102374261B1 (ko) | 2021-11-30 | 2022-03-15 | 주식회사 코메스타 | 임무용 면허대역에서 채널 감지 및 운용 주파수 선택 방법 및 이를 실행하는 시스템 |
Citations (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP2003319398A (ja) * | 2002-04-24 | 2003-11-07 | Nec Corp | 動画像の符号化方法および復号化方法、これを用いた装置とプログラム |
WO2008069073A1 (ja) * | 2006-11-30 | 2008-06-12 | Ntt Docomo, Inc. | 動画像符号化装置、動画像符号化方法、動画像符号化プログラム、動画像復号化装置、動画像復号化方法及び動画像復号化プログラム |
WO2010021108A1 (ja) * | 2008-08-18 | 2010-02-25 | パナソニック株式会社 | 補間フィルタリング方法、画像符号化方法、画像復号方法、補間フィルタリング装置、プログラム、及び集積回路 |
Family Cites Families (12)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
EP0817121A3 (en) | 1996-06-06 | 1999-12-22 | Matsushita Electric Industrial Co., Ltd. | Image coding method and system |
JP2008054267A (ja) | 2006-07-28 | 2008-03-06 | Hitachi Ltd | 画像処理装置、画像符号化装置及び画像復号化装置 |
WO2008053746A1 (fr) * | 2006-10-30 | 2008-05-08 | Nippon Telegraph And Telephone Corporation | Procédé de génération d'informations de référence prédictives, procédé de codage et de décodage d'image dynamiques, leur dispositif, leur programme et support de stockage contenant le programme |
KR101369746B1 (ko) * | 2007-01-22 | 2014-03-07 | 삼성전자주식회사 | 적응적 보간 필터를 이용한 영상 부호화, 복호화 방법 및장치 |
KR20080107965A (ko) * | 2007-06-08 | 2008-12-11 | 삼성전자주식회사 | 객체 경계 기반 파티션을 이용한 영상의 부호화, 복호화방법 및 장치 |
RU2496252C2 (ru) * | 2007-06-29 | 2013-10-20 | Шарп Кабусики Кайся | Устройство кодирования изображения, способ кодирования изображения, устройство декодирования изображения, способ декодирования изображения, программа и запоминающий носитель |
KR101291196B1 (ko) | 2008-01-25 | 2013-07-31 | 삼성전자주식회사 | 영상의 부호화, 복호화 방법 및 장치 |
JP5253835B2 (ja) * | 2008-02-19 | 2013-07-31 | 株式会社キーエンス | 画像生成装置、画像生成方法及びコンピュータプログラム |
EP2266321B1 (en) * | 2008-04-23 | 2017-12-13 | Telefonaktiebolaget LM Ericsson (publ) | Template-based pixel block processing |
JP5271031B2 (ja) * | 2008-08-09 | 2013-08-21 | 株式会社キーエンス | 画像のデータ圧縮方法、画像処理におけるパターンモデルの位置決め方法、画像処理装置、画像処理プログラム及びコンピュータで読み取り可能な記録媒体 |
KR101538704B1 (ko) * | 2009-01-28 | 2015-07-28 | 삼성전자주식회사 | 보간 필터를 적응적으로 사용하여 영상을 부호화 및 복호화하는 방법 및 장치 |
JP5485851B2 (ja) * | 2010-09-30 | 2014-05-07 | 日本電信電話株式会社 | 映像符号化方法,映像復号方法,映像符号化装置,映像復号装置およびそれらのプログラム |
-
2010
- 2010-08-12 JP JP2010180813A patent/JP5524762B2/ja active Active
-
2011
- 2011-08-10 KR KR1020137001194A patent/KR20130055631A/ko active Application Filing
- 2011-08-10 BR BR112013002733A patent/BR112013002733A2/pt not_active IP Right Cessation
- 2011-08-10 CN CN201180035072.4A patent/CN103069803B/zh active Active
- 2011-08-10 KR KR1020157035181A patent/KR20160003869A/ko not_active Application Discontinuation
- 2011-08-10 WO PCT/JP2011/068286 patent/WO2012020800A1/ja active Application Filing
- 2011-08-10 EP EP11816464.9A patent/EP2590410B1/en active Active
- 2011-08-10 CA CA2807327A patent/CA2807327C/en active Active
- 2011-08-10 US US13/814,144 patent/US9609318B2/en active Active
- 2011-08-10 PL PL11816464T patent/PL2590410T3/pl unknown
- 2011-08-12 TW TW100128829A patent/TWI504236B/zh active
Patent Citations (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP2003319398A (ja) * | 2002-04-24 | 2003-11-07 | Nec Corp | 動画像の符号化方法および復号化方法、これを用いた装置とプログラム |
WO2008069073A1 (ja) * | 2006-11-30 | 2008-06-12 | Ntt Docomo, Inc. | 動画像符号化装置、動画像符号化方法、動画像符号化プログラム、動画像復号化装置、動画像復号化方法及び動画像復号化プログラム |
WO2010021108A1 (ja) * | 2008-08-18 | 2010-02-25 | パナソニック株式会社 | 補間フィルタリング方法、画像符号化方法、画像復号方法、補間フィルタリング装置、プログラム、及び集積回路 |
Non-Patent Citations (10)
Title |
---|
E. G. RICHARDSON; G. J. SULLIVAN: "H.264 and MPEG-4 VIDEO COMPRESSION", 2003, WILEY, pages: 172 - 175 |
HIROSHI HARASHIMA; YOSHINORI SAKAI; TOSHIYUKI YOSHIDA: "Video Information Encoding", 2001, OHMSHA, LTD., pages: 135 - 136 |
KAI ZHANG ET AL.: "Single-Pass Encoding Using Multiple Adaptive Interpolation Filters", ITU - TELECOMMUNICATIONS STANDARDIZATION SECTOR STUDY GROUP 16 QUESTION 6 VIDEO CODING EXPERTS GROUP, 15 April 2009 (2009-04-15), XP030003678 * |
RONGGANG WANG ET AL.: "Partition-level adaptive interpolation filter for video coding, Multimedia and Expo, 2009.", IEEE INTERNATIONAL CONFERENCE, 3 July 2009 (2009-07-03), pages 157 - 160, XP031510716 * |
S. WITTMANN; T WEDI: "Separable adaptive interpolation filter for video coding", PROC. ICIP2008, IEEE INTERNATIONAL CONFERENCE ON IMAGE PROCESSING, October 2008 (2008-10-01), pages 2500 - 2503, XP031374548 |
SAKAE OKUBO; SHINYA KADONO; YOSHIHIRO KIKUCHI; TERUHIKO SUZUKI: "H.264/AVC Textbook: 3rd Revised Edition", 2009, IMPRESS R&D, pages: 119 - 123 |
See also references of EP2590410A4 |
SHOHEI MATSUO ET AL.: "Region-based adaptive interpolation filter", JCTVC-B051 (JOINT COLLABORATIVE TEAM ON VIDEO CODING (JCT-VC) OF ITU-T SG16 WP3 AND ISO/IEC JTC1/ SC29/WG11), JCTVC-B051, 16 July 2010 (2010-07-16), XP030007631 * |
SHOHEI MATSUO; SEISHI TAKAMURA; HIROHISA JOZAWA: "Separable adaptive interpolation filter with region dividing technique for motion compensation", THE INSTITUTE OF ELECTRONICS, INFORMATION AND COMMUNICATION ENGINEERS, TECHNICAL COMMITTEE ON IMAGE ENGINEERING, November 2009 (2009-11-01), pages 113 - 116 |
Y VATIS; B. EDLER; D. T NGUYEN; J. OSTERMANN: "Motion-and aliasing-compensated prediction using a two-dimensional non-separable adaptive Wiener interpolation filter", PROC. ICIP2005, IEEE INTERNATIONAL CONFERENCE ON IMAGE PROCESSING, September 2005 (2005-09-01), pages 894 - 897, XP010851198, DOI: doi:10.1109/ICIP.2005.1530200 |
Also Published As
Publication number | Publication date |
---|---|
CN103069803A (zh) | 2013-04-24 |
KR20160003869A (ko) | 2016-01-11 |
US20130128984A1 (en) | 2013-05-23 |
TW201225676A (en) | 2012-06-16 |
CA2807327A1 (en) | 2012-02-16 |
CN103069803B (zh) | 2016-09-28 |
TWI504236B (zh) | 2015-10-11 |
EP2590410A1 (en) | 2013-05-08 |
JP5524762B2 (ja) | 2014-06-18 |
KR20130055631A (ko) | 2013-05-28 |
BR112013002733A2 (pt) | 2016-05-31 |
EP2590410A4 (en) | 2016-11-02 |
JP2012044238A (ja) | 2012-03-01 |
PL2590410T3 (pl) | 2022-03-28 |
EP2590410B1 (en) | 2021-11-24 |
US9609318B2 (en) | 2017-03-28 |
CA2807327C (en) | 2016-03-22 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
JP5563403B2 (ja) | 映像符号化方法,映像復号方法,映像符号化装置,映像復号装置およびそれらのプログラム | |
WO2012020800A1 (ja) | 映像符号化方法,映像復号方法,映像符号化装置,映像復号装置およびそれらのプログラム | |
JP5485851B2 (ja) | 映像符号化方法,映像復号方法,映像符号化装置,映像復号装置およびそれらのプログラム | |
JP5649523B2 (ja) | 映像符号化方法,装置,映像復号方法,装置およびそれらのプログラム | |
JP2016220245A (ja) | 映像復号化装置 | |
Lainema et al. | Angular intra prediction in high efficiency video coding (HEVC) | |
TWI504241B (zh) | 影像編碼方法、裝置、影像解碼方法、裝置及其程式產品 | |
JP2011082725A (ja) | 映像符号化方法,映像符号化装置,映像復号方法,映像復号装置,映像符号化・復号方法,およびプログラム | |
JP5762243B2 (ja) | 映像符号化方法,装置,映像復号方法,装置およびそれらのプログラム |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
WWE | Wipo information: entry into national phase |
Ref document number: 201180035072.4 Country of ref document: CN |
|
121 | Ep: the epo has been informed by wipo that ep was designated in this application |
Ref document number: 11816464 Country of ref document: EP Kind code of ref document: A1 |
|
ENP | Entry into the national phase |
Ref document number: 20137001194 Country of ref document: KR Kind code of ref document: A |
|
ENP | Entry into the national phase |
Ref document number: 2807327 Country of ref document: CA |
|
WWE | Wipo information: entry into national phase |
Ref document number: 13814144 Country of ref document: US Ref document number: 2011816464 Country of ref document: EP |
|
NENP | Non-entry into the national phase |
Ref country code: DE |
|
REG | Reference to national code |
Ref country code: BR Ref legal event code: B01A Ref document number: 112013002733 Country of ref document: BR |
|
ENP | Entry into the national phase |
Ref document number: 112013002733 Country of ref document: BR Kind code of ref document: A2 Effective date: 20130204 |