WO2015190078A1 - 映像符号化装置、映像符号化方法および記録媒体 - Google Patents
映像符号化装置、映像符号化方法および記録媒体 Download PDFInfo
- Publication number
- WO2015190078A1 WO2015190078A1 PCT/JP2015/002837 JP2015002837W WO2015190078A1 WO 2015190078 A1 WO2015190078 A1 WO 2015190078A1 JP 2015002837 W JP2015002837 W JP 2015002837W WO 2015190078 A1 WO2015190078 A1 WO 2015190078A1
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- ctu
- prediction
- rectangular area
- video
- image
- Prior art date
Links
Images
Classifications
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/90—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using coding techniques not provided for in groups H04N19/10-H04N19/85, e.g. fractals
- H04N19/96—Tree coding, e.g. quad-tree coding
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/10—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
- H04N19/102—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the element, parameter or selection affected or controlled by the adaptive coding
- H04N19/115—Selection of the code volume for a coding unit prior to coding
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/10—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
- H04N19/102—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the element, parameter or selection affected or controlled by the adaptive coding
- H04N19/119—Adaptive subdivision aspects, e.g. subdivision of a picture into rectangular or non-rectangular coding blocks
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/10—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
- H04N19/134—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the element, parameter or criterion affecting or controlling the adaptive coding
- H04N19/162—User input
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/10—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
- H04N19/134—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the element, parameter or criterion affecting or controlling the adaptive coding
- H04N19/167—Position within a video image, e.g. region of interest [ROI]
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/30—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using hierarchical techniques, e.g. scalability
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/30—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using hierarchical techniques, e.g. scalability
- H04N19/33—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using hierarchical techniques, e.g. scalability in the spatial domain
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/50—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding
- H04N19/503—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding involving temporal prediction
- H04N19/51—Motion estimation or motion compensation
- H04N19/513—Processing of motion vectors
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/50—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding
- H04N19/503—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding involving temporal prediction
- H04N19/51—Motion estimation or motion compensation
- H04N19/513—Processing of motion vectors
- H04N19/517—Processing of motion vectors by encoding
- H04N19/52—Processing of motion vectors by encoding by predictive encoding
Definitions
- the present invention relates to a coding control technique of a scalable coding system, and relates to a video coding apparatus, a video coding method, and a recording medium using, for example, Scalable High-efficiency Video Coding (SHVC).
- SHVC Scalable High-efficiency Video Coding
- the video coding method of the scalable coding method based on the method described in Non-Patent Document 1 encodes a low resolution image obtained by down-sampling an input image as a low resolution layer (BL: Base Layer). Further, the video encoding method encodes an input image as a high resolution layer (EL: Enhancement Layer). Each frame in the BL and EL of the digitized video is divided into coding tree units (CTU: Coding Tree Unit). Then, each CTU is encoded in the raster scan order.
- CTU Coding Tree Unit
- the CTU is divided into coding units (CU: Coding Unit) in a quad tree structure and is coded.
- CU Coding Unit
- PU Prediction Unit
- TU Transform Unit
- CU is a coding unit of intra prediction / interframe prediction / interlayer prediction.
- intra prediction, interframe prediction, and interlayer prediction will be described.
- Intra prediction is prediction generated from the reconstructed image of the encoding target frame.
- 33 types of angle intra prediction shown in FIG. 15 are defined.
- an intra prediction signal is generated by extrapolating the reconstructed pixels around the encoding target block in any of the 33 types of directions shown in FIG.
- a CU encoded based on intra prediction is referred to as an intra CU.
- Inter-frame prediction is prediction based on an image of a reconstructed frame (reference picture) having a display time different from that of an encoding target frame.
- inter-frame prediction is also referred to as inter prediction.
- FIG. 16 is a diagram for explaining an example of inter-frame prediction.
- the motion vector MV (mv x , mv y ) indicates the parallel movement amount of the reconstructed image block of the reference picture with respect to the encoding target block.
- an inter prediction signal is generated based on a reconstructed image block of a reference picture (using pixel interpolation if necessary).
- AMVP Advanced Motion Vector Prediction
- merge mode is a technique for predicting a motion vector by using a motion vector of a reference picture so that a difference between motion vectors is minimized.
- AMVP a set of a reference picture index, an AMVP index associated with an AMVP motion vector predictor, and an AMVP motion vector predictor is transmitted.
- the merge mode is a technique that uses the motion vector of the reference picture as it is. In the merge mode, a set of a merge flag indicating that merge prediction is valid and a merge candidate index associated with a diverted motion vector is transmitted.
- Inter-layer prediction is inter prediction using an upsampled image of a reconstructed frame of a coded BL.
- FIG. 17 is a diagram for explaining inter-layer prediction.
- an inter-layer prediction signal is generated by up-sampling an encoded BL reconstructed frame to the same resolution as an EL frame.
- inter CU a CU encoded based on inter prediction or inter-layer prediction.
- a frame encoded only by an intra CU is called an I frame (or I picture).
- a frame encoded including not only an intra CU but also an inter CU is called a P frame (or P picture).
- a frame encoded by including not only one reference picture for inter prediction of a block but also an inter CU using two reference pictures at the same time is called a B frame (or B picture).
- BL encoder 900A that encodes BL
- EL encoder 900B that encodes EL
- downsampler 909 a downsampler 909
- multiplexer 910 The video encoder shown in FIG.
- the BL encoder 900A includes an estimator 901A, a predictor 902A, a frequency converter 903A, a quantizer 904A, an inverse frequency transformer / inverse quantizer 905A, a buffer 906A, and an entropy encoder 907A.
- the EL encoder 900B includes an estimator 901B, a predictor 902B, a frequency transformer 903B, a quantizer 904B, an inverse frequency transformer / inverse quantizer 905B, a buffer 906B, an entropy encoder 907B, and an upsampler 908. Have.
- the EL and BL CTUs respectively input to the BL encoder 900A and the EL encoder 900B are divided into variable-size CUs based on a quadtree structure.
- the CTU becomes the CU as it is, and therefore the CTU size becomes the maximum size (maxCUSize) of the CU.
- the maximum size CU and the minimum size CU are referred to as an LCU (Large Coding Unit, maximum coding unit) and an SCU (Smallest Coding Unit, minimum coding unit), respectively.
- FIG. 19 illustrates a CTU partition example of the t-th frame and a CU partition example of the eighth CTU (CTU 8) when the spatial resolution of the frame is CIF (Common Intermediate Format) and the CTU size is 64.
- CTU 8 the eighth CTU
- the numbers assigned to the CUs indicate the processing order of the CUs.
- the t-th frame is also referred to as a frame t.
- FIG. 20 is a diagram for explaining a hierarchical block representation and a quadtree structure corresponding to a CU partitioning example of CTU8.
- CU Depth in the hierarchical block representation shown in FIG. 20 indicates the depth of the divided hierarchy of the CU starting from the CTU.
- the video encoding apparatus transmits a split_cu_flag syntax indicating whether or not to divide a CU in order to signal the CU partition structure of the CTU (send a signal from the encoder to the decoder).
- the value (0 or 1) of the quadtree node in the quadtree structure corresponds to the value of split_cu_flag.
- FIG. 21 is a diagram for explaining an example of PU division of a CU.
- N is a variable indicating the size.
- the shape of a divided PU (also referred to as a PU divided shape) is two patterns of 2N ⁇ 2N and N ⁇ N.
- PU partition shapes are 8 patterns of 2N ⁇ 2N, 2N ⁇ N, N ⁇ 2N, N ⁇ N, 2N ⁇ nU, 2N ⁇ nD, nL ⁇ 2N, and nR ⁇ 2N.
- n represents an arbitrary number
- U, D, L, and R are variables indicating an arbitrary size.
- the numbers assigned to the PUs indicate the processing order of the PUs.
- the video encoding apparatus transmits a parameter (block division shape) indicating which one of intra prediction, inter prediction, and inter-layer prediction is selected and which division pattern is selected when the CU is divided into PUs. Also, the video encoding apparatus transmits parameters based on AMVP or merge mode. Information indicating which prediction is selected from intra prediction, inter prediction, and inter-layer prediction, block partition shape, and parameters based on AMVP or merge mode are combined into block partition / block prediction parameters or simply block prediction Called a parameter.
- the prediction error of each CU is divided into variable-size TUs based on the quadtree structure, similar to the CTU.
- FIG. 22 is a diagram for explaining a TU partition example in the case of an inter CU, a hierarchical block expression and a quad tree structure corresponding to this TU partition example.
- the parent node position of the TU quadtree structure is the CU. Thereby, transform coding across a plurality of PUs in the same CU is possible.
- the TU Depth in the hierarchical block representation shown in FIG. 22 indicates the depth of the divided hierarchy of the TU starting from the CU.
- the video encoding apparatus transmits a split_transform_flag syntax indicating whether or not to split the TU.
- the value (0 or 1) of the quadtree node in the quadtree structure corresponds to the value of split_transform_flag.
- FIG. 23 is a diagram for explaining a TU partition example in the case of an intra CU, a hierarchical block expression and a quadtree structure corresponding to the TU partition example.
- the parent node position of the TU quadtree structure is PU, and is divided into TUs similarly to the inter CU.
- the estimator 901A For each CTU of the low-resolution image obtained by down-sampling the input image by the down-sampler 909, the estimator 901A includes a CU quadtree structure, a PU block prediction parameter (hereinafter referred to as a PU block prediction parameter), and Determine the TU quadtree structure.
- a PU block prediction parameter hereinafter referred to as a PU block prediction parameter
- the predictor 902A generates a prediction signal for the input image signal of the CU based on the CU quadtree structure and the PU block prediction parameter determined by the estimator 901A.
- the prediction signal is generated based on the above-described intra prediction or inter prediction.
- the frequency converter 903A performs frequency conversion on a prediction error signal (hereinafter also referred to as a prediction error image) obtained by subtracting the prediction signal from the input image signal based on the TU quadtree structure determined by the estimator 901A.
- a prediction error image obtained by subtracting the prediction signal from the input image signal based on the TU quadtree structure determined by the estimator 901A.
- the quantizer 904A quantizes the orthogonal transform coefficient (frequency-transformed prediction error image).
- the quantized orthogonal transform coefficient is referred to as a coefficient level.
- a coefficient level having a value other than 0 is called a significant coefficient level.
- the entropy encoder 907A entropy-encodes the split_cu_flag indicating the CU quadtree structure in units of CTUs, the PU block prediction parameter, the split_transform_flag indicating the TU quadtree structure, and the coefficient level.
- a group of parameters to be entropy encoded is called an encoding parameter.
- the inverse frequency transform / inverse quantizer 905A inversely quantizes the coefficient level. Furthermore, the inverse frequency transform / inverse quantizer 905A performs inverse frequency transform on the inversely quantized orthogonal transform coefficient.
- the reconstructed prediction error image subjected to the inverse frequency transform is supplied with a prediction signal and supplied to the buffer 906A as a reconstructed image.
- the buffer 906A stores the reconstructed image.
- the reconstructed image stored in the buffer 906A is acquired from the estimator 901A and the predictor 902A, and is used to determine the CU quadtree structure, the PU block prediction parameter, and the TU quadtree structure, and to generate a prediction signal.
- the estimator 901B determines a CU quadtree structure, a PU block prediction parameter, and a TU quadtree structure for each CTU of the input image.
- the predictor 902B generates a prediction signal for the input image signal of the CU based on the CU quadtree structure and the PU block prediction parameter determined by the estimator 901B.
- the prediction signal is generated based on the above-described intra prediction, inter prediction, or inter-layer prediction.
- the frequency converter 903B performs frequency conversion on the prediction error image obtained by subtracting the prediction signal from the input image signal based on the TU quadtree structure determined by the estimator 901B.
- the quantizer 904B quantizes the orthogonal transform coefficient (frequency-transformed prediction error image).
- the entropy encoder 907B entropy-encodes the split_cu_flag indicating the CU quadtree structure, the PU block prediction parameter, the split_transform_flag indicating the TU quadtree structure, and the coefficient level.
- the inverse frequency transform / inverse quantizer 905B performs inverse quantization on the coefficient level. Further, the inverse frequency transform / inverse quantizer 905B performs inverse frequency transform on the inversely quantized orthogonal transform coefficient.
- the reconstructed prediction error image subjected to inverse frequency conversion is supplied with a prediction signal and supplied to the buffer 906B as a reconstructed image.
- the buffer 906B stores the reconstructed image.
- the buffer 906B also stores an image obtained by up-sampling the BL reconstructed image by the up-sampler 908.
- Data stored in the buffer 906B is acquired from the estimator 901B and the predictor 902B, and is used to determine the CU quadtree structure, PU block prediction parameters, and TU quadtree structure, and to generate a prediction signal.
- the BL encoder 900A generates a BL bit stream that is a sub bit stream.
- the EL encoder 900B generates an EL bit stream that is a sub bit stream.
- a general video encoding apparatus generates a scalable bit stream by multiplexing these sub bit streams with a multiplexer 910.
- Patent Document 1 describes a moving picture coding apparatus that optimizes coding efficiency and prediction efficiency.
- the apparatus described in Patent Document 1 is different from the encoded one in the case where the motion of the block to be processed moves at a constant speed with any reference image in the LX direction and any reference image in the LY direction.
- the motion information of the single direction is By generating the scaled combined motion information candidate by scaling, it is possible to encode only the merge index without encoding the motion information.
- High efficiency video coding HEVC
- JCT-VC Joint Collaborative Team on Video Coding
- an area that compresses with priority on image quality (hereinafter also referred to as an image quality priority compression area) and an area that compresses with priority on the number of bits (hereinafter referred to as bit number priority compression area). If it is attempted to switch a suitable compression process, the calculation amount for the determination to switch the compression process and CTU division control increases.
- the image quality is, for example, spatial resolution.
- the minimum image quality is the image quality of the low-resolution image.
- a general video encoding device cannot compress the image quality priority area designated by the user to high image quality while keeping the minimum image quality of the entire screen constant without increasing the amount of calculation.
- the present invention provides a video encoding device, a video code, and the like that can suppress an increase in the amount of computation and can compress the image quality priority area designated by the user to a high image quality while keeping the minimum image quality of the entire screen constant. It is an object of the present invention to provide a recording medium for storing an encoding method and a video encoding program.
- a video encoding device includes a base layer bit stream in which a low-resolution image obtained by down-sampling an input image is encoded as a base layer, and an enhancement layer bit stream in which the input image is encoded as an enhancement layer.
- Is a video encoding device that outputs a scalable bitstream multiplexed with a rectangular area generation unit that generates a rectangular area that is a multiple of the CTU size, including a specific rectangular area, and the CTU to be encoded is a CTU
- a determination unit that determines whether or not a rectangular area that is a multiple of the size is included in the rectangular area that is the multiple of the size, and if the CTU that is the encoding target is not included in the rectangular area that is a multiple of the CTU size, And then predict each CU with a prediction signal from the base layer of the zero motion vector And a prediction means.
- a video transmission / reception system includes a base layer bit stream in which a low-resolution image obtained by down-sampling an input image is encoded as a base layer, and an enhancement layer bit stream in which the input image is encoded as an enhancement layer.
- a video encoding device that outputs a scalable bitstream multiplexed with each other, a video decoding device that receives and decodes a scalable bitstream output from the video encoding device, a decoded image, and a specific rectangular area
- An image generation unit that generates an image including rectangular area information, wherein the video encoding device includes a rectangular area generation unit that generates a rectangular area that is a multiple of the CTU size, including a specific rectangular area, and an encoding target It is determined whether or not the CTU is included in a rectangular area that is a multiple of the CTU size When the CTU to be encoded is not included in the rectangular area that is a multiple of the CTU size, the CTU to be encoded is divided by the minimum number of CU blocks, and each CU is moved with zero motion.
- Prediction means for predicting with a prediction signal from the vector base layer.
- a display video generation device is a display video generation device that generates a display video based on a decoded video of a scalable bitstream and rectangular area information, and includes a video decoding device and an image generation unit.
- the video decoding device decodes the base layer bit stream from the scalable bit stream, and the image generation unit generates the display video of the base layer bit stream expanded to the display size.
- the video decoding device decodes the enhancement layer bit stream including the base layer bit stream and the rectangular region from the scalable bit stream, and the image generation unit converts the decoded video of the base layer bit stream and the rectangular region.
- Including enhancement layer bit list Generates a decoded image of the over-time, when the user specifies the rectangular area display, the image generating unit superimposes the rectangular area in the decoded video.
- a video encoding method includes a base layer bit stream in which a low-resolution image obtained by down-sampling an input image is encoded as a base layer, and an enhancement layer bit stream in which the input image is encoded as an enhancement layer.
- a video encoding method in a video encoding device that outputs a scalable bitstream multiplexed with a CTU, and generates a rectangular area that is a multiple of the CTU size, including a specific rectangular area, and the CTU to be encoded is a CTU It is determined whether or not the CTU to be encoded is included in the rectangular area that is a multiple of the CTU size, and the CTU to be encoded is divided by the minimum number of CU blocks. Further, each CU is predicted with a prediction signal from the base layer of the zero motion vector.
- a computer-readable recording medium includes a base layer bitstream in which a low-resolution image obtained by down-sampling an input image is encoded as a base layer, and an enhancement layer in which the input image is encoded as an enhancement layer.
- the present invention it is possible to compress the image quality priority area designated by the user to high image quality while suppressing an increase in the amount of calculation and keeping the minimum image quality of the entire screen constant.
- FIG. 1 is a block diagram showing an example of the configuration of a video encoding apparatus according to the first embodiment of the present invention.
- the configuration of a video encoding apparatus according to the first embodiment that outputs a bit stream using each frame of a digitized video as an input image will be described.
- the video encoding apparatus generates a rectangular area that is a multiple of the CTU size including a rectangular area (specific rectangular area) designated from the outside of the apparatus by a CTU align coordinate converter 111 described later. .
- a rectangular area designated from the outside of the apparatus by a CTU align coordinate converter 111 described later.
- the determination of the area to be compressed with priority on image quality and the area to be compressed with priority on the number of bits is made in CTU units, and the switching control of compression processing suitable for each area is simplified.
- block division / block prediction parameters in an area to be compressed with priority given to the number of bits are determined by a bit number priority estimator 101B2 described later.
- the encoding parameter having the minimum number of bits using the base layer is uniquely selected, and the amount of calculation necessary for estimating the encoding parameter is greatly reduced while the image quality is kept constant.
- the 1 includes a BL encoder 100A that encodes BL, an EL encoder 100B that encodes EL, a downsampler 109, and a multiplexer 110.
- the BL encoder 100A is, for example, a BL HEVC encoder.
- the EL encoder 100B is, for example, an EL HEVC encoder.
- the BL encoder 100A includes an estimator 101A, a predictor 102A, a frequency converter 103A, a quantizer 104A, an inverse frequency transformer / inverse quantizer 105A, a buffer 106A, and an entropy encoder 107A.
- the EL encoder 100B includes a predictor 102B, a frequency converter 103B, a quantizer 104B, an inverse frequency transformer / inverse quantizer 105B, a buffer 106B, an entropy encoder 107B, an upsampler 108, and a CTU align coordinate converter. 111, CTU determination unit 112, image quality priority estimator 101B1, and bit number priority estimator 101B2.
- the EL encoder 100B is provided with a CTU align coordinate converter 111, a CTU determiner 112, an image quality priority estimator 101B1, and a bit number priority estimator 101B2.
- Other blocks in the video encoding device shown in FIG. 1 are the same as the blocks in the video encoding device shown in FIG. Therefore, hereinafter, the configuration of the EL encoder 100B, which is a characteristic part of the present embodiment, will be described.
- the CTU align coordinate converter 111 receives an upper left (x, y) coordinate and a lower right (x, y) coordinate (hereinafter also referred to as rectangular area information) of a rectangular area designated by the user as an image quality priority area, and gives priority to image quality. Output the compressed area.
- the CTU align coordinate converter 111 includes an upper left (x, y) coordinate and a lower right (x, y) adjusted to a multiple of the maximum size of the CTU, including the image quality priority area specified by the user. )
- the coordinates are output as an image quality priority compression area.
- the CTU determiner 112 receives the image quality priority compression area output from the CTU align coordinate converter 111 and the current encoding target CTU.
- the CTU determination unit 112 determines whether or not the current encoding target CTU is included in the image quality priority compression region, and outputs a control signal.
- the CTU determination unit 112 When the current encoding target CTU is included in the image quality priority compression region, the CTU determination unit 112 outputs a control signal for controlling the switch so as to satisfy the following (i), (ii), and (iii).
- the current encoding target CTU is input to the image quality priority estimator 101B1.
- the output of the image quality priority estimator 101B1 is input to the predictor 102B and the entropy encoder 107B.
- the image quality priority estimator 101B1 can acquire the data stored in the buffer 106B from the buffer 106B.
- the image quality priority estimator 101B1 determines a CU quadtree structure, a PU block prediction parameter, and a TU quadtree structure for each CTU, as in a general video encoding apparatus.
- the CU quadtree structure is determined so that the rate distortion cost of the CU of the current coding target CTU is minimized, as in a general video coding apparatus.
- the PU block prediction parameter is determined so that the rate distortion cost of each CU is minimized, as in a general video encoding apparatus.
- the TU quadtree structure is determined so that the rate distortion cost of each CU is minimized, as in a general video encoding apparatus.
- the CTU determination unit 112 When the current encoding target CTU is not included in the image quality priority compression region, the CTU determination unit 112 outputs a control signal for controlling the switch so that the following (i), (ii), and (iii) are satisfied. .
- the current encoding target CTU is input to the image quality priority estimator 101B2.
- the output of the image quality priority estimator 101B2 is input to the predictor 102B and the entropy encoder 107B.
- the image quality priority estimator 101B2 can acquire the data stored in the buffer 106B from the buffer 106B.
- the bit number priority estimator 101B2 keeps the image quality of the current encoding target CTU constant, minimizes the number of bits, and increases the efficiency of the encoding process. Determine PU block prediction parameters and TU quadtree structure.
- the bit number priority estimator 101B2 determines PU block prediction parameters so that the number of bits is minimized in each CU. For example, the bit number priority estimator 101B2 determines the partition shape of the PU as 2N ⁇ 2N with a small number of partitions. Furthermore, the bit number priority estimator 101B2 selects inter-layer prediction of a motion vector of zero instead of intra prediction so as to maintain a constant image quality in each PU.
- the predictor 102B outputs a prediction signal for the input image signal of the CU based on the CU quadtree structure and the PU block prediction parameter determined by the image quality priority estimator 101B1 or the bit number priority estimator 101B2.
- the prediction signal is generated based on the above-described intra prediction, inter prediction, or inter-layer prediction.
- the frequency converter 103B performs frequency conversion on the prediction error image obtained by subtracting the prediction signal from the input image signal based on the TU quadtree structure determined by the image quality priority estimator 101B1 or the bit number priority estimator 101B2. Then, the frequency transformer 103B outputs orthogonal transform coefficients (frequency-transformed prediction error images).
- the quantizer 104B quantizes the orthogonal transform coefficient. Then, the quantizer 104B outputs a coefficient level.
- the entropy encoder 107B entropy-encodes the split_cu_flag indicating the CU quadtree structure, the PU block prediction parameter, the split_transform_flag indicating the TU quadtree structure, and the coefficient level. Then, the entropy encoder 107B outputs an EL bit stream.
- the inverse frequency transformer / inverse quantizer 105B performs inverse quantization on the coefficient level. Then, the inverse frequency transform / inverse quantizer 105B performs inverse frequency transform on the orthogonal transform coefficient obtained by inverse quantization and outputs a reconstructed prediction error image.
- the buffer 106B receives an image obtained by up-sampling the BL reconstructed image and a signal obtained by adding a prediction signal to the reconstructed prediction error image, and stores them as an EL reconstructed image.
- the CTU align coordinate converter 111 receives the upper left (x, y) coordinate and lower right (x, y) coordinate of the rectangular area designated as the image quality priority area by the user. Then, the CTU align coordinate converter 111 converts the upper left (x, y) coordinates and lower right (x, y) coordinates, which are adjusted to a multiple of the maximum size of the CTU, including the image quality priority area specified by the user. Output.
- the CTU determination unit 112 determines whether or not the current encoding target CTU is included in the image quality priority compression region (step S102). When included in the image quality priority compression area (Yes in step S102), the EL encoder 100B proceeds to the process of step S103. When not included in the image quality priority compression area (No in step S102), the EL encoder 100B proceeds to the process of step S109.
- the image quality priority estimator 101B1 determines the CU quadtree structure, the PU block prediction parameter, and the TU quadtree structure (step S103).
- the CU quadtree structure is determined such that the CU of the current coding target CTU has a minimum rate distortion cost.
- the PU block prediction parameters are determined so that the rate distortion cost of each CU is minimized.
- the TU quadtree structure is determined such that the rate distortion cost of each CU is minimized. Then, the EL encoder 100B proceeds to the process of step S104.
- the bit number priority estimator 101B2 uniquely determines the CU quadtree structure, the PU block prediction parameter, and the TU quadtree structure (step S109).
- the CU quadtree structure is determined so that the number of CU divisions of the current coding target CTU is minimized.
- the block division shape among the PU block prediction parameters is determined so that the number of bits of the PU parameter is minimized in each CU.
- the bit number priority estimator 101B2 selects inter-layer prediction, not intra prediction or inter prediction, so as to maintain a constant image quality in each PU.
- the TU quadtree structure is determined so that the number of bits of the TU parameter of each CU is minimized. That is, the bit number priority estimator 101B2 selects the maximum size TU parameter. Then, the EL encoder 100B proceeds to the process of step S104.
- step S104 the predictor 102B generates and outputs a prediction signal based on the determined PU block prediction parameter. Then, the EL encoder 100B proceeds to the process of step S105.
- a prediction error image (prediction error signal) is generated by subtracting the prediction signal from the input image signal (step S105).
- a prediction error image that is a difference between the input image signal and the prediction signal is input to the frequency converter 103B. Then, the EL encoder 100B proceeds to the process of step S106.
- the frequency converter 103B performs frequency conversion on the prediction error image based on the determined TU quadtree structure. Then, the frequency transformer 103B outputs orthogonal transform coefficients (frequency-transformed prediction error images). The quantizer 104B quantizes the orthogonal transform coefficient and outputs a coefficient level (step S106). Then, the EL encoder 100B proceeds to the process of step S107.
- the entropy encoder 107B entropy-encodes the split_cu_flag indicating the CU quadtree structure, the PU block prediction parameter, the split_transform_flag indicating the TU quadtree structure, and the coefficient level (step S107). Then, the entropy encoder 107B outputs an EL bit stream. Then, the EL encoder 100B proceeds to the process of step S108.
- the EL encoder 100B determines whether all CTUs included in the input image have been processed (step S108). When all the CTUs have been processed (Yes in step S108), the EL encoder 100B ends the input image encoding process. Otherwise (No in step S108), the EL encoder 100B proceeds to the process in step S102 in order to process the next CTU.
- the video encoding apparatus includes the encoder configuration shown in FIG.
- the video encoding apparatus according to the second embodiment is configured to further include an AMVP estimator 113 in addition to the video encoding apparatus shown in FIG.
- FIG. 3 is a block diagram showing an example of the configuration of the video encoding apparatus according to the second embodiment of the present invention.
- the configuration of a video encoding apparatus according to the second embodiment that outputs a bit stream using each frame of a digitized video as an input image will be described.
- the video encoding apparatus prioritizes the area to be compressed and the number of bits with priority on image quality by means of generating a rectangular area that is a multiple of the CTU size including the rectangular area specified from outside the apparatus.
- the area to be compressed is determined in units of CTU.
- a means for generating a rectangular area that is a multiple of the CTU size corresponds to the CTU align coordinate converter 111.
- the switching control of the compression process suitable for each area is simplified.
- the encoding parameter of the minimum bit number using the base layer is uniquely selected by using AMVP by means for determining the block division / block prediction parameter in the region to be compressed with priority on the bit number.
- the means for determining the block division / block prediction parameters greatly reduces the amount of calculation required for estimating the encoding parameters while maintaining the image quality constant.
- the means for determining the block division / block prediction parameter corresponds to the bit number priority estimator 101B2 and the AMVP estimator 113 described later. As a result, it is possible to effectively utilize the remaining number of bits and the amount of calculation in the area to be compressed with priority on the number of bits, and to compress the area to be compressed with priority on image quality with higher image quality.
- FIG. 3 includes a BL encoder 100A, an EL encoder 200B, a downsampler 109, and a multiplexer 110.
- the configuration of the BL encoder 100A is the same as that in the first embodiment shown in FIG.
- the EL encoder 200B includes an AMVP estimator 113 in addition to the configuration of the EL encoder 100B in the first embodiment shown in FIG.
- the CTU align coordinate converter 111 receives the upper left (x, y) and lower right (x, y) coordinates of the rectangular area designated as the image quality priority area by the user, and outputs the image quality priority compression area. Specifically, the CTU align coordinate converter 111 includes an upper left (x, y) coordinate and a lower right (x, y) adjusted to a multiple of the maximum size of the CTU, including the image quality priority area specified by the user. ) The coordinates are output as an image quality priority compression area.
- the CTU determiner 112 receives the image quality priority compression area output from the CTU align coordinate converter 111 and the current encoding target CTU.
- the CTU determination unit 112 determines whether or not the current encoding target CTU is included in the image quality priority compression region, and outputs a control signal.
- the CTU determination unit 112 When the current encoding target CTU is not included in the image quality priority compression region, the CTU determination unit 112 outputs a control signal for controlling the switch so that the following (i), (ii), and (iii) are satisfied. .
- the current coding target CTU is input to the bit number priority estimator 101B2.
- the output of the AMVP estimator 113 is input to the predictor 102B and the entropy encoder 107B.
- the image quality priority estimator 101B2 can acquire the data stored in the buffer 106B from the buffer 106B.
- bit number priority estimator 101B2 and the AMVP estimator 113 are set for each CTU so that the number of bits is minimized while the image quality of the current CTU to be encoded is kept constant and the encoding process is made efficient. Then, the following (A) to (C) are determined.
- the bit number priority estimator 101B2 determines the block division shape among the PU block prediction parameters so that the number of bits is minimized in each CU. For example, the bit number priority estimator 101B2 determines the partition shape of the PU as 2N ⁇ 2N with a small number of partitions. Furthermore, the bit number priority estimator 101B2 selects inter-layer prediction instead of intra prediction so as to maintain a constant image quality in each PU. Among the PU block prediction parameters, a parameter based on AMVP is determined by the AMVP estimator 113.
- the AMVP estimator 113 outputs the following sets (a) to (c) as parameters based on the AM block PU prediction parameter based on the AMVP.
- the predictor 102B outputs a prediction signal for the input image signal of the CU based on the following (1) or (2).
- the prediction signal is generated based on the above-described intra prediction, inter prediction, or inter-layer prediction.
- the frequency converter 103B performs frequency conversion on the prediction error image obtained by subtracting the prediction signal from the input image signal based on the TU quadtree structure determined by the image quality priority estimator 101B1 or the bit number priority estimator 101B2. Then, the frequency transformer 103B outputs orthogonal transform coefficients (frequency-transformed prediction error images).
- the quantizer 104B quantizes the orthogonal transform coefficient. Then, the quantizer 104B outputs a coefficient level.
- the entropy encoder 107B entropy-encodes the split_cu_flag indicating the CU quadtree structure, the PU block prediction parameter, the split_transform_flag indicating the TU quadtree structure, and the coefficient level. Then, the entropy encoder 107B outputs an EL bit stream.
- the inverse frequency transformer / inverse quantizer 105B performs inverse quantization on the coefficient level. Then, the inverse frequency transform / inverse quantizer 105B performs inverse frequency transform on the orthogonal transform coefficient obtained by inverse quantization and outputs a reconstructed prediction error image.
- the buffer 106B receives an image obtained by up-sampling the BL reconstructed image and a signal obtained by adding a prediction signal to the reconstructed prediction error image, and stores them as an EL reconstructed image.
- the operation of the EL encoder 200B is the same as that of the first embodiment except for step S109.
- the EL encoder 200B according to the present embodiment differs from the EL encoder 100B in the operation of determining the PU block prediction parameter in step S109 described above. Therefore, the operation of the AMVP estimator 113 that determines a parameter based on AMVP among PU block prediction parameters will be described with reference to the flowchart shown in FIG.
- the AMVP estimator 113 determines a reference picture index associated with the base layer (step S201). Then, the AMVP estimator 113 proceeds to the process of step S202.
- the AMVP estimator 113 determines an AMVP index associated with the AMVP predicted motion vector closest to the zero motion vector (step S202). Then, the AMVP estimator 113 proceeds to the process of step S203.
- the AMVP estimator 113 determines a differential motion vector obtained by subtracting the AMVP predicted motion vector closest to the zero motion vector from the zero motion vector (step S203).
- the AMVP estimator 113 determines the following groups (a) to (c) as parameters based on the AMVP of the PU block prediction parameter.
- the AMVP estimator 113 ends the process of determining parameters based on AMVP among the PU block prediction parameters.
- the video encoding apparatus according to the third embodiment has an encoder configuration shown in FIG. Is provided.
- the video encoding apparatus according to the third embodiment is configured to further include a merge prediction estimator 114 in addition to the video encoding apparatus shown in FIG.
- FIG. 5 is a block diagram showing a configuration of a video encoding apparatus according to the third embodiment of the present invention.
- the configuration of a video encoding apparatus according to the third embodiment that outputs a bit stream using each frame of a digitized video as an input image will be described.
- the video encoding apparatus gives priority to the area to be compressed and the number of bits by giving priority to the image quality by means of generating a rectangular area that is a multiple of the CTU size including the rectangular area specified from the outside of the apparatus.
- the area to be compressed is determined in units of CTU.
- a means for generating a rectangular area that is a multiple of the CTU size corresponds to the CTU align coordinate converter 111.
- the coding parameter of the minimum bit number using the base layer is uniquely selected by utilizing the merge prediction by means for determining the block division / block prediction parameter in the area to be compressed with priority on the bit number.
- the means for determining the block division / block prediction parameter corresponds to a bit number priority estimator 101B2, an AMVP estimator 113, and a merge prediction estimator 114 described later.
- FIG. 5 includes a BL encoder 100A, an EL encoder 300B, a downsampler 109, and a multiplexer 110.
- the configuration of the BL encoder 100A is the same as that of the second embodiment shown in FIG.
- the EL encoder 300B has a merge prediction estimator 114 in addition to the configuration of the EL encoder 200B in the second embodiment shown in FIG.
- the CTU align coordinate converter 111 receives the upper left (x, y) and lower right (x, y) coordinates of the rectangular area designated as the image quality priority area by the user, and outputs the image quality priority compression area. Specifically, the CTU align coordinate converter 111 includes an upper left (x, y) coordinate and a lower right (x, y) adjusted to a multiple of the maximum size of the CTU, including the image quality priority area specified by the user. ) The coordinates are output as an image quality priority compression area.
- the CTU determiner 112 receives the image quality priority compression area output from the CTU align coordinate converter 111 and the current encoding target CTU.
- the CTU determination unit 112 determines whether or not the current encoding target CTU is included in the image quality priority compression region, and outputs a control signal.
- the CTU determination unit 112 When the current encoding target CTU is not included in the image quality priority compression region, the CTU determination unit 112 outputs a control signal for controlling the switch so that the following (i), (ii), and (iii) are satisfied. .
- the current coding target CTU is input to the bit number priority estimator 101B2.
- the output of the merge prediction estimator 114 is input to the predictor 102B and the entropy encoder 107B.
- the image quality priority estimator 101B2 can acquire the data stored in the buffer 106B from the buffer 106B.
- the unit 114 determines the following (A) to (C) for each CTU.
- the bit number priority estimator 101B2 determines the block division shape among the PU block prediction parameters so that the number of bits is minimized in each CU. For example, the bit number priority estimator 101B2 determines the partition shape of the PU as 2N ⁇ 2N with a small number of partitions. Furthermore, the bit number priority estimator 101B2 selects inter-layer prediction instead of intra prediction so as to maintain a constant image quality in each PU.
- a parameter based on AMVP is determined by the AMVP estimator 113, and among the PU block prediction parameters, a parameter based on the merge mode is determined by the merge prediction estimator 114.
- the AMVP estimator 113 outputs the following sets (a) to (c) as parameters based on the AM block PU prediction parameter based on the AMVP.
- the merge prediction estimator 114 outputs a combination of a merge flag and the merge candidate index when there is a reference picture index associated with the base layer and a merge candidate index associated with a zero motion vector. This merge flag indicates that merge prediction is valid. Further, the acquisition unit 114 outputs a set of the merge flag and the merge candidate index as a parameter based on the merge mode of the PU block prediction parameter.
- the predictor 102B outputs a prediction signal for the input image signal of the CU based on the following (1) or (2).
- the prediction signal is generated based on the above-described intra prediction, inter prediction, or inter-layer prediction.
- the frequency converter 103B performs frequency conversion on the prediction error image obtained by subtracting the prediction signal from the input image signal based on the TU quadtree structure determined by the image quality priority estimator 101B1 or the bit number priority estimator 101B2. Then, the frequency transformer 103B outputs orthogonal transform coefficients (frequency-transformed prediction error images).
- the quantizer 104B quantizes the orthogonal transform coefficient. Then, the quantizer 104B outputs a coefficient level.
- the entropy encoder 107B entropy-encodes the split_cu_flag indicating the CU quadtree structure, the PU block prediction parameter, the split_transform_flag indicating the TU quadtree structure, and the coefficient level. Then, the entropy encoder 107B outputs an EL bit stream.
- the inverse frequency transformer / inverse quantizer 105B performs inverse quantization on the coefficient level. Then, the inverse frequency transform / inverse quantizer 105B performs inverse frequency transform on the orthogonal transform coefficient obtained by inverse quantization and outputs a reconstructed prediction error image.
- the buffer 106B receives an image obtained by up-sampling the BL reconstructed image and a signal obtained by adding a prediction signal to the reconstructed prediction error image, and stores them as an EL reconstructed image.
- the operation of the EL encoder 300B is the same as that of the second embodiment except for the operation of PU block prediction parameter determination. Therefore, the operation of the merge prediction estimator 114 that determines a parameter based on the merge mode among the PU block prediction parameters will be described with reference to the flowchart shown in FIG.
- the merge prediction estimator 114 executes the following steps S301 to S303 after the AMVP estimator 113 executes the steps S201 to S203.
- the merge prediction estimator 114 confirms whether there is a reference picture index associated with the base layer and a merge candidate index associated with a zero motion vector (step S301). When the reference picture index and the merge candidate index exist, it is determined that the merge prediction is valid. If merge prediction is valid (Yes in step S301), merge prediction estimator 114 proceeds to the process in step S302. If not (No in step S301), the merge prediction estimator 114 ends the process of determining the parameters based on the merge mode among the PU block prediction parameters.
- the merge prediction estimator 114 determines a merge flag indicating that the merge prediction is valid (step S302). Then, the merge prediction estimator 114 proceeds to the process of step S303.
- the merge prediction estimator 114 determines a merge candidate index associated with a zero motion vector to be used for merge prediction (step S303).
- the merge prediction estimator 114 determines a merge flag indicating that the merge prediction is valid and a set of merge candidate indexes as parameters based on the merge mode among the PU block prediction parameters, and sets the PU block prediction parameter. The process of determining parameters based on the merge mode is terminated.
- a video encoding device In order to more reliably guarantee that the number of bits is minimized in the bit number priority compression region than in the first, second, or third embodiment, a video encoding device according to the fourth embodiment is shown in FIG.
- the encoder configuration shown is provided.
- the video encoding apparatus according to the fourth embodiment is further provided with a prediction error truncator 115 in the video encoding apparatus shown in FIG.
- FIG. 7 is a block diagram showing an example of the configuration of a video encoding apparatus according to the fourth embodiment of the present invention.
- a configuration of a video encoding apparatus according to a fourth embodiment that outputs a bit stream using each frame of a digitized video as an input image will be described.
- the video encoding apparatus gives priority to the area to be compressed and the number of bits by giving priority to the image quality by means of generating a rectangular area that is a multiple of the CTU size including the rectangular area specified from outside the apparatus.
- the area to be compressed is determined in units of CTU.
- a means for generating a rectangular area that is a multiple of the CTU size corresponds to the CTU align coordinate converter 111.
- the coding parameter of the minimum bit number using the base layer is uniquely selected by means for determining the block division / block prediction parameter in the area to be compressed with priority on the bit number.
- the calculation amount necessary for estimating the encoding parameter is greatly reduced by the means for determining the block division / block prediction parameter while maintaining the image quality constant.
- the means for determining the block division / block prediction parameter corresponds to the bit number priority estimator 101B2.
- the prediction error truncation means for forcibly setting the prediction error signal to 0 greatly increases the number of bits required for encoding the area to be compressed with priority on the number of bits. Reduced to As a result, it is possible to effectively utilize the remaining number of bits and the amount of calculation in the area to be compressed with priority on the number of bits, and to compress the area to be compressed with priority on image quality with higher image quality.
- FIG. 7 includes a BL encoder 100A, an EL encoder 400B, a downsampler 109, and a multiplexer 110.
- the configuration of the BL encoder 100A is the same as that in the first embodiment shown in FIG.
- the EL encoder 400B includes a prediction error truncator 115 in addition to the configuration of the EL encoder 100B in the first embodiment shown in FIG.
- the CTU align coordinate converter 111 receives the upper left (x, y) and lower right (x, y) coordinates of the rectangular area designated as the image quality priority area by the user, and outputs the image quality priority compression area. Specifically, the CTU align coordinate converter 111 includes an upper left (x, y) coordinate and a lower right (x, y) adjusted to a multiple of the maximum size of the CTU, including the image quality priority area specified by the user. ) The coordinates are output as an image quality priority compression area.
- the CTU align coordinate converter 111 Convert to upper left (0, 0), lower right (512, 320). Then, the CTU align coordinate converter 111 outputs the converted coordinates (image quality priority compression area).
- the CTU determiner 112 receives the image quality priority compression area output from the CTU align coordinate converter 111 and the current encoding target CTU.
- the CTU determination unit 112 determines whether or not the current encoding target CTU is included in the image quality priority compression region, and outputs a control signal.
- the CTU determination unit 112 When the current encoding target CTU is not included in the image quality priority compression region, the CTU determination unit 112 outputs a control signal for controlling the switch so that the following (i), (ii), and (iii) are satisfied. .
- the current coding target CTU is input to the bit number priority estimator 101B2.
- the output of the bit number priority estimator 101B2 is input to the predictor 102B and the entropy encoder 107B.
- the image quality priority estimator 101B2 can acquire the data stored in the buffer 106B from the buffer 106B.
- the bit number priority estimator 101B2 performs the CU quadtree for each CTU so that the number of bits is minimized while the image quality of the current encoding target CTU is kept constant and the encoding process is made efficient.
- the bit number priority estimator 101B2 determines the block division shape among the PU block prediction parameters so that the number of bits is minimized in each CU. For example, the bit number priority estimator 101B2 determines the partition shape of the PU as 2N ⁇ 2N with a small number of partitions. Furthermore, the bit number priority estimator 101B2 selects inter-layer prediction instead of intra prediction so as to maintain a constant image quality in each PU. Note that the bit number priority estimator 101B2 in this embodiment has the functions of the AMVP estimator 113 shown in FIG. 3 and the merge prediction estimator 114 shown in FIG. That is, in the present embodiment, among the PU block prediction parameters, there are the following two parameters based on AMVP or merge mode.
- One is a set of the following (a) to (c) determined based on AMVP.
- A a reference picture index associated with the base layer;
- B the AMVP index associated with the AMVP predicted motion vector closest to the zero motion vector;
- C A difference motion vector obtained by subtracting the AMVP predicted motion vector closest to the zero motion vector from the zero motion vector.
- the other is that when there is a reference picture index associated with the base layer and a merge candidate index associated with a motion vector of zero, a merge flag indicating that merge prediction is valid and the merge candidate index It is a pair.
- the predictor 102B outputs a prediction signal for the input image signal of the CU based on the CU quadtree structure and the PU block prediction parameter determined by the image quality priority estimator 101B1 or the bit number priority estimator 101B2.
- the prediction signal is generated based on the above-described intra prediction, inter prediction, or inter-layer prediction.
- the frequency converter 103B performs frequency conversion on the prediction error image obtained by subtracting the prediction signal from the input image signal based on the TU quadtree structure determined by the image quality priority estimator 101B1 or the bit number priority estimator 101B2. Then, the frequency transformer 103B outputs orthogonal transform coefficients (frequency-transformed prediction error images).
- the quantizer 104B quantizes the orthogonal transform coefficient. Then, the quantizer 104B outputs a coefficient level.
- the prediction error truncator 115 receives the prediction error signal and outputs a prediction error signal that is forced to zero. That is, this process is equivalent to setting all coefficient level values input to the entropy encoder 107B to zero.
- the entropy encoder 107B entropy-encodes the split_cu_flag indicating the CU quadtree structure, the PU block prediction parameter, the split_transform_flag indicating the TU quadtree structure, and the coefficient level, and outputs an EL bitstream.
- the inverse frequency transformer / inverse quantizer 105B performs inverse quantization on the coefficient level. Then, the inverse frequency transform / inverse quantizer 105B performs inverse frequency transform on the orthogonal transform coefficient obtained by inverse quantization and outputs a reconstructed prediction error image.
- the buffer 106B receives an image obtained by up-sampling the BL reconstructed image and a signal obtained by adding a prediction signal to the reconstructed prediction error image, and stores them as an EL reconstructed image.
- the CTU align coordinate converter 111 receives the upper left (x, y) coordinate and lower right (x, y) coordinate of the rectangular area designated as the image quality priority area by the user. Then, the CTU align coordinate converter 111 converts the upper left (x, y) coordinates and lower right (x, y) coordinates, which are adjusted to a multiple of the maximum size of the CTU, including the image quality priority area specified by the user. Output.
- the CTU determination unit 112 determines whether or not the current encoding target CTU is included in the image quality priority compression region (step S402). If included in the image quality priority compression region (Yes in step S402), the EL encoder 400B proceeds to the process in step S403. If not included in the image quality priority compression area (No in step S402), the EL encoder 400B proceeds to the process of step S410.
- the image quality priority estimator 101B1 determines the CU quadtree structure, the PU block prediction parameter, and the TU quadtree structure (step S403).
- the CU quadtree structure is determined such that the CU of the current coding target CTU has a minimum rate distortion cost.
- the PU block prediction parameters are determined so that the rate distortion cost of each CU is minimized.
- the TU quadtree structure is determined such that the rate distortion cost of each CU is minimized. Then, the EL encoder 400B proceeds to the process of step S404.
- the bit number priority estimator 101B2 uniquely determines the CU quadtree structure, the PU block prediction parameter, and the TU quadtree structure (step S410).
- the CU quadtree structure is determined so that the number of CU divisions of the current coding target CTU is minimized.
- the PU block prediction parameter is determined so that the number of bits of the PU parameter is minimized in each CU.
- the bit number priority estimator 101B2 selects inter-layer prediction, not intra prediction or inter prediction, so as to maintain a constant image quality in each PU.
- the TU quadtree structure is determined so that the number of bits of the TU parameter of each CU is minimized. That is, the bit number priority estimator 101B2 selects the maximum size TU parameter. Then, the EL encoder 400B proceeds to the process of step S404.
- step S403 or step S410 the predictor 102B generates and outputs a prediction signal based on the determined PU block prediction parameter (step S404). Then, the EL encoder 400B proceeds to the process of step S405.
- the EL encoder 400B determines whether or not the current encoding target CTU in the CTU determination unit 112 is included in the image quality priority compression region (step S405). If included in the image quality priority compression area (Yes in step S405), the EL encoder 400B proceeds to the process of step S406. If not included in the image quality priority compression area (No in step S405), the EL encoder 400B proceeds to step S411.
- step S405 a prediction error image is generated by subtracting the prediction signal from the input image signal (step S406).
- the CTU determination unit 112 of the EL encoder 400B is configured so that the prediction error signal is input to the frequency converter 103B, and the output of the quantizer 104B is the inverse frequency transform / inverse quantizer 105B and A control signal to be controlled is output so as to be input to the entropy encoder 107B.
- the prediction error image which is the difference between the input image signal and the prediction signal is input to the frequency converter 103B.
- the EL encoder 400B proceeds to the process of step S407.
- the frequency converter 103B performs frequency conversion on the prediction error image based on the determined TU quadtree structure. Then, the frequency transformer 103B outputs orthogonal transform coefficients (frequency-transformed prediction error images). The quantizer 104B quantizes the orthogonal transform coefficient and outputs a coefficient level (step S407). Then, the EL encoder 400B proceeds to the process of step S408.
- step S415 the prediction error truncator 115 forcibly sets the prediction error signal to 0 (step S411).
- the CTU determination unit 112 of the EL encoder 400B allows the prediction error signal to be input to the prediction error truncator 115, and the output of the prediction error truncator 115 is the inverse frequency transform / inverse quantum.
- a control signal to be controlled is output so as to be input to the encoder 105B and the entropy encoder 107B. Then, the EL encoder 400B proceeds to the process of step S408.
- step S407 or step S411 the entropy encoder 107B entropy-encodes the split_cu_flag indicating the CU quadtree structure, the PU block prediction parameter, the split_transform_flag indicating the TU quadtree structure, and the coefficient level (step S408). Then, the entropy encoder 107B outputs an EL bit stream. Then, the EL encoder 400B proceeds to the process of step S409.
- the EL encoder 400B determines whether or not all the CTUs included in the input image have been processed (step S409). If all the CTUs have been processed (Yes in step S409), the EL encoder 400B ends the input image encoding process. Otherwise (No in step S409), the EL encoder 400B proceeds to the process of step S402 in order to process the next CTU.
- the bit number priority estimator 101B2 has the functions of the AMVP estimator 113 and the merge prediction estimator 114 is taken as an example.
- the EL encoder 400B includes the AMVP estimator 113 and the merge prediction estimator 114. It goes without saying that the estimator 114 may be included. That is, it goes without saying that the EL encoder 200B according to the second embodiment or the EL encoder 300B according to the third embodiment may further include a prediction error truncator 115. .
- FIG. 9 is a block diagram showing an example of the configuration of a video transmission / reception system according to the fifth embodiment of the present invention. With reference to FIG. 9, the structure of the video transmission / reception system according to the fifth embodiment will be described.
- the video transmission / reception system of the fifth embodiment includes rectangular area information by an image generation unit (corresponding to an image generation unit 520 described later) that receives rectangular area information specified from the outside, and the image quality of the entire video is improved. It is possible to easily generate a display image with high image quality only in the rectangular area while keeping it constant. As a result, the receiving side can perform display control so that the rectangular area included in the decoded video is easily visible.
- an image generation unit corresponding to an image generation unit 520 described later
- the video transmission / reception system shown in FIG. 9 includes the SHVC encoder 100 on the transmission side, and includes the SHVC decoder 510 and the image generation unit 520 on the reception side.
- the SHVC encoder 100 includes the configuration of the video encoding device according to the first, second, third, or fourth embodiment.
- the SHVC encoder 100 receives video and rectangular area information (hereinafter also referred to as user data) input by the user on the transmission side.
- the SHVC encoder 100 compresses the image quality priority area designated by the user to a higher image quality and outputs a bitstream without increasing the amount of calculation while keeping the minimum image quality of the entire screen constant.
- the SHVC decoder 510 receives a bit stream and outputs a decoded video.
- the SHVC decoder 510 receives the bit stream transmitted from the SHVC encoder 100 via the network.
- the image generation unit 520 receives the decoded video output from the SHVC decoder 510 and the user data, includes rectangular area information, and maintains only the rectangular area corresponding to the rectangular area information while maintaining the image quality of the entire video.
- the display image is output.
- the user data is transmitted from the transmission side to the image generation unit 520 via the network.
- the rectangular area information may be input from the user on the receiving side.
- FIG. 10 is a block diagram showing an example of the configuration of a display video generation apparatus according to the sixth embodiment of the present invention.
- FIG. 10 shows an outline of a display video generation apparatus according to the sixth embodiment of the present invention.
- a configuration of a display video generation apparatus according to the sixth embodiment will be described.
- the display video generation apparatus can easily display a video according to the user's needs by an image generation unit (corresponding to an image generation unit 620 described later) that receives a control signal sent from the user.
- an image generation unit corresponding to an image generation unit 620 described later
- the display video generator 600 shown in FIG. 10 includes a SHVC decoder 610 and an image generator 620.
- the SHVC decoder 610 receives a bit stream and outputs a decoded video.
- the image generation unit 620 outputs the decoded video output from the SHVC decoder 610, user data input by the user (for example, a user on the transmission side in the video transmission / reception system illustrated in FIG. 9), and the user (for example, the video illustrated in FIG. 9).
- a control signal input by a receiving user in the transmission / reception system is input.
- the image generation unit 620 outputs a display video. For example, as shown in FIG. 10, the control signal is input to the display video generation apparatus 600 when the user operates a remote controller or the like.
- the SHVC decoder 610 decodes only the base layer bit stream from the scalable bit stream. Then, the image generation unit 620 outputs the decoded video (left video in FIG. 10) of the base layer bitstream expanded to the display size as a display video to a display device or the like.
- the SHVC decoder 610 decodes the base layer bit stream and the enhancement bit stream including the rectangular area designated by the user data from the scalable bit stream. Then, the image generation unit 620 outputs the decoded video of the base layer bit stream and the decoded video of the enhancement bit stream including the rectangular area (the central video in FIG. 10) as a display video to a display device or the like.
- the image generation unit 620 when the user designates rectangular area display by a control signal, the image generation unit 620 outputs a decoded video (video on the right side in FIG. 10) on which the rectangular area is superimposed to a display device or the like as a display video.
- the image generation unit 620 superimposes the rectangular area information on the decoded video of the base layer bitstream and the decoded video of the enhancement bitstream including the rectangular area, but the present embodiment is not limited to this.
- the image generation unit 620 may superimpose the rectangular area information on the decoded video of the enhancement bitstream including the rectangular area, and display the decoded video on which the rectangular area information is superimposed on the display size. Good.
- each of the above embodiments can be configured by hardware, it can also be realized by a computer program.
- the information processing system illustrated in FIG. 11 includes a processor 1001, a program memory 1002, a storage medium 1003 for storing video data, and a storage medium 1004 for storing a bitstream.
- the storage medium 1003 and the storage medium 1004 may be separate storage media, or may be storage areas composed of the same storage medium.
- a magnetic storage medium such as a hard disk can be used as the storage medium.
- the program memory 1002 has the function of each block (excluding the buffer block) shown in the respective drawings of the first, second, third, or fourth embodiments.
- a program to be realized is stored.
- the processor 1001 executes processing according to a program stored in the program memory 1002, thereby realizing the functions of the video encoding device described in the above embodiments.
- FIG. 12 is a block diagram illustrating an example of a main part of a video encoding device according to each embodiment of the present invention.
- FIG. 13 is a block diagram showing a main part of another video encoding apparatus according to each embodiment of the present invention.
- the video encoding apparatus includes a base layer bit stream obtained by encoding a low resolution image obtained by down-sampling an input image as a base layer, and an input image as an enhancement layer. It is a video encoding device that outputs a scalable bit stream in which an encoded enhancement layer bit stream is multiplexed.
- the video encoding device includes a rectangular area generation unit 11, a determination unit 12, and a prediction unit 13.
- the rectangular area generation unit 11 generates a rectangular area that is a multiple of the CTU size and includes a specific rectangular area.
- the determination unit 12 determines whether or not the CTU to be encoded is included in a rectangular area that is a multiple of the CTU size.
- An example of the determination unit 12 is a CTU determination unit 112 shown in FIG.
- the prediction unit 13 divides the CTU to be encoded by the minimum number of CU blocks, and further, each CU has a zero motion vector. Prediction is performed using a prediction signal from the base layer.
- An example of the prediction unit 13 is the bit number priority estimator 101B2 shown in FIG.
- the bit number priority estimator 101B2 and the AMVP estimator 113 shown in FIG. 3 or the bit number priority estimator 101B2, the AMVP estimator 113, and the merge prediction estimator 114 shown in FIG. Is mentioned.
- the determination of the area to be compressed with priority on image quality and the area to be compressed with priority on the number of bits is made in CTU units, and switching control of compression processing suitable for each area is simplified.
- the encoding parameter having the minimum number of bits using the base layer is uniquely selected, and the amount of calculation required for estimating the encoding parameter is greatly reduced while the image quality is kept constant.
- the video encoding apparatus can effectively use the remaining number of bits and the calculation amount in the area to be compressed with priority on the number of bits, and can compress the area to be compressed with priority on the image quality with higher image quality.
- the prediction unit 13 may determine the following groups (1) to (3) as parameters based on the AMVP of the block prediction parameter.
- (1) a reference picture index associated with base layer prediction; (2) the AMVP index associated with the AMVP predicted motion vector closest to the zero motion vector, and (3) A differential motion vector obtained by subtracting the AMVP predicted motion vector closest to the zero motion vector from the zero motion vector.
- the prediction unit 13 may determine the merge flag and the merge candidate index as parameters. This merge flag indicates that merge prediction is valid.
- the parameter determined by the prediction unit 13 is a parameter based on the merge mode of block prediction parameters. According to such a configuration, it is possible to more reliably ensure that the number of bits is minimized in the bit number priority compression region.
- the video encoding apparatus includes a prediction error truncation unit 14 that forcibly sets a prediction error signal to 0 in a CTU that is not included in a rectangular area that is a multiple of the CTU size (as an example, 7 may be provided as shown in FIG. According to such a configuration, it is possible to more reliably ensure that the number of bits is minimized in the bit number priority compression region.
- FIG. 14 is a block diagram showing a main part of the video transmission / reception system according to each embodiment of the present invention.
- the video transmission / reception system includes a video encoding device 10, a video decoding device 21, and an image generation unit 22.
- An example of the video encoding device 10 is the video encoding device shown in FIG.
- the video decoding device 21 receives and decodes the scalable bit stream output from the video encoding device 10.
- An example of the video decoding device 21 is the SHVC decoder 510 shown in FIG. 9 or the SHVC decoder 610 shown in FIG.
- the image generation unit 22 generates an image including the decoded image and rectangular area information indicating a specific rectangular area.
- a display video generation apparatus is a display video generation apparatus that generates a display video based on a decoded video of a scalable bitstream and rectangular area information.
- the apparatus 21 and the image generation part 22 are provided.
- the video decoding device 21 decodes the base layer bit stream from the scalable bit stream, and the image generation unit 22 generates a display video of the base layer bit stream expanded to the display size.
- the video decoding device 21 decodes the base layer bit stream and the enhancement layer bit stream including the rectangular area from the scalable bit stream.
- the image generation unit 22 generates a decoded video of the base layer bit stream and a decoded video of the enhancement layer bit stream including a rectangular area.
- the image generation unit 22 When the user designates rectangular area display, the image generation unit 22 superimposes the rectangular area on the decoded video.
Landscapes
- Engineering & Computer Science (AREA)
- Multimedia (AREA)
- Signal Processing (AREA)
- Compression Or Coding Systems Of Tv Signals (AREA)
Abstract
Description
以下、本発明の第1の実施形態を、図面を参照して説明する。
(i)画質優先推定器101B1に、現在の符号化対象のCTUが入力される。
(ii)画質優先推定器101B1の出力が予測器102Bおよびエントロピー符号化器107Bに入力される。
(iii)画質優先推定器101B1が、バッファ106Bから、バッファ106Bに格納されたデータを取得可能となる。
そして、画質優先推定器101B1は、一般的な映像符号化装置と同様に、CTU毎に、CUクアッドツリー構造、PUブロック予測パラメータ、およびTUクアッドツリー構造を決定する。CUクアッドツリー構造は、一般的な映像符号化装置と同様に、現在の符号化対象のCTUのCUのレート歪みコストが最小化されるように決定される。PUブロック予測パラメータは、一般的な映像符号化装置と同様に、各CUのレート歪みコストが最小化されるように決定される。さらに、TUクアッドツリー構造は、一般的な映像符号化装置と同様に、各CUのレート歪みコストが最小化されるように決定される。
(i)画質優先推定器101B2に、現在の符号化対象のCTUが入力される。
(ii)画質優先推定器101B2の出力が予測器102Bおよびエントロピー符号化器107Bに入力される。
(iii)画質優先推定器101B2が、バッファ106Bから、バッファ106Bに格納されたデータを取得可能となる。
ビット数優先推定器101B2は、現在の符号化対象のCTUの画質を一定に保ち、そのビット数が最小となり、かつ符号化処理が効率化されるように、CTU毎に、CUクアッドツリー構造、PUブロック予測パラメータ、およびTUクアッドツリー構造を決定する。
以下、本発明の第2の実施形態を図面を参照して説明する。
(i)ビット数優先推定器101B2に、現在の符号化対象のCTUが入力される。
(ii)AMVP推定器113の出力が予測器102Bおよびエントロピー符号化器107Bに入力される。
(iii)画質優先推定器101B2が、バッファ106Bから、バッファ106Bに格納されたデータを取得可能となる。
そして、現在の符号化対象のCTUの画質を一定に保ちつつそのビット数が最小となり、かつ、符号化処理が効率化されるように、ビット数優先推定器101B2およびAMVP推定器113がCTU毎に、以下の(A)~(C)を決定する。
(A)CUクアッドツリー構造、
(B)PUブロック予測パラメータ、および
(C)TUクアッドツリー構造。
(a)ベースレイヤに関連づけられた参照ピクチャインデックス、
(b)ゼロの動きベクトルに最も近いAMVP予測動きベクトルに関連づけられたAMVPインデックス、および、
(c)ゼロの動きベクトルから当該ゼロの動きベクトルに最も近いAMVP予測動きベクトルを減じた差分動きベクトル。
(1)画質優先推定器101B1が決定した、CUクアッドツリー構造およびPUブロック予測パラメータ、
(2)ビット数優先推定器101B2が決定したCUクアッドツリー構造、並びに、ビット数優先推定器101B2およびAMVP推定器113が決定したPUブロック予測パラメータ。
予測信号は、上述したイントラ予測、インター予測、またはレイヤ間予測に基づいて生成される。
(a)ベースレイヤに関連づけられた参照ピクチャインデックス、
(b)ゼロの動きベクトルに最も近いAMVP予測動きベクトルに関連づけられたAMVPインデックス、および、
(c)ゼロの動きベクトルから当該ゼロの動きベクトルに最も近いAMVP予測動きベクトルを減じた差分動きベクトル。
以下、本発明の第3の実施形態を図面を参照して説明する。
(i)ビット数優先推定器101B2に、現在の符号化対象のCTUが入力される。
(ii)マージ予測推定器114の出力が予測器102Bおよびエントロピー符号化器107Bに入力される。
(iii)画質優先推定器101B2が、バッファ106Bから、バッファ106Bに格納されたデータを取得可能となる。
そして、現在の符号化対象のCTUの画質を一定に保ちつつそのビット数が最小となり、かつ符号化処理が効率化されるように、ビット数優先推定器101B2、AMVP推定器113およびマージ予測推定器114がCTU毎に、以下の(A)~(C)を決定する。(A)CUクアッドツリー構造、
(B)PUブロック予測パラメータ、および
(C)TUクアッドツリー構造。
(a)ベースレイヤに関連づけられた参照ピクチャインデックス、
(b)ゼロの動きベクトルに最も近いAMVP予測動きベクトルに関連づけられたAMVPインデックス、および、
(c)ゼロの動きベクトルから当該ゼロの動きベクトルに最も近いAMVP予測動きベクトルを減じた差分動きベクトル。
(1)画質優先推定器101B1が決定した、CUクアッドツリー構造およびPUブロック予測パラメータ、
(2)ビット数優先推定器101B2が決定したCUクアッドツリー構造、並びに、ビット数優先推定器101B2およびAMVP推定器113およびマージ予測推定器114が決定したPUブロック予測パラメータ。
予測信号は、上述したイントラ予測、インター予測、またはレイヤ間予測に基づいて生成される。
以下、本発明の第4の実施形態を図面を参照して説明する。
(i)ビット数優先推定器101B2に、現在の符号化対象のCTUが入力される。
(ii)ビット数優先推定器101B2の出力が予測器102Bおよびエントロピー符号化器107Bに入力される。
(iii)画質優先推定器101B2が、バッファ106Bからバッファ106Bに格納されたデータを取得可能となる。
そして、現在の符号化対象のCTUの画質を一定に保ちつつそのビット数が最小となり、かつ、符号化処理が効率化されるように、ビット数優先推定器101B2がCTU毎に、CUクアッドツリー構造、PUブロック予測パラメータ、およびTUクアッドツリー構造を決定する。
(a)ベースレイヤに関連づけられた参照ピクチャインデックス、
(b)ゼロの動きベクトルに最も近いAMVP予測動きベクトルに関連づけられたAMVPインデックス、および、
(c)ゼロの動きベクトルから当該ゼロの動きベクトルに最も近いAMVP予測動きベクトルを減じた差分動きベクトル。
以下、本発明の第5の実施形態を図面を参照して説明する。
以下、本発明の第6の実施形態を図面を参照して説明する。
(1)ベースレイヤ予測に関連づけられた参照ピクチャインデックス、
(2)ゼロの動きベクトルに最も近いAMVP予測動きベクトルに関連づけられたAMVPインデックス、および、
(3)ゼロの動きベクトルからゼロの動きベクトルに最も近いAMVP予測動きベクトルを減じた差分動きベクトル。
そのような構成によれば、ビット数優先圧縮領域において、ビット数が最小になることをより確実に保証することができる。
この出願は、2014年6月12日に出願された日本出願特願2014-121635を基礎とする優先権を主張し、その開示の全てをここに取り込む。
11 矩形領域生成部
12 判定部
13 予測部
14 予測誤差切捨部
21 映像復号装置
22 画像生成部
100 SHVCエンコーダ
100A、900A BL符号化器
100B、200B、300B、400B、900B EL符号化器
101A、901A、901B 推定器
102A、102B、902A、902B 予測器
103A、103B、903A、903B 周波数変換器
104A、104B、904A、904B 量子化器
105A、105B、905A、905B 逆周波数変換/逆量子化器
106A、106B、906A、906B バッファ
107A、107B、907A、907B エントロピー符号化器
108、908 アップサンプル器
109、909 ダウンサンプル器
111 CTUアライン座標変換器
112 CTU判定器
113 AMVP推定器
114 マージ予測推定器
115 予測誤差切捨器
101B1 画質優先推定器
101B2 ビット数優先推定器
510、610 SHVCデコーダ
520、620 画像生成部
1001 プロセッサ
1002 プログラムメモリ
1003、1004 記憶媒体
Claims (8)
- 入力画像がダウンサンプルされた低解像度画像がベースレイヤとして符号化されたベースレイヤビットストリームと前記入力画像がエンハンスメントレイヤとして符号化されたエンハンスメントレイヤビットストリームとが多重化されたスケーラブルビットストリームを出力する映像符号化装置であって、
特定の矩形領域を含む、CTU(Coding Tree Unit)サイズの倍数の矩形領域を生成する矩形領域生成手段と、
符号化対象のCTUが前記CTUサイズの倍数の矩形領域に含まれるか否かを判定する判定手段と、
前記符号化対象のCTUが前記CTUサイズの倍数の矩形領域に含まれない場合、前記符号化対象のCTUを最小のCUブロック数で分割し、さらに、その各CUをゼロの動きベクトルのベースレイヤからの予測信号で予測する予測手段と、を有する
ことを特徴とする映像符号化装置。 - 前記予測手段は、ベースレイヤ予測に関連づけられた参照ピクチャインデックス、ゼロの動きベクトルに最も近いAMVP予測動きベクトルに関連づけられたAMVPインデックス、および、ゼロの動きベクトルから前記ゼロの動きベクトルに最も近いAMVP予測動きベクトルを減じた差分動きベクトルの組をブロック予測パラメータに含める
請求項1記載の映像符号化装置。 - 前記予測手段は、ベースレイヤ予測に関連づけられた参照ピクチャインデックスとゼロの動きベクトルに関連づけられたマージ候補インデックスとが存在する場合、マージ予測が有効であることを示すマージフラグと前記マージ候補インデックスをブロック予測パラメータに含める
請求項1または請求項2記載の映像符号化装置。 - 前記CTUサイズの倍数の矩形領域に含まれないCTUにおいて、予測誤差信号を強制的に0にする予測誤差切捨手段を有する
請求項1から請求項3のうちのいずれか1項に記載の映像符号化装置。 - 請求項1から請求項4のうちのいずれか1項に記載の映像符号化装置と、
前記映像符号化装置が出力するスケーラブルビットストリームを受信して復号する映像復号装置と、
復号された画像と特定の矩形領域を示す矩形領域情報とを含む画像を生成する画像生成手段とを備える
ことを特徴とする映像送受信システム。 - スケーラブルビットストリームの復号映像と矩形領域情報に基づいて表示映像を生成する表示映像生成装置であって、
映像復号装置と、画像生成手段とを備え、
ユーザが通常表示を指定した場合、前記映像復号装置が、スケーラブルビットストリームからベースレイヤビットストリームを復号し、前記画像生成手段が、ディスプレイサイズに拡大されたベースレイヤビットストリームの表示映像を生成し、
ユーザが詳細表示を指定した場合、前記映像復号装置が、スケーラブルビットストリームからベースレイヤビットストリームと前記矩形領域を含むエンハンスメントレイヤビットストリームを復号し、前記画像生成手段が、ベースレイヤビットストリームの復号映像と前記矩形領域を含むエンハンスメントレイヤビットストリームの復号映像を生成し、
ユーザが矩形領域表示を指定した場合、前記画像生成手段が、復号映像に前記矩形領域を重畳させる
ことを特徴とする表示映像生成装置。 - 入力画像がダウンサンプルされた低解像度画像がベースレイヤとして符号化されたベースレイヤビットストリームと前記入力画像がエンハンスメントレイヤとして符号化されたエンハンスメントレイヤビットストリームとが多重化されたスケーラブルビットストリームを出力する映像符号化装置における映像符号化方法であって、
特定の矩形領域を含む、CTUサイズの倍数の矩形領域を生成し、
符号化対象のCTUが前記CTUサイズの倍数の矩形領域に含まれるか否かを判定し、
前記符号化対象のCTUが前記CTUサイズの倍数の矩形領域に含まれない場合、前記符号化対象のCTUを最小のCUブロック数で分割し、さらに、その各CUをゼロの動きベクトルのベースレイヤからの予測信号で予測する
ことを特徴とする映像符号化方法。 - 入力画像がダウンサンプルされた低解像度画像がベースレイヤとして符号化されたベースレイヤビットストリームと前記入力画像がエンハンスメントレイヤとして符号化されたエンハンスメントレイヤビットストリームとが多重化されたスケーラブルビットストリームを出力する映像符号化装置におけるコンピュータに、
特定の矩形領域を含む、CTUサイズの倍数の矩形領域を生成する処理と、
符号化対象のCTUが前記CTUサイズの倍数の矩形領域に含まれるか否かを判定する処理と、
前記符号化対象のCTUが前記CTUサイズの倍数の矩形領域に含まれない場合、前記符号化対象のCTUを最小のCUブロック数で分割し、さらに、その各CUをゼロの動きベクトルのベースレイヤからの予測信号で予測する処理とを実行させる映像符号化プログラムを記憶するコンピュータ読み取り可能な記録媒体。
Priority Applications (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
JP2016527630A JPWO2015190078A1 (ja) | 2014-06-12 | 2015-06-05 | 映像符号化装置、映像符号化方法および映像符号化プログラム |
US15/314,551 US20170201773A1 (en) | 2014-06-12 | 2015-06-05 | Video coding apparatus, video coding method, and recording medium |
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
JP2014121635 | 2014-06-12 | ||
JP2014-121635 | 2014-06-12 |
Publications (1)
Publication Number | Publication Date |
---|---|
WO2015190078A1 true WO2015190078A1 (ja) | 2015-12-17 |
Family
ID=54833192
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
PCT/JP2015/002837 WO2015190078A1 (ja) | 2014-06-12 | 2015-06-05 | 映像符号化装置、映像符号化方法および記録媒体 |
Country Status (3)
Country | Link |
---|---|
US (1) | US20170201773A1 (ja) |
JP (1) | JPWO2015190078A1 (ja) |
WO (1) | WO2015190078A1 (ja) |
Families Citing this family (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
FR3029333A1 (fr) * | 2014-11-27 | 2016-06-03 | Orange | Procede de codage et decodage d'images, dispositif de codage et decodage et programmes d'ordinateur correspondants |
US9942548B2 (en) * | 2016-02-16 | 2018-04-10 | Google Llc | Entropy coding transform partitioning information |
WO2020220037A1 (en) * | 2019-04-25 | 2020-10-29 | Beijing Dajia Internet Information Technology Co., Ltd. | Methods and apparatuses for video coding with triangle prediction |
Citations (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP2009049979A (ja) * | 2007-07-20 | 2009-03-05 | Fujifilm Corp | 画像処理装置、画像処理方法、画像処理システム、及びプログラム |
Family Cites Families (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP6295951B2 (ja) * | 2012-06-25 | 2018-03-20 | ソニー株式会社 | 画像復号装置及び画像復号方法 |
EP2903285A4 (en) * | 2012-09-28 | 2016-06-01 | Sharp Kk | IMAGE DECODING DEVICE AND IMAGE CODING DEVICE |
US9599353B2 (en) * | 2013-07-26 | 2017-03-21 | Whirlpool Corporation | Split air conditioning system with a single outdoor unit |
-
2015
- 2015-06-05 US US15/314,551 patent/US20170201773A1/en not_active Abandoned
- 2015-06-05 WO PCT/JP2015/002837 patent/WO2015190078A1/ja active Application Filing
- 2015-06-05 JP JP2016527630A patent/JPWO2015190078A1/ja active Pending
Patent Citations (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP2009049979A (ja) * | 2007-07-20 | 2009-03-05 | Fujifilm Corp | 画像処理装置、画像処理方法、画像処理システム、及びプログラム |
Non-Patent Citations (3)
Title |
---|
CHEUNG AUYEUNG ET AL.: "A Use Case and High Level Syntax of SHVC:Scalability Based Region of Interest", JOINT COLLABORATIVE TEAM ON VIDEO CODING(JCT- VC) OF ITU-T SG 16 WP3 AND ISO/IEC JTC1/SC29/ WG11 13TH MEETING, pages 1 - 4 * |
DO-KYOUNG KWON ET AL.: "SCE2:Results of test 1.1 on skipped inter-layer texture prediction(ILTP)block signaling in Inter slice", JOINT COLLABORATIVE TEAM ON VIDEO CODING(JCT- VC) OF ITU-T SG 16 WP3 AND ISO/IEC JTC1/SC29/ WG11 13TH MEETING:INCHEON * |
TOMOYUKI YAMAMOTO ET AL.: "MV-HEVC/SHVC HLS:Skipped slice and use case", JOINT COLLABORATIVE TEAM ON VIDEO CODING(JCT- VC) OF ITU-T SG 16 WP3 AND ISO/IEC JTC1/SC29/ WG11 15TH MEETING, pages 1 - 6 * |
Also Published As
Publication number | Publication date |
---|---|
JPWO2015190078A1 (ja) | 2017-04-20 |
US20170201773A1 (en) | 2017-07-13 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
JP6432661B2 (ja) | 映像符号化装置、映像符号化方法及びプログラム | |
KR102078103B1 (ko) | 인코딩 방법 및 장치, 디코딩 방법 및 장치, 및 컴퓨터 판독가능 저장 매체 | |
JP6497423B2 (ja) | 映像復号装置、映像復号方法及びプログラム | |
KR102209109B1 (ko) | 영상 부호화/복호화 방법 및 장치 | |
JP7164710B2 (ja) | ビデオ復号化方法及びビデオ・デコーダ | |
CN103907346A (zh) | 用于三维视频编码与高效率视频编码的运动矢量及视差矢量导出方法及其装置 | |
CN104838657A (zh) | 可缩放及多视角视频译码中的高级残差预测 | |
CN103281531B (zh) | 面向hevc的质量可伸缩层间预测编码 | |
JP2015130688A (ja) | 動画像符号化装置、動画像符号化方法、動画像復号装置、動画像復号方法及び動画像符号化データ | |
KR20120103517A (ko) | 인트라 예측 방법 및 그 장치 | |
WO2015190078A1 (ja) | 映像符号化装置、映像符号化方法および記録媒体 | |
JP2008536393A (ja) | 少なくとも一のデジタル画像をエンコードする方法、エンコーダ、及びコンピュータプログラム製品 | |
KR102025413B1 (ko) | 복수의 레이어를 지원하는 영상의 부호화 및 복호화 방법 및 이를 이용하는 장치 | |
JP2018032913A (ja) | 映像符号化装置、プログラム及び方法、並びに、映像復号装置、プログラム及び方法、並びに、映像伝送システム | |
KR20100083957A (ko) | 비트플레인 코딩을 이용한 깊이 정보맵 코딩 방법 및 장치 | |
KR102602633B1 (ko) | 영상 부호화/복호화 방법 및 장치 | |
JPWO2016142977A1 (ja) | 映像符号化装置、映像符号化方法および映像符号化プログラム | |
KR20210013254A (ko) | 영상 부호화/복호화 방법 및 장치 | |
KR20150007205A (ko) | 영상의 복호화 방법 및 이를 이용하는 장치 |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
121 | Ep: the epo has been informed by wipo that ep was designated in this application |
Ref document number: 15806714 Country of ref document: EP Kind code of ref document: A1 |
|
ENP | Entry into the national phase |
Ref document number: 2016527630 Country of ref document: JP Kind code of ref document: A |
|
WWE | Wipo information: entry into national phase |
Ref document number: 15314551 Country of ref document: US |
|
NENP | Non-entry into the national phase |
Ref country code: DE |
|
122 | Ep: pct application non-entry in european phase |
Ref document number: 15806714 Country of ref document: EP Kind code of ref document: A1 |