WO2006073116A1 - 映像符号化方法及び装置、映像復号方法及び装置、それらのプログラムおよびそれらプログラムを記録した記録媒体 - Google Patents
映像符号化方法及び装置、映像復号方法及び装置、それらのプログラムおよびそれらプログラムを記録した記録媒体 Download PDFInfo
- Publication number
- WO2006073116A1 WO2006073116A1 PCT/JP2005/024093 JP2005024093W WO2006073116A1 WO 2006073116 A1 WO2006073116 A1 WO 2006073116A1 JP 2005024093 W JP2005024093 W JP 2005024093W WO 2006073116 A1 WO2006073116 A1 WO 2006073116A1
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- image
- information
- encoding
- video
- setting
- Prior art date
Links
Classifications
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/90—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using coding techniques not provided for in groups H04N19/10-H04N19/85, e.g. fractals
- H04N19/91—Entropy coding, e.g. variable length coding [VLC] or arithmetic coding
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N13/00—Stereoscopic video systems; Multi-view video systems; Details thereof
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/10—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
- H04N19/169—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding
- H04N19/17—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object
- H04N19/176—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object the region being a block, e.g. a macroblock
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/50—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding
- H04N19/503—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding involving temporal prediction
- H04N19/51—Motion estimation or motion compensation
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/50—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding
- H04N19/597—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding specially adapted for multi-view video sequence encoding
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/60—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using transform coding
- H04N19/61—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using transform coding in combination with predictive coding
Definitions
- the present invention relates to a video encoding method and apparatus used for encoding a multi-view video, a video encoding program used to realize the video encoding method, and a computer reading recording the program Readable recording medium, video decoding method and apparatus for decoding encoded data generated by the video encoding technique, video decoding program used to realize the video decoding method, and computer recording the program And a possible recording medium.
- Multi-viewpoint moving images are a plurality of moving images obtained by photographing the same subject and background with a plurality of cameras at various positions.
- a moving image taken with one camera is called a “two-dimensional moving image”
- a set of two-dimensional moving images taken of the same subject and background is called a “multi-view moving image”.
- Two-dimensional video images of each camera included in a multi-view video image have a strong correlation in the time direction.
- encoding can be performed using intra-frame correlation in an I frame, and encoding can be performed using inter-frame correlation in a plurality of past frames in a P frame.
- encoding can be performed using inter-frame correlation in a plurality of past frames in a P frame.
- the frame in the past or future frame Signs can be obtained using inter-frame correlation.
- the frame is divided into blocks (this block is a macroblock, the block size is 16 ⁇ 16), and intra prediction is performed in each macroblock.
- each macroblock can be divided into smaller blocks (hereinafter referred to as sub-blocks), and different intra prediction methods can be used for each sub-block.
- intra prediction or inter prediction can be performed in each macroblock.
- Intra prediction for P frames is the same as for I frames.
- motion compensation is performed during inter prediction. Even in motion compensation, a macroblock is divided into smaller blocks, and each subblock has a different motion vector and reference image.
- the ability to perform intra prediction and inter prediction also in the B frame can be used as a motion compensation reference image in addition to the past frame.
- the future frame can be used as a motion compensation reference image in addition to the past frame.
- encoding is performed in the order of I ⁇ P ⁇ B ⁇ B.
- motion compensation can be performed with reference to the I and P frames.
- a prediction residual is obtained when intra or inter prediction is performed, but quantization is performed by performing DCT (discrete cosine transform) on the prediction residual block in each macroblock. Specifically, a macro block having a block size of 16 ⁇ 16 is divided into 4 ⁇ 4 sub-blocks, and 4 ⁇ 4 DCT is performed.
- the quantized value sequence of DCT coefficients is expressed by the following information.
- -Coded block pattern Of the four 8 X 8 blocks that can be defined in a macroblock, there is a DCT coefficient that is non-zero in any block (hereinafter non-zero coefficient).
- Coded block flag Information indicating which 4 X 4 blocks have non-zero coefficients among the 4 4 X 4 blocks in an 8 X 8 block where non-zero coefficients exist
- Significance map Information indicating which of DCT coefficients existing in 4 X 4 block indicated by Coded block flag is non-zero.
- Level information The value of the non-zero coefficient indicated by the Significance map
- Encoding is completed by applying a lossless code that is called an entropy code to the method of dividing each macroblock into sub-blocks, motion vectors, etc., based on the information related to the above DCT coefficients. .
- auxiliary information the quantization value of the pixel region, or an entropy code other than the quantization value of the transform coefficient obtained by orthogonal transformation with respect to the image block (corresponding to the level information in the case of the DCT coefficient) Information that is the subject of the key is called “auxiliary information”.
- auxiliary information other than the DC T coefficient related auxiliary information includes, for example, the following. Such auxiliary information is given in units of macroblocks.
- the macroblock type is an index that specifies the combination of intra prediction or inter prediction for macroblocks, prediction method, block division format, etc. Is.
- the sub-macro block type is an index that expresses combinations of prediction methods in sub-blocks, block division formats, and so on.
- Reference image index The index value of the reference image used for motion compensation in each sub-block.
- Motion vector A motion vector in each sub-block. In H.264, it is expressed as a residual predicted by surrounding motion vectors.
- the entropy code ⁇ is a lossless code ⁇ .
- lossless encoding is a process of converting a symbol to be encoded (which can be interpreted as one value extracted from a set of integers) into a bit string of 1s and 0s. For example, if it is one value in the set of integers with symbol power S ⁇ 0, 1, 2, 3 ⁇ , the symbol power is 00, 1 for 01, and 1 for 2. If it is 10 or 3, encoding to 11 will result in lossless encoding.
- This encoding method is called fixed-length encoding.
- a set of codes for encoding symbols (in this example, ⁇ 00, 01, 10, 11 ⁇ ) is called a “code table”.
- the high-efficiency entropy code can be performed on the premise that the probability distribution of the symbol to be encoded is known. Therefore, conventionally, the probability distribution is determined empirically or the symbol probability distribution is learned while signing. In addition, the probability distribution power of symbols There are conventional methods for obtaining the optimum code table (Huffman code and arithmetic code). Therefore, we treat the probability distribution and the code table as synonymous below.
- CABAC Context—Adaptive Binary Arithmetic Coding
- CABAC CABAC
- the code table is switched with reference to the encoded macroblock types of the macroblock above and to the left.
- FIG. 17 shows a conceptual diagram of this reference relationship.
- the macroblock types of macroblocks described as A and B in the figure have a strong correlation with the target macroblock type.
- CABAC uses this correlation to predict an optimal code table. Specifically, for macroblocks A and B, a code table is maintained for each possible combination of macroblock type values, and the code table (probability) corresponding to the actual macroblock type values of A and B Distribution) is used to entropy code the target macroblock type. Based on the same idea, the other information about the sign Entropy encoded.
- a group of frames called "GOP" is classified into two types of GOPs, Base GOP and InterGOP, and encoded.
- the GOP called Base GOP all frames included in the GOP image of the same camera are encoded by intra or inter prediction.
- the frames included in the GOP called Inter GOP such intra and inter predictions are used.
- parallax direction prediction using GOP images from another camera may be used.
- the parallax direction prediction is to perform the same process as motion compensation using an image of another camera as a reference image when encoding a macroblock of an image of a certain camera.
- FIG. 18 shows an example of the GOP structure in this method.
- each GOP has two images, and each arrow represents a reference relationship between images.
- the code key efficiency can be improved as compared with the method using the correlation only in the time direction.
- Tokubari 1 Detlev Marpe, et.al., "Ontext— Based Adaptive Binary Arithmetic o ding in the H.264 / AVC Video Compression Standard, IEEE Transactions on Circuits and Systems for Video Technology, Vol. 13. No. 7, pp. 620—636, July, 2003.
- Non-Patent Document 2 Hideaki Kimataand Masaki Kitahara, “Preliminary results on multiple view video coding (3DAV),” document Ml 0976 MPEG Redmond Meeting, July, 2004 Disclosure of
- auxiliary information such as a prediction residual, its orthogonal transformation coefficient, a motion vector, and a macroblock type is correlated between cameras. Therefore, it is conceivable to use the correlation between cameras in the sign of these information.
- this is a method of encoding one two-dimensional moving image, assuming that the encoding method is such that the prediction error of motion compensation is quantized in the pixel region and the quantized value is entropy encoded.
- Reference 3 (Taichiro Shiojira, Ichiro Matsuda, Satoshi Ito, “Lossless video coding based on motion compensation and 3D prediction: Study on context modeling”, Information Science and Technology Forum (FIT 2003) It is conceivable to apply the method of No. J-053, pp. 303-304, Sep. 2003).
- the entropy code of the quantized value of the prediction residual is used to use the motion vector obtained in each block, and the code of the previous frame has already been completed.
- the code table is switched by referring to the quantized value.
- code target information other than the prediction residual in the pixel region for example, auxiliary information such as orthogonal transform coefficients, motion vectors, and macroblock types
- auxiliary information such as orthogonal transform coefficients, motion vectors, and macroblock types
- the present invention has been made in view of powerful circumstances, and when encoding a multi-view video, encoding is performed using the correlation of the parallax direction in the entropy code. In this way, the code efficiency cannot be improved by the conventional parallax direction prediction, and the code efficiency is improved for any multi-view video including multi-view video.
- An object of the present invention is to provide a new code technology that can be improved. Means for solving the problem
- the present invention is a video encoding method for encoding a plurality of images as one video
- reference disparity information indicating a viewpoint shift between the first image and the encoded second image is set.
- a code table setting step for setting a code table corresponding to the code object information at the corresponding position
- reference disparity information setting step global disparity information indicating an average viewpoint shift between the first and second images is set as the reference disparity information.
- the parallax between the two images is estimated using the second image and another third image, and the parallax and the imaging devices of the first to third images are estimated.
- the reference disparity information is set based on the positional relationship.
- the disparity information to be referred to (information indicating the magnitude of disparity such as a disparity vector) is not encoded, or data is encoded even if encoded.
- the amount of information can be reduced to a small amount, and moreover, by using the correlation of the parallax direction, the code information to be encoded can be encoded, so that a plurality of videos (for example, multi-view video images) can be displayed. It becomes possible to efficiently sign.
- the parallax information indicating the viewpoint shift between different cameras is estimated because the parallax information is obtained and encoded for each block, or one average parallax is obtained for one image.
- Information global disparity information
- each pixel of the encoding target image is not used without using the encoding target image (first image). It is possible to estimate disparity information (outside disparity).
- the image sequence at the same time of each camera is equivalent to shooting a moving image with the subject moving in parallel, so the reference difference information Local parallax information can be approximated only by
- the code table for the entropy code of the code key target image is switched with reference to the code key target information of another image.
- the code table setting step the encoding target information of the corresponding position in the second image and the position around the predetermined position in the first image are matched.
- the code table corresponding to the code key target information that has been encoded is set.
- the code table is set with reference to code key target information corresponding to positions around the corresponding position in the second image. According to this, it is possible to switch the robust code table.
- the encoding target information is set for each block obtained by dividing an image.
- the corresponding position setting step when the position of the block corresponding to the predetermined position in the second image is shifted based on the reference parallax information, the overlapping area with the block is maximized. Set the block position as the corresponding position.
- the code table setting means switches the code table by referring to the code key target information of the block at the corresponding position.
- the encoding target information is set for each block obtained by dividing an image
- the position of the block corresponding to the predetermined position in the second image is determined from the positions of a plurality of blocks around the shifted position based on the reference disparity information. Set as.
- the code table setting means switches the code table by referring to the code key target information of the plurality of blocks.
- an encoding pair used for setting the code table is used. It is also possible to set a code table corresponding to the mode value of the image information. By adopting this configuration, while preventing the explosive increase in the number of code tables, it is expressed as a numerical value such as a macro block type or prediction mode, but it is a numerical value as an index to the last. Therefore, it is possible to efficiently code the target information of the code key that cannot be defined.
- a code table corresponding to the average value of the code target information used for setting the code table may be set.
- the number of corresponding positions to be referenced is N
- the number of code tables when the number of corresponding positions to be referenced is 1 is c.
- the code table is switched with reference to the average value of the information to be encoded regarding a plurality of corresponding positions, so the number of code tables is only c.
- the present invention is also a video decoding method for decoding encoded data generated by encoding a plurality of images as one video,
- reference disparity information indicating a viewpoint shift between the first image and the encoded second image is set.
- An encoding target information decoding step for decoding the encoding target information in the first image using the code table
- a video decoding method is provided.
- the disparity between both images is estimated using the second image and another third image, and the disparity and the first to third
- the reference parallax information is set based on the positional relationship between the image capturing devices.
- the code table setting step encoding target information of the corresponding position in the second image and decoding for a position around the predetermined position in the first image.
- a code table corresponding to the already completed code target information is set.
- the code table is set with reference to code key target information corresponding to positions around the corresponding position in the second image.
- the encoding target information is set for each block obtained by dividing an image.
- the position of the block corresponding to the predetermined position in the second image is referred to.
- the position of the block that has the largest overlap area with the block when shifted is set as the corresponding position.
- the encoding target information is set for each block obtained by dividing an image.
- the position of the block corresponding to the predetermined position in the second image is determined from the positions of a plurality of blocks around the shifted position based on the reference disparity information. Set as.
- an encoding pair used for setting the code table is used.
- a code table corresponding to the mode value of the image information may be set, or a code table corresponding to the average value of the code target information used for setting the code table may be set.
- the present invention is also a video encoding device for encoding a plurality of images as one video,
- reference disparity information indicating a viewpoint shift between the first image and the encoded second image is set.
- a reference disparity information setting unit
- a corresponding position setting unit that sets a corresponding position associated with the predetermined position in the second image using the reference parallax information
- a code table setting unit for setting a code table corresponding to code target information at the corresponding position
- An encoding target information encoding unit that encodes the encoding target information in the first image using the code table
- a video encoding device is provided.
- the reference disparity information setting unit sets, as the reference disparity information, global disparity information indicating an average viewpoint shift between the first and second images and encodes the global disparity information. It has a global parallax information code key part to be displayed.
- the reference parallax information setting unit estimates the parallax between both images using the second image and another third image, and the parallax and the first to third The reference parallax information is set on the basis of the positional relationship between the image capturing devices of the image.
- the present invention is also a video decoding apparatus for decoding encoded data generated by encoding a plurality of images as one video,
- reference disparity information indicating a viewpoint shift between the first image and the encoded second image is set.
- the second image is associated with the predetermined position.
- a corresponding position setting unit for setting a corresponding position
- a code table setting unit for setting a code table corresponding to code target information at the corresponding position
- An encoding target information decoding unit that decodes the encoding target information in the first image using the code table
- a video decoding device is provided.
- the reference disparity information setting unit decodes global disparity information indicating an average viewpoint shift between the first and second images, and sets this as the reference disparity information.
- the reference parallax information setting unit estimates the parallax between both images using the second image and another third image, and the parallax and the first to third
- the reference parallax information is set on the basis of the positional relationship between the image capturing devices of the image.
- Each method of the present invention can be realized by a computer program, and this computer program is provided by being recorded on a suitable computer-readable recording medium or provided via a network.
- the present invention can be realized by installing and operating on control means such as a CPU.
- the code efficiency cannot be improved by the parallax direction prediction. Since the correlation between the parallax directions can be used, the code efficiency can be improved for all multi-view video images.
- FIG. 1A is a diagram showing an example of a reference relationship of encoding target information for switching code tables.
- FIG. 1B is a diagram illustrating an example of a reference relationship of code target information for switching code tables.
- FIG. 2A is a diagram showing an example of a reference relationship of encoding target information for switching code tables.
- FIG. 2B is a diagram illustrating an example of a reference relationship of code key target information for switching code tables.
- FIG. 3 is a conceptual diagram showing a configuration of a camera in an embodiment of the present invention.
- FIG. 4 is a diagram showing a configuration of a video encoding device in the same embodiment.
- FIG. 5 A diagram showing the configuration of the code object information reference code part in the embodiment.
- ⁇ 6 A diagram showing an operation flow executed by the video encoding device in the same embodiment.
- ⁇ 7 A diagram showing an operation flow of an encoding process of encoding target information in the image of the camera C2 in the embodiment. It is.
- FIG. 8A is a diagram showing an example of a reference relationship of encoding target information for switching code tables.
- FIG. 8B is a diagram illustrating an example of a reference relationship of code target information for switching code tables.
- FIG. 9 is a diagram showing a configuration of a video decoding device in the embodiment.
- FIG. 10 is a diagram showing a configuration of a code key target information reference decoding unit in the same embodiment.
- FIG. 11 is a diagram showing an operation flow executed by the video decoding device in the embodiment.
- FIG. 12 is a diagram showing an operation flow of the decoding process of the encoding target information in the image of the camera C2 in the embodiment.
- FIG. 13A is a diagram showing an example of a reference relationship of encoding target information for switching code tables in the embodiment.
- FIG. 13B is a diagram illustrating an example of a reference relationship of code key target information for switching code tables.
- FIG. 14 is a diagram illustrating a principle of estimating disparity information (disparity vector) regarding each pixel of an image without using an encoding target image.
- FIG. 15 is a diagram showing an operation flow of a coding process for coding target information in an image of a camera C2 when parallax information is estimated.
- FIG. 16 is a diagram showing an operation flow of a decoding process of encoding target information in an image of a camera C2 when disparity information is estimated.
- FIG. 17 is a diagram showing a reference relationship of macroblocks for switching code tables in H.264.
- FIG. 18 is a diagram showing an example of a conventional GOP structure.
- FIGS. 1A and 1B and FIGS. 2A and 2B show an example of a reference relationship of code key target information for switching code tables.
- a global disparity vector global disparity information
- the block 502 to which the encoding target information currently being encoded in the camera C2 belongs corresponds to the position specified by shifting the position 501 in the image of the corresponding camera C 1 by the global parallax vector in the image of the camera C 1.
- the code table is switched according to the code key target information of two blocks (the block described as C in the figure), and the code key target information is encoded using the code table.
- the position 501 in the image of the camera C1 corresponding to the position of the block 502 corresponds to the position specified by shifting the position in the image of the camera C1 by the global parallax vector.
- the position of the block (for example, C) having the largest overlap area with the block 501 is set as the corresponding position.
- the block 502 to which the encoding target information that is the current encoding target in the camera C2 belongs (in this example, block-based encoding such as MPEG-2 is performed). Assuming that the position 501 in the image of the camera C 1 corresponding to the position of the camera C 1 is shifted by the global disparity vector in the image of the camera C1
- the code table is switched in accordance with the code target information of the blocks described as C, D, E, and F, and the code target information is encoded using the code table.
- FIG. 3 shows a conceptual diagram of a camera configuration used in the present embodiment.
- the square figure shown in the figure represents the frame (image) of each camera.
- each row is left-force scanned and processed from the top row.
- the code key target information called a macroblock type is encoded.
- a context-dependent entropy code that uses encoded information about the block corresponding to the encoding target information is used.
- context-dependent entropy coding using the encoded information subject information and the encoded subject information of the frame of another camera is performed. It is assumed that appropriate code tables are prepared in advance.
- FIG. 4 shows the configuration of the video encoding device 1 used in this embodiment.
- the video encoding device 1 shown in this figure includes an image information input unit 10 for inputting the frames of the cameras CI and C2 in the order shown in FIG. 3, an image memory 11 for storing the frames of the camera C1, and a camera.
- Reference disparity information setting unit 12 for obtaining reference disparity information (global disparity vector in this embodiment) between frames of the camera C1 and the camera C2, and code information target information for encoding the information to be encoded by the camera C1
- Non-reference code key unit 13 encoding target information memory 14 for storing block information (encoding target information) of camera C1, and encoding target information reference code for encoding encoding target information of camera C2
- a global parallax code key unit 16 that codes the global parallax vector.
- FIG. 5 shows a detailed configuration of the encoding target information reference encoding unit 15 in the video encoding device 1.
- the encoding target information reference encoding unit 15 performs the global disparity beta.
- the corresponding position setting unit 150 for obtaining the corresponding position from the camera, the reference parameter setting unit 151 for setting the reference parameter for setting the code table based on the encoding target information of the camera C 1 and the camera C2, and the reference parameter.
- a code table setting unit 152 for setting a code table based on it, an encoding target information setting unit 153 for setting encoding target information from an image of camera C2, and an encoding target for encoding encoding target information of camera C2.
- an information encoding unit 154 for setting a code table based on it, an encoding target information setting unit 153 for setting encoding target information from an image of camera C2, and an encoding target for encoding encoding target information of camera C2.
- FIG. 6 shows an operation flow executed by the video coding apparatus 1 configured as described above.
- the encoding target information non-reference encoding unit 13 encodes the encoding target information of each block (S11ZS12).
- the image memory 11 has a memory for one frame, and the latest input frame of the camera C1 is always stored.
- the encoding target information reference encoding unit 15 encodes the encoding target information of each block (S11ZS13). Note that the global disparity vector obtained in advance by the reference disparity information setting unit 12 is used in the sign of the target information of the frame of the camera C2.
- FIG. 7 shows a detailed operation flow of the code key processing of the code key target information of the camera C2.
- the reference parallax information setting unit 12 receives an image of the camera C1 from the image memory 11, and receives an image of the camera C2 from the image information input unit 10. Then, the global disparity vector is obtained by the reference disparity information setting unit 12 (S130).
- the global parallax vector is calculated as an average vector of parallax vectors for each block (for example, macroblock).
- the reference disparity information setting unit 12 performs block matching on each block of the camera C2 based on the following evaluation function.
- the global disparity vector obtained in this way is input to the global disparity code encoding unit 16 and encoded by the global disparity encoding unit 16 (S131).
- the sign of the global disparity vector is assumed to be an irreversible code in the present embodiment.
- the quantization step size is quantized by 1 (equivalent to mapping to the nearest integer) and entropy-coded.
- the encoding target information setting unit 153 obtains the encoding target information (S133). More specifically, information such as motion vectors by motion compensation, MBtype, and encoding target information such as orthogonal transform coefficients are required.
- the encoding target information encoding section 154 encodes the encoding target information.
- the corresponding position setting unit 150 uses the global disparity vector that has been encoded and decoded to determine the force that the corresponding position exists in the frame of the camera C1 (S134).
- the corresponding position is obtained by shifting the position 501 in the image of the camera C 1 corresponding to the block b in the image of the camera C2 by the global parallax vector.
- the position of the four blocks closest to the selected position Specifically, when the position of the upper left corner of the current block b is I, the position of the upper left corner of the block in the image of camera C1 is I +
- the reference parameter setting unit 151 obtains a set of MBtypes of blocks described as A and B in FIG. 8B. For example, a set of MBtypes such as (0, 5) is obtained.
- the code table setting unit 152 sets a code table corresponding to this (0, 5) set of MBtypes for entropy code (S 137).
- the code table used here is the same as that used in camera C1. Also, the code table setting unit 152 holds a code table for each possible combination of MBtype values.
- the code key target information code key unit 154 outputs a code corresponding to the MBtype of the current block in the code table (S138). That is, the encoding target information of the current block is encoded.
- the reference parameter setting unit 151 determines that the corresponding position exists (the blocks described as C, D, E, and F in FIG. 8A), the reference parameter setting unit 151 The mode value of the MBtype related to the corresponding position is obtained, and the final MBtype pair is obtained based on the MBtype and the MBtype pair of blocks described as A and B in Fig. 8B. That is, reference is made to the encoded code target information around the block b that is currently being encoded in the encoding target image and the encoded target information of the corresponding position in the encoded reference image. It is done. For example, if the MBtypes at the four corresponding positions are (0, 5, 5, 4), the mode value is 5. In addition, if the combination of MBtypes of the blocks described as A and B in FIG. 8 is (0, 5), the reference parameter setting unit 151 finally (0, 5) based on these MBtypes. , 5) And!
- the code table setting unit 152 sets a code table corresponding to the MBtype set (0, 5, 5) for the entropy code (S 135).
- the code key target information code key unit 154 In response to the setting of the code table, the code key target information code key unit 154 outputs a code corresponding to the MBtype of the current block in the code table (S136).
- the video encoding device 1 of the present invention performs processing so as to encode the encoding target information that is the target of the enthalpy encoding using the correlation of the parallax directions. .
- FIG. 9 shows the configuration of the video decoding device 2 used in this embodiment.
- the video decoding device 2 shown in this figure accumulates the encoding target information non-reference decoding unit 20 that decodes the encoding target information related to the camera C1, and the encoded target information decoded for the camera C1.
- An image information restoration unit 24 that restores an image based on! /
- an image information output unit 25 that outputs the restored image.
- FIG. 10 shows a detailed configuration of the encoding target information reference decoding unit 22 in the video decoding device 2.
- the encoding target information reference decoding unit 22 includes an encoding target information decoding unit 220 that decodes the encoding target information of the camera C2, and a corresponding position from the decoded global disparity vector.
- a code table setting unit 223 for setting a table.
- FIG. 11 shows an operation flow executed by the video decoding apparatus 2 configured as described above.
- the code data strengths of the cameras C1 and C2 are input in the order shown in FIG. 3 (S20).
- the encoded data of the camera C1 is input to the encoding target information non-reference decoding unit 20, and the encoded data of the camera C2 is input to the encoding target information reference decoding unit 22.
- the encoded data of the global parallax vector is also input to the global parallax decoding unit 23.
- the encoding target information non-reference decoding unit 20 decodes the encoding target information of each block (S21ZS22).
- the encoding target information reference decoding unit 22 decodes the encoding target information of each block (S21ZS23).
- FIG. 12 shows a detailed operation flow of the decoding process of the encoding target information of the camera C2.
- the encoded data of the global disparity vector is input to the global disparity decoding unit 23, and the global disparity decoding unit 23 decodes the global disparity vector. (S230).
- the corresponding position setting unit 221 determines whether a corresponding position corresponding to the current block b in the frame of the camera C2 exists in the frame of the camera C1. (S232).
- the method for determining the corresponding position is the same as in encoding.
- the code table is switched by referring only to the decoded code target information of the camera C2, and the code number of the current block is changed.
- the target information is decrypted (S235ZS236).
- the reference relationship is the same as in the case of sign-up.
- the reference parameter setting unit 222 and the code table setting unit 223 are the same as the reference parameter setting unit 151 and the code table setting unit 152 performed when the corresponding position exists in the video encoding device 1. Operates to decode the MBtype of the current block.
- the image information restoring unit 24 restores the image information of the block based on these pieces of information and outputs the image information from the image information output unit 25. (S237).
- the video decoding device 2 of the present invention performs processing so as to decode the encoded data generated by the video encoding device 1 of the present invention using the correlation in the parallax direction.
- the code processing in the above-described embodiment relates to the macroblock type and the entropy of the information !, but the present invention is subject to other entropy codes.
- the present invention can be applied as it is to the entropy code of information. Some examples are given below.
- the information that is the target of the entropy code is all expressed as a bit string, and is entropy-encoded bit by bit with a context-dependent arithmetic code. That is, when applying the present invention, the entropy code for each bit is used. It can be used to switch the code table (specifically, it is a probability table because it is an arithmetic code).
- Coded block pattern is information indicating which block has a non-zero coefficient in any of the four 8 X 8 blocks in the macroblock. Specifically, it is expressed as a total of 4 bits of information that represents the presence or absence of non-zero coefficients in the 8 X 8 block by 1 bit.
- FIGS. 8A and 8B are diagrams showing the positional relationship of blocks corresponding to FIGS. 8A and 8B, respectively.
- code tables corresponding to all combinations of 8 8 blocks 8, B, and C are prepared, and A, B, C most suitable for the 8 x 8 block b to be encoded.
- the Coded block flag and!? Information indicate which block has non-zero coefficients in the four 4 X 4 blocks in the 8 X 8 block where non-zero coefficients exist. It is information to show. Therefore, the Coded block flag is information in the same format as the Coded block pattern, and the application of the present invention can be performed in the same form.
- the significance map is information indicating which DCT coefficient is non-zero in the 4 ⁇ 4 block indicated by the Coded block flag. Specifically, 1 bit indicates whether a certain DCT coefficient is non-zero, 16 DCT coefficients in the 4 X 4 block are scanned in a certain order, and the DCT coefficient is non-zero. If yes, set to 1, otherwise set to 0. These bits are called Significance bits.
- the DCT coefficient is non-zero and set to 1
- the DCT coefficient is represented by 1 bit as to whether or not the DCT coefficient was the last non-zero coefficient in the scan. Value 1 if yes, value 0 if not) Include in Significance map. These bits are called the Last bit.
- the significan ce bit for subsequent DCT coefficients is not required. Also, whether or not the last DCT coefficient in the scan is non-zero is obvious from the previous Significance and Last bits, so information is particularly necessary for this DCT coefficient. .
- the power code table that independently encodes the Significance bit and the Last bit is switched according to the position in the scan (as described above, information is not available for the last DCT coefficient). Because it is not necessary, 15 code tables are required).
- the present invention for example, as shown in FIGS. 13A and 13B, it is conceivable to switch the code table with reference to the number of non-zero coefficients in the 4 ⁇ 4 block C of another camera.
- the code table used at this time is switched depending on the number of quantized values having an absolute value greater than 1 that occurred during the previous scan in the block. If the absolute value is larger than 1 at a certain point in the scan! / And many quantized values are generated, there is a probability that the quantized value will have a large value at that point. Therefore, a code table that reflects this property is prepared.
- the present invention has been described according to the embodiment, the present invention is not limited to this.
- images were taken by two cameras, cameras CI and C2.
- the present invention has been described with respect to an image as a processing target, the present invention can be directly applied to an image captured by three or more cameras.
- the present invention can also be realized by inputting disparity information obtained externally in advance using various disparity estimation methods and performing encoding processing using the disparity information as reference disparity. .
- a disparity vector is used as the disparity information.
- the present invention may use information other than the disparity vector as the disparity information.
- the video encoding device determines the parameters of the glossy parallax model and encodes these parameters. Also, when obtaining the corresponding position, these parameters are applied to the global parallax model to obtain the global parallax vector, and the corresponding position is determined. On the other hand, in the video decoding apparatus, the corresponding position is determined by applying the decoded parameters of the global parallax model to the global parallax model.
- the force described when there are two cameras and the decoded image of one camera is referred to.
- Camera parameters of each camera (camera position and orientation, lens angle of view, etc.) ) Is known and the decoded images of two or more different cameras can be referred to, it is possible to refer to each pixel of the image to be encoded by the method shown in FIG. 14 without using the image to be encoded.
- Disparity information (disparity vector) can be estimated.
- the viewpoint position of the camera C2 that captures the sign ⁇ contrast image is 620, and two other cameras that capture the subject 600 (that is, output the reference image) C 1, C3
- the viewpoint positions are 610 and 630, respectively.
- the pixel position for which the difference is to be obtained in the encoded reference image is 640.
- the pixel value of the corresponding point corresponding to the pixel position 640 in each reference image should be almost the same value. Therefore, in many parallax estimation methods, the pixel values of corresponding points of each reference image are compared with respect to various depths, and the closest depth (indicated by reference numeral 650 in FIG. 14) and the known camera parameters Based on this, the parallax from the camera C1 or C3 can be estimated. This process can be performed in units of pixels of the encoding target image.
- the global disparity code encoding unit 16 of the video encoding device 1 and the global disparity decoding unit 23 of the video decoding device 2 are unnecessary.
- the code may be performed by the same process as in the above embodiment. Specifically, the process shown in FIG. 15 is performed in the encoding process of the encoding target information.
- step S1300 is provided, and in this step, the cameras CI and C3 captured at the same time as the target images captured by the camera C2 are provided.
- the parallax from the camera C1 or C3 is estimated by the method described above.
- step S2300 is provided in place of the processing of step S230 in FIG. 12, and in this step, decoding of the images of the cameras CI and C3 captured at the same time as the target image captured by the camera C2 is performed. Using the image, the parallax from the camera C1 or C3 is estimated by the same method.
Landscapes
- Engineering & Computer Science (AREA)
- Multimedia (AREA)
- Signal Processing (AREA)
- Compression Or Coding Systems Of Tv Signals (AREA)
- Compression, Expansion, Code Conversion, And Decoders (AREA)
- Testing, Inspecting, Measuring Of Stereoscopic Televisions And Televisions (AREA)
Abstract
Description
Claims
Priority Applications (3)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US10/586,235 US9088802B2 (en) | 2005-01-07 | 2005-12-28 | Video encoding method and apparatus, video decoding method and apparatus, programs therefor, and storage media for storing the programs |
EP05822685.3A EP1835747B1 (en) | 2005-01-07 | 2005-12-28 | Video encoding method and device, video decoding method and device, program thereof, and recording medium containing the program |
JP2006519047A JP4937741B2 (ja) | 2005-01-07 | 2005-12-28 | 映像符号化方法及び装置、映像復号方法及び装置、それらのプログラムおよびそれらプログラムを記録した記録媒体 |
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
JP2005-002423 | 2005-01-07 | ||
JP2005002423 | 2005-01-07 |
Publications (1)
Publication Number | Publication Date |
---|---|
WO2006073116A1 true WO2006073116A1 (ja) | 2006-07-13 |
Family
ID=36647597
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
PCT/JP2005/024093 WO2006073116A1 (ja) | 2005-01-07 | 2005-12-28 | 映像符号化方法及び装置、映像復号方法及び装置、それらのプログラムおよびそれらプログラムを記録した記録媒体 |
Country Status (7)
Country | Link |
---|---|
US (1) | US9088802B2 (ja) |
EP (1) | EP1835747B1 (ja) |
JP (1) | JP4937741B2 (ja) |
KR (1) | KR100946790B1 (ja) |
CN (1) | CN100584013C (ja) |
TW (1) | TWI308028B (ja) |
WO (1) | WO2006073116A1 (ja) |
Cited By (19)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
WO2008035654A1 (fr) * | 2006-09-20 | 2008-03-27 | Nippon Telegraph And Telephone Corporation | Procédés et dispositifs de codage et de décodage d'image, dispositif et programmes de décodage d'image, et support de stockage desdits programmes |
WO2008035665A1 (fr) * | 2006-09-20 | 2008-03-27 | Nippon Telegraph And Telephone Corporation | procédé DE CODAGE D'IMAGE, PROCÉDÉ DE DÉCODAGE, DISPOSITIF associÉ, DISPOSITIF DE DÉCODAGE D'IMAGE, programme associÉ, et support de stockage contenant le programme |
WO2008053746A1 (fr) * | 2006-10-30 | 2008-05-08 | Nippon Telegraph And Telephone Corporation | Procédé de génération d'informations de référence prédictives, procédé de codage et de décodage d'image dynamiques, leur dispositif, leur programme et support de stockage contenant le programme |
WO2008053758A1 (fr) * | 2006-10-30 | 2008-05-08 | Nippon Telegraph And Telephone Corporation | Procédé de codage d'image dynamique, procédé de décodage, leur dispositif, leur programme et support de stockage contenant le programme |
WO2008047300A3 (en) * | 2006-10-16 | 2008-07-03 | Nokia Corp | System and method for using parallelly decodable slices for multi-view video coding |
JP2009502084A (ja) * | 2005-07-18 | 2009-01-22 | エレクトロニクス アンド テレコミュニケーションズ リサーチ インスチチュート | 時間及び視点間参照映像バッファを活用した予測符号化/復号化装置及びその方法 |
JP2010515400A (ja) * | 2007-01-03 | 2010-05-06 | サムスン エレクトロニクス カンパニー リミテッド | 全域差ベクトルを利用した多視点映像の符号化、復号化方法及び装置 |
JP2010516158A (ja) * | 2007-01-11 | 2010-05-13 | サムスン エレクトロニクス カンパニー リミテッド | 多視点映像の符号化、復号化方法及び装置 |
WO2011083520A1 (ja) * | 2010-01-08 | 2011-07-14 | 富士通株式会社 | 動画像符号化装置および動画像復号装置 |
JP2012505457A (ja) * | 2008-10-08 | 2012-03-01 | ミツビシ・エレクトリック・アールアンドディー・センター・ヨーロッパ・ビーヴィ | マルチメディアシグネチャのための符号化及び復号の方法及び装置 |
RU2444856C2 (ru) * | 2007-06-25 | 2012-03-10 | Ниппон Телеграф Энд Телефон Корпорейшн | Способ кодирования видеосигнала и способ декодирования, устройства для этого и носители хранения информации, которые хранят программы для этого |
RU2446618C2 (ru) * | 2006-12-28 | 2012-03-27 | Ниппон Телеграф Энд Телефон Корпорейшн | Способ кодирования и способ декодирования видео, устройства для их осуществления и носители памяти, хранящие программы |
WO2013002342A1 (ja) * | 2011-06-30 | 2013-01-03 | ソニー株式会社 | 画像処理装置および方法 |
RU2480941C2 (ru) * | 2011-01-20 | 2013-04-27 | Корпорация "Самсунг Электроникс Ко., Лтд" | Способ адаптивного предсказания кадра для кодирования многоракурсной видеопоследовательности |
US8520961B2 (en) | 2006-10-13 | 2013-08-27 | Samsung Electronics Co., Ltd. | Method and apparatus for encoding and decoding multi-view image |
US8532412B2 (en) | 2007-04-11 | 2013-09-10 | Samsung Electronics Co., Ltd. | Method and apparatus for encoding and decoding and multi-view image |
WO2013146507A1 (ja) * | 2012-03-30 | 2013-10-03 | ソニー株式会社 | 画像処理装置および方法、並びに記録媒体 |
RU2525850C2 (ru) * | 2009-02-12 | 2014-08-20 | Ниппон Телеграф Энд Телефон Корпорейшн | Способ кодирования многовидового изображения, способ декодирования многовидового изображения, устройство кодирования многовидового изображения, устройство декодирования многовидового изображения, компьютерно-читаемый носитель, содержащий программу кодирования многовидового изображения и компьютерно-читаемый носитель, содержащий программу декодирования многовидового изображения |
US9094687B2 (en) | 2009-12-28 | 2015-07-28 | Fujitsu Limited | Video encoder and video decoder |
Families Citing this family (21)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
EP1703513A1 (en) * | 2005-03-15 | 2006-09-20 | Deutsche Thomson-Brandt Gmbh | Method and apparatus for encoding plural video signals as a single encoded video signal, method and and apparatus for decoding such an encoded video signal |
KR100867995B1 (ko) * | 2006-01-07 | 2008-11-10 | 한국전자통신연구원 | 동영상 데이터 인코딩/디코딩 방법 및 그 장치 |
WO2007110000A1 (en) * | 2006-03-29 | 2007-10-04 | Huawei Technologies Co., Ltd. | A method and device of obtaining disparity vector and its multi-view encoding-decoding |
JP2008227689A (ja) * | 2007-03-09 | 2008-09-25 | Seiko Epson Corp | 符号化装置及び画像記録装置 |
TWI375470B (en) * | 2007-08-03 | 2012-10-21 | Via Tech Inc | Method for determining boundary strength |
KR101511082B1 (ko) * | 2008-05-09 | 2015-04-13 | 삼성전자주식회사 | 최하위 비트를 이용한 엔트로피 부호화 방법과 그 장치 및엔트로피 복호화 방법과 그 장치 |
US20120320985A1 (en) * | 2010-02-09 | 2012-12-20 | Nippon Telegraph And Telephone Corporation | Motion vector predictive encoding method, motion vector predictive decoding method, moving picture encoding apparatus, moving picture decoding apparatus, and programs thereof |
RU2519525C2 (ru) | 2010-02-09 | 2014-06-10 | Ниппон Телеграф Энд Телефон Корпорейшн | Способ кодирования с предсказанием вектора движения, способ декодирования с предсказанием вектора движения, устройство кодирования фильма, устройство декодирования фильма и их программы |
ES2652337T3 (es) | 2010-02-09 | 2018-02-01 | Nippon Telegraph And Telephone Corporation | Procedimiento de codificación predictiva para vector de movimiento, procedimiento de decodificación predictiva para vector de movimiento, dispositivo de codificación de imagen, dispositivo de decodificación de imagen, y programas para ello |
KR101829594B1 (ko) * | 2010-09-30 | 2018-02-14 | 미쓰비시덴키 가부시키가이샤 | 동화상 복호 장치, 동화상 복호 방법, 동화상 부호화 장치, 동화상 부호화 방법 및 기록 매체 |
US8976861B2 (en) * | 2010-12-03 | 2015-03-10 | Qualcomm Incorporated | Separately coding the position of a last significant coefficient of a video block in video coding |
US20120163456A1 (en) | 2010-12-22 | 2012-06-28 | Qualcomm Incorporated | Using a most probable scanning order to efficiently code scanning order information for a video block in video coding |
BR112013021276A2 (pt) * | 2011-02-23 | 2018-06-12 | Koninklijke Philips Nv | dispositivo e método de processamento de sinal para processar dados da imagem tridimensional, gerador de ponto de vista, gerador dos dados de profundidade e programa de computador |
TW201245768A (en) | 2011-03-29 | 2012-11-16 | Sony Corp | Image pickup apparatus, image pickup device, image processing method, aperture control method, and program |
KR20130049524A (ko) * | 2011-11-04 | 2013-05-14 | 오수미 | 인트라 예측 블록 생성 방법 |
TWI618401B (zh) * | 2011-12-28 | 2018-03-11 | Jvc Kenwood Corp | Motion picture coding device, motion picture coding method and memory medium |
JP6080405B2 (ja) * | 2012-06-29 | 2017-02-15 | キヤノン株式会社 | 画像符号化装置、画像符号化方法及びプログラム、画像復号装置、画像復号方法及びプログラム |
US9544566B2 (en) * | 2012-12-14 | 2017-01-10 | Qualcomm Incorporated | Disparity vector derivation |
US10027947B2 (en) * | 2013-06-05 | 2018-07-17 | Sony Corporation | Image processing apparatus and image processing method |
CN111200736B (zh) * | 2013-06-14 | 2022-06-21 | 北京三星通信技术研究有限公司 | 一种视频编/解码中运动信息的获取方法及装置 |
US10904570B2 (en) * | 2016-09-08 | 2021-01-26 | Kaonmedia Co., Ltd. | Method for encoding/decoding synchronized multi-view video by using spatial layout information and apparatus of the same |
Citations (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JPH06225274A (ja) * | 1993-01-21 | 1994-08-12 | Oki Electric Ind Co Ltd | 画像符号化及び復号化装置 |
JPH09261653A (ja) * | 1996-03-18 | 1997-10-03 | Sharp Corp | 多視点画像符号化装置 |
JPH09275578A (ja) * | 1996-04-04 | 1997-10-21 | Sharp Corp | 多視点画像符号化装置および復号装置 |
WO2003063501A1 (en) * | 2002-01-22 | 2003-07-31 | Nokia Corporation | Coding transform coefficients in image/video encoders and/or decoders |
Family Cites Families (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
TW498682B (en) | 2000-03-01 | 2002-08-11 | Soft4D Co Ltd | Method and apparatus for generating sterescopic image using MPEG data |
US6690307B2 (en) | 2002-01-22 | 2004-02-10 | Nokia Corporation | Adaptive variable length coding of digital video |
EP1422928A3 (en) * | 2002-11-22 | 2009-03-11 | Panasonic Corporation | Motion compensated interpolation of digital video signals |
JP3746506B2 (ja) | 2004-03-08 | 2006-02-15 | 一成 江良 | 立体視化パラメータ埋込装置及び立体視画像再生装置 |
-
2005
- 2005-12-28 KR KR20067014435A patent/KR100946790B1/ko active IP Right Grant
- 2005-12-28 CN CN200580002762A patent/CN100584013C/zh active Active
- 2005-12-28 EP EP05822685.3A patent/EP1835747B1/en active Active
- 2005-12-28 US US10/586,235 patent/US9088802B2/en active Active
- 2005-12-28 WO PCT/JP2005/024093 patent/WO2006073116A1/ja active Application Filing
- 2005-12-28 JP JP2006519047A patent/JP4937741B2/ja active Active
-
2006
- 2006-01-05 TW TW95100423A patent/TWI308028B/zh active
Patent Citations (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JPH06225274A (ja) * | 1993-01-21 | 1994-08-12 | Oki Electric Ind Co Ltd | 画像符号化及び復号化装置 |
JPH09261653A (ja) * | 1996-03-18 | 1997-10-03 | Sharp Corp | 多視点画像符号化装置 |
JPH09275578A (ja) * | 1996-04-04 | 1997-10-21 | Sharp Corp | 多視点画像符号化装置および復号装置 |
WO2003063501A1 (en) * | 2002-01-22 | 2003-07-31 | Nokia Corporation | Coding transform coefficients in image/video encoders and/or decoders |
Non-Patent Citations (3)
Title |
---|
KIMATA H., ET AL.: "System Design of Free Viewpoint Video Communication", THE FOURTH INTERNATIONAL CONFERENCE ON COMPUTER AND INFORMATION TECHNOLOGY 2004 (CIT '04), 14 September 2004 (2004-09-14), pages 52 - 59, XP010743221 * |
See also references of EP1835747A4 * |
SHIODERA T. ET AL: "Ugoki Hosho to 3-Jigen Yosoku ni Motoduita Dogazo no Kagyaku Fugoka ~Context Modeling ni Kansuru Kento~", FORUM ON INFORMATION TECHNOLOGY (FIT 2003) KOEN RONBUN, no. J-053, September 2003 (2003-09-01), pages 303 - 304, XP003000402 * |
Cited By (54)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US8369406B2 (en) | 2005-07-18 | 2013-02-05 | Electronics And Telecommunications Research Institute | Apparatus of predictive coding/decoding using view-temporal reference picture buffers and method using the same |
JP2009502084A (ja) * | 2005-07-18 | 2009-01-22 | エレクトロニクス アンド テレコミュニケーションズ リサーチ インスチチュート | 時間及び視点間参照映像バッファを活用した予測符号化/復号化装置及びその方法 |
US9154786B2 (en) | 2005-07-18 | 2015-10-06 | Electronics And Telecommunications Research Institute | Apparatus of predictive coding/decoding using view-temporal reference picture buffers and method using the same |
US8385628B2 (en) | 2006-09-20 | 2013-02-26 | Nippon Telegraph And Telephone Corporation | Image encoding and decoding method, apparatuses therefor, programs therefor, and storage media for storing the programs |
WO2008035665A1 (fr) * | 2006-09-20 | 2008-03-27 | Nippon Telegraph And Telephone Corporation | procédé DE CODAGE D'IMAGE, PROCÉDÉ DE DÉCODAGE, DISPOSITIF associÉ, DISPOSITIF DE DÉCODAGE D'IMAGE, programme associÉ, et support de stockage contenant le programme |
US8290289B2 (en) | 2006-09-20 | 2012-10-16 | Nippon Telegraph And Telephone Corporation | Image encoding and decoding for multi-viewpoint images |
JP4999853B2 (ja) * | 2006-09-20 | 2012-08-15 | 日本電信電話株式会社 | 画像符号化方法及び復号方法、それらの装置、及びそれらのプログラム並びにプログラムを記録した記憶媒体 |
JP4999854B2 (ja) * | 2006-09-20 | 2012-08-15 | 日本電信電話株式会社 | 画像符号化方法及び復号方法、それらの装置、及びそれらのプログラム並びにプログラムを記録した記憶媒体 |
WO2008035654A1 (fr) * | 2006-09-20 | 2008-03-27 | Nippon Telegraph And Telephone Corporation | Procédés et dispositifs de codage et de décodage d'image, dispositif et programmes de décodage d'image, et support de stockage desdits programmes |
KR101023262B1 (ko) | 2006-09-20 | 2011-03-21 | 니폰덴신뎅와 가부시키가이샤 | 화상 부호화 방법 및 복호 방법, 이들의 장치 및 이들의 프로그램과 프로그램을 기록한 기억매체 |
KR101031624B1 (ko) | 2006-09-20 | 2011-04-27 | 니폰덴신뎅와 가부시키가이샤 | 화상 부호화 방법 및 복호 방법, 그 장치 및 그 프로그램을 기록한 기억매체 |
US8520961B2 (en) | 2006-10-13 | 2013-08-27 | Samsung Electronics Co., Ltd. | Method and apparatus for encoding and decoding multi-view image |
KR101101965B1 (ko) * | 2006-10-16 | 2012-01-02 | 노키아 코포레이션 | 멀티뷰 비디오 코딩에 대한 병렬 디코딩가능 슬라이스들을 이용하기 위한 시스템 및 방법 |
CN101627634A (zh) * | 2006-10-16 | 2010-01-13 | 诺基亚公司 | 用于将可并行解码片用于多视点视频编码的系统和方法 |
WO2008047300A3 (en) * | 2006-10-16 | 2008-07-03 | Nokia Corp | System and method for using parallelly decodable slices for multi-view video coding |
US8594203B2 (en) | 2006-10-16 | 2013-11-26 | Nokia Corporation | System and method for using parallelly decodable slices for multi-view video coding |
TWI400959B (zh) * | 2006-10-30 | 2013-07-01 | Nippon Telegraph & Telephone | 活動影像編碼方法及解碼方法、其裝置、其電腦程式以及記錄有該電腦程式之記憶媒體 |
RU2458479C2 (ru) * | 2006-10-30 | 2012-08-10 | Ниппон Телеграф Энд Телефон Корпорейшн | Способ кодирования и способ декодирования видео, устройства для них, программы для них, а также носители хранения данных, которые сохраняют программы |
JP4999860B2 (ja) * | 2006-10-30 | 2012-08-15 | 日本電信電話株式会社 | 動画像符号化方法及び復号方法、それらの装置、及びそれらのプログラム並びにプログラムを記録した記憶媒体 |
US8532190B2 (en) | 2006-10-30 | 2013-09-10 | Nippon Telegraph And Telephone Corporation | Video encoding method and decoding method, apparatuses therefor, programs therefor, and storage media which store the programs |
US8654854B2 (en) | 2006-10-30 | 2014-02-18 | Nippon Telegraph And Telephone Corporation | Video encoding method and decoding method, apparatuses therefor, programs therefor, and storage media which store the programs |
JP4999859B2 (ja) * | 2006-10-30 | 2012-08-15 | 日本電信電話株式会社 | 予測参照情報生成方法、動画像符号化及び復号方法、それらの装置、及びそれらのプログラム並びにプログラムを記録した記憶媒体 |
WO2008053758A1 (fr) * | 2006-10-30 | 2008-05-08 | Nippon Telegraph And Telephone Corporation | Procédé de codage d'image dynamique, procédé de décodage, leur dispositif, leur programme et support de stockage contenant le programme |
US8675735B2 (en) | 2006-10-30 | 2014-03-18 | Nippon Telegraph And Telephone Corporation | Predicted reference information generating method, video encoding and decoding methods, apparatuses therefor, programs therefor, and storage media which store the programs |
US8355438B2 (en) | 2006-10-30 | 2013-01-15 | Nippon Telegraph And Telephone Corporation | Predicted reference information generating method, video encoding and decoding methods, apparatuses therefor, programs therefor, and storage media which store the programs |
WO2008053746A1 (fr) * | 2006-10-30 | 2008-05-08 | Nippon Telegraph And Telephone Corporation | Procédé de génération d'informations de référence prédictives, procédé de codage et de décodage d'image dynamiques, leur dispositif, leur programme et support de stockage contenant le programme |
US9066096B2 (en) | 2006-12-28 | 2015-06-23 | Nippon Telegraph And Telephone Corporation | Video encoding method and decoding method, apparatuses therefor, programs therefor, and storage media which store the programs |
RU2446618C2 (ru) * | 2006-12-28 | 2012-03-27 | Ниппон Телеграф Энд Телефон Корпорейшн | Способ кодирования и способ декодирования видео, устройства для их осуществления и носители памяти, хранящие программы |
JP2010515400A (ja) * | 2007-01-03 | 2010-05-06 | サムスン エレクトロニクス カンパニー リミテッド | 全域差ベクトルを利用した多視点映像の符号化、復号化方法及び装置 |
JP2010516158A (ja) * | 2007-01-11 | 2010-05-13 | サムスン エレクトロニクス カンパニー リミテッド | 多視点映像の符号化、復号化方法及び装置 |
US8670626B2 (en) | 2007-04-11 | 2014-03-11 | Samsung Electronics Co., Ltd. | Method and apparatus for encoding and decoding and multi-view image |
US8611688B2 (en) | 2007-04-11 | 2013-12-17 | Samsung Electronics Co., Ltd. | Method and apparatus for encoding and decoding and multi-view image |
US9088779B2 (en) | 2007-04-11 | 2015-07-21 | Samsung Electronics Co., Ltd. | Method and apparatus for encoding and decoding and multi-view image |
US8532412B2 (en) | 2007-04-11 | 2013-09-10 | Samsung Electronics Co., Ltd. | Method and apparatus for encoding and decoding and multi-view image |
RU2444856C2 (ru) * | 2007-06-25 | 2012-03-10 | Ниппон Телеграф Энд Телефон Корпорейшн | Способ кодирования видеосигнала и способ декодирования, устройства для этого и носители хранения информации, которые хранят программы для этого |
US8204118B2 (en) | 2007-06-25 | 2012-06-19 | Nippon Telegraph And Telephone Corporation | Video encoding method and decoding method, apparatuses therefor, programs therefor, and storage media which store the programs |
JP2012505457A (ja) * | 2008-10-08 | 2012-03-01 | ミツビシ・エレクトリック・アールアンドディー・センター・ヨーロッパ・ビーヴィ | マルチメディアシグネチャのための符号化及び復号の方法及び装置 |
RU2525850C2 (ru) * | 2009-02-12 | 2014-08-20 | Ниппон Телеграф Энд Телефон Корпорейшн | Способ кодирования многовидового изображения, способ декодирования многовидового изображения, устройство кодирования многовидового изображения, устройство декодирования многовидового изображения, компьютерно-читаемый носитель, содержащий программу кодирования многовидового изображения и компьютерно-читаемый носитель, содержащий программу декодирования многовидового изображения |
US9094687B2 (en) | 2009-12-28 | 2015-07-28 | Fujitsu Limited | Video encoder and video decoder |
WO2011083520A1 (ja) * | 2010-01-08 | 2011-07-14 | 富士通株式会社 | 動画像符号化装置および動画像復号装置 |
JP5533885B2 (ja) * | 2010-01-08 | 2014-06-25 | 富士通株式会社 | 動画像符号化装置および動画像復号装置 |
US9078006B2 (en) | 2010-01-08 | 2015-07-07 | Fujitsu Limited | Video encoder and video decoder |
RU2480941C2 (ru) * | 2011-01-20 | 2013-04-27 | Корпорация "Самсунг Электроникс Ко., Лтд" | Способ адаптивного предсказания кадра для кодирования многоракурсной видеопоследовательности |
US9491462B2 (en) | 2011-06-30 | 2016-11-08 | Sony Corporation | High efficiency video coding device and method based on reference picture type |
WO2013002342A1 (ja) * | 2011-06-30 | 2013-01-03 | ソニー株式会社 | 画像処理装置および方法 |
US9560375B2 (en) | 2011-06-30 | 2017-01-31 | Sony Corporation | High efficiency video coding device and method based on reference picture type |
US9648344B2 (en) | 2011-06-30 | 2017-05-09 | Sony Corporation | High efficiency video coding device and method based on reference picture type |
US9788008B2 (en) | 2011-06-30 | 2017-10-10 | Sony Corporation | High efficiency video coding device and method based on reference picture type |
US10158877B2 (en) | 2011-06-30 | 2018-12-18 | Sony Corporation | High efficiency video coding device and method based on reference picture type of co-located block |
US10187652B2 (en) | 2011-06-30 | 2019-01-22 | Sony Corporation | High efficiency video coding device and method based on reference picture type |
US10484704B2 (en) | 2011-06-30 | 2019-11-19 | Sony Corporation | High efficiency video coding device and method based on reference picture type |
US10764600B2 (en) | 2011-06-30 | 2020-09-01 | Sony Corporation | High efficiency video coding device and method based on reference picture type |
US11405634B2 (en) | 2011-06-30 | 2022-08-02 | Sony Corporation | High efficiency video coding device and method based on reference picture type |
WO2013146507A1 (ja) * | 2012-03-30 | 2013-10-03 | ソニー株式会社 | 画像処理装置および方法、並びに記録媒体 |
Also Published As
Publication number | Publication date |
---|---|
EP1835747B1 (en) | 2019-05-08 |
EP1835747A4 (en) | 2011-02-02 |
KR100946790B1 (ko) | 2010-03-11 |
US20070189396A1 (en) | 2007-08-16 |
EP1835747A1 (en) | 2007-09-19 |
CN100584013C (zh) | 2010-01-20 |
US9088802B2 (en) | 2015-07-21 |
TW200629914A (en) | 2006-08-16 |
JP4937741B2 (ja) | 2012-05-23 |
TWI308028B (en) | 2009-03-21 |
KR20070052692A (ko) | 2007-05-22 |
JPWO2006073116A1 (ja) | 2008-06-12 |
CN1910931A (zh) | 2007-02-07 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
JP4937741B2 (ja) | 映像符号化方法及び装置、映像復号方法及び装置、それらのプログラムおよびそれらプログラムを記録した記録媒体 | |
JP5234587B2 (ja) | 映像符号化方法及び復号方法、それらの装置、及びそれらのプログラム並びにプログラムを記録した記憶媒体 | |
JP5368482B2 (ja) | ビデオ符号化および復号におけるインプリシットなブロック分割のための方法および装置 | |
JP4755095B2 (ja) | 映像符号化の符号化側/復号化側に使用される双方向予測方法 | |
JP5061179B2 (ja) | 照明変化補償動き予測符号化および復号化方法とその装置 | |
US8204118B2 (en) | Video encoding method and decoding method, apparatuses therefor, programs therefor, and storage media which store the programs | |
WO2007077942A1 (ja) | 映像符号化方法及び復号方法、それらの装置、及びそれらのプログラム並びにプログラムを記録した記憶媒体 | |
US20070133689A1 (en) | Low-cost motion estimation apparatus and method thereof | |
KR20110073263A (ko) | 인트라 예측 부호화 방법 및 부호화 방법, 그리고 상기 방법을 수행하는 인트라 예측 부호화 장치 및 인트라 예측 복호화 장치 | |
JP4732184B2 (ja) | 動画像符号化装置及びその制御方法、コンピュータプログラム、記憶媒体 | |
JP2011061302A (ja) | 動画像符号化装置 | |
JP2009260421A (ja) | 動画像処理システム、符号化装置、符号化方法、符号化プログラム、復号化装置、復号化方法および復号化プログラム | |
JP2007329528A (ja) | 動きベクトル復号化方法および復号化装置 | |
KR101076272B1 (ko) | 분산 비디오 코딩을 위한 픽셀 도메인 기반의 관심영역 추출방법 및 컴퓨터 판독가능 매체 | |
JP2009049969A (ja) | 動画像符号化装置及び方法並びに動画像復号化装置及び方法 | |
JP4642033B2 (ja) | 参照フレームの数を固定する符号化方式で画像の参照ブロックを取得する方法 | |
JP4789719B2 (ja) | 動きベクトル検出装置、動きベクトル検出方法、コンピュータプログラム及び記憶媒体 | |
JP5885886B2 (ja) | 画像解析装置及び画像解析方法 | |
JP4561701B2 (ja) | 動画像符号化装置 | |
JP4760551B2 (ja) | 動きベクトル復号化方法および復号化装置 | |
US20130215966A1 (en) | Image encoding method, image decoding method, image encoding device, image decoding device | |
JP5310620B2 (ja) | 動画像符号化装置、動画像符号化方法及び動画像符号化用コンピュータプログラムならびに映像伝送装置 | |
KR20120079561A (ko) | 선택적 다중 예측을 통한 적응적 인트라 예측 부호화/복호화 장치 및 방법 | |
JP2010068219A (ja) | 動画像符号化装置 | |
JP2012095099A (ja) | 動画像符号化装置、動画像符号化方法及び動画像符号化プログラム、並びに動画像復号装置、動画像復号方法及び動画像復号プログラム |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
WWE | Wipo information: entry into national phase |
Ref document number: 2006519047 Country of ref document: JP |
|
WWE | Wipo information: entry into national phase |
Ref document number: 2007189396 Country of ref document: US Ref document number: 10586235 Country of ref document: US |
|
WWE | Wipo information: entry into national phase |
Ref document number: 2005822685 Country of ref document: EP Ref document number: 1020067014435 Country of ref document: KR |
|
WWE | Wipo information: entry into national phase |
Ref document number: 200580002762.4 Country of ref document: CN |
|
121 | Ep: the epo has been informed by wipo that ep was designated in this application | ||
NENP | Non-entry into the national phase |
Ref country code: DE |
|
WWP | Wipo information: published in national office |
Ref document number: 10586235 Country of ref document: US |
|
WWP | Wipo information: published in national office |
Ref document number: 2005822685 Country of ref document: EP |