WO2014168121A1 - 画像符号化方法、画像復号方法、画像符号化装置、画像復号装置、画像符号化プログラム、および画像復号プログラム - Google Patents
画像符号化方法、画像復号方法、画像符号化装置、画像復号装置、画像符号化プログラム、および画像復号プログラム Download PDFInfo
- Publication number
- WO2014168121A1 WO2014168121A1 PCT/JP2014/060116 JP2014060116W WO2014168121A1 WO 2014168121 A1 WO2014168121 A1 WO 2014168121A1 JP 2014060116 W JP2014060116 W JP 2014060116W WO 2014168121 A1 WO2014168121 A1 WO 2014168121A1
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- image
- prediction
- encoding
- decoding
- unit
- Prior art date
Links
- 238000000034 method Methods 0.000 title claims abstract description 102
- 239000002131 composite material Substances 0.000 claims description 135
- 239000013598 vector Substances 0.000 claims description 28
- 230000015572 biosynthetic process Effects 0.000 claims description 5
- 238000003786 synthesis reaction Methods 0.000 claims description 5
- 238000012545 processing Methods 0.000 description 40
- 238000010586 diagram Methods 0.000 description 25
- 230000033001 locomotion Effects 0.000 description 25
- 230000000875 corresponding effect Effects 0.000 description 18
- 238000004364 calculation method Methods 0.000 description 14
- 238000006243 chemical reaction Methods 0.000 description 3
- 230000006870 function Effects 0.000 description 3
- 230000014509 gene expression Effects 0.000 description 3
- 230000002123 temporal effect Effects 0.000 description 3
- 238000004891 communication Methods 0.000 description 2
- 230000002596 correlated effect Effects 0.000 description 2
- 238000013139 quantization Methods 0.000 description 2
- ROGUAPYLUCHQGK-UHFFFAOYSA-N 1-piperazinecarboxamide, 4-(3-chloro-2-pyridinyl)-n-[4-(1,1-dimethylethyl)phenyl]- Chemical compound C1=CC(C(C)(C)C)=CC=C1NC(=O)N1CCN(C=2C(=CC=CN=2)Cl)CC1 ROGUAPYLUCHQGK-UHFFFAOYSA-N 0.000 description 1
- 241000406668 Loxodonta cyclotis Species 0.000 description 1
- 238000007792 addition Methods 0.000 description 1
- 238000007796 conventional method Methods 0.000 description 1
- 230000000694 effects Effects 0.000 description 1
- 238000005259 measurement Methods 0.000 description 1
- 230000003287 optical effect Effects 0.000 description 1
- 230000002093 peripheral effect Effects 0.000 description 1
- 230000002441 reversible effect Effects 0.000 description 1
- 238000006467 substitution reaction Methods 0.000 description 1
- 230000001360 synchronised effect Effects 0.000 description 1
Images
Classifications
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/10—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
- H04N19/169—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding
- H04N19/17—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object
- H04N19/172—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object the region being a picture, frame or field
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/10—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
- H04N19/134—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the element, parameter or criterion affecting or controlling the adaptive coding
- H04N19/146—Data rate or code amount at the encoder output
- H04N19/147—Data rate or code amount at the encoder output according to rate distortion criteria
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/10—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
- H04N19/169—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding
- H04N19/17—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object
- H04N19/176—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object the region being a block, e.g. a macroblock
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/44—Decoders specially adapted therefor, e.g. video decoders which are asymmetric with respect to the encoder
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/46—Embedding additional information in the video signal during the compression process
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/50—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding
- H04N19/597—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding specially adapted for multi-view video sequence encoding
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/10—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
- H04N19/102—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the element, parameter or selection affected or controlled by the adaptive coding
- H04N19/132—Sampling, masking or truncation of coding units, e.g. adaptive resampling, frame skipping, frame interpolation or high-frequency transform coefficient masking
Definitions
- the present invention relates to an image encoding method, an image decoding method, an image encoding device, an image decoding device, an image encoding program, and an image decoding program for encoding and decoding a multi-view image.
- multi-view images composed of a plurality of images obtained by photographing the same subject and background with a plurality of cameras are known. These moving images taken by a plurality of cameras are called multi-view moving images (or multi-view images).
- an image (moving image) captured by one camera is referred to as a “two-dimensional image (moving image)”, and a plurality of cameras having the same subject and background with different positions and orientations (hereinafter referred to as viewpoints).
- viewpoints A group of two-dimensional images (two-dimensional moving images) photographed in the above is referred to as “multi-view images (multi-view images)”.
- the two-dimensional moving image has a strong correlation in the time direction, and the encoding efficiency can be increased by using the correlation.
- the encoding efficiency can be increased by using this correlation.
- H. an international encoding standard.
- high-efficiency encoding is performed using techniques such as motion compensation prediction, orthogonal transform, quantization, and entropy encoding.
- H.M. In H.264, encoding using temporal correlation with a plurality of past or future frames is possible.
- H. The details of the motion compensation prediction technique used in H.264 are described in Non-Patent Document 1, for example.
- H. An outline of the motion compensation prediction technique used in H.264 will be described.
- H. H.264 motion compensation prediction divides the encoding target frame into blocks of various sizes, and allows each block to have different motion vectors and different reference frames. By using a different motion vector for each block, it is possible to achieve highly accurate prediction that compensates for different motions for each subject. On the other hand, by using a different reference frame for each block, it is possible to realize highly accurate prediction in consideration of occlusion caused by temporal changes.
- the difference between the multi-view image encoding method and the multi-view image encoding method is that, in addition to the correlation between cameras, the multi-view image has a temporal correlation at the same time. However, in either case, correlation between cameras can be used in the same way. Therefore, here, a method used in encoding a multi-view video is described.
- FIG. 21 is a conceptual diagram showing parallax generated between cameras.
- the image plane of a camera with parallel optical axes is looked down vertically. In this way, the position where the same part on the subject is projected on the image plane of a different camera is generally called a corresponding point.
- each pixel value of the encoding target frame is predicted from the reference frame based on the correspondence relationship, and the prediction residual and the disparity information indicating the correspondence relationship are encoded. Since the parallax changes for each target camera pair and position, it is necessary to encode the parallax information for each region where the parallax compensation prediction is performed. In fact, H. In the H.264 multi-view video encoding scheme, a vector representing disparity information is encoded for each block using disparity compensation prediction.
- Correspondence given by the parallax information can be represented by a one-dimensional quantity indicating the three-dimensional position of the subject instead of a two-dimensional vector based on epipolar geometric constraints by using camera parameters.
- information indicating the three-dimensional position of the subject there are various expressions, but the distance from the reference camera to the subject or the coordinate value on the axis that is not parallel to the image plane of the camera is often used.
- the reciprocal of the distance is used instead of the distance.
- the reciprocal of the distance is information proportional to the parallax, there are cases where two reference cameras are set and expressed as a parallax amount between images taken by these cameras. Since there is no essential difference no matter what expression is used, in the following, information indicating these three-dimensional positions is expressed as depth without distinguishing by expression.
- FIG. 22 is a conceptual diagram of epipolar geometric constraints.
- the point on the image of another camera corresponding to the point on the image of one camera is constrained on a straight line called an epipolar line.
- the corresponding point is uniquely determined on the epipolar line.
- the corresponding point in the second camera image with respect to the subject projected at the position m in the first camera image is on the epipolar line when the subject position in the real space is M ′.
- the subject position in the real space is M ′′, it is projected at the position m ′′ on the epipolar line.
- Non-Patent Document 2 using this property, a synthesized image for the encoding target frame is generated from the reference frame according to the three-dimensional information of each subject given by the depth map (distance image) for the reference frame, By using prediction image candidates, highly accurate prediction is realized, and efficient multi-view video encoding is realized. Note that a composite image generated based on this depth is called a viewpoint composite image, a viewpoint interpolation image, or a parallax compensation image.
- Non-Patent Document 3 for each region, a virtual depth map for the encoding target frame is generated from a depth map for the reference frame, and a corresponding point is obtained using the generated virtual depth map. While using a depth map, it is possible to generate a viewpoint composite image only for a necessary region.
- Non-Patent Document 2 it is possible to realize highly efficient prediction using a viewpoint composite image in which parallax compensation is performed with high accuracy using three-dimensional information of a subject obtained from a depth map. .
- the existing prediction for each area and the prediction based on the viewpoint composite image even when a viewpoint composite image with low accuracy is generated partially due to the influence of depth map quality and occlusion, By selecting whether or not the viewpoint composite image is a predicted image for each region, it is possible to prevent the code amount from increasing.
- Non-Patent Document 2 a viewpoint composite image for one frame must be generated and stored regardless of whether the viewpoint composite image is used as a predicted image. There is a problem that memory consumption increases. In addition, when the parallax between the processing target image (encoding target image or decoding target image) and the reference frame is small, or when the quality of the depth map is high, the quality of the wide area of the processing target image is high. Although a viewpoint composite image is obtained, information indicating whether or not the viewpoint composite image is used as a predicted image has to be encoded for each region, which increases the amount of code.
- Non-Patent Document 3 it is not necessary to generate a viewpoint composite image for an area not used for prediction, so that the problem of processing load and memory consumption can be solved.
- the present invention has been made in view of such circumstances, and suppresses an increase in processing amount and memory consumption when a multi-view video is encoded or decoded while using a view synthesized image as one of the predicted images.
- An object is to provide an image encoding method, an image decoding method, an image encoding device, an image decoding device, an image encoding program, and an image decoding program capable of realizing encoding with a small amount of code. .
- the present invention when encoding a multi-view image consisting of a plurality of different viewpoint images, an encoded reference image for a viewpoint different from the encoding target image, and a reference depth map for a subject in the reference image Is an image encoding device that performs encoding while predicting images between different viewpoints, and generates a viewpoint composite image for the entire encoding target image using the reference image and the reference depth map A prediction synthesized image generation unit that performs prediction for each encoding target block obtained by dividing the encoding target image as a prediction unit, or performs prediction using the viewpoint synthetic image for the entire encoding target image A prediction unit setting unit that selects the prediction unit, a prediction unit information encoding unit that encodes information indicating the selected prediction unit, and the prediction target unit for each encoding target block When performing the measurement is selected, the encoding target image, for each of the encoding target block, and an encoding target image predictive encoding unit for prediction encoding while selecting a prediction image generation method.
- the image encoding device when the prediction unit is selected to perform prediction using the viewpoint composite image for the entire encoding target image as the prediction unit, and the encoding target image and the viewpoint composite image A view synthesis prediction residual encoding unit that encodes the difference may be further included.
- the image encoding device of the present invention estimates an image unit prediction rate distortion cost that is an image unit prediction rate distortion cost that is a rate distortion cost when predicting and encoding the entire encoding target image with the viewpoint composite image.
- a cost estimation unit, and the prediction unit setting unit may set the prediction unit by comparing the image unit prediction rate distortion cost with the block unit prediction rate distortion cost. .
- the image encoding device of the present invention generates a partial viewpoint synthesized image that is a viewpoint synthesized image for the encoding target block, using the reference image and the reference depth map for each encoding target block.
- a composite image generation unit may be further included, and the encoding target image predictive encoding unit may use the partial viewpoint composite image as a predicted image candidate.
- the image encoding device generates prediction information for each encoding target block when the prediction unit is selected to perform prediction using the viewpoint composite image on the entire image as the prediction unit. You may further have an information generation part.
- the prediction information generation unit determines a prediction block size
- the viewpoint composite image generation unit repeats the process of generating a viewpoint composite image for each prediction block size, You may make it produce
- the prediction information generation unit may estimate a disparity vector and generate prediction information as a disparity compensation prediction.
- the prediction information generation unit may determine a prediction method and generate prediction information for the prediction method.
- the present invention provides a decoded reference image for a viewpoint different from the decoding target image and a subject in the reference image when decoding the decoding target image from code data of a multi-viewpoint image including a plurality of different viewpoint images.
- An image decoding apparatus that performs decoding while predicting images between different viewpoints using a reference depth map for the image, and using the reference image and the reference depth map, a viewpoint composite image for the entire decoding target image Whether to perform prediction for each decoding target block obtained by dividing the decoding target image from the code data, or to perform prediction using the viewpoint synthetic image for the entire decoding target image
- a prediction unit information decoding unit that decodes information of a prediction unit indicating the prediction unit, and the prediction unit performs prediction on the entire decoded image using the viewpoint composite image
- a decoding target image setting unit that uses the viewpoint synthesized image as the decoding target image, and when the prediction unit indicates prediction for each decoding target block, from the code data, the decoding target block
- the decoding target image setting unit decodes a difference between the decoding target image and the viewpoint synthesized image from code data, and adds the difference and the viewpoint synthesized image to add the decoding target.
- An image may be generated.
- the image decoding apparatus generates a partial viewpoint synthesized image that generates a partial viewpoint synthesized image that is a viewpoint synthesized image for the decoding target block, using the reference image and the reference depth map for each decoding target block. May be further included, and the decoding target image decoding unit may use the partial viewpoint synthesized image as a predicted image candidate.
- the image decoding apparatus of the present invention generates prediction information for each decoding target block when the prediction unit indicates that prediction is performed on the entire decoded image using the viewpoint synthesized image. It may further have a part.
- the prediction information generation unit determines a prediction block size
- the viewpoint composite image generation unit repeats the process of generating a viewpoint composite image for each prediction block size, thereby decoding the decoding You may make it produce
- the prediction information generation unit may estimate a disparity vector and generate prediction information as a disparity compensation prediction.
- the prediction information generation unit may determine a prediction method and generate prediction information for the prediction method.
- the present invention when encoding a multi-view image consisting of a plurality of different viewpoint images, an encoded reference image for a viewpoint different from the encoding target image, and a reference depth map for a subject in the reference image, Is an image encoding method that performs encoding while predicting images between different viewpoints, and generates a viewpoint composite image for the entire encoding target image using the reference image and the reference depth map.
- a viewpoint synthesized image generation step, and prediction is performed for each encoding target block obtained by dividing the encoding target image, or prediction is performed on the entire encoding target image using the viewpoint synthetic image.
- the present invention provides a decoded reference image for a viewpoint different from the decoding target image and a subject in the reference image when decoding the decoding target image from code data of a multi-viewpoint image including a plurality of different viewpoint images. And a reference synthesized image for the entire decoding target image using the reference image and the reference depth map. Whether to perform prediction for each decoding target block obtained by dividing the decoding target image from the code data, or to perform prediction using the viewpoint combining image for the entire decoding target image A prediction unit information decoding step for decoding information on a prediction unit indicating the prediction unit using the viewpoint composite image for the prediction unit.
- the present invention is an image encoding program for causing a computer to execute the image encoding method.
- the present invention is an image decoding program for causing a computer to execute the image decoding method.
- One aspect of the present invention is a computer-readable recording medium on which the image encoding program is recorded.
- One aspect of the present invention is a computer-readable recording medium on which the image decoding program is recorded.
- the amount of calculation and the memory during decoding are adaptively switched between prediction for the entire encoding target image and prediction for each encoding target block.
- FIG. 7 It is a block diagram which shows the image coding apparatus by the 1st Embodiment of this invention. It is a flowchart which shows operation
- movement of the image coding apparatus shown in FIG. 7 is a flowchart showing another operation of the image encoding device shown in FIG. 1.
- FIG. 7 It is a block diagram which shows the image coding apparatus by the 2nd Embodiment of this invention.
- 5 is a flowchart illustrating an operation of the image encoding device illustrated in FIG. 4.
- 6 is a flowchart showing another operation of the image encoding device shown in FIG. 4.
- FIG. 9 is a flowchart showing a processing operation for constructing and outputting a bit stream for frame-based prediction in the image encoding device shown in FIGS. 7 and 8.
- FIG. It is a block diagram which shows the image decoding apparatus by the 5th Embodiment of this invention. It is a flowchart which shows the operation
- movement of the image decoding apparatus shown in FIG. Fig. 11 is a flowchart showing another operation of the image decoding apparatus shown in Fig. 10. It is a block diagram which shows the image decoding apparatus by the 6th Embodiment of this invention. It is a flowchart which shows operation
- movement of the image decoding apparatus shown in FIG. Fig. 17 is a flowchart showing an operation of the image decoding device shown in Fig. 16. It is a block diagram which shows the image coding apparatus by the 9th Embodiment of this invention. It is a block diagram which shows the image decoding apparatus by the 10th Embodiment of this invention. It is a conceptual diagram which shows the parallax which arises between two cameras. It is a conceptual diagram explaining an epipolar geometric constraint.
- a multi-viewpoint image captured by two cameras a first camera (referred to as camera A) and a second camera (referred to as camera B), is encoded.
- a description will be given assuming that an image of the camera B is encoded or decoded as a reference image.
- information necessary for obtaining the parallax from the depth information is given separately. Specifically, it is an external parameter representing the positional relationship between the camera A and the camera B and an internal parameter representing projection information on the image plane by the camera, but parallax can be obtained from the depth information even in other forms.
- information that can specify the position between the symbols [] is added to an image, video frame, or depth map to add the position. It is assumed that the image signal sampled by the pixels and the depth corresponding thereto are shown.
- the coordinate value or block at a position where the coordinate or block is shifted by the amount of the vector by adding the coordinate value or the index value that can be associated with the block and the vector is represented.
- FIG. 1 is a block diagram showing a configuration of an image encoding device according to the first embodiment of the present invention.
- the image encoding device 100a includes an encoding target image input unit 101, an encoding target image memory 102, a reference image input unit 103, a reference depth map input unit 104, a viewpoint composite image generation unit 105, a viewpoint.
- a composite image memory 106, a frame unit prediction RD (Rate Distortion) cost calculation unit 107, an image encoding unit 108, a block unit prediction RD cost calculation unit 109, a prediction unit determination unit 110, and a bit stream generation unit 111 are provided.
- the encoding target image input unit 101 inputs an image to be encoded.
- the image to be encoded is referred to as an encoding target image.
- an image of camera B is input.
- a camera that captures an encoding target image (camera B in this case) is referred to as an encoding target camera.
- the encoding target image memory 102 stores the input encoding target image.
- the reference image input unit 103 inputs an image to be referred to when generating a viewpoint composite image (parallax compensation image).
- the image input here is referred to as a reference image.
- an image of camera A is input.
- the reference depth map input unit 104 inputs a depth map to be referred to when generating a viewpoint composite image.
- the depth map for the reference image is input, but a depth map for another camera may be input.
- this depth map is referred to as a reference depth map.
- the depth map represents the three-dimensional position of the subject shown in each pixel of the corresponding image. Any information may be used as long as the three-dimensional position can be obtained by information such as separately provided camera parameters. For example, a distance from the camera to the subject, a coordinate value with respect to an axis that is not parallel to the image plane, and a parallax amount with respect to another camera (for example, camera B) can be used.
- a parallax map that directly expresses the amount of parallax may be used instead of the depth map.
- the depth map is passed in the form of an image.
- the image may not be in the form of an image.
- the camera (here, camera A) corresponding to the reference depth map is referred to as a reference depth camera.
- the viewpoint composite image generation unit 105 obtains a correspondence relationship between the pixels of the encoding target image and the pixels of the reference image using the reference depth map, and generates a viewpoint composite image for the encoding target image.
- the viewpoint composite image memory 106 stores a viewpoint composite image for the generated encoding target image.
- the frame unit prediction RD cost calculation unit 107 calculates a rate distortion cost when the encoding target image is predicted on a frame basis using the viewpoint composite image.
- the image encoding unit 108 predictively encodes the encoding target image in units of blocks while using the viewpoint composite image.
- the block unit prediction RD cost calculation unit 109 calculates a rate distortion cost when the encoding target image is predictively encoded in units of blocks while using the viewpoint composite image.
- the prediction unit determination unit 110 determines whether to predict the encoding target image in units of frames or in units of blocks based on the rate distortion cost. Based on the determination by the prediction unit determination unit 110, the bit stream generation unit 111 constructs and outputs a bit stream for the encoding target image.
- FIG. 2 is a flowchart showing the operation of the image encoding device 100a shown in FIG.
- the encoding target image input unit 101 receives the encoding target image Org and stores it in the encoding target image memory 102 (step S101).
- the reference image input unit 103 inputs a reference image
- the reference depth map input unit 104 inputs the reference depth map
- the viewpoint composite image generation unit 105 step S102.
- the reference image and the reference depth map input in step S102 are the same as those obtained on the decoding side, such as those obtained by decoding already encoded ones. This is to suppress the occurrence of coding noise such as drift by using exactly the same information obtained by the decoding device. However, when the generation of such coding noise is allowed, the one that can be obtained only on the coding side, such as the one before coding, may be input.
- the reference depth map in addition to the one already decoded, the depth map estimated by applying stereo matching or the like to the multi-viewpoint images decoded for a plurality of cameras, or decoded
- the depth map estimated using the disparity vector, the motion vector, and the like can also be used as the same one can be obtained on the decoding side.
- the viewpoint composite image generation unit 105 generates a viewpoint composite image Synth for the encoding target image and stores it in the viewpoint composite image memory 106 (step S103).
- the process here may be any method as long as it uses a reference image and a reference depth map to synthesize an image in the encoding target camera.
- the frame unit prediction RD cost calculation unit 107 calculates a rate distortion cost when the entire encoding target image is predicted and encoded with the viewpoint composite image (step S104).
- the rate distortion cost is a value represented by a weighted sum of a generated code amount and distortion caused by encoding, as shown in the following equation (1).
- Cost m is an RD cost
- D m is an encoding result of an image obtained as a result of encoding (more precisely, a decoded image obtained by decoding a bit stream of the encoding result).
- the amount of distortion R m is the code amount of the bit stream obtained as a result of encoding
- ⁇ is a Lagrange undetermined multiplier that depends on the target bit rate and target quality. Any scale may be used as the amount of distortion. For example, signal distortion such as SSD (Sum of Squared Differences) and SAD (Sum of Absolute Differences) And a measure of distortion related to subjective quality such as SSIM (Structural Similarity) can be used.
- m represents a method used for encoding
- frame represents a coding method based on prediction in frame units using a viewpoint composite image.
- any method may be used as long as information indicating generation and selection of a predicted image is not encoded in each region.
- the decoding result for the encoding target image is a viewpoint composite image
- the decoding result for the encoding target image is a viewpoint composite image
- a method of skipping encoding of the encoding target image and using the information indicating the skip as the encoding result will be described.
- the prediction image is a viewpoint composite image in the entire encoding target image
- another method such as a method of transform encoding the prediction residual of the encoding target image for each frame or region may be used.
- the amount of distortion is expressed by SSD, it is assumed that the decoding result for the encoding target image is a viewpoint composite image, and the encoding of the encoding target image is skipped, and information indicating the skip is used as the encoding result.
- the distortion amount D frame in this case is expressed by the following equation (2).
- p is an index representing the pixel position
- sigma p represents the sum for all the pixels in the encoding target image.
- the code amount R frame is assumed to be 1 bit here. Note that a flag having a length of 1 bit or more may be used, or entropy coding may be performed together with flags for other frames to obtain a code amount of less than 1 bit.
- the image encoding unit 108 performs encoding while generating a predicted image for each region (encoding target block) obtained by dividing the encoding target image (step S105).
- Any encoding method may be used as long as the image is divided and encoded for each block.
- the H.P. A method according to H.264 / AVC may be used.
- a method using a viewpoint composite image may be used as a predicted image candidate to be selected for each block, or a method not using it may be used.
- the block unit prediction RD cost calculation unit 109 divides the encoding target image into a plurality of blocks, and rate distortion when encoding while selecting a prediction method for each block
- the cost Cost block is calculated (step S106).
- the block unit prediction RD cost Cost block is the distortion amount D block for the image to be encoded of the image of the encoding result in step S105 (more precisely, the decoded image obtained by decoding the bit stream of the encoding result).
- the code amount R block obtained by adding the code amount of the flag indicating that the encoding of the encoding target image was not skipped to the code amount of the bit stream as the encoding result in step S105 Therefore, calculate.
- the prediction unit determination unit 110 determines to determine a prediction unit by comparing the rate distortion costs (step S107). Note that the rate distortion cost defined by equation (1) indicates that the smaller the value is, the better the coding efficiency is. Therefore, a prediction unit having a small rate distortion cost is selected. If a rate distortion cost indicating that the coding efficiency is better as the value is larger is used, it is necessary to reverse the determination and select a prediction unit having a larger rate distortion cost.
- the bit stream generation unit 111 selects the bit stream for performing the frame unit prediction. Generate (step S108).
- the generated bit stream is an output of the image encoding device 100a.
- a 1-bit flag indicating that the entire decoded image is a viewpoint synthesized image is a bit stream in this case.
- the prediction image is a viewpoint composite image in the entire encoding target image, and the prediction residual of the encoding target image is transform-coded for each frame or block.
- a bit stream in which a bit stream corresponding to the prediction residual is concatenated with the flag is generated.
- a bitstream for the prediction residual may be newly generated, but the bitstream generated in step S104 may be stored in a memory or the like, and the bitstream may be read from the memory or the like and used. .
- the bit stream generation unit 111 generates a bit stream for performing block unit prediction (step S109). ).
- the generated bit stream is an output of the image encoding device 100a.
- a bit stream is generated by concatenating the bit stream generated by the image encoding unit 108 in step S105 to a 1-bit flag indicating that the entire decoded image is not a viewpoint synthesized image. Note that the bit stream generated in step S105 may be stored in a memory or the like, read out and used, or generated again.
- the image encoding device 100a outputs a bit stream for the image signal. That is, a parameter set and a header indicating information such as an image size are separately added to the bit stream output from the image encoding device 100a as necessary.
- FIG. 3 is a flowchart showing a processing operation in the case where determination is performed for each block as an example. Parts that perform the same processing as the processing operation shown in FIG.
- the processing operation shown in FIG. 3 is different from the processing operation shown in FIG. 2 in that after calculating the frame-unit predicted RD cost, for each block, encoding processing, rate distortion cost calculation processing, and prediction unit determination processing. It is a point that repeats. That is, first, the variable blk indicating the index of the block, which is a unit for performing the encoding process for dividing the image to be encoded, is set to zero, and the block unit prediction rate distortion cost Cost block is initialized with ⁇ (step 110).
- step S114 while adding the variable blk by 1 (step S114), the following processing (steps S111 to 113, step S107) is repeated until the variable blk reaches the number of blocks numBlks in the encoding target image (step S115). .
- Cost block is initialized with ⁇ . However, it is necessary to perform initialization with an appropriate value according to the bit amount of information representing the prediction unit and the unit of the code amount when calculating the RD cost. There is.
- the information indicating the prediction unit is 1 bit
- the code amount in the RD cost calculation is in units of bits.
- the image encoding unit 108 encodes the encoding target image for the block represented by the variable blk (step S111). Any method may be used for encoding as long as decoding is possible on the decoding side.
- MPEG-2 and H.264 In general video encoding or image encoding such as H.264 and JPEG, a prediction image is generated by selecting one mode from a plurality of prediction modes for each block, and an encoding target image, a prediction image, Are subjected to frequency conversion such as DCT. Next, encoding is performed by sequentially applying quantization, binarization, and entropy encoding to the value obtained as a result of the frequency conversion. Note that in encoding, a viewpoint composite image may be used as one of predicted image candidates.
- step S112 calculate the rate distortion cost Cost blk for block blk (step S112).
- the processing here is the same as the processing in step S106 described above except that the range of the target image is different. That is, the calculation is performed according to the equation (1) from the distortion amount D blk and the code amount R blk of the block blk. Then, the rate distortion cost for the block blk obtained by calculation is added to the cost block (step S113), and the prediction unit is determined by comparing with the cost frame (step S107).
- the cost block is equal to or higher than the cost frame , it is determined that the prediction for each frame is used, and the processing for each block is ended. Since the determination is performed for each block, when the processing is completed for all the blocks, it is determined that the prediction of the block unit is used without determining the prediction unit again.
- the same viewpoint composite image is used for the case where the prediction is performed in frame units and the case where the prediction is performed in block units, but the viewpoint composite images may be generated by different methods.
- the amount of memory for accumulating viewpoint composite images is reduced by referring to the information of already encoded blocks, and the quality of the viewpoint composite images is improved. It doesn't matter.
- the quality of the decoded image obtained on the decoding side may be improved by performing synthesis in consideration of consistency and objective quality in the entire frame.
- FIG. 4 is a block diagram illustrating a configuration of an image encoding device when a viewpoint composite image is generated by a different method for each prediction unit.
- the difference between the image encoding device 100a shown in FIG. 1 and the image encoding device 100b shown in FIG. 4 is that in the image encoding device 100b, a frame unit viewpoint composite image generation unit 114 and a block unit viewpoint composite image generation unit 115 are two.
- One viewpoint composite image generation unit is provided, and a viewpoint composite image memory is not necessarily provided.
- symbol is attached
- the frame-unit viewpoint composite image generation unit 114 obtains a correspondence relationship between the pixels of the encoding target image and the pixels of the reference image using the reference depth map, and generates a viewpoint composite image for the entire encoding target image.
- the block-unit viewpoint composite image generation unit 115 generates a viewpoint composite image for each block on which the encoding process of the encoding target image is performed using the reference depth map.
- FIGS. 5 and 6 are flowcharts showing the operation of the image encoding device 100b shown in FIG.
- FIG. 5 is a processing operation in the case where the prediction unit is determined after performing the encoding using the prediction in block units for all the blocks, and FIG. Processing operation.
- FIG. 5 or FIG. 6 the same reference numerals are given to the same processes as those in the flowchart shown in FIG. 2 or FIG.
- a viewpoint composite image for that block is generated for each block separately from the viewpoint composite image generated for prediction in units of frames. This is a point to be processed (step S117).
- any method may be used as the process for generating the viewpoint composite image for each block.
- the method described in Non-Patent Document 3 may be used.
- the prediction information is information used for generating a prediction image such as a prediction block size, a prediction mode, a motion / disparity vector, and decoding a prediction residual.
- 7 and 8 can be used to generate prediction information for each block obtained by dividing the image to be encoded when it is determined to perform frame-by-frame prediction and to encode another frame.
- It is a block diagram which shows the structure of the image coding apparatus in the case of doing.
- the image encoding device 100c shown in FIG. 7 corresponds to the image encoding device 100a shown in FIG. 1, and the image encoding device 100d shown in FIG. 8 is a block diagram corresponding to the image encoding device 100b shown in FIG.
- the difference is that a block unit prediction information generation unit 116 is further provided.
- symbol is attached
- the block unit prediction information generation unit 116 When it is determined that prediction in units of frames is performed, the block unit prediction information generation unit 116 generates prediction information for each block obtained by dividing the encoding target image, and encodes another frame. Output. Note that when another frame is encoded by the same image encoding device, the generated information is passed to the image encoding unit 108.
- the processing operations executed by the image encoding device 100c shown in FIG. 7 and the image encoding device 100d shown in FIG. 8 are basically the same as those described above, and construct and output a bit stream for frame prediction.
- the processing to be performed (step S108) only executes the processing operation shown in FIG.
- FIG. 9 is a flowchart showing a processing operation for constructing and outputting a bit stream of frame unit prediction.
- a bit stream for frame prediction is constructed and output (step S1801). This process is the same as step S108 described above.
- the block unit prediction information generation unit 116 generates and outputs prediction information for each block obtained by dividing the encoding target image (step S1802).
- the prediction information can be generated as long as the decoding side can generate the same information.
- the predicted block size may be as large as possible or as small as possible.
- a different block size may be set for each block by making a determination based on the used depth map and the generated viewpoint composite image.
- the block size may be adaptively determined so as to be as large as possible a set of pixels having similar pixel values and depth values.
- mode information and motion / disparity vectors indicating prediction using a viewpoint composite image may be set for all blocks when prediction is performed for each block. Further, the mode information corresponding to the inter-viewpoint prediction mode and the disparity vector obtained from the depth or the like may be set as the mode information and the motion / disparity vector, respectively.
- the disparity vector may be obtained by searching the reference image using the viewpoint composite image for the block as a template.
- an optimal block size and prediction mode may be estimated and generated by analyzing the viewpoint synthesized image as an encoding target image.
- the prediction mode intra-screen prediction, motion compensation prediction, or the like may be selectable.
- the encoding efficiency of another frame can be improved. This is because when similar frames are encoded, such as frames that are temporally continuous or frames of the same subject, the motion vectors and the prediction modes are also correlated, so the redundancy is removed using these correlations. It is because it can.
- FIG. 10 is a block diagram showing the configuration of the image decoding apparatus according to this embodiment.
- the image decoding apparatus 200a includes a bit stream input unit 201, a bit stream memory 202, a reference image input unit 203, a reference depth map input unit 204, a viewpoint synthesized image generation unit 205, a viewpoint synthesized image memory 206, A prediction unit information decoding unit 207 and an image decoding unit 208 are provided.
- the bit stream input unit 101 inputs a bit stream of an image to be decoded.
- the image to be decoded is referred to as a decoding target image.
- the image of the camera B is indicated.
- a camera that captures a decoding target image (camera B in this case) is referred to as a decoding target camera.
- the bit stream memory 202 stores a bit stream for the input decoding target image.
- the reference image input unit 203 inputs an image to be referred to when generating a viewpoint composite image (parallax compensation image).
- the image input here is referred to as a reference image.
- the reference depth map input unit 204 inputs a depth map to be referred to when generating a viewpoint composite image.
- the depth map for the reference image is input, but a depth map for another camera may be input.
- this depth map is referred to as a reference depth map.
- the depth map represents the three-dimensional position of the subject shown in each pixel of the corresponding image. Any information may be used as long as the three-dimensional position can be obtained by information such as separately provided camera parameters. For example, a distance from the camera to the subject, a coordinate value with respect to an axis that is not parallel to the image plane, and a parallax amount with respect to another camera (for example, camera B) can be used.
- a parallax map that directly expresses the amount of parallax may be used instead of the depth map.
- the depth map is passed in the form of an image.
- the image may not be in the form of an image.
- the camera (here, camera A) corresponding to the reference depth map is referred to as a reference depth camera.
- the viewpoint synthesized image generation unit 205 uses the reference depth map to obtain a correspondence relationship between the pixels of the decoding target image and the pixels of the reference image, and generates a viewpoint synthesized image for the decoding target image.
- the viewpoint synthesized image memory 206 stores viewpoint synthesized images for the generated decoding target images.
- the prediction unit information decoding unit 207 decodes information indicating whether the decoding target image is predicted in units of frames or in units of blocks from the bitstream.
- the image decoding unit 208 decodes the decoding target image from the bit stream based on the information decoded by the prediction unit information decoding unit 207 and outputs the decoded image.
- FIG. 11 is a flowchart showing the operation of the image decoding device 200a shown in FIG.
- the bit stream input unit 201 inputs a bit stream obtained by encoding a decoding target image and stores it in the bit stream memory 202 (step S201).
- the reference image input unit 203 inputs a reference image
- the reference depth map input unit 204 inputs the reference depth map, and outputs it to the viewpoint composite image generation unit 205 (step S202).
- the reference image and reference depth map input in step S202 are the same as those used on the encoding side. This is to suppress the occurrence of coding noise such as drift by using exactly the same information as that obtained by the image coding apparatus. However, if such encoding noise is allowed to occur, a different one from that used at the time of encoding may be input.
- the reference depth map in addition to those separately decoded, a depth map estimated by applying stereo matching or the like to multi-viewpoint images decoded for a plurality of cameras, decoded parallax vectors, and motion vectors In some cases, a depth map or the like estimated using the above is used.
- the viewpoint synthesized image generation unit 205 generates a viewpoint synthesized image Synth for the decoding target image and stores it in the viewpoint synthesized image memory 206 (step S203).
- the processing here is the same as step S103 at the time of encoding described above.
- it is necessary to use the same method as that used at the time of coding. A method different from the method used at the time of conversion may be used.
- the prediction unit information decoding unit 207 decodes information representing the prediction unit from the bit stream (step S204). For example, when the prediction unit is represented by the first bit of the bitstream for the decoding target image, the prediction unit is determined by reading the 1 bit.
- the image decoding unit 208 decodes the decoding target image according to the obtained prediction unit.
- the obtained decoding target image is an output of the image decoding device 200a. If the decoding target image is used when decoding other frames, such as when the present invention is used for moving image decoding or multi-viewpoint image decoding, the decoding target image is stored in a separately determined decoding image memory. .
- a method corresponding to the method used at the time of encoding is used.
- decoding is performed by setting a viewpoint composite image as a decoded image.
- prediction in units of blocks is performed, the decoding target image is decoded while generating a prediction image by a designated method for each region (decoding target block) obtained by dividing the decoding target image. For example, the H.P.
- the bit stream for the image signal is input to the image decoding apparatus 200a. That is, a parameter set or header indicating information such as image size is interpreted outside the image decoding device 200a as necessary, and information necessary for decoding is notified to the image decoding device 200a.
- FIG. 12 is a flowchart showing a processing operation for generating a viewpoint composite image only when the prediction unit is a frame unit.
- the processing operation shown in FIG. 12 differs from the processing operation shown in FIG. 11 in that the reference image and reference depth map are input (step S202) and the viewpoint composite image is generated based on the determination of the prediction unit (step S206). Whether or not (Step S203) is executed is determined.
- the same viewpoint composite image is used for the case where the prediction is performed in units of frames and the case where the prediction is performed in units of blocks, but the viewpoint composite images may be generated by different methods.
- the amount of memory for storing the viewpoint composite image is reduced by referring to the information of the already decoded blocks, thereby improving the quality of the viewpoint composite image. It doesn't matter.
- the quality of the viewpoint synthesized image and the decoding target image may be improved by performing synthesis in consideration of the consistency and objective quality in the entire frame.
- FIG. 13 is a block diagram illustrating a configuration of an image decoding device when a viewpoint composite image is generated by a different method for each prediction unit.
- the image decoding device 200b shown in FIG. 13 is different from the image decoding device 200a shown in FIG. 10 in that two viewpoint composite image generation units, a frame unit viewpoint composite image generation unit 209 and a block unit viewpoint composite image generation unit 210, The switch 211 is provided, and the viewpoint composite image memory is not necessarily provided. Note that the same components as those of the image decoding device 200a are denoted by the same reference numerals and description thereof is omitted.
- the frame-unit viewpoint composite image generation unit 209 obtains a correspondence relationship between the pixels of the decoding target image and the pixels of the reference image using the reference depth map, and generates a viewpoint composite image for the entire decoding target image.
- the block-unit viewpoint composite image generation unit 210 generates a viewpoint composite image for each block on which the decoding process of the decoding target image is performed using the reference depth map.
- the switch 211 switches the viewpoint composite image input to the image decoding unit 208 according to the prediction unit output by the prediction unit information decoding unit 207.
- FIG. 14 is a flowchart showing the processing operation of the image decoding apparatus 100b shown in FIG.
- the processing operation shown in FIG. 14 is different from the processing operation shown in FIG. 11 or FIG. 12 in that the viewpoint composite image to be generated is switched according to the prediction unit obtained by decoding (step S206). .
- the process of generating a block-unit viewpoint composite image (step S210) and the process of decoding the decoding target image (step S211) are repeated for each block.
- the variable indicating the index for the block to be decoded is blk, and is represented by the number of blocks numBlks in the decoding target image.
- step S207 The process for generating the viewpoint composite image for the entire frame (step S207) is the same as step S203 described above.
- any method may be used as a method for generating the viewpoint composite image for each block.
- the method described in Non-Patent Document 3 may be used.
- the process of decoding the decoding target image (step S208 and step S211) is the same as step S205 described above, except that the prediction unit is fixed and the unit to be processed is different.
- prediction information when prediction in frame units is performed, only information indicating the prediction unit is generated for the decoding target image, and prediction information is not generated for each block of the decoding target image.
- prediction information for each block not included in the bitstream may be generated so that it can be referred to when another frame is decoded.
- the prediction information is information used for generating a prediction image such as a prediction block size, a prediction mode, a motion / disparity vector, and decoding a prediction residual.
- FIG. 15 and FIG. 15 and FIG. 16 show a case where prediction information is generated for each block obtained by dividing a decoding target image so that it can be referred to when another frame is decoded when it is determined to perform prediction in units of frames.
- It is a block diagram which shows the structure of the image decoding apparatus.
- An image decoding device 200c shown in FIG. 15 corresponds to the image decoding device 200a shown in FIG. 10, and an image decoding device 200d shown in FIG. 16 is a block diagram corresponding to the image decoding device 200b shown in FIG.
- the difference is that a block unit prediction information generation unit 212 is further provided.
- symbol is attached
- the block unit prediction information generation unit 212 When it is determined that prediction in units of frames is performed, the block unit prediction information generation unit 212 generates prediction information for each decoding target block obtained by dividing the decoding target image, and to an image encoding device that decodes another frame. Output. Note that when another frame is encoded by the same image decoding apparatus, the generated information is passed to the image decoding unit 208.
- FIGS. 17 and 18 are flowcharts showing the processing operations of the image decoding device 200c shown in FIG. 15 and the image decoding device 200d shown in FIG. Since the basic processing is the same as the processing operation shown in FIGS. 11 and 14, the same reference numerals are given to the steps for performing the same processing as described above, and the description thereof is omitted.
- a process (step S214) for generating and outputting prediction information for each block when the prediction unit is a frame unit is added.
- the prediction information is generated in the same way as the information generated on the encoding side, any information may be generated.
- the predicted block size may be as large as possible or as small as possible.
- a different block size may be set for each block by making a determination based on the used depth map and the generated viewpoint composite image.
- the block size may be adaptively determined so as to be as large as possible a set of pixels having similar pixel values and depth values.
- mode information and motion / disparity vectors indicating prediction using a viewpoint composite image may be set for all blocks when prediction is performed for each block. Further, the mode information corresponding to the inter-viewpoint prediction mode and the disparity vector obtained from the depth or the like may be set as the mode information and the motion / disparity vector, respectively.
- the disparity vector may be obtained by searching the reference image using the viewpoint composite image for the block as a template.
- an optimal block size and prediction mode may be estimated and generated by analyzing the viewpoint synthesized image as an encoding target image.
- the prediction mode intra-screen prediction, motion compensation prediction, or the like may be selectable.
- the process of encoding and decoding one frame has been described. However, it can be applied to moving picture encoding by repeating a plurality of frames. It can also be applied only to some frames or some blocks of a moving image. For example, you may apply only to the one part area
- the reference depth map is described as a depth map for an image captured by a camera different from the encoding target camera or the decoding target camera, but the encoding target image or the decoding target image
- the depth maps for images taken by the encoding target camera or the decoding target camera at different times may be used as the reference depth map.
- FIG. 19 is a block diagram illustrating a hardware configuration when the above-described image encoding device 100 is configured by a computer and a software program.
- the system shown in FIG. 19 includes a CPU 50, a memory 51 such as a RAM, an encoding target image input unit 52, a reference image input unit 53, a reference depth map input unit 54, a program storage device 55, and a bit stream output.
- the unit 56 is connected by a bus.
- the CPU 50 executes a program.
- a memory 51 such as a RAM stores programs and data accessed by the CPU 50.
- the encoding target image input unit 52 inputs an encoding target image signal from a camera or the like (the encoding target image input unit 52 may be a storage unit that stores an image signal from a disk device or the like).
- the reference image input unit 53 inputs an image signal to be referred from a camera or the like (the reference image input unit 53 may be a storage unit that stores an image signal from a disk device or the like).
- the reference depth map input unit 54 inputs a depth map for a camera having a position and orientation different from that of the camera that captured the encoding target image from the depth camera or the like (the reference depth map input unit 54 is a depth map by a disk device or the like). May also be a storage unit that stores.
- the program storage device 55 stores an image encoding program 551 that is a software program that causes the CPU 50 to execute an image encoding process.
- the bit stream output unit 56 outputs a bit stream generated by the CPU 50 executing the image encoding program 551 loaded in the memory 51 via a network, for example (the bit stream output unit 56 is a disk device or the like). May be a storage unit that stores a bitstream of
- FIG. 20 is a block diagram showing a hardware configuration when the above-described image decoding apparatus 200 is configured by a computer and a software program.
- the system shown in FIG. 20 includes a CPU 60, a memory 51 such as a RAM, a bitstream input unit 62, a reference image input unit 63, a reference depth map input unit 64, a program storage device 65, and a decoding target image output unit. 66 are connected by a bus.
- the CPU 60 executes a program.
- a memory 51 such as a RAM stores programs and data accessed by the CPU 60.
- the bit stream input unit 62 inputs a bit stream encoded by the image encoding device according to this method (the bit stream input unit 62 may be a storage unit that stores an image signal from a disk device or the like).
- the reference image input unit 63 inputs a reference target image signal from a camera or the like (the reference image input unit 63 may be a storage unit that stores an image signal from a disk device or the like).
- the reference depth map input unit 64 inputs a depth map for a camera having a position and orientation different from that of the camera that captured the decoding target from the depth camera or the like (the reference depth map input unit 64 stores depth information by a disk device or the like). It may be a storage unit).
- the program storage device 65 stores an image decoding program 651 that is a software program that causes the CPU 60 to execute an image decoding process.
- the decoding target image output unit 66 outputs the decoding target image obtained by decoding the bitstream to the playback device or the like by the CPU 60 executing the image decoding program 651 loaded in the memory 61 (decoding target image).
- the output unit 66 may be a storage unit that stores an image signal from a disk device or the like).
- the image encoding device 100 and the image decoding device 200 in the above-described embodiment may be realized by a computer.
- a program for realizing this function may be recorded on a computer-readable recording medium, and the program recorded on this recording medium may be read into a computer system and executed.
- the “computer system” includes an OS and hardware such as peripheral devices.
- the “computer-readable recording medium” refers to a storage device such as a flexible medium, a magneto-optical disk, a portable medium such as a ROM or a CD-ROM, and a hard disk incorporated in a computer system.
- the “computer-readable recording medium” dynamically holds a program for a short time like a communication line when transmitting a program via a network such as the Internet or a communication line such as a telephone line.
- a volatile memory inside a computer system serving as a server or a client in that case may be included and a program held for a certain period of time.
- the program may be for realizing a part of the functions described above, and may be a program capable of realizing the functions described above in combination with a program already recorded in the computer system. It may be realized using hardware such as PLD (Programmable Logic Device) or FPGA (Field Programmable Gate Array).
- the present invention can be applied to an application in which it is indispensable to achieve high coding efficiency without increasing the calculation amount and memory consumption during decoding.
- Reference depth map input unit 205... Viewpoint composite image generation unit, 206 ... viewpoint composite image memory, 207 ... Prediction unit information decoding unit, 208: Image decoding unit, 209 ... Frame-unit viewpoint composite image generation unit, 211... Switch 212: A block unit prediction information generation unit.
Landscapes
- Engineering & Computer Science (AREA)
- Multimedia (AREA)
- Signal Processing (AREA)
- Compression Or Coding Systems Of Tv Signals (AREA)
- Testing, Inspecting, Measuring Of Stereoscopic Televisions And Televisions (AREA)
Abstract
Description
本願は、2013年4月11日に、日本に出願された特願2013-82956号に基づき優先権を主張し、その内容をここに援用する。
なお、デプス情報から視差を得るために必要となる情報は別途与えられているものとする。具体的には、カメラAとカメラBの位置関係を表す外部パラメータや、カメラによる画像平面への投影情報を表す内部パラメータであるが、これら以外の形態であってもデプス情報から視差が得られるものであれば、別の情報が与えられていてもよい。これらのカメラパラメータに関する詳しい説明は、例えば、文献「Oliver Faugeras, "Three-Dimension Computer Vision", MIT Press; BCTC/UFF-006.37 F259 1993, ISBN:0-262-06158-9.」に記載されている。この文献には、複数のカメラの位置関係を示すパラメータや、カメラによる画像平面への投影情報を表すパラメータに関する説明が記載されている。
フレーム単位予測RDコスト計算部107は、視点合成画像を用いて符号化対象画像をフレーム単位で予測した際のレート歪みコストを計算する。画像符号化部108は、視点合成画像を用いながら、符号化対象画像をブロック単位で予測符号化する。ブロック単位予測RDコスト計算部109は、視点合成画像を用いながら、符号化対象画像をブロック単位で予測符号化した際のレート歪みコストを計算する。予測単位判定部110は、レート歪みコストに基づいて、符号化対象画像をフレーム単位で予測するか、ブロック単位で予測符号化するかを決定する。ビットストリーム生成部111は、予測単位決定部110の決定に基づき、符号化対象画像に対するビットストリームを構築して出力する。
図5及び図6は、図4に示す画像符号化装置100bの動作を示すフローチャートである。図5は、ブロック単位の予測を用いる符号化を全てのブロックに対して行った後に、予測単位の判定を行う場合の処理動作であり、図6はブロックごとに符号化と判定を繰り返す場合の処理動作である。図5または図6において、図2または図3に示すフローチャートと同じ処理を行う部分には同じ符号を付し、その説明を省略する。
図7および図8は、フレーム単位の予測を行うと判定された際に、符号化対象画像を分割したブロックごとに予測情報を生成して、別のフレームを符号化する際に参照できるようにする場合の画像符号化装置の構成を示すブロック図である。図7に示す画像符号化装置100cは図1に示す画像符号化装置100aに対応し、図8に示す画像符号化装置100dは図4に示す画像符号化装置100bに対応するブロック図である。それぞれで異なる点は、ブロック単位予測情報生成部116を更に備えている点である。なお、それぞれ同じ構成には同じ符号を付して、その説明を省略する。
ここではカメラAの画像が入力されるものとする。
図14に示す処理動作が、図11または図12に示す処理動作と異なる点は、復号して得られた予測単位に応じて(ステップS206)、生成する視点合成画像を切り替えている点である。なお、ブロック単位の予測を行う場合は、ブロックごとに、ブロック単位視点合成画像を生成する処理(ステップS210)と復号対象画像を復号する処理(ステップS211)とを繰り返すことになる。このフローチャートでは、復号するブロックに対するインデックスを示す変数をblkとし、復号対象画像内のブロック数numBlksで示した。
CPU50は、プログラムを実行する。RAM等のメモリ51は、CPU50がアクセスするプログラムやデータが格納する。符号化対象画像入力部52は、カメラ等からの符号化対象の画像信号を入力する(符号化対象画像入力部52は、ディスク装置等による画像信号を記憶する記憶部でもよい)。参照画像入力部53は、カメラ等からの参照対象の画像信号を入力する(参照画像入力部53は、ディスク装置等による画像信号を記憶する記憶部でもよい)。参照デプスマップ入力部54は、デプスカメラ等からの符号化対象画像を撮影したカメラとは異なる位置や向きのカメラに対するデプスマップを入力する(参照デプスマップ入力部54は、ディスク装置等によるデプスマップを記憶する記憶部でもよい)。プログラム記憶装置55は、画像符号化処理をCPU50に実行させるソフトウェアプログラムである画像符号化プログラム551を格納する。ビットストリーム出力部56は、CPU50がメモリ51にロードされた画像符号化プログラム551を実行することにより生成されたビットストリームを、例えばネットワークを介して出力する(ビットストリーム出力部56は、ディスク装置等によるビットストリームを記憶する記憶部でもよい)。
CPU60は、プログラムを実行する。RAM等のメモリ51は、CPU60がアクセスするプログラムやデータを格納する。ビットストリーム入力部62は、画像符号化装置が本手法により符号化したビットストリームを入力する(ビットストリーム入力部62は、ディスク装置等による画像信号を記憶する記憶部でもよい)。参照画像入力部63は、カメラ等からの参照対象の画像信号を入力する(参照画像入力部63は、ディスク装置等による画像信号を記憶する記憶部でもよい)。参照デプスマップ入力部64は、デプスカメラ等からの復号対象を撮影したカメラとは異なる位置や向きのカメラに対するデプスマップを入力する(参照デプスマップ入力部64は、ディスク装置等によるデプス情報を記憶する記憶部でもよい)。プログラム記憶装置65は、画像復号処理をCPU60に実行させるソフトウェアプログラムである画像復号プログラム651を格納する。復号対象画像出力部66は、CPU60がメモリ61にロードされた画像復号プログラム651を実行することにより、ビットストリームを復号して得られた復号対象画像を、再生装置などに出力する(復号対象画像出力部66は、ディスク装置等による画像信号を記憶する記憶部でもよい)。
102・・・符号化対象画像メモリ、
103・・・参照画像入力部、
104・・・参照デプスマップ入力部、
105・・・視点合成画像生成部、
106・・・視点合成画像メモリ、
107・・・フレーム単位予測RDコスト計算部、
108・・・画像符号化部、
109・・・ブロック単位予測RDコスト計算部、
110・・・予測単位判定部、
111・・・ビットストリーム生成部、
112・・・参照画像メモリ、
113・・・参照デプスマップメモリ、
114・・・フレーム単位視点合成画像生成部、
115・・・ブロック単位視点合成画像生成部、
116・・・ブロック単位予測情報生成部、
201・・・ビットストリーム入力部、
202・・・ビットストリームメモリ、
203・・・参照画像入力部、
204・・・参照デプスマップ入力部、
205・・・視点合成画像生成部、
206・・・視点合成画像メモリ、
207・・・予測単位情報復号部、
208・・・画像復号部、
209・・・フレーム単位視点合成画像生成部、
211・・・スイッチ、
212・・・ブロック単位予測情報生成部。
Claims (19)
- 複数の異なる視点の画像からなる多視点画像を符号化する際に、符号化対象画像とは異なる視点に対する符号化済みの参照画像と、前記参照画像中の被写体に対する参照デプスマップとを用いて、異なる視点間で画像を予測しながら符号化を行う画像符号化装置であって、
前記参照画像と前記参照デプスマップとを用いて、前記符号化対象画像全体に対する視点合成画像を生成する視点合成画像生成部と、
予測単位として、前記符号化対象画像を分割した符号化対象ブロックごとに予測を行うか、前記符号化対象画像全体に対して前記視点合成画像を用いて予測を行うかを選択する予測単位設定部と、
選択された前記予測単位を示す情報を符号化する予測単位情報符号化部と、
前記予測単位として前記符号化対象ブロックごとに予測を行うことが選択された場合に、前記符号化対象画像を、前記符号化対象ブロックごとに、予測画像生成方法を選択しながら予測符号化する符号化対象画像予測符号化部と
を備える画像符号化装置。 - 前記予測単位として前記符号化対象画像全体に対して前記視点合成画像を用いて予測を行うことが選択された場合に、前記符号化対象画像と前記視点合成画像の差分を符号化する視点合成予測残差符号化部を更に有する請求項1に記載の画像符号化装置。
- 前記符号化対象画像全体を前記視点合成画像で予測して符号化する場合のレート歪みコストである画像単位予測レート歪みコストを推定する画像単位予測レート歪みコスト推定部と、
前記符号化対象ブロックごとに、予測画像生成方法を選択しながら前記符号化対象画像を予測符号化する場合のレート歪みコストであるブロック単位予測レート歪みコストを推定するブロック単位予測レート歪みコスト推定部とを更に有し、
前記予測単位設定部では、前記画像単位予測レート歪みコストと前記ブロック単位予測レート歪みコストとを比較して前記予測単位を設定する請求項1または請求項2に記載の画像符号化装置。 - 前記符号化対象ブロックごとに、前記参照画像と前記参照デプスマップとを用いて、前記符号化対象ブロックに対する視点合成画像である部分視点合成画像を生成する部分視点合成画像生成部を更に有し、
前記符号化対象画像予測符号化部では、前記部分視点合成画像を予測画像の候補として用いる請求項1から3のいずれか1項に記載の画像符号化装置。 - 前記予測単位として画像全体に対して前記視点合成画像を用いて予測を行うことが選択された場合に、前記符号化対象ブロックごとに、予測情報を生成する予測情報生成部を更に有する請求項1から4のいずれか1項に記載の画像符号化装置。
- 前記予測情報生成部では、予測ブロックサイズを決定し、
前記視点合成画像生成部では、前記予測ブロックサイズごとに視点合成画像を生成する処理を繰り返すことで、前記符号化対象画像全体に対する前記視点合成画像を生成する請求項5に記載の画像符号化装置。 - 前記予測情報生成部では、視差ベクトルを推定し、視差補償予測としての予測情報を生成する請求項5に記載の画像符号化装置。
- 前記予測情報生成部では、予測方法を決定し、当該予測方法に対する予測情報を生成する請求項5に記載の画像符号化装置。
- 複数の異なる視点の画像からなる多視点画像の符号データから、復号対象画像を復号する際に、復号対象画像とは異なる視点に対する復号済みの参照画像と、前記参照画像中の被写体に対する参照デプスマップとを用いて、異なる視点間で画像を予測しながら復号を行う画像復号装置であって、
前記参照画像と前記参照デプスマップとを用いて、前記復号対象画像全体に対する視点合成画像を生成する視点合成画像生成部と、
前記符号データから前記復号対象画像を分割した復号対象ブロックごとに予測を行うか、前記復号対象画像全体に対して前記視点合成画像を用いて予測を行うかを示す予測単位の情報を復号する予測単位情報復号部と、
前記予測単位が前記復号画像全体に対して前記視点合成画像を用いて予測を行うことを示す場合に、前記視点合成画像を前記復号対象画像とする復号対象画像設定部と、
前記予測単位が前記復号対象ブロックごとに予測を行うことを示す場合に、前記符号データから、前記復号対象ブロックごとに予測画像を生成しながら、復号対象画像を復号する復号対象画像復号部と
を備える画像復号装置。 - 前記復号対象画像設定部において、符号データから前記復号対象画像と前記視点合成画像の差分を復号し、当該差分と前記視点合成画像とを足し合わせることで前記復号対象画像を生成する請求項9に記載の画像復号装置。
- 前記復号対象ブロックごとに、前記参照画像と前記参照デプスマップとを用いて、前記復号対象ブロックに対する視点合成画像である部分視点合成画像を生成する部分視点合成画像生成部を更に有し、
前記復号対象画像復号部では、前記部分視点合成画像を予測画像の候補として用いる請求項9または請求項10に記載の画像復号装置。 - 前記予測単位が前記復号画像全体に対して前記視点合成画像を用いて予測を行うことを示す場合に、前記復号対象ブロックごとに、予測情報を生成する予測情報生成部を更に有する請求項9から請求項11のいずれか1項に記載の画像復号装置。
- 前記予測情報生成部では、予測ブロックサイズを決定し、
前記視点合成画像生成部では、前記予測ブロックサイズごとに視点合成画像を生成する処理を繰り返すことで、前記復号対象画像全体に対する前記視点合成画像を生成する請求項12に記載の画像復号装置。 - 前記予測情報生成部では、視差ベクトルを推定し、視差補償予測としての予測情報を生成する請求項12に記載の画像復号装置。
- 前記予測情報生成部では、予測方法を決定し、当該予測方法に対する予測情報を生成する請求項12に記載の画像復号装置。
- 複数の異なる視点の画像からなる多視点画像を符号化する際に、符号化対象画像とは異なる視点に対する符号化済みの参照画像と、前記参照画像中の被写体に対する参照デプスマップとを用いて、異なる視点間で画像を予測しながら符号化を行う画像符号化方法であって、
前記参照画像と前記参照デプスマップとを用いて、前記符号化対象画像全体に対する視点合成画像を生成する視点合成画像生成ステップと、
予測単位として、前記符号化対象画像を分割した符号化対象ブロックごとに予測を行うか、前記符号化対象画像全体に対して前記視点合成画像を用いて予測を行うかを選択する予測単位設定ステップと、
選択された前記予測単位を示す情報を符号化する予測単位情報符号化ステップと、
前記予測単位として前記符号化対象ブロックごとに予測を行うことが選択された場合に、前記符号化対象画像を、前記符号化対象ブロックごとに、予測画像生成方法を選択しながら予測符号化する符号化対象画像予測符号化ステップと
を有することを特徴とする画像符号化方法。 - 複数の異なる視点の画像からなる多視点画像の符号データから、復号対象画像を復号する際に、復号対象画像とは異なる視点に対する復号済みの参照画像と、前記参照画像中の被写体に対する参照デプスマップとを用いて、異なる視点間で画像を予測しながら復号を行う画像復号方法であって、
前記参照画像と前記参照デプスマップとを用いて、前記復号対象画像全体に対する視点合成画像を生成する視点合成画像生成ステップと、
前記符号データから前記復号対象画像を分割した復号対象ブロックごとに予測を行うか、前記復号対象画像全体に対して前記視点合成画像を用いて予測を行うかを示す予測単位の情報を復号する予測単位情報復号ステップと、
前記予測単位が前記復号画像全体に対して前記視点合成画像を用いて予測を行うことを示す場合に、前記視点合成画像を前記復号対象画像とする復号対象画像設定ステップと、
前記予測単位が前記復号対象ブロックごとに予測を行うことを示す場合に、前記符号データから、前記復号対象ブロックごとに予測画像を生成しながら、前記復号対象画像を復号する復号対象画像復号ステップと
を有することを特徴とする画像復号方法。 - コンピュータに、請求項16に記載の画像符号化方法を実行させるための画像符号化プログラム。
- コンピュータに、請求項17に記載の画像復号方法を実行させるための画像復号プログラム。
Priority Applications (4)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US14/782,050 US20160037172A1 (en) | 2013-04-11 | 2014-04-07 | Image encoding method, image decoding method, image encoding apparatus, image decoding apparatus, image encoding program, and image decoding program |
KR1020157025982A KR20150122706A (ko) | 2013-04-11 | 2014-04-07 | 화상 부호화 방법, 화상 복호 방법, 화상 부호화 장치, 화상 복호 장치, 화상 부호화 프로그램, 및 화상 복호 프로그램 |
CN201480020084.3A CN105075257A (zh) | 2013-04-11 | 2014-04-07 | 图像编码方法、图像解码方法、图像编码装置、图像解码装置、图像编码程序、以及图像解码程序 |
JP2015511254A JP5926451B2 (ja) | 2013-04-11 | 2014-04-07 | 画像符号化方法、画像復号方法、画像符号化装置、画像復号装置、画像符号化プログラム、および画像復号プログラム |
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
JP2013082956 | 2013-04-11 | ||
JP2013-082956 | 2013-04-11 |
Publications (1)
Publication Number | Publication Date |
---|---|
WO2014168121A1 true WO2014168121A1 (ja) | 2014-10-16 |
Family
ID=51689528
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
PCT/JP2014/060116 WO2014168121A1 (ja) | 2013-04-11 | 2014-04-07 | 画像符号化方法、画像復号方法、画像符号化装置、画像復号装置、画像符号化プログラム、および画像復号プログラム |
Country Status (5)
Country | Link |
---|---|
US (1) | US20160037172A1 (ja) |
JP (1) | JP5926451B2 (ja) |
KR (1) | KR20150122706A (ja) |
CN (1) | CN105075257A (ja) |
WO (1) | WO2014168121A1 (ja) |
Cited By (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
WO2020105520A1 (ja) * | 2018-11-21 | 2020-05-28 | 日本電信電話株式会社 | 評価装置、評価方法、及びプログラム。 |
Families Citing this family (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US10051279B2 (en) * | 2014-07-01 | 2018-08-14 | Samsung Display Co., Ltd. | High quality display system combining compressed frame buffer and temporal compensation technique |
CN112313938B (zh) * | 2018-07-10 | 2022-06-21 | 奥林巴斯株式会社 | 摄像装置、图像校正方法以及计算机可读记录介质 |
US11107268B2 (en) * | 2018-09-07 | 2021-08-31 | Cognex Corporation | Methods and apparatus for efficient data processing of initial correspondence assignments for three-dimensional reconstruction of an object |
CN112114760A (zh) * | 2020-08-13 | 2020-12-22 | 西安万像电子科技有限公司 | 图像处理方法及装置 |
CN114079779B (zh) * | 2022-01-12 | 2022-05-17 | 深圳传音控股股份有限公司 | 图像处理方法、智能终端及存储介质 |
Citations (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
WO2007013194A1 (ja) * | 2005-07-26 | 2007-02-01 | National University Corporation Nagoya University | 画像情報圧縮方法及び自由視点テレビシステム |
JP2011519226A (ja) * | 2008-04-25 | 2011-06-30 | トムソン ライセンシング | 奥行きを用いた視点間スキップモード |
WO2011105297A1 (ja) * | 2010-02-23 | 2011-09-01 | 日本電信電話株式会社 | 動きベクトル推定方法、多視点映像符号化方法、多視点映像復号方法、動きベクトル推定装置、多視点映像符号化装置、多視点映像復号装置、動きベクトル推定プログラム、多視点映像符号化プログラム、及び多視点映像復号プログラム |
JP2012124564A (ja) * | 2010-12-06 | 2012-06-28 | Nippon Telegr & Teleph Corp <Ntt> | 多視点画像符号化方法,多視点画像復号方法,多視点画像符号化装置,多視点画像復号装置およびそれらのプログラム |
Family Cites Families (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
KR101059178B1 (ko) * | 2006-12-28 | 2011-08-25 | 니폰덴신뎅와 가부시키가이샤 | 영상 부호화 방법 및 복호방법, 그들의 장치, 그들의 프로그램을 기록한 기억매체 |
JP2009272702A (ja) * | 2008-04-30 | 2009-11-19 | Canon Inc | 動きベクトル探索装置及びその制御方法 |
-
2014
- 2014-04-07 JP JP2015511254A patent/JP5926451B2/ja active Active
- 2014-04-07 CN CN201480020084.3A patent/CN105075257A/zh active Pending
- 2014-04-07 KR KR1020157025982A patent/KR20150122706A/ko not_active Application Discontinuation
- 2014-04-07 US US14/782,050 patent/US20160037172A1/en not_active Abandoned
- 2014-04-07 WO PCT/JP2014/060116 patent/WO2014168121A1/ja active Application Filing
Patent Citations (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
WO2007013194A1 (ja) * | 2005-07-26 | 2007-02-01 | National University Corporation Nagoya University | 画像情報圧縮方法及び自由視点テレビシステム |
JP2011519226A (ja) * | 2008-04-25 | 2011-06-30 | トムソン ライセンシング | 奥行きを用いた視点間スキップモード |
WO2011105297A1 (ja) * | 2010-02-23 | 2011-09-01 | 日本電信電話株式会社 | 動きベクトル推定方法、多視点映像符号化方法、多視点映像復号方法、動きベクトル推定装置、多視点映像符号化装置、多視点映像復号装置、動きベクトル推定プログラム、多視点映像符号化プログラム、及び多視点映像復号プログラム |
JP2012124564A (ja) * | 2010-12-06 | 2012-06-28 | Nippon Telegr & Teleph Corp <Ntt> | 多視点画像符号化方法,多視点画像復号方法,多視点画像符号化装置,多視点画像復号装置およびそれらのプログラム |
Cited By (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
WO2020105520A1 (ja) * | 2018-11-21 | 2020-05-28 | 日本電信電話株式会社 | 評価装置、評価方法、及びプログラム。 |
Also Published As
Publication number | Publication date |
---|---|
JPWO2014168121A1 (ja) | 2017-02-16 |
US20160037172A1 (en) | 2016-02-04 |
JP5926451B2 (ja) | 2016-05-25 |
KR20150122706A (ko) | 2015-11-02 |
CN105075257A (zh) | 2015-11-18 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
JP5947977B2 (ja) | 画像符号化方法、画像復号方法、画像符号化装置、画像復号装置、画像符号化プログラム及び画像復号プログラム | |
JP6232076B2 (ja) | 映像符号化方法、映像復号方法、映像符号化装置、映像復号装置、映像符号化プログラム及び映像復号プログラム | |
JP5833757B2 (ja) | 画像符号化方法、画像復号方法、画像符号化装置、画像復号装置、画像符号化プログラム、画像復号プログラム及び記録媒体 | |
JP5926451B2 (ja) | 画像符号化方法、画像復号方法、画像符号化装置、画像復号装置、画像符号化プログラム、および画像復号プログラム | |
JP6307152B2 (ja) | 画像符号化装置及び方法、画像復号装置及び方法、及び、それらのプログラム | |
WO2014050830A1 (ja) | 画像符号化方法、画像復号方法、画像符号化装置、画像復号装置、画像符号化プログラム、画像復号プログラム及び記録媒体 | |
KR101648094B1 (ko) | 화상 부호화 방법, 화상 복호 방법, 화상 부호화 장치, 화상 복호 장치, 화상 부호화 프로그램, 화상 복호 프로그램 및 기록매체 | |
JP6053200B2 (ja) | 画像符号化方法、画像復号方法、画像符号化装置、画像復号装置、画像符号化プログラム及び画像復号プログラム | |
JP6027143B2 (ja) | 画像符号化方法、画像復号方法、画像符号化装置、画像復号装置、画像符号化プログラム、および画像復号プログラム | |
JP6232075B2 (ja) | 映像符号化装置及び方法、映像復号装置及び方法、及び、それらのプログラム | |
KR101750421B1 (ko) | 동화상 부호화 방법, 동화상 복호 방법, 동화상 부호화 장치, 동화상 복호 장치, 동화상 부호화 프로그램, 및 동화상 복호 프로그램 | |
JP5706291B2 (ja) | 映像符号化方法,映像復号方法,映像符号化装置,映像復号装置およびそれらのプログラム | |
JP6386466B2 (ja) | 映像符号化装置及び方法、及び、映像復号装置及び方法 | |
JP5759357B2 (ja) | 映像符号化方法、映像復号方法、映像符号化装置、映像復号装置、映像符号化プログラム及び映像復号プログラム | |
WO2015098827A1 (ja) | 映像符号化方法、映像復号方法、映像符号化装置、映像復号装置、映像符号化プログラム及び映像復号プログラム | |
JPWO2015141549A1 (ja) | 動画像符号化装置及び方法、及び、動画像復号装置及び方法 |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
WWE | Wipo information: entry into national phase |
Ref document number: 201480020084.3 Country of ref document: CN |
|
121 | Ep: the epo has been informed by wipo that ep was designated in this application |
Ref document number: 14782611 Country of ref document: EP Kind code of ref document: A1 |
|
ENP | Entry into the national phase |
Ref document number: 2015511254 Country of ref document: JP Kind code of ref document: A |
|
ENP | Entry into the national phase |
Ref document number: 20157025982 Country of ref document: KR Kind code of ref document: A |
|
WWE | Wipo information: entry into national phase |
Ref document number: 14782050 Country of ref document: US |
|
NENP | Non-entry into the national phase |
Ref country code: DE |
|
122 | Ep: pct application non-entry in european phase |
Ref document number: 14782611 Country of ref document: EP Kind code of ref document: A1 |