WO2011105337A1 - 多視点映像符号化方法、多視点映像復号方法、多視点映像符号化装置、多視点映像復号装置、及びプログラム - Google Patents
多視点映像符号化方法、多視点映像復号方法、多視点映像符号化装置、多視点映像復号装置、及びプログラム Download PDFInfo
- Publication number
- WO2011105337A1 WO2011105337A1 PCT/JP2011/053742 JP2011053742W WO2011105337A1 WO 2011105337 A1 WO2011105337 A1 WO 2011105337A1 JP 2011053742 W JP2011053742 W JP 2011053742W WO 2011105337 A1 WO2011105337 A1 WO 2011105337A1
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- viewpoint
- frame
- decoding
- image
- view
- Prior art date
Links
- 238000000034 method Methods 0.000 title claims abstract description 124
- 238000012937 correction Methods 0.000 claims abstract description 135
- 238000012545 processing Methods 0.000 claims abstract description 53
- 239000002131 composite material Substances 0.000 claims description 116
- 238000003702 image correction Methods 0.000 claims description 35
- 230000002194 synthesizing effect Effects 0.000 claims description 8
- 230000008569 process Effects 0.000 description 31
- 230000006870 function Effects 0.000 description 23
- 238000004364 calculation method Methods 0.000 description 22
- 230000015572 biosynthetic process Effects 0.000 description 20
- 238000003786 synthesis reaction Methods 0.000 description 20
- 230000008859 change Effects 0.000 description 13
- 239000013598 vector Substances 0.000 description 9
- 238000010586 diagram Methods 0.000 description 8
- 239000000203 mixture Substances 0.000 description 4
- 230000007423 decrease Effects 0.000 description 3
- 230000000694 effects Effects 0.000 description 3
- 238000012886 linear function Methods 0.000 description 3
- 238000013139 quantization Methods 0.000 description 3
- 238000006243 chemical reaction Methods 0.000 description 2
- 238000007796 conventional method Methods 0.000 description 2
- 230000001419 dependent effect Effects 0.000 description 2
- 238000007792 addition Methods 0.000 description 1
- 238000004458 analytical method Methods 0.000 description 1
- 238000013459 approach Methods 0.000 description 1
- 230000006835 compression Effects 0.000 description 1
- 238000007906 compression Methods 0.000 description 1
- 239000006185 dispersion Substances 0.000 description 1
- 239000011159 matrix material Substances 0.000 description 1
- 230000007246 mechanism Effects 0.000 description 1
- 238000012986 modification Methods 0.000 description 1
- 230000004048 modification Effects 0.000 description 1
- 238000011084 recovery Methods 0.000 description 1
- 230000009467 reduction Effects 0.000 description 1
- 230000004044 response Effects 0.000 description 1
- 238000006467 substitution reaction Methods 0.000 description 1
- 230000002123 temporal effect Effects 0.000 description 1
Images
Classifications
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/10—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
- H04N19/102—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the element, parameter or selection affected or controlled by the adaptive coding
- H04N19/103—Selection of coding mode or of prediction mode
- H04N19/105—Selection of the reference unit for prediction within a chosen coding or prediction mode, e.g. adaptive choice of position and number of pixels used for prediction
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/50—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding
- H04N19/597—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding specially adapted for multi-view video sequence encoding
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/46—Embedding additional information in the video signal during the compression process
- H04N19/463—Embedding additional information in the video signal during the compression process by compressing encoding parameters before transmission
Definitions
- the present invention relates to a multi-view video encoding method and multi-view video encoding apparatus for encoding a multi-view image or a multi-view video, a multi-view video decoding method and a multi-view video for decoding a multi-view image or a multi-view video.
- the present invention relates to a decoding device and a program.
- a multi-view image is a plurality of images obtained by photographing the same subject and its background with a plurality of cameras
- a multi-view video (multi-view video) is a moving image.
- efficient encoding is realized using motion compensation prediction using a high correlation existing between frames having different shooting times in a video.
- Motion compensated prediction is described in H.264. This is a technique adopted in recent international standards for video coding systems represented by H.264. That is, motion compensated prediction generates an image by compensating for the motion of a subject between an encoding target frame and an already encoded reference frame, and between the generated image and the encoding target frame. This is a method of taking a difference and encoding the difference signal and a motion vector.
- parallax compensation prediction that takes an inter-frame difference between an image (frame) generated by compensating for disparity between viewpoints instead of motion and an encoding target frame, and encodes a difference signal and a disparity vector. Is used.
- the parallax compensation prediction is performed in H.264. H.264 Annex. H is adopted as an international standard (for example, see Non-Patent Document 1).
- the parallax used here is a difference between positions at which the same position on the subject is projected on an image plane of a camera arranged in a different position or orientation. In the parallax compensation prediction, this is expressed by a two-dimensional vector and encoded. As shown in FIG. 7, since the parallax is information generated depending on the viewpoint position of the camera and the distance (depth) of the subject from the camera, it is called viewpoint synthesis prediction (view interpolation prediction) using this principle. There is a method.
- View synthesis prediction (view interpolation prediction) is encoded or decoded using a part of a multi-view video that has already been processed and obtained a decoding result according to the three-dimensional positional relationship between the camera and the subject.
- frames for different viewpoints to be processed are combined (interpolated), and an image obtained by combining is used as a predicted image (for example, see Non-Patent Document 2).
- Depth maps (sometimes called distance images, parallax images, and disparity maps) that represent the distance (depth) from the camera to the subject for each pixel are used to represent the three-dimensional position of the subject. There are many.
- polygon information of the subject and voxel information of the subject space can also be used.
- the method for obtaining the depth map can be broadly divided into a method for generating a depth map by measuring using infrared pulses, etc., and the principle of triangulation from the point of view of the same subject on a multi-viewpoint image. There is a method of generating a depth map after estimating the depth using. Which method to use the depth map obtained by is not a big problem in viewpoint synthesis prediction. If a depth map can be obtained, it is not a big problem where to estimate.
- parallax compensation prediction and viewpoint synthesis prediction there are individual differences in the response of the image sensor of the camera, gain control and gamma correction are performed for each camera, and there are direction-dependent lighting effects in the scene. Coding efficiency is degraded. This is because the prediction is performed on the assumption that the color of the subject is the same between the encoding target frame and the reference frame.
- Non-Patent Document 1 describes the H.P. H.264 employs Weighted Prediction that is corrected using a linear function.
- a method of correcting using a color table has been proposed (see, for example, Non-Patent Document 3).
- mismatch between the luminance and color cameras of these subjects is a subject-dependent local one, so that correction is normally performed using locally different correction parameters (correction parameters). It is better to do this.
- correction parameters correction parameters
- these mismatches occur according to a slightly complicated model such as not only a simple gain difference but also a focus difference. For this reason, it is desirable to use a complex correction model obtained by modeling a projection process or the like instead of a simple correction model.
- the present invention has been made in consideration of such circumstances, and its purpose is to separately encode a correction parameter even in a multi-view video including local brightness and color mismatch between cameras.
- a first aspect of the present invention is a multi-view video encoding method that encodes a multi-view video, the encoding target frame at the encoding target viewpoint of the multi-view video, and View synthesis that synthesizes a view synthesized image corresponding to the encoding target frame of the encoding target viewpoint from a reference viewpoint frame that has already been encoded at a reference view different from the encoding target viewpoint, which is captured at the same time.
- An image generation step and a reference region estimation step for searching for a reference region on a reference frame that has already been encoded in the encoding target viewpoint, corresponding to the viewpoint composite image, for each processing unit region having a predetermined size.
- a correction parameter for correcting an inter-camera mismatch from the viewpoint composite image for the processing unit region and the reference frame for the reference region a correction parameter for correcting an inter-camera mismatch from the viewpoint composite image for the processing unit region and the reference frame for the reference region.
- a correction parameter estimation step for determining, a viewpoint composite image correction step for correcting the viewpoint composite image for the processing unit region using the estimated correction parameter, and the code using the corrected viewpoint composite image A multi-view video encoding method including an image encoding step for predictively encoding a video of a conversion target viewpoint.
- the method further includes a reliability setting step of setting a reliability indicating the certainty of the viewpoint synthesized image for each pixel of the viewpoint synthesized image, wherein the reference area estimating step includes the reliability
- the matching cost of each pixel when searching the reference area on the reference frame corresponding to the viewpoint composite image may be weighted.
- the correction parameter estimation step may weight the matching cost of each pixel when estimating the correction parameter based on the reliability.
- the method further includes an estimation accuracy setting step for setting an estimation accuracy indicating whether or not the reference region has been correctly estimated for each pixel of the viewpoint composite image
- the correction parameter estimation step includes the step of A matching cost of each pixel when estimating the correction parameter may be weighted based on either or both of the estimation accuracy and the reliability.
- a second aspect of the present invention is a multi-view video decoding method for decoding multi-view video, which is the same time as a decoding target frame in the decoding target viewpoint of the multi-view video.
- a reference region estimation step for searching a reference region on a reference frame that has already been decoded, corresponding to the viewpoint composite image, and the viewpoint composite image for the processing unit region; From the reference frame for the reference region, a correction parameter estimation algorithm for estimating a correction parameter for correcting a mismatch between cameras.
- a viewpoint synthesized image correction step for correcting the viewpoint synthesized image for the processing unit region using the estimated correction parameter, and using the corrected viewpoint synthesized image as a prediction signal, the decoding target
- An image decoding step including decoding a decoding target frame that has been predictively encoded at a viewpoint from encoded video data for the decoding target viewpoint.
- the method further includes a reliability setting step of setting a reliability indicating the certainty of the viewpoint synthesized image for each pixel of the viewpoint synthesized image, wherein the reference area estimating step includes the reliability
- the matching cost of each pixel when searching the reference area on the reference frame corresponding to the viewpoint composite image may be weighted.
- the correction parameter estimation step may weight the matching cost of each pixel when estimating the correction parameter based on the reliability.
- the method further includes an estimation accuracy setting step for setting an estimation accuracy indicating whether or not the reference region has been correctly estimated for each pixel of the viewpoint composite image
- the correction parameter estimation step includes the step of A matching cost of each pixel when estimating the correction parameter may be weighted based on either or both of the estimation accuracy and the reliability.
- a third aspect of the present invention is a multi-view video encoding apparatus that encodes a multi-view video, and the encoding target at the encoding target viewpoint of the multi-view video.
- a view synthesized image corresponding to the encoding target frame of the encoding target viewpoint is synthesized from a reference viewpoint frame that has already been encoded at a reference viewpoint different from the encoding target viewpoint, which is captured at the same time as the frame.
- Reference already encoded in the encoding target viewpoint corresponding to the viewpoint synthesized image synthesized by the viewpoint synthesized image generating means for each processing unit area of a predetermined size with the viewpoint synthesized image generating means A reference area estimating means for searching for a reference area on the frame; the viewpoint synthesized image for the processing unit area; and the reference area searched for by the reference area estimating means.
- Correction parameter estimation means for estimating a correction parameter for correcting an inter-camera mismatch from the reference frame, and the viewpoint composite image for the processing unit area using the correction parameter estimated by the correction parameter estimation means.
- Multi-viewpoint video encoding comprising: a viewpoint composite image correction unit that corrects; and an image encoding unit that predictively encodes the video of the encoding target viewpoint using the viewpoint composite image corrected by the viewpoint composite image correction unit.
- the image processing apparatus further includes a reliability setting unit that sets a reliability indicating the certainty of the viewpoint synthesized image for each pixel of the viewpoint synthesized image synthesized by the viewpoint synthesized image generating unit,
- the reference area estimation means calculates the matching cost of each pixel when searching the reference area on the reference frame corresponding to the viewpoint synthesized image based on the reliability set by the reliability setting means. You may make it attach a weight.
- the correction parameter estimation unit weights the matching cost of each pixel when estimating the correction parameter based on the reliability set by the reliability setting unit. It may be.
- estimation accuracy setting means for setting an estimation accuracy indicating whether or not the reference region has been correctly estimated for each pixel of the viewpoint synthesized image synthesized by the viewpoint synthesized image generating means.
- the correction parameter estimation means determines the correction parameter based on one or both of the estimation accuracy set by the estimation accuracy setting means and the reliability set by the reliability setting means.
- a weight may be given to the matching cost of each pixel when estimating.
- a fourth aspect of the present invention is a multi-view video decoding device that decodes a multi-view video, and is the same time as a decoding target frame in the decoding target viewpoint of the multi-view video.
- a viewpoint synthesized image generating means for synthesizing a viewpoint synthesized image corresponding to the decoding target frame of the decoding target viewpoint from a reference viewpoint frame at a reference viewpoint different from the decoding target viewpoint captured in Reference area estimation means for searching for a reference area on a reference frame that has already been decoded in the decoding target viewpoint, corresponding to the viewpoint synthesized image synthesized by the viewpoint synthesized image generation means for each processing unit area; From the viewpoint composite image for the processing unit area and the reference frame for the reference area searched by the reference area estimation means, a camera Correction parameter estimation means for estimating a correction parameter for correcting an inter-miscue, and viewpoint composite image correction means for correcting the viewpoint composite image for the processing unit region using the correction parameter estimated by the correction parameter estimation means; , Using the viewpoint synthesized image corrected by the viewpoint synthesized image correcting unit as a prediction signal, decoding the decoding target frame that is predictively encoded at the decoding target viewpoint from the encoded data of the video for the decoding target viewpoint A
- a fifth aspect of the present invention provides a computer of a multi-view video encoding apparatus that encodes a multi-view video to an encoding target at the encoding target viewpoint of the multi-view video.
- a view synthesized image corresponding to the encoding target frame of the encoding target viewpoint is synthesized from a reference viewpoint frame that has already been encoded at a reference viewpoint different from the encoding target viewpoint, which is captured at the same time as the frame.
- View synthesized image generation function for each processing unit area of a predetermined size, a reference area estimation that searches for a reference area on a reference frame that has already been coded in the view to be coded that corresponds to the view synthesized image Function, a correction parameter for correcting an inter-camera mismatch from the viewpoint composite image for the processing unit area and the reference frame for the reference area
- a correction parameter estimation function to estimate, a viewpoint composite image correction function to correct the viewpoint composite image for the processing unit region using the estimated correction parameter, and the encoding target using the corrected viewpoint composite image
- a sixth aspect of the present invention provides a computer of a multi-view video decoding device that decodes a multi-view video at the same time as a decoding target frame in the decoding target viewpoint of the multi-view video.
- a view synthesized image generation function for synthesizing a view synthesized image corresponding to the decoding target frame of the decoding target viewpoint from a reference viewpoint frame in a reference view different from the decoding target viewpoint taken in For each processing unit region, a reference region estimation function for searching for a reference region on a reference frame that has already been decoded at the decoding target viewpoint corresponding to the viewpoint composite image, the viewpoint composite image and the reference for the processing unit region
- a correction parameter estimation function for estimating a correction parameter for correcting a mismatch between cameras from the reference frame for the region;
- a viewpoint composite image correction function for correcting the viewpoint composite image for the processing unit region using the estimated correction parameter, and predictive encoding at the decoding target viewpoint using the corrected viewpoint composite image as a prediction signal
- the encoding of an efficient multi-view image or multi-view video is not performed without separately encoding / decoding correction parameters. Can be realized.
- a corresponding area on a frame that has already been encoded with respect to the area being processed is obtained, and a video signal in the corresponding area in the encoded frame is used as a reference.
- the brightness and color of the composite image are corrected.
- correction is performed using the assumption that color and luminance mismatches that depend on the subject do not change significantly in time, rather than the assumption that the same subject appears in the adjacent region used by the conventional method. Find the parameters. In general, since a plurality of subjects are included in a frame, there is always an area that deviates from the conventional assumption.
- the embodiment of the present invention functions effectively. That is, even in a region where the conventional method has failed to be corrected, it is possible to perform correction to reduce mismatch, and it is possible to realize efficient multi-view video coding.
- FIG. 1 is a block diagram showing a configuration of a multi-view video encoding apparatus according to the first embodiment of the present invention.
- a multi-view video encoding apparatus 100 includes an encoding target frame input unit 101, an encoding target image memory 102, a reference viewpoint frame input unit 103, a reference viewpoint image memory 104, a viewpoint synthesis unit 105, and a viewpoint synthesized image memory.
- a reliability setting unit 107 a view synthesized image correction unit 108, a prediction residual encoding unit 109, a prediction residual decoding unit 110, a decoded image memory 111, a prediction residual calculation unit 112, and a decoded image calculation unit 113.
- the encoding target frame input unit 101 inputs a video frame (encoding target frame) to be encoded.
- the encoding target image memory 102 stores the input encoding target frame.
- the reference view frame input unit 103 inputs a reference video frame (reference view frame) for a view (reference view) different from the encoding target frame.
- the reference viewpoint image memory 104 stores the input reference viewpoint frame.
- the view synthesis unit 105 generates a view synthesized image for the encoding target frame using the reference view frame.
- the viewpoint synthesized image memory 106 stores the generated viewpoint synthesized image.
- the reliability setting unit 107 sets the reliability for each pixel of the generated viewpoint composite image.
- the viewpoint synthesized image correction unit 108 corrects the mismatch between the cameras in the viewpoint synthesized image and outputs a corrected viewpoint synthesized image.
- the prediction residual calculation unit 112 generates a difference (prediction residual signal) between the encoding target frame and the corrected viewpoint synthesized image.
- the prediction residual encoding unit 109 encodes the generated prediction residual signal and outputs encoded data.
- the prediction residual decoding unit 110 decodes encoded data of the prediction residual signal.
- the decoded image calculation unit 113 adds the decoded prediction residual signal and the corrected viewpoint synthesized image to generate a decoded image of the encoding target frame.
- the decoded image memory 111 stores the generated decoded image.
- FIG. 2 is a block diagram showing a configuration of the viewpoint synthesized image correction unit 108 of the multi-view video encoding apparatus 100 according to the first embodiment.
- the view synthesized image correction unit 108 of the first embodiment includes a reference area setting unit 1081 that searches for a block on a reference frame corresponding to an encoding target block as a reference area using the view synthesized image, and a reference.
- An estimation accuracy setting unit 1082 for setting an estimation accuracy indicating whether or not a corresponding region has been correctly set for each pixel of the region
- a correction parameter estimation unit 1083 for estimating a parameter for correcting an inter-camera mismatch in a viewpoint composite image
- An image correction unit 1084 that corrects the viewpoint composite image in accordance with the obtained correction parameter.
- FIG. 3 is a flowchart for explaining the operation of the multi-view video encoding apparatus 100 according to the first embodiment. The process executed by the multi-view video encoding apparatus 100 will be described in detail according to this flowchart.
- the viewpoint synthesis unit 105 synthesizes an image shot at the same viewpoint at the same time as the encoding target frame from the information of the reference viewpoint frame, and accumulates the generated viewpoint synthesized image Syn in the viewpoint synthesized image memory 106.
- Step Sa2 Any method may be used as a method of generating the viewpoint composite image Syn. For example, if the depth information for the reference viewpoint frame is given in addition to the video information of the reference viewpoint frame, Non-Patent Document 2 and Non-Patent Document 5 (Y.YMori, N. Fukushima, T.
- Fujii, and M .Tanimoto, “View Generation Generation with 3D Warping” Using “Depth Information” for “FTV,” “Proceedings” of “3DTV-CON2008,” pp. 229-232, “May” 2008.) can be used.
- Non-Patent Document 6 S. Yea and A. Vetro, “View Synthesis Prediction for Rate-Overhead Reduction in FTV,” Proceedings of 3DTV-CON2008, pp 145-148, ⁇ May 2008.
- Non-Patent Document 7 J. Sun, N. Zheng, and H. Shum, “Stereo Matching Using Belief Propagation,” IEEE Transactions on Pattern Analysis and Machine Intelligence, Vol. 25, No. .7, pp.
- Non-Patent Document 8 S. Shimizu, Y. Tonomura, H. Kimata, and Y. Ohtani, “Improved View Interpolation Prediction for Side Information in Multiview Distributed) Video Coding, ”Proceedings of ICDSC2009, August 2009.).
- the reliability setting unit 107 generates, for each pixel of the viewpoint composite image, a reliability ⁇ indicating how accurately the composition for the pixel can be realized (step Sa3).
- the reliability ⁇ is a real number from 0 to 1.
- the reliability can be expressed in any way. I do not care.
- the reliability may be expressed by an 8-bit integer of 1 or more.
- the reliability ⁇ may be any value as long as it can indicate how accurately the composition is performed.
- the reliability is represented using the following formula (1) and formula (2). It is possible.
- max is a function that returns the maximum value for a given set.
- the other functions are expressed by the following mathematical formula (3).
- the reference viewpoint frame is clustered by the pixel value of the corresponding pixel, and the variance value, the maximum value and the minimum value are compared with the pixel value of the corresponding pixel of the reference viewpoint frame belonging to the largest cluster. You may calculate and use the difference.
- the error at the corresponding points between the viewpoints follows a normal distribution or a Laplace distribution, and each of the values obtained by diff in the above formula (4) using the average value of the distribution and the variance value as parameters.
- the reliability may be defined using a probability value corresponding to the error amount of the pixel.
- a distribution model, an average value, and a variance value may be determined in advance, or information on the used model may be encoded and transmitted. In general, if the subject is completely diffusely reflected, the average value of the distribution can be considered as 0 theoretically, so the model may be simplified.
- the error from the change in the error amount when the depth is changed slightly.
- the probability that the error is within a certain range when the error occurrence probability follows the error distribution is used as the reliability.
- the definition using the error distribution model and the pixel value of the corresponding pixel on the reference viewpoint frame at the time of generating the viewpoint composite image is the reference at the time of generating the viewpoint composite image when the error distribution probability follows the estimated error distribution.
- the probability of occurrence of a situation represented by the pixel value of the corresponding pixel on the viewpoint frame is used as the reliability.
- parallax (depth) obtained when using a method called Belief Propagation (the above-mentioned non-patent document 7) when estimating the parallax (depth) required when performing viewpoint synthesis.
- the probability value for can be used as the reliability.
- Belief Propagation it is possible to use the information as reliability as long as it is a depth estimation algorithm that internally calculates the probability of the solution for each pixel of the viewpoint composite image.
- a part of processing for obtaining corresponding point information and depth information may be the same as part of reliability calculation. In such a case, it is possible to reduce the amount of calculation by simultaneously performing the viewpoint composite image generation and the reliability calculation.
- the encoding target frame is divided into blocks, and the video signal of the encoding target frame is corrected for each area by the viewpoint synthesized image correction unit 108 while correcting the camera-to-camera mismatch of the viewpoint synthesized image.
- Is encoded steps Sa4 to Sa12). That is, assuming that the encoding target block index is blk and the total number of encoding target blocks is numBlks, blk is initialized to 0 (step Sa4), and then 1 is added to blk (step Sa11). The following processing (steps Sa5 to Sa10) is repeated until numBlks is reached (step Sa12).
- those processes can also be performed as part of the process repeated for each encoding target block. is there. For example, this corresponds to the case where depth information for the encoding target block is given.
- the reference area setting unit 1081 uses the viewpoint synthesized image to find a reference area that is a block on the reference frame corresponding to the block blk (step Sa5).
- the reference frame is a local decoded image obtained by decoding data that has already been encoded.
- the local decoded image data is data stored in the decoded image memory 111.
- the reason for using the local decoded image is to prevent the occurrence of coding distortion called drift by using the same data that can be acquired at the same timing on the decoding side.
- an input frame encoded before the encoding target frame may be used instead of the local decoded image.
- the process for obtaining the reference area is a process for obtaining on the local decoded image stored in the decoded image memory 111 a corresponding block that maximizes the fitness or minimizes the divergence using the viewpoint synthesized image Syn [blk] as a template. It is.
- a matching cost indicating the degree of divergence is used. Specific examples of the matching cost indicating the degree of divergence include the following formula (5) and formula (6).
- vec is a vector between corresponding blocks
- t is an index value indicating one of the local decoded images Dec stored in the decoded image memory 111.
- DCT Discrete Cosine Transform
- a matrix A it can be represented by the following formula (7) or formula (8).
- represents the norm of X.
- the process for obtaining a block that minimizes the matching cost is to obtain a set of (best_vec, best_t) represented by the following formula (9).
- argmin indicates a process for obtaining a parameter that minimizes a given function.
- the set of parameters to be derived is a set given at the lower part of argmin.
- Any method may be used as the method for determining the number of frames to be searched, the search range, the search order, and censoring.
- the search range and the truncation method greatly affect the calculation cost.
- there is a method of appropriately setting a search center As one example, there is a method in which the corresponding point represented by the motion vector used in the corresponding region on the reference viewpoint frame is set as the search center.
- a frame determination method to be searched may be determined in advance. For example, this corresponds to a method in which the most recently encoded frame is the search target.
- a method for limiting the search target frame there is a method of encoding information indicating which frame is targeted and notifying the decoding side. In this case, the decoding side needs to have a mechanism for decoding information such as an index value indicating the search target frame and determining the search target frame based on the decoded information.
- the necessary data is a predicted value of the video signal of the encoding target block expressed using video signals of temporally different frames. For this reason, a corresponding pixel may be obtained for each pixel in the encoding target block blk, and a video signal created by arranging them in a block shape may be used as the reference region. Also, a plurality of blocks corresponding to the encoding target block blk may be set, and a video signal represented by an average value of the video signals in the plurality of blocks may be used as the reference area. By doing so, when noise is superimposed on the search target frame or when the search accuracy is low, it is possible to reduce the influence thereof and set the reference region more robustly.
- the estimation accuracy ⁇ indicating how correctly the reference area is obtained for each pixel of the reference area Ref [blk] by the estimation accuracy setting unit 1082 Is set (step Sa6).
- Any value may be used for the estimation accuracy, but a value depending on an error amount between corresponding pixels in the viewpoint synthesized image and the reference frame can be used. For example, minus the square error represented by Equation (10) or Equation (11), the reciprocal of the absolute value error, the square error represented by Equation (12) or Equation (13), or the absolute value error. There is a value multiplied by.
- Equation (14) is an example in which a Laplace distribution with an average of 0 is used, and ⁇ is a parameter.
- the correction parameter estimation unit 1083 estimates a correction parameter for correcting the viewpoint composite image Syn [blk] (step Sa7). Any method may be used for the correction method and the correction parameter estimation, but it is necessary to use the same method as that used on the decoding side.
- correction methods include correction using an offset value, correction using a linear function, and gamma correction. If the value before correction is in and the value after correction is out, they can be expressed by the following equations (15), (16), and (17).
- offset, ( ⁇ , ⁇ ), and ( ⁇ , a, b) are correction parameters, respectively.
- the value before correction is the image signal of the viewpoint composite image
- the ideal value after correction is the image signal of the reference region. It becomes. That is, highly accurate correction can be performed by obtaining the correction parameter so that the matching cost represented by the degree of divergence between the two image signals is reduced.
- the matching cost is expressed by the degree of matching between the two image signals, the parameter is obtained so as to maximize the matching cost.
- the process for obtaining the correction parameter can be expressed by the following equation (18).
- par F indicates a set of correction parameters of the correction method F
- argmin indicates a process for obtaining a parameter that minimizes a given function.
- the set of parameters to be derived is a set given at the lower part of argmin. Any matching cost may be used. For example, the square of the difference between two signals can be used. Also, in the matching cost, weighting may be performed for each pixel using the reliability of the viewpoint composite image, the estimation accuracy of the reference region, or both.
- the following mathematical formulas (19), (20), (21), and (22) are the reliability of the viewpoint composite image when weighting is not performed at all when the square of the difference between two signals is used as the divergence.
- the offset can be obtained using the following formula (23).
- correction parameters may be obtained for each luminance and color difference signal, or for each color channel such as RGB. It is also possible to subdivide each channel and perform different corrections for each fixed range (for example, correction using different correction parameters for R channels 0 to 127 and 128 to 255).
- the image correction unit 1084 corrects the viewpoint composite image for the block blk according to the correction parameter, and generates the corrected viewpoint composite image Pred (step Sa8).
- the viewpoint composite image may be input to the correction model into which the correction parameter is substituted. For example, when correction using an offset value is performed, a corrected viewpoint composite image Pred is generated according to the following formula (24).
- the encoding target frame Org [blk] is predictively encoded using the corrected viewpoint composite image Pred as a prediction image (step Sa9). That is, the difference between the encoding target frame Org [blk] and the corrected viewpoint synthesized image Pred is generated as a prediction residual by the prediction residual calculation unit 112, and the prediction residual is encoded by the prediction residual encoding unit 109.
- Any encoding method may be used. In a general encoding method such as H.264, encoding is performed by applying DCT, quantization, binarization, and entropy encoding to the prediction residual.
- the bit stream of the encoding result becomes an output of the multi-view video encoding apparatus 100, and is decoded by the prediction residual decoding unit 110 for each block, and the decoded result and the corrected viewpoint synthesized image Pred are decoded by the decoded image calculation unit 113.
- the local decoded image Dec cur [blk] is constructed by adding them.
- the constructed local decoded image is stored in the decoded image memory 111 for use in future prediction (step Sa10).
- FIG. 4 is a block diagram showing the configuration of the multi-view video decoding apparatus according to the second embodiment.
- the multi-view video decoding apparatus 200 includes an encoded data input unit 201, an encoded data memory 202, a reference viewpoint frame input unit 203, a reference viewpoint image memory 204, a viewpoint synthesis unit 205, a viewpoint synthesized image memory 206, a trust A degree setting unit 207, a viewpoint synthesized image correction unit 208, a prediction residual decoding unit 210, a decoded image memory 211, and a decoded image calculation unit 212.
- the encoded data input unit 201 inputs encoded data for a video frame to be decoded (decoding target frame).
- the encoded data memory 202 stores input encoded data.
- the reference viewpoint frame input unit 203 inputs a reference viewpoint frame that is a video frame for a viewpoint different from the decoding target frame.
- the reference viewpoint image memory 204 stores the input reference viewpoint frame.
- the view synthesis unit 205 generates a view synthesized image for the decoding target frame using the reference view frame.
- the viewpoint composite image memory 206 stores the generated viewpoint composite image.
- the reliability setting unit 207 sets the reliability for each pixel of the generated viewpoint composite image.
- the viewpoint composite image correction unit 208 corrects the camera-to-camera mismatch of the viewpoint composite image and outputs a corrected viewpoint composite image.
- the prediction residual decoding unit 210 decodes the difference between the decoding target frame and the corrected viewpoint synthesized image from the encoded data as a prediction residual signal.
- the decoded image memory 211 stores a decoded image of the decoding target frame obtained by adding the decoded prediction residual signal and the corrected viewpoint synthesized image by the decoded image calculation unit 212.
- the prediction error decoding unit 210, and the decoded image memory 211 are respectively the reference view frame input unit 103, the reference view image memory 104, the view synthesis unit 105, and the view synthesis in the multi-view video encoding device 100 according to the first embodiment. This is the same as the image memory 106, the reliability setting unit 107, the viewpoint composite image correction unit 108, the prediction error decoding unit 110, and the decoded image memory 111.
- the configuration of the viewpoint composite image correction unit 208 is the same as that of the viewpoint composite image correction unit 108 (FIG. 2) of the multi-view video encoding apparatus 100 according to the first embodiment described above. However, in the following description, as illustrated in FIG. 5, the reference area setting unit 2081, the estimation accuracy setting unit 2082, the correction parameter estimation unit 2083, and the image correction unit 2084 will be described.
- FIG. 6 is a flowchart for explaining the operation of the multi-view video decoding apparatus 200 according to the second embodiment. The processing executed by the multi-view video decoding apparatus 200 will be described in detail according to this flowchart.
- the reference viewpoint frame input here is a separately decoded image.
- drift it is necessary to input the same one used in the coding apparatus.
- a different one from that used in the encoding apparatus may be input.
- n is an index indicating a reference viewpoint
- N is the number of reference viewpoints available here.
- the viewpoint synthesis unit 205 synthesizes an image shot from the same viewpoint at the same time as the decoding target frame from the information of the reference viewpoint frame, and accumulates the generated viewpoint synthesized image Syn in the viewpoint synthesized image memory 206.
- the reliability setting unit 207 generates, for each pixel of the viewpoint composite image, a reliability ⁇ indicating how much the composition for the pixel can be realized (step Sb3).
- the video signal of the decoding target frame is decoded while correcting the mismatch between the cameras of the viewpoint composite image by the viewpoint composite image correction unit 208 for each predetermined block (steps Sb4 to Sb12). . That is, if the decoding target block index is represented by blk and the total number of decoding target blocks is represented by numBlks, blk is initialized with 0 (step Sb4), and then 1 is added to blk (step Sb11), and blk is added to numBlks. The following processing (steps Sb5 to Sb10) is repeated until it becomes (step Sb12).
- step Sb9 described later may be performed for all blocks in advance without being performed for each block, and the result may be accumulated and used. However, in that case, a memory for storing the decoded prediction residual signal is required.
- the reference region setting unit 2081 uses the viewpoint synthesized image and uses a reference frame as a block on the reference frame corresponding to the block blk. Ref [blk] is found (step Sb5).
- the reference frame is data that has already been decoded and accumulated in the decoded image memory 211.
- step Sa5 This process is the same as step Sa5 of the first embodiment.
- Matching costs for search, determination method of search target frame, generation method of video signal for reference area, etc. can be used to prevent noise generation by using the same method as used in the encoding device. It is.
- the estimation accuracy setting unit 2082 determines how accurate the reference region is for each pixel of the reference region Ref [blk]. Estimated accuracy ⁇ indicating whether it has been obtained is set (step Sb6). Thereafter, the correction parameter estimation unit 2083 ( ⁇ correction parameter estimation unit 1083) estimates a correction parameter for correcting the viewpoint composite image Syn [blk] (step Sb7). Next, the image correction unit 2084 ( ⁇ image correction unit 1084) corrects the viewpoint composite image for the block blk according to the correction parameter, and generates the corrected viewpoint composite image Pred (step Sb8). These processes are the same as steps Sa6, Sa7, and Sa8 of the first embodiment, respectively.
- the prediction error decoding unit 210 decodes the prediction residual signal for the block blk from the encoded data (step Sb9).
- the decoding process here is a process corresponding to the encoding method. For example, H.M.
- decoding is performed by performing IDCT (Inverse Discrete Cosine Transform), inverse quantization, multi-value quantization, entropy decoding, and the like. .
- decoded decoded residual signal DecRes and the corrected viewpoint synthesized image Pred are added together by the decoded image calculation unit 212 to construct a decoding target frame Dec cur [blk].
- the constructed decoding target frame is stored in the decoded image memory 211 and used as an output of the multi-view video decoding device 200 for use in future prediction (step Sb10).
- a corresponding area on a frame that has already been encoded with respect to the area being processed is obtained, and an image of the corresponding area in the encoded frame is obtained.
- the luminance and color of the viewpoint composite image are corrected using the signal as a reference.
- a reliability indicating the certainty of the composition process is set for each pixel of the viewpoint composite image, and the matching cost is weighted for each pixel based on the reliability.
- step Sa5 of the first embodiment and step Sb5 of the second embodiment described above the correspondence on the reference frame corresponding to the viewpoint composite image Syn [blk] of the processing target frame (the encoding target frame or the decoding target frame).
- the block is obtained using the reference frame Dec.
- the viewpoint composite image RefSyn of the reference frame is obtained, the corresponding block may be obtained using the viewpoint composite image RefSyn instead of the reference frame Dec.
- the set of (best_vec, best_t) shown in equation (9) is obtained, so that the corresponding block on the reference frame is obtained. You can ask for it.
- the reference area Ref is generated using the reference frame Dec.
- the view synthesized image RefSyn and the reference frame Dec are considered to be equal. Therefore, even if the corresponding block is searched using the view synthesized image RefSyn in this way, the effect of the embodiment of the present invention is achieved. Is similarly obtained.
- the viewpoint composite image RefSyn When using the viewpoint composite image RefSyn, it is necessary to input a reference viewpoint frame photographed at the same time as the reference frame, generate a viewpoint composite image of the reference frame, and store it.
- the viewpoint synthesized image is continuously accumulated in the viewpoint synthesized image memory while the processed frames are accumulated in the decoded image memory.
- the processed region stored in the decoded image memory in the corresponding region search (step Sa5 in the first embodiment and step Sb5 in the second embodiment) is not required, and therefore the corresponding region search is performed.
- This process need not be performed in synchronization with the encoding process or the decoding process. As a result, parallel calculation or the like can be performed, and an effect of reducing the entire calculation time can be obtained.
- the viewpoint composite image and the reference frame are used as they are.
- the accuracy of the corresponding area search decreases due to the influence of noise such as film grain and coding distortion generated in the viewpoint composite image and the reference frame. Since these noises are specific frequency components (especially high-frequency components), search is performed after applying a band-pass filter (or low-pass filter if the noise is high-frequency) to the frame (image) used for the corresponding region search By performing the above, it is possible to reduce the influence of noise.
- the accuracy of the corresponding area search decreases due to the influence of noise or the like, the spatial correlation of the vector indicating the corresponding area decreases.
- the accuracy of the corresponding region search may be improved by applying an average value filter or a median filter to the motion vector estimated for each block to increase the spatial correlation.
- the processing target blocks have been described as having the same size as the corresponding region search block, but it is clear that they need not be the same size. Since the temporal change of the video is non-linear, the change of the video signal can be predicted more accurately by finding the corresponding area for each smaller block. However, when a small block is used, not only the calculation amount increases, but also the influence of noise included in the video signal increases. In order to cope with this problem, when searching for a corresponding region for a small region, it is also a process in a range that can be easily analogized to reduce the influence of noise by using several pixels around the small region for the search.
- the process of encoding or decoding one frame of one camera has been described. By repeating this process for each frame, encoding of a multi-view video is performed. Or decoding can be realized. Furthermore, by repeating the processing for each camera, it is possible to realize encoding or decoding of multi-view video images of a plurality of cameras.
- the correction parameter is obtained using the assumption that the color and luminance mismatch depending on the subject does not change greatly in time. Therefore, when the scene changes suddenly due to a scene change or the like, the mismatch changes with time. In such a case, in the embodiment of the present invention, an appropriate correction parameter cannot be estimated, and there is a possibility that the difference between the viewpoint synthesized image and the processing target frame is increased by the correction. Therefore, it is possible to determine whether or not there is a sudden change such as a scene change and correct the viewpoint composite image only when it is determined that there is no such a sudden change in video. As a method for determining such a sudden video change, the value of the degree of divergence of the corresponding area obtained as a result of the corresponding area search is checked. You may use the method of determining with having generate
- the processing described above can also be realized by a computer and a software program. Further, the program can be provided by being recorded on a computer-readable recording medium or can be provided through a network.
- the multi-view video encoding device and the multi-view video decoding device have been mainly described.
- the multi-view video encoding method and the multi-view video decoding method of the present invention can be realized by steps corresponding to the operations of the respective units of the multi-view video encoding device and the multi-view video decoding device.
- the present invention is used, for example, for encoding and decoding a multi-view image or a multi-view video. According to the present invention, even when a luminance or color mismatch between cameras occurs locally, the encoding of an efficient multi-view image or multi-view video is not performed without separately encoding / decoding correction parameters. Can be realized.
- Multiview video encoding apparatus 101 Encoding object frame input part 102 Encoding object image memory 103 Reference viewpoint frame input part 104 Reference viewpoint image memory 105 Viewpoint synthesis part 106 Viewpoint synthetic image memory 107 Reliability setting part 108 Viewpoint synthetic image correction Unit 109 prediction residual encoding unit 110 prediction residual decoding unit 111 decoded image memory 112 prediction residual calculation unit 113 decoded image calculation unit 1081 reference region setting unit 1082 estimation accuracy setting unit 1083 correction parameter estimation unit 1084 image correction unit 200 multi Viewpoint video decoding apparatus 201 Encoded data input unit 202 Encoded data memory 203 Reference view frame input unit 204 Reference view image memory 205 Viewpoint synthesis unit 206 Viewpoint synthesized image memory 207 Reliability setting unit 208 Viewpoint synthesized image correction unit 210 Prediction residual Recovery No. 211 Decoded image memory 212 Decoded image calculation unit
Landscapes
- Engineering & Computer Science (AREA)
- Multimedia (AREA)
- Signal Processing (AREA)
- Compression Or Coding Systems Of Tv Signals (AREA)
- Testing, Inspecting, Measuring Of Stereoscopic Televisions And Televisions (AREA)
Abstract
Description
本願は、2010年2月24日に日本へ出願された日本特願2010-038680号に対して優先権を主張し、その内容をここに援用する。
なお、以下の説明では、映像(フレーム)に記号[]で挟まれた位置を特定可能な情報(座標値、もしくは座標値に対応付け可能なインデックス)を付加することで、その位置の画素に関してサンプリングされた映像信号を示すものとする。
まず、本発明の第1実施形態について説明する。
図1は、本発明の第1実施形態による多視点映像符号化装置の構成を示すブロック図である。図1において、多視点映像符号化装置100は、符号化対象フレーム入力部101、符号化対象画像メモリ102、参照視点フレーム入力部103、参照視点画像メモリ104、視点合成部105、視点合成画像メモリ106、信頼度設定部107、視点合成画像補正部108、予測残差符号化部109、予測残差復号部110、復号画像メモリ111、予測残差算出部112、及び復号画像算出部113を備えている。
つまり、補正処理を表す関数をFとし、2つの画像信号の乖離度を表すマッチングコスト関数をCとすると、補正パラメータを求める処理は、次の数式(18)で表すことができる。
マッチングコストにはどのようなものを用いても構わないが、例えば、2つの信号の差の二乗を用いることができる。また、マッチングコストにおいて、視点合成画像の信頼度や、参照領域の推定精度や、その両方を用いて画素ごとに重み付けを行っても構わない。次の数式(19),(20),(21),(22)は、2つの信号の差の二乗を乖離度とする際に、それぞれ、全く重み付けを行わない場合、視点合成画像の信頼度で重み付けを行う場合、参照領域の推定精度で重み付けを行う場合、視点合成画像の信頼度と参照領域の推定精度の両方で重み付けを行う場合のマッチングコスト関数の例を示す。
次に、本発明の第2実施形態について説明する。
図4は、本第2実施形態による多視点映像復号装置の構成を示すブロック図である。図4において、多視点映像復号装置200は、符号化データ入力部201、符号化データメモリ202、参照視点フレーム入力部203、参照視点画像メモリ204、視点合成部205、視点合成画像メモリ206、信頼度設定部207、視点合成画像補正部208、予測残差復号部210、復号画像メモリ211及び復号画像算出部212を備えている。
101 符号化対象フレーム入力部
102 符号化対象画像メモリ
103 参照視点フレーム入力部
104 参照視点画像メモリ
105 視点合成部
106 視点合成画像メモリ
107 信頼度設定部
108 視点合成画像補正部
109 予測残差符号化部
110 予測残差復号部
111 復号画像メモリ
112 予測残差算出部
113 復号画像算出部
1081 参照領域設定部
1082 推定精度設定部
1083 補正パラメータ推定部
1084 画像補正部
200 多視点映像復号装置
201 符号化データ入力部
202 符号化データメモリ
203 参照視点フレーム入力部
204 参照視点画像メモリ
205 視点合成部
206 視点合成画像メモリ
207 信頼度設定部
208 視点合成画像補正部
210 予測残差復号部
211 復号画像メモリ
212 復号画像算出部
Claims (15)
- 多視点映像を符号化する多視点映像符号化方法であって、
前記多視点映像の符号化対象視点における符号化対象フレームと同時刻に撮影された、前記符号化対象視点とは異なる参照視点における既に符号化済みの参照視点フレームから、前記符号化対象視点の前記符号化対象フレームに対応する視点合成画像を合成する視点合成画像生成ステップと、
予め定められた大きさの処理単位領域毎に、前記視点合成画像に対応する、前記符号化対象視点における既に符号化済みの参照フレーム上の参照領域を探索する参照領域推定ステップと、
前記処理単位領域に対する前記視点合成画像と前記参照領域に対する前記参照フレームとから、カメラ間ミスマッチを補正する補正パラメータを推定する補正パラメータ推定ステップと、
前記推定された補正パラメータを用いて、前記処理単位領域に対する前記視点合成画像を補正する視点合成画像補正ステップと、
前記補正された視点合成画像を用いて、前記符号化対象視点の映像を予測符号化する画像符号化ステップと
を含む多視点映像符号化方法。 - 前記視点合成画像の各画素について、前記視点合成画像の確からしさを示す信頼度を設定する信頼度設定ステップを更に含み、
前記参照領域推定ステップは、
前記信頼度に基づいて、前記視点合成画像に対応する、前記参照フレーム上の前記参照領域を探索する際の各画素のマッチングコストに重みをつける
請求項1に記載の多視点映像符号化方法。 - 前記補正パラメータ推定ステップは、
前記信頼度に基づいて、前記補正パラメータを推定する際の各画素のマッチングコストに重みをつける
請求項2に記載の多視点映像符号化方法。 - 前記視点合成画像の各画素について、前記参照領域が正しく推定できたかどうかを示す推定精度を設定する推定精度設定ステップを更に含み、
前記補正パラメータ推定ステップは、
前記推定精度、および前記信頼度のいずれか一方、または双方に基づいて、前記補正パラメータを推定する際の各画素のマッチングコストに重みをつける
請求項2に記載の多視点映像符号化方法。 - 多視点映像を復号する多視点映像復号方法であって、
前記多視点映像の復号対象視点における復号対象フレームと同時刻に撮影された、前記復号対象視点とは異なる参照視点における参照視点フレームから、前記復号対象視点の前記復号対象フレームに対応する視点合成画像を合成する視点合成画像生成ステップと、
予め定められた大きさの処理単位領域毎に、前記視点合成画像に対応する、前記復号対象視点における既に復号済みの参照フレーム上の参照領域を探索する参照領域推定ステップと、
前記処理単位領域に対する前記視点合成画像と前記参照領域に対する前記参照フレームとから、カメラ間ミスマッチを補正する補正パラメータを推定する補正パラメータ推定ステップと、
前記推定された補正パラメータを用いて、前記処理単位領域に対する前記視点合成画像を補正する視点合成画像補正ステップと、
前記補正された視点合成画像を予測信号として用いて、前記復号対象視点における予測符号化されている復号対象フレームを、前記復号対象視点に対する映像の符号化データから復号する画像復号ステップと
を含む多視点映像復号方法。 - 前記視点合成画像の各画素について、前記視点合成画像の確からしさを示す信頼度を設定する信頼度設定ステップを更に含み、
前記参照領域推定ステップは、
前記信頼度に基づいて、前記視点合成画像に対応する、前記参照フレーム上の前記参照領域を探索する際の各画素のマッチングコストに重みをつける
請求項5に記載の多視点映像復号方法。 - 前記補正パラメータ推定ステップは、
前記信頼度に基づいて、前記補正パラメータを推定する際の各画素のマッチングコストに重みをつける
請求項6に記載の多視点映像復号方法。 - 前記視点合成画像の各画素について、前記参照領域が正しく推定できたかどうかを示す推定精度を設定する推定精度設定ステップを更に含み、
前記補正パラメータ推定ステップは、
前記推定精度、および前記信頼度のいずれか一方、または双方に基づいて、前記補正パラメータを推定する際の各画素のマッチングコストに重みをつける
請求項6に記載の多視点映像復号方法。 - 多視点映像を符号化する多視点映像符号化装置であって、
前記多視点映像の符号化対象視点における符号化対象フレームと同時刻に撮影された、前記符号化対象視点とは異なる参照視点における既に符号化済みの参照視点フレームから、前記符号化対象視点の前記符号化対象フレームに対応する視点合成画像を合成する視点合成画像生成手段と、
予め定められた大きさの処理単位領域毎に、前記視点合成画像生成手段により合成された前記視点合成画像に対応する、前記符号化対象視点における既に符号化済みの参照フレーム上の参照領域を探索する参照領域推定手段と、
前記処理単位領域に対する前記視点合成画像と前記参照領域推定手段により探索された前記参照領域に対する前記参照フレームとから、カメラ間ミスマッチを補正する補正パラメータを推定する補正パラメータ推定手段と、
前記補正パラメータ推定手段により推定された前記補正パラメータを用いて、前記処理単位領域に対する前記視点合成画像を補正する視点合成画像補正手段と、
前記視点合成画像補正手段により補正された視点合成画像を用いて、前記符号化対象視点の映像を予測符号化する画像符号化手段と
を備える多視点映像符号化装置。 - 前記視点合成画像生成手段により合成された前記視点合成画像の各画素について、前記視点合成画像の確からしさを示す信頼度を設定する信頼度設定手段を更に備え、
前記参照領域推定手段は、
前記信頼度設定手段により設定された前記信頼度に基づいて、前記視点合成画像に対応する、前記参照フレーム上の前記参照領域を探索する際の各画素のマッチングコストに重みをつける
請求項9に記載の多視点映像符号化装置。 - 前記補正パラメータ推定手段は、
前記信頼度設定手段により設定された前記信頼度に基づいて、前記補正パラメータを推定する際の各画素のマッチングコストに重みをつける
請求項10に記載の多視点映像符号化装置。 - 前記視点合成画像生成手段により合成された前記視点合成画像の各画素について、前記参照領域が正しく推定できたかどうかを示す推定精度を設定する推定精度設定手段を更に備え、
前記補正パラメータ推定手段は、
前記推定精度設定手段により設定された前記推定精度、および前記信頼度設定手段により設定された前記信頼度のいずれか一方、または双方に基づいて、前記補正パラメータを推定する際の各画素のマッチングコストに重みをつける
請求項10に記載の多視点映像符号化装置。 - 多視点映像を復号する多視点映像復号装置であって、
前記多視点映像の復号対象視点における復号対象フレームと同時刻に撮影された、前記復号対象視点とは異なる参照視点における参照視点フレームから、前記復号対象視点の前記復号対象フレームに対応する視点合成画像を合成する視点合成画像生成手段と、
予め定められた大きさの処理単位領域毎に、前記視点合成画像生成手段により合成された前記視点合成画像に対応する、前記復号対象視点における既に復号済みの参照フレーム上の参照領域を探索する参照領域推定手段と、
前記処理単位領域に対する前記視点合成画像と前記参照領域推定手段により探索された前記参照領域に対する前記参照フレームとから、カメラ間ミスマッチを補正する補正パラメータを推定する補正パラメータ推定手段と、
前記補正パラメータ推定手段により推定された前記補正パラメータを用いて、前記処理単位領域に対する前記視点合成画像を補正する視点合成画像補正手段と、
前記視点合成画像補正手段により補正された視点合成画像を予測信号として用いて、前記復号対象視点における予測符号化されている復号対象フレームを、前記復号対象視点に対する映像の符号化データから復号する画像復号手段と
を有する多視点映像復号装置。 - 多視点映像を符号化する多視点映像符号化装置のコンピュータに、
前記多視点映像の符号化対象視点における符号化対象フレームと同時刻に撮影された、前記符号化対象視点とは異なる参照視点における既に符号化済みの参照視点フレームから、前記符号化対象視点の前記符号化対象フレームに対応する視点合成画像を合成する視点合成画像生成機能、
予め定められた大きさの処理単位領域毎に、前記視点合成画像に対応する、前記符号化対象視点における既に符号化済みの参照フレーム上の参照領域を探索する参照領域推定機能、
前記処理単位領域に対する前記視点合成画像と前記参照領域に対する前記参照フレームとから、カメラ間ミスマッチを補正する補正パラメータを推定する補正パラメータ推定機能、
前記推定された補正パラメータを用いて、前記処理単位領域に対する前記視点合成画像を補正する視点合成画像補正機能、
前記補正された視点合成画像を用いて、前記符号化対象視点の映像を予測符号化する画像符号化機能
を実行させるプログラム。 - 多視点映像を復号する多視点映像復号装置のコンピュータに、
前記多視点映像の復号対象視点における復号対象フレームと同時刻に撮影された、前記復号対象視点とは異なる参照視点における参照視点フレームから、前記復号対象視点の前記復号対象フレームに対応する視点合成画像を合成する視点合成画像生成機能、
予め定められた大きさの処理単位領域毎に、前記視点合成画像に対応する、前記復号対象視点における既に復号済みの参照フレーム上の参照領域を探索する参照領域推定機能、
前記処理単位領域に対する前記視点合成画像と前記参照領域に対する前記参照フレームとから、カメラ間ミスマッチを補正する補正パラメータを推定する補正パラメータ推定機能、
前記推定された補正パラメータを用いて、前記処理単位領域に対する前記視点合成画像を補正する視点合成画像補正機能、
前記補正された視点合成画像を予測信号として用いて、前記復号対象視点における予測符号化されている復号対象フレームを、前記復号対象視点に対する映像の符号化データから復号する画像復号機能
を実行させるプログラム。
Priority Applications (8)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CA2790268A CA2790268A1 (en) | 2010-02-24 | 2011-02-21 | Multiview video encoding method, multiview video decoding method, multiview video encoding apparatus, multiview video decoding apparatus, and program |
BR112012020993A BR112012020993A2 (pt) | 2010-02-24 | 2011-02-21 | método de codificação de vídeo de múltiplas visões , método de decodificação de vídeo de múltiplas visões, aparelho de codificação de vídeo de múltiplas visões, aparelho de decodificação de vídeo de múltiplas visões , e programa |
US13/579,675 US20120314776A1 (en) | 2010-02-24 | 2011-02-21 | Multiview video encoding method, multiview video decoding method, multiview video encoding apparatus, multiview video decoding apparatus, and program |
JP2012501773A JP5303754B2 (ja) | 2010-02-24 | 2011-02-21 | 多視点映像符号化方法、多視点映像復号方法、多視点映像符号化装置、多視点映像復号装置、及びプログラム |
KR1020127021725A KR101374812B1 (ko) | 2010-02-24 | 2011-02-21 | 다시점 영상 부호화 방법, 다시점 영상 복호 방법, 다시점 영상 부호화 장치, 다시점 영상 복호 장치 및 프로그램 |
EP11747298A EP2541943A1 (en) | 2010-02-24 | 2011-02-21 | Multiview video coding method, multiview video decoding method, multiview video coding device, multiview video decoding device, and program |
RU2012135682/08A RU2527737C2 (ru) | 2010-02-24 | 2011-02-21 | Способ кодирования многопроекционного видео, способ декодирования многопроекционного видео, устройство кодирования многопроекционного видео, устройство декодирования многопроекционного видео, и программа |
CN201180010368.0A CN102918846B (zh) | 2010-02-24 | 2011-02-21 | 多视点视频编码方法、多视点视频解码方法、多视点视频编码装置、多视点视频解码装置 |
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
JP2010-038680 | 2010-02-24 | ||
JP2010038680 | 2010-02-24 |
Publications (1)
Publication Number | Publication Date |
---|---|
WO2011105337A1 true WO2011105337A1 (ja) | 2011-09-01 |
Family
ID=44506745
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
PCT/JP2011/053742 WO2011105337A1 (ja) | 2010-02-24 | 2011-02-21 | 多視点映像符号化方法、多視点映像復号方法、多視点映像符号化装置、多視点映像復号装置、及びプログラム |
Country Status (10)
Country | Link |
---|---|
US (1) | US20120314776A1 (ja) |
EP (1) | EP2541943A1 (ja) |
JP (1) | JP5303754B2 (ja) |
KR (1) | KR101374812B1 (ja) |
CN (1) | CN102918846B (ja) |
BR (1) | BR112012020993A2 (ja) |
CA (1) | CA2790268A1 (ja) |
RU (1) | RU2527737C2 (ja) |
TW (1) | TWI436637B (ja) |
WO (1) | WO2011105337A1 (ja) |
Cited By (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN102761765A (zh) * | 2012-07-16 | 2012-10-31 | 清华大学 | 一种用于三维立体视频的深度快速插帧方法 |
CN103079083A (zh) * | 2012-12-06 | 2013-05-01 | 上海大学 | 一种已标定平行摄像机阵列多视图像校正方法 |
WO2013087880A1 (en) | 2011-12-14 | 2013-06-20 | Thomson Licensing | Method and system for interpolating a virtual image from a first and a second input images |
CN103379349A (zh) * | 2012-04-25 | 2013-10-30 | 浙江大学 | 一种视点合成预测编码方法、解码方法、对应的装置及码流 |
JP2016513925A (ja) * | 2013-04-09 | 2016-05-16 | 聯發科技股▲ふん▼有限公司Mediatek Inc. | 3dビデオ符号化におけるビュー合成予測の方法と装置 |
Families Citing this family (13)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US7983835B2 (en) | 2004-11-03 | 2011-07-19 | Lagassey Paul J | Modular intelligent transportation system |
US20130329800A1 (en) * | 2012-06-07 | 2013-12-12 | Samsung Electronics Co., Ltd. | Method of performing prediction for multiview video processing |
US9854138B2 (en) * | 2012-09-20 | 2017-12-26 | Gyrus Acmi, Inc. | Fixed pattern noise reduction |
US9615089B2 (en) | 2012-12-26 | 2017-04-04 | Samsung Electronics Co., Ltd. | Method of encoding and decoding multiview video sequence based on adaptive compensation of local illumination mismatch in inter-frame prediction |
JP6027143B2 (ja) | 2012-12-27 | 2016-11-16 | 日本電信電話株式会社 | 画像符号化方法、画像復号方法、画像符号化装置、画像復号装置、画像符号化プログラム、および画像復号プログラム |
WO2014178051A2 (en) * | 2013-04-30 | 2014-11-06 | Mantisvision Ltd. | Selective 3d registration |
CN103402097B (zh) * | 2013-08-15 | 2016-08-10 | 清华大学深圳研究生院 | 一种自由视点视频深度图编码方法及其失真预测方法 |
CN103763567B (zh) * | 2013-12-31 | 2017-01-18 | 华中科技大学 | 一种应用于监控视频隐私保护的压缩域失真漂移补偿方法 |
CN105184780B (zh) * | 2015-08-26 | 2018-06-05 | 京东方科技集团股份有限公司 | 一种立体视觉深度的预测方法和系统 |
CN105430397B (zh) * | 2015-11-20 | 2018-04-17 | 清华大学深圳研究生院 | 一种3d图像体验质量预测方法及装置 |
US10304468B2 (en) | 2017-03-20 | 2019-05-28 | Qualcomm Incorporated | Target sample generation |
US10891960B2 (en) * | 2017-09-11 | 2021-01-12 | Qualcomm Incorproated | Temporal offset estimation |
DE102021200225A1 (de) | 2021-01-12 | 2022-07-14 | Rheinisch-Westfälische Technische Hochschule (RWTH) Aachen, Körperschaft des öffentlichen Rechts | Verfahren zur Wiedergabe eines Videostreams durch einen Client |
Citations (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
WO2008048487A2 (en) * | 2006-10-18 | 2008-04-24 | Thomson Licensing | Local illumination and color compensation without explicit signaling |
WO2009001791A1 (ja) * | 2007-06-25 | 2008-12-31 | Nippon Telegraph And Telephone Corporation | 映像符号化方法及び復号方法、それらの装置、それらのプログラム並びにプログラムを記録した記録媒体 |
JP2009523355A (ja) * | 2006-01-12 | 2009-06-18 | エルジー エレクトロニクス インコーポレイティド | 多視点ビデオの処理 |
JP2010038680A (ja) | 2008-08-04 | 2010-02-18 | Nireco Corp | 超音波式厚み検出装置、及び、超音波式エッジ位置検出装置 |
Family Cites Families (23)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US7085409B2 (en) * | 2000-10-18 | 2006-08-01 | Sarnoff Corporation | Method and apparatus for synthesizing new video and/or still imagery from a collection of real video and/or still imagery |
US20020131500A1 (en) * | 2001-02-01 | 2002-09-19 | Gandhi Bhavan R. | Method for determining a motion vector for a video signal |
US6961055B2 (en) * | 2001-05-09 | 2005-11-01 | Free Radical Design Limited | Methods and apparatus for constructing virtual environments |
US6859494B2 (en) * | 2001-07-27 | 2005-02-22 | General Instrument Corporation | Methods and apparatus for sub-pixel motion estimation |
WO2003032628A1 (en) * | 2001-10-08 | 2003-04-17 | Koninklijke Philips Electronics N.V. | Device and method for motion estimation |
EP1442428A2 (en) * | 2001-10-25 | 2004-08-04 | Koninklijke Philips Electronics N.V. | Method and apparatus for motion estimation |
US7489342B2 (en) * | 2004-12-17 | 2009-02-10 | Mitsubishi Electric Research Laboratories, Inc. | Method and system for managing reference pictures in multiview videos |
CA2524031C (en) * | 2003-05-20 | 2015-07-07 | Interlego Ag | Method and system for manipulating a digital representation of a three-dimensional object |
WO2004114224A1 (ja) * | 2003-06-20 | 2004-12-29 | Nippon Telegraph And Telephone Corporation | 仮想視点画像生成方法及び3次元画像表示方法並びに装置 |
US7728877B2 (en) * | 2004-12-17 | 2010-06-01 | Mitsubishi Electric Research Laboratories, Inc. | Method and system for synthesizing multiview videos |
US7468745B2 (en) * | 2004-12-17 | 2008-12-23 | Mitsubishi Electric Research Laboratories, Inc. | Multiview video decomposition and encoding |
US7671894B2 (en) * | 2004-12-17 | 2010-03-02 | Mitsubishi Electric Research Laboratories, Inc. | Method and system for processing multiview videos for view synthesis using skip and direct modes |
RU2322771C2 (ru) * | 2005-04-25 | 2008-04-20 | Святослав Иванович АРСЕНИЧ | Стереопроекционная система |
JP2007180981A (ja) * | 2005-12-28 | 2007-07-12 | Victor Co Of Japan Ltd | 画像符号化装置、画像符号化方法、及び画像符号化プログラム |
CN101375593A (zh) * | 2006-01-12 | 2009-02-25 | Lg电子株式会社 | 处理多视图视频 |
KR20090099546A (ko) * | 2007-01-04 | 2009-09-22 | 톰슨 라이센싱 | 멀티 뷰 코딩된 비디오에서 높은 레벨의 신택스 레퍼런스 뷰들을 이용하여 비디오 에러 은닉을 행하는 방법 및 장치 |
US8320456B2 (en) * | 2007-01-17 | 2012-11-27 | Lg Electronics Inc. | Method and apparatus for processing a video signal |
KR100801968B1 (ko) * | 2007-02-06 | 2008-02-12 | 광주과학기술원 | 변위를 측정하는 방법, 중간화면 합성방법과 이를 이용한다시점 비디오 인코딩 방법, 디코딩 방법, 및 인코더와디코더 |
EP2061248A1 (en) * | 2007-11-13 | 2009-05-20 | IBBT vzw | Motion estimation and compensation process and device |
US8351685B2 (en) * | 2007-11-16 | 2013-01-08 | Gwangju Institute Of Science And Technology | Device and method for estimating depth map, and method for generating intermediate image and method for encoding multi-view video using the same |
US8639046B2 (en) * | 2009-05-04 | 2014-01-28 | Mamigo Inc | Method and system for scalable multi-user interactive visualization |
US9124874B2 (en) * | 2009-06-05 | 2015-09-01 | Qualcomm Incorporated | Encoding of three-dimensional conversion information with two-dimensional video sequence |
BR112012020856A2 (pt) * | 2010-02-23 | 2019-09-24 | Nippon Telegraph & Telephone | ver anexo. |
-
2011
- 2011-02-21 KR KR1020127021725A patent/KR101374812B1/ko active IP Right Grant
- 2011-02-21 EP EP11747298A patent/EP2541943A1/en not_active Withdrawn
- 2011-02-21 US US13/579,675 patent/US20120314776A1/en not_active Abandoned
- 2011-02-21 JP JP2012501773A patent/JP5303754B2/ja active Active
- 2011-02-21 CN CN201180010368.0A patent/CN102918846B/zh active Active
- 2011-02-21 RU RU2012135682/08A patent/RU2527737C2/ru active
- 2011-02-21 WO PCT/JP2011/053742 patent/WO2011105337A1/ja active Application Filing
- 2011-02-21 BR BR112012020993A patent/BR112012020993A2/pt not_active IP Right Cessation
- 2011-02-21 CA CA2790268A patent/CA2790268A1/en not_active Abandoned
- 2011-02-22 TW TW100105721A patent/TWI436637B/zh active
Patent Citations (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP2009523355A (ja) * | 2006-01-12 | 2009-06-18 | エルジー エレクトロニクス インコーポレイティド | 多視点ビデオの処理 |
WO2008048487A2 (en) * | 2006-10-18 | 2008-04-24 | Thomson Licensing | Local illumination and color compensation without explicit signaling |
WO2009001791A1 (ja) * | 2007-06-25 | 2008-12-31 | Nippon Telegraph And Telephone Corporation | 映像符号化方法及び復号方法、それらの装置、それらのプログラム並びにプログラムを記録した記録媒体 |
JP2010038680A (ja) | 2008-08-04 | 2010-02-18 | Nireco Corp | 超音波式厚み検出装置、及び、超音波式エッジ位置検出装置 |
Non-Patent Citations (8)
Title |
---|
ADVANCED VIDEO CODING FOR GENERIC AUDIOVISUAL SERVICES, March 2009 (2009-03-01) |
J. SUN; N. ZHENG; H. SHUM: "Stereo Matching Using Belief Propagation", IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, vol. 25, no. 7, July 2003 (2003-07-01), pages 787 - 800 |
K. YAMAMOTO; M. KITAHARA; H. KIMATA; T. YENDO; T. FUJII; M. TANIMOTO; S. SHIMIZU; K. KAMIKURA; Y. YASHIMA: "Multiview Video Coding Using View Interpolation and Color Correction", IEEE TRANSACTIONS ON CIRCUITS AND SYSTEM FOR VIDEO TECHNOLOGY, vol. 17, no. 11, November 2007 (2007-11-01), pages 1436 - 1449 |
S. SHIMIZU; H. KIMATA; Y. OHTANI: "Adaptive Appearance Compensated View Synthesis Prediction for Multiview Video Coding", PROCEEDINGS OF ICIP2009, November 2009 (2009-11-01), pages 2949 - 2952 |
S. SHIMIZU; M. KITAHARA, H. KIMATA; K. KAMIKURA; Y. YASHIMA: "View Scalable Multiview Video Coding Using 3-D Warping with Depth Map", IEEE TRANSACTIONS ON CIRCUITS AND SYSTEM FOR VIDEO TECHNOLOGY, vol. 17, no. 11, November 2007 (2007-11-01), pages 1485 - 1495 |
S. SHIMIZU; Y. TONOMURA; H. KIMATA; Y. OHTANI: "Improved View Interpolation Prediction for Side Information in Multiview Distributed Video Coding", PROCEEDINGS OF ICDSC2009, August 2009 (2009-08-01) |
S. YEA; A. VETRO: "View Synthesis Prediction for Rate-Overhead Reduction in FTV", PROCEEDINGS OF 3DTV-CON2008, May 2008 (2008-05-01), pages 145 - 148 |
Y. MORI; N. FUKUSHIMA; T. FUJII; M. TANIMOTO: "View Generation with 3D Warping Using Depth Information for FTV", PROCEEDINGS OF 3DTV-CON2008, May 2008 (2008-05-01), pages 229 - 232 |
Cited By (7)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
WO2013087880A1 (en) | 2011-12-14 | 2013-06-20 | Thomson Licensing | Method and system for interpolating a virtual image from a first and a second input images |
CN103379349A (zh) * | 2012-04-25 | 2013-10-30 | 浙江大学 | 一种视点合成预测编码方法、解码方法、对应的装置及码流 |
CN102761765A (zh) * | 2012-07-16 | 2012-10-31 | 清华大学 | 一种用于三维立体视频的深度快速插帧方法 |
CN102761765B (zh) * | 2012-07-16 | 2014-08-20 | 清华大学 | 一种用于三维立体视频的深度快速插帧方法 |
CN103079083A (zh) * | 2012-12-06 | 2013-05-01 | 上海大学 | 一种已标定平行摄像机阵列多视图像校正方法 |
CN103079083B (zh) * | 2012-12-06 | 2015-05-06 | 上海大学 | 一种已标定平行摄像机阵列多视图像校正方法 |
JP2016513925A (ja) * | 2013-04-09 | 2016-05-16 | 聯發科技股▲ふん▼有限公司Mediatek Inc. | 3dビデオ符号化におけるビュー合成予測の方法と装置 |
Also Published As
Publication number | Publication date |
---|---|
CN102918846A (zh) | 2013-02-06 |
CN102918846B (zh) | 2015-09-09 |
KR101374812B1 (ko) | 2014-03-18 |
TWI436637B (zh) | 2014-05-01 |
JP5303754B2 (ja) | 2013-10-02 |
JPWO2011105337A1 (ja) | 2013-06-20 |
KR20120117888A (ko) | 2012-10-24 |
RU2012135682A (ru) | 2014-03-27 |
TW201218745A (en) | 2012-05-01 |
BR112012020993A2 (pt) | 2016-05-03 |
EP2541943A1 (en) | 2013-01-02 |
US20120314776A1 (en) | 2012-12-13 |
CA2790268A1 (en) | 2011-09-01 |
RU2527737C2 (ru) | 2014-09-10 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
JP5303754B2 (ja) | 多視点映像符号化方法、多視点映像復号方法、多視点映像符号化装置、多視点映像復号装置、及びプログラム | |
JP4414379B2 (ja) | 映像符号化方法、映像復号方法、映像符号化プログラム、映像復号プログラム及びそれらのプログラムを記録したコンピュータ読み取り可能な記録媒体 | |
KR101451286B1 (ko) | 움직임 벡터 추정 방법, 다시점 영상 부호화 방법, 다시점 영상 복호 방법, 움직임 벡터 추정 장치, 다시점 영상 부호화 장치, 다시점 영상 복호 장치, 움직임 벡터 추정 프로그램, 다시점 영상 부호화 프로그램 및 다시점 영상 복호 프로그램 | |
JP5934375B2 (ja) | 画像符号化方法、画像復号方法、画像符号化装置、画像復号装置、画像符号化プログラム、画像復号プログラム及び記録媒体 | |
JP6027143B2 (ja) | 画像符号化方法、画像復号方法、画像符号化装置、画像復号装置、画像符号化プログラム、および画像復号プログラム | |
JP6053200B2 (ja) | 画像符号化方法、画像復号方法、画像符号化装置、画像復号装置、画像符号化プログラム及び画像復号プログラム | |
JP6307152B2 (ja) | 画像符号化装置及び方法、画像復号装置及び方法、及び、それらのプログラム | |
WO2014050827A1 (ja) | 画像符号化方法、画像復号方法、画像符号化装置、画像復号装置、画像符号化プログラム、画像復号プログラム及び記録媒体 | |
JP4851564B2 (ja) | 映像符号化方法、映像復号方法、映像符号化プログラム、映像復号プログラム及びそれらのプログラムを記録したコンピュータ読み取り可能な記録媒体 | |
JP4944046B2 (ja) | 映像符号化方法,復号方法,符号化装置,復号装置,それらのプログラムおよびコンピュータ読み取り可能な記録媒体 | |
JP4851563B2 (ja) | 映像符号化方法、映像復号方法、映像符号化プログラム、映像復号プログラム及びそれらのプログラムを記録したコンピュータ読み取り可能な記録媒体 | |
US20160286212A1 (en) | Video encoding apparatus and method, and video decoding apparatus and method |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
WWE | Wipo information: entry into national phase |
Ref document number: 201180010368.0 Country of ref document: CN |
|
121 | Ep: the epo has been informed by wipo that ep was designated in this application |
Ref document number: 11747298 Country of ref document: EP Kind code of ref document: A1 |
|
DPE1 | Request for preliminary examination filed after expiration of 19th month from priority date (pct application filed from 20040101) | ||
WWE | Wipo information: entry into national phase |
Ref document number: 2012501773 Country of ref document: JP |
|
ENP | Entry into the national phase |
Ref document number: 2790268 Country of ref document: CA |
|
WWE | Wipo information: entry into national phase |
Ref document number: 2011747298 Country of ref document: EP |
|
WWE | Wipo information: entry into national phase |
Ref document number: 13579675 Country of ref document: US Ref document number: 7170/CHENP/2012 Country of ref document: IN |
|
ENP | Entry into the national phase |
Ref document number: 20127021725 Country of ref document: KR Kind code of ref document: A |
|
NENP | Non-entry into the national phase |
Ref country code: DE |
|
WWE | Wipo information: entry into national phase |
Ref document number: 2012135682 Country of ref document: RU |
|
REG | Reference to national code |
Ref country code: BR Ref legal event code: B01A Ref document number: 112012020993 Country of ref document: BR |
|
ENP | Entry into the national phase |
Ref document number: 112012020993 Country of ref document: BR Kind code of ref document: A2 Effective date: 20120821 |