US20130235155A1 - Method of converting 2d into 3d based on image motion information - Google Patents
Method of converting 2d into 3d based on image motion information Download PDFInfo
- Publication number
- US20130235155A1 US20130235155A1 US13/818,101 US201113818101A US2013235155A1 US 20130235155 A1 US20130235155 A1 US 20130235155A1 US 201113818101 A US201113818101 A US 201113818101A US 2013235155 A1 US2013235155 A1 US 2013235155A1
- Authority
- US
- United States
- Prior art keywords
- image
- depth
- pixel
- value
- converting
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Abandoned
Links
- 238000000034 method Methods 0.000 title claims abstract description 39
- 239000013598 vector Substances 0.000 claims description 10
- 229910003460 diamond Inorganic materials 0.000 claims description 5
- 239000010432 diamond Substances 0.000 claims description 5
- 238000010845 search algorithm Methods 0.000 claims description 4
- 230000000007 visual effect Effects 0.000 abstract description 7
- 238000006243 chemical reaction Methods 0.000 abstract description 6
- 238000009825 accumulation Methods 0.000 abstract description 3
- 238000001914 filtration Methods 0.000 description 2
- 239000006002 Pepper Substances 0.000 description 1
- 230000009286 beneficial effect Effects 0.000 description 1
- 230000000694 effects Effects 0.000 description 1
- 238000009877 rendering Methods 0.000 description 1
- 230000009466 transformation Effects 0.000 description 1
- 230000016776 visual perception Effects 0.000 description 1
Images
Classifications
-
- H04N13/0022—
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N13/00—Stereoscopic video systems; Multi-view video systems; Details thereof
- H04N13/10—Processing, recording or transmission of stereoscopic or multi-view image signals
- H04N13/106—Processing image signals
- H04N13/128—Adjusting depth or disparity
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
- G06T7/50—Depth or shape recovery
- G06T7/55—Depth or shape recovery from multiple images
- G06T7/579—Depth or shape recovery from multiple images from motion
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
- G06T7/20—Analysis of motion
- G06T7/223—Analysis of motion using block-matching
- G06T7/238—Analysis of motion using block-matching using non-full search, e.g. three-step search
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N13/00—Stereoscopic video systems; Multi-view video systems; Details thereof
- H04N13/20—Image signal generators
- H04N13/261—Image signal generators with monoscopic-to-stereoscopic image conversion
- H04N13/264—Image signal generators with monoscopic-to-stereoscopic image conversion using the relative movement of objects in two video frames or fields
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/20—Special algorithmic details
- G06T2207/20021—Dividing image into blocks, subimages or windows
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N2213/00—Details of stereoscopic systems
- H04N2213/003—Aspects relating to the "2D+depth" image format
Definitions
- the present application relates to the field of conversion from 2D into 3D, and in particular to a method of converting 2D into 3D based on image motion information.
- 3D (Three Dimensions) TVs have swept the world and become a new trend in the global TV industry. Every major TV manufacturer has launched its own 3D TV. The application of 3D has become more and more popular in people's life. Although 3D films are kept shooting all the time, the 3D resources are still unable to meet the current market needs.
- the conversion from 2D into 3D is to generate the second view video based on 2D view content, and the conversion process comprises two aspects of treatment: one is depth estimation for the purpose of obtaining a depth map/image; the other is Depth Image Based Rendering, DIBR.
- the depth image stores the depth information as grey values in 8 bits (Grey value 0 represents the farthest value, and grey value 255 represents the nearest value).
- the algorithm based on motion estimation is commonly used, which obtains the depth image of the input image by the method of motion estimation.
- the wide application of the said method has been limited, because a depth image requires considerable density and precision, but the depth image achieved by the current algorithm converting 2D into 3D based on the motion estimation are sparse, thus different objects cannot be distinguished at the position where they are decomposed, hence the image quality achieved by means of DIBR and thereby the promotion of the related method have been hindered.
- the technical problems to be solved by the present invention is to improve the image quality generated by the method of converting 2D into 3D based on image motion information.
- a method of converting 2D into 3D based on motion estimation comprising:
- the step of S1 further comprises:
- the depth value is calculated by a formula below:
- the method of motion estimation is the diamond search algorithm.
- the step of S2 further comprises:
- sum ′ sum sidth * height ;
- the step of S2.1 further comprises:
- D ( x,y )′ min( D ( x ⁇ 1 ,y )′+
- D ( x,y )′ min( D ( x ⁇ 1 ,y )′+
- SCALE 0.1.
- DEPTH_SCALE 120.
- the step of S3 further comprises:
- xl and xr are the positions in left eye image and right eye image corresponding to the position xc of the input 2D image respectively; f is the focal length of the eye; tx is the distance between the two eyes; Z is the distance between the pixel point and human eye; Dzero is the position of zero plane with a value interval [0,255];
- Dzero 255.
- the depth image provided in the method described herein is continuous and dense, which improves the quality of the reconstructed image and the 3D visual effect.
- FIG. 1 is a flow chart of the method of converting 2D into 3D based on image motion information according to one embodiment of the present application
- FIG. 2 is a schematic view of the visual model of a dual-camera.
- the method of converting 2D into 3D based on image motion information comprises:
- step of S1 further comprises:
- the depth value is calculated from a formula below:
- step S1.1 To enhance the search precision of step S1.1 and to lessen the influence on the precision of motion search caused by noise (in particular those salt-and-pepper noise added in some video resource), before carrying out the motion search of step S1.1, a de-noising processing can be conducted on the input 2D image.
- This processing is commonly known by those skilled in this art and herein no further details will be given thereto.
- the present application conducts an accumulation of the depth values obtained by computing the motion vector according to the luminous information of each pixel.
- step of S2 further comprises:
- D ( x,y )′ min( D ( x ⁇ 1 ,y )′+
- D ( x,y )′ min( D ( x ⁇ 1 ,y )′+
- I (x,y) is the luminance value of the pixel at the position (x,y) with a value interval [0, 255];
- width is the width value of the input 2D image;
- height is the height value of the input 2D image;
- sum ′ sum sidth * height ( 7 )
- the depth values should keep continuous as far as possible in the horizontal direction to avoid the influence of excessive noise caused by the motion search. Therefore, the present application does not apply the horizontal gradient value to the scale motion for achieving the depth value.
- the visual perception of 70% people relies heavily on the right eye, and 20% on the left eye.
- the present invention only reconstructs the eye on which is not heavily relied, herein defaulting to the left eye.
- the quality of a reconstructed frame in this case is poor, it does not affect the 3D visual effect. Consequently, the step of S3 in this embodiment takes the left eye image as an example, namely, in the step of S3, the left eye image is reconstructed based on DIBR according to the depth image obtained in the step of S2.
- Cc is the input 2D image
- Cl is the reconstructed left eye image
- Cr is the reconstructed right eye image
- f is the focus length of the eye
- tx is the baseline distance, i.e., the distance between the two eyes
- Z is the distance between the observed pixel point and the human eye, which is computed in accordance with the formula (11)
- Dzero is the position of zero plane with a value interval [0,255], in this embodiment a value of 255 is taken.
- Formula (9), (10) are projection geometrical relationship in FIG. 2 corresponding to the same pixel in Cl, Cr and Cc.
- the value of xl or xr corresponding to the position xc of the input 2D image is computed, and then the pixel value at the position (xc, y) is copied to the corresponding position (xl, y) or (xr, y). (copied to (xl, y) in this embodiment).
- step of S3 further comprises:
- xl and xr are the positions in left eye image and right eye image corresponding to the position xc of the input 2D image respectively;
- f is the focal length of the eye;
- tx is the distance between the two eyes;
- Z is the distance between the pixel point and the human eye;
- Dzero is the position of zero plane with a value interval [0,255];
- the input 2D image is scaled in the horizontal direction firstly, in order to enhance the pixel precision at the time of projection.
- the image is stretched in the horizontal direction to be four times of its original size.
- the value x of 1 ⁇ 4 pixel precision to which every xl in each row corresponds is computed.
- the pixel value at the position xl is obtained based on interpolation; if there are multiple xl corresponding to the same x, then take the xl which makes D(x,y)′′ largest, then the pixel values of other xl are obtained based on interpolation; if there is an exclusive x to which xl corresponds, then the pixel value at the position xl is the pixel value at the position x in the input 2D image.
- the reconstructed images obtained by the method of converting 2D into 3D based on image motion information described herein have high image quality, excellent 3D visual effect, and hence the present method is of great importance for the market development in impelling the automatic conversion from 2D resource into 3D.
Landscapes
- Engineering & Computer Science (AREA)
- Multimedia (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Physics & Mathematics (AREA)
- General Physics & Mathematics (AREA)
- Theoretical Computer Science (AREA)
- Signal Processing (AREA)
- Testing, Inspecting, Measuring Of Stereoscopic Televisions And Televisions (AREA)
- Processing Or Creating Images (AREA)
Abstract
The present invention relates to the field of 2D to 3D conversion, in particular discloses a method of converting 2D into 3D based on image motion information. The method comprises: S1, obtaining a depth value of each pixel of the input 2D image based on a method of motion estimation; S2, accumulating the depth value of each pixel in accordance with a luminance value of each pixel to obtain a depth image of the input 2D image; S3, reconstructing a left eye and/or a right eye image based on a reconstruction of depth image in accordance with the depth image obtained in the step of S2; S4, combining the left eye image and the right eye image obtained in the step of S4 and outputting a combined image to obtain the 3D image. In the method herein, due to the accumulation process of the depth value obtained by the motion estimation, the resulted depth image is continuous and dense, which improves the quality of the reconstructed image and the 3D visual effect.
Description
- The present application relates to the field of conversion from 2D into 3D, and in particular to a method of converting 2D into 3D based on image motion information.
- 3D (Three Dimensions) TVs have swept the world and become a new trend in the global TV industry. Every major TV manufacturer has launched its own 3D TV. The application of 3D has become more and more popular in people's life. Although 3D films are kept shooting all the time, the 3D resources are still unable to meet the current market needs. A new market desire to convert the resources of 2D (Two Dimensions) into that of 3D automatically has been created. The conversion from 2D into 3D is to generate the second view video based on 2D view content, and the conversion process comprises two aspects of treatment: one is depth estimation for the purpose of obtaining a depth map/image; the other is Depth Image Based Rendering, DIBR. The depth image stores the depth information as grey values in 8 bits (Grey value 0 represents the farthest value, and grey value 255 represents the nearest value). In the past few years, there have been numerous algorithms proposed in the field of 2D to 3D conversion. The algorithm based on motion estimation is commonly used, which obtains the depth image of the input image by the method of motion estimation. However, the wide application of the said method has been limited, because a depth image requires considerable density and precision, but the depth image achieved by the current algorithm converting 2D into 3D based on the motion estimation are sparse, thus different objects cannot be distinguished at the position where they are decomposed, hence the image quality achieved by means of DIBR and thereby the promotion of the related method have been hindered.
- The technical problems to be solved by the present invention is to improve the image quality generated by the method of converting 2D into 3D based on image motion information.
- To solve the aforementioned problem, a method of converting 2D into 3D based on motion estimation is provided, comprising:
- S1, obtaining a depth value of each pixel of the
input 2D image based on a method of motion estimation; - S2, accumulating the depth value of each pixel in accordance with a luminance value of each pixel to obtain a depth image of the
input 2D image; - S3, reconstructing a left eye image and/or a right eye image based on a reconstruction of depth image in accordance with the depth image obtained in the step of S2;
- S4, combining the left eye image and the right eye image obtained in the step of S3 and outputting a combined image to obtain a 3D image;
- Preferably, the step of S1 further comprises:
- S1.1, computing a motion vector of each pixel based on the method of motion estimation;
- S1.2, computing the depth value of each pixel respectively according to the motion vector obtained in the step of S1.1.
- Preferably, the depth value is calculated by a formula below:
-
D(x,y)=C*√{square root over (MV x 2 +MV y 2)}; - Preferably, the method of motion estimation is the diamond search algorithm.
- Preferably, the step of S2 further comprises:
- S2.1, accumulating the depth value of each pixel beginning from the first row of the
input 2D image to obtain an accumulated depth value D(x, y) of each pixel; - S2.2, obtaining a normalized depth value D(x,y)″ by normalizing the accumulated depth value to an interval [0, 255] according to the formula below:
-
- wherein, I (x,y) is the luminance value of the pixel at the position (x,y) with a value interval [0, 255]; SCALE is the scaling factor of the luminance value; width is the width value of the
input 2D image, height is the height value of theinput 2D image; DEPTH_SCALE is the scaling factor of the depth value; -
- Preferably, the step of S2.1 further comprises:
- S2.11, if y is zero, then D(x,y)′=0, otherwise, carrying out the step of S2.12;
- S2.12, if y is an odd number and x is zero, then D(x,y)′=D(x,y−1)′+D(x,y);
- if x is not zero, then
-
D(x,y)′=min(D(x−1,y)′+|I(x+1,y)−I(x−1,y)|*SCALE,D(x,y−1)′)+D(x,y)*(1+|I(x,y−1)−I(x,y+1)|*SCALE); - otherwise, carrying out the step of S2.13;
- S2.13, if x=width−1, then D(x,y)′=D(x,y−1)′+D(x,y); otherwise,
-
D(x,y)′=min(D(x−1,y)′+|I(x+1,y)−I(x−1,y)|*SCALE,D(x,y−1)′)+D(x,y)*(1+|I(x,y−1)−I(x,y+1)|*SCALE); - S2.14, if y<height, then returning to the step of S2.11,
- Otherwise, outputting the result D(x,y)′ of the step of S2.12 or S2.13.
- Preferably, SCALE=0.1.
- Preferably, DEPTH_SCALE=120.
- Preferably, the step of S3 further comprises:
- S3.1, reconstructing the left eye or right eye image according to the formula below:
-
- wherein, xl and xr are the positions in left eye image and right eye image corresponding to the position xc of the
input 2D image respectively; f is the focal length of the eye; tx is the distance between the two eyes; Z is the distance between the pixel point and human eye; Dzero is the position of zero plane with a value interval [0,255]; - S3.2, copying the pixel value at the position (xc,y) to the corresponding position (xl,y) or (xr,y);
- Preferably, Dzero=255.
- Due to the accumulation process of the depth value obtained by the motion estimation, the depth image provided in the method described herein is continuous and dense, which improves the quality of the reconstructed image and the 3D visual effect.
-
FIG. 1 is a flow chart of the method of converting 2D into 3D based on image motion information according to one embodiment of the present application; -
FIG. 2 is a schematic view of the visual model of a dual-camera. - Hereinafter the method of converting 2D into 3D based on image motion information provided by the present invention will be described in detail with reference to the accompanying drawings and embodiments.
- As shown in FIG. 1/2, the method of converting 2D into 3D based on image motion information according to one embodiment of the present application comprises:
- S1, obtaining a depth value of each pixel of the
input 2D image based on a method of motion estimation; - S2, accumulating the depth value of each pixel in accordance with a luminance value of each pixel to obtain a depth image of the
input 2D image; - S3, reconstructing a left eye and/or a right eye image based on a reconstruction of depth image in accordance with the depth image obtained in the step of S2;
- S4, combining the left eye image and the right eye image obtained in the step of S3 and outputting a combined image to obtain the 3D image.
- In the method of this embodiment, the step of S1 further comprises:
- S1.1, computing a motion vector of each pixel based on the method of motion estimation, wherein, the method of motion estimation adopts the diamond search algorithm. It begins with big diamond search which is followed by small diamond search, and ends with the resulted motion vector with integral pixel precision. Certainly, other search algorithms are also applicable, without limiting the method described herein.
- S1.2, computing the depth value of each pixel respectively according to the motion vector obtained in the step of S1.1.
- wherein, the depth value is calculated from a formula below:
-
D(x,y)=C*√{square root over (MV x 2 +MV y 2)} (1) - y is the row where the pixel locates; x is the column where the pixel locates; D (x,y) is the depth value of the pixel at an unknown position (x,y); MVx and MVy are motion vectors in the horizontal direction and vertical direction of the pixel, respectively; C is a constant, in this embodiment C=1.
- To enhance the search precision of step S1.1 and to lessen the influence on the precision of motion search caused by noise (in particular those salt-and-pepper noise added in some video resource), before carrying out the motion search of step S1.1, a de-noising processing can be conducted on the
input 2D image. This processing is commonly known by those skilled in this art and herein no further details will be given thereto. - Since the motion vector obtained by the motion search is discontinuous, the depth image obtained by direct computation is quite sparse, while the actual depth image should be dense. Therefore, the present application conducts an accumulation of the depth values obtained by computing the motion vector according to the luminous information of each pixel.
- In this embodiment, the step of S2 further comprises:
- S2.1, accumulating the depth value of each pixel beginning from the first row of the
input 2D image to obtain an accumulated depth value D(x,y)′ of each pixel, further comprising: - S2.11, if y is zero, then D(x,y)′=0, otherwise, carrying out the step of S2.12;
- S2.12, if y is an odd number and x is zero, then D(x,y)′=D(x,y−1)′=D(x,y), if x is not zero, then
-
D(x,y)′=min(D(x−1,y)′+|I(x+1,y)−I(x−1,y)|*SCALE,D(x,y−1)′)+D(x,y)*(1+|I(x,y−1)−I(x,y+1)|*SCALE); - otherwise, carrying out the step of S2.13;
- S2.13, if x-width-1, then D(x,y)′=D(x,y−1)′+D(x,y), otherwise,
-
D(x,y)′=min(D(x−1,y)′+|I(x+1,y)−I(x−1,y)|*SCALE,D(x,y−1)′)+D(x,y)*(1+|I(x,y−1)−I(x,y+1)|*SCALE) - S2.14, if y<height, then returning to the step of S2.11, otherwise outputting the result D(x,y)′ of the step of S2.12 or S2.13;
- S2.2, obtaining a normalized depth value D(x,y)″ and hence obtaining a continuous and dense depth image by normalizing the accumulated depth value to an interval [0, 255] according to the formula below:
-
- wherein, I (x,y) is the luminance value of the pixel at the position (x,y) with a value interval [0, 255]; SCALE is the scaling factor of the luminance value, in this embodiment SCALE=0.1; width is the width value of the
input 2D image; height is the height value of theinput 2D image; DEPTH_SCALE is the scaling factor of the depth value, in this embodiment, DEPTH_SCALE=120; -
- S2.3, conducting an asymmetric Gaussian filtering on the normalized depth value D(x,y)″ obtained in the step of S2.2 to obtain an ultimate depth value Dz(x,y)′. The asymmetric Gaussian filtering is commonly known by those skilled in this art and herein no further details will be given thereto.
- As a projection transformation will be conducted in the horizontal direction of the image, the depth values should keep continuous as far as possible in the horizontal direction to avoid the influence of excessive noise caused by the motion search. Therefore, the present application does not apply the horizontal gradient value to the scale motion for achieving the depth value.
- Due to the human visual property, the visual perception of 70% people relies heavily on the right eye, and 20% on the left eye. To reduce the computation amount, when using DIBR to reconstruct image, the present invention only reconstructs the eye on which is not heavily relied, herein defaulting to the left eye. Moreover, although the quality of a reconstructed frame in this case is poor, it does not affect the 3D visual effect. Consequently, the step of S3 in this embodiment takes the left eye image as an example, namely, in the step of S3, the left eye image is reconstructed based on DIBR according to the depth image obtained in the step of S2.
- As shown in FIG. 2/2, wherein Cc is the
input 2D image; Cl is the reconstructed left eye image; Cr is the reconstructed right eye image; f is the focus length of the eye; tx is the baseline distance, i.e., the distance between the two eyes; Z is the distance between the observed pixel point and the human eye, which is computed in accordance with the formula (11); Dzero is the position of zero plane with a value interval [0,255], in this embodiment a value of 255 is taken. Formula (9), (10) are projection geometrical relationship inFIG. 2 corresponding to the same pixel in Cl, Cr and Cc. According to the formula (9), (10), the value of xl or xr corresponding to the position xc of theinput 2D image is computed, and then the pixel value at the position (xc, y) is copied to the corresponding position (xl, y) or (xr, y). (copied to (xl, y) in this embodiment). - Namely the step of S3 further comprises:
- S3.1, reconstructing the left eye or right eye image according to the formula below:
-
- wherein, xl and xr are the positions in left eye image and right eye image corresponding to the position xc of the
input 2D image respectively; f is the focal length of the eye; tx is the distance between the two eyes; Z is the distance between the pixel point and the human eye; Dzero is the position of zero plane with a value interval [0,255]; - S3.2, copying the pixel value at the position (xc,y) to the corresponding position (xl,y) or (xr,y).
- To lessen the zigzagging effect of the reconstructed image, the
input 2D image is scaled in the horizontal direction firstly, in order to enhance the pixel precision at the time of projection. In this embodiment, the image is stretched in the horizontal direction to be four times of its original size. In line with the aforementioned visual relation of human eye, the value x of ¼ pixel precision to which every xl in each row corresponds is computed. If the value x to which xl corresponds exceeds the boundary of the image, then the pixel value at the position xl is obtained based on interpolation; if there are multiple xl corresponding to the same x, then take the xl which makes D(x,y)″ largest, then the pixel values of other xl are obtained based on interpolation; if there is an exclusive x to which xl corresponds, then the pixel value at the position xl is the pixel value at the position x in theinput 2D image. - The aforementioned embodiments of the present invention are disclosed for illustrative purpose only but not limiting the scope thereof. Those skilled in the art will appreciate that various changes and variants can be made thereto without departing from the scope and spirit of the invention. Therefore all equivalent technical solutions also fall within the scope of the present invention.which should be defined by the appended claims.
- The reconstructed images obtained by the method of converting 2D into 3D based on image motion information described herein have high image quality, excellent 3D visual effect, and hence the present method is of great importance for the market development in impelling the automatic conversion from 2D resource into 3D.
Claims (9)
1. A method of converting 2D into 3D based on image motion information, characterized in that, the method comprises the following steps:
S1, obtaining a depth value of each pixel of the input 2D image based on a method of motion estimation;
S2, accumulating the depth value of each pixel in accordance with a luminance value of each pixel to obtain a depth image of the input 2D image;
S3, reconstructing a left eye image and/or a right eye image based on a image reconstruction of depth image in accordance with the depth image obtained in the step of S2;
S4, combining the left eye image and the right eye image obtained in the step of S4, and outputting a combined image to obtain the 3D image.
2. The method of converting 2D into 3D based on image motion information of claim 1 , characterized in that, the step of S1 further comprises:
S1.1, computing a motion vector of each pixel based on the method of motion estimation;
S1.2, computing the depth value of each pixel respectively according to the motion vector obtained in the step of S1.1.
3. The method of converting 2D into 3D based on image motion information of claim 2 , characterized in that, the method of motion estimation is the diamond search algorithm.
4. The method of converting 2D into 3D based on image motion information of claim 3 , characterized in that, the step of S2 further comprises:
S2.1, accumulating the depth value of each pixel beginning from the first row of the input 2D image to obtain an accumulated depth value D(x,y)′ of each pixel;
S2.2, obtaining a normalized depth value D(x,y)″ by normalizing the accumulated depth value to an interval [0, 255] according to the formula below:
wherein, I (x, y) is the luminance value of the pixel at the position (x, y) with a value interval [0, 255]; SCALE is the scaling factor of the luminance value;
width is the width value of the input 2D image; height is the height value of the input 2D image; DEPTH_SCALE is the scaling factor of the depth value;
5. The method of converting 2D into 3D based on image motion information of claim 4 , characterized in that, the step of S2.1 further comprises:
S2.11, if y is zero, then D(x,y)′=0, otherwise, carrying out the step of S2.12;
S2.12, if y is an odd number and x is zero, then D(x,y)′=D(x,y−1)′+D(x,y);
if x is not zero, then
D(x,y)′=min(D(x−1,y)′+|I(x+1,y)−I(x−1,y)|*SCALE,D(x,y−1)′)+D(x,y)*(1+|I(x,y−1)−I(x,y+1)|*SCALE);
D(x,y)′=min(D(x−1,y)′+|I(x+1,y)−I(x−1,y)|*SCALE,D(x,y−1)′)+D(x,y)*(1+|I(x,y−1)−I(x,y+1)|*SCALE);
otherwise, carrying out the step of S2.13;
S2.13, if x=width−1, then D(x,y)′=D(x,y−1)′+D(x,y); otherwise,
D(x,y)′=min(D(x−1,y)′+|I(x+1,y)−I(x−1,y)|*SCALE,D(x,y−1)′)+D(x,y)*(1+|I(x,y−1)−I(x,y+1)|*SCALE);
D(x,y)′=min(D(x−1,y)′+|I(x+1,y)−I(x−1,y)|*SCALE,D(x,y−1)′)+D(x,y)*(1+|I(x,y−1)−I(x,y+1)|*SCALE);
S2.14, if y<height, then returning to the step of S2.11, otherwise outputting the result D(x,y)′ of the step of S2.12 or S2.13.
6. The method of converting 2D into 3D based on image motion information of claim 5 , characterized in that, SCALE=0.1.
7. The method of converting 2D into 3D based on image motion information of claim 5 , characterized in that, DEPTH_SCALE=120.
8. The method of converting 2D into 3D based on image motion information of claim 5 , characterized in that, the step of S3 further comprises:
S3.1, reconstructing the left eye or right eye image according to the formula below:
wherein, xl and xr are the positions in left eye image and right eye image corresponding to the position xc of the input 2D image respectively; f is the focal length of the eye; tx is the distance between the two eyes; Z is the distance between the pixel point and human eye; Dzero is the position of zero plane with a value interval [0,255];
S3.2, copying the pixel value at the position (xc, y) to the corresponding position (xl, y) or (xr, y).
9. The method of converting 2D into 3D based on image motion information of claim 8 , characterized in that, Dzero=255.
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
PCT/CN2011/001377 WO2013023325A1 (en) | 2011-08-18 | 2011-08-18 | Method for converting 2d into 3d based on image motion information |
Publications (1)
Publication Number | Publication Date |
---|---|
US20130235155A1 true US20130235155A1 (en) | 2013-09-12 |
Family
ID=47714669
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US13/818,101 Abandoned US20130235155A1 (en) | 2011-08-18 | 2011-08-18 | Method of converting 2d into 3d based on image motion information |
Country Status (5)
Country | Link |
---|---|
US (1) | US20130235155A1 (en) |
EP (1) | EP2629531A4 (en) |
JP (1) | JP2014504468A (en) |
CN (1) | CN103053165B (en) |
WO (1) | WO2013023325A1 (en) |
Cited By (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20130076858A1 (en) * | 2011-09-26 | 2013-03-28 | Samsung Electronics Co., Ltd. | Method and apparatus for converting 2d content into 3d content |
US20140363100A1 (en) * | 2011-02-28 | 2014-12-11 | Sony Corporation | Method and apparatus for real-time conversion of 2-dimensional content to 3-dimensional content |
US20220286658A1 (en) * | 2021-03-03 | 2022-09-08 | Acer Incorporated | Stereo image generation method and electronic apparatus using the same |
Families Citing this family (7)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN104113745A (en) | 2013-04-17 | 2014-10-22 | 咏传电子科技(上海)有限公司 | Display device and image display method thereof |
JP5858254B2 (en) * | 2013-06-06 | 2016-02-10 | ソニー株式会社 | Method and apparatus for real-time conversion of 2D content to 3D content |
CN103533329B (en) * | 2013-10-09 | 2016-04-27 | 上海大学 | A kind of 2D turns the video automatic evaluation method of 3D |
CN103826032B (en) * | 2013-11-05 | 2017-03-15 | 四川长虹电器股份有限公司 | Depth map post-processing method |
CN105989326B (en) * | 2015-01-29 | 2020-03-03 | 北京三星通信技术研究有限公司 | Method and device for determining three-dimensional position information of human eyes |
CN109274951B (en) * | 2017-07-13 | 2020-11-10 | 富泰华工业(深圳)有限公司 | Depth calculation method and device |
CN111369612B (en) * | 2018-12-25 | 2023-11-24 | 北京欣奕华科技有限公司 | Three-dimensional point cloud image generation method and device |
Citations (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20100104219A1 (en) * | 2008-06-24 | 2010-04-29 | Samsung Electronics Co., Ltd. | Image processing method and apparatus |
US20110001883A1 (en) * | 2009-07-01 | 2011-01-06 | Mstar Semiconductor, Inc. | Motion Estimation Method and Apparatus Thereof |
Family Cites Families (14)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JPH08331607A (en) * | 1995-03-29 | 1996-12-13 | Sanyo Electric Co Ltd | Three-dimensional display image generating method |
JP2001016609A (en) * | 1999-06-05 | 2001-01-19 | Soft Foo Deii:Kk | Stereoscopic video image generator and its method using mpeg data |
JP2001103513A (en) * | 1999-09-27 | 2001-04-13 | Sanyo Electric Co Ltd | Method for converting two-dimensional video image into three-dimensional video image |
JP4898459B2 (en) * | 2004-02-17 | 2012-03-14 | コーニンクレッカ フィリップス エレクトロニクス エヌ ヴィ | Depth map generation |
JP2007300169A (en) * | 2006-04-27 | 2007-11-15 | Toshiba Corp | Motion vector detector |
CN101720480B (en) * | 2007-07-03 | 2012-07-18 | 皇家飞利浦电子股份有限公司 | Computing a depth map |
CN101271578B (en) * | 2008-04-10 | 2010-06-02 | 清华大学 | Depth sequence generation method of technology for converting plane video into stereo video |
KR101468267B1 (en) * | 2008-10-02 | 2014-12-15 | 프라운호퍼-게젤샤프트 추르 푀르데룽 데어 안제반텐 포르슝 에 파우 | Intermediate view synthesis and multi-view data signal extraction |
KR20100040236A (en) * | 2008-10-09 | 2010-04-19 | 삼성전자주식회사 | Two dimensional image to three dimensional image converter and conversion method using visual attention analysis |
JP5428454B2 (en) * | 2009-03-30 | 2014-02-26 | 凸版印刷株式会社 | Image generation method |
KR20100135032A (en) * | 2009-06-16 | 2010-12-24 | 삼성전자주식회사 | Conversion device for two dimensional image to three dimensional image and method thereof |
CN101631256B (en) * | 2009-08-13 | 2011-02-09 | 浙江大学 | Method for converting 2D video into 3D video in three-dimensional television system |
US8610758B2 (en) * | 2009-12-15 | 2013-12-17 | Himax Technologies Limited | Depth map generation for a video conversion system |
CN102075780B (en) * | 2011-02-25 | 2014-02-26 | 福建华映显示科技有限公司 | Stereoscopic image generating device and method |
-
2011
- 2011-08-18 US US13/818,101 patent/US20130235155A1/en not_active Abandoned
- 2011-08-18 EP EP11870997.1A patent/EP2629531A4/en not_active Withdrawn
- 2011-08-18 WO PCT/CN2011/001377 patent/WO2013023325A1/en active Application Filing
- 2011-08-18 JP JP2013540213A patent/JP2014504468A/en active Pending
- 2011-08-18 CN CN201180028889.9A patent/CN103053165B/en active Active
Patent Citations (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20100104219A1 (en) * | 2008-06-24 | 2010-04-29 | Samsung Electronics Co., Ltd. | Image processing method and apparatus |
US20110001883A1 (en) * | 2009-07-01 | 2011-01-06 | Mstar Semiconductor, Inc. | Motion Estimation Method and Apparatus Thereof |
Cited By (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20140363100A1 (en) * | 2011-02-28 | 2014-12-11 | Sony Corporation | Method and apparatus for real-time conversion of 2-dimensional content to 3-dimensional content |
US9483836B2 (en) * | 2011-02-28 | 2016-11-01 | Sony Corporation | Method and apparatus for real-time conversion of 2-dimensional content to 3-dimensional content |
US20130076858A1 (en) * | 2011-09-26 | 2013-03-28 | Samsung Electronics Co., Ltd. | Method and apparatus for converting 2d content into 3d content |
US9154772B2 (en) * | 2011-09-26 | 2015-10-06 | Samsung Electronics Co., Ltd. | Method and apparatus for converting 2D content into 3D content |
US20220286658A1 (en) * | 2021-03-03 | 2022-09-08 | Acer Incorporated | Stereo image generation method and electronic apparatus using the same |
TWI784428B (en) * | 2021-03-03 | 2022-11-21 | 宏碁股份有限公司 | Stereo image generation method and electronic apparatus using the same |
Also Published As
Publication number | Publication date |
---|---|
EP2629531A4 (en) | 2015-01-21 |
CN103053165A (en) | 2013-04-17 |
CN103053165B (en) | 2015-02-11 |
JP2014504468A (en) | 2014-02-20 |
EP2629531A1 (en) | 2013-08-21 |
WO2013023325A1 (en) | 2013-02-21 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US20130235155A1 (en) | Method of converting 2d into 3d based on image motion information | |
US9547887B2 (en) | Visual-experience-optimized super-resolution frame generator | |
JP5583850B2 (en) | Vision-based quality metrics for 3D video | |
US9324191B2 (en) | Method and arrangement for image model construction | |
US10565691B2 (en) | Method of multi-view deblurring for 3D shape reconstruction, recording medium and device for performing the method | |
CN103402098B (en) | A kind of video frame interpolation method based on image interpolation | |
CN109963048B (en) | Noise reduction method, noise reduction device and noise reduction circuit system | |
CN109978774B (en) | Denoising fusion method and device for multi-frame continuous equal exposure images | |
CN103400376B (en) | A kind of method for registering of mammary gland dynamic contrast-enhanced magnetic resonance image sequence | |
US20180089806A1 (en) | Robust regression method for image-space denoising | |
US20120141045A1 (en) | Method and apparatus for reducing block artifacts during image processing | |
Tan et al. | Multipoint filtering with local polynomial approximation and range guidance | |
Choi et al. | 2D-plus-depth based resolution and frame-rate up-conversion technique for depth video | |
US8718402B2 (en) | Depth generation method and apparatus using the same | |
TW201203172A (en) | Depth map enhancing method and computer-readable medium therefor | |
US8995755B2 (en) | Two-dimensional to stereoscopic conversion systems and methods | |
CN105488760A (en) | Virtual image stitching method based on flow field | |
WO2023160426A1 (en) | Video frame interpolation method and apparatus, training method and apparatus, and electronic device | |
CN105282400A (en) | An efficient video stabilization method based on geometric interpolation | |
CN111369435B (en) | Color image depth up-sampling method and system based on self-adaptive stable model | |
CN111405264B (en) | 3D video comfort level improving method based on depth adjustment | |
US8976175B2 (en) | Depth estimation data generating device, computer readable recording medium having depth estimation data generating program recorded thereon, and pseudo-stereo image display device | |
TWI410141B (en) | Image processing method | |
Gsaxner et al. | DeepDR: Deep Structure-Aware RGB-D Inpainting for Diminished Reality | |
Kim et al. | Reconstruction of stereoscopic imagery for visual comfort |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
AS | Assignment |
Owner name: BEIJING GOLAND TECH CO., LTD., CHINA Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:FENG, TAO;ZHANG, YANDING;YANG, DONG;REEL/FRAME:029851/0302 Effective date: 20130204 |
|
STCB | Information on status: application discontinuation |
Free format text: ABANDONED -- FAILURE TO RESPOND TO AN OFFICE ACTION |