WO2011017308A1 - Systèmes et procédés de génération de vidéo tridimensionnelle - Google Patents

Systèmes et procédés de génération de vidéo tridimensionnelle Download PDF

Info

Publication number
WO2011017308A1
WO2011017308A1 PCT/US2010/044227 US2010044227W WO2011017308A1 WO 2011017308 A1 WO2011017308 A1 WO 2011017308A1 US 2010044227 W US2010044227 W US 2010044227W WO 2011017308 A1 WO2011017308 A1 WO 2011017308A1
Authority
WO
WIPO (PCT)
Prior art keywords
training image
video
rules
internal data
generating
Prior art date
Application number
PCT/US2010/044227
Other languages
English (en)
Inventor
Haohong Wang
Glenn Adler
Original Assignee
Shenzhen Tcl New Technology Ltd.
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Shenzhen Tcl New Technology Ltd. filed Critical Shenzhen Tcl New Technology Ltd.
Priority to EP10807016.0A priority Critical patent/EP2462536A4/fr
Publication of WO2011017308A1 publication Critical patent/WO2011017308A1/fr

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/50Depth or shape recovery
    • G06T7/55Depth or shape recovery from multiple images
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N13/00Stereoscopic video systems; Multi-view video systems; Details thereof
    • H04N13/20Image signal generators
    • H04N13/261Image signal generators with monoscopic-to-stereoscopic image conversion
    • H04N13/264Image signal generators with monoscopic-to-stereoscopic image conversion using the relative movement of objects in two video frames or fields

Definitions

  • This disclosure relates to systems and methods for three- dimensional video generation.
  • 3D TV has recently been foreseen as part of a next wave of promising technologies for consumer electronics.
  • 3D technologies incorporate a third dimension of depth into an image, which may provide a stereographic perception to a viewer of the image.
  • a system for generating three-dimensional (3D) video based on a two-dimensional (2D) input image sequence including at least one 2D input image comprising: a rule generator configured to generate rules for 2D-to-3D conversion; and a 3D video converter coupled to the rule generator, and configured to obtain the rules from the rule generator and automatically convert the 2D input image sequence to the 3D video based on the obtained rules.
  • a computer-implemented method for generating three-dimensional (3D) video based on a two-dimensional (2D) input image sequence including at least one 2D input image comprising: generating rules for 2D-to-3D conversion; and automatically converting the 2D input image sequence to the 3D video based on the rules.
  • a computer-readable medium including instructions, executable by a processor of a three-dimensional (3D) video generating system, for performing a method for generating 3D video based on a two-dimensional (2D) input image sequence including at least one 2D input image, the method comprising:
  • FIG. 1 illustrates a block diagram of a system for generating 3D video, according to an exemplary embodiment.
  • Fig. 2 illustrates a block diagram of a rule generator, according to an exemplary embodiment.
  • Fig. 3 illustrates a block diagram of a 3D video convertor, according to an exemplary embodiment.
  • Fig. 4 illustrates a block diagram of a 3D video generator
  • Fig. 5 illustrates a block diagram of a 3D video generator
  • FIG. 6 illustrates a flowchart of a method for generating 3D video based on a 2D input image sequence, according to an exemplary embodiment.
  • Fig. 1 illustrates a block diagram of a system 100 for generating a three-dimensional (3D) video, according to an exemplary embodiment.
  • the system 100 may include a two-dimensional (2D) video content source, such as a video storage medium 102 or a media server 104 connected with a network 106.
  • the system 100 also includes a video device 108, a 3D video generator 110, and a display device 112.
  • the video storage medium 102 may be any medium for storing video content.
  • the video storage medium 102 may be provided as a compact disc (CD), a digital video disc (DVD), a hard disk, a magnetic tape, a flash memory card/drive, a volatile or non-volatile memory, a holographic data storage, or any other storage medium.
  • the video storage medium 102 may be located within the video device 108, local to the video device 108, or remote from the video device 108.
  • the media server 104 may be a computer server that receives a request for 2D video content from the video device 108, processes the request, and provides 2D video content to the video device 108 through the network 106.
  • the media server 104 may be a web server, an enterprise server, or any other type of computer server.
  • the media server 104 is configured to accept requests from the video device 108 based on, e.g., a hypertext transfer protocol (HTTP) or other protocols that may initiate a video session, and to serve the video device 108 with 2D video content.
  • HTTP hypertext transfer protocol
  • the network 106 may include a wide area network (WAN), a local area network (LAN), a wireless network suitable for packet-type communications, such as Internet communications, a broadcast network, or any combination thereof.
  • the network 106 is configured to distribute digital or non-digital video content.
  • the video device 108 is a hardware device such as a computer, a personal digital assistant (PDA), a mobile phone, a laptop, a desktop, a videocassette recorder (VCR), a laserdisc player, a DVD player, a blue ray disc player, or any electronic device configured to output 2D video, i.e., a 2D image sequence.
  • the video device 108 may include software applications that allow the video device 108 to communicate with and receive 2D video content from, e.g., the video storage medium 102 or the media server 104.
  • the video device 108 may, by means of included software
  • the 3D video generator 110 is configured to generate 3D video based on the 2D image sequence outputted by the video device 108.
  • the generator 110 may be implemented as a hardware device that is either stand-alone or incorporated into the video device 108, or software applications installed on the video device 108, or a combination thereof.
  • the 3D video generator 110 may include a processor to generate the 3D video.
  • the 3D video generator 110 may include a rule generator 110-1 configured to generate rules, e.g., semantic rules, for automatic 2D-to-3D conversion, based on a 2D training image sequence, and a 3D video convertor 110-2 configured to automatically convert a 2D input image sequence to 3D video based on the rules.
  • the rule generator 110-1 is also referred to as an interactive 2D-to-3D conversion environment (i3DV)
  • the 3D video convertor 110-2 is also referred to as an automatic 2D-to-3D conversion solution (a3DC).
  • i3DV interactive 2D-to-3D conversion environment
  • a3DC automatic 2D-to-3D conversion solution
  • the display device 112 is configured to display images in the training image sequence or the 2D input image sequence, and to present the generated 3D video.
  • the display device 112 may be provided as a monitor, a projector, or any other video display device.
  • the display device 112 may also be a part of the video device 108.
  • the user may view the images in the training image sequence to provide inputs or feedbacks to the 3D video generator 110.
  • the user may also watch the generated 3D video displayed by the display device 112. It is to be understood that devices shown in Fig. 1 are for illustrative purposes. Certain devices may be removed or combined, and additional devices may be added.
  • Fig. 2 illustrates a block diagram of a rule generator 200, according to an exemplary embodiment.
  • the rule generator 200 may be the rule generator 110-1 (Fig. 1).
  • the rule generator 200 is configured to receive a 2D training image sequence including at least one 2D training image, and to generate 3D video from the training image sequence for preview.
  • the rule generator 200 may further generate rules, e.g., semantic rules, for automatic 2D- to-3D conversion based on the training image sequence.
  • the rule generator 200 may include one or more of the following components: an image sequence analyzing module 202, a scene structure detection module 204, an object segmentation module 206, an object tracking module 208, an object classification module 210, and an object depth estimation module 212.
  • the rule generator 200 may further include a quality evaluation module 214, a user interface 216, an object evolving detection module 218, a depth-image based rendering (DIBR) module 220, a memory device 222 for storing internal data, and a rule database 224.
  • DIBR depth-image based rendering
  • the image sequence analyzing module 202 is a hardware device or software configured to perform analysis of the training image sequence, to determine one or more key frames in the training image sequence. For example, the image sequence analyzing module 202 may analyze the training image sequence to detect scene changes therein, and break the training image sequence into one or more chunks each starting with a key frame where a scene change occurs. The image sequence analyzing module 202 may then output key frame indexes 230 for the determined key frames, and store the key frame indexes 230 as the internal data in the memory device 222.
  • the scene structure detection module 204 is a hardware device or software configured to analyze the training image in the training image sequence to obtain a scene structure 232 for the training image, which is also stored as the internal data in the memory device 222.
  • the scene structure detection module 204 may perform the analysis based on linear perspective properties of the training image, by detecting a vanishing point and vanishing lines in the training image.
  • a vanishing point represents a farthest point in a scene shown in a 2D image, and vanishing lines each represent a direction of increasing depth. The vanishing lines converge at the vanishing point in the 2D image.
  • the scene structure detection module 204 obtains a projected far-to-near direction in the training image and, thus, obtains the scene structure 232.
  • the object segmentation module 206, the object tracking module 208, the object classification module 210, and the object depth estimation module 212 are configured to generate object data 234 regarding semantic objects in the training image, and store the object data 234 as the internal data in the memory device 222.
  • the object segmentation module 206 is a hardware device or software configured to separate scene content in the training image into one or more constituent parts, i.e., semantic objects, referred to hereafter as objects.
  • objects corresponds to a region or shape in the training image that represents an entity with certain semantic meaning, such as a tree, a lake, a house, etc.
  • the object segmentation module 206 may detect the objects in the training image and segment out the detected objects, thereby generating the object data 234.
  • the object segmentation module 206 may group pixels of the training image into different regions based on a homogeneous low-level feature, such as color, motion, or texture, each of the regions representing one of the objects. As a result, the object segmentation module 206 detects and segments out the objects in the training image. In addition, the object segmentation module 206 may cut one or more foreground objects out of the training image, and refine boundaries of the foreground objects.
  • a homogeneous low-level feature such as color, motion, or texture
  • the object tracking module 208 is a hardware device or software configured to use an object tracking algorithm to track the objects in the training image sequence. For example, the object tracking module 208 may detect how positions/sizes of the objects change with time in the training image, thereby generating the object data 234. In such manner, temporal coherence may be provided for generating the 3D video from the training image sequence.
  • the object classification module 210 is a hardware device or software configured to perform object classification for the objects in the training image based on a plurality of training data sets in an object database (not shown).
  • Each of the training data sets includes a group of training images previously processed by the rule generator 200 and representing an object category, such as a building category, a grass category, a tree category, a cow category, a sky category, a face category, a car category, a bicycle category, etc.
  • the object classification module 210 may classify or identify the objects in the present training image, and assign category labels for the classified objects.
  • a training data set may initially include no training images or a relatively small number of training images.
  • the object classification module 210 operates together with the user interface 216 to perform user interactive object classification, as described below. As a result, the object classification module 210 learns additional object categories.
  • the object depth estimation module 212 is a hardware device or software configured to estimate depths of the objects in the training image. Typically, each pixel of the objects in a 2D image
  • a depth of a pixel of an object represents a distance between a viewer and a part of the object corresponding to that pixel.
  • the object depth estimation module 212 may detect the vanishing point and the vanishing lines in the training image, and generate a depth map of the training image accordingly. For example, the object depth estimation module 212 may generate different depth gradient planes relative to the vanishing point and the vanishing lines. The object depth estimation module 212 then assigns depth levels to pixels of the training image according to the depth gradient planes, thereby generating the object data 234. The object depth estimation module 212 may additionally estimate orientation or thicknesses of the objects in the training image.
  • the quality evaluation module 214 is a hardware device or software configured to provide evaluation of visual quality of the 3D video generated from the training image sequence.
  • the quality evaluation module 214 may provide a quality value 236 representing the visual quality of the generated 3D video, and store the quality value 236 in the memory device 222.
  • the quality value 236 is also part of the internal data in the memory device 222.
  • the user interface 216 is configured to receive user input and to output the 3D video generated from the training image sequence for preview, such that the rule generator 200 operates in a user interactive manner.
  • the user interface 216 may be implemented with a hardware device, such as a keyboard or a mouse, to receive user input, and/or a software application to process the user input. By involving user interaction in the rule generating process, efficiency and accuracy may be improved.
  • the user interface 216 receives user input regarding key frame adjustment (240). For example, a user may view the training image sequence on a display device, such as the display device 112 (Fig. 1), and make adjustment to the key frames determined by the image sequence analyzing module 202. The user may add additional key frames or remove ones of the key frames determined by the image sequence analyzing module 202. As a result, the user interface 216 receives the user input regarding the key frame adjustment, and the key frame indexes 230 are interactively specified or adjusted by the user.
  • the user interface 216 receives user input regarding specification or adjustment of the vanishing point/vanishing lines/depth gradient planes in the training image (242).
  • the user may view the training image on the display device and provide scene structure information by specifying or adjusting the vanishing point/vanishing lines/depth gradient planes in the training image.
  • the user interface 216 receives the user input regarding the specification or adjustment of the vanishing point/vanishing lines/depth gradient planes in the training image, and the scene structure 232 is interactively specified or adjusted by the user.
  • the user interface 216 receives user input regarding object selection, object editing, specification of object depth and orientation, object labeling, and/or specification of a topological relationship for the objects in the training image (244).
  • the user may view the training image on the display device and provide semantic object information for the training image.
  • the user may select an object in the training image and edit a shape of the selected object.
  • the user may also specify depth and orientation of the selected object, and classify the selected object by labeling the selected object with a category.
  • the user may additionally specify the topological relationship for the objects in the training image.
  • the user interface 216 receives user input regarding the object selection, the object editing, the specification of the object depth and orientation, the object labeling, and/or the specification of the topological relationship for the objects in the training image, and the object data 234 is interactively specified or adjusted by the user.
  • the user interface 216 receives user input regarding an evaluation of the visual quality of the 3D video generated from the training image sequence (246). For example, the user may view the generated 3D video on the display device and check its visual quality. The user may determine the generated 3D video has relatively good or relatively poor visual quality, and specify/adjust the quality value 236. As a result, the user interface 216 receives the user input regarding the evaluation of the visual quality of the generated 3D video, and the quality value 236 is interactively specified or adjusted by the user.
  • the object evolving detection module 218 is a hardware device or software configured to detect for each of the objects in the training image sequence an object evolving path in the time domain, based on the internal data stored in the memory device 222, such as the key frame indexes 230, the scene structure 232, and/or the object data 234.
  • the object evolving detection module 218 is also configured to provide the detected object evolving path for user preview through the user interface 216.
  • the object evolving detection module 218 manages lifecycles of all the objects that have appeared in the scene, from their first appearance, being partially/wholly occluded, being moved or deformed, until final disappearance from the scene.
  • the DIBR module 220 is a hardware device or software configured to apply DIBR algorithms to generate the 3D video for the training image sequence, based on the internal data stored in the memory device 222, such as the key frame indexes 230, the scene structure 232, and/or the object data 234.
  • the DIBR module 220 is also configured to provide the generated 3D video for user preview through the user interface 216.
  • the DIBR algorithms may include 3D image warping.
  • 3D image warping changes a view direction and a viewpoint of an object, and transforms pixels in a reference image of the object to a destination view in a 3D environment based on depth levels of the pixels.
  • a function may be used to map pixels from the reference image to the destination view.
  • the DIBR module 220 may adjust and reconstruct the destination view to achieve a better effect.
  • the DIBR algorithms may also include plenoptic image modeling.
  • Plenoptic image modeling provides 3D scene information of an image visible from arbitrary viewpoints.
  • the 3D scene information may be obtained by a function based on a set of reference images with depth information. These reference images are warped and combined to form 3D representations of the scene from a particular viewpoint.
  • the DIBR module 220 may adjust and reconstruct the 3D scene information. Based on the 3D scene information, the DIBR module 220 may generate multi-view video frames for 3D displaying.
  • the rule database 224 may then generate rules for automatic 2D-to-3D conversion based on the internal data stored in the memory device 222, such as the key frame indexes 230, the scene structure 232, the object data 234, and/or the quality value 236.
  • the rule database 224 also stores the generated rules for further use.
  • Fig. 3 illustrates a block diagram of a 3D video convertor 300, according to an exemplary embodiment.
  • the 3D video convertor 300 may be the 3D video convertor 110-2 (Fig. 1).
  • the 3D video convertor 300 is configured to receive a 2D input image sequence including at least one 2D input image to be converted to 3D video, and to perform automatic 2D-to-3D conversion on the input image sequence based on the rules generated by the rule generator 200 (Fig. 2).
  • the 3D video convertor 300 may include a complexity control module 302, an object segmentation module 304, an object depth estimation module 306, and a DIBR module 308.
  • the 3D video convertor 300 may further include an interface 310, e.g., a plug-in interface, for connecting the 3D video convertor 300 to the rule generator 200 (Fig. 2) and for obtaining the rules therefrom.
  • the complexity control module 302 is a hardware device or software configured to specify a computational complexity for the 2D-to-3D conversion.
  • the user may configure or customize algorithms adopted by the object segmentation module 304 or the object depth estimation module 306.
  • the user may customize an algorithm according to a set of complexity levels, or provide a specific number of logical operation limitations as well as selected features in the algorithm, which may automatically generate a configuration for the algorithm that provides a balance between computational complexity and system performance.
  • the object segmentation module 304 is a hardware device or software configured to separate scene content in the input image into one or more objects, based on the rules obtained from the rule generator 200 (Fig. 2).
  • the object segmentation module 304 may detect the objects in the input image and segment out the detected objects, similar to the above description in connection with the object segmentation module 206 (Fig. 2).
  • the object depth estimation module 306 is a hardware device or software configured to estimate depths of the objects in the input image.
  • the object depth estimation module 306 may use a gravity-based depth generation algorithm to perform automatic 2D-to-3D conversion when, initially, the rules from the rule generator 200 (Fig. 2) are not sufficient for performing the conversion.
  • the object depth estimation module 306 may generate different depth gradient planes relative to a vanishing point and vanishing lines in the input image, thereby deriving a depth map for the input image, similar to the above description in connection with the object depth estimation module 212 (Fig. 2).
  • the DIBR module 308 is a hardware device or software configured to apply DIBR algorithms to generate 3D video for the input image sequence.
  • the DIBR algorithms may produce a 3D
  • the DIBR module 308 may utilize depth information of one or more neighboring images in the input image sequence.
  • the rule generator 200 (Fig. 2) and the 3D video convertor 300 (Fig. 3) may be connected in a loose connection manner or in a tight connection manner.
  • the 3D video convertor 300 is a well established, e.g., independent, system that may obtain the rules from the rule generator 200 to perform automatic 2D-to-3D conversion.
  • the 3D video convertor 300 may be software applications automatically developed by the rule generator 200 or may share one or more modules of the rule generator 200, to perform automatic 2D-to-3D conversion.
  • Fig. 4 illustrates a block diagram of a 3D video generator 400 configured to operate in the loose connection manner, according to an exemplary embodiment.
  • the 3D video generator 400 includes a rule generator 402, similar to the rule generator 200 (Fig. 2) described above, and a 3D video convertor 404, similar to the 3D video convertor 300 (Fig. 3) described above.
  • the 3D video convertor 404 is connected to the rule generator 402 through, e.g., a plug-in interface (not shown).
  • the object segmentation module and the object depth estimation module in the 3D video convertor 404 may obtain rules from the rule database in the rule generator 402, and perform automatic 2D-to-3D conversion on a 2D input image sequence based on the obtained rules.
  • Fig. 5 illustrates a block diagram of a 3D video generator 500 configured to operate in the tight connection manner, according to an exemplary embodiment.
  • the 3D video generator 500 includes a rule generator 502, similar to the rule generator 200 (Fig. 2) described above, and 3D video convertors 504-1 , 504-2, ... and 504-N, each similar to the 3D video convertor 300 (Fig. 3) described above.
  • the rule generator 502 further includes a software application, referred to herein as a
  • 3D-video-convertor generator 506 to generate additional software applications or computer programs for automatic 2D-to-3D conversion, referred to herein as 3D video convertors 504-1 , 504-2, ... and 504-N.
  • the 3D-video-convertor generator 506 may generate the 3D video convertors 504-1 , 504-2, ... and 504-N based on rules in the rule database in the rule generator 502.
  • Each of the 3D video convertors 504-1 , 504-2, ... and 504-N may perform automatic 2D-to-3D conversion for a different scenario in a 2D input image sequence.
  • Fig. 6 illustrates a flowchart of a method 600 for generating 3D video based on a 2D input image sequence including at least one input 2D image, according to an exemplary embodiment.
  • rules e.g., semantic rules, for automatic 2D-to-3D conversion are generated (602).
  • the 2D input image sequence may then be automatically converted to the 3D video based on the rules (604).
  • internal data is generated based on a 2D training image sequence including at least one 2D training image.
  • 3D video from the 2D training image sequence is then generated based on the internal data and depth-image based rendering. Quality evaluation for the generated 3D video may be further provided and an evaluation result is included in the internal data.
  • the rules are generated based on the internal data including the evaluation result, and are stored in a rule database.
  • the generating of the internal data may include at least one of: determining key frame indexes for the 2D training image sequence and storing the key frame indexes as the internal data; detecting a scene structure in the 2D training image and storing the scene structure as the internal data;
  • segmenting out objects in the 2D training image and storing a segmentation result as the internal data tracking the objects in the 2D training image sequence and storing a tracking result as the internal data; classifying the objects in the 2D training image and storing a classification result as the internal data; or
  • the rules are obtained, and objects in the 2D input image are segmented out based on the obtained rules.
  • Depth information for the 2D input image is also estimated based on the obtained rules.
  • 3D video is automatically generated from the 2D input image sequence based on the object segmentation and the depth information estimation.
  • the method disclosed herein may be implemented as a computer program product, i.e., a computer program tangibly embodied in an information carrier, e.g., a machine readable storage device, for execution by a data processing apparatus, e.g., a programmable processor, a computer, or multiple computers.
  • the computer program may be written in any form of programming language, including compiled or interpreted languages, and may be deployed in any form, including stand-alone program, module, subroutine, or other unit suitable for use in a computing environment.
  • the computer program may be deployed to be executed on one computer, or on multiple computers.
  • a computer- readable medium including instructions, executable by a processor in a 3D video generating system, for performing the above described method for generating 3D video based on a 2D input image sequence.
  • a portion or all of the method disclosed herein may also be implemented by an application specific integrated circuit (ASIC), a field- programmable gate array (FPGA), a complex programmable logic device (CPLD), a printed circuit board (PCB), a digital signal processor (DSP), a combination of programmable logic components and programmable interconnects, a single central processing unit (CPU) chip, or a CPU chip combined on a motherboard.
  • ASIC application specific integrated circuit
  • FPGA field- programmable gate array
  • CPLD complex programmable logic device
  • PCB printed circuit board
  • DSP digital signal processor
  • CPU central processing unit
  • CPU central processing unit
  • CPU central processing unit

Landscapes

  • Engineering & Computer Science (AREA)
  • Multimedia (AREA)
  • Signal Processing (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Theoretical Computer Science (AREA)
  • Processing Or Creating Images (AREA)

Abstract

L'invention concerne un système de génération de vidéo tridimensionnelle (3D) sur la base d’une suite d’images d’entrée bidimensionnelles (2D) comprenant au moins une image d’entrée 2D. Le système comprend : un générateur de règles configuré pour générer des règles de conversion 2D en 3D; et un convertisseur vidéo 3D couplé au générateur de règles et configuré pour obtenir les règles à partir du générateur de règles et convertir automatiquement la suite d’images d’entrée 2D en vidéo 3D sur la base des règles obtenues.
PCT/US2010/044227 2009-08-04 2010-08-03 Systèmes et procédés de génération de vidéo tridimensionnelle WO2011017308A1 (fr)

Priority Applications (1)

Application Number Priority Date Filing Date Title
EP10807016.0A EP2462536A4 (fr) 2009-08-04 2010-08-03 Systèmes et procédés de génération de vidéo tridimensionnelle

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
US23128609P 2009-08-04 2009-08-04
US61/231,286 2009-08-04

Publications (1)

Publication Number Publication Date
WO2011017308A1 true WO2011017308A1 (fr) 2011-02-10

Family

ID=43544623

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/US2010/044227 WO2011017308A1 (fr) 2009-08-04 2010-08-03 Systèmes et procédés de génération de vidéo tridimensionnelle

Country Status (2)

Country Link
EP (1) EP2462536A4 (fr)
WO (1) WO2011017308A1 (fr)

Cited By (10)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2012061549A2 (fr) * 2010-11-03 2012-05-10 3Dmedia Corporation Procédés, systèmes et produits logiciels informatiques pour créer des séquences vidéo tridimensionnelles
US8274552B2 (en) 2010-12-27 2012-09-25 3Dmedia Corporation Primary and auxiliary image capture devices for image processing and related methods
WO2012157886A2 (fr) 2011-05-17 2012-11-22 Samsung Electronics Co., Ltd. Appareil et procédé de conversion de contenu 2d en contenu 3d, et support de mémoire lisible par ordinateur
US8436893B2 (en) 2009-07-31 2013-05-07 3Dmedia Corporation Methods, systems, and computer-readable storage media for selecting image capture positions to generate three-dimensional (3D) images
US8508580B2 (en) 2009-07-31 2013-08-13 3Dmedia Corporation Methods, systems, and computer-readable storage media for creating three-dimensional (3D) images of a scene
US9344701B2 (en) 2010-07-23 2016-05-17 3Dmedia Corporation Methods, systems, and computer-readable storage media for identifying a rough depth map in a scene and for determining a stereo-base distance for three-dimensional (3D) content creation
US9380292B2 (en) 2009-07-31 2016-06-28 3Dmedia Corporation Methods, systems, and computer-readable storage media for generating three-dimensional (3D) images of a scene
EP2568439A3 (fr) * 2011-09-08 2016-12-28 Samsung Electronics Co., Ltd. Appareil et procédé pour générer des informations de profondeur
US10200671B2 (en) 2010-12-27 2019-02-05 3Dmedia Corporation Primary and auxiliary image capture devices for image processing and related methods
CN110119663A (zh) * 2018-02-06 2019-08-13 耐能有限公司 低功率消耗的脸部辨识方法及低功率消耗的脸部辨识系统

Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20060067573A1 (en) * 2000-03-08 2006-03-30 Parr Timothy C System, method, and apparatus for generating a three-dimensional representation from one or more two-dimensional images
US20080181486A1 (en) * 2007-01-26 2008-07-31 Conversion Works, Inc. Methodology for 3d scene reconstruction from 2d image sequences
US20080205791A1 (en) * 2006-11-13 2008-08-28 Ramot At Tel-Aviv University Ltd. Methods and systems for use in 3d video generation, storage and compression
US20090116732A1 (en) * 2006-06-23 2009-05-07 Samuel Zhou Methods and systems for converting 2d motion pictures for stereoscopic 3d exhibition

Patent Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20060067573A1 (en) * 2000-03-08 2006-03-30 Parr Timothy C System, method, and apparatus for generating a three-dimensional representation from one or more two-dimensional images
US20090116732A1 (en) * 2006-06-23 2009-05-07 Samuel Zhou Methods and systems for converting 2d motion pictures for stereoscopic 3d exhibition
US20080205791A1 (en) * 2006-11-13 2008-08-28 Ramot At Tel-Aviv University Ltd. Methods and systems for use in 3d video generation, storage and compression
US20080181486A1 (en) * 2007-01-26 2008-07-31 Conversion Works, Inc. Methodology for 3d scene reconstruction from 2d image sequences

Cited By (23)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US8508580B2 (en) 2009-07-31 2013-08-13 3Dmedia Corporation Methods, systems, and computer-readable storage media for creating three-dimensional (3D) images of a scene
US11044458B2 (en) 2009-07-31 2021-06-22 3Dmedia Corporation Methods, systems, and computer-readable storage media for generating three-dimensional (3D) images of a scene
US9380292B2 (en) 2009-07-31 2016-06-28 3Dmedia Corporation Methods, systems, and computer-readable storage media for generating three-dimensional (3D) images of a scene
US12034906B2 (en) 2009-07-31 2024-07-09 3Dmedia Corporation Methods, systems, and computer-readable storage media for generating three-dimensional (3D) images of a scene
US8436893B2 (en) 2009-07-31 2013-05-07 3Dmedia Corporation Methods, systems, and computer-readable storage media for selecting image capture positions to generate three-dimensional (3D) images
US8810635B2 (en) 2009-07-31 2014-08-19 3Dmedia Corporation Methods, systems, and computer-readable storage media for selecting image capture positions to generate three-dimensional images
US9344701B2 (en) 2010-07-23 2016-05-17 3Dmedia Corporation Methods, systems, and computer-readable storage media for identifying a rough depth map in a scene and for determining a stereo-base distance for three-dimensional (3D) content creation
WO2012061549A3 (fr) * 2010-11-03 2012-07-12 3Dmedia Corporation Procédés, systèmes et produits logiciels informatiques pour créer des séquences vidéo tridimensionnelles
WO2012061549A2 (fr) * 2010-11-03 2012-05-10 3Dmedia Corporation Procédés, systèmes et produits logiciels informatiques pour créer des séquences vidéo tridimensionnelles
US9185388B2 (en) 2010-11-03 2015-11-10 3Dmedia Corporation Methods, systems, and computer program products for creating three-dimensional video sequences
US11388385B2 (en) 2010-12-27 2022-07-12 3Dmedia Corporation Primary and auxiliary image capture devices for image processing and related methods
US10200671B2 (en) 2010-12-27 2019-02-05 3Dmedia Corporation Primary and auxiliary image capture devices for image processing and related methods
US8441520B2 (en) 2010-12-27 2013-05-14 3Dmedia Corporation Primary and auxiliary image capture devcies for image processing and related methods
US8274552B2 (en) 2010-12-27 2012-09-25 3Dmedia Corporation Primary and auxiliary image capture devices for image processing and related methods
US10911737B2 (en) 2010-12-27 2021-02-02 3Dmedia Corporation Primary and auxiliary image capture devices for image processing and related methods
EP2695384A4 (fr) * 2011-05-17 2014-10-08 Samsung Electronics Co Ltd Appareil et procédé de conversion de contenu 2d en contenu 3d, et support de mémoire lisible par ordinateur
EP2695384A2 (fr) * 2011-05-17 2014-02-12 Samsung Electronics Co., Ltd. Appareil et procédé de conversion de contenu 2d en contenu 3d, et support de mémoire lisible par ordinateur
CN103548343A (zh) * 2011-05-17 2014-01-29 三星电子株式会社 用于将2d内容转换为3d内容的设备和方法及其计算机可读存储介质
WO2012157886A2 (fr) 2011-05-17 2012-11-22 Samsung Electronics Co., Ltd. Appareil et procédé de conversion de contenu 2d en contenu 3d, et support de mémoire lisible par ordinateur
KR101862543B1 (ko) * 2011-09-08 2018-07-06 삼성전자 주식회사 깊이정보 생성장치, 깊이정보 생성방법 및 그 저장매체
EP2568439A3 (fr) * 2011-09-08 2016-12-28 Samsung Electronics Co., Ltd. Appareil et procédé pour générer des informations de profondeur
CN110119663A (zh) * 2018-02-06 2019-08-13 耐能有限公司 低功率消耗的脸部辨识方法及低功率消耗的脸部辨识系统
CN110119663B (zh) * 2018-02-06 2022-12-30 耐能有限公司 低功率消耗的脸部辨识方法及低功率消耗的脸部辨识系统

Also Published As

Publication number Publication date
EP2462536A1 (fr) 2012-06-13
EP2462536A4 (fr) 2014-10-22

Similar Documents

Publication Publication Date Title
WO2011017308A1 (fr) Systèmes et procédés de génération de vidéo tridimensionnelle
EP2481023B1 (fr) Conversion vidéo de 2d en 3d
US10529086B2 (en) Three-dimensional (3D) reconstructions of dynamic scenes using a reconfigurable hybrid imaging system
US9846960B2 (en) Automated camera array calibration
JP4938093B2 (ja) 2d−to−3d変換のための2d画像の領域分類のシステム及び方法
US9414048B2 (en) Automatic 2D-to-stereoscopic video conversion
Karsch et al. Depth transfer: Depth extraction from video using non-parametric sampling
Karsch et al. Depth extraction from video using non-parametric sampling
US8861836B2 (en) Methods and systems for 2D to 3D conversion from a portrait image
US9237330B2 (en) Forming a stereoscopic video
CN101689299B (zh) 用于图像的立体匹配的系统和方法
US8897542B2 (en) Depth map generation based on soft classification
US20100220920A1 (en) Method, apparatus and system for processing depth-related information
Jang et al. Efficient disparity map estimation using occlusion handling for various 3D multimedia applications
JP7184748B2 (ja) 場面の積層深度データを生成するための方法
US20120068996A1 (en) Safe mode transition in 3d content rendering
US9661307B1 (en) Depth map generation using motion cues for conversion of monoscopic visual content to stereoscopic 3D
Jain et al. Efficient stereo-to-multiview synthesis
US20150030233A1 (en) System and Method for Determining a Depth Map Sequence for a Two-Dimensional Video Sequence
JPWO2011030399A1 (ja) 画像処理方法及び装置
WO2008152607A1 (fr) Procede, appareil, systeme et programme informatique de propagation d'informations relatives a la profondeur
KR101511315B1 (ko) 스테레오스코픽 컨텐츠를 위한 다이나믹 플로팅 윈도우 생성 방법 및 시스템
EP2462539A1 (fr) Systèmes et procédés de génération de vidéo tridimensionnelle
Lee et al. Estimating scene-oriented pseudo depth with pictorial depth cues
CN117475105A (zh) 一种基于单目图像的开放世界三维场景重建及感知方法

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 10807016

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: DE

REEP Request for entry into the european phase

Ref document number: 2010807016

Country of ref document: EP

WWE Wipo information: entry into national phase

Ref document number: 2010807016

Country of ref document: EP