WO2009157701A2 - Image generating method and apparatus and image processing method and apparatus - Google Patents

Image generating method and apparatus and image processing method and apparatus Download PDF

Info

Publication number
WO2009157701A2
WO2009157701A2 PCT/KR2009/003383 KR2009003383W WO2009157701A2 WO 2009157701 A2 WO2009157701 A2 WO 2009157701A2 KR 2009003383 W KR2009003383 W KR 2009003383W WO 2009157701 A2 WO2009157701 A2 WO 2009157701A2
Authority
WO
WIPO (PCT)
Prior art keywords
image
video data
shot
information
frames
Prior art date
Application number
PCT/KR2009/003383
Other languages
French (fr)
Other versions
WO2009157701A3 (en
Inventor
Kil-Soo Jung
Hyun-Kwon Chung
Dae-Jong Lee
Original Assignee
Samsung Electronics Co,. Ltd.
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Samsung Electronics Co,. Ltd. filed Critical Samsung Electronics Co,. Ltd.
Priority to CN200980123639.6A priority Critical patent/CN102067615B/en
Priority to JP2011514502A priority patent/JP2011525745A/en
Priority to EP09770373.0A priority patent/EP2289247A4/en
Publication of WO2009157701A2 publication Critical patent/WO2009157701A2/en
Publication of WO2009157701A3 publication Critical patent/WO2009157701A3/en

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T9/00Image coding
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N13/00Stereoscopic video systems; Multi-view video systems; Details thereof
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T15/003D [Three Dimensional] image rendering
    • G06T15/005General purpose rendering architectures
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T5/00Image enhancement or restoration
    • G06T5/40Image enhancement or restoration using histogram techniques
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T5/00Image enhancement or restoration
    • G06T5/90Dynamic range modification of images or parts thereof
    • G06T5/92Dynamic range modification of images or parts thereof based on global image properties
    • GPHYSICS
    • G11INFORMATION STORAGE
    • G11BINFORMATION STORAGE BASED ON RELATIVE MOVEMENT BETWEEN RECORD CARRIER AND TRANSDUCER
    • G11B27/00Editing; Indexing; Addressing; Timing or synchronising; Monitoring; Measuring tape travel
    • G11B27/02Editing, e.g. varying the order of information signals recorded on, or reproduced from, record carriers
    • G11B27/031Electronic editing of digitised analogue information signals, e.g. audio or video signals
    • G11B27/034Electronic editing of digitised analogue information signals, e.g. audio or video signals on discs
    • GPHYSICS
    • G11INFORMATION STORAGE
    • G11BINFORMATION STORAGE BASED ON RELATIVE MOVEMENT BETWEEN RECORD CARRIER AND TRANSDUCER
    • G11B27/00Editing; Indexing; Addressing; Timing or synchronising; Monitoring; Measuring tape travel
    • G11B27/10Indexing; Addressing; Timing or synchronising; Measuring tape travel
    • GPHYSICS
    • G11INFORMATION STORAGE
    • G11BINFORMATION STORAGE BASED ON RELATIVE MOVEMENT BETWEEN RECORD CARRIER AND TRANSDUCER
    • G11B27/00Editing; Indexing; Addressing; Timing or synchronising; Monitoring; Measuring tape travel
    • G11B27/10Indexing; Addressing; Timing or synchronising; Measuring tape travel
    • G11B27/19Indexing; Addressing; Timing or synchronising; Measuring tape travel by using information detectable on the record carrier
    • G11B27/28Indexing; Addressing; Timing or synchronising; Measuring tape travel by using information detectable on the record carrier by using information signals recorded by the same method as the main recording
    • G11B27/32Indexing; Addressing; Timing or synchronising; Measuring tape travel by using information detectable on the record carrier by using information signals recorded by the same method as the main recording on separate auxiliary tracks of the same or an auxiliary record carrier
    • G11B27/322Indexing; Addressing; Timing or synchronising; Measuring tape travel by using information detectable on the record carrier by using information signals recorded by the same method as the main recording on separate auxiliary tracks of the same or an auxiliary record carrier used signal is digitally coded
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N13/00Stereoscopic video systems; Multi-view video systems; Details thereof
    • H04N13/10Processing, recording or transmission of stereoscopic or multi-view image signals
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N13/00Stereoscopic video systems; Multi-view video systems; Details thereof
    • H04N13/10Processing, recording or transmission of stereoscopic or multi-view image signals
    • H04N13/106Processing image signals
    • H04N13/111Transformation of image signals corresponding to virtual viewpoints, e.g. spatial image interpolation
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N13/00Stereoscopic video systems; Multi-view video systems; Details thereof
    • H04N13/10Processing, recording or transmission of stereoscopic or multi-view image signals
    • H04N13/106Processing image signals
    • H04N13/122Improving the 3D impression of stereoscopic images by modifying image signal contents, e.g. by filtering or adding monoscopic depth cues
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N13/00Stereoscopic video systems; Multi-view video systems; Details thereof
    • H04N13/10Processing, recording or transmission of stereoscopic or multi-view image signals
    • H04N13/106Processing image signals
    • H04N13/133Equalising the characteristics of different image components, e.g. their average brightness or colour balance
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N13/00Stereoscopic video systems; Multi-view video systems; Details thereof
    • H04N13/10Processing, recording or transmission of stereoscopic or multi-view image signals
    • H04N13/106Processing image signals
    • H04N13/139Format conversion, e.g. of frame-rate or size
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N13/00Stereoscopic video systems; Multi-view video systems; Details thereof
    • H04N13/10Processing, recording or transmission of stereoscopic or multi-view image signals
    • H04N13/106Processing image signals
    • H04N13/158Switching image signals
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N13/00Stereoscopic video systems; Multi-view video systems; Details thereof
    • H04N13/10Processing, recording or transmission of stereoscopic or multi-view image signals
    • H04N13/106Processing image signals
    • H04N13/161Encoding, multiplexing or demultiplexing different image signal components
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N13/00Stereoscopic video systems; Multi-view video systems; Details thereof
    • H04N13/10Processing, recording or transmission of stereoscopic or multi-view image signals
    • H04N13/106Processing image signals
    • H04N13/172Processing image signals image signals comprising non-image signal components, e.g. headers or format information
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N13/00Stereoscopic video systems; Multi-view video systems; Details thereof
    • H04N13/10Processing, recording or transmission of stereoscopic or multi-view image signals
    • H04N13/106Processing image signals
    • H04N13/172Processing image signals image signals comprising non-image signal components, e.g. headers or format information
    • H04N13/178Metadata, e.g. disparity information
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N13/00Stereoscopic video systems; Multi-view video systems; Details thereof
    • H04N13/10Processing, recording or transmission of stereoscopic or multi-view image signals
    • H04N13/106Processing image signals
    • H04N13/172Processing image signals image signals comprising non-image signal components, e.g. headers or format information
    • H04N13/183On-screen display [OSD] information, e.g. subtitles or menus
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N13/00Stereoscopic video systems; Multi-view video systems; Details thereof
    • H04N13/10Processing, recording or transmission of stereoscopic or multi-view image signals
    • H04N13/189Recording image signals; Reproducing recorded image signals
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N13/00Stereoscopic video systems; Multi-view video systems; Details thereof
    • H04N13/20Image signal generators
    • H04N13/261Image signal generators with monoscopic-to-stereoscopic image conversion
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N13/00Stereoscopic video systems; Multi-view video systems; Details thereof
    • H04N13/30Image reproducers
    • H04N13/332Displays for viewing with the aid of special glasses or head-mounted displays [HMD]
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N13/00Stereoscopic video systems; Multi-view video systems; Details thereof
    • H04N13/30Image reproducers
    • H04N13/332Displays for viewing with the aid of special glasses or head-mounted displays [HMD]
    • H04N13/339Displays for viewing with the aid of special glasses or head-mounted displays [HMD] using spatial multiplexing
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N13/00Stereoscopic video systems; Multi-view video systems; Details thereof
    • H04N13/30Image reproducers
    • H04N13/332Displays for viewing with the aid of special glasses or head-mounted displays [HMD]
    • H04N13/341Displays for viewing with the aid of special glasses or head-mounted displays [HMD] using temporal multiplexing
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N13/00Stereoscopic video systems; Multi-view video systems; Details thereof
    • H04N13/30Image reproducers
    • H04N13/356Image reproducers having separate monoscopic and stereoscopic modes
    • H04N13/359Switching between monoscopic and stereoscopic modes
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N13/00Stereoscopic video systems; Multi-view video systems; Details thereof
    • H04N13/30Image reproducers
    • H04N13/361Reproducing mixed stereoscopic images; Reproducing mixed monoscopic and stereoscopic images, e.g. a stereoscopic image overlay window on a monoscopic image background
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/50Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding
    • H04N19/597Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding specially adapted for multi-view video sequence encoding
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N5/00Details of television systems
    • H04N5/14Picture signal circuitry for video frequency region
    • H04N5/144Movement detection
    • H04N5/145Movement estimation
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/14Digital output to display device ; Cooperation and interconnection of the display device with other functional units
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/10Image acquisition modality
    • G06T2207/10016Video; Image sequence
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/20Special algorithmic details
    • G06T2207/20172Image enhancement details
    • G06T2207/20208High dynamic range [HDR] image processing
    • GPHYSICS
    • G09EDUCATION; CRYPTOGRAPHY; DISPLAY; ADVERTISING; SEALS
    • G09GARRANGEMENTS OR CIRCUITS FOR CONTROL OF INDICATING DEVICES USING STATIC MEANS TO PRESENT VARIABLE INFORMATION
    • G09G2320/00Control of display operating conditions
    • G09G2320/06Adjustment of display parameters
    • G09G2320/0626Adjustment of display parameters for control of overall brightness
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N13/00Stereoscopic video systems; Multi-view video systems; Details thereof
    • H04N13/10Processing, recording or transmission of stereoscopic or multi-view image signals
    • H04N13/106Processing image signals
    • H04N13/156Mixing image signals
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N13/00Stereoscopic video systems; Multi-view video systems; Details thereof
    • H04N13/10Processing, recording or transmission of stereoscopic or multi-view image signals
    • H04N13/194Transmission of image signals
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N13/00Stereoscopic video systems; Multi-view video systems; Details thereof
    • H04N13/20Image signal generators
    • H04N13/286Image signal generators having separate monoscopic and stereoscopic modes
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N2213/00Details of stereoscopic systems
    • H04N2213/005Aspects relating to the "3D+depth" image format

Definitions

  • aspects of the present invention generally relate to an image generating method and apparatus and an image processing method and apparatus, and more particularly, to an image generating method and apparatus and an image processing method and apparatus in which video data is output as a two-dimensional (2D) image or a three-dimensional (3D) image by using metadata associated with the video data.
  • 3D image technology expresses a more realistic image by adding depth information to a two-dimensional (2D) image.
  • the 3D image technology can be classified into technology to generate video data as a 3D image and technology to convert video data generated as a 2D image into a 3D image. Both technologies have been studied together.
  • aspects of the present invention provide an image processing method and apparatus to output video data as a two-dimensional image or a three-dimensional image by using metadata associated with the video data can be provided.
  • video data can be output as a 2D image at a shot change point.
  • it is determined for each shot whether to output video data as a 2D image or a 3D image and the video data is output according to a result of the determination, thereby reducing the amount of computation that may increase due to conversion of total video data into a 3D image.
  • FIG. 1 is a block diagram of an image generating apparatus according to an embodiment of the present invention
  • FIG. 2 illustrates metadata generated by the image generating apparatus illustrated in FIG. 1;
  • FIGs. 3A through 3C are views to explain a depth map generated by using background depth information
  • FIG. 4 is a block diagram of an image processing apparatus according to an embodiment of the present invention.
  • FIG. 5 is a block diagram of an image processing apparatus according to another embodiment of the present invention.
  • FIG. 6 is a flowchart illustrating an image processing method according to an embodiment of the present invention.
  • FIG. 7 is a flowchart illustrating in detail an operation illustrated in FIG. 6 where video data is output as a two-dimensional (2D) image or a three-dimensional (3D) image.
  • an image processing method to output video data being a two-dimensional (2D) image as the 2D image or a three-dimensional (3D) image the image processing method including: extracting information about the video data from metadata associated with the video data; and outputting the video data as the 2D image or the 3D image by using the extracted information about the video data, wherein the information about the video data includes information to classify frames of the video data into predetermined units.
  • the information to classify the frames of the video data as the predetermined units may be shot information to classify a group of frames in which a background composition of a current frame is predictable by using a previous frame preceding the current frame as a shot.
  • the shot information may include output moment information of a frame being output first and output moment information of a frame being output last from among the group of frames classified as the shot.
  • the metadata may include shot type information indicating whether the frames classified as the shot are to be output as the 2D image or the 3D image
  • the outputting of the video data may include outputting the frames classified as the shot as the 2D image or the 3D image by using the shot type information.
  • the outputting of the video data may include determining, by using the metadata, whether a background composition of a current frame is not predictable by using a previous frame preceding the current frame and thus the current frame is classified as a new shot, outputting the current frame as the 2D image when the current frame is classified as the new shot, and converting the remaining frames of the frames classified as the new shot into the 3D image and outputting the converted 3D image.
  • the outputting of the video data may include determining, by using the metadata, whether a background composition of a current frame is not predictable by using a previous frame preceding the current frame and thus the current frame is classified as a new shot, extracting background depth information to be applied to the current frame classified as the new shot from the metadata when the current frame is classified as the new shot, and generating a depth map for the current frame by using the background depth information.
  • the generating of the depth map for the current frame may include generating the depth map for a background of the current frame by using coordinate point values of the background of the current frame, depth values corresponding to the coordinate point values, and a panel position value, in which the coordinate point values, the depth value, and the panel position value are included in the background depth information.
  • the image processing method may further include reading the metadata from a disc recorded with the video data or downloading the metadata from a server through a communication network.
  • the metadata may include identification information to identify the video data
  • the identification information may include a disc identifier (ID) to identify a disc recorded with the video data and a title ID to indicate a title including the video data among a plurality of titles recorded in the disc identified by the disc ID.
  • ID disc identifier
  • an image generating method including: receiving video data being a two-dimensional (2D) image; and generating metadata associated with the video data, the metadata including information to classify frames of the video data as predetermined units, wherein the information to classify the frames of the video data as the predetermined units is shot information to classify a group of frames in which a background composition of a current frame is predictable by using a previous frame preceding the current frame as a shot.
  • 2D two-dimensional
  • the shot information may include output moment information of a frame being output first and output moment information of a frame being output last from among the frames classified as the shot, and/or may include shot type information indicating whether the frames classified as the shot are to be output as the 2D image or a three-dimensional (3D) image.
  • the metadata may include background depth information for frames classified as a predetermined shot and the background depth information may include coordinate point values of a background of the frame classified as the predetermined shot, depth values corresponding to the coordinate point values, and a panel position value.
  • an image processing apparatus to output video data being a two-dimensional (2D) image as the 2D image or a three-dimensional (3D) image
  • the image processing apparatus including: a metadata analyzing unit to determine whether the video data is to be output as the 2D image or the 3D image by using metadata associated with the video data; a 3D image converting unit to convert the video data into the 3D image when the video data is to be output as the 3D image; and an output unit to output the video data as the 2D image or the 3D image, wherein the metadata includes information to classify frames of the video data into predetermined units.
  • the information to classify the frames of the video data into the predetermined units may be shot information to classify a group of frames in which a background composition of a current frame is predictable by using a previous frame preceding the current frame as a shot.
  • the shot information may include output moment information of a frame being output first and output moment information of a frame being output last from among the frames classified as the shot.
  • the metadata may include shot type information indicating whether the frames classified as the shot are to be output as the 2D image or the 3D image.
  • the metadata may include background depth information for a frame classified as a predetermined shot, and the background depth information may include coordinate point values of a background of the frame classified as the predetermined shot, depth values corresponding to the coordinate point values, and a panel position value.
  • an image generating apparatus including: a video data encoding unit to encode video data being a two-dimensional (2D) image; a metadata generating unit to generate metadata associated with the video data, the metadata including information to classify frames of the video data into predetermined units; and a metadata encoding unit to encode the metadata, in which the information to classify the frames of the video data into the predetermined units is shot information to classify a group of frames in which a background composition of a current frame is predictable by using a previous frame preceding the current frame as a shot.
  • 2D two-dimensional
  • a computer-readable information storage medium including video data being a two-dimensional (2D) image and metadata associated with the video data, the metadata including information to classify frames of the video data into predetermined units, wherein the information to classify the frames of the video data into the predetermined units is shot information to classify a group of frames in which a background composition of a current frame is predictable by using a previous frame preceding the current frame as a shot.
  • 2D two-dimensional
  • a computer-readable information storage medium having recorded thereon a program to execute an image processing method to output video data being a two-dimensional (2D) image as the 2D image or a three-dimensional (3D) image
  • the image processing method including: extracting information about the video data from metadata associated with the video data; and outputting the video data as the 2D image or the 3D image by using the extracted information about the video data, wherein the information about the video data includes information to classify frames of the video data into predetermined units.
  • a system to output video data as a two-dimensional (2D) image or a three-dimensional (3D) image including: an image generating apparatus including: a video data encoding unit to encode the video data being the 2D image, a metadata generating unit to generate metadata associated with the video data, the metadata comprising information to classify frames of the video data as predetermined units and used to determine whether each of the classified frames is to be converted to the 3D image; and an image processing apparatus to receive the encoded video data and the generated metadata, and to output the video data as the 2D image or the 3D image, the image processing apparatus including: a metadata analyzing unit to determine whether the video data is to be output as the 2D image or the 3D image by using the information to classify the frames of the video data comprised in the received metadata associated with the video data, a 3D image converting unit to convert the video data into the 3D image when the metadata analyzing unit determines that the video data is to be output as the 3D image, and an image generating apparatus including: a video data encoding unit
  • a computer-readable information storage medium including: metadata associated with video data comprising two-dimensional (2D) frames, the metadata comprising information used by an image processing apparatus to classify the frames of the video data as predetermined units and used by the image processing apparatus to determine whether each of the classified frames is to be converted by the image processing apparatus to a three-dimensional (3D) image, wherein the information to classify the frames of the video data as the predetermined units comprises shot information to classify, as a shot, a group of frames in which a background composition of a current frame is predictable by using a previous frame preceding the current frame in the group of frames.
  • an image processing method to output video data having two-dimensional (2D) images as the 2D images or three-dimensional (3D) images, the image processing method including: determining, by an image processing apparatus, whether metadata associated with the video data exists on a disc comprising the video data; reading, by the image processing apparatus, the metadata from the disc if the metadata is determined to exist on the disc; retrieving, by the image processing apparatus, the metadata from a server if the metadata is determined to not exist on the disc; and outputting, by the image processing apparatus, the video data as selectable between the 2D image and the 3D image according to the metadata.
  • FIG. 1 is a block diagram of an image generating apparatus 100 according to an embodiment of the present invention.
  • the image generating apparatus 100 includes a video data generating unit 110, a video data encoding unit 120, a metadata generating unit 130, a metadata encoding unit 140, and a multiplexing unit 150.
  • the video data generating unit 110 generates video data and outputs the generated video data to the video data encoding unit 120.
  • the video data encoding unit 120 encodes the input video data and outputs the encoded video data (OUT1) to the multiplexing unit 150, and/or to an image processing apparatus (not shown) through a communication network, though it is understood that the video data encoding unit 120 may output the encoded video data to the image processing apparatus through any wired and/or wireless connection (such as IEEE 1394, universal serial bus, a Bluetooth, an infrared, etc.).
  • the image generating apparatus 100 may be a computer, a workstation, a camera device, a mobile device, a stand-alone device, etc.
  • each of the units 110, 120, 130, 140, 150 can be one or more processors or processing elements on one or more chips or integrated circuits.
  • the metadata generating unit 130 analyzes the video data generated by the video data generating unit 110 to generate metadata including information about frames of the video data.
  • the metadata includes information to convert the generated video data from a two-dimensional (2D) image into a three-dimensional (3D) image.
  • the metadata also includes information to classify the frames of the video data as predetermined units.
  • the metadata generated by the metadata generating unit 130 will be described in more detail with reference to FIG. 2.
  • the metadata generating unit 130 outputs the generated metadata to the metadata encoding unit 140.
  • the metadata encoding unit 140 encodes the input metadata and outputs the encoded metadata (OUT3) to the multiplexing unit 150 and/or to the image processing apparatus.
  • the multiplexing unit 150 multiplexes the encoded video data (OUT1) and the encoded metadata (OUT3) and transmits the multiplexing result (OUT2) to the image processing apparatus through a wired and/or wireless communication network, or any wired and/or wireless connection, as described above.
  • the metadata encoding unit 140 may transmit the encoded metadata (OUT3), separately from the encoded video data (OUT1), to the image processing apparatus, instead of to or in addition to the multiplexing unit 150. In this way, the image generating apparatus 100 generates metadata associated with video data, the metadata including information to convert the video data from a 2D image into a 3D image.
  • FIG. 2 illustrates metadata generated by the image generating apparatus 100 illustrated in FIG. 1.
  • the metadata includes information about video data.
  • disc identification information to identify a disc in which the video data is recorded is included in the metadata, though it is understood that the metadata does not include the disc identification information in other embodiments.
  • the disc identification information may include a disc identifier (ID) to identify the disc recorded with the video data and a title ID to identify a title including the video data among a plurality of titles recorded in the disc identified by the disc ID.
  • ID disc identifier
  • the metadata includes information about the frames.
  • the information about the frames may include information to classify the frames according to a predetermined criterion. Assuming that a group of similar frames is a unit, total frames of the video data can be classified as a plurality of units.
  • information to classify the frames of the video data as predetermined units is included in the metadata. Specifically, a group of frames having similar background compositions in which a background composition of a current frame can be predicted by using a previous frame preceding the current frame is classified as a shot.
  • the metadata generating unit 130 classifies the frames of the video data as a predetermined shot and incorporates information about the shot (i.e., shot information) into the metadata. When the background composition of the current frame is different from that of the previous frame due to a significant change in the frame background composition, the current frame and the previous frame are classified as different shots.
  • the shot information includes information about output moments of frames classified within the shot. For example, such information includes output moment information of a frame being output first (shot start moment information in FIG. 2) and output moment information of a frame being output last (shot end moment information in FIG. 2) among the frames classified as each shot, though aspects of the present invention are not limited thereto.
  • the shot information includes the shot start moment information and information on a number of frames included in the shot.
  • the metadata further includes shot type information about frames classified as a shot. The shot type information indicates for each shot whether frames classified as a shot are to be output as a 2D image or a 3D image.
  • the metadata also includes background depth information, which will be described in detail with reference to FIGs. 3A through 3C.
  • FIGs. 3A through 3C are views to explain a depth map generated by using the background depth information.
  • FIG. 3A illustrates a 2D image
  • FIG. 3B illustrates a depth map to be applied to the 2D image illustrated in FIG. 3A
  • FIG. 3C illustrates a result of applying the depth map to the 2D image.
  • a sense of depth is given to the 2D image.
  • an image projected on the screen is formed in each of the user's two eyes.
  • a distance between two points of the images formed in the eyes is called parallax, and the parallax can be classified into positive parallax, zero parallax, and negative parallax.
  • the positive parallax refers to parallax corresponding to a case when the image appears to be formed inside the screen, and the positive parallax is less than or equal to a distance between the eyes. As the positive parallax increases, more cubic effect by which the image appears to lie behind the screen is given. When the image appears to be two-dimensionally formed on the screen plane, a parallax is 0 (i.e., zero parallax). In the case of the zero parallax, the user cannot feel a cubic effect because the image is formed on the screen plane.
  • the negative parallax refers to parallax corresponding to a case when the image appears to lie in front of the screen. This parallax is generated when lines of sight to the user's eyes intersect. The negative parallax gives a cubic effect by which the image appears to protrude forward.
  • a motion of a current frame may be predicted by using a previous frame and the sense of depth may be added to an image of the current frame by using the predicted motion.
  • a depth map for a frame may be generated by using a composition of the frame and the sense of depth may be added to the frame by using the depth map.
  • Metadata includes information to classify frames of video data as predetermined shots.
  • a composition of a current frame cannot be predicted by using a previous frame due to no similarity in composition between the current frame and the previous frame, the current frame and the previous frame are classified as different shots.
  • the metadata includes information about compositions to be applied to frames classified as a shot due to their similarity in composition, and/or includes information about a composition to be applied to each shot.
  • the metadata includes background depth information to indicate a composition of a corresponding frame.
  • the background depth information may include type information of a background included in a frame, coordinate point information of the background, and a depth value of the background corresponding to a coordinate point.
  • the type information of the background may be an ID indicating a composition of the background from among a plurality of compositions.
  • a frame includes a background including the ground and the sky.
  • the horizon where the ground and the sky meet is the farthest point from the perspective of a viewer, and an image corresponding to the bottom portion of the ground is the nearest point from the perspective of the viewer.
  • the image generating apparatus 100 determines that a composition of a type illustrated in FIG. 3B is to be applied to the frame illustrated in FIG. 3A, and generates metadata including type information indicative of the composition illustrated in FIG. 3B for the frame illustrated in FIG. 3A.
  • Coordinate point values refer to values of a coordinate point of a predetermined position in 2D images.
  • a depth value refers to the degree of depth of an image.
  • the depth value may be one of 256 values ranging from 0 to 255. As the depth value decreases, the depth becomes greater and thus an image appears to be farther from a viewer. Conversely, as the depth value increases, an image appears nearer to a viewer. Referring to FIGs. 3B and 3C, it can be seen that a portion where the ground and the sky meets (i.e., the horizon portion) has a smallest depth value and the bottom portion of the ground has a largest depth value in the frame.
  • the image processing apparatus extracts the background depth information included in the metadata, generates the depth map as illustrated in FIG. 3C by using the extracted depth information, and outputs a 2D image as a 3D image by using the depth map.
  • FIG. 4 is a block diagram of an image processing apparatus 400 according to an embodiment of the present invention.
  • the image processing apparatus 400 includes a video data decoding unit 410, a metadata analyzing unit 420, and a 3D image converting unit 430, and an output unit 440 to output a 3D image to a screen.
  • the image processing apparatus 400 need not include the output unit 440 in all embodiments, and/or the output unit 440 may be provided separately from the image processing apparatus 400.
  • the image processing apparatus 400 may be a computer, a mobile device, a set-top box, a workstation, etc.
  • the output unit 440 may be a cathode ray tube device, a liquid crystal display device, a plasma display device, an organic light emitting diode display device, etc. and/or be connected to the same and or connected to goggles through wired and/or wireless protocols.
  • the video data decoding unit 410 reads video data (IN2) from a disc (such as a DVD, Blu-ray, etc.), a local storage, transmitted the image generating device 100 of FIG. 1, or any external storage device (such as a hard disk drive, a flash memory, etc.) and decodes the read video data.
  • the metadata analyzing unit 420 decodes metadata (IN3) to extract information about frames of the read video data from the metadata, and analyzes the extracted information. By using the metadata, the metadata analyzing unit 420 controls a switching unit 433 included in the 3D image converting unit 430 in order to output a frame as a 2D image or a 3D image.
  • the metadata analyzing unit 420 receives the metadata IN3 from a disc, a local storage, transmitted from the image generating device 100 of FIG. 1, or any external storage device (such as a hard disk drive, a flash memory, etc.).
  • the metadata need not be stored with the video data in all aspects of the invention.
  • the 3D image converting unit 430 converts the video data from a 2D image received from the video data decoding unit 410 into a 3D image.
  • the 3D image converting unit 430 estimates a motion of a current frame by using a previous frame in order to generate a 3D image for the current frame.
  • the metadata analyzing unit 420 extracts, from the metadata, output moment information of a frame being output first and/or output moment information of a frame being output last among frames classified as a shot, and determines whether a current frame being currently decoded by the video data decoding unit 410 is classified as a new shot, based on the extracted output moment information.
  • the metadata analyzing unit 420 determines that the current frame is classified as a new shot
  • the metadata analyzing unit 420 controls the switching unit 433 in order to not convert the current frame into a 3D image such that a motion estimating unit 434 does not estimate the motion of the current frame by using a previous frame stored in a previous frame storing unit 432.
  • the switch unit 433 disconnects the storing unit 432 to prevent use of the previous frame, but aspects of the invention are not limited thereto.
  • the metadata includes the shot type information indicating that frames of the video data are to be output as a 2D image.
  • the metadata analyzing unit 420 determines whether the video data is to be output as a 2D image or a 3D image for each shot using the shot type information and controls the switching unit 433 depending on a result of the determination.
  • the metadata analyzing unit 420 determines, based on the shot type information, that video data classified as a predetermined shot does is not to be converted into a 3D image
  • the metadata analyzing unit 420 controls the switching unit 433 such that the 3D image converting unit 430 does not estimate the motion of the current frame by using the previous frame by disconnected the storing unit 432 from the motion estimating unit 434.
  • the metadata analyzing unit 420 determines, based on the shot type information, that video data classified as a predetermined shot is to be converted into a 3D image
  • the metadata analyzing unit 420 controls the switching unit 433 such that the image converting unit 430 converts the current frame into a 3D image by using the previous frame by connecting the storing unit 432 and the motion estimating unit 434.
  • the 3D image converting unit 430 converts the video data being a 2D image received from the video data decoding unit 410 into the 3D image.
  • the 3D image converting unit 430 includes an image block unit 431, the previous frame storing unit 432, the motion estimating unit 434, a block synthesizing unit 435, a left-/right-view image determining unit 436, and the switching unit 433.
  • the image block unit 431 divides a frame of video data, which is a 2D image, into blocks of a predetermined size.
  • the previous frame storing unit 432 stores a predetermined number of previous frames preceding a current frame. Under the control of the metadata analyzing unit 420, the switching unit 433 enables or disables outputting of previous frames stored in the previous frame storing unit 432 to the motion estimating unit 434.
  • the motion estimating unit 434 obtains a per-block motion vector regarding the amount and direction of motion using a block of a current frame and a block of a previous frame.
  • the block synthesizing unit 435 synthesizes blocks selected by using the motion vectors obtained by the motion estimating unit 434 from among predetermined blocks of previous frames in order to generate a new frame.
  • the motion estimating unit 434 outputs the current frame received from the image block unit 431 to the block synthesizing unit 435.
  • the generated new frame or the current frame is input to the left-/right-view image determining unit 436.
  • the left-/right-view image determining unit 436 determines a left-view image and a right-view image by using the frame received from the block synthesizing unit 435 and a frame received from the video data decoding unit 410.
  • the metadata analyzing unit 420 controls the switching unit 433 to not convert video data into a 3D image
  • the left-/right-view image determining unit 436 generates the left-view image and the right-view image that are the same as each other by using the frame with a 2D image received from the block synthesizing unit 435 and the frame with a 2D image received from the video data decoding unit 410.
  • the left-/right-view image determining unit 436 outputs the left-view image and the right-view image to the output unit 440, an external output device, and/or an external terminal (such as a computer, an external display device, a server, etc.).
  • the image processing apparatus 400 further includes the output unit 440 to output the left-view image and the right-view image (OUT2) determined by the left-/right-view image determining unit 436 to the screen alternately at lest every 1/120 second.
  • the image processing apparatus 400 does not convert video data corresponding to a shot change point or video data for which 3D image conversion is not required according to the determination based on the shot information provided in metadata, thereby reducing unnecessary computation and complexity of the apparatus 400.
  • the output image OUT2 can be received at a receiving unit through which a user sees the screen, such as goggles, through wired and/or wireless protocols.
  • FIG. 5 is a block diagram of an image processing apparatus 500 according to another embodiment of the present invention.
  • the image processing apparatus 500 includes a video data decoding unit 510, a metadata analyzing unit 520, a 3D image converting unit 530, and an output unit 540.
  • the image processing apparatus 500 need not include the output unit 540 in all embodiments, and/or the output unit 540 may be provided separately from the image processing apparatus 500.
  • the image processing apparatus 500 may be a computer, a mobile device, a set-top box, a workstation, etc.
  • the output unit 540 may be a cathode ray tube device, a liquid crystal display device, a plasma display device, an organic light emitting diode display device, etc. and/or connected to the same or connected to goggles through wired and/or wireless protocols.
  • each of the units 510, 520, 530 can be one or more processors or processing elements on one or more chips or integrated circuits.
  • the video data decoding unit 510 and the metadata analyzing unit 520 read the video data (IN4) and the metadata (IN5) from the loaded disc.
  • the metadata may be recorded in a lead-in region, a user data region, and/or a lead-out region of the disc.
  • aspects of the present invention are not limited to receiving the video data and the metadata from a disc.
  • the image processing apparatus 500 may further include a communicating unit (not shown) to communicate with an external server or an external terminal (for example, through a communication network and/or any wired/wireless connection).
  • the image processing apparatus 500 may download video data and/or metadata associated therewith from the external server or the external terminal and store the downloaded data in a local storage (not shown).
  • the image processing apparatus 500 may receive the video data and/or metadata from any external storage device different from the disc (for example, a flash memory).
  • the video data decoding unit 510 reads the video data from the disc, the external storage device, the external terminal, or the local storage and decodes the read video data.
  • the metadata analyzing unit 520 reads the metadata associated with the video data from the disc, the external storage device, the external terminal, or the local storage and analyzes the read metadata.
  • the metadata analyzing unit 520 extracts, from the metadata, a disc ID to identify the disc recorded with the video data and a title ID indicating titles including the video data among a plurality of titles in the disc, and determines which video data the metadata is associated with by using the extracted disc ID and title ID.
  • the metadata analyzing unit 520 analyzes the metadata to extract information about frames of the video data classified as a predetermined shot.
  • the metadata analyzing unit 520 determines whether a current frame is video data corresponding to a shot change point (i.e., is classified as a new shot), in order to control a depth map generating unit 531.
  • the metadata analyzing unit 520 determines whether the frames classified as the predetermined shot are to be output as a 2D image or a 3D image by using shot type information, and controls the depth map generating unit 531 according to a result of the determination. Furthermore, the metadata analyzing unit 520 extracts depth information from the metadata and outputs the depth information to the depth map generating unit 531.
  • the stereo rendering unit 533 generates a left-view image and a right-view image by using the video data received from the video data decoding unit 510 and the depth map received from the depth map generating unit 531. Accordingly, the stereo rendering unit 533 generates a 3D-format image including both the generated left-view image and the generated right-view image.
  • a frame received from the depth map generating unit 531 and a frame received from the video data decoding unit 510 are the same as each other, and thus the left-view image and the right-view image generated by the stereo rendering unit 533 are also the same as each other.
  • the 3D format may be a top-and-down format, a side-by-side format, or an interlaced format.
  • the stereo rendering unit 533 outputs the left-view image and the right-view image to the output unit 540, an external output device, and/or an external terminal (such as a computer, an external display device, a server, etc.).
  • the image processing apparatus 500 further includes the output unit 540 that operates as an output device.
  • the output unit 540 sequentially outputs the left-view image and the right-view image received from the stereo rendering unit 533 to the screen.
  • a viewer perceives that an image is sequentially and seamlessly reproduced when the image is output at a frame rate of at least 60 Hz as viewed from a single eye. Therefore, the output unit 540 outputs the screen at a frame rate of at least 120 Hz so that the viewer can perceive that a 3D image is seamlessly reproduced.
  • the output unit 540 sequentially outputs the left-view image and the right-view image (OUT3) included in a frame to the screen at least every 1/120 second. The viewer can have his/her view selectively blocked using goggles to alternate which eye receives the image and/or using polarized light.
  • FIG. 6 is a flowchart illustrating an image processing method according to an embodiment of the present invention.
  • the image processing apparatus 400 or 500 determines whether metadata associated with read video data exists in operation 610. For example, when the video data and metadata are provided on a disc and the disc is loaded and the image processing apparatus 400 or 500 is instructed to output a predetermined title of the loaded disc, the image processing apparatus 400 or 500 determines whether metadata associated with the title exists therein by using a disc ID and a title ID in operation 610. If the image processing apparatus 400 or 500 determines that the disc does not have the metadata therein, the image forming apparatus 400 or 500 may download the metadata from an external server or the like through a communication network in operation 620.
  • existing video such as movies on DVD and Blu-ray discs or computer games
  • the disc could only contain the metadata, and when the metadata for a particular video is selected, the video is downloaded from the server.
  • the image processing apparatus 400 or 500 extracts information about a unit in which the video data is classified from the metadata associated with the video data in operation 630.
  • the information about a unit may be information about a shot (i.e., shot information) in some aspects of the present invention.
  • the shot information indicates whether a current frame is classified as the same shot as a previous frame, and may include shot type information indicating whether the current frame is to be output as a 2D image or a 3D image.
  • the image processing apparatus 400 or 500 determines whether to output frames as a 2D image or a 3D image by using the shot information, and outputs frames classified as a predetermined shot as a 2D image or a 3D image according to a result of the determination in operation 640.
  • FIG. 7 is a flowchart illustrating in detail operation 640 of FIG. 6.
  • the image processing apparatus 400 or 500 when outputting video data, determines whether a current frame has a different composition from a previous frame and is, thus, classified as a new shot in operation 710.
  • the image processing apparatus 400 or 500 determines that the current frame is classified as the new shot, the image processing apparatus 400 or 500 outputs an initial frame included in the new shot as a 2D image without converting the initial frame into a 3D image in operation 720.
  • the image processing apparatus 400 or 500 determines whether to output the remaining frames following the initial frame among total frames classified as the new shot as a 2D image or a 3D image by using shot type information regarding the new shot, provided in metadata, in operation 730.
  • the image processing apparatus 400 or 500 converts the video data classified as the new shot into a 3D image in operation 740.
  • the image processing apparatus 400 or 500 determines a left-view image and a right-view image from the video data converted into the 3D image and the video data being a 2D image and outputs the video data classified as the new shot as a 3D image in operation 740.
  • the image processing apparatus 500 When the image processing apparatus 500 generates a 3D image by using composition information as in FIG. 5, the image processing apparatus 500 extracts background depth information to be applied to a current frame classified as a new shot from metadata and generates a depth map for the current frame by using the background depth information.
  • the image processing apparatus 400 and 500 When the shot type information regarding the new shot indicates that the video data classified as the new shot is to be output as a 2D image (operation 730), the image processing apparatus 400 and 500 outputs the video data as a 2D image without converting the video data into a 3D image in operation 750. The image processing apparatus 400 or 500 determines whether the entire video data has been completely output in operation 760. If not, the image processing apparatus 400 or 500 repeats operation 710.
  • video data can be output as a 2D image at a shot change point.
  • it is determined for each shot whether to output video data as a 2D image or a 3D image and the video data is output according to a result of the determination, thereby reducing the amount of computation that may increase due to conversion of total video data into a 3D image.
  • aspects of the present invention can also be embodied as computer-readable code on a computer-readable recording medium.
  • the computer-readable recording medium is any data storage device that can store data that can be thereafter read by a computer system. Examples of the computer-readable recording medium include read-only memory (ROM), random-access memory (RAM), CD-ROMs, magnetic tapes, floppy disks, and optical data storage devices.
  • the computer-readable recording medium can also be distributed over network-coupled computer systems so that the computer-readable code is stored and executed in a distributed fashion.
  • aspects of the present invention may also be realized as a data signal embodied in a carrier wave and comprising a program readable by a computer and transmittable over the Internet.
  • one or more units of the image processing apparatus 400 and 500 can include a processor or microprocessor executing a computer program stored in a computer-readable medium, such as a local storage (not shown).
  • a processor or microprocessor executing a computer program stored in a computer-readable medium, such as a local storage (not shown).
  • the image generating apparatus 100 and the image processing apparatus 400 or 500 may be provided in a single apparatus in some embodiments.

Landscapes

  • Engineering & Computer Science (AREA)
  • Multimedia (AREA)
  • Signal Processing (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Library & Information Science (AREA)
  • Computer Graphics (AREA)
  • Human Computer Interaction (AREA)
  • Testing, Inspecting, Measuring Of Stereoscopic Televisions And Televisions (AREA)
  • Controls And Circuits For Display Device (AREA)
  • Processing Or Creating Images (AREA)
  • Two-Way Televisions, Distribution Of Moving Picture Or The Like (AREA)
  • Television Signal Processing For Recording (AREA)
  • Apparatus For Radiation Diagnosis (AREA)
  • Measuring And Recording Apparatus For Diagnosis (AREA)
  • Image Processing (AREA)

Abstract

An image processing method and apparatus and an image generating method and apparatus, the image processing method to output a video data being a two-dimensional (2D) image as the 2D image or a three-dimensional (3D) image including: extracting information about the video data from metadata associated with the video data; and outputting the video data as the 2D image or the 3D image by using the extracted information about the video data.

Description

IMAGE GENERATING METHOD AND APPARATUS AND IMAGE PROCESSING METHOD AND APPARATUS Technical Field
Aspects of the present invention generally relate to an image generating method and apparatus and an image processing method and apparatus, and more particularly, to an image generating method and apparatus and an image processing method and apparatus in which video data is output as a two-dimensional (2D) image or a three-dimensional (3D) image by using metadata associated with the video data.
Background Art
With the development of digital technology, three-dimensional (3D) image technology has widely spread. The 3D image technology expresses a more realistic image by adding depth information to a two-dimensional (2D) image. The 3D image technology can be classified into technology to generate video data as a 3D image and technology to convert video data generated as a 2D image into a 3D image. Both technologies have been studied together.
Technical Solution
Aspects of the present invention provide an image processing method and apparatus to output video data as a two-dimensional image or a three-dimensional image by using metadata associated with the video data can be provided.
Advantageous Effects
In this way, according to aspects of the present invention, by using shot information included in metadata, video data can be output as a 2D image at a shot change point. Moreover, according to an embodiment of the present invention, it is determined for each shot whether to output video data as a 2D image or a 3D image and the video data is output according to a result of the determination, thereby reducing the amount of computation that may increase due to conversion of total video data into a 3D image.
Description of Drawings
FIG. 1 is a block diagram of an image generating apparatus according to an embodiment of the present invention;
FIG. 2 illustrates metadata generated by the image generating apparatus illustrated in FIG. 1;
FIGs. 3A through 3C are views to explain a depth map generated by using background depth information;
FIG. 4 is a block diagram of an image processing apparatus according to an embodiment of the present invention;
FIG. 5 is a block diagram of an image processing apparatus according to another embodiment of the present invention;
FIG. 6 is a flowchart illustrating an image processing method according to an embodiment of the present invention; and
FIG. 7 is a flowchart illustrating in detail an operation illustrated in FIG. 6 where video data is output as a two-dimensional (2D) image or a three-dimensional (3D) image.
Best Mode
According to an aspect of the present invention, there is provided an image processing method to output video data being a two-dimensional (2D) image as the 2D image or a three-dimensional (3D) image, the image processing method including: extracting information about the video data from metadata associated with the video data; and outputting the video data as the 2D image or the 3D image by using the extracted information about the video data, wherein the information about the video data includes information to classify frames of the video data into predetermined units.
According to an aspect of the present invention, the information to classify the frames of the video data as the predetermined units may be shot information to classify a group of frames in which a background composition of a current frame is predictable by using a previous frame preceding the current frame as a shot.
According to an aspect of the present invention, the shot information may include output moment information of a frame being output first and output moment information of a frame being output last from among the group of frames classified as the shot.
According to an aspect of the present invention, the metadata may include shot type information indicating whether the frames classified as the shot are to be output as the 2D image or the 3D image, and the outputting of the video data may include outputting the frames classified as the shot as the 2D image or the 3D image by using the shot type information.
According to an aspect of the present invention, the outputting of the video data may include determining, by using the metadata, whether a background composition of a current frame is not predictable by using a previous frame preceding the current frame and thus the current frame is classified as a new shot, outputting the current frame as the 2D image when the current frame is classified as the new shot, and converting the remaining frames of the frames classified as the new shot into the 3D image and outputting the converted 3D image.
According to an aspect of the present invention, the outputting of the video data may include determining, by using the metadata, whether a background composition of a current frame is not predictable by using a previous frame preceding the current frame and thus the current frame is classified as a new shot, extracting background depth information to be applied to the current frame classified as the new shot from the metadata when the current frame is classified as the new shot, and generating a depth map for the current frame by using the background depth information.
According to an aspect of the present invention, the generating of the depth map for the current frame may include generating the depth map for a background of the current frame by using coordinate point values of the background of the current frame, depth values corresponding to the coordinate point values, and a panel position value, in which the coordinate point values, the depth value, and the panel position value are included in the background depth information.
According to an aspect of the present invention, the image processing method may further include reading the metadata from a disc recorded with the video data or downloading the metadata from a server through a communication network.
According to an aspect of the present invention, the metadata may include identification information to identify the video data, and the identification information may include a disc identifier (ID) to identify a disc recorded with the video data and a title ID to indicate a title including the video data among a plurality of titles recorded in the disc identified by the disc ID.
According to another aspect of the present invention, there is provided an image generating method including: receiving video data being a two-dimensional (2D) image; and generating metadata associated with the video data, the metadata including information to classify frames of the video data as predetermined units, wherein the information to classify the frames of the video data as the predetermined units is shot information to classify a group of frames in which a background composition of a current frame is predictable by using a previous frame preceding the current frame as a shot.
According to an aspect of the present invention, the shot information may include output moment information of a frame being output first and output moment information of a frame being output last from among the frames classified as the shot, and/or may include shot type information indicating whether the frames classified as the shot are to be output as the 2D image or a three-dimensional (3D) image.
According to an aspect of the present invention, the metadata may include background depth information for frames classified as a predetermined shot and the background depth information may include coordinate point values of a background of the frame classified as the predetermined shot, depth values corresponding to the coordinate point values, and a panel position value.
According to another aspect of the present invention, there is provided an image processing apparatus to output video data being a two-dimensional (2D) image as the 2D image or a three-dimensional (3D) image, the image processing apparatus including: a metadata analyzing unit to determine whether the video data is to be output as the 2D image or the 3D image by using metadata associated with the video data; a 3D image converting unit to convert the video data into the 3D image when the video data is to be output as the 3D image; and an output unit to output the video data as the 2D image or the 3D image, wherein the metadata includes information to classify frames of the video data into predetermined units.
According to an aspect of the present invention, the information to classify the frames of the video data into the predetermined units may be shot information to classify a group of frames in which a background composition of a current frame is predictable by using a previous frame preceding the current frame as a shot.
According to an aspect of the present invention, the shot information may include output moment information of a frame being output first and output moment information of a frame being output last from among the frames classified as the shot.
According to an aspect of the present invention, the metadata may include shot type information indicating whether the frames classified as the shot are to be output as the 2D image or the 3D image.
According to an aspect of the present invention, the metadata may include background depth information for a frame classified as a predetermined shot, and the background depth information may include coordinate point values of a background of the frame classified as the predetermined shot, depth values corresponding to the coordinate point values, and a panel position value.
According to another aspect of the present invention, there is provided an image generating apparatus including: a video data encoding unit to encode video data being a two-dimensional (2D) image; a metadata generating unit to generate metadata associated with the video data, the metadata including information to classify frames of the video data into predetermined units; and a metadata encoding unit to encode the metadata, in which the information to classify the frames of the video data into the predetermined units is shot information to classify a group of frames in which a background composition of a current frame is predictable by using a previous frame preceding the current frame as a shot.
According to yet another aspect of the present invention, there is provided a computer-readable information storage medium including video data being a two-dimensional (2D) image and metadata associated with the video data, the metadata including information to classify frames of the video data into predetermined units, wherein the information to classify the frames of the video data into the predetermined units is shot information to classify a group of frames in which a background composition of a current frame is predictable by using a previous frame preceding the current frame as a shot.
According to still another aspect of the present invention, there is provided a computer-readable information storage medium having recorded thereon a program to execute an image processing method to output video data being a two-dimensional (2D) image as the 2D image or a three-dimensional (3D) image, the image processing method including: extracting information about the video data from metadata associated with the video data; and outputting the video data as the 2D image or the 3D image by using the extracted information about the video data, wherein the information about the video data includes information to classify frames of the video data into predetermined units.
According to an aspect of the present invention, there is provided a system to output video data as a two-dimensional (2D) image or a three-dimensional (3D) image, the system including: an image generating apparatus including: a video data encoding unit to encode the video data being the 2D image, a metadata generating unit to generate metadata associated with the video data, the metadata comprising information to classify frames of the video data as predetermined units and used to determine whether each of the classified frames is to be converted to the 3D image; and an image processing apparatus to receive the encoded video data and the generated metadata, and to output the video data as the 2D image or the 3D image, the image processing apparatus including: a metadata analyzing unit to determine whether the video data is to be output as the 2D image or the 3D image by using the information to classify the frames of the video data comprised in the received metadata associated with the video data, a 3D image converting unit to convert the video data into the 3D image when the metadata analyzing unit determines that the video data is to be output as the 3D image, and an output unit to output the video data as the 2D image or the 3D image according to the determination of the metadata analyzing unit, wherein the information to classify the frames of the video data as the predetermined units is shot information to classify a group of frames in which a background composition of a current frame is predictable by using a previous frame preceding the current frame in the group of frames as a shot.
According to another aspect of the present invention, there is provided a computer-readable information storage medium including: metadata associated with video data comprising two-dimensional (2D) frames, the metadata comprising information used by an image processing apparatus to classify the frames of the video data as predetermined units and used by the image processing apparatus to determine whether each of the classified frames is to be converted by the image processing apparatus to a three-dimensional (3D) image, wherein the information to classify the frames of the video data as the predetermined units comprises shot information to classify, as a shot, a group of frames in which a background composition of a current frame is predictable by using a previous frame preceding the current frame in the group of frames.
According to another aspect of the present invention, there is provided an image processing method to output video data having two-dimensional (2D) images as the 2D images or three-dimensional (3D) images, the image processing method including: determining, by an image processing apparatus, whether metadata associated with the video data exists on a disc comprising the video data; reading, by the image processing apparatus, the metadata from the disc if the metadata is determined to exist on the disc; retrieving, by the image processing apparatus, the metadata from a server if the metadata is determined to not exist on the disc; and outputting, by the image processing apparatus, the video data as selectable between the 2D image and the 3D image according to the metadata.
Additional aspects and/or advantages of the invention will be set forth in part in the description which follows and, in part, will be obvious from the description, or may be learned by practice of the invention.
Mode for Invention
This application claims the benefit of U.S. Provisional Application No. 61/075,184, filed on June 24, 2008 in the U.S. Patent and Trademark Office, and the benefit of Korean Patent Application No. 10-2008-0091269, filed on September 17, 2008 in the Korean Intellectual Property Office, the disclosures of which are incorporated herein in their entirety by reference.
Reference will now be made in detail to the present embodiments of the present invention, examples of which are illustrated in the accompanying drawings, wherein like reference numerals refer to the like elements throughout. The embodiments are described below in order to explain the present invention by referring to the figures.
FIG. 1 is a block diagram of an image generating apparatus 100 according to an embodiment of the present invention. Referring to FIG. 1, the image generating apparatus 100 includes a video data generating unit 110, a video data encoding unit 120, a metadata generating unit 130, a metadata encoding unit 140, and a multiplexing unit 150. The video data generating unit 110 generates video data and outputs the generated video data to the video data encoding unit 120. The video data encoding unit 120 encodes the input video data and outputs the encoded video data (OUT1) to the multiplexing unit 150, and/or to an image processing apparatus (not shown) through a communication network, though it is understood that the video data encoding unit 120 may output the encoded video data to the image processing apparatus through any wired and/or wireless connection (such as IEEE 1394, universal serial bus, a Bluetooth, an infrared, etc.). The image generating apparatus 100 may be a computer, a workstation, a camera device, a mobile device, a stand-alone device, etc. Moreover, while not required, each of the units 110, 120, 130, 140, 150 can be one or more processors or processing elements on one or more chips or integrated circuits.
The metadata generating unit 130 analyzes the video data generated by the video data generating unit 110 to generate metadata including information about frames of the video data. The metadata includes information to convert the generated video data from a two-dimensional (2D) image into a three-dimensional (3D) image. The metadata also includes information to classify the frames of the video data as predetermined units. The metadata generated by the metadata generating unit 130 will be described in more detail with reference to FIG. 2. The metadata generating unit 130 outputs the generated metadata to the metadata encoding unit 140.
The metadata encoding unit 140 encodes the input metadata and outputs the encoded metadata (OUT3) to the multiplexing unit 150 and/or to the image processing apparatus. The multiplexing unit 150 multiplexes the encoded video data (OUT1) and the encoded metadata (OUT3) and transmits the multiplexing result (OUT2) to the image processing apparatus through a wired and/or wireless communication network, or any wired and/or wireless connection, as described above. The metadata encoding unit 140 may transmit the encoded metadata (OUT3), separately from the encoded video data (OUT1), to the image processing apparatus, instead of to or in addition to the multiplexing unit 150. In this way, the image generating apparatus 100 generates metadata associated with video data, the metadata including information to convert the video data from a 2D image into a 3D image.
FIG. 2 illustrates metadata generated by the image generating apparatus 100 illustrated in FIG. 1. The metadata includes information about video data. In order to indicate with which video data the information included in the metadata is associated, disc identification information to identify a disc in which the video data is recorded is included in the metadata, though it is understood that the metadata does not include the disc identification information in other embodiments. The disc identification information may include a disc identifier (ID) to identify the disc recorded with the video data and a title ID to identify a title including the video data among a plurality of titles recorded in the disc identified by the disc ID.
Since the video data has a series of frames, the metadata includes information about the frames. The information about the frames may include information to classify the frames according to a predetermined criterion. Assuming that a group of similar frames is a unit, total frames of the video data can be classified as a plurality of units. In the present embodiment, information to classify the frames of the video data as predetermined units is included in the metadata. Specifically, a group of frames having similar background compositions in which a background composition of a current frame can be predicted by using a previous frame preceding the current frame is classified as a shot. The metadata generating unit 130 classifies the frames of the video data as a predetermined shot and incorporates information about the shot (i.e., shot information) into the metadata. When the background composition of the current frame is different from that of the previous frame due to a significant change in the frame background composition, the current frame and the previous frame are classified as different shots.
The shot information includes information about output moments of frames classified within the shot. For example, such information includes output moment information of a frame being output first (shot start moment information in FIG. 2) and output moment information of a frame being output last (shot end moment information in FIG. 2) among the frames classified as each shot, though aspects of the present invention are not limited thereto. For example, according to other aspects, the shot information includes the shot start moment information and information on a number of frames included in the shot. The metadata further includes shot type information about frames classified as a shot. The shot type information indicates for each shot whether frames classified as a shot are to be output as a 2D image or a 3D image. The metadata also includes background depth information, which will be described in detail with reference to FIGs. 3A through 3C.
FIGs. 3A through 3C are views to explain a depth map generated by using the background depth information. FIG. 3A illustrates a 2D image, FIG. 3B illustrates a depth map to be applied to the 2D image illustrated in FIG. 3A, and FIG. 3C illustrates a result of applying the depth map to the 2D image. In order to add a cubic effect to a 2D image, a sense of depth is given to the 2D image. When a user sees a screen, an image projected on the screen is formed in each of the user's two eyes. A distance between two points of the images formed in the eyes is called parallax, and the parallax can be classified into positive parallax, zero parallax, and negative parallax. The positive parallax refers to parallax corresponding to a case when the image appears to be formed inside the screen, and the positive parallax is less than or equal to a distance between the eyes. As the positive parallax increases, more cubic effect by which the image appears to lie behind the screen is given. When the image appears to be two-dimensionally formed on the screen plane, a parallax is 0 (i.e., zero parallax). In the case of the zero parallax, the user cannot feel a cubic effect because the image is formed on the screen plane. The negative parallax refers to parallax corresponding to a case when the image appears to lie in front of the screen. This parallax is generated when lines of sight to the user's eyes intersect. The negative parallax gives a cubic effect by which the image appears to protrude forward.
In order to generate a 3D image by adding the sense of depth to a 2D image, a motion of a current frame may be predicted by using a previous frame and the sense of depth may be added to an image of the current frame by using the predicted motion. For the same purpose, a depth map for a frame may be generated by using a composition of the frame and the sense of depth may be added to the frame by using the depth map. The former will be described in detail with reference to FIG. 4, and the latter will be described in detail with reference to FIG. 5.
As stated previously, metadata includes information to classify frames of video data as predetermined shots. When a composition of a current frame cannot be predicted by using a previous frame due to no similarity in composition between the current frame and the previous frame, the current frame and the previous frame are classified as different shots. The metadata includes information about compositions to be applied to frames classified as a shot due to their similarity in composition, and/or includes information about a composition to be applied to each shot.
Background compositions of frames may vary. The metadata includes background depth information to indicate a composition of a corresponding frame. The background depth information may include type information of a background included in a frame, coordinate point information of the background, and a depth value of the background corresponding to a coordinate point. The type information of the background may be an ID indicating a composition of the background from among a plurality of compositions.
Referring to FIG. 3A, a frame includes a background including the ground and the sky. In this frame, the horizon where the ground and the sky meet is the farthest point from the perspective of a viewer, and an image corresponding to the bottom portion of the ground is the nearest point from the perspective of the viewer. The image generating apparatus 100 determines that a composition of a type illustrated in FIG. 3B is to be applied to the frame illustrated in FIG. 3A, and generates metadata including type information indicative of the composition illustrated in FIG. 3B for the frame illustrated in FIG. 3A.
Coordinate point values refer to values of a coordinate point of a predetermined position in 2D images. A depth value refers to the degree of depth of an image. In aspects of the present invention, the depth value may be one of 256 values ranging from 0 to 255. As the depth value decreases, the depth becomes greater and thus an image appears to be farther from a viewer. Conversely, as the depth value increases, an image appears nearer to a viewer. Referring to FIGs. 3B and 3C, it can be seen that a portion where the ground and the sky meets (i.e., the horizon portion) has a smallest depth value and the bottom portion of the ground has a largest depth value in the frame. The image processing apparatus (not shown) extracts the background depth information included in the metadata, generates the depth map as illustrated in FIG. 3C by using the extracted depth information, and outputs a 2D image as a 3D image by using the depth map.
FIG. 4 is a block diagram of an image processing apparatus 400 according to an embodiment of the present invention. Referring to FIG. 4, the image processing apparatus 400 includes a video data decoding unit 410, a metadata analyzing unit 420, and a 3D image converting unit 430, and an output unit 440 to output a 3D image to a screen. However, it is understood that the image processing apparatus 400 need not include the output unit 440 in all embodiments, and/or the output unit 440 may be provided separately from the image processing apparatus 400. Moreover, the image processing apparatus 400 may be a computer, a mobile device, a set-top box, a workstation, etc. The output unit 440 may be a cathode ray tube device, a liquid crystal display device, a plasma display device, an organic light emitting diode display device, etc. and/or be connected to the same and or connected to goggles through wired and/or wireless protocols.
The video data decoding unit 410 reads video data (IN2) from a disc (such as a DVD, Blu-ray, etc.), a local storage, transmitted the image generating device 100 of FIG. 1, or any external storage device (such as a hard disk drive, a flash memory, etc.) and decodes the read video data. The metadata analyzing unit 420 decodes metadata (IN3) to extract information about frames of the read video data from the metadata, and analyzes the extracted information. By using the metadata, the metadata analyzing unit 420 controls a switching unit 433 included in the 3D image converting unit 430 in order to output a frame as a 2D image or a 3D image. The metadata analyzing unit 420 receives the metadata IN3 from a disc, a local storage, transmitted from the image generating device 100 of FIG. 1, or any external storage device (such as a hard disk drive, a flash memory, etc.). The metadata need not be stored with the video data in all aspects of the invention.
The 3D image converting unit 430 converts the video data from a 2D image received from the video data decoding unit 410 into a 3D image. In FIG. 4, the 3D image converting unit 430 estimates a motion of a current frame by using a previous frame in order to generate a 3D image for the current frame.
The metadata analyzing unit 420 extracts, from the metadata, output moment information of a frame being output first and/or output moment information of a frame being output last among frames classified as a shot, and determines whether a current frame being currently decoded by the video data decoding unit 410 is classified as a new shot, based on the extracted output moment information. When the metadata analyzing unit 420 determines that the current frame is classified as a new shot, the metadata analyzing unit 420 controls the switching unit 433 in order to not convert the current frame into a 3D image such that a motion estimating unit 434 does not estimate the motion of the current frame by using a previous frame stored in a previous frame storing unit 432. This is because motion information of a current frame is extracted by referring to a previous frame in order to convert video data from a 2D image into a 3D image. However, if the current frame and the previous frame are classified as different shots, the current frame and the previous frame do not have sufficient similarity therebetween, and thus a composition of the current frame cannot be predicted by using the previous frame. As shown, the switch unit 433 disconnects the storing unit 432 to prevent use of the previous frame, but aspects of the invention are not limited thereto.
When the video data is not to be converted into a 3D image (for example, when the video data is a warning sentence, a menu screen, an ending credit, etc.), the metadata includes the shot type information indicating that frames of the video data are to be output as a 2D image. The metadata analyzing unit 420 determines whether the video data is to be output as a 2D image or a 3D image for each shot using the shot type information and controls the switching unit 433 depending on a result of the determination. Specifically, when the metadata analyzing unit 420 determines, based on the shot type information, that video data classified as a predetermined shot does is not to be converted into a 3D image, the metadata analyzing unit 420 controls the switching unit 433 such that the 3D image converting unit 430 does not estimate the motion of the current frame by using the previous frame by disconnected the storing unit 432 from the motion estimating unit 434. When the metadata analyzing unit 420 determines, based on the shot type information, that video data classified as a predetermined shot is to be converted into a 3D image, the metadata analyzing unit 420 controls the switching unit 433 such that the image converting unit 430 converts the current frame into a 3D image by using the previous frame by connecting the storing unit 432 and the motion estimating unit 434.
When the video data is classified as a predetermined shot and is to be output as a 3D image, the 3D image converting unit 430 converts the video data being a 2D image received from the video data decoding unit 410 into the 3D image. The 3D image converting unit 430 includes an image block unit 431, the previous frame storing unit 432, the motion estimating unit 434, a block synthesizing unit 435, a left-/right-view image determining unit 436, and the switching unit 433. The image block unit 431 divides a frame of video data, which is a 2D image, into blocks of a predetermined size. The previous frame storing unit 432 stores a predetermined number of previous frames preceding a current frame. Under the control of the metadata analyzing unit 420, the switching unit 433 enables or disables outputting of previous frames stored in the previous frame storing unit 432 to the motion estimating unit 434.
The motion estimating unit 434 obtains a per-block motion vector regarding the amount and direction of motion using a block of a current frame and a block of a previous frame. The block synthesizing unit 435 synthesizes blocks selected by using the motion vectors obtained by the motion estimating unit 434 from among predetermined blocks of previous frames in order to generate a new frame. When the motion estimating unit 434 does not use a previous frame due to the control of the switching unit 433 by the metadata analyzing unit 420, the motion estimating unit 434 outputs the current frame received from the image block unit 431 to the block synthesizing unit 435.
The generated new frame or the current frame is input to the left-/right-view image determining unit 436. The left-/right-view image determining unit 436 determines a left-view image and a right-view image by using the frame received from the block synthesizing unit 435 and a frame received from the video data decoding unit 410. When the metadata analyzing unit 420 controls the switching unit 433 to not convert video data into a 3D image, the left-/right-view image determining unit 436 generates the left-view image and the right-view image that are the same as each other by using the frame with a 2D image received from the block synthesizing unit 435 and the frame with a 2D image received from the video data decoding unit 410. The left-/right-view image determining unit 436 outputs the left-view image and the right-view image to the output unit 440, an external output device, and/or an external terminal (such as a computer, an external display device, a server, etc.).
The image processing apparatus 400 further includes the output unit 440 to output the left-view image and the right-view image (OUT2) determined by the left-/right-view image determining unit 436 to the screen alternately at lest every 1/120 second. As such, by using the shot information included in the metadata, the image processing apparatus 400 according to an embodiment of the present invention does not convert video data corresponding to a shot change point or video data for which 3D image conversion is not required according to the determination based on the shot information provided in metadata, thereby reducing unnecessary computation and complexity of the apparatus 400. While not required, the output image OUT2 can be received at a receiving unit through which a user sees the screen, such as goggles, through wired and/or wireless protocols.
FIG. 5 is a block diagram of an image processing apparatus 500 according to another embodiment of the present invention. Referring to FIG. 5, the image processing apparatus 500 includes a video data decoding unit 510, a metadata analyzing unit 520, a 3D image converting unit 530, and an output unit 540. However, it is understood that the image processing apparatus 500 need not include the output unit 540 in all embodiments, and/or the output unit 540 may be provided separately from the image processing apparatus 500. Moreover, the image processing apparatus 500 may be a computer, a mobile device, a set-top box, a workstation, etc. The output unit 540 may be a cathode ray tube device, a liquid crystal display device, a plasma display device, an organic light emitting diode display device, etc. and/or connected to the same or connected to goggles through wired and/or wireless protocols. Moreover, while not required, each of the units 510, 520, 530 can be one or more processors or processing elements on one or more chips or integrated circuits.
When video data that is a 2D image and metadata associated with the video data are recorded in a disc (not shown) in a multiplexed state or separately from each other, upon loading of the disc recorded with the video data and the metadata into the image processing apparatus 500, the video data decoding unit 510 and the metadata analyzing unit 520 read the video data (IN4) and the metadata (IN5) from the loaded disc. The metadata may be recorded in a lead-in region, a user data region, and/or a lead-out region of the disc. However, it is understood that aspects of the present invention are not limited to receiving the video data and the metadata from a disc. For example, according to other aspects, the image processing apparatus 500 may further include a communicating unit (not shown) to communicate with an external server or an external terminal (for example, through a communication network and/or any wired/wireless connection). The image processing apparatus 500 may download video data and/or metadata associated therewith from the external server or the external terminal and store the downloaded data in a local storage (not shown). Furthermore, the image processing apparatus 500 may receive the video data and/or metadata from any external storage device different from the disc (for example, a flash memory).
The video data decoding unit 510 reads the video data from the disc, the external storage device, the external terminal, or the local storage and decodes the read video data. The metadata analyzing unit 520 reads the metadata associated with the video data from the disc, the external storage device, the external terminal, or the local storage and analyzes the read metadata. When the video data is recorded in the disc, the metadata analyzing unit 520 extracts, from the metadata, a disc ID to identify the disc recorded with the video data and a title ID indicating titles including the video data among a plurality of titles in the disc, and determines which video data the metadata is associated with by using the extracted disc ID and title ID.
The metadata analyzing unit 520 analyzes the metadata to extract information about frames of the video data classified as a predetermined shot. The metadata analyzing unit 520 determines whether a current frame is video data corresponding to a shot change point (i.e., is classified as a new shot), in order to control a depth map generating unit 531. The metadata analyzing unit 520 determines whether the frames classified as the predetermined shot are to be output as a 2D image or a 3D image by using shot type information, and controls the depth map generating unit 531 according to a result of the determination. Furthermore, the metadata analyzing unit 520 extracts depth information from the metadata and outputs the depth information to the depth map generating unit 531.
The 3D image converting unit 530 generates a 3D image for video data. The 3D image converting unit 530 includes the depth map generating unit 531 and a stereo rendering unit 533. The depth map generating unit 531 generates a depth map for a frame by using the background depth information received from the metadata analyzing unit 520. The background depth information includes coordinate point values of a background included in a current frame, a depth value corresponding to the coordinate point values, and a panel position value that represents a depth value of the screen on which an image is output. The depth map generating unit 531 generates a depth map for the background of the current frame by using the background depth information and outputs the generated depth map to the stereo rendering unit 533. However, when the current frame is to be output as a 2D image, the depth map generating unit 531 outputs the current frame to the stereo rendering unit 533 without generating the depth map for the current frame.
The stereo rendering unit 533 generates a left-view image and a right-view image by using the video data received from the video data decoding unit 510 and the depth map received from the depth map generating unit 531. Accordingly, the stereo rendering unit 533 generates a 3D-format image including both the generated left-view image and the generated right-view image. When the current frame is to be output as a 2D image, a frame received from the depth map generating unit 531 and a frame received from the video data decoding unit 510 are the same as each other, and thus the left-view image and the right-view image generated by the stereo rendering unit 533 are also the same as each other. The 3D format may be a top-and-down format, a side-by-side format, or an interlaced format. The stereo rendering unit 533 outputs the left-view image and the right-view image to the output unit 540, an external output device, and/or an external terminal (such as a computer, an external display device, a server, etc.).
In the present embodiment, the image processing apparatus 500 further includes the output unit 540 that operates as an output device. In this case, the output unit 540 sequentially outputs the left-view image and the right-view image received from the stereo rendering unit 533 to the screen. A viewer perceives that an image is sequentially and seamlessly reproduced when the image is output at a frame rate of at least 60 Hz as viewed from a single eye. Therefore, the output unit 540 outputs the screen at a frame rate of at least 120 Hz so that the viewer can perceive that a 3D image is seamlessly reproduced. Accordingly, the output unit 540 sequentially outputs the left-view image and the right-view image (OUT3) included in a frame to the screen at least every 1/120 second. The viewer can have his/her view selectively blocked using goggles to alternate which eye receives the image and/or using polarized light.
FIG. 6 is a flowchart illustrating an image processing method according to an embodiment of the present invention. Referring to FIG. 6, the image processing apparatus 400 or 500 determines whether metadata associated with read video data exists in operation 610. For example, when the video data and metadata are provided on a disc and the disc is loaded and the image processing apparatus 400 or 500 is instructed to output a predetermined title of the loaded disc, the image processing apparatus 400 or 500 determines whether metadata associated with the title exists therein by using a disc ID and a title ID in operation 610. If the image processing apparatus 400 or 500 determines that the disc does not have the metadata therein, the image forming apparatus 400 or 500 may download the metadata from an external server or the like through a communication network in operation 620. In this manner, existing video (such as movies on DVD and Blu-ray discs or computer games) can become 3D by merely downloading the corresponding metadata. Alternatively, the disc could only contain the metadata, and when the metadata for a particular video is selected, the video is downloaded from the server.
The image processing apparatus 400 or 500 extracts information about a unit in which the video data is classified from the metadata associated with the video data in operation 630. As previously described, the information about a unit may be information about a shot (i.e., shot information) in some aspects of the present invention. The shot information indicates whether a current frame is classified as the same shot as a previous frame, and may include shot type information indicating whether the current frame is to be output as a 2D image or a 3D image. The image processing apparatus 400 or 500 determines whether to output frames as a 2D image or a 3D image by using the shot information, and outputs frames classified as a predetermined shot as a 2D image or a 3D image according to a result of the determination in operation 640.
FIG. 7 is a flowchart illustrating in detail operation 640 of FIG. 6. Referring to FIG. 7, the image processing apparatus 400 or 500, when outputting video data, determines whether a current frame has a different composition from a previous frame and is, thus, classified as a new shot in operation 710. When the image processing apparatus 400 or 500 determines that the current frame is classified as the new shot, the image processing apparatus 400 or 500 outputs an initial frame included in the new shot as a 2D image without converting the initial frame into a 3D image in operation 720.
The image processing apparatus 400 or 500 determines whether to output the remaining frames following the initial frame among total frames classified as the new shot as a 2D image or a 3D image by using shot type information regarding the new shot, provided in metadata, in operation 730. When the shot type information regarding the new shot indicates that video data classified as the new shot is to be output as a 3D image, the image processing apparatus 400 or 500 converts the video data classified as the new shot into a 3D image in operation 740. Specifically, the image processing apparatus 400 or 500 determines a left-view image and a right-view image from the video data converted into the 3D image and the video data being a 2D image and outputs the video data classified as the new shot as a 3D image in operation 740. When the image processing apparatus 500 generates a 3D image by using composition information as in FIG. 5, the image processing apparatus 500 extracts background depth information to be applied to a current frame classified as a new shot from metadata and generates a depth map for the current frame by using the background depth information.
When the shot type information regarding the new shot indicates that the video data classified as the new shot is to be output as a 2D image (operation 730), the image processing apparatus 400 and 500 outputs the video data as a 2D image without converting the video data into a 3D image in operation 750. The image processing apparatus 400 or 500 determines whether the entire video data has been completely output in operation 760. If not, the image processing apparatus 400 or 500 repeats operation 710.
In this way, according to aspects of the present invention, by using shot information included in metadata, video data can be output as a 2D image at a shot change point. Moreover, according to an embodiment of the present invention, it is determined for each shot whether to output video data as a 2D image or a 3D image and the video data is output according to a result of the determination, thereby reducing the amount of computation that may increase due to conversion of total video data into a 3D image.
While not restricted thereto, aspects of the present invention can also be embodied as computer-readable code on a computer-readable recording medium. The computer-readable recording medium is any data storage device that can store data that can be thereafter read by a computer system. Examples of the computer-readable recording medium include read-only memory (ROM), random-access memory (RAM), CD-ROMs, magnetic tapes, floppy disks, and optical data storage devices. The computer-readable recording medium can also be distributed over network-coupled computer systems so that the computer-readable code is stored and executed in a distributed fashion. Aspects of the present invention may also be realized as a data signal embodied in a carrier wave and comprising a program readable by a computer and transmittable over the Internet. Moreover, while not required in all aspects, one or more units of the image processing apparatus 400 and 500 can include a processor or microprocessor executing a computer program stored in a computer-readable medium, such as a local storage (not shown). Furthermore, it is understood that the image generating apparatus 100 and the image processing apparatus 400 or 500 may be provided in a single apparatus in some embodiments.
Although a few embodiments of the present invention have been shown and described, it would be appreciated by those skilled in the art that changes may be made in this embodiment without departing from the principles and spirit of the invention, the scope of which is defined in the claims and their equivalents.

Claims (15)

1. An image processing method to output video data having two-dimensional (2D) images as the 2D images or three-dimensional (3D) images, the image processing method comprises:
extracting, by an image processing apparatus, information about the video data from metadata associated with the video data; and
outputting, by the image processing apparatus, the video data as selectable between the 2D image and the 3D image according to the extracted information about the video data,
wherein the information about the video data includes information to classify frames of the video data into predetermined units.
2. The image processing method as claimed in claim 1, wherein the information to classify the frames of the video data into the predetermined units is shot information to classify, as a shot, a group of frames in which a background composition of a current frame is predictable by using a previous frame preceding the current frame in the group of frames.
3. The image processing method as claimed in claim 2, wherein the shot information comprises output moment information of a frame being output first from among the group of frames classified as the shot and/or output moment information of a frame being output last from among the group of frames classified as the shot.
4. The image processing method as claimed in claim 2, wherein:
the metadata comprises shot type information indicating whether the group of frames classified as the shot are to be output as the 2D image or the 3D image; and
the outputting of the video data comprises outputting the group of frames classified as the shot as the 2D image or the 3D image according to the shot type information.
5. The image processing method as claimed in claim 2, wherein the outputting of the video data comprises:
according to the metadata, determining that a current frame is classified as a new shot as compared to a previous frame preceding the current frame when a background composition of the current frame is not predictable by using the previous frame;
when the current frame is classified as the new shot, outputting the current frame as the 2D image; and
converting other frames of a group of frames classified as the new shot into the 3D image and outputting the converted 3D image.
6. The image processing method as claimed in claim 2, wherein the outputting of the video data comprises:
according to the metadata, determining that a current frame is classified as a new shot as compared to a previous frame preceding the current frame when a background composition of the current frame is not predictable by using the previous frame;
when the current frame is classified as the new shot, extracting background depth information to be applied to the current frame classified as the new shot from the metadata; and
when the current frame is classified as the new shot, generating a depth map for the current frame by using the background depth information.
7. The image processing method as claimed in claim 6, wherein:
the background depth information comprises coordinate point values of a background of the current frame, depth values respectively corresponding to the coordinate point values, and a panel position value; and
the generating of the depth map for the current frame comprises generating the depth map for the background of the current frame by using the coordinate point values, the depth values, and the panel position value that represents a depth value of an output screen.
8. The image processing method as claimed in claim 1, further comprising reading the metadata from a disc recorded with the video data or downloading the metadata from a server through a communication network.
9. The image processing method as claimed in claim 1, wherein the metadata comprises identification information to identify the video data, and the identification information comprises a disc identifier (ID) to identify a disc recorded with the video data and a title ID to indicate a title including the video data from among a plurality of titles recorded in the disc identified by the disc ID.
10. An image generating method comprising:
receiving, by an image generating apparatus, video data as two-dimensional (2D) images; and
generating, by the image generating apparatus, metadata associated with the video data, the metadata comprising information to classify frames of the video data as predetermined units and used to determine whether each of the classified frames is to be converted to a three-dimensional (3D) image,
wherein the information to classify the frames of the video data as the predetermined units comprises shot information to classify a group of frames, as a shot, in which a background composition of a current frame is predictable by using a previous frame preceding the current frame in the group of frames.
11. The image generating method as claimed in claim 10, wherein the shot information comprises output moment information of a frame being output first from among the group of frames classified as the shot, output moment information of a frame being output last from among the group of frames classified as the shot, and/or shot type information indicating whether the group of frames classified as the shot are to be output as the 2D image or the 3D image.
12. The image generating method as claimed in claim 10, wherein:
the metadata further comprises background depth information for the group of frames classified as the predetermined shot; and
the background depth information comprises coordinate point values of a background of the group of frames classified as the predetermined shot, depth values corresponding to the coordinate point values, and a panel position value that represents a depth value of an output screen.
13. An image processing apparatus to output video data having two-dimensional (2D) images as the 2D images or three-dimensional (3D) images, the image processing apparatus comprising:
a metadata analyzing unit to determine whether the video data is to be output as the 2D image or the 3D image by using metadata associated with the video data;
a 3D image converting unit to convert the video data into the 3D image when the metadata analyzing unit determines that the video data is to be output as the 3D image; and
an output unit to output the video data as the 2D image or the 3D image according to the determination of the metadata analyzing unit,
wherein the metadata includes information to classify frames of the video data into predetermined units.
14. An image generating apparatus comprising:
a video data encoding unit to encode video data as two-dimensional (2D) images;
a metadata generating unit to generate metadata associated with the video data, the metadata comprising information to classify frames of the video data as predetermined units and used to determine whether each of the classified frames is to be converted to a three-dimensional (3D) image; and
a metadata encoding unit to encode the metadata,
wherein the information to classify the frames of the video data as the predetermined units comprises shot information to classify, as a shot, a group of frames in which a background composition of a current frame is predictable by using a previous frame preceding the current frame in the group of frames.
15. A computer-readable information storage medium comprising:
video data recorded as two-dimensional (2D) images; and
metadata associated with the video data, the metadata comprising information used by an image processing apparatus to classify frames of the video data as predetermined units and used by the image processing apparatus to determine whether each of the classified frames is to be converted by the image processing apparatus to a three-dimensional (3D) image,
wherein the information to classify the frames of the video data as the predetermined units comprises shot information to classify, as a shot, a group of frames in which a background composition of a current frame is predictable by using a previous frame preceding the current frame in the group of frames, and
wherein the shot information comprises output moment information of a frame being output first from among the group of frames classified as the shot, output moment information of a frame being output last from among the group of frames classified as the shot, and/or shot type information indicating whether the group of frames classified as the shot are to be output as the 2D image or the 3D image.
PCT/KR2009/003383 2008-06-24 2009-06-24 Image generating method and apparatus and image processing method and apparatus WO2009157701A2 (en)

Priority Applications (3)

Application Number Priority Date Filing Date Title
CN200980123639.6A CN102067615B (en) 2008-06-24 2009-06-24 Image generating method and apparatus and image processing method and apparatus
JP2011514502A JP2011525745A (en) 2008-06-24 2009-06-24 Video generation method, video processing method and apparatus
EP09770373.0A EP2289247A4 (en) 2008-06-24 2009-06-24 Image generating method and apparatus and image processing method and apparatus

Applications Claiming Priority (4)

Application Number Priority Date Filing Date Title
US7518408P 2008-06-24 2008-06-24
US61/075,184 2008-06-24
KR10-2008-0091269 2008-09-17
KR1020080091269A KR20100002032A (en) 2008-06-24 2008-09-17 Image generating method, image processing method, and apparatus thereof

Publications (2)

Publication Number Publication Date
WO2009157701A2 true WO2009157701A2 (en) 2009-12-30
WO2009157701A3 WO2009157701A3 (en) 2010-04-15

Family

ID=41812276

Family Applications (4)

Application Number Title Priority Date Filing Date
PCT/KR2009/003235 WO2009157668A2 (en) 2008-06-24 2009-06-17 Method and apparatus for outputting and displaying image data
PCT/KR2009/003399 WO2009157708A2 (en) 2008-06-24 2009-06-24 Method and apparatus for processing 3d video image
PCT/KR2009/003406 WO2009157714A2 (en) 2008-06-24 2009-06-24 Method and apparatus for processing three dimensional video data
PCT/KR2009/003383 WO2009157701A2 (en) 2008-06-24 2009-06-24 Image generating method and apparatus and image processing method and apparatus

Family Applications Before (3)

Application Number Title Priority Date Filing Date
PCT/KR2009/003235 WO2009157668A2 (en) 2008-06-24 2009-06-17 Method and apparatus for outputting and displaying image data
PCT/KR2009/003399 WO2009157708A2 (en) 2008-06-24 2009-06-24 Method and apparatus for processing 3d video image
PCT/KR2009/003406 WO2009157714A2 (en) 2008-06-24 2009-06-24 Method and apparatus for processing three dimensional video data

Country Status (7)

Country Link
US (6) US20090315884A1 (en)
EP (4) EP2292019A4 (en)
JP (4) JP2011525743A (en)
KR (9) KR101539935B1 (en)
CN (4) CN102077600A (en)
MY (1) MY159672A (en)
WO (4) WO2009157668A2 (en)

Cited By (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2012231254A (en) * 2011-04-25 2012-11-22 Toshiba Corp Stereoscopic image generating apparatus and stereoscopic image generating method
US9426441B2 (en) 2010-03-08 2016-08-23 Dolby Laboratories Licensing Corporation Methods for carrying and transmitting 3D z-norm attributes in digital TV closed captioning
US9519994B2 (en) 2011-04-15 2016-12-13 Dolby Laboratories Licensing Corporation Systems and methods for rendering 3D image independent of display size and viewing distance

Families Citing this family (207)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
EP2289235A4 (en) 2008-05-20 2011-12-28 Pelican Imaging Corp Capturing and processing of images using monolithic camera array with hetergeneous imagers
US11792538B2 (en) 2008-05-20 2023-10-17 Adeia Imaging Llc Capturing and processing of images including occlusions focused on an image sensor by a lens stack array
US8866920B2 (en) 2008-05-20 2014-10-21 Pelican Imaging Corporation Capturing and processing of images using monolithic camera array with heterogeneous imagers
US8520979B2 (en) 2008-08-19 2013-08-27 Digimarc Corporation Methods and systems for content processing
US20100045779A1 (en) * 2008-08-20 2010-02-25 Samsung Electronics Co., Ltd. Three-dimensional video apparatus and method of providing on screen display applied thereto
JP2010088092A (en) * 2008-09-02 2010-04-15 Panasonic Corp Three-dimensional video transmission system, video display device and video output device
JP2010062695A (en) * 2008-09-02 2010-03-18 Sony Corp Image processing apparatus, image processing method, and program
WO2010028107A1 (en) * 2008-09-07 2010-03-11 Dolby Laboratories Licensing Corporation Conversion of interleaved data sets, including chroma correction and/or correction of checkerboard interleaved formatted 3d images
CN102224737B (en) * 2008-11-24 2014-12-03 皇家飞利浦电子股份有限公司 Combining 3D video and auxiliary data
US8599242B2 (en) * 2008-12-02 2013-12-03 Lg Electronics Inc. Method for displaying 3D caption and 3D display apparatus for implementing the same
US20110249757A1 (en) * 2008-12-19 2011-10-13 Koninklijke Philips Electronics N.V. Method and device for overlaying 3d graphics over 3d video
CN102576412B (en) 2009-01-13 2014-11-05 华为技术有限公司 Method and system for image processing to classify an object in an image
KR20100112940A (en) * 2009-04-10 2010-10-20 엘지전자 주식회사 A method for processing data and a receiving system
TW201119353A (en) 2009-06-24 2011-06-01 Dolby Lab Licensing Corp Perceptual depth placement for 3D objects
WO2010151555A1 (en) * 2009-06-24 2010-12-29 Dolby Laboratories Licensing Corporation Method for embedding subtitles and/or graphic overlays in a 3d or multi-view video data
US9479766B2 (en) * 2009-07-10 2016-10-25 Dolby Laboratories Licensing Corporation Modifying images for a 3-dimensional display mode
JP2011029849A (en) * 2009-07-23 2011-02-10 Sony Corp Receiving device, communication system, method of combining caption with stereoscopic image, program, and data structure
CN102474638B (en) * 2009-07-27 2015-07-01 皇家飞利浦电子股份有限公司 Combining 3D video and auxiliary data
KR20110013693A (en) 2009-08-03 2011-02-10 삼성모바일디스플레이주식회사 Organic light emitting display and driving method thereof
KR101056281B1 (en) * 2009-08-03 2011-08-11 삼성모바일디스플레이주식회사 Organic electroluminescent display and driving method thereof
JP5444955B2 (en) 2009-08-31 2014-03-19 ソニー株式会社 Stereoscopic image display system, parallax conversion device, parallax conversion method, and program
US8614737B2 (en) * 2009-09-11 2013-12-24 Disney Enterprises, Inc. System and method for three-dimensional video capture workflow for dynamic rendering
US20110063298A1 (en) * 2009-09-15 2011-03-17 Samir Hulyalkar Method and system for rendering 3d graphics based on 3d display capabilities
US8988495B2 (en) * 2009-11-03 2015-03-24 Lg Eletronics Inc. Image display apparatus, method for controlling the image display apparatus, and image display system
JP2011109398A (en) * 2009-11-17 2011-06-02 Sony Corp Image transmission method, image receiving method, image transmission device, image receiving device, and image transmission system
US8514491B2 (en) 2009-11-20 2013-08-20 Pelican Imaging Corporation Capturing and processing of images using monolithic camera array with heterogeneous imagers
JP5502436B2 (en) * 2009-11-27 2014-05-28 パナソニック株式会社 Video signal processing device
US20110134217A1 (en) * 2009-12-08 2011-06-09 Darren Neuman Method and system for scaling 3d video
TWI491243B (en) * 2009-12-21 2015-07-01 Chunghwa Picture Tubes Ltd Image processing method
JP2011139261A (en) * 2009-12-28 2011-07-14 Sony Corp Image processing device, image processing method, and program
KR20120123087A (en) * 2010-01-13 2012-11-07 톰슨 라이센싱 System and method for combining 3d text with 3d content
US8687116B2 (en) * 2010-01-14 2014-04-01 Panasonic Corporation Video output device and video display system
WO2011091309A1 (en) * 2010-01-21 2011-07-28 General Instrument Corporation Stereoscopic video graphics overlay
WO2011098936A2 (en) * 2010-02-09 2011-08-18 Koninklijke Philips Electronics N.V. 3d video format detection
US9025933B2 (en) * 2010-02-12 2015-05-05 Sony Corporation Information processing device, information processing method, playback device, playback method, program and recording medium
JP2011166666A (en) * 2010-02-15 2011-08-25 Sony Corp Image processor, image processing method, and program
WO2011102818A1 (en) 2010-02-19 2011-08-25 Thomson Licensing Stereo logo insertion
KR101445777B1 (en) * 2010-02-19 2014-11-04 삼성전자 주식회사 Reproducing apparatus and control method thereof
CN102771129A (en) * 2010-02-24 2012-11-07 汤姆森特许公司 Subtitling for stereoscopic images
KR20110098420A (en) * 2010-02-26 2011-09-01 삼성전자주식회사 Display device and driving method thereof
US20110216083A1 (en) * 2010-03-03 2011-09-08 Vizio, Inc. System, method and apparatus for controlling brightness of a device
MX2012010268A (en) * 2010-03-05 2012-10-05 Gen Instrument Corp Method and apparatus for converting two-dimensional video content for insertion into three-dimensional video content.
US8830300B2 (en) * 2010-03-11 2014-09-09 Dolby Laboratories Licensing Corporation Multiscalar stereo video format conversion
US8730301B2 (en) * 2010-03-12 2014-05-20 Sony Corporation Service linkage to caption disparity data transport
JP2011199388A (en) * 2010-03-17 2011-10-06 Sony Corp Reproducing device, reproduction control method, and program
JP2011217361A (en) * 2010-03-18 2011-10-27 Panasonic Corp Device and method of reproducing stereoscopic image and integrated circuit
KR20110107151A (en) * 2010-03-24 2011-09-30 삼성전자주식회사 Method and apparatus for processing 3d image in mobile terminal
US20130010064A1 (en) * 2010-03-24 2013-01-10 Panasonic Corporation Video processing device
JP5526929B2 (en) * 2010-03-30 2014-06-18 ソニー株式会社 Image processing apparatus, image processing method, and program
BR112012026162A2 (en) * 2010-04-12 2017-07-18 Fortem Solutions Inc method, medium and system for three-dimensional rederization of a three-dimensional visible area
JP5960679B2 (en) * 2010-04-14 2016-08-02 サムスン エレクトロニクス カンパニー リミテッド Bitstream generation method, generation apparatus, reception method, and reception apparatus
US20110255003A1 (en) * 2010-04-16 2011-10-20 The Directv Group, Inc. Method and apparatus for presenting on-screen graphics in a frame-compatible 3d format
US9237366B2 (en) 2010-04-16 2016-01-12 Google Technology Holdings LLC Method and apparatus for distribution of 3D television program materials
KR101697184B1 (en) 2010-04-20 2017-01-17 삼성전자주식회사 Apparatus and Method for generating mesh, and apparatus and method for processing image
US9414042B2 (en) * 2010-05-05 2016-08-09 Google Technology Holdings LLC Program guide graphics and video in window for 3DTV
KR20120119927A (en) * 2010-05-11 2012-11-01 삼성전자주식회사 3-Dimension glasses and System for wireless power transmission
KR101824672B1 (en) 2010-05-12 2018-02-05 포토네이션 케이맨 리미티드 Architectures for imager arrays and array cameras
KR101082234B1 (en) * 2010-05-13 2011-11-09 삼성모바일디스플레이주식회사 Organic light emitting display device and driving method thereof
JP2011249895A (en) * 2010-05-24 2011-12-08 Panasonic Corp Signal processing system and signal processing apparatus
US20110292038A1 (en) * 2010-05-27 2011-12-01 Sony Computer Entertainment America, LLC 3d video conversion
KR101699875B1 (en) * 2010-06-03 2017-01-25 엘지디스플레이 주식회사 Apparatus and method for three- dimension liquid crystal display device
US9030536B2 (en) 2010-06-04 2015-05-12 At&T Intellectual Property I, Lp Apparatus and method for presenting media content
JP5682149B2 (en) * 2010-06-10 2015-03-11 ソニー株式会社 Stereo image data transmitting apparatus, stereo image data transmitting method, stereo image data receiving apparatus, and stereo image data receiving method
US8402502B2 (en) 2010-06-16 2013-03-19 At&T Intellectual Property I, L.P. Method and apparatus for presenting media content
US9053562B1 (en) 2010-06-24 2015-06-09 Gregory S. Rabin Two dimensional to three dimensional moving image converter
US8640182B2 (en) 2010-06-30 2014-01-28 At&T Intellectual Property I, L.P. Method for detecting a viewing apparatus
US8593574B2 (en) 2010-06-30 2013-11-26 At&T Intellectual Property I, L.P. Apparatus and method for providing dimensional media content based on detected display capability
US9787974B2 (en) 2010-06-30 2017-10-10 At&T Intellectual Property I, L.P. Method and apparatus for delivering media content
US8918831B2 (en) 2010-07-06 2014-12-23 At&T Intellectual Property I, Lp Method and apparatus for managing a presentation of media content
KR101645404B1 (en) 2010-07-06 2016-08-04 삼성디스플레이 주식회사 Organic Light Emitting Display
JP5609336B2 (en) * 2010-07-07 2014-10-22 ソニー株式会社 Image data transmitting apparatus, image data transmitting method, image data receiving apparatus, image data receiving method, and image data transmitting / receiving system
KR101279660B1 (en) * 2010-07-07 2013-06-27 엘지디스플레이 주식회사 3d image display device and driving method thereof
US9049426B2 (en) 2010-07-07 2015-06-02 At&T Intellectual Property I, Lp Apparatus and method for distributing three dimensional media content
US8848038B2 (en) * 2010-07-09 2014-09-30 Lg Electronics Inc. Method and device for converting 3D images
US9232274B2 (en) 2010-07-20 2016-01-05 At&T Intellectual Property I, L.P. Apparatus for adapting a presentation of media content to a requesting device
US9032470B2 (en) 2010-07-20 2015-05-12 At&T Intellectual Property I, Lp Apparatus for adapting a presentation of media content according to a position of a viewing apparatus
US9560406B2 (en) 2010-07-20 2017-01-31 At&T Intellectual Property I, L.P. Method and apparatus for adapting a presentation of media content
IT1401367B1 (en) * 2010-07-28 2013-07-18 Sisvel Technology Srl METHOD TO COMBINE REFERENCE IMAGES TO A THREE-DIMENSIONAL CONTENT.
US8994716B2 (en) 2010-08-02 2015-03-31 At&T Intellectual Property I, Lp Apparatus and method for providing media content
JPWO2012017687A1 (en) * 2010-08-05 2013-10-03 パナソニック株式会社 Video playback device
KR101674688B1 (en) * 2010-08-12 2016-11-09 엘지전자 주식회사 A method for displaying a stereoscopic image and stereoscopic image playing device
JP2012044625A (en) * 2010-08-23 2012-03-01 Sony Corp Stereoscopic image data transmission device, stereoscopic image data transmission method, stereoscopic image data reception device and stereoscopic image data reception method
US8438502B2 (en) 2010-08-25 2013-05-07 At&T Intellectual Property I, L.P. Apparatus for controlling three-dimensional images
KR101218815B1 (en) * 2010-08-26 2013-01-21 주식회사 티스마트 3D user interface processing method and set-top box using the same
US8994792B2 (en) * 2010-08-27 2015-03-31 Broadcom Corporation Method and system for creating a 3D video from a monoscopic 2D video and corresponding depth information
JP5058316B2 (en) * 2010-09-03 2012-10-24 株式会社東芝 Electronic device, image processing method, and image processing program
EP2426931A1 (en) * 2010-09-06 2012-03-07 Advanced Digital Broadcast S.A. A method and a system for determining a video frame type
WO2012031406A1 (en) * 2010-09-10 2012-03-15 青岛海信信芯科技有限公司 Display method and equipment for 3d tv interface
JP2012094111A (en) * 2010-09-29 2012-05-17 Sony Corp Image processing device, image processing method and program
WO2012044272A1 (en) * 2010-09-29 2012-04-05 Thomson Licensing Automatically switching between three dimensional and two dimensional contents for display
JP5543892B2 (en) * 2010-10-01 2014-07-09 日立コンシューマエレクトロニクス株式会社 REPRODUCTION DEVICE, REPRODUCTION METHOD, DISPLAY DEVICE, AND DISPLAY METHOD
US8947511B2 (en) 2010-10-01 2015-02-03 At&T Intellectual Property I, L.P. Apparatus and method for presenting three-dimensional media content
US8941724B2 (en) * 2010-10-01 2015-01-27 Hitachi Maxell Ltd. Receiver
TWI420151B (en) * 2010-10-07 2013-12-21 Innolux Corp Display method
KR101232086B1 (en) * 2010-10-08 2013-02-08 엘지디스플레이 주식회사 Liquid crystal display and local dimming control method of thereof
US20120092327A1 (en) * 2010-10-14 2012-04-19 Sony Corporation Overlaying graphical assets onto viewing plane of 3d glasses per metadata accompanying 3d image
JP5550520B2 (en) * 2010-10-20 2014-07-16 日立コンシューマエレクトロニクス株式会社 Playback apparatus and playback method
KR20120047055A (en) * 2010-11-03 2012-05-11 삼성전자주식회사 Display apparatus and method for providing graphic image
CN102469319A (en) * 2010-11-10 2012-05-23 康佳集团股份有限公司 Three-dimensional menu generation method and three-dimensional display device
JP5789960B2 (en) * 2010-11-18 2015-10-07 セイコーエプソン株式会社 Display device, display device control method, and program
JP5786315B2 (en) * 2010-11-24 2015-09-30 セイコーエプソン株式会社 Display device, display device control method, and program
CN101980545B (en) * 2010-11-29 2012-08-01 深圳市九洲电器有限公司 Method for automatically detecting 3DTV video program format
CN101984671B (en) * 2010-11-29 2013-04-17 深圳市九洲电器有限公司 Method for synthesizing video images and interface graphs by 3DTV receiving system
US8878950B2 (en) 2010-12-14 2014-11-04 Pelican Imaging Corporation Systems and methods for synthesizing high resolution images using super-resolution processes
JP2012129845A (en) * 2010-12-16 2012-07-05 Jvc Kenwood Corp Image processing device
JP5611807B2 (en) * 2010-12-27 2014-10-22 Necパーソナルコンピュータ株式会社 Video display device
US8600151B2 (en) * 2011-01-03 2013-12-03 Apple Inc. Producing stereoscopic image
KR101814798B1 (en) * 2011-01-26 2018-01-04 삼성전자주식회사 Apparatus for processing three dimension image and method for the same
CN105554551A (en) * 2011-03-02 2016-05-04 华为技术有限公司 Method and device for acquiring three-dimensional (3D) format description information
CN102157012B (en) * 2011-03-23 2012-11-28 深圳超多维光电子有限公司 Method for three-dimensionally rendering scene, graphic image treatment device, equipment and system
KR101801141B1 (en) * 2011-04-19 2017-11-24 엘지전자 주식회사 Apparatus for displaying image and method for operating the same
KR20120119173A (en) * 2011-04-20 2012-10-30 삼성전자주식회사 3d image processing apparatus and method for adjusting three-dimensional effect thereof
EP2708019B1 (en) 2011-05-11 2019-10-16 FotoNation Limited Systems and methods for transmitting and receiving array camera image data
US9445046B2 (en) 2011-06-24 2016-09-13 At&T Intellectual Property I, L.P. Apparatus and method for presenting media content with telepresence
US8947497B2 (en) 2011-06-24 2015-02-03 At&T Intellectual Property I, Lp Apparatus and method for managing telepresence sessions
US9602766B2 (en) 2011-06-24 2017-03-21 At&T Intellectual Property I, L.P. Apparatus and method for presenting three dimensional objects with telepresence
US9030522B2 (en) 2011-06-24 2015-05-12 At&T Intellectual Property I, Lp Apparatus and method for providing media content
CN102231829B (en) * 2011-06-27 2014-12-17 深圳超多维光电子有限公司 Display format identification method and device of video file as well as video player
US9294752B2 (en) 2011-07-13 2016-03-22 Google Technology Holdings LLC Dual mode user interface system and method for 3D video
US8587635B2 (en) 2011-07-15 2013-11-19 At&T Intellectual Property I, L.P. Apparatus and method for providing media services with telepresence
WO2013023325A1 (en) * 2011-08-18 2013-02-21 北京世纪高蓝科技有限公司 Method for converting 2d into 3d based on image motion information
CN103002297A (en) * 2011-09-16 2013-03-27 联咏科技股份有限公司 Method and device for generating dynamic depth values
US20130070060A1 (en) 2011-09-19 2013-03-21 Pelican Imaging Corporation Systems and methods for determining depth from multiple views of a scene that include aliasing using hypothesized fusion
US8952996B2 (en) * 2011-09-27 2015-02-10 Delta Electronics, Inc. Image display system
WO2013049699A1 (en) 2011-09-28 2013-04-04 Pelican Imaging Corporation Systems and methods for encoding and decoding light field image files
US8813109B2 (en) 2011-10-21 2014-08-19 The Nielsen Company (Us), Llc Methods and apparatus to identify exposure to 3D media presentations
US8687470B2 (en) 2011-10-24 2014-04-01 Lsi Corporation Optical disk playback device with three-dimensional playback functionality
JP5289538B2 (en) * 2011-11-11 2013-09-11 株式会社東芝 Electronic device, display control method and program
CN102413350B (en) * 2011-11-30 2014-04-16 四川长虹电器股份有限公司 Method for processing blue-light 3D (three-dimensional) video
FR2983673A1 (en) * 2011-12-02 2013-06-07 Binocle CORRECTION METHOD FOR ALTERNATE PROJECTION OF STEREOSCOPIC IMAGES
US9412206B2 (en) 2012-02-21 2016-08-09 Pelican Imaging Corporation Systems and methods for the manipulation of captured light field image data
US8479226B1 (en) * 2012-02-21 2013-07-02 The Nielsen Company (Us), Llc Methods and apparatus to identify exposure to 3D media presentations
US8713590B2 (en) 2012-02-21 2014-04-29 The Nielsen Company (Us), Llc Methods and apparatus to identify exposure to 3D media presentations
US10445398B2 (en) * 2012-03-01 2019-10-15 Sony Corporation Asset management during production of media
EP2836657B1 (en) 2012-04-10 2017-12-06 Dirtt Environmental Solutions, Ltd. Tamper evident wall cladding system
US20150109411A1 (en) * 2012-04-26 2015-04-23 Electronics And Telecommunications Research Institute Image playback apparatus for 3dtv and method performed by the apparatus
CN104508681B (en) 2012-06-28 2018-10-30 Fotonation开曼有限公司 For detecting defective camera array, optical device array and the system and method for sensor
US20140002674A1 (en) 2012-06-30 2014-01-02 Pelican Imaging Corporation Systems and Methods for Manufacturing Camera Modules Using Active Alignment of Lens Stack Arrays and Sensors
US8619082B1 (en) 2012-08-21 2013-12-31 Pelican Imaging Corporation Systems and methods for parallax detection and correction in images captured using array cameras that contain occlusions using subsets of images to perform depth estimation
US20140055632A1 (en) 2012-08-23 2014-02-27 Pelican Imaging Corporation Feature based high resolution motion estimation from low resolution images captured using an array source
KR20140039649A (en) 2012-09-24 2014-04-02 삼성전자주식회사 Multi view image generating method and multi view image display apparatus
KR20140049834A (en) * 2012-10-18 2014-04-28 삼성전자주식회사 Broadcast receiving apparatus and method of controlling the same, and user terminal device and method of providing the screen.
US9143711B2 (en) 2012-11-13 2015-09-22 Pelican Imaging Corporation Systems and methods for array camera focal plane control
WO2014084613A2 (en) * 2012-11-27 2014-06-05 인텔렉추얼 디스커버리 주식회사 Method for encoding and decoding image using depth information, and device and image system using same
KR101430985B1 (en) * 2013-02-20 2014-09-18 주식회사 카몬 System and Method on Providing Multi-Dimensional Content
WO2014130849A1 (en) 2013-02-21 2014-08-28 Pelican Imaging Corporation Generating compressed light field representation data
US9253380B2 (en) 2013-02-24 2016-02-02 Pelican Imaging Corporation Thin form factor computational array cameras and modular array cameras
WO2014138695A1 (en) 2013-03-08 2014-09-12 Pelican Imaging Corporation Systems and methods for measuring scene information while capturing images using array cameras
US8866912B2 (en) 2013-03-10 2014-10-21 Pelican Imaging Corporation System and methods for calibration of an array camera using a single captured image
WO2014164909A1 (en) 2013-03-13 2014-10-09 Pelican Imaging Corporation Array camera architecture implementing quantum film sensors
US9124831B2 (en) 2013-03-13 2015-09-01 Pelican Imaging Corporation System and methods for calibration of an array camera
WO2014159779A1 (en) 2013-03-14 2014-10-02 Pelican Imaging Corporation Systems and methods for reducing motion blur in images or video in ultra low light with array cameras
US9992021B1 (en) 2013-03-14 2018-06-05 GoTenna, Inc. System and method for private and point-to-point communication between computing devices
WO2014153098A1 (en) 2013-03-14 2014-09-25 Pelican Imaging Corporation Photmetric normalization in array cameras
US10122993B2 (en) 2013-03-15 2018-11-06 Fotonation Limited Autofocus system for a conventional camera that uses depth information from an array camera
US9445003B1 (en) 2013-03-15 2016-09-13 Pelican Imaging Corporation Systems and methods for synthesizing high resolution images using image deconvolution based on motion and depth information
US9497429B2 (en) 2013-03-15 2016-11-15 Pelican Imaging Corporation Extended color processing on pelican array cameras
WO2014145856A1 (en) 2013-03-15 2014-09-18 Pelican Imaging Corporation Systems and methods for stereo imaging with camera arrays
CN104079941B (en) * 2013-03-27 2017-08-25 中兴通讯股份有限公司 A kind of depth information decoding method, device and Video processing playback equipment
CN104469338B (en) * 2013-09-25 2016-08-17 联想(北京)有限公司 A kind of control method and device
US9898856B2 (en) 2013-09-27 2018-02-20 Fotonation Cayman Limited Systems and methods for depth-assisted perspective distortion correction
US10491916B2 (en) * 2013-10-01 2019-11-26 Advanced Micro Devices, Inc. Exploiting camera depth information for video encoding
CN103543953B (en) * 2013-11-08 2017-01-04 深圳市汉普电子技术开发有限公司 The method of the 3D film source that broadcasting identifies without 3D and touch apparatus
JP2015119464A (en) * 2013-11-12 2015-06-25 セイコーエプソン株式会社 Display device and control method of the same
WO2015074078A1 (en) 2013-11-18 2015-05-21 Pelican Imaging Corporation Estimating depth from projected texture using camera arrays
US9456134B2 (en) 2013-11-26 2016-09-27 Pelican Imaging Corporation Array camera configurations incorporating constituent array cameras and constituent cameras
WO2015134996A1 (en) 2014-03-07 2015-09-11 Pelican Imaging Corporation System and methods for depth regularization and semiautomatic interactive matting using rgb-d images
CN104143308B (en) * 2014-07-24 2016-09-07 京东方科技集团股份有限公司 The display methods of a kind of 3-D view and device
US10228751B2 (en) 2014-08-06 2019-03-12 Apple Inc. Low power mode
US9647489B2 (en) 2014-08-26 2017-05-09 Apple Inc. Brownout avoidance
WO2016054089A1 (en) 2014-09-29 2016-04-07 Pelican Imaging Corporation Systems and methods for dynamic calibration of array cameras
US10708391B1 (en) 2014-09-30 2020-07-07 Apple Inc. Delivery of apps in a media stream
US10231033B1 (en) * 2014-09-30 2019-03-12 Apple Inc. Synchronizing out-of-band content with a media stream
CN105095895B (en) * 2015-04-23 2018-09-25 广州广电运通金融电子股份有限公司 Valuable file identification device self-correction recognition methods
CN105376546A (en) * 2015-11-09 2016-03-02 中科创达软件股份有限公司 2D-to-3D method, device and mobile terminal
CN105472374A (en) * 2015-11-19 2016-04-06 广州华多网络科技有限公司 3D live video realization method, apparatus, and system
US20170150138A1 (en) * 2015-11-25 2017-05-25 Atheer, Inc. Method and apparatus for selective mono/stereo visual display
US20170150137A1 (en) * 2015-11-25 2017-05-25 Atheer, Inc. Method and apparatus for selective mono/stereo visual display
CN105872519B (en) * 2016-04-13 2018-03-27 万云数码媒体有限公司 A kind of 2D plus depth 3D rendering transverse direction storage methods based on RGB compressions
US10433025B2 (en) * 2016-05-10 2019-10-01 Jaunt Inc. Virtual reality resource scheduling of process in a cloud-based virtual reality processing system
CN106101681A (en) * 2016-06-21 2016-11-09 青岛海信电器股份有限公司 3-D view display processing method, signal input device and television terminal
CN106982367A (en) * 2017-03-31 2017-07-25 联想(北京)有限公司 Video transmission method and its device
US10038500B1 (en) * 2017-05-11 2018-07-31 Qualcomm Incorporated Visible light communication
US10735707B2 (en) * 2017-08-15 2020-08-04 International Business Machines Corporation Generating three-dimensional imagery
US10482618B2 (en) 2017-08-21 2019-11-19 Fotonation Limited Systems and methods for hybrid depth regularization
CN107589989A (en) 2017-09-14 2018-01-16 晨星半导体股份有限公司 Display device and its method for displaying image based on Android platform
US11363133B1 (en) 2017-12-20 2022-06-14 Apple Inc. Battery health-based power management
US10817307B1 (en) 2017-12-20 2020-10-27 Apple Inc. API behavior modification based on power source health
EP3644604A1 (en) * 2018-10-23 2020-04-29 Koninklijke Philips N.V. Image generating apparatus and method therefor
CN109257585B (en) * 2018-10-25 2021-04-06 京东方科技集团股份有限公司 Brightness correction device and method, display device, display system and method
CN109274949A (en) * 2018-10-30 2019-01-25 京东方科技集团股份有限公司 A kind of method of video image processing and its device, display equipment
CN112188181B (en) * 2019-07-02 2023-07-04 中强光电股份有限公司 Image display device, stereoscopic image processing circuit and synchronization signal correction method thereof
CN114600165A (en) 2019-09-17 2022-06-07 波士顿偏振测定公司 System and method for surface modeling using polarization cues
EP4042101A4 (en) 2019-10-07 2023-11-22 Boston Polarimetrics, Inc. Systems and methods for surface normals sensing with polarization
KR102558903B1 (en) 2019-11-30 2023-07-24 보스턴 폴라리메트릭스, 인크. System and Method for Segmenting Transparent Objects Using Polarized Signals
KR102241615B1 (en) * 2020-01-15 2021-04-19 한국과학기술원 Method to identify and video titles using metadata in video webpage source code, and apparatuses performing the same
JP7462769B2 (en) 2020-01-29 2024-04-05 イントリンジック イノベーション エルエルシー System and method for characterizing an object pose detection and measurement system - Patents.com
CN115428028A (en) 2020-01-30 2022-12-02 因思创新有限责任公司 System and method for synthesizing data for training statistical models in different imaging modalities including polarized images
WO2021243088A1 (en) 2020-05-27 2021-12-02 Boston Polarimetrics, Inc. Multi-aperture polarization optical systems using beam splitters
CN112004162B (en) * 2020-09-08 2022-06-21 宁波视睿迪光电有限公司 Online 3D content playing system and method
US12069227B2 (en) 2021-03-10 2024-08-20 Intrinsic Innovation Llc Multi-modal and multi-spectral stereo camera arrays
US12020455B2 (en) 2021-03-10 2024-06-25 Intrinsic Innovation Llc Systems and methods for high dynamic range image reconstruction
US11290658B1 (en) 2021-04-15 2022-03-29 Boston Polarimetrics, Inc. Systems and methods for camera exposure control
US11954886B2 (en) 2021-04-15 2024-04-09 Intrinsic Innovation Llc Systems and methods for six-degree of freedom pose estimation of deformable objects
US12067746B2 (en) 2021-05-07 2024-08-20 Intrinsic Innovation Llc Systems and methods for using computer vision to pick up small objects
US11689813B2 (en) 2021-07-01 2023-06-27 Intrinsic Innovation Llc Systems and methods for high dynamic range imaging using crossed polarizers
US11770513B1 (en) * 2022-07-13 2023-09-26 Rovi Guides, Inc. Systems and methods for reducing a number of focal planes used to display three-dimensional objects

Family Cites Families (123)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US4523226A (en) * 1982-01-27 1985-06-11 Stereographics Corporation Stereoscopic television system
US4667228A (en) * 1983-10-14 1987-05-19 Canon Kabushiki Kaisha Image signal processing apparatus
JPS63116593A (en) * 1986-11-04 1988-05-20 Matsushita Electric Ind Co Ltd Stereoscopic picture reproducing device
US5262879A (en) * 1988-07-18 1993-11-16 Dimensional Arts. Inc. Holographic image conversion method for making a controlled holographic grating
US5058992A (en) * 1988-09-07 1991-10-22 Toppan Printing Co., Ltd. Method for producing a display with a diffraction grating pattern and a display produced by the method
JP2508387B2 (en) * 1989-10-16 1996-06-19 凸版印刷株式会社 Method of manufacturing display having diffraction grating pattern
US5291317A (en) * 1990-07-12 1994-03-01 Applied Holographics Corporation Holographic diffraction grating patterns and methods for creating the same
JP3081675B2 (en) * 1991-07-24 2000-08-28 オリンパス光学工業株式会社 Image recording device and image reproducing device
US5740274A (en) * 1991-09-12 1998-04-14 Fuji Photo Film Co., Ltd. Method for recognizing object images and learning method for neural networks
US6011581A (en) * 1992-11-16 2000-01-04 Reveo, Inc. Intelligent method and system for producing and displaying stereoscopically-multiplexed images of three-dimensional objects for use in realistic stereoscopic viewing thereof in interactive virtual reality display environments
US6084978A (en) * 1993-12-16 2000-07-04 Eastman Kodak Company Hierarchical storage and display of digital images used in constructing three-dimensional image hard copy
CN1113320C (en) * 1994-02-01 2003-07-02 三洋电机株式会社 Method of converting two-dimensional images into three-dimensional images
US5739844A (en) * 1994-02-04 1998-04-14 Sanyo Electric Co. Ltd. Method of converting two-dimensional image into three-dimensional image
US5684890A (en) * 1994-02-28 1997-11-04 Nec Corporation Three-dimensional reference image segmenting method and apparatus
US6104828A (en) * 1994-03-24 2000-08-15 Kabushiki Kaisha Topcon Ophthalmologic image processor
JP2846840B2 (en) * 1994-07-14 1999-01-13 三洋電機株式会社 Method for generating 3D image from 2D image
KR100374463B1 (en) * 1994-09-22 2003-05-09 산요 덴키 가부시키가이샤 How to convert 2D image to 3D image
US6985168B2 (en) * 1994-11-14 2006-01-10 Reveo, Inc. Intelligent method and system for producing and displaying stereoscopically-multiplexed images of three-dimensional objects for use in realistic stereoscopic viewing thereof in interactive virtual reality display environments
JPH09116931A (en) * 1995-10-18 1997-05-02 Sanyo Electric Co Ltd Method and identifying left and right video image for time division stereoscopic video signal
US5917940A (en) * 1996-01-23 1999-06-29 Nec Corporation Three dimensional reference image segmenting method and device and object discrimination system
JPH09322199A (en) * 1996-05-29 1997-12-12 Olympus Optical Co Ltd Stereoscopic video display device
US5986781A (en) * 1996-10-28 1999-11-16 Pacific Holographics, Inc. Apparatus and method for generating diffractive element using liquid crystal display
JPH10224822A (en) * 1997-01-31 1998-08-21 Sony Corp Video display method and display device
JPH10313417A (en) * 1997-03-12 1998-11-24 Seiko Epson Corp Digital gamma correction circuit, liquid crystal display device using the same and electronic device
DE19806547C2 (en) * 1997-04-30 2001-01-25 Hewlett Packard Co System and method for generating stereoscopic display signals from a single computer graphics pipeline
JPH11113028A (en) * 1997-09-30 1999-04-23 Toshiba Corp Three-dimension video image display device
ID27878A (en) * 1997-12-05 2001-05-03 Dynamic Digital Depth Res Pty IMAGE IMPROVED IMAGE CONVERSION AND ENCODING ENGINEERING
US6850631B1 (en) * 1998-02-20 2005-02-01 Oki Electric Industry Co., Ltd. Photographing device, iris input device and iris image input method
JP4149037B2 (en) * 1998-06-04 2008-09-10 オリンパス株式会社 Video system
US6704042B2 (en) * 1998-12-10 2004-03-09 Canon Kabushiki Kaisha Video processing apparatus, control method therefor, and storage medium
JP2000298246A (en) * 1999-02-12 2000-10-24 Canon Inc Device and method for display, and storage medium
JP2000275575A (en) * 1999-03-24 2000-10-06 Sharp Corp Stereoscopic video display device
KR100334722B1 (en) * 1999-06-05 2002-05-04 강호석 Method and the apparatus for generating stereoscopic image using MPEG data
JP2001012946A (en) * 1999-06-30 2001-01-19 Toshiba Corp Dynamic image processor and processing method
US6839663B1 (en) * 1999-09-30 2005-01-04 Texas Tech University Haptic rendering of volumetric soft-bodies objects
WO2001045425A1 (en) * 1999-12-14 2001-06-21 Scientific-Atlanta, Inc. System and method for adaptive decoding of a video signal with coordinated resource allocation
US6968568B1 (en) * 1999-12-20 2005-11-22 International Business Machines Corporation Methods and apparatus of disseminating broadcast information to a handheld device
US20020009137A1 (en) * 2000-02-01 2002-01-24 Nelson John E. Three-dimensional video broadcasting system
JP4635403B2 (en) * 2000-04-04 2011-02-23 ソニー株式会社 Stereoscopic image creation method and apparatus
JP2001320693A (en) * 2000-05-12 2001-11-16 Sony Corp Service providing device and method, reception terminal and method, service providing system
AU2001266862A1 (en) * 2000-06-12 2001-12-24 Vrex, Inc. Electronic stereoscopic media delivery system
JP3667620B2 (en) * 2000-10-16 2005-07-06 株式会社アイ・オー・データ機器 Stereo image capturing adapter, stereo image capturing camera, and stereo image processing apparatus
US6762755B2 (en) 2000-10-16 2004-07-13 Pixel Science, Inc. Method and apparatus for creating and displaying interactive three dimensional computer images
GB0100563D0 (en) * 2001-01-09 2001-02-21 Pace Micro Tech Plc Dynamic adjustment of on-screen displays to cope with different widescreen signalling types
US6678323B2 (en) * 2001-01-24 2004-01-13 Her Majesty The Queen In Right Of Canada, As Represented By The Minister Of Industry Through Communications Research Centre Bandwidth reduction for stereoscopic imagery and video signals
KR20020096203A (en) * 2001-06-18 2002-12-31 (주)디지털국영 The Method for Enlarging or Reducing Stereoscopic Images
JP2003157292A (en) * 2001-11-20 2003-05-30 Nec Corp System and method for managing layout of product
KR100397511B1 (en) * 2001-11-21 2003-09-13 한국전자통신연구원 The processing system and it's method for the stereoscopic/multiview Video
GB0129992D0 (en) * 2001-12-14 2002-02-06 Ocuity Ltd Control of optical switching apparatus
US20040218269A1 (en) * 2002-01-14 2004-11-04 Divelbiss Adam W. General purpose stereoscopic 3D format conversion system and method
US7319720B2 (en) * 2002-01-28 2008-01-15 Microsoft Corporation Stereoscopic video
JP2003284099A (en) * 2002-03-22 2003-10-03 Olympus Optical Co Ltd Video information signal recording medium and video display apparatus
US6771274B2 (en) * 2002-03-27 2004-08-03 Sony Corporation Graphics and video integration with alpha and video blending
CA2380105A1 (en) * 2002-04-09 2003-10-09 Nicholas Routhier Process and system for encoding and playback of stereoscopic video sequences
EP2202978A1 (en) * 2002-04-12 2010-06-30 Mitsubishi Denki Kabushiki Kaisha Hint information describing method for manipulating metadata
JP4652389B2 (en) * 2002-04-12 2011-03-16 三菱電機株式会社 Metadata processing method
US20050248561A1 (en) * 2002-04-25 2005-11-10 Norio Ito Multimedia information generation method and multimedia information reproduction device
JP4154569B2 (en) * 2002-07-10 2008-09-24 日本電気株式会社 Image compression / decompression device
WO2004008768A1 (en) * 2002-07-16 2004-01-22 Electronics And Telecommunications Research Institute Apparatus and method for adapting 2d and 3d stereoscopic video signal
KR100488804B1 (en) * 2002-10-07 2005-05-12 한국전자통신연구원 System for data processing of 2-view 3dimention moving picture being based on MPEG-4 and method thereof
JP2004186863A (en) * 2002-12-02 2004-07-02 Amita Technology Kk Stereophoscopic vision display unit and stereophoscopic vision signal processing circuit
JP4183499B2 (en) * 2002-12-16 2008-11-19 三洋電機株式会社 Video file processing method and video processing method
JP2004246066A (en) * 2003-02-13 2004-09-02 Fujitsu Ltd Virtual environment creating method
JP2004274125A (en) * 2003-03-05 2004-09-30 Sony Corp Image processing apparatus and method
JP4677175B2 (en) * 2003-03-24 2011-04-27 シャープ株式会社 Image processing apparatus, image pickup system, image display system, image pickup display system, image processing program, and computer-readable recording medium recording image processing program
JP2004309868A (en) * 2003-04-08 2004-11-04 Sony Corp Imaging device and stereoscopic video generating device
KR100556826B1 (en) * 2003-04-17 2006-03-10 한국전자통신연구원 System and Method of Internet Broadcasting for MPEG4 based Stereoscopic Video
CN101841728B (en) * 2003-04-17 2012-08-08 夏普株式会社 Three-dimensional image processing apparatus
JP2004357156A (en) * 2003-05-30 2004-12-16 Sharp Corp Video reception apparatus and video playback apparatus
JP2005026800A (en) * 2003-06-30 2005-01-27 Konica Minolta Photo Imaging Inc Image processing method, imaging apparatus, image processing apparatus, and image recording apparatus
ITRM20030345A1 (en) * 2003-07-15 2005-01-16 St Microelectronics Srl METHOD TO FIND A DEPTH MAP
US7411611B2 (en) * 2003-08-25 2008-08-12 Barco N. V. Device and method for performing multiple view imaging by means of a plurality of video processing devices
EP1510940A1 (en) * 2003-08-29 2005-03-02 Sap Ag A method of providing a visualisation graph on a computer and a computer for providing a visualisation graph
WO2005055607A1 (en) * 2003-12-08 2005-06-16 Electronics And Telecommunications Research Institute System and method for encoding and decoding an image using bitstream map and recording medium thereof
JP2005175997A (en) * 2003-12-12 2005-06-30 Sony Corp Decoding apparatus, electronic apparatus, computer, decoding method, program, and recording medium
KR100544677B1 (en) * 2003-12-26 2006-01-23 한국전자통신연구원 Apparatus and method for the 3D object tracking using multi-view and depth cameras
JP3746506B2 (en) * 2004-03-08 2006-02-15 一成 江良 Stereoscopic parameter embedding device and stereoscopic image reproducing device
JP4230959B2 (en) * 2004-05-19 2009-02-25 株式会社東芝 Media data playback device, media data playback system, media data playback program, and remote operation program
KR100543219B1 (en) * 2004-05-24 2006-01-20 한국과학기술연구원 Method for generating haptic vector field and 3d-height map in 2d-image
JP4227076B2 (en) * 2004-05-24 2009-02-18 株式会社東芝 Display device for displaying stereoscopic image and display method for displaying stereoscopic image
KR100708838B1 (en) * 2004-06-30 2007-04-17 삼성에스디아이 주식회사 Stereoscopic display device and driving method thereof
JP2006041811A (en) * 2004-07-26 2006-02-09 Kddi Corp Free visual point picture streaming method
KR20040077596A (en) * 2004-07-28 2004-09-04 손귀연 Stereoscopic Image Display Device Based on Flat Panel Display
CN100573231C (en) * 2004-09-08 2009-12-23 日本电信电话株式会社 3 D displaying method, device
KR100694069B1 (en) * 2004-11-29 2007-03-12 삼성전자주식회사 Recording apparatus including plurality of data blocks of different sizes, file managing method using the same and printing apparatus including the same
KR100656575B1 (en) 2004-12-31 2006-12-11 광운대학교 산학협력단 Three-dimensional display device
TWI261099B (en) * 2005-02-17 2006-09-01 Au Optronics Corp Backlight modules
KR20060122672A (en) * 2005-05-26 2006-11-30 삼성전자주식회사 Storage medium including application for obtaining meta data, apparatus for obtaining meta data, and method therefor
KR100828358B1 (en) * 2005-06-14 2008-05-08 삼성전자주식회사 Method and apparatus for converting display mode of video, and computer readable medium thereof
US7404645B2 (en) * 2005-06-20 2008-07-29 Digital Display Innovations, Llc Image and light source modulation for a digital display system
KR100813977B1 (en) * 2005-07-08 2008-03-14 삼성전자주식회사 High resolution 2D-3D switchable autostereoscopic display apparatus
US8384763B2 (en) * 2005-07-26 2013-02-26 Her Majesty the Queen in right of Canada as represented by the Minster of Industry, Through the Communications Research Centre Canada Generating a depth map from a two-dimensional source image for stereoscopic and multiview imaging
JP4717728B2 (en) * 2005-08-29 2011-07-06 キヤノン株式会社 Stereo display device and control method thereof
US9113147B2 (en) * 2005-09-27 2015-08-18 Qualcomm Incorporated Scalability techniques based on content information
CN101292538B (en) * 2005-10-19 2012-11-28 汤姆森特许公司 Multi-view video coding using scalable video coding
KR100739764B1 (en) * 2005-11-28 2007-07-13 삼성전자주식회사 Apparatus and method for processing 3 dimensional video signal
KR100793750B1 (en) * 2006-02-14 2008-01-10 엘지전자 주식회사 The display device for storing the various configuration data for displaying and the method for controlling the same
KR100780701B1 (en) 2006-03-28 2007-11-30 (주)오픈브이알 Apparatus automatically creating three dimension image and method therefore
KR20070098364A (en) * 2006-03-31 2007-10-05 (주)엔브이엘소프트 Apparatus and method for coding and saving a 3d moving image
KR101137347B1 (en) * 2006-05-11 2012-04-19 엘지전자 주식회사 apparatus for mobile telecommunication and method for displaying an image using the apparatus
JP2007304325A (en) * 2006-05-11 2007-11-22 Necディスプレイソリューションズ株式会社 Liquid crystal display device and liquid crystal panel driving method
US7953315B2 (en) * 2006-05-22 2011-05-31 Broadcom Corporation Adaptive video processing circuitry and player using sub-frame metadata
US20070294737A1 (en) * 2006-06-16 2007-12-20 Sbc Knowledge Ventures, L.P. Internet Protocol Television (IPTV) stream management within a home viewing network
CA2653815C (en) * 2006-06-23 2016-10-04 Imax Corporation Methods and systems for converting 2d motion pictures for stereoscopic 3d exhibition
KR100761022B1 (en) * 2006-08-14 2007-09-21 광주과학기술원 Haptic rendering method based on depth image, device therefor, and haptic broadcasting system using them
KR100716142B1 (en) * 2006-09-04 2007-05-11 주식회사 이시티 Method for transferring stereoscopic image data
EP1901474B1 (en) 2006-09-13 2011-11-30 Stmicroelectronics Sa System for synchronizing modules in an integrated circuit in mesochronous clock domains
US20100091012A1 (en) * 2006-09-28 2010-04-15 Koninklijke Philips Electronics N.V. 3 menu display
US8711203B2 (en) * 2006-10-11 2014-04-29 Koninklijke Philips N.V. Creating three dimensional graphics data
JP4755565B2 (en) * 2006-10-17 2011-08-24 シャープ株式会社 Stereoscopic image processing device
KR101362941B1 (en) * 2006-11-01 2014-02-17 한국전자통신연구원 Method and Apparatus for decoding metadata used for playing stereoscopic contents
TWI324477B (en) * 2006-11-03 2010-05-01 Quanta Comp Inc Stereoscopic image format transformation method applied to display system
US8325278B2 (en) * 2006-11-29 2012-12-04 Panasonic Corporation Video display based on video signal and audio output based on audio signal, video/audio device network including video/audio signal input/output device and video/audio reproduction device, and signal reproducing method
KR100786468B1 (en) * 2007-01-02 2007-12-17 삼성에스디아이 주식회사 2d and 3d image selectable display device
US8488868B2 (en) * 2007-04-03 2013-07-16 Her Majesty The Queen In Right Of Canada, As Represented By The Minister Of Industry, Through The Communications Research Centre Canada Generation of a depth map from a monoscopic color image for rendering stereoscopic still and video images
US8213711B2 (en) * 2007-04-03 2012-07-03 Her Majesty The Queen In Right Of Canada As Represented By The Minister Of Industry, Through The Communications Research Centre Canada Method and graphical user interface for modifying depth maps
JP4564512B2 (en) 2007-04-16 2010-10-20 富士通株式会社 Display device, display program, and display method
KR100839429B1 (en) * 2007-04-17 2008-06-19 삼성에스디아이 주식회사 Electronic display device and the method thereof
WO2008140190A1 (en) * 2007-05-14 2008-11-20 Samsung Electronics Co, . Ltd. Method and apparatus for encoding and decoding multi-view image
JP4462288B2 (en) * 2007-05-16 2010-05-12 株式会社日立製作所 Video display device and three-dimensional video display device using the same
CN103281589A (en) * 2007-10-10 2013-09-04 韩国电子通信研究院 Non-transient computer readable storage medium
US20090315981A1 (en) * 2008-06-24 2009-12-24 Samsung Electronics Co., Ltd. Image processing method and apparatus
US8482654B2 (en) * 2008-10-24 2013-07-09 Reald Inc. Stereoscopic image format with depth information

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
See references of EP2289247A4 *

Cited By (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US9426441B2 (en) 2010-03-08 2016-08-23 Dolby Laboratories Licensing Corporation Methods for carrying and transmitting 3D z-norm attributes in digital TV closed captioning
US9519994B2 (en) 2011-04-15 2016-12-13 Dolby Laboratories Licensing Corporation Systems and methods for rendering 3D image independent of display size and viewing distance
JP2012231254A (en) * 2011-04-25 2012-11-22 Toshiba Corp Stereoscopic image generating apparatus and stereoscopic image generating method

Also Published As

Publication number Publication date
JP2011526103A (en) 2011-09-29
MY159672A (en) 2017-01-13
CN102067614A (en) 2011-05-18
US20100103168A1 (en) 2010-04-29
CN102067614B (en) 2014-06-11
EP2289248A2 (en) 2011-03-02
US20090315979A1 (en) 2009-12-24
KR20100002049A (en) 2010-01-06
KR20100002033A (en) 2010-01-06
CN102067615A (en) 2011-05-18
EP2289247A2 (en) 2011-03-02
WO2009157708A3 (en) 2010-04-15
KR20100002031A (en) 2010-01-06
EP2292019A4 (en) 2014-04-30
KR20100002035A (en) 2010-01-06
WO2009157701A3 (en) 2010-04-15
KR20100002037A (en) 2010-01-06
US20100104219A1 (en) 2010-04-29
US20090317061A1 (en) 2009-12-24
KR20100002032A (en) 2010-01-06
EP2279625A2 (en) 2011-02-02
KR20100002048A (en) 2010-01-06
JP2011525743A (en) 2011-09-22
EP2279625A4 (en) 2013-07-03
WO2009157708A2 (en) 2009-12-30
KR20100002038A (en) 2010-01-06
US8488869B2 (en) 2013-07-16
CN102067615B (en) 2015-02-25
WO2009157714A3 (en) 2010-03-25
WO2009157668A2 (en) 2009-12-30
WO2009157714A2 (en) 2009-12-30
EP2289248A4 (en) 2014-07-02
WO2009157668A3 (en) 2010-03-25
JP2011525746A (en) 2011-09-22
CN102067613B (en) 2016-04-13
CN102077600A (en) 2011-05-25
KR101539935B1 (en) 2015-07-28
EP2292019A2 (en) 2011-03-09
US20090315884A1 (en) 2009-12-24
KR20100002036A (en) 2010-01-06
JP5547725B2 (en) 2014-07-16
JP2011525745A (en) 2011-09-22
US20090315977A1 (en) 2009-12-24
EP2289247A4 (en) 2014-05-28
CN102067613A (en) 2011-05-18

Similar Documents

Publication Publication Date Title
WO2009157701A2 (en) Image generating method and apparatus and image processing method and apparatus
JP4755565B2 (en) Stereoscopic image processing device
TWI644559B (en) Method of encoding a video data signal for use with a multi-view rendering device
US7136415B2 (en) Method and apparatus for multiplexing multi-view three-dimensional moving picture
WO2009157710A2 (en) Image processing method and apparatus
JP4952657B2 (en) Pseudo stereoscopic image generation apparatus, image encoding apparatus, image encoding method, image transmission method, image decoding apparatus, and image decoding method
KR101362941B1 (en) Method and Apparatus for decoding metadata used for playing stereoscopic contents
US20060269226A1 (en) Image receiving apparatus and image reproducing apparatus
CA2713857C (en) Apparatus and method for generating and displaying media files
WO2009157707A2 (en) Image processing method and apparatus
JP4251864B2 (en) Image data creating apparatus and image data reproducing apparatus for reproducing the data
US20090199100A1 (en) Apparatus and method for generating and displaying media files
KR101750047B1 (en) Method for providing and processing 3D image and apparatus for providing and processing 3D image
EP1587330A1 (en) Image data creation device and image data reproduction device for reproducing the data
JP2005110121A (en) Image data display device
US20130070052A1 (en) Video procesing device, system, video processing method, and video processing program capable of changing depth of stereoscopic video images
WO2009157713A2 (en) Image processing method and apparatus
US20150334369A1 (en) Method of encoding a video data signal for use with a multi-view stereoscopic display device
AU2013216395A1 (en) Encoding device and encoding method, and decoding device and decoding method
KR101382618B1 (en) Method for making a contents information and apparatus for managing contens using the contents information
JP2009194758A (en) Video recording apparatus, video reproducing apparatus, video recording/reproducing system, method and program
WO2010050692A2 (en) Image processing method and apparatus
JP2012054733A (en) Reproduction apparatus and reproduction method
EP2685730A1 (en) Playback device, playback method, and program
JP2009194759A (en) Multiplexing device, separation device, processing system, method and program of video stream

Legal Events

Date Code Title Description
WWE Wipo information: entry into national phase

Ref document number: 200980123639.6

Country of ref document: CN

121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 09770373

Country of ref document: EP

Kind code of ref document: A2

WWE Wipo information: entry into national phase

Ref document number: 2009770373

Country of ref document: EP

ENP Entry into the national phase

Ref document number: 2011514502

Country of ref document: JP

Kind code of ref document: A

NENP Non-entry into the national phase

Ref country code: DE