EP2790034A1 - 3D image acquisition apparatus and method of generating depth image in the 3D image acquisition apparatus - Google Patents
3D image acquisition apparatus and method of generating depth image in the 3D image acquisition apparatus Download PDFInfo
- Publication number
- EP2790034A1 EP2790034A1 EP20140163725 EP14163725A EP2790034A1 EP 2790034 A1 EP2790034 A1 EP 2790034A1 EP 20140163725 EP20140163725 EP 20140163725 EP 14163725 A EP14163725 A EP 14163725A EP 2790034 A1 EP2790034 A1 EP 2790034A1
- Authority
- EP
- European Patent Office
- Prior art keywords
- image
- light
- group
- images
- cis
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Granted
Links
- 238000000034 method Methods 0.000 title claims abstract description 56
- 230000005540 biological transmission Effects 0.000 claims abstract description 30
- 230000000295 complement effect Effects 0.000 claims description 3
- 229910044991 metal oxide Inorganic materials 0.000 claims description 3
- 150000004706 metal oxides Chemical class 0.000 claims description 3
- 239000004065 semiconductor Substances 0.000 claims description 3
- 230000003287 optical effect Effects 0.000 description 17
- 230000006870 function Effects 0.000 description 8
- 238000002366 time-of-flight method Methods 0.000 description 7
- 230000008569 process Effects 0.000 description 6
- 238000012545 processing Methods 0.000 description 6
- 230000008859 change Effects 0.000 description 4
- 238000010586 diagram Methods 0.000 description 4
- 239000007787 solid Substances 0.000 description 4
- 101100139845 Caenorhabditis elegans rac-2 gene Proteins 0.000 description 3
- 229910001218 Gallium arsenide Inorganic materials 0.000 description 3
- 230000001934 delay Effects 0.000 description 3
- 238000002310 reflectometry Methods 0.000 description 3
- 238000004891 communication Methods 0.000 description 2
- 239000000463 material Substances 0.000 description 2
- 238000005259 measurement Methods 0.000 description 2
- 230000000737 periodic effect Effects 0.000 description 2
- 230000006978 adaptation Effects 0.000 description 1
- 238000013459 approach Methods 0.000 description 1
- 238000010420 art technique Methods 0.000 description 1
- 238000004364 calculation method Methods 0.000 description 1
- 238000012937 correction Methods 0.000 description 1
- 230000008878 coupling Effects 0.000 description 1
- 238000010168 coupling process Methods 0.000 description 1
- 238000005859 coupling reaction Methods 0.000 description 1
- 238000013500 data storage Methods 0.000 description 1
- 230000003111 delayed effect Effects 0.000 description 1
- 238000001514 detection method Methods 0.000 description 1
- 238000005516 engineering process Methods 0.000 description 1
- 238000000605 extraction Methods 0.000 description 1
- 238000007667 floating Methods 0.000 description 1
- 230000007246 mechanism Effects 0.000 description 1
- 238000012986 modification Methods 0.000 description 1
- 230000004048 modification Effects 0.000 description 1
- 230000010363 phase shift Effects 0.000 description 1
- 230000001737 promoting effect Effects 0.000 description 1
- 238000011160 research Methods 0.000 description 1
- 230000004044 response Effects 0.000 description 1
- 238000012360 testing method Methods 0.000 description 1
Images
Classifications
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N13/00—Stereoscopic video systems; Multi-view video systems; Details thereof
- H04N13/20—Image signal generators
- H04N13/271—Image signal generators wherein the generated image signals comprise depth maps or disparity maps
-
- G—PHYSICS
- G01—MEASURING; TESTING
- G01S—RADIO DIRECTION-FINDING; RADIO NAVIGATION; DETERMINING DISTANCE OR VELOCITY BY USE OF RADIO WAVES; LOCATING OR PRESENCE-DETECTING BY USE OF THE REFLECTION OR RERADIATION OF RADIO WAVES; ANALOGOUS ARRANGEMENTS USING OTHER WAVES
- G01S17/00—Systems using the reflection or reradiation of electromagnetic waves other than radio waves, e.g. lidar systems
- G01S17/02—Systems using the reflection of electromagnetic waves other than radio waves
- G01S17/06—Systems determining position data of a target
- G01S17/08—Systems determining position data of a target for measuring distance only
- G01S17/32—Systems determining position data of a target for measuring distance only using transmission of continuous waves, whether amplitude-, frequency-, or phase-modulated, or unmodulated
- G01S17/36—Systems determining position data of a target for measuring distance only using transmission of continuous waves, whether amplitude-, frequency-, or phase-modulated, or unmodulated with phase comparison between the received signal and the contemporaneously transmitted signal
-
- G—PHYSICS
- G01—MEASURING; TESTING
- G01S—RADIO DIRECTION-FINDING; RADIO NAVIGATION; DETERMINING DISTANCE OR VELOCITY BY USE OF RADIO WAVES; LOCATING OR PRESENCE-DETECTING BY USE OF THE REFLECTION OR RERADIATION OF RADIO WAVES; ANALOGOUS ARRANGEMENTS USING OTHER WAVES
- G01S17/00—Systems using the reflection or reradiation of electromagnetic waves other than radio waves, e.g. lidar systems
- G01S17/88—Lidar systems specially adapted for specific applications
- G01S17/89—Lidar systems specially adapted for specific applications for mapping or imaging
- G01S17/894—3D imaging with simultaneous measurement of time-of-flight at a 2D array of receiver pixels, e.g. time-of-flight cameras or flash lidar
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T5/00—Image enhancement or restoration
- G06T5/50—Image enhancement or restoration using two or more images, e.g. averaging or subtraction
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T5/00—Image enhancement or restoration
- G06T5/73—Deblurring; Sharpening
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N23/00—Cameras or camera modules comprising electronic image sensors; Control thereof
- H04N23/60—Control of cameras or camera modules
- H04N23/68—Control of cameras or camera modules for stable pick-up of the scene, e.g. compensating for camera body vibrations
- H04N23/682—Vibration or motion blur correction
-
- G—PHYSICS
- G01—MEASURING; TESTING
- G01S—RADIO DIRECTION-FINDING; RADIO NAVIGATION; DETERMINING DISTANCE OR VELOCITY BY USE OF RADIO WAVES; LOCATING OR PRESENCE-DETECTING BY USE OF THE REFLECTION OR RERADIATION OF RADIO WAVES; ANALOGOUS ARRANGEMENTS USING OTHER WAVES
- G01S7/00—Details of systems according to groups G01S13/00, G01S15/00, G01S17/00
- G01S7/48—Details of systems according to groups G01S13/00, G01S15/00, G01S17/00 of systems according to group G01S17/00
- G01S7/491—Details of non-pulse systems
- G01S7/4912—Receivers
- G01S7/4913—Circuits for detection, sampling, integration or read-out
- G01S7/4914—Circuits for detection, sampling, integration or read-out of detector arrays, e.g. charge-transfer gates
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/10—Image acquisition modality
- G06T2207/10016—Video; Image sequence
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/10—Image acquisition modality
- G06T2207/10028—Range image; Depth image; 3D point clouds
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/20—Special algorithmic details
- G06T2207/20172—Image enhancement details
- G06T2207/20201—Motion blur correction
Definitions
- Exemplary embodiments relate to three-dimensional (3D) image acquisition apparatuses and methods of generating a depth image by using the 3D image acquisition apparatuses.
- the TOF method In the TOF method, light of a predetermined wavelength, e.g., near-infrared (NIR) (850 nm) is projected to a subject using a light-emitting diode (LED) or a laser diode (LD), and light of the same wavelength reflected from the subject is measured or captured using a photodiode or a camera.
- NIR near-infrared
- the TOF method may extract a depth image.
- Various TOF technologies have been introduced which include a series of various optical processes, such as projection of light from a light source, reflection of a subject, optical modulation, capturing, and processing.
- an optical shutter method is a method in which light of a predetermined wavelength (e.g., NIR 850 nm) is projected to a subject using an LED or LD, and an optical image obtained from light of the same wavelength reflected from the subject is optically modulated using an image intensifier or a predetermined solid state modulator.
- An image is captured using an image sensor, such as a photodiode (PD), a charge-coupled device (CCD), or complementary metal oxide semiconductor (CMOS) image sensor (CIS), included at a rear end of the image intensifier or the solid state modulator.
- An intensity value of the image is processed to obtain a depth image.
- an image intensifier may be used which includes a multi-channel plate (MCP) or a GaAs-based solid state modulator.
- MCP multi-channel plate
- GaAs-based solid state modulator In a related art, a GaAs-based modulator device has been released with improved characteristics and a thin-type modulator using an electrooptic material has been released.
- a method of driving a light source and a light modulator a method of using a particular waveform, such as a triangular wave (ramp wave), and a method using a sine wave have been introduced.
- Noise or signal distortion in a light source, a light modulator, or a capturing device (CCD or CIS) used in 3D cameras usually occurs. Therefore, the noise or signal distortion may cause an error in a depth image.
- a noise component has to be removed using an appropriate device.
- a method of using a higher order of a sine wave, which is a particular waveform, a method of using a code division multiple access (CDMA) method, or a method of removing random error is known to remove noise or signal distortion.
- a 3D camera using a TOF method generates one depth image by sequentially obtaining four infrared (IR) images using driving signals basically having different phases, e.g., 0, 90, 180, and 270 degrees.
- Motion blur is caused when capturing times of the four IR images are different from one another.
- a subject of some pixels (or images) may change, or an object of different depth information instead of the same depth information might be captured, during the whole capturing time of the four images.
- a frequency of motion blur is high on a boundary surface of the moving object. Further, the frequency of motion blur is also high when the object moves fast or when capturing occurs relatively slowly.
- the motion blur in a depth image has different consequences in comparison to a typical motion blur.
- a consequence of the typical motion blur is an average image at the boundary surface of the moving object, e.g., average luminance or color with respect to other objects such as the object and the background thereof.
- divergence is generated in calculation operations in regard to motion blur of a camera for obtaining the depth image.
- unexpected results are derived in regard to the motion blur of the camera. Accordingly, it is important to remove the motion blur.
- Exemplary embodiments may provide three-dimensional (3D) image acquisition apparatuses for improving a rate of capturing a depth image by generating one depth image using a combination of two groups of images.
- Exemplary embodiments may provide methods of generating a depth image in a 3D image acquisition apparatus.
- a depth image is generated after detecting and restoring portions of motion blur, such as an edge, which is inevitably caused, so as to prevent divergence of the depth image due to the motion blur.
- exemplary embodiments may include a light source, a light modulator, or a capturing device, which are used in related art operations for generating a depth image, so no additional costs of additional devices are incurred. Since no additional devices are used, additional computations may be reduced.
- a method of generating a depth image includes: sequentially projecting a light transmission signal, which is generated from a light source, to a subject; modulating reflected light, which is reflected by the subject, using a light modulation signal; calculating a phase delay using a combination of a first plurality of images of two groups, from among a second plurality of images of all groups obtained by capturing the modulated reflected light; and generating a depth image based on the phase delay.
- a method of generating a depth image includes: sequentially projecting a light transmission signal, which is generated in a light source, to a subject; modulating reflected light, which is reflected by the subject, using a light modulation signal; obtaining a first image corresponding to 0 degrees, a second image corresponding to 180 degrees, a third image corresponding to 90 degrees, and a fourth image corresponding to 270 degrees, wherein 0 degrees, 180 degrees, 90 degrees, and 270 degrees are phases of the modulated reflected light; determining whether a motion blur image generated by motion of the subject is present by determining whether a difference between a first sum of images in the first group which includes the first image and the second image and a second sum of images in the second group which includes the third image and the fourth image is equal to or greater than a first critical value; determining, when there is a motion blur image, a group including the motion blur image by comparing a difference of images in the first group and a difference of images in
- a three-dimensional (3D) image acquisition apparatus includes: a light source configured to sequentially project a light transmission signal to a subject; a light modulator configured to modulate reflected light which is reflected by the subject, using a light modulation signal having a gain waveform; a capturing device configured to capture the modulated reflected light, which is modulated by the light modulator, to generate a plurality of second images; and a depth image processor configured to calculate a phase delay using a combination of a plurality of first images of two groups, from among the second plurality of images of all groups generated using the capturing device, and generating a depth image based on the phase delay.
- a method of restoring an image which includes motion blur includes: determining whether a motion blur image occurs in a reflected light signal by determining whether a difference between a first sum of images in a first group and a second sum of images in a second group is equal to or greater than a first critical value; determining a group which includes the motion blur image by comparing a difference of images in the first group and a difference of images in the second group in response to the motion blur image occurring in the reflected light signal; and restoring images of the group which includes the motion blur image using images of the other group which is not the group including the motion blur image.
- the first group includes a first image and a second image of the reflected light signal
- the second group includes a third image and a fourth image of the reflected light signal.
- the first image, the second image, the third image, and the fourth image each have different phases.
- a moving window approach may be used in which one group of images is captured and a previous group of images is removed, from the plurality of sequentially obtained images. It also enables an effective motion blur correction to be implemented.
- FIG. 1 is a schematic structural diagram illustrating a three-dimensional (3D) image acquisition apparatus 100 for generating a depth image using a time-of-flight (TOF) method, according to an embodiment.
- 3D three-dimensional
- the 3D image acquisition apparatus 100 may include a light source 101 that emits light of a predetermined wavelength, a light source driving unit 102 that drives the light source 101, a light modulator 103 that modulates light reflected by a subject 200, a light modulator driving unit 104 that drives the light modulator 103, a capturing device 105 that generates an image from the light modulated by the light modulator 103, a depth image processor 107 that generates a depth image based on an output of the capturing device 105, and a control unit 106 that controls operations of the light source driving unit 102, the light modulator driving unit 104, the capturing device 105, and the depth image processor 107.
- a first lens 108 through which the reflected light is to be focused in an area of the light modulator 103, and a filter 109 that passes through only light having a predetermined wavelength and removes background light or miscellaneous light may be further disposed on a light incident surface of the light modulator 103.
- a second lens 110 that focuses the modulated light in an area of the capturing device 105 may be further disposed between the light modulator 103 and the capturing device 105.
- the light source 101 may be, e.g., a light-emitting diode (LED) or a laser diode (LD), which may emit near infrared (NIR) light having a wavelength of about 850 nm that is not visible to human eyes, for safety.
- NIR near infrared
- the light source driving unit 102 may drive the light source 101 according to a control signal received from the control unit 106, e.g., in an amplitude modulation manner or a phase modulation manner.
- a light transmission signal that is projected to the subject 200 from the light source 101 may have a periodic continuous function having a predetermined period according to a driving signal of the light source driving unit 102.
- the light transmission signal may have a waveform that is specifically defined, such as a sine wave, a ramp wave, or a square wave, but may also be a typical, undefined waveform.
- the light modulator 103 modulates light reflected by the subject 200 according to a control of the light modulator driving unit 104.
- the light modulator driving unit 104 drives the light modulator 103 according to a control signal received from the control unit 106.
- the light modulator 103 may modulate an amplitude of the reflected light by varying a gain of the reflected light according to a light modulation signal having a predetermined wavelength that is provided by the light modulator driving unit 104. Therefore, the light modulator 103 has a variable gain.
- the light modulator 103 may operate at a relatively high light modulation speed of about several tens to several hundreds MHz in order to identify a phase difference or traveling time of light according to distance.
- the light modulator 103 may be, e.g., an image intensifier tube including a multi-channel plate (MCP), a GaAs-based solid state modulator, or a thin-type modulator formed using an electrooptic material. While the light modulator 103 of a transmission type is illustrated in FIG. 1 , a reflective-type light modulator may also be used.
- MCP multi-channel plate
- GaAs-based solid state modulator GaAs-based solid state modulator
- thin-type modulator formed using an electrooptic material. While the light modulator 103 of a transmission type is illustrated in FIG. 1 , a reflective-type light modulator may also be used.
- the capturing device 105 generates an image by detecting the reflected light that is modulated by the light modulator 103 according to a control of the control unit 106.
- the capturing device 105 may be, e.g., a single optical sensor, such as a photodiode or an integrator.
- the capturing device 105 may have one dimension or a two-dimensional array including a plurality of photodiodes or other optical detectors.
- the capturing device 105 may be a charge-coupled device (CCD) image sensor or a complementary metal oxide semiconductor (CMOS) image sensor (CIS) having a two-dimensional array.
- CCD charge-coupled device
- CMOS complementary metal oxide semiconductor
- the depth image processor 107 generates a depth image according to a depth image generating algorithm, based on an output of the capturing device 105.
- the depth image processor 107 may be, e.g., an exclusive integrated circuit (IC), or may be software installed in the 3D image acquisition apparatus 100. When the depth image processor 107 is software, the depth image processor 107 may be stored in a separate portable storage medium.
- IC exclusive integrated circuit
- transmitted light or a light transmission signal of the light source 101 is in the form of a continuous function having a predetermined period, such as a sine wave or a square wave, and has different phases.
- the transmitted light or the light transmission signal is controlled to have the same waveforms but different phases according to a control signal of the control unit 106.
- transmission light of the light source 101 formed to have four different phases has been described above, transmission light of the light source 101 having the same phase and light modulation signals of the light modulator 103 having four different phases may also be formed.
- the light transmission signals of four different phases are sequentially projected in the order of 0, 180, 90, and 270 degrees. This description will focus on sequential projection of light transmission signals having four different phases in the order of 0, 180, 90, and 270 degrees.
- the light source 101 sequentially projects four different beams of transmitted light having predetermined periods and waveforms to the subject 200 according to a control of the control unit 106 and the light source driving unit 102.
- transmitted light 1 (0 degrees) is generated and projected to the subject 200 during a time period T1.
- transmitted light 2 180 degrees is generated and projected to the subject 200 during a next time period T2.
- transmitted light 3 (90 degrees) is generated and projected to the subject 200 during a time period T3.
- transmitted light 4 (270 degrees) is generated and projected to the subject 200 during a time period T4.
- the transmitted lights which are sequentially projected to the subject 200 may be in the form of a continuous function having a predetermined period, such as a sine wave.
- the transmitted light projected to the subject 200 is reflected on a surface of the subject 200
- the transmitted light is incident to the first lens 108.
- the subject 200 typically has multiple surfaces having different distances from the 3D image acquisition apparatus 100, i.e., different depths.
- the subject 200 illustrated in FIG. 1 has five surfaces P1 to P5 with different depths.
- the transmitted light is reflected by each of the five surfaces P1 to P5 having different depths, five reflected light beams that are differently delayed in time (i.e., of different phases) are respectively generated.
- the transmitted light 1 is reflected on the five surfaces P1 to P5 of the subject 200, five beams of the reflected light 1 having different phases are generated.
- the transmitted light 2 is reflected on the five surfaces P1 to P5 of the subject 200, five beams of the reflected light 2 having different phases are generated.
- the transmitted light 4 is reflected by the five surfaces P1 to P5 to generate five beams of the reflected light 4 having different phases.
- the reflected light reflected by the surface P1 that is the farthest from the 3D image acquisition apparatus 100 is phase-delayed by ⁇ P1 to arrive at the first lens 108.
- the reflected light reflected on the surface P5 that is the closest to the 3D image acquisition apparatus 100 is phase-delayed by ⁇ P5 which is less than ⁇ P1 to arrive at the first lens 108.
- the first lens 108 focuses the reflected light in an area of the light modulator 103.
- the filter 109 that transmits only light having a predetermined wavelength, may be disposed between the first lens 108 and the light modulator 103 to remove background light or miscellaneous light other than light of a wavelength used.
- the filter 109 may be an IR band pass filter that transmits the NIR wavelength band of about 850 nm. Accordingly, light incident to the light modulator 103 may be mainly light that is emitted from the light source 101 and reflected by the subject 200.
- the filter 109 is disposed between the first lens 108 and the light modulator 103, the positions of the first lens 108 and the filter 109 may be exchanged. For example, NIR light that has first transmitted through the filter 109 may be focused on the light modulator 103 through the first lens 108.
- the light modulator 103 modulates the reflected light to a light modulation signal having a predetermined wavelength.
- a period of a gain waveform in the light modulator 103 may be the same as a period of a waveform.
- the light modulator 103 may modulate the five beams of reflected light 1 that are respectively reflected by the five surfaces P1 to P5 of the subject 200 and provide the same to the capturing device 105.
- the light modulator 103 may also sequentially modulate the five beams of each of the reflected light 2 through 5 and provide the same to the capturing device 105.
- the capturing device 105 generates an image by receiving the modulated light during an exposure time. For example, as indicated in FIG. 2A , the capturing device 105 receives the five beams of the reflected light 1, which are respectively reflected on the five surfaces P1 to P5 of the subject 200 and then modulated, for a predetermined period of exposure time to generate an image 1 (CIS image 1). Next, as illustrated in FIG.
- the capturing device 105 receives the five beams of the reflected light 2, which are respectively reflected on the five surfaces P1 to P5 of the subject 200 and then modulated, for a predetermined period of exposure time to generate an image 2 (CIS image 2).
- the capturing device 105 receives the five beams of the reflected light 4, which are respectively reflected on the five surfaces P1 to P5 of the subject 200 and then modulated, for a predetermined period of exposure time to generate an image 4 (CIS image 4).
- the images 1 through 4 may each be a sub-frame image for generating an image of a frame having depth information. For example, when assuming a period of a frame of depth information to be Td, an exposure time in the capturing device 105 to obtain the four images 1 through 4 may be about Td/4.
- the transmitted light 1 projected from the light source 101 to the subject 200 is reflected by the five surfaces P1 to P5 of the subject 200 to generate five beams of the reflected light 1.
- the five beams of the reflected light 1 are modulated using the light modulator 103, and then arrive at the capturing device 105.
- the capturing device 105 includes only five pixels respectively corresponding to the five surfaces P1 to P5. Accordingly, the five beams of the reflected light 1 may be respectively incident on the corresponding five pixels. As illustrated in FIG.
- the five beams of the reflected light 1 respectively reflected on the surfaces P1 to P5 have different phase delays ⁇ P1 to ⁇ P5 according to the distance from the 3D image acquisition apparatus 100 to the surfaces P1 to P5.
- the capturing device 105 may generate an image 1 by capturing the reflected light 1 during an exposure time of about Td/4.
- different phase delays ⁇ P1 to ⁇ P5 are generated on the five surfaces P1 to P5 with different depths from a second sub-frame to a fourth sub-frame.
- FIG. 3 illustrates an operation of generating four different images (sub-frames) after modulating the reflected light using four different light modulation signals in the light modulator 103, according to another embodiment.
- the reflected light reflected by the subject 200 has the same waveform and phase in each sub-frame.
- phase delays ⁇ P1 to ⁇ P5 in the reflected light of each sub-frame there are different phase delays ⁇ P1 to ⁇ P5 in the reflected light of each sub-frame according to the surfaces P1 to P5 of the subject 200.
- the light modulator 103 modulates the reflected light using a light modulation signal 1 (0 degrees).
- the light modulator 103 modulates the reflected light using a light modulation signal 2 (180 degrees) having a different phase from that of the light modulation signal 1.
- the light modulator 103 modulates the reflected light to a light modulation signal 3 (90 degrees) having another different phase.
- the light modulator 103 modulates the reflected light to another light modulation signal 4 (270 degrees).
- the light modulation signals 1 through 4 are signals having different phases of 0, 180, 90, and 270 degrees.
- the light modulation signals 1 through 4 may have the same periods and waveforms.
- four different images 1 through 4 may be obtained.
- the modulated images are labeled as CIS images 1 through 4, as described above, the capturing device 105 is not limited to a CIS.
- the four images obtained by projecting the light transmission signals having different phases (0, 180, 90, and 270 degrees) as illustrated in FIG. 2 or by using the light modulation signals 1 through 4 having different phases (0, 180, 90, and 270 degrees) as illustrated in FIG. 3 , are transmitted to the depth image processor 107.
- the depth image processor 107 generates a depth image or depth information according to an algorithm, based on the four images. Also, according to another embodiment, in order to effectively remove motion blur that has occurred due to motion of a subject, the depth image processor 107 performs an algorithm for determining whether motion blur has occurred in the four images (the first image corresponding to 0 degrees, the second image corresponding to 180 degrees, the third image corresponding to the third image, and the fourth image corresponding to 270 degrees). When motion blur is generated, an algorithm is performed in the depth image processor 107 for restoring an image or a pixel in which the motion blur is generated.
- FIG. 4 illustrates an operation of generating a depth image by using a moving window method according to another embodiment.
- the depth image processor 107 uses the depth image processor 107 to generate a depth image.
- the four captured images I CIS 0 , I CIS 180 , I CIS 90 , I CIS 270 have different phases of 0, 180, 90, and 270 degrees in regard to a light source and are labeled CIS ("CMOS Image Sensor").
- the capturing device 105 illustrated in FIG. 1 (not limited to a CIS) generates a depth image based on the four intensity images 1 through 4.
- a subject is captured in an order of 0, 180, 90, and 270 degrees to obtain consecutive images.
- the four images including I CIS 0 , I CIS 180 , I CIS 90 , I CIS 270 are combined (400) such that two new images are captured and two prior images are sequentially removed.
- the four images obtained by using the moving window method are divided into two groups, i.e., first and second groups U and V (410).
- the first group U is divided into I CIS 0 and I CIS 180
- the second group V is divided into I CIS 90 and I CIS 270 .
- I CIS 0 and I CIS 180 form a first group of images (the first group having a first plurality, i.e. 2, of images) and I CIS 90 and I CIS 270 form a second group of images (the second group also having a first plurality, i.e. 2, of images).
- the images are all from a second plurality of images, i.e. the set of all four images.
- Intermediate images 411 and 412 illustrated in FIG. 4 are formed based on a difference between images of the first group U and images of the second group V.
- a depth image is calculated according to Equation 3 using the combination of the images of Equation 2.
- the first group U includes I CIS 0 and I CIS 180
- the second group V includes I CIS 90 and I CIS 270 .
- ⁇ depth c 4 ⁇ ⁇ f ⁇ tan - 1 ⁇ - I CIS 90 + I CIS 270 I CIS 0 - I CIS 180
- one depth image may be obtained in a time period for obtaining two images (i.e., in a moving window method in which two images are captured and two previously captured images are removed from among a plurality of sequentially obtained images). This is advantageous in that a depth image may be obtained at a higher speed than an example where one depth image is obtained after four images are captured in the related art.
- a usage range of the embodiments is not limited by whether pixels are arranged in a one-dimensional array or whether a single pixel is used.
- a photodiode or an integrator may be used instead of a CIS.
- a method of removing motion blur in a 3D image acquisition apparatus that is capable of generating a TOF-based depth image.
- the essential reason that motion blur is caused is an abrupt change in luminance of a subject or surface in an IR image capturing section of four IR images.
- a method of detecting (determining) a pixel (or image) in which motion blur has occurred will be described.
- a method of detecting (determining) a portion of the section where the subject does not change to generate a depth image based on the detected portion and remove divergence in an operation due to the motion blur will be described.
- FIG. 5 illustrates a flow of an optical process according to another embodiment.
- FIGS. 6A and 6B are graphs for explaining output waveforms of a light source and a light modulator according to another embodiment.
- Infrared light that has transmitted through an IR filter (e.g., a central wavelength of 850 nm) arrives at the light modulator 103 and is modulated to arrive at the capturing device 105 at a rear end of the 3D image acquisition apparatus 100.
- an IR filter e.g., a central wavelength of 850 nm
- FIG. 6A is a waveform diagram of a light output waveform
- FIG. 6B is a driving waveform diagram of a light modulator.
- the waveform may be various periodic waves, such as a sine wave, a square wave, or a ramp wave.
- a light output wave light transmission signal
- a driving waveform of a light modulator is a sine wave
- the infrared light output is in the form of combination of a square waveform (AC component) with DC offset of output light, which is a DC component.
- AC component square waveform
- DC offset of output light which is a DC component.
- s is 1 to 4 and the infrared light has different phases such as 0 degrees, 180 degrees, 90 degrees, and 270 degrees.
- the reflected light returns after being reflected on a surface of a subject.
- the reflected light is in the form of a square of reflectivity r in which surface reflectivity of the subject, a lens size, etc., are in a comprehensive consideration, and a phase difference, due to a TOF method is present, and external light is also present, similar to the infrared light.
- a modulation waveform (gain) or a light modulation signal of a light modulator (optical shutter) is expressed in Equation 6 below:
- G t c ⁇ sin ⁇ t + G ⁇ ave
- the light modulation signal is in the form of a combination of a sine wave (AC component) and a DC component.
- An output signal (or output image) emitted from the capturing device is in the form obtained by integrating the optical signal of Equation 7, and is expressed as Equation 8 below:
- Equation 9 rA cos ⁇ TOF + rB
- I CIS 180 - rA cos ⁇ TOF + rB
- I CIS 90 - rA sin ⁇ TOF + rB
- I CIS 270 rA sin ⁇ TOF + rB
- Equations 9 through 15 are formulas that are established under the assumption of an unchanged subject and surface reflectivity.
- Equations 9 through 12 are used in detecting and restoring motion blur.
- FIGS. 7A and 7B are schematic views for explaining a type of motion blur generated in sub-frames captured by using a capturing device according to another embodiment. A method of detecting a motion blur pixel (image) will be described with reference to FIGS. 7A and 7B .
- FIG. 7A illustrates an example where motion blur is generated in a pixel while a subject changes from an object A to an object B.
- motion blur may be divided as illustrated in FIG. 7B .
- a stable depth image which is the same as a still image, may be generated without motion blur when four images of the same object, either the object A or the object B is captured (case 0 and case 5), are captured.
- Equation 18 a discriminant for a motion blur pixel (image) in Equation 18 below (an example where no motion blur occurs) and Equation 19 below (an example where motion blur occurs) may be obtained.
- FIGS. 8A and 8B are schematic views for explaining an operation of restoring motion blur by using an image processor according to another embodiment. A method of restoring a motion blur pixel will be described with reference to FIGS. 8A and 8B .
- a constant k refers to a demodulation contrast which indicates modulation characteristics of a TOF system.
- Equations 9 through 12 a size of an infrared ray image due to a phase delay is expressed by sine and cosine functions. Further, an AC value and DC offset values are read by floating the functions to experimentally calculate A and B and a ratio k between A and B.
- I CIS 90 - I CIS 270 may be expressed in Equations 27 and 28 below based on I CIS 0 - I CIS 180 and the known system constant k:
- I CIS 90 - I CIS 270 2 k 2 ⁇ I CIS 0 + I CIS 180 2 - I CIS 0 - I CIS 180 2
- I CIS 90 - I CIS 270 sign ⁇ I CIS 90 - I CIS 270 ⁇ k 2 ⁇ I CIS 0 + I CIS 180 2 - I CIS 0 - I CIS 180 2
- Equation 31 it is determined that motion blur has occurred in I CIS 0 or I CIS 180 . If Equation 32 is satisfied, it is determined that motion blur has occurred in I CIS 90 or I CIS 180 .
- FIG. 9 is a flowchart illustrating a method of generating a depth image in a 3D image acquisition apparatus, according to another embodiment.
- a plurality of light transmission signals generated in a light source is sequentially transmitted to a subject.
- the light transmission signal may have different phases, such as 0, 180, 90, and 270 degrees or may have the same phase.
- light that is reflected by the subject is modulated using a light modulation signal.
- light modulation signals may have the same phase. If the light transmission signal is projected at the same phase, the reflected light is modulated using light modulation signals corresponding to the different phases, i.e., 0, 180, 90, and 270 degrees.
- a phase delay is calculated using a combination of two groups of images from among a plurality of images obtained by capturing the modulated reflected light.
- a phase delay according to a depth is calculated using a combination of two images corresponding to 0 degrees and 180 degrees and two images corresponding to 90 degrees and 270 degrees from among the modulated light.
- a depth image is generated based on the phase delay.
- FIG. 10 is a flowchart illustrating a method of generating a depth image in a 3D image acquisition apparatus, according to another embodiment.
- a plurality of light transmission signals generated in a light source is sequentially projected to a subject.
- light reflected by the subject is modulated using a light modulation signal.
- a light transmission signal from the light source or a light modulation signal from a light modulator may have different phases, such as 0, 180, 90, and 270 degrees or may have the same phases.
- first through fourth images corresponding to phases of the modulated reflected light i.e., 0, 180, 90, and 270 degrees
- a fourth image corresponding to 270 degrees are sequentially obtained.
- four images are maintained such that two images (e.g., images corresponding to 90 degrees and 270 degrees) are captured, and two previously captured images (e.g., images corresponding to 0 degrees and 180 degrees) are removed.
- a difference, between a sum of images in the first group including the first and second images and a sum of images in the second group including the third and fourth images, is equal to or greater than a first critical value is determined.
- the first critical value may be optionally determined according to a noise level of system or scenes.
- the sum of images in the first group and the sum of images in the second group are not identical.
- the difference between the first group and the second group is equal to or greater than the first critical value, it indicates that motion blur is present. If there is no motion blur (if the sum of images in the first group and the sum of images in the second group are identical), the method proceeds to operations 1014 and 1016 to calculate a phase delay using the images of the first group and the second group to generate a depth image.
- operations 1008 and 1010 if there is motion blur, it is determined in which of the first group and the second group the motion blur is present. In operation 1006, if there is motion blur in the first through fourth images, it is determined in which of the first group and the second group the motion blur is present.
- the presence of motion blur indicates a large variation in luminance of an image. Therefore, it is determined that the motion blur is generated in a group with a large variation in luminance. Variation in luminance is determined in a manner above with reference to Equations 31 and 32.
- images of the group with motion blur are restored using the images of the other group.
- a depth image is generated by including the images of the group with motion blur, divergence is generated in resultant images. Thus, a totally different result is obtained. Accordingly, a phase delay and a depth image are generated after restoring the images with motion blur.
- the demodulation constant which is a constant of Equation 24 above, or the system variable k as described above may be measured and calculated in an image test.
- a size of an image due to a phase delay as in Equations 9 through 12 is expressed by sine and cosine functions, and the image is expressed by a graph to read AC and DC offset values thereof, thereby experimentally calculating A and B and a ratio k therebetween.
- a phase delay is calculated and a depth image is generated.
- a result of examining the method of removing motion blur according to the current embodiment by applying the same to an actual image will be described.
- a demodulation constant k of a system was measured, and a depth image was generated using continuous infrared ray images. Examples where motion blur is not removed and motion blur is removed were respectively calculated, thereby comparing the accuracy of the depth image and whether the depth image is converged.
- FIGS. 11A and 11B are images for explaining motion blur generated when capturing a moving subject. Referring to FIG. 11A , images corresponding to 0, 180, 90, and 270 degrees are illustrated, showing a bar-shaped object in a rotated state. As can be seen from the overlapped images of FIG. 11B , an image around an edge is blurred.
- FIG. 12 illustrates a result of applying Equation 19 to the images of FIGS. 11A and 11B in order to determine motion blur, according to another embodiment.
- FIG. 12 shows a result of applying the discriminant for determining motion blur in Equation 19, to the image illustrated in FIG. 11 .
- the image has mostly 0 values but relatively high values in edge portions thereof, thereby accurately searching motion blurred portions.
- FIGS. 13A and 13B illustrate a depth image in which motion blur is generated and a depth image for which the motion blur is restored, according to another embodiment.
- FIG. 13A illustrates an original image before motion blur is restored
- FIG. 13B is an image with restored motion blur according to an embodiment.
- the depth image is significantly diverged before a motion blur restoration algorithm is applied.
- an unexpected value is given (see FIG. 13A ).
- the depth image is stabilized and a physically meaningful result is given (see FIG. 13B ).
- a depth image is generated based on a combination of two groups of images, thereby increasing a rate of capturing the depth image.
- a depth image is generated after detecting and restoring inevitable motion blur in portions such as an edge.
- divergence of the depth image due to the motion blur may be prevented.
- a light source, a light modulator, or a capturing device used in related art operations for generating a depth image are used, no additional costs for additional devices are necessary, and additional amount of computation may be reduced.
- the device described herein may comprise a processor, a memory for storing program data and executing it, a permanent storage such as a disk drive, a communications port for handling communication with external devices, and user interface devices, including a display, keys, etc.
- software modules When software modules are involved, these software modules may be stored as program instructions or computer-readable codes executable on the processor on a computer-readable media such as read-only memory (ROM), random-access memory (RAM), CD-ROMs, magnetic tapes, floppy disks, and optical data storage devices.
- ROM read-only memory
- RAM random-access memory
- CD-ROMs compact discs
- magnetic tapes magnetic tapes
- floppy disks floppy disks
- optical data storage devices optical data storage devices.
- the computer-readable recording medium can also be distributed over network-coupled computer systems so that the computer-readable code is stored and executed in a distributed fashion. This media can be read by the computer, stored in the memory, and executed by the processor.
- Exemplary embodiments may be described in terms of functional block components and various processing steps. Such functional blocks may be realized by any number of hardware and/or software components configured to perform the specified functions.
- the exemplary embodiments may employ various integrated circuit components, e.g., memory elements, processing elements, logic elements, look-up tables, etc., which may carry out a variety of functions under the control of one or more microprocessors or other control devices.
- the exemplary embodiments may be implemented with any programming or scripting language such as C, C++, Java, assembler, etc., with the various algorithms being implemented with any combination of data structures, objects, processes, routines or other programming elements.
Landscapes
- Engineering & Computer Science (AREA)
- Physics & Mathematics (AREA)
- General Physics & Mathematics (AREA)
- Electromagnetism (AREA)
- Computer Networks & Wireless Communication (AREA)
- Radar, Positioning & Navigation (AREA)
- Remote Sensing (AREA)
- Theoretical Computer Science (AREA)
- Signal Processing (AREA)
- Multimedia (AREA)
- Length Measuring Devices By Optical Means (AREA)
- Optical Radar Systems And Details Thereof (AREA)
- Testing, Inspecting, Measuring Of Stereoscopic Televisions And Televisions (AREA)
- Measurement Of Optical Distance (AREA)
Abstract
Description
- Exemplary embodiments relate to three-dimensional (3D) image acquisition apparatuses and methods of generating a depth image by using the 3D image acquisition apparatuses.
- Research has been increasing in the fields of three-dimensional (3D) cameras or light detection and ranging (LIDAR) techniques for capturing distance information of an object. Further, one of these techniques is a time-of-flight (TOF) method which measures a traveling time of light by measuring a distance between a capturing unit and a subject (hereinafter referred to as "depth image").
- In the TOF method, light of a predetermined wavelength, e.g., near-infrared (NIR) (850 nm) is projected to a subject using a light-emitting diode (LED) or a laser diode (LD), and light of the same wavelength reflected from the subject is measured or captured using a photodiode or a camera. Thus, the TOF method may extract a depth image. Various TOF technologies have been introduced which include a series of various optical processes, such as projection of light from a light source, reflection of a subject, optical modulation, capturing, and processing.
- For example, an optical shutter method is a method in which light of a predetermined wavelength (e.g., NIR 850 nm) is projected to a subject using an LED or LD, and an optical image obtained from light of the same wavelength reflected from the subject is optically modulated using an image intensifier or a predetermined solid state modulator. An image is captured using an image sensor, such as a photodiode (PD), a charge-coupled device (CCD), or complementary metal oxide semiconductor (CMOS) image sensor (CIS), included at a rear end of the image intensifier or the solid state modulator. An intensity value of the image is processed to obtain a depth image. To identify a phase difference or traveled time according to the distance of light, a high light modulation rate of about several tens to several hundreds of MHz is required. To obtain the high light modulation rate, an image intensifier may be used which includes a multi-channel plate (MCP) or a GaAs-based solid state modulator. In a related art, a GaAs-based modulator device has been released with improved characteristics and a thin-type modulator using an electrooptic material has been released.
- In regard to an optical process for extracting a depth image or depth information, a method of driving a light source and a light modulator, a method of using a particular waveform, such as a triangular wave (ramp wave), and a method using a sine wave have been introduced. Methods of driving a light source and a light modulator according to respective waveforms and methods of extracting and calculating a depth image based on intensity values of captured images, i.e., various depth algorithms, have been introduced. Noise or signal distortion in a light source, a light modulator, or a capturing device (CCD or CIS) used in 3D cameras usually occurs. Therefore, the noise or signal distortion may cause an error in a depth image. To remove the noise or signal distortion, a noise component has to be removed using an appropriate device. In the related art, a method of using a higher order of a sine wave, which is a particular waveform, a method of using a code division multiple access (CDMA) method, or a method of removing random error is known to remove noise or signal distortion.
- In a related art, a 3D camera using a TOF method generates one depth image by sequentially obtaining four infrared (IR) images using driving signals basically having different phases, e.g., 0, 90, 180, and 270 degrees. Motion blur is caused when capturing times of the four IR images are different from one another. In other words, as four IR images of a moving object are sequentially captured, a subject of some pixels (or images) may change, or an object of different depth information instead of the same depth information might be captured, during the whole capturing time of the four images. A frequency of motion blur is high on a boundary surface of the moving object. Further, the frequency of motion blur is also high when the object moves fast or when capturing occurs relatively slowly.
- The motion blur in a depth image has different consequences in comparison to a typical motion blur. A consequence of the typical motion blur is an average image at the boundary surface of the moving object, e.g., average luminance or color with respect to other objects such as the object and the background thereof. However, instead of the average of depth images obtained for the typical motion blur, divergence is generated in calculation operations in regard to motion blur of a camera for obtaining the depth image. Thus, unexpected results are derived in regard to the motion blur of the camera. Accordingly, it is important to remove the motion blur.
- Exemplary embodiments may provide three-dimensional (3D) image acquisition apparatuses for improving a rate of capturing a depth image by generating one depth image using a combination of two groups of images.
- Exemplary embodiments may provide methods of generating a depth image in a 3D image acquisition apparatus. When a moving object is captured using a time-of-flight (TOF) method, a depth image is generated after detecting and restoring portions of motion blur, such as an edge, which is inevitably caused, so as to prevent divergence of the depth image due to the motion blur. Further, exemplary embodiments may include a light source, a light modulator, or a capturing device, which are used in related art operations for generating a depth image, so no additional costs of additional devices are incurred. Since no additional devices are used, additional computations may be reduced.
- Additional aspects will be set forth in part in the description which follows and, in part, will be apparent from the description, or may be learned by practice of the presented embodiments.
- According to an aspect of the exemplary embodiments, a method of generating a depth image, includes: sequentially projecting a light transmission signal, which is generated from a light source, to a subject; modulating reflected light, which is reflected by the subject, using a light modulation signal; calculating a phase delay using a combination of a first plurality of images of two groups, from among a second plurality of images of all groups obtained by capturing the modulated reflected light; and generating a depth image based on the phase delay.
- According to another aspect of the exemplary embodiments, a method of generating a depth image, includes: sequentially projecting a light transmission signal, which is generated in a light source, to a subject; modulating reflected light, which is reflected by the subject, using a light modulation signal; obtaining a first image corresponding to 0 degrees, a second image corresponding to 180 degrees, a third image corresponding to 90 degrees, and a fourth image corresponding to 270 degrees, wherein 0 degrees, 180 degrees, 90 degrees, and 270 degrees are phases of the modulated reflected light; determining whether a motion blur image generated by motion of the subject is present by determining whether a difference between a first sum of images in the first group which includes the first image and the second image and a second sum of images in the second group which includes the third image and the fourth image is equal to or greater than a first critical value; determining, when there is a motion blur image, a group including the motion blur image by comparing a difference of images in the first group and a difference of images in the second group, and restoring the images of the group including the motion blur image using images of the other group; calculating a phase delay using the images of the other group and the images of the restored group; and generating a depth image based on the phase delay.
- According to another aspect of the exemplary embodiments, a three-dimensional (3D) image acquisition apparatus includes: a light source configured to sequentially project a light transmission signal to a subject; a light modulator configured to modulate reflected light which is reflected by the subject, using a light modulation signal having a gain waveform; a capturing device configured to capture the modulated reflected light, which is modulated by the light modulator, to generate a plurality of second images; and a depth image processor configured to calculate a phase delay using a combination of a plurality of first images of two groups, from among the second plurality of images of all groups generated using the capturing device, and generating a depth image based on the phase delay.
- According to another aspect of the exemplary embodiments, a method of restoring an image which includes motion blur includes: determining whether a motion blur image occurs in a reflected light signal by determining whether a difference between a first sum of images in a first group and a second sum of images in a second group is equal to or greater than a first critical value; determining a group which includes the motion blur image by comparing a difference of images in the first group and a difference of images in the second group in response to the motion blur image occurring in the reflected light signal; and restoring images of the group which includes the motion blur image using images of the other group which is not the group including the motion blur image. The first group includes a first image and a second image of the reflected light signal, and the second group includes a third image and a fourth image of the reflected light signal. The first image, the second image, the third image, and the fourth image each have different phases.
- By dividing the images into two groups, and combining the groups, it becomes possible to reduce the time required to obtain a depth image. For example, a moving window approach may be used in which one group of images is captured and a previous group of images is removed, from the plurality of sequentially obtained images. It also enables an effective motion blur correction to be implemented.
- These and/or other aspects will become apparent and more readily appreciated from the following description of the embodiments, taken in conjunction with the accompanying drawings in which:
-
FIG. 1 is a schematic structural diagram illustrating a three-dimensional (3D) image acquisition apparatus for generating a depth image by using a time-of-flight (TOF) method, according to an embodiment; -
FIGS. 2A through 2D illustrate an operation of generating four different images (sub-frames) by using a capturing device after projecting four different light transmission signals to a subject from a light source, according to another embodiment; -
FIGS. 3A through 3D illustrate an operation of generating four different images (sub-frames) after modulating reflected light by using four different light modulation signals by using a light modulator, according to another embodiment; -
FIG. 4 illustrates an operation of generating a depth image by using a moving window method according to another embodiment; -
FIG. 5 illustrates a flow of an optical process according to another embodiment; -
FIGS. 6A and 6B are graphs for explaining output waveforms of a light source and a light modulator according to another embodiment; -
FIGS. 7A and 7B are schematic views for explaining types of motion blur generated in sub-frames captured by using a capturing device according to another embodiment; -
FIGS. 8A and8B are schematic views for explaining an operation of restoring motion blur by using an image processor according to another embodiment ; -
FIG. 9 is a flowchart illustrating a method of generating a depth image by using a 3D image acquisition apparatus, according to another embodiment; -
FIG. 10 is a flowchart illustrating a method of generating a depth image by using a 3D image acquisition apparatus, according to another embodiment; -
FIGS. 11A and11B are images for explaining motion blur occurring when capturing a moving subject; -
FIG. 12 illustrates a result of applying Equation 19 to the images ofFIGS. 11A and11B in order to determine motion blur according to another embodiment; and -
FIGS. 13A and 13B illustrate a depth image in which motion blur is generated and a depth image for which the motion blur is restored, according to another embodiment. - Reference will now be made in detail to embodiments, examples of which are illustrated in the accompanying drawings, wherein like reference numerals refer to like elements throughout. In this regard, the present embodiments may have different forms and should not be construed as being limited to the descriptions set forth herein. Accordingly, the embodiments are merely described below, by referring to the figures, to explain aspects of the present description. As used herein, the term "and/or" includes any and all combinations of one or more of the associated listed items.
- As the exemplary embodiments allow for various changes and numerous embodiments, particular embodiments will be illustrated in the drawings and described in detail in the written description. However, this is not intended to limit the exemplary embodiments to particular modes of practice, and it is to be appreciated that all changes, equivalents, and substitutes that do not depart from the spirit and technical scope are encompassed in the exemplary embodiments. In the description of the exemplary embodiments, certain detailed explanations of related art are omitted when it is deemed that they may unnecessarily obscure the essence of the exemplary embodiments.
- While such terms as "first, "second", etc., may be used to describe various components, such components must not be limited to the above terms. The above terms are used only to distinguish one component from another.
- The terms used in the present specification are merely used to describe particular embodiments, and are not intended to limit the exemplary embodiments. An expression used in the singular encompasses the expression of the plural, unless it has a clearly different meaning in the context. In the present specification, it is to be understood that the terms such as "including" or "having," etc., are intended to indicate the existence of the features, numbers, steps, actions, components, parts, or combinations thereof disclosed in the specification, and are not intended to preclude the possibility that one or more other features, numbers, steps, actions, components, parts, or combinations thereof may exist or may be added.
- The embodiments will be described below in more detail with reference to the accompanying drawings. Those components that are the same or are in correspondence are rendered the same reference numeral regardless of the figure number, and redundant explanations are omitted.
-
FIG. 1 is a schematic structural diagram illustrating a three-dimensional (3D)image acquisition apparatus 100 for generating a depth image using a time-of-flight (TOF) method, according to an embodiment. - Referring to
FIG. 1 , the 3Dimage acquisition apparatus 100 may include alight source 101 that emits light of a predetermined wavelength, a lightsource driving unit 102 that drives thelight source 101, alight modulator 103 that modulates light reflected by a subject 200, a lightmodulator driving unit 104 that drives thelight modulator 103, acapturing device 105 that generates an image from the light modulated by thelight modulator 103, adepth image processor 107 that generates a depth image based on an output of thecapturing device 105, and acontrol unit 106 that controls operations of the lightsource driving unit 102, the lightmodulator driving unit 104, thecapturing device 105, and thedepth image processor 107. Also, afirst lens 108, through which the reflected light is to be focused in an area of thelight modulator 103, and afilter 109 that passes through only light having a predetermined wavelength and removes background light or miscellaneous light may be further disposed on a light incident surface of thelight modulator 103. In addition, asecond lens 110 that focuses the modulated light in an area of thecapturing device 105 may be further disposed between thelight modulator 103 and thecapturing device 105. - The
light source 101 may be, e.g., a light-emitting diode (LED) or a laser diode (LD), which may emit near infrared (NIR) light having a wavelength of about 850 nm that is not visible to human eyes, for safety. However, the bands of wavelengths and types of light sources are not limited thereto. The lightsource driving unit 102 may drive thelight source 101 according to a control signal received from thecontrol unit 106, e.g., in an amplitude modulation manner or a phase modulation manner. A light transmission signal that is projected to the subject 200 from thelight source 101 may have a periodic continuous function having a predetermined period according to a driving signal of the lightsource driving unit 102. For example, the light transmission signal may have a waveform that is specifically defined, such as a sine wave, a ramp wave, or a square wave, but may also be a typical, undefined waveform. - The
light modulator 103 modulates light reflected by the subject 200 according to a control of the lightmodulator driving unit 104. The lightmodulator driving unit 104 drives thelight modulator 103 according to a control signal received from thecontrol unit 106. For example, thelight modulator 103 may modulate an amplitude of the reflected light by varying a gain of the reflected light according to a light modulation signal having a predetermined wavelength that is provided by the lightmodulator driving unit 104. Therefore, thelight modulator 103 has a variable gain. Thelight modulator 103 may operate at a relatively high light modulation speed of about several tens to several hundreds MHz in order to identify a phase difference or traveling time of light according to distance. Thelight modulator 103 may be, e.g., an image intensifier tube including a multi-channel plate (MCP), a GaAs-based solid state modulator, or a thin-type modulator formed using an electrooptic material. While thelight modulator 103 of a transmission type is illustrated inFIG. 1 , a reflective-type light modulator may also be used. - The
capturing device 105 generates an image by detecting the reflected light that is modulated by thelight modulator 103 according to a control of thecontrol unit 106. When only a distance to a predetermined point of the subject 200 is to be measured, thecapturing device 105 may be, e.g., a single optical sensor, such as a photodiode or an integrator. However, when distances to multiple points on the subject 200 are to be measured at the same time, thecapturing device 105 may have one dimension or a two-dimensional array including a plurality of photodiodes or other optical detectors. For example, thecapturing device 105 may be a charge-coupled device (CCD) image sensor or a complementary metal oxide semiconductor (CMOS) image sensor (CIS) having a two-dimensional array. - The
depth image processor 107 generates a depth image according to a depth image generating algorithm, based on an output of thecapturing device 105. Thedepth image processor 107 may be, e.g., an exclusive integrated circuit (IC), or may be software installed in the 3Dimage acquisition apparatus 100. When thedepth image processor 107 is software, thedepth image processor 107 may be stored in a separate portable storage medium. - Hereinafter, an operation of the 3D
image acquisition apparatus 100 having the above-described structure will be briefly described. According to an embodiment, transmitted light or a light transmission signal of thelight source 101 is in the form of a continuous function having a predetermined period, such as a sine wave or a square wave, and has different phases. In other words, the transmitted light or the light transmission signal is controlled to have the same waveforms but different phases according to a control signal of thecontrol unit 106. While the transmitted light of thelight source 101 formed to have four different phases has been described above, transmission light of thelight source 101 having the same phase and light modulation signals of thelight modulator 103 having four different phases may also be formed. Also, the light transmission signals of four different phases are sequentially projected in the order of 0, 180, 90, and 270 degrees. This description will focus on sequential projection of light transmission signals having four different phases in the order of 0, 180, 90, and 270 degrees. - The
light source 101 sequentially projects four different beams of transmitted light having predetermined periods and waveforms to the subject 200 according to a control of thecontrol unit 106 and the lightsource driving unit 102. For example, transmitted light 1 (0 degrees) is generated and projected to the subject 200 during a time period T1. Then, transmitted light 2 (180 degrees) is generated and projected to the subject 200 during a next time period T2. Then, transmitted light 3 (90 degrees) is generated and projected to the subject 200 during a time period T3. Then, transmitted light 4 (270 degrees) is generated and projected to the subject 200 during a time period T4. The transmitted lights which are sequentially projected to the subject 200 may be in the form of a continuous function having a predetermined period, such as a sine wave. - After the transmitted light projected to the subject 200 is reflected on a surface of the subject 200, the transmitted light is incident to the
first lens 108. The subject 200 typically has multiple surfaces having different distances from the 3Dimage acquisition apparatus 100, i.e., different depths. For a simplified description, the subject 200 illustrated inFIG. 1 has five surfaces P1 to P5 with different depths. As the transmitted light is reflected by each of the five surfaces P1 to P5 having different depths, five reflected light beams that are differently delayed in time (i.e., of different phases) are respectively generated. For example, as the transmittedlight 1 is reflected on the five surfaces P1 to P5 of the subject 200, five beams of the reflectedlight 1 having different phases are generated. As the transmittedlight 2 is reflected on the five surfaces P1 to P5 of the subject 200, five beams of the reflectedlight 2 having different phases are generated. Also, the transmittedlight 4 is reflected by the five surfaces P1 to P5 to generate five beams of the reflectedlight 4 having different phases. The reflected light reflected by the surface P1 that is the farthest from the 3Dimage acquisition apparatus 100 is phase-delayed by ΦP1 to arrive at thefirst lens 108. The reflected light reflected on the surface P5 that is the closest to the 3Dimage acquisition apparatus 100 is phase-delayed by ΦP5 which is less than ΦP1 to arrive at thefirst lens 108. - The
first lens 108 focuses the reflected light in an area of thelight modulator 103. Thefilter 109, that transmits only light having a predetermined wavelength, may be disposed between thefirst lens 108 and thelight modulator 103 to remove background light or miscellaneous light other than light of a wavelength used. For example, when thelight source 101 emits light having an NIR wavelength of about 850 nm, thefilter 109 may be an IR band pass filter that transmits the NIR wavelength band of about 850 nm. Accordingly, light incident to thelight modulator 103 may be mainly light that is emitted from thelight source 101 and reflected by the subject 200. While thefilter 109 is disposed between thefirst lens 108 and thelight modulator 103, the positions of thefirst lens 108 and thefilter 109 may be exchanged. For example, NIR light that has first transmitted through thefilter 109 may be focused on thelight modulator 103 through thefirst lens 108. - The
light modulator 103 modulates the reflected light to a light modulation signal having a predetermined wavelength. A period of a gain waveform in thelight modulator 103 may be the same as a period of a waveform. Referring toFIG. 1 , thelight modulator 103 may modulate the five beams of reflected light 1 that are respectively reflected by the five surfaces P1 to P5 of the subject 200 and provide the same to thecapturing device 105. Thelight modulator 103 may also sequentially modulate the five beams of each of the reflectedlight 2 through 5 and provide the same to thecapturing device 105. - Light, whose amplitude is modulated using the
light modulator 103, passes through thesecond lens 110. Here, a magnification of the light is adjusted and the light is re-focused. Then, the light arrives at thecapturing device 105. Accordingly, the modulated light is focused in an area of thecapturing device 105 through thesecond lens 110. Thecapturing device 105 generates an image by receiving the modulated light during an exposure time. For example, as indicated inFIG. 2A , thecapturing device 105 receives the five beams of the reflectedlight 1, which are respectively reflected on the five surfaces P1 to P5 of the subject 200 and then modulated, for a predetermined period of exposure time to generate an image 1 (CIS image 1). Next, as illustrated inFIG. 2B , thecapturing device 105 receives the five beams of the reflectedlight 2, which are respectively reflected on the five surfaces P1 to P5 of the subject 200 and then modulated, for a predetermined period of exposure time to generate an image 2 (CIS image 2). Finally, by repeating the above operations, as illustrated inFIG. 2C , thecapturing device 105 receives the five beams of the reflectedlight 4, which are respectively reflected on the five surfaces P1 to P5 of the subject 200 and then modulated, for a predetermined period of exposure time to generate an image 4 (CIS image 4). In this manner, as shown inFIG. 2D , four different images may be sequentially obtained. Theimages 1 through 4 may each be a sub-frame image for generating an image of a frame having depth information. For example, when assuming a period of a frame of depth information to be Td, an exposure time in thecapturing device 105 to obtain the fourimages 1 through 4 may be about Td/4. - Referring to
FIG. 2A , in a first sub-frame, the transmittedlight 1 projected from thelight source 101 to the subject 200 is reflected by the five surfaces P1 to P5 of the subject 200 to generate five beams of the reflectedlight 1. The five beams of the reflectedlight 1 are modulated using thelight modulator 103, and then arrive at thecapturing device 105. InFIG. 2 , for convenience of description, thecapturing device 105 includes only five pixels respectively corresponding to the five surfaces P1 to P5. Accordingly, the five beams of the reflectedlight 1 may be respectively incident on the corresponding five pixels. As illustrated inFIG. 2A , the five beams of the reflectedlight 1 respectively reflected on the surfaces P1 to P5 have different phase delays ΦP1 to ΦP5 according to the distance from the 3Dimage acquisition apparatus 100 to the surfaces P1 to P5. For example, thecapturing device 105 may generate animage 1 by capturing the reflectedlight 1 during an exposure time of about Td/4. In the same manner as described above, as illustrated inFIGS. 2B and 2C , different phase delays ΦP1 to ΦP5 are generated on the five surfaces P1 to P5 with different depths from a second sub-frame to a fourth sub-frame. - While four different light transmission signals are transmitted to a subject and four different images (sub-frames) are generated using a capturing device in
FIGS. 2A through 2D , identical transmission light may be used in each sub-frame, and thelight modulator 103 may modulate reflected light to different gain waveforms for each of the sub-frames.FIG. 3 illustrates an operation of generating four different images (sub-frames) after modulating the reflected light using four different light modulation signals in thelight modulator 103, according to another embodiment. Referring toFIG. 3 , the reflected light reflected by the subject 200 has the same waveform and phase in each sub-frame. As described above, there are different phase delays ΦP1 to ΦP5 in the reflected light of each sub-frame according to the surfaces P1 to P5 of the subject 200. As illustrated inFIGS. 3A through 3C , in the first sub-frame, thelight modulator 103 modulates the reflected light using a light modulation signal 1 (0 degrees). In the second sub-frame, thelight modulator 103 modulates the reflected light using a light modulation signal 2 (180 degrees) having a different phase from that of thelight modulation signal 1. Although not illustrated inFIGS. 3A through 3D , in the third sub-frame, thelight modulator 103 modulates the reflected light to a light modulation signal 3 (90 degrees) having another different phase. In the fourth sub-frame, thelight modulator 103 modulates the reflected light to another light modulation signal 4 (270 degrees). Here, the light modulation signals 1 through 4 are signals having different phases of 0, 180, 90, and 270 degrees. Also, the light modulation signals 1 through 4 may have the same periods and waveforms. Then, as illustrated inFIG. 3D , fourdifferent images 1 through 4 may be obtained. InFIGS. 2 and3 , while the modulated images are labeled asCIS images 1 through 4, as described above, thecapturing device 105 is not limited to a CIS. - The four images, obtained by projecting the light transmission signals having different phases (0, 180, 90, and 270 degrees) as illustrated in
FIG. 2 or by using the light modulation signals 1 through 4 having different phases (0, 180, 90, and 270 degrees) as illustrated inFIG. 3 , are transmitted to thedepth image processor 107. Thedepth image processor 107 generates a depth image or depth information according to an algorithm, based on the four images. Also, according to another embodiment, in order to effectively remove motion blur that has occurred due to motion of a subject, thedepth image processor 107 performs an algorithm for determining whether motion blur has occurred in the four images (the first image corresponding to 0 degrees, the second image corresponding to 180 degrees, the third image corresponding to the third image, and the fourth image corresponding to 270 degrees). When motion blur is generated, an algorithm is performed in thedepth image processor 107 for restoring an image or a pixel in which the motion blur is generated. -
FIG. 4 illustrates an operation of generating a depth image by using a moving window method according to another embodiment. - As illustrated in
FIGS. 2 and3 , four sequentially capturedintensity images 1 through 4 are used by thedepth image processor 107 in generating a depth image. The four captured imagescapturing device 105 illustrated inFIG. 1 (not limited to a CIS) generates a depth image based on the fourintensity images 1 through 4. -
- In other words, a subject is captured in an order of 0, 180, 90, and 270 degrees to obtain consecutive images. As illustrated in
FIG. 4 , the four images includingEquation 2 below:
where p is an arbitrary number. -
-
-
Intermediate images FIG. 4 are formed based on a difference between images of the first group U and images of the second group V. -
- As expressed in
Equation 3 above, by generating a depth image using the combination of two groups from among four images, one depth image may be obtained in a time period for obtaining two images (i.e., in a moving window method in which two images are captured and two previously captured images are removed from among a plurality of sequentially obtained images). This is advantageous in that a depth image may be obtained at a higher speed than an example where one depth image is obtained after four images are captured in the related art. - In the above-described embodiment, image extraction by using pixels arranged in a 2D array has been described. However, according to another embodiment, a usage range of the embodiments is not limited by whether pixels are arranged in a one-dimensional array or whether a single pixel is used. For example, when measuring a distance of a single pixel, a photodiode or an integrator may be used instead of a CIS.
- Hereinafter, a method of removing motion blur in a 3D image acquisition apparatus that is capable of generating a TOF-based depth image will be described. The essential reason that motion blur is caused is an abrupt change in luminance of a subject or surface in an IR image capturing section of four IR images. Thus, a method of detecting (determining) a pixel (or image) in which motion blur has occurred will be described. Then, a method of detecting (determining) a portion of the section where the subject does not change to generate a depth image based on the detected portion and remove divergence in an operation due to the motion blur will be described.
-
FIG. 5 illustrates a flow of an optical process according to another embodiment.FIGS. 6A and 6B are graphs for explaining output waveforms of a light source and a light modulator according to another embodiment. - Referring to
FIGS. 5 ,6A, and 6B , an optical processing operation of infrared light in the 3Dimage acquisition apparatus 100 is illustrated. Infrared light that has transmitted through an IR filter (e.g., a central wavelength of 850 nm) arrives at thelight modulator 103 and is modulated to arrive at thecapturing device 105 at a rear end of the 3Dimage acquisition apparatus 100. This series of optical processes are developed as described below. -
FIG. 6A is a waveform diagram of a light output waveform, andFIG. 6B is a driving waveform diagram of a light modulator. The waveform may be various periodic waves, such as a sine wave, a square wave, or a ramp wave. For convenience, a formula for an embodiment where a light output wave (light transmission signal) is a square wave and a driving waveform of a light modulator is a sine wave will be described. Further, variables and symbols are defined as follows:
θ (s) Phase Shift of the s-th Emitting Light
P ave Emitting Light DC offset
Pr (s) s-th Receiving Light Optical Power
P a Receiving Ambient Light
r Light Attenuation of the Object Surface
G Shutter Gain
G ave Shutter Gain DC offset
ω Operating Frequency
φTOF Phsse Delay due to TOF (Time of Flight) -
- In other words, the infrared light output is in the form of combination of a square waveform (AC component) with DC offset of output light, which is a DC component. According to the current embodiment, s is 1 to 4 and the infrared light has different phases such as 0 degrees, 180 degrees, 90 degrees, and 270 degrees.
-
- The reflected light returns after being reflected on a surface of a subject. Thus, the reflected light is in the form of a square of reflectivity r in which surface reflectivity of the subject, a lens size, etc., are in a comprehensive consideration, and a phase difference, due to a TOF method is present, and external light is also present, similar to the infrared light.
-
- The light modulation signal is in the form of a combination of a sine wave (AC component) and a DC component.
-
-
-
-
-
-
- In Equations 9 through 15, a still image is assumed. In other words, Equations 9 through 15 are formulas that are established under the assumption of an unchanged subject and surface reflectivity. In particular, Equations 9 through 12 are used in detecting and restoring motion blur.
-
FIGS. 7A and 7B are schematic views for explaining a type of motion blur generated in sub-frames captured by using a capturing device according to another embodiment. A method of detecting a motion blur pixel (image) will be described with reference toFIGS. 7A and 7B . -
FIG. 7A illustrates an example where motion blur is generated in a pixel while a subject changes from an object A to an object B. When the 3Dimage acquisition apparatus 100 sequentially captures four imagesFIG. 7B . A stable depth image, which is the same as a still image, may be generated without motion blur when four images of the same object, either the object A or the object B is captured (case 0 and case 5), are captured. However, in the four imagescases 1 through 4). Equations 9 through 11 described above are established incases -
-
- When no motion blur is generated, results of Equations 16 and 17 should be identical. Thus, a discriminant for a motion blur pixel (image) in Equation 18 below (an example where no motion blur occurs) and Equation 19 below (an example where motion blur occurs) may be obtained.
where ε is a constant which may be appropriately determined according to a noise level of a three-dimensional (3D) image acquisition apparatus or scene. -
FIGS. 8A and8B are schematic views for explaining an operation of restoring motion blur by using an image processor according to another embodiment. A method of restoring a motion blur pixel will be described with reference toFIGS. 8A and8B . -
-
-
-
- A constant k refers to a demodulation contrast which indicates modulation characteristics of a TOF system. The constant k is defined by Equation 24 below, and is expressed as a system constant, by a ratio between a variation value of an infrared ray image and a DC offset value. This variation value is a known value measured by calibration before capturing.
- As can be seen from Equations 9 through 12, a size of an infrared ray image due to a phase delay is expressed by sine and cosine functions. Further, an AC value and DC offset values are read by floating the functions to experimentally calculate A and B and a ratio k between A and B.
-
-
- A phase delay due to TOF according to
Equation 15 may be calculated using the restored pixel (image) usingEquations 25 through 28. Further, Equations 26 and 28 have complicated formulas and a large amount of computations. Thus, simplified approximation formulas, such as Equations 29 and 30 below, may also be applied: - The standards for determining whether motion blur has occurred in
case 1 orcase 2 illustrated inFIG. 8A ) or incase 3 orcase 4 illustrated inFIG. 8A ) will be described. In other words, it is determined in which of the first group (Equations 25 through 30. -
-
- As illustrated in
FIG. 8B , incase 1 or case 2 (where motion blur occurs incases FIG. 8B , in all of sections where four images are captured, a depth image may be stably generated as if only a single object, either anobject 1 or anobject 2, is being captured. -
FIG. 9 is a flowchart illustrating a method of generating a depth image in a 3D image acquisition apparatus, according to another embodiment. - Referring to
FIG. 9 , inoperation 900, a plurality of light transmission signals generated in a light source is sequentially transmitted to a subject. The light transmission signal may have different phases, such as 0, 180, 90, and 270 degrees or may have the same phase. - In
operation 902, light that is reflected by the subject is modulated using a light modulation signal. When the light transmission signal has different phases, light modulation signals may have the same phase. If the light transmission signal is projected at the same phase, the reflected light is modulated using light modulation signals corresponding to the different phases, i.e., 0, 180, 90, and 270 degrees. - In
operation 904, a phase delay is calculated using a combination of two groups of images from among a plurality of images obtained by capturing the modulated reflected light. Inoperation 902, a phase delay according to a depth is calculated using a combination of two images corresponding to 0 degrees and 180 degrees and two images corresponding to 90 degrees and 270 degrees from among the modulated light. - In
operation 906, a depth image is generated based on the phase delay. -
FIG. 10 is a flowchart illustrating a method of generating a depth image in a 3D image acquisition apparatus, according to another embodiment. - Referring to
FIG. 10 , inoperation 1000, a plurality of light transmission signals generated in a light source is sequentially projected to a subject. Inoperation 1002, light reflected by the subject is modulated using a light modulation signal. A light transmission signal from the light source or a light modulation signal from a light modulator may have different phases, such as 0, 180, 90, and 270 degrees or may have the same phases. - In operation S1004, first through fourth images corresponding to phases of the modulated reflected light, i.e., 0, 180, 90, and 270 degrees, are obtained. Further, a first image corresponding to 0 degrees, a second image corresponding to 180 degrees, a third image corresponding to 90 degrees, and a fourth image corresponding to 270 degrees are sequentially obtained. In this case, four images are maintained such that two images (e.g., images corresponding to 90 degrees and 270 degrees) are captured, and two previously captured images (e.g., images corresponding to 0 degrees and 180 degrees) are removed.
- In
operation 1006, whether a difference, between a sum of images in the first group including the first and second images and a sum of images in the second group including the third and fourth images, is equal to or greater than a first critical value is determined. The first critical value may be optionally determined according to a noise level of system or scenes. In the case of a still image or when motion blur is generated, the sum of images in the first group and the sum of images in the second group are not identical. Thus, when the difference between the first group and the second group is equal to or greater than the first critical value, it indicates that motion blur is present. If there is no motion blur (if the sum of images in the first group and the sum of images in the second group are identical), the method proceeds tooperations - In
operations operation 1006, if there is motion blur in the first through fourth images, it is determined in which of the first group and the second group the motion blur is present. The presence of motion blur indicates a large variation in luminance of an image. Therefore, it is determined that the motion blur is generated in a group with a large variation in luminance. Variation in luminance is determined in a manner above with reference to Equations 31 and 32. - In
operation 1012, images of the group with motion blur are restored using the images of the other group. When a depth image is generated by including the images of the group with motion blur, divergence is generated in resultant images. Thus, a totally different result is obtained. Accordingly, a phase delay and a depth image are generated after restoring the images with motion blur. To restore the motion blur, the images of the group with no motion blur and the above-described system variant k are used. The demodulation constant, which is a constant of Equation 24 above, or the system variable k as described above may be measured and calculated in an image test. A size of an image due to a phase delay as in Equations 9 through 12 is expressed by sine and cosine functions, and the image is expressed by a graph to read AC and DC offset values thereof, thereby experimentally calculating A and B and a ratio k therebetween. - In
operations - A result of examining the method of removing motion blur according to the current embodiment by applying the same to an actual image will be described. A demodulation constant k of a system was measured, and a depth image was generated using continuous infrared ray images. Examples where motion blur is not removed and motion blur is removed were respectively calculated, thereby comparing the accuracy of the depth image and whether the depth image is converged.
-
FIGS. 11A and11B are images for explaining motion blur generated when capturing a moving subject. Referring toFIG. 11A , images corresponding to 0, 180, 90, and 270 degrees are illustrated, showing a bar-shaped object in a rotated state. As can be seen from the overlapped images ofFIG. 11B , an image around an edge is blurred. -
FIG. 12 illustrates a result of applying Equation 19 to the images ofFIGS. 11A and11B in order to determine motion blur, according to another embodiment. -
FIG. 12 shows a result of applying the discriminant for determining motion blur in Equation 19, to the image illustrated inFIG. 11 . The image has mostly 0 values but relatively high values in edge portions thereof, thereby accurately searching motion blurred portions. -
FIGS. 13A and 13B illustrate a depth image in which motion blur is generated and a depth image for which the motion blur is restored, according to another embodiment. -
FIG. 13A illustrates an original image before motion blur is restored, andFIG. 13B is an image with restored motion blur according to an embodiment. Referring to edge portions where the motion blur has occurred, the depth image is significantly diverged before a motion blur restoration algorithm is applied. Thus, an unexpected value is given (seeFIG. 13A ). However, after the algorithm is applied, the depth image is stabilized and a physically meaningful result is given (seeFIG. 13B ). - According to the 3D image acquisition apparatus of the embodiments, a depth image is generated based on a combination of two groups of images, thereby increasing a rate of capturing the depth image.
- Also, according to the 3D image acquisition apparatus according to another embodiment, when a moving object is captured using a TOF method, a depth image is generated after detecting and restoring inevitable motion blur in portions such as an edge. Thus, divergence of the depth image due to the motion blur may be prevented. Also, as a light source, a light modulator, or a capturing device used in related art operations for generating a depth image are used, no additional costs for additional devices are necessary, and additional amount of computation may be reduced.
- The device described herein may comprise a processor, a memory for storing program data and executing it, a permanent storage such as a disk drive, a communications port for handling communication with external devices, and user interface devices, including a display, keys, etc. When software modules are involved, these software modules may be stored as program instructions or computer-readable codes executable on the processor on a computer-readable media such as read-only memory (ROM), random-access memory (RAM), CD-ROMs, magnetic tapes, floppy disks, and optical data storage devices. The computer-readable recording medium can also be distributed over network-coupled computer systems so that the computer-readable code is stored and executed in a distributed fashion. This media can be read by the computer, stored in the memory, and executed by the processor.
- All references, including publications, patent applications, and patents, cited herein are hereby incorporated by reference to the same extent as if each reference were individually and specifically indicated to be incorporated by reference and were set forth in its entirety herein.
- For the purposes of promoting an understanding of the principles of the exemplary embodiments, reference has been made to embodiments illustrated in the drawings, and specific language has been used to describe these embodiments. However, no limitation of the scope is intended by this specific language, and the exemplary embodiments should be construed to encompass all embodiments that would normally occur to one of ordinary skill in the art.
- Exemplary embodiments may be described in terms of functional block components and various processing steps. Such functional blocks may be realized by any number of hardware and/or software components configured to perform the specified functions. For example, the exemplary embodiments may employ various integrated circuit components, e.g., memory elements, processing elements, logic elements, look-up tables, etc., which may carry out a variety of functions under the control of one or more microprocessors or other control devices. Similarly, where the elements of the exemplary embodiments are implemented using software programming or software elements, the exemplary embodiments may be implemented with any programming or scripting language such as C, C++, Java, assembler, etc., with the various algorithms being implemented with any combination of data structures, objects, processes, routines or other programming elements. Functional aspects may be implemented in algorithms that execute on one or more processors. Furthermore, the exemplary embodiments could employ any number of related art techniques for electronics configuration, signal processing and/or control, data processing, etc. The words "mechanism" and "element" are used broadly, and are not limited to mechanical or physical embodiments, but can include software routines in conjunction with processors, etc.
- The particular implementations shown and described herein are illustrative examples of the exemplary embodiments and are not intended to otherwise limit the scope in any way. For the sake of brevity, conventional electronics, control systems, software development and other functional aspects of the systems (and components of the individual operating components of the systems) may not be described in detail. Furthermore, the connecting lines, or connectors shown in the various figures presented are intended to represent exemplary functional relationships and/or physical or logical couplings between the various elements. It should be noted that many alternative or additional functional relationships, physical connections or logical connections may be present in a practical device. Moreover, no item or component is essential to the practice of the invention unless the element is specifically described as "essential" or "critical".
- The use of the terms "a", "an", "the" and similar references in the context of describing the exemplary embodiments (especially in the context of the following claims) are to be construed to cover both the singular and the plural. Furthermore, recitation of ranges of values herein are merely intended to serve as a shorthand method of referring individually to each separate value falling within the range, unless otherwise indicated herein, and each separate value is incorporated into the specification as if it were individually recited herein. Finally, the steps of all methods described herein can be performed in any suitable order unless otherwise indicated herein or otherwise clearly contradicted by context. The use of any and all examples, or exemplary language (e.g., "such as") provided herein, is intended merely to better illuminate the exemplary embodiments, and does not pose a limitation on the scope unless otherwise claimed. Numerous modifications and adaptations will be readily apparent to those skilled in this art without departing from the invention as claimed.
- It should be understood that the exemplary embodiments described therein should be considered in a descriptive sense only and not for purposes of limitation. Descriptions of features or aspects within each embodiment should typically be considered as available for other similar features or aspects in other embodiments.
Claims (15)
- A method of generating a depth image, the method comprising:sequentially projecting a light transmission signal, which is generated from a light source, to a subject;modulating reflected light, which is reflected by the subject, using a light modulation signal;calculating a phase delay using a combination of a first group of images and a second group of images, the first and second groups together defining a plurality of images obtained by capturing the modulated reflected light; andgenerating a depth image based on the phase delay.
- The method of claim 1, wherein the light transmission signal or the light modulation signal has different phases, and is sequentially transmitted in an order of 0, 180, 90, and 270 degrees, and
wherein the phase delay is calculated using the first group which comprises a first image corresponding to 0 degrees and a second image corresponding to 180 degrees, and the second group which comprises a third image corresponding to 90 degrees and a fourth image corresponding to 270 degrees. - The method of claim 1 or 2, further comprising:determining whether there is a motion blur image generated by motion of the subject, from among the plurality of images,wherein the modulated reflected light is captured in an order of phases of 0, 180, 90, and 270 degrees, and the plurality of images comprise a first image corresponding to 0 degrees, a second image corresponding to 180 degrees, a third image corresponding to 90 degrees, and a fourth image corresponding to 270 degrees, andwherein the determining whether there is the motion blur image generated by motion of the subject comprises determining that the motion blur image is generated when a difference between a first sum of images in the first group which comprises the first image and the second image and a second sum of images in the second group which comprises the third image and the fourth image is equal to or greater than a first critical value.
- The method of claim 3, further comprising:determining a group where the motion blur image is generated, from among the first group and the second group if the motion blur image is generated as a result of the determining,wherein the determining the group where the motion blur image is generated comprises determining that the motion blur image is generated in the group where a range of variation in luminance of images is relatively large from among the first group and the second group.
- The method of claim 4, further comprising:restoring the motion blur image,wherein the restoring the motion blur image comprises restoring images of the determined group including the motion blur image, using images of the other group,wherein the restoring the motion blur image optionally further comprises using a demodulation constant which is a ratio between an AC variation and a DC offset value of an image due to the phase delay.
- The method of any one of claims 2 to 5, wherein a first depth image is generated by calculating the phase delay with respect to the first group and the second group, and a second depth image is generated by calculating the phase delay with respect to the second group and a next first group.
- A three-dimensional (3D) image acquisition apparatus comprising:a light source configured to sequentially project a light transmission signal to a subject;a light modulator configured to modulate reflected light which is reflected by the subject, using a light modulation signal having a gain waveform;a capturing device configured to capture the modulated reflected light, which is modulated by the light modulator, to generate a plurality of second images; anda depth image processor configured to calculate a phase delay using a combination of a first group of images and second group of images, the first and second groups together defining a plurality of images generated using the capturing device, and generating a depth image based on the phase delay.
- The 3D image acquisition apparatus of claim 7, further comprising:a light source driver configured to control a waveform of the light transmission signal by driving the light source;a light modulator driver configured to control the gain waveform by driving the light modulator; anda controller configured to control operations of the light source driver, the light modulator driver, and the capturing device.
- The 3D image acquisition apparatus of claim 8, wherein the controller is configured to control the light source driver or the light modulator driver such that the light transmission signal or the light modulation signal has different phases, and is sequentially projected in an order of 0, 180, 90, and 270 degrees,
wherein the depth image processor is configured to calculate the phase delay using the first group which comprises a first image corresponding to 0 degrees and a second image corresponding to 180 degrees, and the second group which comprises a third image corresponding to 90 degrees and a fourth image corresponding to 270 degrees. - The 3D image acquisition apparatus of claim 9, wherein the depth image processor is configured to determine whether there is a motion blur image generated by motion of the subject, from among the plurality of images, and
wherein the depth image processor is configured to determine that there is a motion blur image when a difference between a first sum of images in the first group which comprises the first image and the second image and a second sum of images in the second group which comprises the third image and the fourth image is equal to or greater than a first critical value. - The 3D image acquisition apparatus of claim 10, wherein the depth image processor is configured to restore images of the determined group including the motion blur image, using images of the other group and a demodulation constant which is a ratio between an AC variation and a DC offset value of an image due to the phase delay.
- The 3D image acquisition apparatus of claim 10 or 11, wherein the depth image processor is configured to generate a first depth image by calculating distances of the first group to the subject and the second group to the subject, and a second depth image by calculating the phase delay with respect to the second group and a next first group.
- The 3D image acquisition apparatus of claim 10, 11 or 12, further comprising:a first lens configured to focus the reflected light on a light incident surface of the light modulator within an area of the light modulator;a filter configured to transmit light having a predetermined wavelength between the first lens and the light modulator, and configured to remove light of wavelengths other than the predetermined wavelength; anda second lens configured to focus the modulated light between the light modulator and the capturing device within an area of the capturing device.
- The 3D image acquisition apparatus of any one of claims 10 to 13, wherein the capturing device comprises at least one of a charge-coupled device (CCD) image sensor which comprises a one-dimensional array or a two-dimensional array, a complementary metal oxide semiconductor (CMOS) image sensor, a photodiode array, and one photodiode, and
wherein the capturing device is configured to measure a distance to a single point. - The 3D image acquisition apparatus of any one of claims 10 to 14, wherein the light source is one of a solid-state light-emitting device which comprises a laser diode (LD) and a light-emitting diode (LED) in an infrared ray band.
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
KR1020130038287A KR102040152B1 (en) | 2013-04-08 | 2013-04-08 | An 3D image apparatus and method for generating a depth image in the 3D image apparatus |
Publications (2)
Publication Number | Publication Date |
---|---|
EP2790034A1 true EP2790034A1 (en) | 2014-10-15 |
EP2790034B1 EP2790034B1 (en) | 2018-06-13 |
Family
ID=50442385
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
EP14163725.6A Active EP2790034B1 (en) | 2013-04-08 | 2014-04-07 | 3D image acquisition apparatus and method of generating depth image in the 3D image acquisition apparatus |
Country Status (4)
Country | Link |
---|---|
US (1) | US9749613B2 (en) |
EP (1) | EP2790034B1 (en) |
KR (1) | KR102040152B1 (en) |
CN (1) | CN104104941B (en) |
Cited By (7)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
EP3163316A1 (en) * | 2015-10-29 | 2017-05-03 | Samsung Electronics Co., Ltd. | Apparatus and method for obtaining a depth image |
EP3185053A1 (en) * | 2015-12-21 | 2017-06-28 | Sick Ag | Optical modulator |
EP3907543A4 (en) * | 2019-01-04 | 2022-10-12 | LG Innotek Co., Ltd. | Optical system, and camera module comprising same |
US11650044B2 (en) | 2015-04-20 | 2023-05-16 | Samsung Electronics Co., Ltd. | CMOS image sensor for 2D imaging and depth measurement with ambient light rejection |
US11725933B2 (en) | 2015-04-20 | 2023-08-15 | Samsung Electronics Co., Ltd. | CMOS image sensor for RGB imaging and depth measurement with laser sheet scan |
US11736832B2 (en) | 2015-04-20 | 2023-08-22 | Samsung Electronics Co., Ltd. | Timestamp calibration of the 3D camera with epipolar line laser point scanning |
US11924545B2 (en) | 2015-04-20 | 2024-03-05 | Samsung Electronics Co., Ltd. | Concurrent RGBZ sensor and system |
Families Citing this family (37)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US8948497B2 (en) * | 2012-09-04 | 2015-02-03 | Digital Signal Corporation | System and method for increasing resolution of images obtained from a three-dimensional measurement system |
KR20150010230A (en) * | 2013-07-18 | 2015-01-28 | 삼성전자주식회사 | Method and apparatus for generating color image and depth image of an object using singular filter |
RU2014116610A (en) * | 2014-04-24 | 2015-10-27 | ЭлЭсАй Корпорейшн | DEPTH IMAGE GENERATION USING PSEUDOFRAMES, EACH OF WHICH CONTAINS A LOT OF PHASE IMAGES |
KR102194237B1 (en) | 2014-08-29 | 2020-12-22 | 삼성전자주식회사 | Method and apparatus for generating depth image |
US9773155B2 (en) | 2014-10-14 | 2017-09-26 | Microsoft Technology Licensing, Llc | Depth from time of flight camera |
CN104301707B (en) * | 2014-10-22 | 2016-07-06 | 中国科学院合肥物质科学研究院 | Depth image acquisition device and method based on DMD scanning and integrated phase shift range finding |
KR101693571B1 (en) | 2014-12-04 | 2017-01-09 | 김용기 | Container for growing mushroom |
US9519061B2 (en) * | 2014-12-26 | 2016-12-13 | Here Global B.V. | Geometric fingerprinting for localization of a device |
US10061029B2 (en) * | 2015-01-06 | 2018-08-28 | Samsung Electronics Co., Ltd. | Correction of depth images from T-O-F 3D camera with electronic-rolling-shutter for light modulation changes taking place during light integration |
KR102473740B1 (en) * | 2015-04-20 | 2022-12-05 | 삼성전자주식회사 | Concurrent rgbz sensor and system |
US10250833B2 (en) | 2015-04-20 | 2019-04-02 | Samsung Electronics Co., Ltd. | Timestamp calibration of the 3D camera with epipolar line laser point scanning |
JP2017050830A (en) * | 2015-09-04 | 2017-03-09 | パナソニックIpマネジメント株式会社 | Luminaire, illumination system and program |
KR102475891B1 (en) * | 2015-10-08 | 2022-12-12 | 삼성전자주식회사 | Edge light emitting laser source, apparatus for obtaining 3D image having the same |
KR102523972B1 (en) * | 2015-10-27 | 2023-04-20 | 삼성전자주식회사 | Method of measuring distance using 3D depth sensor including optical shutter |
KR102372087B1 (en) * | 2015-10-28 | 2022-03-08 | 삼성전자주식회사 | Method and apparatus for generating depth image |
KR102473735B1 (en) * | 2015-11-09 | 2022-12-05 | 삼성전자주식회사 | Operation method of imaging apparatus |
KR102610830B1 (en) * | 2015-12-24 | 2023-12-06 | 삼성전자주식회사 | Method and device for acquiring distance information |
CN106997582A (en) * | 2016-01-22 | 2017-08-01 | 北京三星通信技术研究有限公司 | The motion blur removing method and equipment of flight time three-dimension sensor |
US10229502B2 (en) | 2016-02-03 | 2019-03-12 | Microsoft Technology Licensing, Llc | Temporal time-of-flight |
KR102373926B1 (en) * | 2016-02-05 | 2022-03-14 | 삼성전자주식회사 | Vehicle and recognizing method of vehicle's position based on map |
US20170244482A1 (en) * | 2016-02-24 | 2017-08-24 | Qualcomm Incorporated | Light-based communication processing |
KR102454228B1 (en) * | 2016-04-29 | 2022-10-14 | 엘지전자 주식회사 | Multi-vision device |
EP3470880B1 (en) * | 2016-06-08 | 2020-08-19 | Panasonic Semiconductor Solutions Co., Ltd. | Distance-measuring system and distance-measuring method |
US10337691B2 (en) * | 2016-06-10 | 2019-07-02 | Magic Leap, Inc. | Integrating point source for texture projecting bulb |
KR20180021509A (en) * | 2016-08-22 | 2018-03-05 | 삼성전자주식회사 | Method and device for acquiring distance information |
DE102016124594A1 (en) * | 2016-12-16 | 2018-06-21 | Jena-Optronik Gmbh | Method for detecting a 3D scene using a LIDAR system and LIDAR system for this purpose |
KR102457891B1 (en) * | 2017-10-30 | 2022-10-25 | 삼성전자주식회사 | Method and apparatus for image processing |
US11598849B2 (en) * | 2017-12-03 | 2023-03-07 | Munro Design & Technologies, Llc | Signal generating systems for three-dimensional imaging systems and methods thereof |
EP3775979B1 (en) * | 2018-04-01 | 2024-01-17 | Opsys Tech Ltd. | Noise adaptive solid-state lidar system |
KR102590900B1 (en) * | 2018-08-27 | 2023-10-19 | 엘지이노텍 주식회사 | Image processing apparatus and image processing method |
KR102553555B1 (en) * | 2018-09-21 | 2023-07-10 | 엘지이노텍 주식회사 | Camera module |
KR102562360B1 (en) | 2018-10-05 | 2023-08-02 | 엘지이노텍 주식회사 | Method and camera module for acquiring depth information |
DE102018222518A1 (en) * | 2018-12-20 | 2020-06-25 | Zf Friedrichshafen Ag | Camera system with a high update rate |
CN109521434B (en) * | 2018-12-27 | 2020-09-08 | 合肥泰禾光电科技股份有限公司 | Laser measurement method and control processor |
KR102274033B1 (en) * | 2020-03-04 | 2021-07-08 | (주)미래컴퍼니 | Depth image acquisition apparatus reducing motion blur using multi frequency |
US11443447B2 (en) | 2020-04-17 | 2022-09-13 | Samsung Electronics Co., Ltd. | Three-dimensional camera system |
CN111798506B (en) * | 2020-06-30 | 2024-07-23 | 上海数迹智能科技有限公司 | Image processing method, terminal and computer readable storage medium |
Citations (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20110164132A1 (en) * | 2010-01-06 | 2011-07-07 | Mesa Imaging Ag | Demodulation Sensor with Separate Pixel and Storage Arrays |
US20120162197A1 (en) * | 2010-12-23 | 2012-06-28 | Samsung Electronics Co., Ltd. | 3-dimensional image acquisition apparatus and method of extracting depth information in the 3d image acquisition apparatus |
Family Cites Families (38)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US5081530A (en) | 1987-06-26 | 1992-01-14 | Antonio Medina | Three dimensional camera and range finder |
US4935616A (en) | 1989-08-14 | 1990-06-19 | The United States Of America As Represented By The Department Of Energy | Range imaging laser radar |
DE69635858T2 (en) | 1995-06-22 | 2006-11-30 | 3Dv Systems Ltd. | TELECENTRIC 3D CAMERA AND RELATED METHOD |
IL114278A (en) | 1995-06-22 | 2010-06-16 | Microsoft Internat Holdings B | Camera and method |
US6088086A (en) | 1995-09-11 | 2000-07-11 | Sandia Corporation | Range determination for scannerless imaging |
AU5876798A (en) | 1998-02-08 | 1999-08-23 | 3Dv Systems Ltd. | Large aperture optical image shutter |
JP3840341B2 (en) | 1998-10-15 | 2006-11-01 | 浜松ホトニクス株式会社 | Three-dimensional information detection method and apparatus |
US6118946A (en) | 1999-06-29 | 2000-09-12 | Eastman Kodak Company | Method and apparatus for scannerless range image capture using photographic film |
US6856355B1 (en) | 1999-11-30 | 2005-02-15 | Eastman Kodak Company | Method and apparatus for a color scannerless range image system |
US6794628B2 (en) | 2000-01-03 | 2004-09-21 | 3Dv Systems, Ltd. | Solid state optical shutter |
WO2005036372A2 (en) | 2003-10-09 | 2005-04-21 | Honda Motor Co., Ltd. | Systems and methods for determining depth using shuttered light pulses |
JP4161910B2 (en) | 2004-01-28 | 2008-10-08 | 株式会社デンソー | Distance image data generation device, generation method, and program |
US8629976B2 (en) * | 2007-10-02 | 2014-01-14 | Microsoft Corporation | Methods and systems for hierarchical de-aliasing time-of-flight (TOF) systems |
US8203699B2 (en) * | 2008-06-30 | 2012-06-19 | Microsoft Corporation | System architecture design for time-of-flight system having reduced differential pixel size, and time-of-flight systems so designed |
US8619354B2 (en) | 2008-12-24 | 2013-12-31 | Samsung Electronics Co., Ltd. | High speed optical shutter, method of operating the same and apparatus including the same |
KR101520029B1 (en) | 2008-12-31 | 2015-05-15 | 삼성전자주식회사 | Optical Modulator with Pixelization Patterns |
KR101547327B1 (en) | 2009-01-15 | 2015-09-07 | 삼성전자주식회사 | Optical image modulator and optical apparatus comprising the same and methods of manufacturing and operating optical image modulator |
KR101603778B1 (en) | 2009-01-19 | 2016-03-25 | 삼성전자주식회사 | Optical image shutter |
KR101613133B1 (en) * | 2009-05-14 | 2016-04-18 | 삼성전자주식회사 | Method and apparatus for processing 3-dimensional image |
US8988661B2 (en) | 2009-05-29 | 2015-03-24 | Microsoft Technology Licensing, Llc | Method and system to maximize space-time resolution in a time-of-flight (TOF) system |
KR101623960B1 (en) | 2009-06-04 | 2016-05-25 | 삼성전자주식회사 | Optoelectronic shutter, method of operating the same and optical apparatus employing the optoelectronic shutter |
KR101590767B1 (en) | 2009-06-09 | 2016-02-03 | 삼성전자주식회사 | Image processing apparatus and method |
KR101638974B1 (en) | 2009-06-17 | 2016-07-13 | 삼성전자주식회사 | Optical modulator, methods of manufacturing and operating the same and optical apparatus comprising optical modulator |
KR101706354B1 (en) | 2009-06-24 | 2017-02-13 | 삼성전자주식회사 | High speed optical modulator and method of modulating light using the same |
KR101675111B1 (en) | 2010-01-08 | 2016-11-11 | 삼성전자주식회사 | Optical image shutter and method of fabricating the same |
KR101675112B1 (en) * | 2010-01-21 | 2016-11-22 | 삼성전자주식회사 | Method of extractig depth information and optical apparatus employing the method |
KR101638973B1 (en) | 2010-01-22 | 2016-07-12 | 삼성전자주식회사 | Optical modulator and method of fabricating the same |
JP2012049773A (en) | 2010-08-26 | 2012-03-08 | Sony Corp | Imaging apparatus and method, and program |
KR101753312B1 (en) | 2010-09-17 | 2017-07-03 | 삼성전자주식회사 | Apparatus and method for generating depth image |
US9194953B2 (en) * | 2010-10-21 | 2015-11-24 | Sony Corporation | 3D time-of-light camera and method |
KR101598547B1 (en) | 2010-12-03 | 2016-03-02 | 삼성전자주식회사 | Optical image modulator and method of manufacturing the same |
KR101660408B1 (en) | 2010-12-20 | 2016-09-27 | 삼성전자주식회사 | Optical image modulator having correction function for incident angle of incident light and optical apparatus comprising the same |
KR20120075182A (en) | 2010-12-28 | 2012-07-06 | 삼성전자주식회사 | Optical modulator using multiple fabry-perot resonant modes and 3d image acquisition apparatus |
KR101669219B1 (en) | 2010-12-30 | 2016-10-26 | 삼성전자주식회사 | Light modulator and optical apparatus employing the same |
KR101788032B1 (en) * | 2011-03-24 | 2017-10-19 | 삼성전자주식회사 | Depth sensor, depth information error compensation method thereof, and signal processing system having the depth sensor |
KR20120111092A (en) | 2011-03-31 | 2012-10-10 | 삼성전자주식회사 | Image pick-up apparatus |
KR101799521B1 (en) | 2011-05-24 | 2017-11-20 | 삼성전자 주식회사 | Light modulator with photonic crystal and 3D image acquisition apparatus employing the same |
US20140049767A1 (en) * | 2012-08-15 | 2014-02-20 | Microsoft Corporation | Methods and systems for geometric phase unwrapping in time of flight systems |
-
2013
- 2013-04-08 KR KR1020130038287A patent/KR102040152B1/en active IP Right Grant
-
2014
- 2014-04-07 EP EP14163725.6A patent/EP2790034B1/en active Active
- 2014-04-08 US US14/247,907 patent/US9749613B2/en active Active
- 2014-04-08 CN CN201410138665.9A patent/CN104104941B/en active Active
Patent Citations (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20110164132A1 (en) * | 2010-01-06 | 2011-07-07 | Mesa Imaging Ag | Demodulation Sensor with Separate Pixel and Storage Arrays |
US20120162197A1 (en) * | 2010-12-23 | 2012-06-28 | Samsung Electronics Co., Ltd. | 3-dimensional image acquisition apparatus and method of extracting depth information in the 3d image acquisition apparatus |
Non-Patent Citations (1)
Title |
---|
RADKE R J ET AL: "Image change detection algorithms: a systematic survey", IEEE TRANSACTIONS ON IMAGE PROCESSING, IEEE SERVICE CENTER, PISCATAWAY, NJ, US, vol. 14, no. 3, 1 March 2005 (2005-03-01), pages 294 - 307, XP008104908, ISSN: 1057-7149, DOI: 10.1109/TIP.2004.838698 * |
Cited By (8)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US11650044B2 (en) | 2015-04-20 | 2023-05-16 | Samsung Electronics Co., Ltd. | CMOS image sensor for 2D imaging and depth measurement with ambient light rejection |
US11725933B2 (en) | 2015-04-20 | 2023-08-15 | Samsung Electronics Co., Ltd. | CMOS image sensor for RGB imaging and depth measurement with laser sheet scan |
US11736832B2 (en) | 2015-04-20 | 2023-08-22 | Samsung Electronics Co., Ltd. | Timestamp calibration of the 3D camera with epipolar line laser point scanning |
US11924545B2 (en) | 2015-04-20 | 2024-03-05 | Samsung Electronics Co., Ltd. | Concurrent RGBZ sensor and system |
EP3163316A1 (en) * | 2015-10-29 | 2017-05-03 | Samsung Electronics Co., Ltd. | Apparatus and method for obtaining a depth image |
US10760906B2 (en) | 2015-10-29 | 2020-09-01 | Samsung Electronics Co., Ltd. | Apparatus and method for obtaining three-dimensional depth image |
EP3185053A1 (en) * | 2015-12-21 | 2017-06-28 | Sick Ag | Optical modulator |
EP3907543A4 (en) * | 2019-01-04 | 2022-10-12 | LG Innotek Co., Ltd. | Optical system, and camera module comprising same |
Also Published As
Publication number | Publication date |
---|---|
EP2790034B1 (en) | 2018-06-13 |
US20140300701A1 (en) | 2014-10-09 |
KR102040152B1 (en) | 2019-12-05 |
CN104104941A (en) | 2014-10-15 |
KR20140121710A (en) | 2014-10-16 |
US9749613B2 (en) | 2017-08-29 |
CN104104941B (en) | 2018-12-21 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
EP2790034B1 (en) | 3D image acquisition apparatus and method of generating depth image in the 3D image acquisition apparatus | |
US8611610B2 (en) | Method and apparatus for calculating a distance between an optical apparatus and an object | |
US9310488B2 (en) | Apparatus and method for generating depth image | |
KR101722641B1 (en) | 3D image acquisition apparatus and method of extractig depth information in the 3D image acquisition apparatus | |
US9746547B2 (en) | Method and apparatus for generating depth image | |
US9123164B2 (en) | 3D image acquisition apparatus and method of extracting depth information in 3D image acquisition apparatus | |
US10356380B2 (en) | Apparatus and method for acquiring depth information | |
US9894347B2 (en) | 3D image acquisition apparatus and method of driving the same | |
KR101854188B1 (en) | 3D image acquisition apparatus and method of acqiring depth information in the 3D image acquisition apparatus | |
US8369575B2 (en) | 3D image processing method and apparatus for improving accuracy of depth measurement of an object in a region of interest | |
US9194953B2 (en) | 3D time-of-light camera and method | |
EP2477043A1 (en) | 3D time-of-flight camera and method | |
US10205933B2 (en) | Depth image acquisition apparatus and method of acquiring depth information | |
EP3287807A1 (en) | Method and device for acquiring distance information | |
US10264240B2 (en) | Method and apparatus for generating depth image | |
US20120162370A1 (en) | Apparatus and method for generating depth image | |
KR20170076477A (en) | Method and device for acquiring distance information | |
KR20150133086A (en) | Method for generating depth image and image generating apparatus using thereof |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PUAI | Public reference made under article 153(3) epc to a published international application that has entered the european phase |
Free format text: ORIGINAL CODE: 0009012 |
|
17P | Request for examination filed |
Effective date: 20140407 |
|
AK | Designated contracting states |
Kind code of ref document: A1 Designated state(s): AL AT BE BG CH CY CZ DE DK EE ES FI FR GB GR HR HU IE IS IT LI LT LU LV MC MK MT NL NO PL PT RO RS SE SI SK SM TR |
|
AX | Request for extension of the european patent |
Extension state: BA ME |
|
R17P | Request for examination filed (corrected) |
Effective date: 20150415 |
|
RBV | Designated contracting states (corrected) |
Designated state(s): AL AT BE BG CH CY CZ DE DK EE ES FI FR GB GR HR HU IE IS IT LI LT LU LV MC MK MT NL NO PL PT RO RS SE SI SK SM TR |
|
17Q | First examination report despatched |
Effective date: 20150824 |
|
STAA | Information on the status of an ep patent application or granted ep patent |
Free format text: STATUS: EXAMINATION IS IN PROGRESS |
|
GRAP | Despatch of communication of intention to grant a patent |
Free format text: ORIGINAL CODE: EPIDOSNIGR1 |
|
STAA | Information on the status of an ep patent application or granted ep patent |
Free format text: STATUS: GRANT OF PATENT IS INTENDED |
|
INTG | Intention to grant announced |
Effective date: 20171218 |
|
GRAS | Grant fee paid |
Free format text: ORIGINAL CODE: EPIDOSNIGR3 |
|
GRAA | (expected) grant |
Free format text: ORIGINAL CODE: 0009210 |
|
STAA | Information on the status of an ep patent application or granted ep patent |
Free format text: STATUS: THE PATENT HAS BEEN GRANTED |
|
AK | Designated contracting states |
Kind code of ref document: B1 Designated state(s): AL AT BE BG CH CY CZ DE DK EE ES FI FR GB GR HR HU IE IS IT LI LT LU LV MC MK MT NL NO PL PT RO RS SE SI SK SM TR |
|
REG | Reference to a national code |
Ref country code: GB Ref legal event code: FG4D |
|
REG | Reference to a national code |
Ref country code: CH Ref legal event code: EP Ref country code: AT Ref legal event code: REF Ref document number: 1009057 Country of ref document: AT Kind code of ref document: T Effective date: 20180615 |
|
REG | Reference to a national code |
Ref country code: DE Ref legal event code: R096 Ref document number: 602014026867 Country of ref document: DE |
|
REG | Reference to a national code |
Ref country code: IE Ref legal event code: FG4D |
|
REG | Reference to a national code |
Ref country code: NL Ref legal event code: MP Effective date: 20180613 |
|
REG | Reference to a national code |
Ref country code: LT Ref legal event code: MG4D |
|
PG25 | Lapsed in a contracting state [announced via postgrant information from national office to epo] |
Ref country code: CY Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT Effective date: 20180613 Ref country code: LT Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT Effective date: 20180613 Ref country code: NO Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT Effective date: 20180913 Ref country code: SE Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT Effective date: 20180613 Ref country code: ES Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT Effective date: 20180613 Ref country code: FI Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT Effective date: 20180613 Ref country code: BG Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT Effective date: 20180913 |
|
PG25 | Lapsed in a contracting state [announced via postgrant information from national office to epo] |
Ref country code: LV Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT Effective date: 20180613 Ref country code: RS Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT Effective date: 20180613 Ref country code: GR Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT Effective date: 20180914 Ref country code: HR Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT Effective date: 20180613 |
|
REG | Reference to a national code |
Ref country code: AT Ref legal event code: MK05 Ref document number: 1009057 Country of ref document: AT Kind code of ref document: T Effective date: 20180613 |
|
PG25 | Lapsed in a contracting state [announced via postgrant information from national office to epo] |
Ref country code: NL Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT Effective date: 20180613 |
|
PG25 | Lapsed in a contracting state [announced via postgrant information from national office to epo] |
Ref country code: AT Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT Effective date: 20180613 Ref country code: PL Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT Effective date: 20180613 Ref country code: EE Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT Effective date: 20180613 Ref country code: IS Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT Effective date: 20181013 Ref country code: SK Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT Effective date: 20180613 Ref country code: RO Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT Effective date: 20180613 Ref country code: CZ Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT Effective date: 20180613 |
|
REG | Reference to a national code |
Ref country code: CH Ref legal event code: PK Free format text: BERICHTIGUNGEN |
|
RIC2 | Information provided on ipc code assigned after grant |
Ipc: H04N 13/02 20060101ALI20140710BHEP Ipc: G01S 17/36 20060101AFI20140710BHEP |
|
PG25 | Lapsed in a contracting state [announced via postgrant information from national office to epo] |
Ref country code: IT Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT Effective date: 20180613 Ref country code: SM Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT Effective date: 20180613 |
|
REG | Reference to a national code |
Ref country code: DE Ref legal event code: R097 Ref document number: 602014026867 Country of ref document: DE |
|
PLBE | No opposition filed within time limit |
Free format text: ORIGINAL CODE: 0009261 |
|
STAA | Information on the status of an ep patent application or granted ep patent |
Free format text: STATUS: NO OPPOSITION FILED WITHIN TIME LIMIT |
|
26N | No opposition filed |
Effective date: 20190314 |
|
PG25 | Lapsed in a contracting state [announced via postgrant information from national office to epo] |
Ref country code: SI Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT Effective date: 20180613 Ref country code: DK Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT Effective date: 20180613 |
|
PG25 | Lapsed in a contracting state [announced via postgrant information from national office to epo] |
Ref country code: AL Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT Effective date: 20180613 |
|
REG | Reference to a national code |
Ref country code: CH Ref legal event code: PL |
|
REG | Reference to a national code |
Ref country code: BE Ref legal event code: MM Effective date: 20190430 |
|
PG25 | Lapsed in a contracting state [announced via postgrant information from national office to epo] |
Ref country code: LU Free format text: LAPSE BECAUSE OF NON-PAYMENT OF DUE FEES Effective date: 20190407 Ref country code: MC Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT Effective date: 20180613 |
|
PG25 | Lapsed in a contracting state [announced via postgrant information from national office to epo] |
Ref country code: LI Free format text: LAPSE BECAUSE OF NON-PAYMENT OF DUE FEES Effective date: 20190430 Ref country code: CH Free format text: LAPSE BECAUSE OF NON-PAYMENT OF DUE FEES Effective date: 20190430 |
|
PG25 | Lapsed in a contracting state [announced via postgrant information from national office to epo] |
Ref country code: BE Free format text: LAPSE BECAUSE OF NON-PAYMENT OF DUE FEES Effective date: 20190430 |
|
PG25 | Lapsed in a contracting state [announced via postgrant information from national office to epo] |
Ref country code: TR Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT Effective date: 20180613 |
|
PG25 | Lapsed in a contracting state [announced via postgrant information from national office to epo] |
Ref country code: IE Free format text: LAPSE BECAUSE OF NON-PAYMENT OF DUE FEES Effective date: 20190407 |
|
PG25 | Lapsed in a contracting state [announced via postgrant information from national office to epo] |
Ref country code: PT Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT Effective date: 20181015 |
|
PG25 | Lapsed in a contracting state [announced via postgrant information from national office to epo] |
Ref country code: MT Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT Effective date: 20180613 Ref country code: HU Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT; INVALID AB INITIO Effective date: 20140407 |
|
PG25 | Lapsed in a contracting state [announced via postgrant information from national office to epo] |
Ref country code: MK Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT Effective date: 20180613 |
|
P01 | Opt-out of the competence of the unified patent court (upc) registered |
Effective date: 20230530 |
|
PGFP | Annual fee paid to national office [announced via postgrant information from national office to epo] |
Ref country code: GB Payment date: 20240321 Year of fee payment: 11 |
|
PGFP | Annual fee paid to national office [announced via postgrant information from national office to epo] |
Ref country code: FR Payment date: 20240321 Year of fee payment: 11 |
|
PGFP | Annual fee paid to national office [announced via postgrant information from national office to epo] |
Ref country code: DE Payment date: 20240319 Year of fee payment: 11 |