WO2012032826A1 - 立体撮像装置および立体撮像方法 - Google Patents
立体撮像装置および立体撮像方法 Download PDFInfo
- Publication number
- WO2012032826A1 WO2012032826A1 PCT/JP2011/063880 JP2011063880W WO2012032826A1 WO 2012032826 A1 WO2012032826 A1 WO 2012032826A1 JP 2011063880 W JP2011063880 W JP 2011063880W WO 2012032826 A1 WO2012032826 A1 WO 2012032826A1
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- amount
- blur
- aperture
- unit
- image
- Prior art date
Links
Images
Classifications
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N13/00—Stereoscopic video systems; Multi-view video systems; Details thereof
- H04N13/20—Image signal generators
- H04N13/204—Image signal generators using stereoscopic image cameras
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N13/00—Stereoscopic video systems; Multi-view video systems; Details thereof
- H04N13/10—Processing, recording or transmission of stereoscopic or multi-view image signals
- H04N13/106—Processing image signals
- H04N13/128—Adjusting depth or disparity
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N13/00—Stereoscopic video systems; Multi-view video systems; Details thereof
- H04N13/20—Image signal generators
- H04N13/204—Image signal generators using stereoscopic image cameras
- H04N13/207—Image signal generators using stereoscopic image cameras using a single 2D image sensor
- H04N13/218—Image signal generators using stereoscopic image cameras using a single 2D image sensor using spatial multiplexing
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N13/00—Stereoscopic video systems; Multi-view video systems; Details thereof
- H04N13/20—Image signal generators
- H04N13/257—Colour aspects
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N23/00—Cameras or camera modules comprising electronic image sensors; Control thereof
- H04N23/60—Control of cameras or camera modules
- H04N23/67—Focus control based on electronic image sensor signals
- H04N23/672—Focus control based on electronic image sensor signals based on the phase difference signals
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N23/00—Cameras or camera modules comprising electronic image sensors; Control thereof
- H04N23/60—Control of cameras or camera modules
- H04N23/67—Focus control based on electronic image sensor signals
- H04N23/673—Focus control based on electronic image sensor signals based on contrast or high frequency components of image signals, e.g. hill climbing method
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N23/00—Cameras or camera modules comprising electronic image sensors; Control thereof
- H04N23/60—Control of cameras or camera modules
- H04N23/68—Control of cameras or camera modules for stable pick-up of the scene, e.g. compensating for camera body vibrations
- H04N23/681—Motion detection
- H04N23/6812—Motion detection based on additional sensors, e.g. acceleration sensors
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N23/00—Cameras or camera modules comprising electronic image sensors; Control thereof
- H04N23/60—Control of cameras or camera modules
- H04N23/68—Control of cameras or camera modules for stable pick-up of the scene, e.g. compensating for camera body vibrations
- H04N23/682—Vibration or motion blur correction
- H04N23/685—Vibration or motion blur correction performed by mechanical compensation
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N2209/00—Details of colour television systems
- H04N2209/04—Picture signal generators
- H04N2209/041—Picture signal generators using solid-state devices
- H04N2209/042—Picture signal generators using solid-state devices having a single pick-up sensor
- H04N2209/045—Picture signal generators using solid-state devices having a single pick-up sensor using mosaic colour filter
Definitions
- the present invention relates to a stereoscopic imaging apparatus and a stereoscopic imaging method, and more particularly, to a technique for acquiring different viewpoint images by imaging subject images that have passed through different regions in two directions of a photographic lens, respectively.
- Patent Document 1 a monocular stereoscopic imaging device having an optical system shown in FIG. 11 is known.
- a subject image that has passed through different regions in the left-right direction of the main lens 1 and the relay lens 2 is divided into pupils by a mirror 4 and imaged on imaging elements 7 and 8 via imaging lenses 5 and 6, respectively. I am doing so.
- FIG. 12 (A) to 12 (C) are diagrams illustrating separation states of images formed on the image sensor due to differences in the front pin, in-focus (best focus), and rear pin, respectively.
- the mirror 4 shown in FIG. 11 is omitted in order to compare the difference in separation due to focus.
- the focused image among the images divided by the pupil is formed (matched) at the same position on the image sensor, but FIGS. 12A and 12C.
- the images to be the front pin and the rear pin are formed (separated) at different positions on the image sensor.
- Patent Document 2 is an example of a compound eye stereoscopic imaging device. This imaging apparatus measures camera shake from a captured 3D image, determines that it is unsuitable for stereoscopic vision when the amount of camera shake correction exceeds a predetermined amount, and causes the 2D data creation unit to output the captured image data.
- Patent Document 3 is an example of a camera shake correction mechanism of a compound-eye stereoscopic imaging device.
- Patent Document 4 is an example of a camera shake correction mechanism of a monocular camera equipped with a stereo adapter.
- Patent Document 5 is an example of a camera provided with an optical camera shake correction mechanism.
- the amount of camera shake ⁇ x in the yaw direction and the amount of camera shake ⁇ y in the pitch direction are respectively determined from the acceleration a1x ⁇ a2x in the x direction and the acceleration a1y ⁇ a2y in the y direction, the displacement angle ⁇ x around the y axis, the angular velocity ⁇ x, It is obtained by calculating the displacement angle ⁇ y and the angular velocity ⁇ y.
- the image captured by the image sensor is blurred due to the aberration of the optical system compared to the original object, and the image quality is reduced. It is known to deteriorate.
- the technique for removing the blur of the optical system in this way by signal processing and obtaining an ideal image is called “restoration”, “deconvolution”, or “deconvolution” of the image.
- the restoration filter based on the point image intensity distribution (PSF) is information relating to image degradation at the time of imaging, such as shooting conditions (exposure time, exposure amount, distance to the subject, focal length, etc.) and imaging device characteristic information (lens It is generated in consideration of optical characteristics, imaging device identification information, etc. (Patent Document 6).
- the deterioration model due to blur can be expressed by a function.
- the blur phenomenon can be expressed by a normal distribution having a parameter (image height) from the center pixel as a parameter (Patent Document 7).
- Patent Document 8 shows an example of aperture control of an imaging apparatus.
- the camera is controlled so as to increase the amount of light to the solid-state imaging device using an extra open aperture when photographing with the electronic zoom.
- the shutter speed can be increased and camera shake can be prevented.
- camera shake tends to occur.
- the amount of blurring (blurring amount) of the image due to camera shake is calculated, and compared with the calculated blur amount and image performance data with ⁇ extra open, and when the amount of blurring due to camera shake exceeds the effect of image performance degradation Using an extra open aperture, the shutter speed is increased, and shooting is performed at a point that balances camera shake and image performance.
- a monocular stereoscopic imaging device that picks up parallax images by selectively acquiring image information that has passed through different pupil positions of the image imaging means, and generates a stereoscopic image
- the parallax at the in-focus position is 0, and the non-focusing is performed.
- a parallax corresponding to the defocus amount occurs. For this reason, there is a problem that the amount of parallax of the stereoscopic image varies because the defocus amount at the out-of-focus position is affected by a shake caused by a pan operation at the time of shooting.
- the present invention has been made in view of such circumstances, and an object thereof is to generate an appropriate stereoscopic image corresponding to a blurring state at the time of photographing.
- the present invention divides a luminous flux from a subject into two luminous fluxes by pupil division, images the two luminous fluxes on two corresponding photoelectric conversion element groups, and performs photoelectric conversion, thereby pairing images having parallax.
- the diaphragm unit that can adjust the light amount of the light beam formed on the photoelectric conversion element group of the image pickup unit, the light amount detection unit that detects the light amount of the light beam,
- a diaphragm control unit capable of executing a first diaphragm control for controlling an opening amount of the diaphragm unit so as to obtain a desired parallax; and a blur detection unit for detecting a blur amount of a subject image when an image is captured by the imaging unit;
- the aperture control unit executes the second aperture control for performing the second aperture control so that the larger the blur amount detected by the blur detection unit is, the smaller the aperture amount required for the first aperture control is.
- a shake correction unit that optically corrects the blur of the subject image according to the shake amount detected by the shake detection unit, and whether or not the shake amount detected by the shake detection unit is within a shake correction range of the shake correction unit.
- the diaphragm control unit executes the second diaphragm control when the blur amount determination unit determines that the blur amount is out of the range in which the blur correction is possible.
- the blur correction unit optically corrects the blur of the subject image when the blur amount determination unit determines that the blur amount is within a range in which the blur correction is possible, and the aperture control unit determines that the blur amount determination unit When it is determined that the range is within the range in which blur correction is possible, first aperture control is executed.
- a blur direction detection unit that detects a blur direction of the subject image
- a blur direction determination unit that determines whether or not the blur direction detected by the blur direction detection unit matches the direction of pupil division
- the aperture control unit includes: If the blur direction determination unit determines that the blur direction matches the direction of pupil division, second aperture control is executed.
- the diaphragm control unit executes the first diaphragm control when the blur direction determination unit determines that the blur direction does not match the pupil division direction.
- a stereoscopic imaging device divides a light beam from a subject into two light beams by pupil division, images the two light beams on two corresponding photoelectric conversion element groups, and photoelectrically converts the parallax.
- a step of outputting a pair of images having, a step of adjusting a light amount of a light beam formed on the photoelectric conversion element group by an opening amount of a diaphragm portion, a step of detecting a light amount of the light beam, and a desired amount according to the detected light amount
- a first aperture control for controlling an opening amount of the aperture unit so as to obtain a parallax, a step for detecting a blur amount of a subject image at the time of capturing an image, and a larger aperture unit as the detected blur amount increases.
- a step of executing a second aperture control for further reducing the amount of the aperture to be less than the amount of the aperture required for the first aperture control.
- a stereoscopic imaging device divides a light beam from a subject into two light beams by pupil division, images the two light beams on two corresponding photoelectric conversion element groups, and photoelectrically converts the parallax.
- a step of outputting a pair of images having, a step of adjusting a light amount of a light beam formed on the photoelectric conversion element group by an opening amount of a diaphragm portion, a step of detecting a light amount of the light beam, and a desired amount according to the detected light amount
- a first aperture control for controlling an opening amount of the aperture unit so as to obtain a parallax, a step for detecting a blur amount of a subject image at the time of capturing an image, and a larger aperture unit as the detected blur amount increases.
- a step of executing a second aperture control for further reducing the aperture amount of the first aperture control to be smaller than that required for the first aperture control.
- the depth of focus is increased and the amount of parallax is decreased, so that the stereoscopic effect is weakened when the blur is large, and the influence on the observer can be reduced.
- Block diagram of a monocular stereoscopic imaging device The figure which shows the structural example of the pupil division parallax image acquisition image pick-up element CCD.
- a diagram showing the main and sub-pixels one by one 3 is an enlarged view of the main part.
- FIG. 1 is a block diagram showing an embodiment of a monocular stereoscopic imaging device 1 according to the first embodiment.
- the monocular stereoscopic imaging device 1 records a captured image on a memory card 54, and the operation of the entire device is centrally controlled by a central processing unit (CPU) 40.
- CPU central processing unit
- the monocular stereoscopic imaging device 1 is provided with operation units 38 such as a shutter button, a mode dial, a playback button, a MENU / OK key, a cross key, and a BACK key.
- operation units 38 such as a shutter button, a mode dial, a playback button, a MENU / OK key, a cross key, and a BACK key.
- a signal from the operation unit 38 is input to the CPU 40, and the CPU 40 controls each circuit of the monocular stereoscopic imaging device 1 based on the input signal. For example, lens driving control, aperture driving control, photographing operation control, image processing control, Image data recording / reproduction control, display control of the liquid crystal monitor 30 for stereoscopic display, and the like are performed.
- the shutter button is an operation button for inputting an instruction to start shooting, and is configured by a two-stroke switch having an S1 switch that is turned on when half-pressed and an S2 switch that is turned on when fully pressed.
- the mode dial is selection means for selecting one of an auto shooting mode for shooting a still image, a manual shooting mode, a scene position such as a person, a landscape, a night view, and a moving image mode for shooting a moving image.
- the playback button is a button for switching to a playback mode in which a still image or a moving image of a stereoscopic image (3D image) or a planar image (2D image) that has been recorded is displayed on the liquid crystal monitor 30.
- the MENU / OK key is an operation key that has both a function as a menu button for instructing to display a menu on the screen of the liquid crystal monitor 30 and a function as an OK button for instructing confirmation and execution of selection contents. It is.
- the cross key is an operation unit for inputting instructions in four directions, up, down, left, and right, and functions as a button (cursor moving operation means) for selecting an item from the menu screen or instructing selection of various setting items from each menu. To do.
- the up / down key of the cross key functions as a zoom switch for shooting or a playback zoom switch in playback mode
- the left / right key functions as a frame advance (forward / reverse feed) button in playback mode.
- the BACK key is used to delete a desired object such as a selection item, cancel an instruction content, or return to the previous operation state.
- image light indicating a subject is a light-receiving surface of a solid-state imaging device (hereinafter referred to as “CCD”) 16 that is a phase difference image sensor capable of acquiring a pupil-divided parallax image via an imaging lens 12 and a diaphragm 14.
- CCD solid-state imaging device
- the imaging lens 12 is driven by a lens driving unit 36 controlled by the CPU 40, and focus control, zoom (focal length) control, and the like are performed.
- the diaphragm 14 is composed of, for example, five diaphragm blades, and is driven by a diaphragm drive unit 34 controlled by the CPU 40.
- the diaphragm value (F value) F2.8 to F11 is controlled in five steps in increments of 1AV.
- the CPU 40 controls the diaphragm 14 via the diaphragm driving unit 34, and performs charge accumulation time (shutter speed) in the CCD 16, image signal readout control from the CCD 16, and the like via the CCD control unit 32.
- FIG. 2 is a diagram illustrating a configuration example of the CCD 16.
- the CCD 16 has odd-line pixels (main pixels) and even-line pixels (sub-pixels) arranged in a matrix, respectively, and two surfaces that are photoelectrically converted by these main and sub-pixels, respectively. Minute image signals can be read out independently.
- the plurality of light receiving elements corresponding to each pixel group form an effective pixel region for obtaining an effective imaging signal and an optical black region (hereinafter referred to as “OB region”) for obtaining a black level reference signal.
- the OB region is actually formed so as to surround the effective pixel region.
- pixels of GRGR. .. are arranged alternately, and pixels of even lines (2, 4, 6,...) Are similar to the lines of GRGR... , BGBG... Are arranged alternately, and the pixels are arranged so as to be shifted in the line direction by a half pitch with respect to the pixels of the even lines.
- FIG. 3 is a diagram showing each of the main and sub-pixels of the imaging lens 12, the diaphragm 14, and the CCD 16, and FIG. 4 is an enlarged view of the main part of FIG.
- the light beam passing through the exit pupil enters the normal CCD pixel (photodiode PD) through the microlens L without being restricted.
- a light shielding member 16A is formed in the main pixel and subpixel of the CCD 16.
- the light shielding member 16A shields the right half or the left half of the light receiving surface of the main pixel and sub-pixel (photodiode PD).
- the light shielding member 16A shields the upper half or the lower half of the light receiving surface of the main pixel and subpixel (photodiode PD).
- the opening 16B of the light shielding member 16A is located at a position displaced by a predetermined amount ⁇ from the optical axis Ic of the microlens L in the right, left, upper, or lower direction (for example, downward from the optical axis Ic in FIG. 4B). Is provided.
- the light beam passes through the opening 16B and reaches the light receiving surface of the photodiode PD. That is, the light shielding member 16A functions as a pupil division member.
- the main pixel and the sub-pixel have different regions (right half / left half or upper half / lower half) where the light flux is restricted by the light shielding member 16A.
- the right viewpoint image is obtained from the main pixel and the left viewpoint image is obtained from the subpixel.
- the upper half of the luminous flux is restricted at the main pixel and the lower half of the luminous flux is restricted at the subpixel, the lower viewpoint image is obtained from the main pixel and the upper viewpoint image is obtained from the subpixel.
- the CCD 16 configured as described above is configured so that the main pixel and the sub-pixel have different regions (right half and left half) where the light flux is limited by the light shielding member 16A.
- the configuration of the CCD 16 is not limited to this.
- the microlens L and the photodiode PD may be relatively shifted in the left-right direction, and the light flux incident on the photodiode PD may be limited by the shifting direction.
- the light flux incident on each pixel may be limited, or the pupil may be divided by a mirror (for example, FIG. 11). Further, the number of pupil divisions may be two or more (for example, 4, 9, 16,).
- the signal charge accumulated in the CCD 16 is read out as a voltage signal corresponding to the signal charge based on the readout signal applied from the CCD control unit 32.
- the voltage signal read from the CCD 16 is applied to the analog signal processing unit 18 where the R, G, and B signals for each pixel are sampled and held, amplified, and then applied to the A / D converter 20.
- the A / D converter 20 converts R, G, and B signals that are sequentially input into digital R, G, and B signals and outputs them to the image input controller 22.
- the digital signal processing unit 24 performs predetermined processing such as offset control, gain control processing including white balance correction and sensitivity correction, gamma correction processing, YC processing, etc., on the digital image signal input via the image input controller 22. Perform signal processing.
- predetermined processing such as offset control, gain control processing including white balance correction and sensitivity correction, gamma correction processing, YC processing, etc.
- the main image data read from the main pixels of the odd lines of the CCD 16 is processed as left viewpoint image data, and is read from the sub pixels of the even lines.
- the data is processed as right viewpoint image data.
- the left viewpoint image data and right viewpoint image data (3D image data) processed by the digital signal processing unit 24 are input to the VRAM 50.
- the VRAM 50 includes an A area and a B area each storing 3D image data representing a 3D image for one frame.
- 3D image data representing a 3D image for one frame is rewritten alternately in the A area and the B area.
- the written 3D image data is read from an area other than the area in which the 3D image data is rewritten in the A area and the B area of the VRAM 50.
- the 3D image data read from the VRAM 50 is encoded by the video encoder 28 and output to a stereoscopic display liquid crystal monitor (LCD) 30 provided on the back of the camera, whereby a 3D subject image is displayed on the liquid crystal monitor 30. Displayed on the display screen.
- LCD liquid crystal monitor
- the liquid crystal monitor 30 is a stereoscopic display means capable of displaying a stereoscopic image (left viewpoint image and right viewpoint image) as a directional image having a predetermined directivity by a parallax barrier.
- a stereoscopic image left viewpoint image and right viewpoint image
- the configuration of the liquid crystal monitor 30 is not limited to this, and the left viewpoint image and the right viewpoint image can be individually viewed by using special glasses such as those using a lenticular lens or polarized glasses or liquid crystal shutter glasses. You can do it.
- the CPU 40 starts the AF operation and the AE operation, and the focus lens in the imaging lens 12 is aligned via the lens driving unit 36. Control to come to the focal position.
- the image data output from the A / D converter 20 when the shutter button is half-pressed is taken into the AE detection unit 44.
- the AE detection unit 44 integrates the G signals of the entire screen or integrates the G signals that are weighted differently in the central portion and the peripheral portion of the screen, and outputs the integrated value to the CPU 40.
- the CPU 40 calculates the brightness (photographing Ev value) of the subject from the integrated value input from the AE detection unit 44.
- the CPU 40 determines the aperture value of the aperture 14 and the electronic shutter (shutter speed) of the CCD 16 for obtaining appropriate exposure based on the photographing Ev value according to a predetermined program diagram.
- the CPU 40 controls the aperture 14 via the aperture drive unit 34 based on the determined aperture value (normal aperture control), and accumulates charges in the CCD 16 via the CCD control unit 32 based on the determined shutter speed. Control the time. Note that the brightness of the subject may be calculated based on an external photometric sensor.
- the predetermined program diagram corresponds to a combination of the aperture value of the aperture 14 and the shutter speed of the CCD 16 or a combination of these and the imaging sensitivity (ISO sensitivity) corresponding to the brightness of the subject (imaging EV value).
- the following shooting (exposure) conditions are designed. By shooting under the shooting conditions determined according to the program diagram, it is possible to take a main image and a sub image having a desired parallax regardless of the brightness of the subject.
- the F value is fixed at a constant value, a main image and a sub image having a desired parallax can be taken.
- the shooting EV value is larger than 12 (when the shutter speed reaches the maximum value)
- the exposure is overexposed and the shooting cannot be performed.
- the monocular stereoscopic imaging device 1 has a configuration in which the ND filter is automatically inserted to reduce the light amount. If added, shooting is possible even if the shooting EV value is greater than 12.
- the AF processing unit 42 is a part that performs contrast AF processing or phase difference AF processing.
- the contrast AF process by extracting the high frequency component of the image data in the predetermined focus area from the image data of at least one of the left viewpoint image data and the right viewpoint image data, and integrating the high frequency component
- An AF evaluation value indicating the in-focus state is calculated.
- the AF control is performed by controlling the focus lens in the imaging lens 12 so that the AF evaluation value is maximized.
- the phase difference AF process the phase difference between the image data corresponding to the main pixel and the sub pixel in the predetermined focus area in the left viewpoint image data and the right viewpoint image data is detected, and this phase difference is detected.
- the defocus amount is obtained based on the information indicating.
- AF control is performed by controlling the focus lens in the imaging lens 12 so that the defocus amount becomes zero.
- the two pieces of image data temporarily stored in the memory 48 are appropriately read out by the digital signal processing unit 24, where predetermined signals including generation processing (YC processing) of luminance data and color difference data of the image data are performed. Processing is performed.
- the YC processed image data (YC data) is stored in the memory 48 again. Subsequently, the two pieces of YC data are respectively output to the compression / decompression processing unit 26 and subjected to predetermined compression processing such as JPEG (joint photographic experts group), and then stored in the memory 48 again.
- a multi-picture file (MP file: a file in a format in which a plurality of images are connected) is generated from two pieces of YC data (compressed data) stored in the memory 48, and the MP file is generated by the media controller 52. It is read and recorded in the memory card 54.
- MP file a file in a format in which a plurality of images are connected
- the defocus map creating unit 61 not only calculates the phase difference corresponding to the main pixel and the sub-pixel for each of the small areas included in the predetermined focus area, but also substantially covers the entire effective pixel area. Is calculated for each of the small regions.
- the plurality of small areas that substantially cover the entire effective pixel area need not completely cover the entire effective pixel area, and may be densely or roughly arranged over the entire effective pixel area. That's fine. For example, each of the divided areas obtained by dividing the effective pixel area in a matrix by a predetermined unit (for example, 8 ⁇ 8 pixels), or less (for example, 1 ⁇ 1 pixel), or more (for example, 10 ⁇ 10 pixels).
- the phase difference is calculated.
- the phase difference is calculated for each divided region in a predetermined unit separated from the outer edge of the effective pixel region by a predetermined pitch (for example, one divided region or more or less).
- a predetermined pitch for example, one divided region or more or less.
- the defocus map creation unit 61 obtains a defocus amount corresponding to each of the small areas based on the phase difference calculated for each of the small areas.
- a set of defocus amounts corresponding to each of the small areas obtained over the entire effective pixel area is referred to as a defocus map.
- the defocus map creation unit 61 has a volatile storage medium such as a RAM, and temporarily stores the obtained defocus map.
- the defocus map creation unit 61 may detect feature points and corresponding points between the viewpoint images, and create a defocus map based on a difference in position information between the feature points and the corresponding points.
- the restoration filter storage unit 62 is configured by a nonvolatile storage medium such as a ROM, and the image height (distance from the image center, typically the optical axis center L of the imaging lens 12) of each small region in each viewpoint image. ) And a restoration filter corresponding to the defocus amount (or subject distance).
- the restoration unit 63 deconvolves the small area with the restoration filter selected for each small area of each viewpoint image, and restores the corresponding small area of the viewpoint image.
- the shake amount detection unit 71 is a shake amount detection unit employed in an optical camera shake correction mechanism.
- the shake amount detection unit 71 includes a vibration gyro sensor or the like, and detects the shake amount by detecting the angular velocity around the horizontal axis and the angular velocity around the vertical axis of the camera.
- FIG. 5 is a flowchart of imaging processing executed by the monocular stereoscopic imaging device 1.
- creation of a defocus map, storage of a restoration filter, restoration processing of a viewpoint image using the restoration filter, output processing of the restored viewpoint image, and the like can be performed by an information processing apparatus other than the monocular stereoscopic imaging apparatus 1, such as a personal computer.
- a program for causing the monocular stereoscopic imaging apparatus 1 and other information processing apparatuses to execute the processing shown in the following flowchart is recorded in a non-transitory computer-readable recording medium such as a ROM or a CDROM. This process starts in response to half-pressing of the shutter button.
- the CPU 40 controls the shake amount detection unit 71 to detect the shake amount.
- the detected shake amount includes a shake amount ⁇ x in the yaw direction (horizontal direction) and a shake amount ⁇ y in the pitch direction (vertical direction).
- the CPU 40 controls the diaphragm drive unit 34 to set the diaphragm amount corresponding to the blur amount detected by the blur amount detection unit 71.
- the aperture drive unit 34 controls the aperture of the aperture 14 so as to reach the set aperture amount.
- the diaphragm amount Y depending on the blur amount ( ⁇ x, ⁇ y) is such that the larger the blur scalar amount, the smaller the aperture amount becomes (smaller aperture side) than the aperture amount according to a predetermined program diagram. It is in.
- the relationship between the blur amount and the aperture amount is expressed by a function Y having a negative first-order differential coefficient and parameters ( ⁇ x, ⁇ y).
- Y [ ⁇ x ⁇ x + ⁇ y ⁇ y] + A, where ⁇ x ⁇ ⁇ y is a positive constant.
- A is an aperture amount determined according to normal aperture control. That is, as the blurring scalar amount increases, the aperture amount becomes larger than the normal aperture amount.
- the aperture amount Y is expressed by an F value, the F value becomes larger than the value determined by the normal aperture control as the blurring scalar amount increases. For example, if the F value used in normal aperture control is F1.4 (strong parallax), the range used in aperture amount control in this step is F2.8 (standard parallax) to F5.6 (weak) Parallax).
- the CPU 40 controls the CCD 16 to acquire the viewpoint image data for output in response to an instruction to start imaging by fully pressing the shutter button.
- the restoration unit 63 specifies the image height and the defocus amount of each small area set in each viewpoint image.
- the image height of each small area can be specified by calculating the shortest distance from the image center to each small area.
- the defocus amount of each small area is specified for each viewpoint image based on the defocus map created by the defocus map creation unit 61.
- the restoration unit 63 selects a restoration filter corresponding to the image height and defocus amount of the small area specified for each viewpoint image from the restoration filter storage unit 62.
- the restoration unit 63 deconvolves the small area with the restoration filter selected for each small area of each viewpoint image, and restores the corresponding small area of the viewpoint image.
- the restoration unit 63 generates a stereoscopic image from the viewpoint image whose deterioration has been restored, and stores it again in the VRAM 50 as a restored stereoscopic image.
- the deteriorated viewpoint image may be output to the liquid crystal monitor 30 (S5), or may be output to the memory card 54, that is, compressed and recorded (S6).
- the stereoscopic image after restoration of the VRAM 50 is output to the liquid crystal monitor 30, and a good 3D subject image is displayed on the display screen of the liquid crystal monitor 30. That is, the restoration timing is preferably performed before the output of the liquid crystal monitor 30 for the stereoscopic image.
- the stereoscopic image after restoration of the VRAM 50 may be output to a stereoscopic display other than the liquid crystal monitor 30.
- the viewpoint image whose deterioration has been restored is output to the memory card 54, that is, compressed and recorded. If the restored stereoscopic image recorded in the memory card 54 is reproduced on various stereoscopic displays such as the liquid crystal monitor 30, a good 3D subject image can be displayed in the same manner.
- FIGS. 6A and 6B are image diagrams showing how the parallax between the main image and the sub-image varies depending on the F value of the aperture, and FIG. 6A shows the case where the F value is large (dark). FIG. 6B shows a case where the F value is small (bright).
- the parallax (phase difference) between the main image and the sub-image of the main subject (a person in this example) at the focused position (focus position) is 0.
- the background parallax is smaller as the F value is darker (FIG. 6A), and is larger as the F value is brighter (FIG. 6B).
- the depth of focus is increased and the amount of parallax is decreased, thereby reducing the three-dimensional effect when the blur is large and reducing the influence on the observer.
- FIG. 7 is a block diagram of the monocular stereoscopic imaging device 2 according to the second embodiment. Portions common to the first embodiment shown in FIG. 1 are denoted by the same reference numerals, and detailed description thereof is omitted.
- the monocular stereoscopic imaging device 2 further includes a shake amount range determination control unit 72 and a shake correction unit 73.
- the blur correction unit 73 optically corrects the blur of the subject image corresponding to the blur amount in the yaw direction / pitch direction detected by the blur amount detection unit 71 according to the control of the blur amount range determination control unit 72.
- the blur correction unit 73 controls the yaw direction actuator and the pitch direction actuator, and moves the lens 12 in a direction that cancels the detected blur amount to prevent image blur (lens shift method).
- Various types of optical blur correction methods of the blur correction unit 73 can be employed.
- the image blur may be prevented by controlling the yaw direction actuator and the pitch direction actuator and moving the CCD 16 in the direction to cancel the detected blur amount (image sensor shift method).
- FIG. 8 is a flowchart of imaging processing executed by the monocular stereoscopic imaging device 2. This process starts in response to half-pressing of the shutter button.
- S11 is the same as S1.
- the shake amount range determination control unit 72 determines whether or not the detected shake amount is within a range in which shake correction is possible. If yes, go to S13, if no, go to S14.
- the range in which blur correction is possible is the operating range of optical blur correction. For example, the operating range of the yaw direction actuator and the pitch direction actuator.
- the blur amount range determination control unit 72 controls the blur correction unit 73 to correct image blur corresponding to the detected blur amount. Then, normal aperture control is performed.
- S14 is the same as S2. In this case, no blur correction is performed.
- S15 to S18 are the same as S3 to S6.
- shake correction is performed within the range of the amount of shake that is effective for shake correction. Prevent image blur.
- FIG. 9 is a block diagram of the monocular stereoscopic imaging device 3 according to the third embodiment. Portions common to the first and second embodiments shown in FIG. 1 are denoted by the same reference numerals, and detailed description thereof is omitted.
- the monocular stereoscopic imaging device 3 further includes a shake direction detection unit 74 and a shake direction determination control unit 75.
- the shake direction detection unit 74 detects the direction in which the shake occurs (yaw direction and / or pitch direction).
- FIG. 10 is a flowchart of imaging processing executed by the monocular stereoscopic imaging device 2.
- S21 is the same as S1.
- the shake direction detection unit 74 detects the direction corresponding to the larger shake amount among the shake amounts in the yaw direction and the pitch direction detected by the shake amount detection unit 71 as the shake direction. Note that the blur direction detection unit 74 detects the pupil division direction as the blur direction when both blur amounts are the same.
- the shake direction determination control unit 75 determines whether or not the shake direction detected by the shake direction detection unit 74 is different from the pupil division direction. If Yes (both are different), the process proceeds to S24, and if No (both match), the process proceeds to S25. For example, when the pupil division direction is the horizontal direction, if the blur direction is the yaw direction, the process proceeds to S24, and if the blur direction is the pitch direction, the process proceeds to S25.
- S24 to S29 are the same as S13 to S18.
- normal aperture control is performed. That is, the photographing Ev value is calculated from the electrical signal output from the CCD 16, the aperture amount is controlled according to the photographing Ev value and the program diagram, and the luminance level of the subject image is controlled to an appropriate luminance level.
- the blur amount range determination control unit 72 may control the blur correction unit 73 to correct the image blur corresponding to the detected blur amount.
- the diaphragm is further narrowed to reduce the influence of the blur on the parallax.
- the blurring occurs in a direction perpendicular to the pupil division direction, it is considered that the blurring has little influence on the visual perception of the stereoscopic image, so only blurring correction is performed.
- the shake direction determination control unit 75 determines whether or not the shake direction detected by the shake direction detection unit 74 is different from the pupil division direction. If the blur direction determination control unit 75 determines that the blur direction is not equal to the pupil division direction, the process proceeds to S13. The subsequent processing is the same as in the second embodiment.
- the shake amount range determination control unit 72 determines whether or not the detected shake amount is within a range in which shake correction is possible. In the case of Yes, it progresses to S13.
- the subsequent processing is the same as in the second embodiment. If no, the process proceeds to S25.
- the subsequent processing is the same as in the third embodiment.
- the aperture can be reduced more than usual, and the influence on parallax can be reduced.
Landscapes
- Engineering & Computer Science (AREA)
- Multimedia (AREA)
- Signal Processing (AREA)
- Testing, Inspecting, Measuring Of Stereoscopic Televisions And Televisions (AREA)
- Studio Devices (AREA)
- Focusing (AREA)
- Stereoscopic And Panoramic Photography (AREA)
- Automatic Focus Adjustment (AREA)
- Adjustment Of Camera Lenses (AREA)
Abstract
Description
g=f*h+n (*は畳み込み積分) ・・・(A)
で表される。g、h、nを既知として、式(A)から元の物体の輝度分布fを求めることができる。このようにして、光学系のボケを信号処理により取り除き、理想的な像を得る技術は、像の「復元」、「逆畳み込み」、あるいは「デコンボリューション(Deconvolution )」と呼ばれている。点像強度分布(PSF)に基づく復元フィルタは、撮像時の画像の劣化に関する情報、例えば撮影条件(露光時間、露光量、被写体までの距離、焦点距離等)や撮像装置の特性情報(レンズの光学特性、撮像装置の識別情報等)等を考慮して生成される(特許文献6)。
図1は第1実施形態に係る単眼立体撮像装置1の実施の形態を示すブロック図である。
図2はCCD16の構成例を示す図である。
。
図7は第2実施形態に係る単眼立体撮像装置2のブロック図である。図1に示した第1実施形態と共通する部分には同一の符号を付し、その詳細な説明は省略する。
図9は第3実施形態に係る単眼立体撮像装置3のブロック図である。図1に示した第1・2実施形態と共通する部分には同一の符号を付し、その詳細な説明は省略する。
第2実施形態と第3実施形態の処理は組み合わせることができる。すなわち、まず、ぶれ方向判定制御部75は、ぶれ方向検出部74の検出したぶれ方向が、瞳分割方向と異なるか否かを判断する。ぶれ方向判定制御部75が、ぶれ方向≠瞳分割方向と判定した場合、S13に進む。この後の処理は第2実施形態と同様である。
Claims (7)
- 被写体からの光束を瞳分割により2つの光束に分割し、前記2つの光束をそれぞれ対応する2つの光電変換素子群に結像して光電変換することで、視差を有する1対の画像を出力可能な撮像部と、
前記撮像部の光電変換素子群に結像する光束の光量を調整可能な絞り部と、
前記光束の光量を検出する光量検出部と、
前記光量検出部の検出した光量に応じ、所望の視差が得られるよう前記絞り部の開放量を制御する第1の絞り制御を実行可能な絞り制御部と、
前記撮像部による画像の撮影時における被写体像のぶれ量を検出するぶれ検出部と、を備え、
前記絞り制御部は前記ぶれ検出部の検出したぶれ量が大きいほど前記絞り部の開放量を前記第1の絞り制御に必要な開放量よりもさらに小さくする第2の絞り制御を実行する立体撮像装置。 - 前記ぶれ検出部の検出したぶれ量に応じて被写体像のぶれを光学的に補正するぶれ補正部と、
前記ぶれ検出部の検出したぶれ量が前記ぶれ補正部のぶれ補正可能な範囲内であるか否かを判定するぶれ量判定部を備え、
前記絞り制御部は、前記ぶれ量判定部が前記ぶれ量が前記ぶれ補正可能な範囲外であると判定した場合、前記第2の絞り制御を実行する請求項1に記載の立体撮像装置。 - 前記ぶれ補正部は、前記ぶれ量判定部が前記ぶれ量が前記ぶれ補正可能な範囲内であると判定した場合、前記被写体像のぶれを光学的に補正し、
前記絞り制御部は、前記ぶれ量判定部が前記ぶれ量が前記ぶれ補正可能な範囲内であると判定した場合、前記第1の絞り制御を実行する請求項2に記載の立体撮像装置。 - 前記被写体像のぶれ方向を検出するぶれ方向検出部と、
前記ぶれ方向検出部の検出したぶれ方向が前記瞳分割の方向と一致するか否かを判定するぶれ方向判定部と、を備え、
前記絞り制御部は、前記ぶれ方向判定部が前記ぶれ方向が前記瞳分割の方向と一致すると判定した場合、前記第2の絞り制御を実行する請求項1~3のいずれかに記載の立体撮像装置。 - 前記絞り制御部は、前記ぶれ方向判定部が前記ぶれ方向が前記瞳分割の方向と不一致であると判定した場合、前記第1の絞り制御を実行する請求項4に記載の立体撮像装置。
- 立体撮像装置が、
被写体からの光束を瞳分割により2つの光束に分割し、前記2つの光束をそれぞれ対応する2つの光電変換素子群に結像して光電変換することで、視差を有する1対の画像を出力するステップと、
絞り部の開放量で前記光電変換素子群に結像する光束の光量を調整するステップと、
前記光束の光量を検出するステップと、
前記検出した光量に応じ、所望の視差が得られるよう前記絞り部の開放量を制御する第1の絞り制御を実行するステップと、
前記画像の撮影時における被写体像のぶれ量を検出するステップと、
前記検出したぶれ量が大きいほど前記絞り部の開放量を前記第1の絞り制御に必要な開放量よりもさらに小さくする第2の絞り制御を実行するステップと、
を実行する立体撮像方法。 - 立体撮像装置が、
被写体からの光束を瞳分割により2つの光束に分割し、前記2つの光束をそれぞれ対応する2つの光電変換素子群に結像して光電変換することで、視差を有する1対の画像を出力するステップと、
絞り部の開放量で前記光電変換素子群に結像する光束の光量を調整するステップと、
前記光束の光量を検出するステップと、
前記検出した光量に応じ、所望の視差が得られるよう前記絞り部の開放量を制御する第1の絞り制御を実行するステップと、
前記画像の撮影時における被写体像のぶれ量を検出するステップと、
前記検出したぶれ量が大きいほど前記絞り部の開放量を前記第1の絞り制御に必要な開放量よりもさらに小さくする第2の絞り制御を実行するステップと、
を実行する立体撮像プログラム。
Priority Applications (3)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
JP2012532885A JP5502205B2 (ja) | 2010-09-10 | 2011-06-17 | 立体撮像装置および立体撮像方法 |
CN201180043307.4A CN103098457B (zh) | 2010-09-10 | 2011-06-17 | 立体成像装置和立体成像方法 |
US13/789,016 US9282316B2 (en) | 2010-09-10 | 2013-03-07 | Stereoscopic imaging device and stereoscopic imaging method |
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
JP2010-203394 | 2010-09-10 | ||
JP2010203394 | 2010-09-10 |
Related Child Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US13/789,016 Continuation US9282316B2 (en) | 2010-09-10 | 2013-03-07 | Stereoscopic imaging device and stereoscopic imaging method |
Publications (1)
Publication Number | Publication Date |
---|---|
WO2012032826A1 true WO2012032826A1 (ja) | 2012-03-15 |
Family
ID=45810431
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
PCT/JP2011/063880 WO2012032826A1 (ja) | 2010-09-10 | 2011-06-17 | 立体撮像装置および立体撮像方法 |
Country Status (4)
Country | Link |
---|---|
US (1) | US9282316B2 (ja) |
JP (1) | JP5502205B2 (ja) |
CN (1) | CN103098457B (ja) |
WO (1) | WO2012032826A1 (ja) |
Cited By (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
WO2013146066A1 (ja) * | 2012-03-30 | 2013-10-03 | 富士フイルム株式会社 | 画像処理装置及び方法並びに撮像装置 |
WO2013190814A1 (ja) * | 2012-06-22 | 2013-12-27 | 株式会社ニコン | 画像処理装置、撮像装置および画像処理プログラム |
JP2014229974A (ja) * | 2013-05-20 | 2014-12-08 | キヤノン株式会社 | 撮像装置および撮像装置の制御方法 |
JP2015015629A (ja) * | 2013-07-05 | 2015-01-22 | 株式会社ニコン | 画像処理装置および撮像装置 |
US9288472B2 (en) | 2012-05-09 | 2016-03-15 | Fujifilm Corporation | Image processing device and method, and image capturing device |
JP2020009255A (ja) * | 2018-07-10 | 2020-01-16 | 凸版印刷株式会社 | 三次元形状モデル生成装置、三次元モデル生成方法及びプログラム |
Families Citing this family (8)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP2013061440A (ja) * | 2011-09-13 | 2013-04-04 | Canon Inc | 撮像装置および撮像装置の制御方法 |
JP6210824B2 (ja) | 2013-10-02 | 2017-10-11 | オリンパス株式会社 | 焦点調節装置および焦点調節方法 |
JP6478511B2 (ja) | 2014-08-01 | 2019-03-06 | キヤノン株式会社 | 画像処理方法、画像処理装置、複眼撮像装置、画像処理プログラム、および、記憶媒体 |
JP6494333B2 (ja) * | 2015-03-04 | 2019-04-03 | キヤノン株式会社 | 画像処理装置及び画像処理方法、及び撮像装置 |
WO2016168788A2 (en) * | 2015-04-17 | 2016-10-20 | Tulip Interfaces, Inc. | Containerized communications gateway |
CN105208366A (zh) * | 2015-09-16 | 2015-12-30 | 云南师范大学 | 一种用于近视患者立体视觉增强的方法 |
CN109906604B (zh) * | 2016-11-09 | 2021-04-20 | 富士胶片株式会社 | 摄像装置、摄像方法及记录介质 |
JP6776392B2 (ja) * | 2019-03-29 | 2020-10-28 | エーエーシー コミュニケーション テクノロジーズ(ジョウシュウ)カンパニーリミテッド | カメラ用レンズ駆動装置 |
Citations (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP2005045328A (ja) * | 2003-07-22 | 2005-02-17 | Sharp Corp | 3次元画像撮像装置 |
JP2008187385A (ja) * | 2007-01-29 | 2008-08-14 | Fujifilm Corp | 撮影装置 |
Family Cites Families (24)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US3712199A (en) * | 1970-09-23 | 1973-01-23 | Video West Inc | Three-dimensional color photographic process, apparatus and product |
US6163337A (en) * | 1996-04-05 | 2000-12-19 | Matsushita Electric Industrial Co., Ltd. | Multi-view point image transmission method and multi-view point image display method |
EP0817125B1 (en) * | 1996-06-26 | 2004-09-08 | Matsushita Electric Industrial Co., Ltd. | Stereoscopic computer graphics moving image generating apparatus |
JP3397602B2 (ja) * | 1996-11-11 | 2003-04-21 | 富士通株式会社 | 画像表示装置及び方法 |
AUPO894497A0 (en) * | 1997-09-02 | 1997-09-25 | Xenotech Research Pty Ltd | Image processing method and apparatus |
ID27878A (id) * | 1997-12-05 | 2001-05-03 | Dynamic Digital Depth Res Pty | Konversi image yang ditingkatkan dan teknik mengenkodekan |
JP2000020691A (ja) | 1998-07-01 | 2000-01-21 | Canon Inc | 画像処理装置及びその方法、撮像装置及びその制御方法並びにメモリ媒体 |
US6822758B1 (en) | 1998-07-01 | 2004-11-23 | Canon Kabushiki Kaisha | Image processing method, system and computer program to improve an image sensed by an image sensing apparatus and processed according to a conversion process |
JP2001166258A (ja) * | 1999-12-10 | 2001-06-22 | Canon Inc | 立体画像撮影用光学系及びそれを用いた立体画像撮影装置 |
JP4305598B2 (ja) | 2000-06-05 | 2009-07-29 | 富士フイルム株式会社 | カメラの絞り制御方法及び装置、並びにカメラ |
JP4871942B2 (ja) | 2000-06-05 | 2012-02-08 | 富士フイルム株式会社 | カメラの絞り制御方法及び装置、並びにカメラ |
NZ505513A (en) * | 2000-06-30 | 2002-05-31 | Marc Dawson | Anaglyphic 3-D colour imaging with selective colour filtering |
IL150131A (en) * | 2002-06-10 | 2007-03-08 | Rafael Advanced Defense Sys | A method for turning a series of monoscopic images into a series of stereoscopic images |
JP2005167310A (ja) | 2003-11-28 | 2005-06-23 | Sharp Corp | 撮影装置 |
WO2006089417A1 (en) * | 2005-02-23 | 2006-08-31 | Craig Summers | Automatic scene modeling for the 3d camera and 3d video |
JP5426174B2 (ja) | 2006-02-13 | 2014-02-26 | スリーエム イノベイティブ プロパティズ カンパニー | 単眼三次元撮像 |
US7372642B2 (en) | 2006-02-13 | 2008-05-13 | 3M Innovative Properties Company | Three-channel camera systems with non-collinear apertures |
US8619865B2 (en) * | 2006-02-16 | 2013-12-31 | Vidyo, Inc. | System and method for thinning of scalable video coding bit-streams |
JP2008020543A (ja) | 2006-07-11 | 2008-01-31 | Fujifilm Corp | 撮像装置 |
JP2008172321A (ja) | 2007-01-09 | 2008-07-24 | Olympus Imaging Corp | 電気的画像回復処理を行う撮像装置 |
JP2008311743A (ja) * | 2007-06-12 | 2008-12-25 | Canon Inc | 撮像装置 |
GB0716776D0 (en) * | 2007-08-29 | 2007-10-10 | Setred As | Rendering improvement for 3D display |
JP2010136269A (ja) | 2008-12-08 | 2010-06-17 | Fujifilm Corp | デジタルカメラ |
JP5503438B2 (ja) * | 2009-07-21 | 2014-05-28 | 富士フイルム株式会社 | 3次元画像表示装置及び3次元画像表示方法 |
-
2011
- 2011-06-17 CN CN201180043307.4A patent/CN103098457B/zh not_active Expired - Fee Related
- 2011-06-17 JP JP2012532885A patent/JP5502205B2/ja not_active Expired - Fee Related
- 2011-06-17 WO PCT/JP2011/063880 patent/WO2012032826A1/ja active Application Filing
-
2013
- 2013-03-07 US US13/789,016 patent/US9282316B2/en not_active Expired - Fee Related
Patent Citations (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP2005045328A (ja) * | 2003-07-22 | 2005-02-17 | Sharp Corp | 3次元画像撮像装置 |
JP2008187385A (ja) * | 2007-01-29 | 2008-08-14 | Fujifilm Corp | 撮影装置 |
Cited By (16)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP5740045B2 (ja) * | 2012-03-30 | 2015-06-24 | 富士フイルム株式会社 | 画像処理装置及び方法並びに撮像装置 |
WO2013146066A1 (ja) * | 2012-03-30 | 2013-10-03 | 富士フイルム株式会社 | 画像処理装置及び方法並びに撮像装置 |
CN104205827B (zh) * | 2012-03-30 | 2016-03-16 | 富士胶片株式会社 | 图像处理装置及方法、以及摄像装置 |
CN104205827A (zh) * | 2012-03-30 | 2014-12-10 | 富士胶片株式会社 | 图像处理装置及方法、以及摄像装置 |
US9277201B2 (en) | 2012-03-30 | 2016-03-01 | Fujifilm Corporation | Image processing device and method, and imaging device |
US9288472B2 (en) | 2012-05-09 | 2016-03-15 | Fujifilm Corporation | Image processing device and method, and image capturing device |
CN104322061B (zh) * | 2012-06-22 | 2017-01-18 | 株式会社尼康 | 图像处理装置、摄像装置及图像处理方法 |
US20150062308A1 (en) * | 2012-06-22 | 2015-03-05 | Nikon Corporation | Image processing apparatus, image-capturing apparatus and image processing method |
JPWO2013190814A1 (ja) * | 2012-06-22 | 2016-02-08 | 株式会社ニコン | 画像処理装置、撮像装置および画像処理プログラム |
CN104322061A (zh) * | 2012-06-22 | 2015-01-28 | 株式会社尼康 | 图像处理装置、摄像装置及图像处理程序 |
WO2013190814A1 (ja) * | 2012-06-22 | 2013-12-27 | 株式会社ニコン | 画像処理装置、撮像装置および画像処理プログラム |
US9723288B2 (en) * | 2012-06-22 | 2017-08-01 | Nikon Corporation | Image processing apparatus, image-capturing apparatus and image processing method |
JP2014229974A (ja) * | 2013-05-20 | 2014-12-08 | キヤノン株式会社 | 撮像装置および撮像装置の制御方法 |
JP2015015629A (ja) * | 2013-07-05 | 2015-01-22 | 株式会社ニコン | 画像処理装置および撮像装置 |
JP2020009255A (ja) * | 2018-07-10 | 2020-01-16 | 凸版印刷株式会社 | 三次元形状モデル生成装置、三次元モデル生成方法及びプログラム |
JP7135517B2 (ja) | 2018-07-10 | 2022-09-13 | 凸版印刷株式会社 | 三次元形状モデル生成装置、三次元モデル生成方法及びプログラム |
Also Published As
Publication number | Publication date |
---|---|
CN103098457A (zh) | 2013-05-08 |
JP5502205B2 (ja) | 2014-05-28 |
US20130182082A1 (en) | 2013-07-18 |
JPWO2012032826A1 (ja) | 2014-01-20 |
US9282316B2 (en) | 2016-03-08 |
CN103098457B (zh) | 2016-04-13 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
JP5502205B2 (ja) | 立体撮像装置および立体撮像方法 | |
JP5192096B2 (ja) | 立体撮像装置 | |
JP5180407B2 (ja) | 立体撮像装置および視差画像復元方法 | |
JP5640143B2 (ja) | 撮像装置及び撮像方法 | |
JP5421829B2 (ja) | 撮像装置 | |
JP5368350B2 (ja) | 立体撮像装置 | |
JP5690396B2 (ja) | 撮像装置およびシェーディング補正方法 | |
JP5469258B2 (ja) | 撮像装置および撮像方法 | |
JP5871989B2 (ja) | 撮影装置、撮影方法及びプログラム | |
WO2013069445A1 (ja) | 立体撮像装置及び画像処理方法 | |
WO2013018471A1 (ja) | 撮像装置 | |
JP5595505B2 (ja) | 立体撮像装置および立体撮像方法 | |
WO2012002298A1 (ja) | 撮像装置および撮像方法 | |
JP5457240B2 (ja) | 立体撮像装置 | |
JP2012124650A (ja) | 撮像装置および撮像方法 | |
JP2014063190A (ja) | 撮像装置 | |
JP5649837B2 (ja) | 立体撮像装置 |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
WWE | Wipo information: entry into national phase |
Ref document number: 201180043307.4 Country of ref document: CN |
|
121 | Ep: the epo has been informed by wipo that ep was designated in this application |
Ref document number: 11823305 Country of ref document: EP Kind code of ref document: A1 |
|
WWE | Wipo information: entry into national phase |
Ref document number: 2012532885 Country of ref document: JP |
|
NENP | Non-entry into the national phase |
Ref country code: DE |
|
122 | Ep: pct application non-entry in european phase |
Ref document number: 11823305 Country of ref document: EP Kind code of ref document: A1 |