WO2011118089A1 - 立体撮像装置 - Google Patents
立体撮像装置 Download PDFInfo
- Publication number
- WO2011118089A1 WO2011118089A1 PCT/JP2010/071473 JP2010071473W WO2011118089A1 WO 2011118089 A1 WO2011118089 A1 WO 2011118089A1 JP 2010071473 W JP2010071473 W JP 2010071473W WO 2011118089 A1 WO2011118089 A1 WO 2011118089A1
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- image
- image processing
- correction
- processing unit
- stereoscopic
- Prior art date
Links
Images
Classifications
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N13/00—Stereoscopic video systems; Multi-view video systems; Details thereof
-
- G—PHYSICS
- G03—PHOTOGRAPHY; CINEMATOGRAPHY; ANALOGOUS TECHNIQUES USING WAVES OTHER THAN OPTICAL WAVES; ELECTROGRAPHY; HOLOGRAPHY
- G03B—APPARATUS OR ARRANGEMENTS FOR TAKING PHOTOGRAPHS OR FOR PROJECTING OR VIEWING THEM; APPARATUS OR ARRANGEMENTS EMPLOYING ANALOGOUS TECHNIQUES USING WAVES OTHER THAN OPTICAL WAVES; ACCESSORIES THEREFOR
- G03B35/00—Stereoscopic photography
- G03B35/08—Stereoscopic photography by simultaneous recording
- G03B35/10—Stereoscopic photography by simultaneous recording having single camera with stereoscopic-base-defining system
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N13/00—Stereoscopic video systems; Multi-view video systems; Details thereof
- H04N13/10—Processing, recording or transmission of stereoscopic or multi-view image signals
- H04N13/106—Processing image signals
- H04N13/133—Equalising the characteristics of different image components, e.g. their average brightness or colour balance
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N23/00—Cameras or camera modules comprising electronic image sensors; Control thereof
- H04N23/60—Control of cameras or camera modules
- H04N23/67—Focus control based on electronic image sensor signals
- H04N23/672—Focus control based on electronic image sensor signals based on the phase difference signals
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N23/00—Cameras or camera modules comprising electronic image sensors; Control thereof
- H04N23/60—Control of cameras or camera modules
- H04N23/67—Focus control based on electronic image sensor signals
- H04N23/673—Focus control based on electronic image sensor signals based on contrast or high frequency components of image signals, e.g. hill climbing method
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N23/00—Cameras or camera modules comprising electronic image sensors; Control thereof
- H04N23/60—Control of cameras or camera modules
- H04N23/69—Control of means for changing angle of the field of view, e.g. optical zoom objectives or electronic zooming
Definitions
- the present invention relates to a stereoscopic imaging apparatus, and more particularly to a technique for obtaining a left viewpoint image and a right viewpoint image by forming subject images that have passed through different regions in the left-right direction of an imaging optical system on imaging elements, respectively.
- Patent Document 1 Japanese Patent Document 1
- a subject image that has passed through different regions in the left-right direction of the main lens 1 and the relay lens 2 is divided into pupils by a mirror 4 and formed on imaging elements 7 and 8 via imaging lenses 5 and 6, respectively. I am doing so.
- FIG. 17 are diagrams showing separation states of images formed on the image sensor due to differences in the front pin, in-focus (best focus), and rear pin, respectively.
- the mirror 4 shown in FIG. 16 is omitted in order to compare the difference in separation due to focus.
- the focused image of the pupil-divided images is formed (matched) at the same position on the image sensor, but FIG. 17A and FIG. As shown in C), the images to be the front pin and the rear pin are formed (separated) at different positions on the image sensor.
- Patent Document 2 there is one described in Patent Document 2 as a prior art related to a stereoscopic imaging device.
- Patent Document 3 a digital camera that extracts a person image area from an image and performs different sharpness correction on the extracted person image area and the other background image area has been proposed as image processing for producing a sense of depth and stereoscopic effect.
- the left viewpoint image and the right viewpoint image acquired by the stereoscopic imaging device described in Patent Documents 1 and 2 and the like correspond to subject images that are pupil-divided in the left-right direction. For this reason, there is a problem in that, due to the influence of the difference in optical performance (chromatic aberration, distortion, etc.) in the left-right direction of the photographic optical system, the blurring method becomes unbalanced and stereoscopic viewing is difficult.
- the digital camera described in Patent Document 3 applies different sharpness corrections to different areas within a single screen (for each subject distance such as a person image area and other background image areas) to change the sense of depth and stereoscopic effect. ing.
- the sharpness correction in Patent Document 3 is not a sharpness correction for a stereoscopic image including a left viewpoint image and a right viewpoint image in the first place.
- the present invention has been made in view of such circumstances, and a stereoscopic imaging apparatus capable of correcting an imbalance between a left viewpoint image and a right viewpoint image and acquiring a stereoscopic image suitable for stereoscopic viewing.
- the purpose is to provide.
- a stereoscopic imaging apparatus includes a single photographing optical system and first and second regions having different predetermined directions of the photographing optical system.
- An imaging device that forms an image of a passed subject image by dividing the pupil, and photoelectrically converts the subject images that have passed through the first and second regions to output a first image and a second image, respectively.
- an image processing unit that performs a first image process on the first image and a second image process different from the first image process on the second image.
- an image processing unit that performs the first image processing and the second image processing so as to reduce a difference in image quality between the first image and the second image after each processing.
- the difference in image quality between the two images is reduced. Unbalance including a difference in the blurring method can be corrected), and thereby a stereoscopic image that can be easily stereoscopically viewed can be obtained.
- the stereoscopic imaging apparatus is configured such that, in the first aspect, the image processing unit performs different sharpness corrections on the first image and the second image, respectively. ing.
- the stereoscopic imaging apparatus includes the parallax amount calculation unit that calculates a parallax amount based on the first image and the second image in the second aspect, and the image processing unit. Is configured to weaken the sharpness correction for an image in a large range of the calculated parallax amount. Since an image with a large amount of parallax is blurred, the image quality cannot be improved even if a large sharpness correction is applied to the image within this range, and the blurred range is a difference in optical characteristics (such as distortion). This is because when the sharpness correction is applied, the difference in image quality becomes more remarkable and the image quality becomes unfavorable.
- the image processing unit performs different chromatic aberration correction on the first image and the second image, respectively. It is configured.
- the image processing unit performs different distortion correction on the first image and the second image, respectively. It is configured.
- a stereoscopic imaging apparatus in the first to fifth aspects, includes a first correction parameter and a second correction used for image processing of the first image and the second image.
- a storage unit that stores parameters in advance, and the image processing unit performs image processing on the first image and the second image based on the stored first correction parameter and second correction parameter, respectively. It is configured as follows.
- the photographing optical system is a zoom lens
- the storage unit corresponds to a zoom position of the zoom lens or zooms.
- the first correction parameter and the second correction parameter are stored in correspondence with the position and the aperture value, respectively, and the image processing unit reads from the storage unit based on the zoom position at the time of shooting or the zoom position and the aperture value.
- Corresponding first correction parameters and second correction parameters are read, and image processing of the first image and the second image is performed based on the read first correction parameters and second correction parameters, respectively. It is configured.
- the optical performance in the horizontal direction of the photographic optical system varies depending on the zoom position of the photographic optical system (zoom lens) or the zoom position and the aperture value, but it corresponds to the zoom position of the zoom lens or the zoom position and the aperture value. Since correction is performed using the stored first correction parameter and second correction parameter, more appropriate image processing can be performed on the first image and the second image.
- a stereoscopic imaging apparatus provides a stereoscopic image recording medium including the first image and the second image processed by the image processing unit in the first to seventh aspects.
- a recording unit for recording is provided.
- the image processing unit is configured to capture the planar image obtained by adding the first image and the second image. It is configured to perform image processing different from the first image processing and the second image processing.
- a stereoscopic imaging device in the first to seventh aspects, includes information indicating a stereoscopic image together with the first image and the second image before processing by the image processing unit.
- a recording unit that records information indicating a planar image in association with the recording medium, and the image processing unit performs the first image and the second image read from the recording medium when outputting the stereoscopic image with respect to the first image and the second image.
- the first image processing and the second image processing are performed, and image processing different from the first image processing and the second image processing is performed at the time of outputting a planar image.
- a stereoscopic imaging apparatus is the first to tenth aspects, in which the imaging elements are each a first group of photoelectric conversion pixels and a second group of pixels arranged in a matrix.
- a first group of pixels that are limited in the light receiving direction of the light beam so as to receive only a subject image that has passed through the first region of the photographing optical system; and a second region of the photographing optical system
- a second group of pixels that are limited in the light receiving direction of the light beam so as to receive only the subject image that has passed through the first image and the second group of pixels from the first group of pixels and the second group of pixels.
- the second image can be read out.
- the present invention it is possible to correct the imbalance of the left and right viewpoint images by performing separate image processing on the first and second images (left and right viewpoint images) acquired by the stereoscopic imaging device, Thereby, a stereoscopic image suitable for stereoscopic vision can be acquired.
- FIG. 4 is a diagram illustrating a configuration example of a phase difference CCD (subpixel).
- the figure which showed one pixel each of the main and sub pixel of an imaging optical system and phase contrast CCD Fig. 3 is an enlarged view of the main part (normal pixel). 3 is an enlarged view of the main part (phase difference pixel).
- the figure which shows the 3rd example of the aberration of the lens of an imaging optical system 6 is a flowchart showing a shooting operation of the stereoscopic imaging apparatus according to the first embodiment of the present invention.
- Diagram showing point image before correction Diagram showing point image after correction
- a diagram showing the difference in point images depending on the presence or absence of parallax (no parallax) A diagram showing the difference in point images depending on the presence or absence of parallax (with parallax)
- movement of the stereo imaging device of the 5th Embodiment of this invention The figure used for demonstrating the effect
- the figure which shows an example of the optical system of the conventional stereoscopic imaging device The figure which shows the principle by which an image with a phase difference is imaged with a stereo imaging device
- FIG. 1 is a block diagram showing an embodiment of a stereoscopic imaging apparatus 10 according to the present invention.
- the stereoscopic imaging apparatus 10 records the captured image on the memory card 54, and the overall operation of the apparatus is centrally controlled by a central processing unit (CPU: Central Processing Unit) 40.
- CPU Central Processing Unit
- the stereoscopic imaging device 10 is provided with operation units 38 such as a shutter button, a mode dial, a playback button, a MENU / OK key, a cross key, and a BACK key.
- operation units 38 such as a shutter button, a mode dial, a playback button, a MENU / OK key, a cross key, and a BACK key.
- a signal from the operation unit 38 is input to the CPU 40, and the CPU 40 controls each circuit of the stereoscopic imaging device 10 based on the input signal. For example, lens driving control, aperture driving control, photographing operation control, image processing control, image processing Data recording / reproduction control, display control of the liquid crystal monitor 30 for stereoscopic display, and the like are performed.
- the shutter button is an operation button for inputting an instruction to start shooting, and is configured by a two-stroke switch having an S1 switch that is turned on when half-pressed and an S2 switch that is turned on when fully pressed.
- the mode dial is a selection operation member that selects a 2D shooting mode, a 3D shooting mode, an auto shooting mode, a manual shooting mode, a scene position such as a person, a landscape, a night view, and a moving image mode for shooting a moving image.
- the playback button is a button for switching to a playback mode in which a still image or a moving image of a stereoscopic image (3D image) or a planar image (2D image) that has been recorded is displayed on the liquid crystal monitor 30.
- the MENU / OK key is an operation key having both a function as a menu button for instructing to display a menu on the screen of the liquid crystal monitor 30 and a function as an OK button for instructing confirmation and execution of the selection contents. It is.
- the cross key is an operation unit for inputting instructions in four directions, up, down, left, and right, and is a button (cursor moving operation member) for selecting an item from a menu screen or instructing selection of various setting items from each menu. Function.
- the up / down key of the cross key functions as a zoom switch for shooting or a playback zoom switch in playback mode
- the left / right key functions as a frame advance (forward / reverse feed) button in playback mode.
- the BACK key is used to delete a desired object such as a selection item, cancel an instruction content, or return to the previous operation state.
- phase difference CCD charge-coupled device
- the photographing optical system 12 is driven by a lens driving unit 36 controlled by the CPU 40, and performs focus control, zoom control, and the like.
- the diaphragm 14 is composed of, for example, five diaphragm blades, and is driven by the diaphragm driving unit 34 controlled by the CPU 40. For example, the diaphragm 14 is controlled in five stages from the aperture value F2.8 to F11 in increments of 1AV.
- the CPU 40 controls the diaphragm 14 via the diaphragm driving unit 34, and controls the charge accumulation time (shutter speed) in the phase difference CCD 16 and the image signal readout control from the phase difference CCD 16 via the CCD control unit 36. Etc.
- ⁇ Configuration example of phase difference CCD> 2A to 2C are diagrams showing an example of the configuration of the phase difference CCD 16.
- the phase difference CCD 16 has odd-numbered lines of pixels (main pixels) and even-numbered lines of pixels (sub-pixels) arranged in a matrix, and photoelectric conversion is performed in these main and sub-pixels.
- the image signals for the two surfaces can be read independently.
- the pixels of all lines of the phase difference CCD 16 may be sequentially read, and then the main image composed of pixels of odd lines and the sub-image composed of pixels of even lines may be separated.
- the odd lines (1, 3, 5,...) Of the phase difference CCD 16 include pixels of R (red), G (green), and B (blue) color filters.
- the GRGR... Pixel array line and the BGBG... Pixel array line are alternately provided, while the even line (2, 4, 6,%) Pixels have the GRGR.
- the pixel array lines and BGBG... Pixel array lines are alternately provided, and the pixels are arranged so as to be shifted in the line direction by a half pitch with respect to the even-line pixels.
- FIG. 3 is a diagram showing each of the main and sub-pixels of the photographing optical system 12, the diaphragm 14, and the phase difference CCD 16, and FIGS. 4A and 4B are enlarged views of the main part of FIG.
- the light beam passing through the exit pupil enters the normal CCD pixel (photodiode PD) through the microlens L without being restricted.
- a light shielding member 16A is formed in the main pixel and subpixel of the phase difference CCD 16, and the right half or the left half of the light receiving surface of the main pixel and subpixel (photodiode PD) is shielded by the light shielding member 16A. ing. That is, the light shielding member 16A functions as a pupil division member.
- the phase difference CCD 16 having the above configuration is configured so that the main pixel and the sub-pixel have different regions (right half and left half) in which the light flux is restricted by the light shielding member 16A. Without providing the light shielding member 16A, the microlens L and the photodiode PD may be relatively shifted in the left-right direction, and the light flux incident on the photodiode PD may be limited by the shifting direction. By providing one microlens for each pixel (main pixel and subpixel), the light flux incident on each pixel may be limited.
- the signal charge accumulated in the phase difference CCD 16 is read out as a voltage signal corresponding to the signal charge based on the readout signal applied from the CCD control unit 32.
- the voltage signal read from the phase difference CCD 16 is applied to the analog signal processing unit 18 where the R, G, and B signals for each pixel are sampled and held and then added to the A / D converter 20. It is done.
- the A / D converter 20 converts R, G, and B signals that are sequentially input into digital R, G, and B signals and outputs them to the image input controller 22.
- the digital signal processing unit 24 performs gain control processing including offset processing, white balance correction, sensitivity correction, gamma correction processing, distortion correction processing, and chromatic aberration correction on a digital image signal input via the image input controller 22. Predetermined signal processing such as processing, synchronization processing, YC processing, and sharpness correction is performed.
- ROM Read-Only Memory; for example, EEPROM (Electrically Erasable Programmable Read-Only Memory)
- EEPROM Electrical Erasable Programmable Read-Only Memory
- the main image data read from the odd-line main pixels of the phase difference CCD 16 is processed as left viewpoint image data
- the sub-image data read from the even-line sub-pixels is Are processed as right viewpoint image data.
- the left viewpoint image data and right viewpoint image data (3D image data) processed by the digital signal processing unit 24 are input to a VRAM (Video Random Access Memory) 50.
- the VRAM 50 includes an A area and a B area each storing 3D image data representing a 3D image for one frame.
- 3D image data representing a 3D image for one frame is rewritten alternately in the A area and the B area.
- the written 3D image data is read from an area other than the area in which the 3D image data is rewritten in the A area and the B area of the VRAM 50.
- the 3D image data read from the VRAM 50 is encoded by the video encoder 28 and is output to the stereoscopic display liquid crystal monitor 30 provided on the back of the camera, whereby the 3D subject image is displayed on the display screen of the liquid crystal monitor 30. Is displayed.
- the liquid crystal monitor 30 is a stereoscopic display device that can display a stereoscopic image (left viewpoint image and right viewpoint image) as a directional image having a predetermined directivity by a parallax barrier.
- the left viewpoint image and the right viewpoint image may be viewed separately by using a lens, or by wearing dedicated glasses such as polarized glasses or liquid crystal shutter glasses.
- the phase difference CCD 16 starts an AF (Automatic Focus) operation and an AE (Automatic Exposure) operation, via the lens driving unit 36. Then, control is performed so that the focus lens in the photographing optical system 12 comes to the in-focus position.
- the image data output from the A / D converter 20 when the shutter button is half-pressed is taken into the AE detection unit 44.
- the AE detection unit 44 integrates the G signals of the entire screen or integrates the G signals that are weighted differently in the central portion and the peripheral portion of the screen, and outputs the integrated value to the CPU 40.
- the CPU 40 calculates the brightness of the subject (shooting Ev value) from the integrated value input from the AE detection unit 44, and based on this shooting Ev value, calculates the aperture value of the aperture 14 and the electronic shutter (shutter speed) of the phase difference CCD 16. It is determined according to a predetermined program diagram, the aperture 14 is controlled via the aperture drive unit 34 based on the determined aperture value, and the phase difference CCD 16 via the CCD control unit 36 based on the determined shutter speed. To control the charge accumulation time.
- the AF processing unit 42 is a part that performs contrast AF processing or phase AF processing.
- contrast AF processing by extracting a high frequency component of image data in a predetermined focus area from at least one of left viewpoint image data and right viewpoint image data, and integrating the high frequency component
- An AF evaluation value indicating the in-focus state is calculated.
- AF control is performed by controlling the focus lens in the photographic optical system 12 so that the AF evaluation value is maximized.
- the phase difference AF process the phase difference between the image data corresponding to the main pixel and the sub pixel in the predetermined focus area in the left viewpoint image data and the right viewpoint image data is detected, and this phase difference is detected.
- the defocus amount is obtained based on the information indicating.
- AF control is performed by controlling the focus lens in the photographing optical system 12 so that the defocus amount becomes zero.
- Image data for two images is input from an image input controller 22 to a memory (SDRAM: Synchronous Random Access Memory) 48 and temporarily stored.
- SDRAM Synchronous Random Access Memory
- the two pieces of image data temporarily stored in the memory 48 are appropriately read out by the digital signal processing unit 24, where predetermined signals including generation processing (YC processing) of luminance data and color difference data of the image data are performed. Processing is performed.
- the YC processed image data (YC data) is stored in the memory 48 again. Subsequently, two pieces of YC data are respectively output to the compression / decompression processing unit 26, and after a predetermined compression process such as JPEG (Joint Photographic Experts Group) is executed, it is stored in the memory 48 again.
- a multi-picture file (MP file: a file in a format in which a plurality of images are connected) is generated from two pieces of YC data (compressed data) stored in the memory 48, and the MP file is generated by the media controller 52. It is read and recorded in the memory card 54.
- MP file a file in a format in which a plurality of images are connected
- 5D to 5F are graphs showing examples of aberrations of the photographing optical system.
- the horizontal axis indicates the pupil position, and the vertical axis indicates the aberration.
- the photographing optical system has aberrations such as spherical aberration, chromatic aberration, coma aberration, astigmatism, and field curvature.
- the stereoscopic imaging apparatus 10 acquires image data of main pixels and sub-pixels from a subject image (subject image that passes through different regions in the left-right direction of the photographing optical system) divided in the left-right direction.
- Directional aberration affects the image quality of image data of the main pixel and subpixel.
- FIG. 5A to FIG. 5C show examples of point images of main pixels and sub-pixels obtained from the stereoscopic imaging device having the photographing optical system shown in FIG. 5D to FIG. 5F, respectively.
- the aberration does not affect the peak and the inclination of the point image of the main pixel and sub-pixel, but the aberration as shown in FIG. 5D.
- the aberration greatly affects the peak and spread (PSF: Point-Spread-Function) of the point image of the main pixel and sub-pixel.
- aberrations and the like of the imaging optical system 12 are inspected before product shipment, and separate correction parameters for sharpness correction are created for the image data of the main pixel and subpixel, and are written in the ROM 46. Leave it in.
- the digital signal processing unit 24 reads out different correction parameters from the ROM 46, and corrects the image data of the main pixel and the sub-pixel with the corresponding correction parameters. Adjust the balance of both images.
- sharpness correction for example, a convolution filter of a predetermined size (3 ⁇ 3, 5 ⁇ 5) can be used, and the strength of sharpness correction can be adjusted by appropriately setting the filter coefficient (correction parameter). can do. Also, by providing correction parameters for each area within the shooting angle of view, it is possible to correct the sharpness of the main image and the sub-image for each area.
- FIG. 6 is a flowchart showing the photographing operation of the stereoscopic imaging apparatus 10 according to the first embodiment of the present invention.
- step S10 When the shutter button is fully pressed, the CPU 40 performs exposure for main photographing (step S10).
- the main image and the sub image are read from the phase difference CCD 16 and temporarily held in the memory 48 (step S12).
- the main image and the sub image held in the memory 48 are subjected to various signal processing such as white balance correction, gamma correction processing, synchronization processing, YC processing, and the like by the digital signal processing unit 24. In FIG. Is omitted.
- the digital signal processing unit 24 reads out different correction parameters for sharpness correction corresponding to the main image and the sub-image from the ROM 46 (step S14). Then, sharpness correction is performed on the main image and the sub-pixel using different correction parameters (step S16). The sharpness correction is normally performed on the Y signal (luminance signal) generated by the YC process.
- the main image and the sub image corrected as described above are subjected to sharpness correction by separate correction parameters, and as shown in FIG. 7B, compared to FIG.
- the peaks of the corresponding point images can be aligned, and the balance of the left and right viewpoint images is improved.
- the main and sub-images (YC signal) subjected to image processing as described above are each compressed and stored in an MP file, and the MP file is recorded on the memory card 54 via the media controller 52 ( Step S18).
- FIG. 8 is a flowchart showing the shooting operation of the stereoscopic imaging apparatus 10 according to the second embodiment of the present invention.
- the same step number is attached
- the second embodiment is different from the first embodiment in that the processes of step S20 and step S22 are added.
- Step S20 calculates the amount of parallax between corresponding points of the main image and the sub-image acquired in step S12.
- the calculation of the amount of parallax is based on one image (for example, main image), and the corresponding pixel of the other image (sub-image) is obtained.
- a method for obtaining the corresponding pixel for example, a block matching method can be applied.
- a parallax map indicating the parallax for one screen is created by obtaining the parallax with the corresponding pixel on the sub-image.
- sharpness correction is performed in an area where there is no parallax (or a small area) as shown in FIG. 9A.
- sharpness correction can be weakly applied in a region with a large parallax (blurred region).
- FIG. 10 is a flowchart showing the shooting operation of the stereoscopic imaging apparatus 10 according to the third embodiment of the present invention.
- the same step number is attached
- the third embodiment is different from the first embodiment in that the processing of step S30 and step S32 is performed instead of step S14 and step S16 of FIG.
- step S30 separate correction parameters for correcting chromatic aberration corresponding to the main image and the sub-image are read from the ROM 46.
- step S32 chromatic aberration correction is performed on the R, G, and B signals of the main image and the R, G, and B signals of the sub-pixels using the read color-specific correction parameters and the main and sub-image correction parameters.
- FIG. 11 shows an example of the difference in chromatic aberration around the angle of view between the main image and the sub-image.
- the third embodiment it is possible to individually correct chromatic aberration around the angle of view of the main image and the sub-image, and it is possible to obtain the main image and the sub-image without any color shift.
- FIG. 12 is a flowchart showing the photographing operation of the stereoscopic imaging apparatus 10 according to the fourth embodiment of the present invention.
- the same step number is attached
- the third embodiment is different from the first embodiment in that steps S40 and S42 are performed instead of steps S14 and S16 in FIG.
- step S40 correction parameters for correcting different distortion (distortion aberration) corresponding to the main image and the sub-image are read from the ROM 46.
- step S42 distortion correction is performed on the main image and sub-pixels using the read-out separate correction parameters.
- FIG. 13 is a flowchart showing the photographing operation of the stereoscopic imaging apparatus 10 according to the fifth embodiment of the present invention.
- the same step number is attached
- image processing is switched between when a 2D image is captured and when a 3D image is captured.
- the stereoscopic imaging apparatus 10 performs different processes as described below depending on whether the 2D shooting mode is selected or the 3D shooting mode is selected. The same processing is performed until the main image and the sub image are read from the phase difference CCD 16 up to steps S10 and S12.
- step S50 it is determined whether or not the camera is in 3D shooting mode (2D shooting mode). If the camera is in 3D shooting mode ("YES”), the process proceeds to step S52, and if it is in 2D shooting mode ("NO"). )), The process proceeds to step S58.
- the processing for the main image and the sub image in the 3D shooting mode reads out correction parameters for the main image and the sub image from the ROM 46 (steps).
- step S54 image processing of the main image and the sub-image is performed using the read different correction parameters (step S54). Then, the corrected main image and sub-image are stored in the MP file and recorded on the memory card 54 (step S56).
- the main image and the sub-image can be balanced regardless of the difference in the optical characteristics of the photographing optical system 12 in the left-right direction.
- the correction parameters for 2D images which are different correction parameters for the main image and the sub-image and different from the correction parameters for the 3D image, are read from the ROM 46 (step S58).
- the features of the correction parameters for 2D images will be described later.
- step S60 the main image and the sub image are corrected using the read-out separate correction parameters
- step S62 corresponding pixels of the corrected main image and sub image are added to generate a 2D image
- the generated 2D image is stored in an image file (Exif file) for a digital camera and recorded on the memory card 54 (step S64).
- FIG. 14A before correction
- FIG. 14B after correction
- the main image and the sub-image are corrected so that their point images are equal.
- FIG. 14C shows the point images of the main image and the sub image corrected by the correction parameter for 2D image, and even if these point images are added, there is no 2-line blur.
- substantially the same correction parameters can be used for the correction parameters for the main image and the sub-image of 2D image shooting. Therefore, the main image and the sub-image may be added before correction, and the added image may be corrected using the correction parameters for 2D images.
- FIG. 15 is a flowchart showing the reproduction operation of the stereoscopic imaging apparatus 10 according to the sixth embodiment of the present invention.
- the MP file storing the 3D image or the Exif file storing the 2D image as described above. Is recorded in the memory card 54.
- the stereoscopic imaging device 10 is configured not to perform image processing such as sharpness correction on the 3D image or 2D image during recording on the memory card 54.
- step S70 when the stereoscopic imaging apparatus 10 is switched to the reproduction mode, the 3D image or 2D image recorded on the memory card 54 is read (step S70).
- step S72 it is determined whether or not the read image is a 3D image. This determination can be made from the MP file, the extension indicating the Exif file, or tag information indicating the type of image in the MP file.
- step S74 When the image read from the memory card 54 is determined to be a 3D image, different correction parameters are read from the ROM 46 for the main image and sub-image constituting the 3D image (step S74). The main image and the sub image are processed using the correction parameter (step S74). Then, the corrected main image and sub image are output to the liquid crystal monitor 30, an external 3D display, a 3D printer, or the like (step S78).
- the correction parameter for the 2D image is read from the ROM 46 (step S80), and image processing of the 2D image is performed using the read correction parameter. This is performed (step S82). Then, the corrected 2D image is output to the liquid crystal monitor 30, an external 3D display, a 3D printer, or the like (step S84).
- image processing such as sharpness correction is not performed on a 3D image or 2D image during recording on the memory card 54.
- the present invention is not limited to this. It may be given. In this case, only when the 3D image is reproduced, different image processing may be performed on the main image and the sub image to correct the imbalance between the two images.
- correction parameters for correcting the imbalance between the main image and the sub image are stored for each zoom position and aperture value. It is desirable to use correction parameters corresponding to the zoom position and aperture value in accordance with the zoom position and aperture value during 3D shooting.
- the stereoscopic imaging apparatus 10 of this embodiment uses one phase difference CCD 16, the apparatus can be reduced in size as compared with the apparatus using the two imaging elements 7 and 8 shown in FIG.
- the present invention is not limited to the one using one image sensor, but can be applied to the one having the conventional optical system and the image sensor shown in FIG.
- the image sensor is not limited to the CCD sensor of this embodiment, and may be an image sensor such as a CMOS (Complementary Metal-Oxide Semiconductor) sensor.
- CMOS Complementary Metal-Oxide Semiconductor
- the main image and the sub image showing the subject image divided in the left-right direction are obtained.
- the number of subject images divided in the pupil is not limited to two.
- the dividing direction is not limited to the horizontal direction, and pupil division may be performed in the vertical and horizontal directions.
- SYMBOLS 10 ... Stereo imaging device, 12 ... Imaging optical system, 14 ... Diaphragm, 16 ... Imaging device (phase difference CCD), 30 ... Liquid crystal monitor, 32 ... CCD control part, 34 ... Diaphragm drive part, 36 ... Lens drive part, 38 ... Operating section, 40 ... Central processing unit (CPU), 42 ... AF processing section, 44 ... AE detecting section, 46 ... ROM
Landscapes
- Engineering & Computer Science (AREA)
- Multimedia (AREA)
- Signal Processing (AREA)
- Physics & Mathematics (AREA)
- General Physics & Mathematics (AREA)
- Testing, Inspecting, Measuring Of Stereoscopic Televisions And Televisions (AREA)
- Stereoscopic And Panoramic Photography (AREA)
- Studio Devices (AREA)
- Camera Data Copying Or Recording (AREA)
Abstract
Description
図1は本発明に係る立体撮像装置10の実施の形態を示すブロック図である。
図2Aから図2Cは位相差CCD16の構成例を示す図である。
次に、位相差CCD16から読み出された主画素、副画素の画像データに対するデジタル信号処理部24での信号処理(画像処理)について説明する。
図8は本発明の第2の実施形態の立体撮像装置10の撮影動作を示すフローチャートである。尚、図6に示した第1の実施形態と共通する部分には同一のステップ番号を付し、その詳細な説明は省略する。
図10は本発明の第3の実施形態の立体撮像装置10の撮影動作を示すフローチャートである。尚、図6に示した第1の実施形態と共通する部分には同一のステップ番号を付し、その詳細な説明は省略する。
図12は本発明の第4の実施形態の立体撮像装置10の撮影動作を示すフローチャートである。尚、図6に示した第1の実施形態と共通する部分には同一のステップ番号を付し、その詳細な説明は省略する。
図13は本発明の第5の実施形態の立体撮像装置10の撮影動作を示すフローチャートである。尚、図6に示した第1の実施形態と共通する部分には同一のステップ番号を付し、その詳細な説明は省略する。
図15は本発明の第6の実施形態の立体撮像装置10の再生動作を示すフローチャートである。
撮影光学系12の左右方向の光学特性はズーム位置により異なるため、主画像と副画像のアンバランスを補正するための補正パラメータを各ズーム位置毎に記憶保持し、3D撮影時のズーム位置(ズーム倍率)に応じて、そのズーム位置に対応する補正パラメータを使用する。
Claims (11)
- 単一の撮影光学系と、
前記撮影光学系の予め定められた方向の異なる第1、第2の領域を通過した被写体像が瞳分割されてそれぞれ結像される撮像素子であって、前記第1、第2の領域を通過した被写体像をそれぞれ光電変換して第1の画像及び第2の画像を出力する撮像素子と、
前記第1の画像に対して第1の画像処理を行うとともに、前記第2の画像に対して前記第1の画像処理と異なる第2の画像処理を行う画像処理部であって、各処理後の第1の画像と第2の画像との画質の差を低減するように前記第1の画像処理及び第2の画像処理を行う画像処理部と、
を備える立体撮像装置。 - 前記画像処理部は、前記第1の画像及び第2の画像に対してそれぞれ異なるシャープネス補正を行う、請求項1に記載の立体撮像装置。
- 前記第1の画像及び第2の画像に基づいて視差量を算出する視差量算出部を備え、
前記画像処理部は、前記算出した視差量の大きな範囲の画像に対して前記シャープネス補正を弱くする、請求項2に記載の立体撮像装置。 - 前記画像処理部は、前記第1の画像及び第2の画像に対してそれぞれ異なる色収差補正を行う、請求項1から3のいずれかに記載の立体撮像装置。
- 前記画像処理部は、前記第1の画像及び第2の画像に対してそれぞれ異なるディストーション補正を行う、請求項1から4のいずれかに記載の立体撮像装置。
- 前記第1の画像及び第2の画像の画像処理に使用する第1の補正パラメータ及び第2の補正パラメータを予め記憶する記憶部を備え、
前記画像処理部は、前記記憶された第1の補正パラメータ及び第2の補正パラメータに基づいてそれぞれ前記第1の画像及び第2の画像の画像処理を行う、請求項1から5のいずれかに記載の立体撮像装置。 - 前記撮影光学系はズームレンズであり、
前記記憶部には、前記ズームレンズのズーム位置に対応して、又はズーム位置と絞り値に対応してそれぞれ前記第1の補正パラメータ及び第2の補正パラメータが記憶され、
前記画像処理部は、撮影時のズーム位置、又はズーム位置と絞り値に基づいて前記記憶部から対応する第1の補正パラメータ及び第2の補正パラメータを読み出し、読み出した第1の補正パラメータ及び第2の補正パラメータに基づいてそれぞれ前記第1の画像及び第2の画像の画像処理を行う、請求項6に記載の立体撮像装置。 - 前記画像処理部により処理された前記第1の画像及び第2の画像からなる立体視画像を記録媒体に記録する記録部を備える請求項1から7のいずれかに記載の立体撮像装置。
- 前記画像処理部は、前記第1の画像及び第2の画像を加算した平面画像の撮影時には、前記第1の画像処理及び第2の画像処理とは異なる画像処理を行う、請求項1から8のいずれかに記載の立体撮像装置。
- 前記画像処理部による処理前の前記第1の画像及び第2の画像とともに、立体画像を示す情報と平面画像を示す情報を関連付けて記録媒体に記録する記録部を備え、
前記画像処理部は、立体視画像の出力時に前記記録媒体から読み出した第1の画像及び第2の画像に対して前記第1の画像処理及び第2の画像処理を行い、平面画像の出力時に前記第1の画像処理及び第2の画像処理とは異なる画像処理を行う、請求項1から7のいずれかに記載の立体撮像装置。 - 前記撮像素子は、それぞれマトリクス状に配列された光電変換用の第1群の画素及び第2群の画素であって、前記撮影光学系の第1の領域を通過した被写体像のみを受光するように光束の受光方向の制限を受けた第1群の画素と、前記撮影光学系の第2の領域を通過した被写体像のみを受光するように光束の受光方向の制限を受けた第2群の画素とを有し、前記第1群の画素及び第2群の画素から前記第1の画像及び第2の画像の読み出しが可能な撮像素子である、請求項1から10のいずれかに記載の立体撮像装置。
Priority Applications (3)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US13/636,608 US8520059B2 (en) | 2010-03-24 | 2010-12-01 | Stereoscopic image taking apparatus |
CN201080065711.7A CN102812714B (zh) | 2010-03-24 | 2010-12-01 | 立体图像获取装置 |
JP2012506771A JP5192096B2 (ja) | 2010-03-24 | 2010-12-01 | 立体撮像装置 |
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
JP2010-068240 | 2010-03-24 | ||
JP2010068240 | 2010-03-24 |
Publications (1)
Publication Number | Publication Date |
---|---|
WO2011118089A1 true WO2011118089A1 (ja) | 2011-09-29 |
Family
ID=44672673
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
PCT/JP2010/071473 WO2011118089A1 (ja) | 2010-03-24 | 2010-12-01 | 立体撮像装置 |
Country Status (4)
Country | Link |
---|---|
US (1) | US8520059B2 (ja) |
JP (1) | JP5192096B2 (ja) |
CN (1) | CN102812714B (ja) |
WO (1) | WO2011118089A1 (ja) |
Cited By (21)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP2013080025A (ja) * | 2011-10-03 | 2013-05-02 | Sony Corp | 撮像装置および映像記録再生システム |
JP2013118464A (ja) * | 2011-12-02 | 2013-06-13 | Canon Inc | 撮像装置およびその制御方法 |
WO2013100039A1 (ja) * | 2011-12-27 | 2013-07-04 | 富士フイルム株式会社 | カラー撮像素子及び撮像装置 |
JP2013187878A (ja) * | 2012-03-12 | 2013-09-19 | Canon Inc | 撮像装置および撮像装置の制御方法 |
WO2013146067A1 (ja) * | 2012-03-30 | 2013-10-03 | 富士フイルム株式会社 | 画像処理装置、撮像装置、画像処理方法、記録媒体及びプログラム |
WO2013146996A1 (ja) | 2012-03-29 | 2013-10-03 | 富士フイルム株式会社 | 画像処理装置、撮像装置および画像処理方法 |
JP2013207678A (ja) * | 2012-03-29 | 2013-10-07 | Canon Inc | 撮像装置およびその制御方法 |
WO2013168667A1 (ja) * | 2012-05-09 | 2013-11-14 | 富士フイルム株式会社 | 画像処理装置及び方法並びに撮像装置 |
JP2013236207A (ja) * | 2012-05-08 | 2013-11-21 | Nikon Corp | 撮像装置、撮像方法、およびプログラム |
JP2013243602A (ja) * | 2012-05-22 | 2013-12-05 | Canon Inc | 撮像装置およびその制御方法 |
WO2014024500A1 (ja) * | 2012-08-10 | 2014-02-13 | 株式会社ニコン | 画像処理方法、画像処理装置、撮像装置および画像処理プログラム |
JP2014138279A (ja) * | 2013-01-17 | 2014-07-28 | Canon Inc | 撮像装置および撮像装置の制御方法 |
JP2014146962A (ja) * | 2013-01-29 | 2014-08-14 | Canon Inc | 撮像装置およびその制御方法 |
CN104205827A (zh) * | 2012-03-30 | 2014-12-10 | 富士胶片株式会社 | 图像处理装置及方法、以及摄像装置 |
JP2016106466A (ja) * | 2015-12-25 | 2016-06-16 | キヤノン株式会社 | 撮像装置およびその制御方法、並びに画像処理装置およびその制御方法 |
JP2016154360A (ja) * | 2016-03-30 | 2016-08-25 | キヤノン株式会社 | 撮像装置およびその制御方法 |
US9521395B2 (en) | 2011-10-04 | 2016-12-13 | Canon Kabushiki Kaisha | Imaging apparatus and method for controlling same |
JP2016219985A (ja) * | 2015-05-19 | 2016-12-22 | キヤノン株式会社 | 画像処理装置、撮像装置、画像処理方法および画像処理プログラム |
JP2017077005A (ja) * | 2016-12-01 | 2017-04-20 | キヤノン株式会社 | 撮像装置およびその制御方法 |
JP2017229075A (ja) * | 2017-07-13 | 2017-12-28 | キヤノン株式会社 | 画像処理装置およびその制御方法、撮像装置 |
JP2018007111A (ja) * | 2016-07-05 | 2018-01-11 | 株式会社ニコン | 撮像装置、及び、画像処理装置 |
Families Citing this family (14)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP2011182075A (ja) * | 2010-02-26 | 2011-09-15 | Hitachi Consumer Electronics Co Ltd | 画像処理装置 |
CN103796003B (zh) * | 2014-01-21 | 2016-03-02 | 深圳市掌网立体时代视讯技术有限公司 | 一种立体摄像的图像修正方法及系统 |
US9729806B2 (en) * | 2015-01-06 | 2017-08-08 | Semiconductor Components Industries, Llc | Imaging systems with phase detection pixels |
CN104574404B (zh) * | 2015-01-14 | 2017-09-19 | 宁波大学 | 一种立体图像重定位方法 |
CN104754221B (zh) * | 2015-03-11 | 2018-11-16 | 广东欧珀移动通信有限公司 | 一种拍照方法及移动终端 |
CN104754229B (zh) * | 2015-03-27 | 2019-01-15 | 联想(北京)有限公司 | 一种图像处理方法及电子设备 |
CN104811684B (zh) * | 2015-04-02 | 2017-06-16 | 广东欧珀移动通信有限公司 | 一种图像的三维美颜方法及装置 |
JP6755737B2 (ja) * | 2016-07-27 | 2020-09-16 | キヤノン株式会社 | 距離測定装置、撮像装置、および距離測定方法 |
CN107124548A (zh) * | 2017-04-25 | 2017-09-01 | 深圳市金立通信设备有限公司 | 一种拍照方法及终端 |
JP2018189443A (ja) | 2017-04-28 | 2018-11-29 | キヤノン株式会社 | 距離測定装置、距離測定方法及び撮像装置 |
KR20190012812A (ko) * | 2017-07-28 | 2019-02-11 | 에스케이하이닉스 주식회사 | 위상차 검출 픽셀을 구비한 이미지 센서 |
JP7027064B2 (ja) | 2017-08-10 | 2022-03-01 | キヤノン株式会社 | 画像処理装置および画像処理方法、撮像装置、プログラム |
GB2608496B (en) | 2021-05-07 | 2024-04-24 | Canon Kk | Image processing apparatus and method, and image capturing apparatus and control method thereof, program, and storage medium |
CN115118881A (zh) * | 2022-06-24 | 2022-09-27 | 维沃移动通信有限公司 | 信号处理电路、图像传感器、电子设备及图像处理方法 |
Citations (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JPH0735989A (ja) * | 1993-07-26 | 1995-02-07 | Olympus Optical Co Ltd | 立体視内視鏡 |
JP2004207985A (ja) * | 2002-12-25 | 2004-07-22 | Fuji Photo Film Co Ltd | デジタルカメラ |
JP2009527007A (ja) * | 2006-02-13 | 2009-07-23 | スリーエム イノベイティブ プロパティズ カンパニー | 単眼三次元撮像 |
JP2009168995A (ja) * | 2008-01-15 | 2009-07-30 | Nikon Corp | 測距装置および撮像装置 |
Family Cites Families (8)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP4574022B2 (ja) * | 2001-01-17 | 2010-11-04 | キヤノン株式会社 | 撮像装置及びシェーディング補正方法 |
US7596286B2 (en) * | 2003-08-06 | 2009-09-29 | Sony Corporation | Image processing apparatus, image processing system, imaging apparatus and image processing method |
WO2006068129A1 (ja) * | 2004-12-22 | 2006-06-29 | Matsushita Electric Industrial Co., Ltd. | 撮像装置およびその製造方法 |
KR100739764B1 (ko) * | 2005-11-28 | 2007-07-13 | 삼성전자주식회사 | 입체 영상 신호 처리 장치 및 방법 |
US7372642B2 (en) | 2006-02-13 | 2008-05-13 | 3M Innovative Properties Company | Three-channel camera systems with non-collinear apertures |
US7646550B2 (en) | 2006-02-13 | 2010-01-12 | 3M Innovative Properties Company | Three-channel camera systems with collinear apertures |
US7819591B2 (en) | 2006-02-13 | 2010-10-26 | 3M Innovative Properties Company | Monocular three-dimensional imaging |
JP5034429B2 (ja) * | 2006-10-16 | 2012-09-26 | ソニー株式会社 | レンズ装置、撮像装置及び収差補正方法 |
-
2010
- 2010-12-01 CN CN201080065711.7A patent/CN102812714B/zh not_active Expired - Fee Related
- 2010-12-01 US US13/636,608 patent/US8520059B2/en active Active
- 2010-12-01 WO PCT/JP2010/071473 patent/WO2011118089A1/ja active Application Filing
- 2010-12-01 JP JP2012506771A patent/JP5192096B2/ja not_active Expired - Fee Related
Patent Citations (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JPH0735989A (ja) * | 1993-07-26 | 1995-02-07 | Olympus Optical Co Ltd | 立体視内視鏡 |
JP2004207985A (ja) * | 2002-12-25 | 2004-07-22 | Fuji Photo Film Co Ltd | デジタルカメラ |
JP2009527007A (ja) * | 2006-02-13 | 2009-07-23 | スリーエム イノベイティブ プロパティズ カンパニー | 単眼三次元撮像 |
JP2009168995A (ja) * | 2008-01-15 | 2009-07-30 | Nikon Corp | 測距装置および撮像装置 |
Cited By (44)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US10070114B2 (en) | 2011-10-03 | 2018-09-04 | Sony Corporation | Splitting light to generate parallax image |
JP2013080025A (ja) * | 2011-10-03 | 2013-05-02 | Sony Corp | 撮像装置および映像記録再生システム |
US9924155B2 (en) | 2011-10-04 | 2018-03-20 | Canon Kabushiki Kaisha | Imaging apparatus and method for controlling same |
US9521395B2 (en) | 2011-10-04 | 2016-12-13 | Canon Kabushiki Kaisha | Imaging apparatus and method for controlling same |
US10587860B2 (en) | 2011-10-04 | 2020-03-10 | Canon Kabushiki Kaisha | Imaging apparatus and method for controlling same |
JP2013118464A (ja) * | 2011-12-02 | 2013-06-13 | Canon Inc | 撮像装置およびその制御方法 |
WO2013100039A1 (ja) * | 2011-12-27 | 2013-07-04 | 富士フイルム株式会社 | カラー撮像素子及び撮像装置 |
JPWO2013100039A1 (ja) * | 2011-12-27 | 2015-05-11 | 富士フイルム株式会社 | カラー撮像素子及び撮像装置 |
JP5621059B2 (ja) * | 2011-12-27 | 2014-11-05 | 富士フイルム株式会社 | カラー撮像素子及び撮像装置 |
US9100558B2 (en) | 2011-12-27 | 2015-08-04 | Fujifilm Corporation | Color imaging element and imaging apparatus |
JP2013187878A (ja) * | 2012-03-12 | 2013-09-19 | Canon Inc | 撮像装置および撮像装置の制御方法 |
JP2013207678A (ja) * | 2012-03-29 | 2013-10-07 | Canon Inc | 撮像装置およびその制御方法 |
EP2833638A1 (en) * | 2012-03-29 | 2015-02-04 | Fujifilm Corporation | Image processing device, imaging device, and image processing method |
US9167224B2 (en) | 2012-03-29 | 2015-10-20 | Fujifilm Corporation | Image processing device, imaging device, and image processing method |
EP2833638A4 (en) * | 2012-03-29 | 2015-11-25 | Fujifilm Corp | IMAGE PROCESSING DEVICE, IMAGING APPARATUS AND IMAGE PROCESSING METHOD |
CN104221370A (zh) * | 2012-03-29 | 2014-12-17 | 富士胶片株式会社 | 图像处理装置、摄像装置以及图像处理方法 |
WO2013146996A1 (ja) | 2012-03-29 | 2013-10-03 | 富士フイルム株式会社 | 画像処理装置、撮像装置および画像処理方法 |
JP5655174B2 (ja) * | 2012-03-29 | 2015-01-14 | 富士フイルム株式会社 | 画像処理装置、撮像装置および画像処理方法 |
CN104205827A (zh) * | 2012-03-30 | 2014-12-10 | 富士胶片株式会社 | 图像处理装置及方法、以及摄像装置 |
US9609302B2 (en) | 2012-03-30 | 2017-03-28 | Fujifilm Corporation | Image processing device, imaging device, image processing method, and recording medium |
JPWO2013146067A1 (ja) * | 2012-03-30 | 2015-12-10 | 富士フイルム株式会社 | 画像処理装置、撮像装置、画像処理方法、記録媒体及びプログラム |
US9277201B2 (en) | 2012-03-30 | 2016-03-01 | Fujifilm Corporation | Image processing device and method, and imaging device |
CN104205827B (zh) * | 2012-03-30 | 2016-03-16 | 富士胶片株式会社 | 图像处理装置及方法、以及摄像装置 |
WO2013146067A1 (ja) * | 2012-03-30 | 2013-10-03 | 富士フイルム株式会社 | 画像処理装置、撮像装置、画像処理方法、記録媒体及びプログラム |
JP2013236207A (ja) * | 2012-05-08 | 2013-11-21 | Nikon Corp | 撮像装置、撮像方法、およびプログラム |
JP5687803B2 (ja) * | 2012-05-09 | 2015-03-25 | 富士フイルム株式会社 | 画像処理装置及び方法並びに撮像装置 |
CN104272732A (zh) * | 2012-05-09 | 2015-01-07 | 富士胶片株式会社 | 图像处理装置、方法及摄像装置 |
WO2013168667A1 (ja) * | 2012-05-09 | 2013-11-14 | 富士フイルム株式会社 | 画像処理装置及び方法並びに撮像装置 |
US9288472B2 (en) | 2012-05-09 | 2016-03-15 | Fujifilm Corporation | Image processing device and method, and image capturing device |
CN104272732B (zh) * | 2012-05-09 | 2016-06-01 | 富士胶片株式会社 | 图像处理装置、方法及摄像装置 |
JP2013243602A (ja) * | 2012-05-22 | 2013-12-05 | Canon Inc | 撮像装置およびその制御方法 |
US9509978B2 (en) | 2012-08-10 | 2016-11-29 | Nikon Corporation | Image processing method, image processing apparatus, image-capturing apparatus, and image processing program |
JP5904281B2 (ja) * | 2012-08-10 | 2016-04-13 | 株式会社ニコン | 画像処理方法、画像処理装置、撮像装置および画像処理プログラム |
CN104429056A (zh) * | 2012-08-10 | 2015-03-18 | 株式会社尼康 | 图像处理方法、图像处理装置、摄像装置及图像处理程序 |
CN104429056B (zh) * | 2012-08-10 | 2017-11-14 | 株式会社尼康 | 图像处理方法、图像处理装置、摄像装置及图像处理程序 |
WO2014024500A1 (ja) * | 2012-08-10 | 2014-02-13 | 株式会社ニコン | 画像処理方法、画像処理装置、撮像装置および画像処理プログラム |
JP2014138279A (ja) * | 2013-01-17 | 2014-07-28 | Canon Inc | 撮像装置および撮像装置の制御方法 |
JP2014146962A (ja) * | 2013-01-29 | 2014-08-14 | Canon Inc | 撮像装置およびその制御方法 |
JP2016219985A (ja) * | 2015-05-19 | 2016-12-22 | キヤノン株式会社 | 画像処理装置、撮像装置、画像処理方法および画像処理プログラム |
JP2016106466A (ja) * | 2015-12-25 | 2016-06-16 | キヤノン株式会社 | 撮像装置およびその制御方法、並びに画像処理装置およびその制御方法 |
JP2016154360A (ja) * | 2016-03-30 | 2016-08-25 | キヤノン株式会社 | 撮像装置およびその制御方法 |
JP2018007111A (ja) * | 2016-07-05 | 2018-01-11 | 株式会社ニコン | 撮像装置、及び、画像処理装置 |
JP2017077005A (ja) * | 2016-12-01 | 2017-04-20 | キヤノン株式会社 | 撮像装置およびその制御方法 |
JP2017229075A (ja) * | 2017-07-13 | 2017-12-28 | キヤノン株式会社 | 画像処理装置およびその制御方法、撮像装置 |
Also Published As
Publication number | Publication date |
---|---|
JP5192096B2 (ja) | 2013-05-08 |
JPWO2011118089A1 (ja) | 2013-07-04 |
US20130010078A1 (en) | 2013-01-10 |
US8520059B2 (en) | 2013-08-27 |
CN102812714B (zh) | 2014-04-30 |
CN102812714A (zh) | 2012-12-05 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
JP5192096B2 (ja) | 立体撮像装置 | |
JP5180407B2 (ja) | 立体撮像装置および視差画像復元方法 | |
JP5640143B2 (ja) | 撮像装置及び撮像方法 | |
JP5421829B2 (ja) | 撮像装置 | |
JP5690396B2 (ja) | 撮像装置およびシェーディング補正方法 | |
JP5368350B2 (ja) | 立体撮像装置 | |
JP5502205B2 (ja) | 立体撮像装置および立体撮像方法 | |
JP5243666B2 (ja) | 撮像装置、撮像装置本体およびシェーディング補正方法 | |
JP5871989B2 (ja) | 撮影装置、撮影方法及びプログラム | |
JP5284306B2 (ja) | 立体撮像装置、ゴースト像処理装置およびゴースト像処理方法 | |
WO2012036019A1 (ja) | 単眼立体撮像装置、単眼立体撮像装置用シェーディング補正方法及び単眼立体撮像装置用プログラム | |
US9124875B2 (en) | Stereoscopic imaging apparatus | |
WO2013069445A1 (ja) | 立体撮像装置及び画像処理方法 | |
JP5628914B2 (ja) | 撮像装置および撮像方法 | |
JP5595505B2 (ja) | 立体撮像装置および立体撮像方法 | |
JP5457240B2 (ja) | 立体撮像装置 | |
JP2012124650A (ja) | 撮像装置および撮像方法 | |
JP2014063190A (ja) | 撮像装置 | |
JP5649837B2 (ja) | 立体撮像装置 |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
WWE | Wipo information: entry into national phase |
Ref document number: 201080065711.7 Country of ref document: CN |
|
121 | Ep: the epo has been informed by wipo that ep was designated in this application |
Ref document number: 10848481 Country of ref document: EP Kind code of ref document: A1 |
|
WWE | Wipo information: entry into national phase |
Ref document number: 2012506771 Country of ref document: JP |
|
WWE | Wipo information: entry into national phase |
Ref document number: 13636608 Country of ref document: US |
|
NENP | Non-entry into the national phase |
Ref country code: DE |
|
122 | Ep: pct application non-entry in european phase |
Ref document number: 10848481 Country of ref document: EP Kind code of ref document: A1 |