WO2017047080A1 - 映像処理装置、撮影装置、映像処理方法及びプログラム記録媒体 - Google Patents
映像処理装置、撮影装置、映像処理方法及びプログラム記録媒体 Download PDFInfo
- Publication number
- WO2017047080A1 WO2017047080A1 PCT/JP2016/004195 JP2016004195W WO2017047080A1 WO 2017047080 A1 WO2017047080 A1 WO 2017047080A1 JP 2016004195 W JP2016004195 W JP 2016004195W WO 2017047080 A1 WO2017047080 A1 WO 2017047080A1
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- signal
- color
- infrared
- pattern
- video
- Prior art date
Links
- 238000012545 processing Methods 0.000 title claims abstract description 87
- 238000003384 imaging method Methods 0.000 title claims abstract description 19
- 238000003672 processing method Methods 0.000 title claims description 4
- 230000005540 biological transmission Effects 0.000 claims description 36
- 238000000034 method Methods 0.000 claims description 31
- 230000003287 optical effect Effects 0.000 claims description 30
- 238000010586 diagram Methods 0.000 description 38
- 238000004364 calculation method Methods 0.000 description 19
- 230000003595 spectral effect Effects 0.000 description 16
- 239000003086 colorant Substances 0.000 description 10
- 230000006870 function Effects 0.000 description 10
- 230000035945 sensitivity Effects 0.000 description 10
- 230000014509 gene expression Effects 0.000 description 9
- 238000011156 evaluation Methods 0.000 description 7
- 239000011159 matrix material Substances 0.000 description 6
- 238000012986 modification Methods 0.000 description 6
- 230000004048 modification Effects 0.000 description 6
- 230000004456 color vision Effects 0.000 description 4
- 238000002834 transmittance Methods 0.000 description 4
- 238000012937 correction Methods 0.000 description 3
- 241000282412 Homo Species 0.000 description 2
- 230000006399 behavior Effects 0.000 description 2
- XUIMIQQOPSSXEZ-UHFFFAOYSA-N Silicon Chemical compound [Si] XUIMIQQOPSSXEZ-UHFFFAOYSA-N 0.000 description 1
- 230000002411 adverse Effects 0.000 description 1
- 230000000903 blocking effect Effects 0.000 description 1
- 230000000694 effects Effects 0.000 description 1
- 238000009434 installation Methods 0.000 description 1
- 230000003252 repetitive effect Effects 0.000 description 1
- 238000005070 sampling Methods 0.000 description 1
- 229910052710 silicon Inorganic materials 0.000 description 1
- 239000010703 silicon Substances 0.000 description 1
- 230000017105 transposition Effects 0.000 description 1
Images
Classifications
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N23/00—Cameras or camera modules comprising electronic image sensors; Control thereof
- H04N23/10—Cameras or camera modules comprising electronic image sensors; Control thereof for generating image signals from different wavelengths
- H04N23/11—Cameras or camera modules comprising electronic image sensors; Control thereof for generating image signals from different wavelengths for generating image signals from visible and infrared light wavelengths
-
- G—PHYSICS
- G02—OPTICS
- G02B—OPTICAL ELEMENTS, SYSTEMS OR APPARATUS
- G02B5/00—Optical elements other than lenses
- G02B5/20—Filters
- G02B5/208—Filters for use with infrared or ultraviolet radiation, e.g. for separating visible light from infrared and/or ultraviolet radiation
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N23/00—Cameras or camera modules comprising electronic image sensors; Control thereof
- H04N23/10—Cameras or camera modules comprising electronic image sensors; Control thereof for generating image signals from different wavelengths
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N23/00—Cameras or camera modules comprising electronic image sensors; Control thereof
- H04N23/10—Cameras or camera modules comprising electronic image sensors; Control thereof for generating image signals from different wavelengths
- H04N23/13—Cameras or camera modules comprising electronic image sensors; Control thereof for generating image signals from different wavelengths with multiple sensors
- H04N23/16—Optical arrangements associated therewith, e.g. for beam-splitting or for colour correction
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N23/00—Cameras or camera modules comprising electronic image sensors; Control thereof
- H04N23/45—Cameras or camera modules comprising electronic image sensors; Control thereof for generating image signals from two or more image sensors being of different type or operating in different modes, e.g. with a CMOS sensor for moving images in combination with a charge-coupled device [CCD] for still images
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N23/00—Cameras or camera modules comprising electronic image sensors; Control thereof
- H04N23/80—Camera processing pipelines; Components thereof
- H04N23/84—Camera processing pipelines; Components thereof for processing colour signals
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N25/00—Circuitry of solid-state image sensors [SSIS]; Control thereof
- H04N25/10—Circuitry of solid-state image sensors [SSIS]; Control thereof for transforming different wavelengths into image signals
- H04N25/11—Arrangement of colour filter arrays [CFA]; Filter mosaics
- H04N25/13—Arrangement of colour filter arrays [CFA]; Filter mosaics characterised by the spectral characteristics of the filter elements
- H04N25/131—Arrangement of colour filter arrays [CFA]; Filter mosaics characterised by the spectral characteristics of the filter elements including elements passing infrared wavelengths
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N23/00—Cameras or camera modules comprising electronic image sensors; Control thereof
- H04N23/50—Constructional details
- H04N23/55—Optical parts specially adapted for electronic image sensors; Mounting thereof
Definitions
- the present invention relates to video processing.
- a color image input device such as a digital still camera or a video camera generally has a configuration in which optical filters of three colors of red (R), green (G), and blue (B) are incorporated in an image sensor.
- the light incident on the color image input device is decomposed by the three-color optical filter and converted into R, G, and B color signals by the image sensor.
- the image sensor used in the color image input device is a silicon sensor
- the sensor sensitivity ranges from the visible region to the near infrared region.
- near-infrared light components may adversely affect color reproduction.
- the three-color optical filter guarantees a certain transmittance in the wavelength region corresponding to each color, the light transmission characteristics are not necessarily guaranteed in regions other than the visible region such as the near infrared region.
- FIG. 17 is a diagram illustrating the spectral transmittance of the RGB three-color optical filter.
- each color filter is expected to transmit light having wavelengths in the vicinity of 400 to 500 nm (B), 500 to 600 nm (G), and 600 to 700 nm (R).
- B 500 nm
- G 500 to 600 nm
- R 600 to 700 nm
- each color filter may also have a characteristic of transmitting light in a region other than the visible region.
- the spectral sensitivity characteristics of image sensors using photodiodes which are often used in image input devices, have sensitivity even in the wavelength region of 700 nm or more. If this is done, a problem may arise from the viewpoint of color reproducibility only by applying a three-color optical filter having spectral sensitivity characteristics as shown in FIG. 17 to a general image sensor. Therefore, the image sensor is provided with an infrared cut filter when high color reproducibility is required.
- FIG. 18 is a diagram showing the color matching functions of the XYZ color system relating to human color perception. As shown in FIG. 18, human color perception is insensitive to light having a wavelength of 700 nm or more. Therefore, light having power in a wavelength region of 700 nm or more does not affect the perceived color that is a psychophysical quantity.
- an infrared cut filter having a spectral transmittance for removing the influence of near-infrared light of 700 nm or more as shown in FIG. Is used. Specifically, as shown in FIG. 21, by providing an infrared cut filter 610 in the optical system of the color image input device, the incidence of near infrared light to the three-color optical filter 620 and the image sensor 630 is blocked. The By doing so, light having no power in the near infrared region can be incident on the three-color optical filter 620 and the image sensor 630.
- the simplest method of using near-infrared light for high-sensitivity shooting is to move the infrared cut filter set in the optical system mechanically during high-sensitivity shooting and temporarily move the infrared cut filter from the optical system. It is a method to remove.
- this method has not only a problem of an increase in the number of parts, that is, an increase in cost, but also a problem that the possibility of a failure is increased by requiring a mechanical operation for moving the infrared cut filter.
- Non-Patent Document 1 discloses a method of photographing without requiring a mechanical operation. Specifically, Non-Patent Document 1 describes a photographing method using two cameras for photographing a color image and a near-infrared image.
- Non-Patent Document 2 includes an image sensor 700 in which a four-color optical filter in which an IR (infrared) filter that transmits near-infrared light is added to an RGB three-color optical filter is incorporated. Disclosure.
- FIG. 2 of Non-Patent Document 2 describes spectral sensitivity characteristics of R, G, B, and IR optical filters. The spectral sensitivity characteristics of the R, G, and B optical filters have the same spectral sensitivity as that of the IR filter in the near infrared region. In order to realize high color reproducibility in daytime shooting, it is necessary to suppress or remove the influence of near-infrared light included in the R, G, and B color signals.
- Non-Patent Document 2 removes IR components included in R, G, and B color signals in daytime shooting, and IR signals obtained by passing through an IR filter in nighttime shooting. In addition, a black and white image is obtained using not only the IR components included in the R, G, and B color signals.
- Patent Document 1 uses R, G, and B three-color optical filters that transmit near-infrared (NIR) light, and uses a photosensor that detects near-infrared light, so that R, G , B, and NIR color imaging devices that generate signals are disclosed.
- This photosensor has a visible light sensor portion at a shallow position with respect to the incident direction of light, and has an invisible light sensor portion at a deep position with respect to the direction.
- Non-Patent Document 3 does not use an IR cut filter and uses a different color filter array such as two types of filters having different spectral transmission characteristics with respect to a G filter of an RGB Bayer type color filter array (CFA).
- CFA RGB Bayer type color filter array
- Non-Patent Document 1 generates an RGB image and a near-infrared image using two cameras. Note that the method described in Non-Patent Document 1 requires two optical paths and two image sensors for RGB and near-infrared images, although it is not impossible to configure with a single device. .
- Non-Patent Document 2 and Patent Document 1 are special image sensors for generating a near-infrared image, and can be said not to have a general configuration.
- the color filter array described in Non-Patent Document 3 requires two different G filters.
- An object of the present invention is to provide a technique that enables image processing of visible light and near-infrared light using an imaging device having a general configuration.
- the present invention uses acquisition means for acquiring an image signal representing an image including near-infrared light having an intensity corresponding to a pattern having a predetermined geometric shape, and pattern information defining the pattern. And a signal processing unit that outputs a color signal representing a visible light component corresponding to the acquired video signal and a near infrared signal representing a near infrared light component corresponding to the video signal. provide.
- the present invention includes an optical filter that transmits near-infrared light in a pattern having a predetermined geometric shape, and a video signal representing an image including near-infrared light that has passed through the optical filter.
- a near-infrared light representing a color signal representing a visible light component corresponding to the generated video signal and a near-infrared light component corresponding to the video signal by using a light receiving means to generate and pattern information defining the pattern
- a photographing apparatus including signal processing means for outputting a signal.
- the present invention obtains a video signal representing an image including near-infrared light having an intensity corresponding to a pattern having a predetermined geometric shape, and uses pattern information that defines the pattern,
- a video processing method for outputting a color signal representing a visible light component corresponding to the acquired video signal and a near infrared signal representing a near infrared light component corresponding to the video signal.
- a computer obtains a video signal representing an image including near-infrared light having an intensity corresponding to a pattern having a predetermined geometric shape, and a pattern defining the pattern Using the information, a process for outputting a color signal representing a visible light component corresponding to the acquired video signal and a near infrared signal representing a near infrared light component corresponding to the video signal is executed.
- a computer-readable program recording medium recording a program is provided.
- FIG. 1 is a block diagram illustrating the configuration of the video processing apparatus.
- FIG. 2 is a diagram illustrating an NIR cut filter.
- FIG. 3 is a block diagram illustrating the configuration of the video processing apparatus.
- FIG. 4 is a block diagram illustrating the configuration of the second color signal estimation unit.
- FIG. 5 is a flowchart illustrating an outline of processing executed by the video processing apparatus.
- FIG. 6 is a diagram illustrating a near infrared signal.
- FIG. 7A is a diagram illustrating near-infrared light according to the pattern of the NIR cut filter.
- FIG. 7B is a diagram illustrating near-infrared light according to the pattern of the NIR cut filter.
- FIG. 8 is a diagram for explaining the value of equation (2).
- FIG. 8 is a diagram for explaining the value of equation (2).
- FIG. 9 is a diagram illustrating the relationship between the intensity of near infrared light and the distance from the center.
- FIG. 10 is a schematic view illustrating the configuration of the photographing apparatus.
- FIG. 11 is a schematic diagram illustrating the behavior of near-infrared light incident on the light receiving unit.
- FIG. 12 is a diagram illustrating the configuration of the color filter.
- FIG. 13 is a diagram for explaining an example of the demosaicing process.
- FIG. 14 is a diagram illustrating the correspondence between the infrared transmission part of the NIR cut filter and the color filter.
- FIG. 15 is a schematic view illustrating the configuration of the photographing apparatus.
- FIG. 16 is a schematic view illustrating the configuration of the photographing apparatus.
- FIG. 17 is a diagram illustrating the spectral transmittance of the RGB three-color optical filter.
- FIG. 18 is a diagram showing color matching functions of the XYZ color system relating to human color perception.
- FIG. 19 is a diagram illustrating a spectral intensity distribution of certain color light.
- FIG. 20 is a diagram illustrating an example of spectral characteristics of an infrared cut filter.
- FIG. 21 is a diagram illustrating a configuration example of a color image input apparatus.
- FIG. 22 is a diagram illustrating a configuration example of a color image input apparatus.
- FIG. 1 is a block diagram illustrating the configuration of a video processing apparatus according to an embodiment of the present invention.
- the video processing apparatus 100 is an apparatus that acquires a video signal representing an image including near-infrared light and outputs a color signal and a near-infrared signal corresponding to the acquired video signal.
- the video processing apparatus 100 is an apparatus that separates a color signal and a near-infrared signal from an image captured in a state including visible light and near-infrared light.
- the arrows shown in the block diagrams after this figure show an example of the signal flow, and are not intended to limit the signal flow to a specific direction.
- the image means an image taken in through an optical system such as a lens, and may be either a still image or a moving image.
- the color signal is a signal representing a visible light component in the video signal.
- the near-infrared signal is a signal representing a near-infrared light component in the video signal.
- the color signal and the near-infrared signal represent, for example, the luminance of the pixel, but are not limited to luminance.
- the color signal and the near-infrared signal represent the brightness of each pixel of the image at a specific time point of the video when the color signal and the near-infrared signal are still images or videos.
- the visible region means a wavelength region of 400 to 700 nm.
- the wavelength range of 400 to 500 nm is blue (B)
- 500 to 600 nm is green (G)
- 600 to 700 nm is red (R).
- the near infrared region refers to a wavelength region of 700 nm to 2.5 ⁇ m.
- the wavelength region classification shown here is merely an example.
- the video processing apparatus 100 includes an acquisition unit 110 and a signal processing unit 120.
- the video processing device 100 is connected to an external device that supplies a video signal.
- the external device is, for example, an imaging device having an image sensor.
- the acquisition unit 110 acquires a video signal from the external device.
- the signal processing unit 120 outputs a color signal and a near-infrared signal based on the video signal acquired by the acquisition unit 110.
- the acquisition unit 110 acquires a video signal representing an image including near-infrared light having an intensity corresponding to a pattern having a predetermined geometric shape.
- the pattern here is a pattern in which figures such as a circle and a rectangle are repeatedly arranged regularly.
- Such a video signal is obtained, for example, by providing an optical filter (hereinafter referred to as “NIR cut filter”) that cuts near infrared light in the image sensor.
- NIR cut filter optical filter
- FIG. 2 is a diagram illustrating an NIR cut filter, and is a diagram illustrating the NIR cut filter perpendicular to the incident direction of light.
- the NIR cut filter 10 shown in FIG. 2 has a configuration in which a plurality of infrared transmission parts 12 are provided in the filter part 11.
- the infrared transmission parts 12 are circular holes arranged at equal intervals.
- the NIR cut filter 10 transmits near-infrared light without blocking in the infrared transmission part 12, and the near-infrared light is cut by a predetermined ratio or more in the filter part 11.
- the pattern formed on the video by the near-infrared light component included in the video signal does not necessarily match the pattern of the NIR cut filter. This is because near infrared light diffracts after passing through the NIR cut filter.
- Each pattern of the near-infrared light component pattern that appears as an image in the video signal is larger than the pattern formed on the NIR cut filter (corresponding to the infrared transmission part 12 in FIG. 2).
- the signal processing unit 120 acquires such a video signal from the acquisition unit 110, and outputs a color signal and a near-infrared signal.
- the color signal is typically output as a three-component signal of R, G, and B, but is not necessarily limited thereto.
- the signal processing unit 120 can output a color signal and a near-infrared signal by executing predetermined arithmetic processing using pattern information that defines a pattern of near-infrared light.
- the pattern information is information for specifying the near-infrared light pattern on the video.
- the pattern information is data representing the position and shape of the pattern in the NIR cut filter.
- the pattern information may be data describing the coordinates of the center of the circle and the radius of the circle of the infrared transmission part 12 as parameters, but the pattern can be specified. Any data can be used.
- the pattern information may be stored in advance in the video processing apparatus 100 or may be input by a user or the like. For example, the pattern information may be obtained in advance by the user performing calibration.
- the video processing apparatus 100 separates the color signal and the near infrared signal from the video signal by using the video signal representing the video including the near infrared light in a predetermined pattern and the pattern information. Is possible. Therefore, according to the video processing apparatus 100, it is possible to execute visible light video processing and near infrared light video processing based on a video signal including a color signal and a near infrared signal, respectively.
- the image pickup apparatus when such an image signal is obtained by an image pickup apparatus, the image pickup apparatus only needs to be provided with an NIR cut filter as shown in FIG. 2 in a general image pickup apparatus, so that a general configuration can be used.
- the use of the output near-infrared signal is not particularly limited.
- FIG. 3 is a block diagram illustrating the configuration of a video processing apparatus according to another embodiment of the invention.
- the video processing device 200 shown in FIG. 3 includes a video data acquisition unit 210, a first color signal acquisition unit 220, a pattern storage unit 230, a second color signal estimation unit 240, a near infrared signal calculation unit 250, And an output unit 260.
- the video processing device 200 has the same function as the video processing device 100 of the first embodiment.
- the video data acquisition unit 210 acquires video data.
- the video data acquisition unit 210 can acquire video data from the same external device as in the first embodiment.
- the video data includes at least a plurality of color signals.
- the plurality of color signals are color signals expressed by being separated into three color components of R, G, and B, and each pixel is represented by a value of a predetermined number of bits.
- the color signal here represents an image in a state where a near-infrared light component is superimposed on a visible light component.
- a color signal is also referred to as a “first color signal”.
- the first color signal is a signal obtained by adding a second color signal described later and a near-infrared signal.
- the first color signal acquisition unit 220 acquires the first color signal from the video data acquisition unit 210.
- the first color signal acquisition unit 220 acquires a first color signal for each color.
- the pattern storage unit 230 stores pattern information.
- the pattern storage unit 230 is configured by a storage medium such as a hard disk or a flash memory. Note that the same data as in the first embodiment can be used as the pattern information in this embodiment.
- the pattern information can use common data for each color.
- the second color signal estimation unit 240 estimates a second color signal that is a color signal obtained by removing the near-infrared light component from the first color signal. Further, the second color signal estimation unit 240 estimates the intensity ratio between the second color signal and the near infrared signal in addition to the second color signal. The second color signal estimation unit 240 calculates the second color signal and the intensity ratio of each color based on the first color signal acquired by the first color signal acquisition unit 220 and the pattern information stored in the pattern storage unit 230. presume.
- the near infrared signal calculation unit 250 calculates a near infrared signal for each color.
- the near-infrared signal calculation unit 250 can calculate a near-infrared signal using the second color signal estimated by the second color signal estimation unit 240 and the intensity ratio of the second color signal and the near-infrared signal. it can.
- the output unit 260 outputs the second color signal and the near infrared signal.
- the output unit 260 performs a predetermined calculation (for example, addition) on the near-infrared signal for each color calculated by the near-infrared signal calculation unit 250 and outputs the result.
- first color signal acquisition unit 220 the second color signal estimation unit 240, and the near-infrared signal calculation unit 250 may execute the processing of each color sequentially or in parallel.
- FIG. 4 is a block diagram showing the configuration of the second color signal estimation unit 240 in more detail.
- the second color signal estimation unit 240 includes an initial value estimation unit 241, an estimated value selection unit 242, a smoothness evaluation unit 243, a first color signal estimation unit 244, an error calculation unit 245, and an estimated value update unit 246. With.
- the initial value estimation unit 241 calculates an initial value of the second color signal and the estimated value of the intensity ratio between the second color signal and the near-infrared signal.
- the initial value estimation unit 241 calculates an initial value of the estimated value of the second color signal and an estimated value of the intensity ratio based on the first color signal.
- the estimated value selection unit 242 selects the second color signal and the estimated value of the intensity ratio, respectively.
- the estimated value selection unit 242 repeats the process of selecting these estimated values.
- the estimated value selection unit 242 selects the initial value calculated by the initial value estimation unit 241 in the first selection process, while the estimated value updated by the estimated value update unit 246 in the second and subsequent selection processes. Select.
- the smoothness evaluation unit 243 evaluates the smoothness of the second color signal and the estimated value of the intensity ratio.
- smoothness means the degree of spatial variation of values and is also referred to as “smoothness”.
- the smoothness of the estimated value of the second color signal means that the difference between the maximum value and the minimum value of the estimated value of pixels within a certain range that constitutes the video is equal to or less than a predetermined threshold value.
- the smoothness evaluation unit 243 calculates a smoothness evaluation value according to a predetermined algorithm.
- the first color signal estimation unit 244 estimates the first color signal.
- the first color signal estimation unit 244 calculates an estimated value of the first color signal based on the estimated value selected by the estimated value selection unit 242 and the pattern information stored in the pattern storage unit 230.
- the error calculation unit 245 compares the estimated value of the first color signal with the actual first color signal, and calculates the error. That is, the error calculation unit 245 compares the first color signal estimated by the first color signal estimation unit 244 with the first color signal acquired by the first color signal acquisition unit 220.
- the estimated value update unit 246 updates the estimated value of the second color signal and the intensity ratio.
- the estimated value updating unit 246 updates these estimated values based on the evaluation value calculated by the smoothness evaluating unit 243 and the error calculated by the error calculating unit 245.
- the estimated value update unit 246 compares the estimated values before and after the update, and ends the update when the update amount of each estimated value becomes sufficiently small. Specifically, the estimated value update unit 246 compares the update amount of the estimated value with a predetermined threshold value, and ends the update when the update amount is equal to or less than the threshold value. The estimated value updating unit 246 uses the estimated value at the time when the updating is completed as the output value of the second color signal estimating unit 240.
- the estimated value update unit 246 supplies the estimated value to the estimated value selection unit 242 when the update amount exceeds the threshold value.
- the estimated value selection unit 242, the smoothness evaluation unit 243, the first color signal estimation unit 244, the error calculation unit 245, and the estimated value update unit 246 execute the above-described process again using the updated estimated value. This is repeated until the update of the estimated value is completed.
- the configuration of the video processing apparatus 200 is as described above. In this configuration, when the video processing device 200 acquires video data, the video processing device 200 outputs a color signal and a near-infrared signal.
- the specific operation of the video processing apparatus 200 is as follows. Here, it is assumed that color signals of R, G, and B colors are set for all the pixels of the video data.
- FIG. 5 is a flowchart illustrating an outline of processing executed by the video processing apparatus 200.
- the video processing apparatus 200 does not necessarily need to execute processing as shown in FIG.
- the video processing apparatus 200 may execute the processes in steps S3 and S4 in parallel for the color signals of R, G, and B colors.
- the video data acquisition unit 210 acquires video data (step S1).
- the first color signal acquisition unit 220 selects any one of a plurality of first color signals included in the video data acquired by the video data acquisition unit 210 (step S2).
- the first color signal acquisition unit 220 selects a first color signal that has not yet been processed in steps S3 and S4 described below.
- the second color signal estimation unit 240 is close to the second color signal and the second color signal based on the selected first color signal. Infrared signal intensity ratios are estimated (step S3). In other words, the second color signal estimation unit 240 calculates an estimated value of the second color signal and an estimated value of the intensity ratio, respectively. Next, the near-infrared signal calculation unit 250 calculates a near-infrared signal based on these estimated values (step S4).
- the output unit 260 outputs the necessary second color signal and near-infrared signal when they are obtained. That is, the output unit 260 determines whether or not the processing in steps S2 to S4 has been performed for all colors (step S5), and if the processing for all colors has been completed (step S5: YES), the second color signal. And a near-infrared signal is output (step S6).
- step S5 when there is a color that has not been subjected to the processes of steps S2 to S4 (step S5: NO), the first color signal acquisition unit 220 selects an unprocessed first color signal (step S2).
- the second color signal estimation unit 240 and the near-infrared signal calculation unit 250 execute the processes of steps S3 and S4 again according to the selection of step S2.
- steps S3 and S4 are as follows in more detail. Note that the following description is described using G color signals for convenience, but the processing of other colors is the same.
- FIG. 6 is a diagram illustrating a near-infrared signal in the present embodiment, and represents a near-infrared signal that has passed through a circular infrared transmission portion provided in the NIR cut filter.
- the X axis and the Y axis correspond to orthogonal coordinates defined for the video represented by the video data.
- the Z axis indicates the luminance (brightness) of the near-infrared signal.
- the near-infrared signal has a significant value in a range wider than the actual area of the infrared transmission part due to the influence of diffraction of the near-infrared light, and the value goes from the center of the infrared transmission part to the outside thereof. And gradually decrease.
- a component derived from one infrared transmission part and a component derived from another infrared transmission part may be mixed.
- FIG. 7A and 7B are diagrams exemplifying near infrared light corresponding to the pattern of the NIR cut filter, and show the range of near infrared light irradiated on the image sensor.
- FIG. 7A shows a case where the infrared light from the respective infrared transmission portions does not overlap.
- FIG. 7B shows a case where the infrared light from the respective infrared transmission parts overlap.
- the intensity I (w) on the image sensor when near-infrared light having a wavelength ⁇ and an incident intensity I 0 is incident on a certain infrared transmission part on the NIR cut filter is expressed by the following formula ( Represented by 1).
- J 1 (w) is a first-order Bessel function of degree 1
- C is a predetermined correction coefficient.
- the correction coefficient C is a coefficient for adjusting the intensity I (w) so that it matches the pattern formed in the actual video. W is as shown in the following formula (2).
- a represents the radius of the infrared transmission part.
- q and R are respectively the distance between the point p and the point where the perpendicular line dropped from the center of the infrared transmission part to the image sensor intersects the image sensor, where p is an arbitrary point on the image sensor, This corresponds to the distance between the center of the infrared transmission part and the point p.
- FIG. 8 illustrates a, q, and R in the formula (2).
- FIG. 9 is a diagram illustrating the relationship between the intensity and the distance from the center of one of the patterns formed on the image sensor by the near infrared light being diffracted by the NIR cut filter.
- the correction coefficient C in Expression (1) is determined so that the intensity I (w) matches such a pattern.
- the intensity of the near-infrared signal at the position X on the image corresponding to the near-infrared light transmission part is I NIR_G (X)
- the near-red light transmitted through the same infrared transmission part The intensity I NIR — G (X, x) observed at the pixel at the position x by external light is expressed as the following equation (3).
- k X ⁇ x is a coefficient calculated using the formulas (1) and (2) from the distance between the position X and the position x on the image sensor.
- the calculation method of the coefficient k X ⁇ x is not limited to this.
- a calculation method of the coefficient k X ⁇ x when the spectral distribution of the near-infrared signal I NIR_G (X) is known, it is calculated using the equations (1) and (2) at each wavelength.
- the coefficient k X ⁇ x can also be obtained by calculating based on a standard near-infrared spectral distribution in the installation scene or by using another means in advance.
- the light reaching the pixel at the position x is a mixed light of visible light and near infrared light.
- the first color signal in the pixel at the position x is transmitted to the pixel from the second color signal, which is a color signal based only on the visible light component, and a plurality of infrared transmission parts constituting the pattern on the NIR cut filter.
- This is a sum of signals representing near-infrared light that arrives. Therefore, this first color signal can be expressed by I G_NIR (x) shown in Expression (4).
- I G (x) corresponds to the second color signal representing the G component of the visible light components.
- I G represents a vector having the second color signal (I G (x)) of each pixel as an element.
- S represents a sampling matrix for extracting the second color signal in a portion that transmits near-infrared light.
- D (M G ) represents a diagonal matrix having each element of the vector M G having the values of the intensity ratios (m G ) of the portions that transmit near infrared light as diagonal elements.
- K represents a matrix having the values of the coefficient k X ⁇ x as elements.
- the second color signal I G and the intensity ratio M G are obtained by calculating a value that minimizes the energy function E represented by the following equation (8).
- the first term on the right side of Equation (8) is a value greater than 0 if the second color signal I G and the intensity ratio M G do not satisfy the relationship of Equation (7).
- the second and third terms on the right side of Equation (8) are regularizations to prevent the energy minimization of the first term on the right side derived from Equation (7) from falling into an ill-posed state.
- Term (regularization term) are coefficients ⁇ 1 set in advance in the cost functions C 1 (I G ) and C 2 (M G ) for evaluating the spatial smoothness of the second color signal and the intensity ratio. , ⁇ 2 .
- the cost function for evaluating the smoothness is expressed by C (p) shown in equations (9) and (10), for example.
- ⁇ p represents a set of pixels constituting the image sensor
- N (x) represents a function indicating the positions of spatially adjacent pixel groups.
- p (x) represents data (second color signal I G or intensity ratio M G ) corresponding to the pixel at the position x on the image sensor
- p (y) is spatially applied to the pixel at the position x.
- Data (second color signal I G or intensity ratio M G ) corresponding to any arbitrary pixel in the adjacent pixel group is represented.
- the second color signal I G and the intensity ratio M G are calculated by the second color signal estimation unit 240 updating values using repetitive calculations.
- I 0 G is set as I G_NIR
- M 0 G is set as a matrix having values of all elements of 1.0
- the update expressions represented by the equations (11) and (12) are set. This calculation is repeated until the update amount becomes sufficiently small.
- I t G and M t G represent I G and M G at the repetition count t, respectively.
- K t represents a coefficient for adjusting the update amount at the repetition count t and satisfies 0 ⁇ k t ⁇ 1.
- E represents a unit matrix, and superscript T represents transposition of the matrix.
- the near-infrared signal I NIR_G can be calculated.
- the near-infrared signal calculating section 250 calculates the near-infrared signal I NIR_G by substituting the second color signal I G and the intensity ratio M G in equation (5).
- the second color signal and the near-infrared signal can be calculated similarly to the G component described above. That is, the R component second color signal I R , the near infrared signal I NIR_R, the B component second color signal I B , and the near infrared signal I NIR_B can be calculated by the same calculation as the G component.
- the near-infrared signal output from the video processing device 200 is obtained by adding the near-infrared signals of the R, G, and B components. That is, if the near-infrared signal to be output is I NIR , I NIR is as shown in the following equation (13).
- Image processing apparatus 200 by executing such a processing, it is possible to output the image data including the near-infrared signals I NIR and second color signals I R, I G, I B .
- the video processing apparatus 200 can obtain the second color signal and the near-infrared signal from the first color signal only by preparing the NIR cut filter and the corresponding pattern information. In this case, the imaging device does not require a special configuration other than the NIR cut filter.
- FIG. 10 is a schematic view illustrating the configuration of an imaging apparatus according to yet another embodiment of the invention.
- An imaging apparatus 300 illustrated in FIG. 10 includes a light receiving unit 310 and a video processing unit 320. More specifically, the light receiving unit 310 includes an NIR cut filter 311, a color filter 312, and a photosensor 313. Light including visible light and near-infrared light is incident on the photographing apparatus 300 via an optical system such as a lens.
- the NIR cut filter 311 is an optical filter having the same configuration as the NIR cut filter of the first embodiment and the second embodiment.
- the NIR cut filter 311 is provided on the front side of the traveling direction of the incident light with respect to the color filter 312 and the photosensor 313.
- the NIR cut filter 311 is provided at a predetermined distance from the color filter 312 and the photosensor 313 so that the near infrared light diffused by diffraction is received by the photosensor 313.
- the NIR cut filter 311 may be configured to be detachable or movable.
- FIG. 11 is a schematic diagram showing the behavior of near-infrared light incident on the light receiving unit 310. As shown in FIG. 11, near-infrared light passes through a part (infrared transmission part) of the NIR cut filter 311, but is cut off at other parts. However, since near-infrared light is diffracted when passing through the infrared transmission part, it enters the photosensor 313 in a wider range than the infrared transmission part.
- the color filter 312 is a three-color optical filter having a general configuration.
- the color filter 312 has, for example, the spectral characteristics shown in FIG.
- the photosensor 313 includes a plurality of photoelectric elements (that is, sensors) that generate signals according to the intensity of incident light.
- the photosensor 313 may have a configuration similar to that of a general image input device or photographing device.
- the video processing unit 320 acquires the signal generated by the photosensor 313 and executes video processing.
- the video processing unit 320 has a function common to the video processing apparatus 200 of the second embodiment and a function of executing a demosaicing process described later.
- FIG. 12 is a diagram partially showing the configuration of the color filter 312. As shown in FIG. 12, the color filter 312 has a so-called Bayer type arrangement. The color filter 312 is provided so that each filter corresponds to each sensor (that is, pixel) of the photosensor 313.
- the pattern of the NIR cut filter 311 may have a corresponding relationship with the pixel arrangement of the photosensor 313.
- the correspondence here is, for example, that the interval between the infrared transmission parts of the NIR cut filter 311 is equal to the interval between the pixels of the photosensor 313 or is an integer multiple.
- the infrared transmission part of the NIR cut filter 311 may be provided so as to overlap with pixels corresponding to a specific color of the photosensor 313.
- the pattern of the NIR cut filter 311 does not necessarily have a corresponding relationship with the pixel arrangement of the photosensor 313.
- the configuration of the photographing apparatus 300 is as described above. Under this configuration, the imaging apparatus 300 generates video data represented by four colors (four components) of R, G, B, and NIR, based on video data represented by three colors of R, G, and B. Is possible.
- the operation of the imaging device 300 is common to the operation of the video processing device 200 of the second embodiment in its main points. However, the imaging apparatus 300 performs a demosaicing process prior to the operation described in the second embodiment.
- FIG. 13 is a diagram for explaining an example of the demosaicing process, and is a diagram illustrating a correspondence relationship between pixels and coordinates.
- coordinates of (1, 1), (1, 2), (2, 1), and (2, 2) are assigned to the pixels in 2 rows and 2 columns shown in FIG. Note that the pixel at coordinates (1, 1) corresponds to the R component. A pixel at coordinates (2, 2) corresponds to the B component. The remaining pixels correspond to the G component.
- color information (color signal values) of RGB colors at coordinates (i, j) will be expressed as R (i, j), G (i, j), and B (i, j), respectively.
- R (1,1) represents the color information of the R component of the pixel at coordinates (1,1).
- the color information at the time of execution of the demosaicing process actually includes an NIR component.
- the NIR component of the color information is not considered.
- the pixel at the coordinates (1, 1) corresponds to the R component. Therefore, the color information of the R component at the coordinates (1, 1) is expressed by the following formula (14).
- R (1,1) R (1,1) (14)
- the pixel at the coordinates (1, 1) does not receive other color components. For this reason, the color information of the G component and B component of the pixel at the coordinates (1, 1) is obtained by interpolation from the surrounding pixels as shown in equations (15) and (16).
- G (1,1) (G (2,1) + G (1,2)) / 2 (15)
- B (1,1) B (2,2) (16)
- the color information of the pixel at coordinates (2, 1) is obtained in the same manner as the color information of the pixel at coordinates (1, 2).
- the color information of the pixel at coordinates (2, 2) is obtained in the same manner as the color information of the pixel at coordinates (1, 1).
- the video processing unit 320 executes such processing for all the pixels and acquires color information of each color. Subsequently, the video processing unit 320 calculates a near-infrared signal by the operation described in the second embodiment. Note that the demosaicing process is not limited to the method described above, and may be executed using, for example, the methods disclosed in Non-Patent Documents 4 to 6.
- the imaging device 300 can achieve the same effects as the video processing device 200 of the second embodiment.
- the imaging apparatus 300 can disperse infrared light by diffraction in the NIR cut filter 311. Thereby, the imaging apparatus 300 can reduce the intensity of the near-infrared signal per pixel even when infrared light having an intensity that saturates the near-infrared signal is incident on the photosensor 313. It makes it possible to increase the apparent dynamic range.
- FIG. 14 is a diagram illustrating a preferable correspondence relationship between the infrared transmission unit 311 a of the NIR cut filter 311 and the color filter 312.
- FIG. 14 shows the positional relationship when the NIR cut filter 311 and the color filter 312 are viewed from the direction in which light enters.
- the infrared transmission part 311a shown in FIG. 14 is in a position overlapping with the pixel corresponding to the G component.
- I NIR_R and I NIR_B are ignored compared to the case where the NIR cut filter 311 does not have such a pattern (for example, when the infrared transmission unit 311a overlaps with a pixel corresponding to the R component or B component). It is possible to reduce the influence of the error.
- Embodiments of the present invention are not limited to the first to third embodiments described above.
- the present invention can be implemented by the following modified embodiments.
- the present invention may be implemented in a mode in which the first to third embodiments and modifications are appropriately combined.
- the specific shape of the NIR cut filter pattern is not limited as long as it can be described as pattern information.
- the infrared transmission portion does not have to be circular, and all the infrared transmission portions do not necessarily have the same shape.
- the visible light component is not limited to the three components R, G, and B.
- the visible light component for example, three components of cyan (C), magenta (M), and yellow (Y) may be used. Further, the visible light component need not be composed of three components, and may be more or less than this.
- Modification 3 15 and 16 are diagrams showing another example of the photographing apparatus.
- FIG. 15 is a diagram showing an imaging apparatus 400 having a so-called three-plate type, that is, a configuration in which sensors corresponding to R, G, and B colors are independent.
- FIG. 16 is a diagram showing an imaging apparatus 500 including a so-called stacked sensor. The present invention can also be applied to an imaging apparatus having such a configuration.
- the photographing apparatus 400 includes a prism 410, photosensors 420, 430, and 440, an NIR cut filter 450, and an image processing unit 460.
- the prism 410 decomposes the incident light and emits the light in directions corresponding to the R, G, and B components.
- the photo sensors 420 (R), 430 (G), and 440 (B) generate signals corresponding to the intensity of incident light of each color.
- the NIR cut filter 450 is an optical filter similar to the NIR cut filter 311 of the third embodiment.
- the NIR cut filter 450 is not necessarily provided in all of the photosensors 420, 430, and 440, and may be provided in any of these (the photosensor 420 in FIG. 15) according to the spectral characteristics of the prism 410.
- the near infrared light incident on the photosensors 430 and 440 is sufficiently smaller than the near infrared light incident on the photosensor 420.
- an optical filter that cuts near infrared light is provided in front of the photosensors 430 and 440. Good.
- the video processing unit 460 may have the same configuration as the video processing unit 320 described in the third embodiment. However, in the example shown in FIG. 15, the color signal including the near infrared light component is only the R component. Therefore, the video processing unit 460 only needs to perform the process of separating the near-infrared signal from the color signal only on the R component color signal.
- the photographing apparatus 500 includes an NIR cut filter 510, a stacked sensor 520, and a video processing unit 530.
- the NIR cut filter 510 and the video processing unit 530 may have the same configuration as the NIR cut filter 450 and the video processing unit 460 shown in FIG.
- the stacked sensor 520 is a sensor in which sensors 521, 522, and 523 are stacked.
- the sensor 521 has sensitivity in the wavelength region of the B component.
- the sensor 522 has sensitivity in the wavelength region of the G component.
- the sensor 523 has sensitivity in the wavelength region of the R component and the near infrared light component.
- the video processing devices 100 and 200 and the video processing unit 320 can be realized by a processing device (processor) such as a CPU (Central Processing Unit) and a memory.
- processor processing device
- the present invention may be realized by a general-purpose processor or a processor dedicated to video processing.
- the present invention may be provided in the form of a program executable by a computer.
- This program may be provided in the form of being downloaded from another device (server or the like) via a network, or may be provided in the form of a computer-readable recording medium.
- the present invention can be provided as a video processing method in addition to a video processing device, a photographing device, a program, and a recording medium.
- Video processing device 110 Acquisition unit 120 Signal processing unit 210 Video data acquisition unit 220 First color signal acquisition unit 230 Pattern storage unit 240 Second color signal estimation unit 241 Initial value estimation unit 242 Estimation value selection unit 243 Smoothness evaluation unit 244 First color signal estimation unit 245 Error calculation unit 246 Estimated value update unit 250 Near infrared signal calculation unit 260 Output unit 300, 400, 500 Imaging device 310 Light receiving unit 312 Color filter 313 Photo sensor 320 Image processing unit
Landscapes
- Engineering & Computer Science (AREA)
- Multimedia (AREA)
- Signal Processing (AREA)
- Physics & Mathematics (AREA)
- Spectroscopy & Molecular Physics (AREA)
- Health & Medical Sciences (AREA)
- Toxicology (AREA)
- General Physics & Mathematics (AREA)
- Optics & Photonics (AREA)
- Human Computer Interaction (AREA)
- Color Television Image Signal Generators (AREA)
- Optical Filters (AREA)
Abstract
Description
図1は、本発明の一実施形態に係る映像処理装置の構成を例示するブロック図である。映像処理装置100は、近赤外光を含む映像を表す映像信号を取得し、取得した映像信号に対応する色信号及び近赤外信号を出力する装置である。換言すれば、映像処理装置100は、可視光及び近赤外光を含んだ状態で撮像された映像から色信号と近赤外信号を分離する装置である。なお、本図以降のブロック図に示した矢印は、信号の流れの一例を示すものであって、信号の流れが特定の方向に限定されることを意図したものではない。
図3は、本発明の別の実施形態に係る映像処理装置の構成を例示するブロック図である。図3に示す映像処理装置200は、映像データ取得部210と、第1色信号取得部220と、パターン記憶部230と、第2色信号推定部240と、近赤外信号算出部250と、出力部260とを備える。映像処理装置200は、第1実施形態の映像処理装置100と同様の機能を有する。
図10は、本発明のさらに別の実施形態に係る撮影装置の構成を例示する模式図である。図10に示す撮影装置300は、受光部310と、映像処理部320とを備える。受光部310は、より詳細には、NIRカットフィルタ311と、カラーフィルタ312と、フォトセンサ313とを備える。撮影装置300には、レンズ等の光学系を介して、可視光と近赤外光とを含んだ光が入射される。
R(1,1)=R(1,1) (14)
G(1,1)=(G(2,1)+G(1,2))/2 (15)
B(1,1)=B(2,2) (16)
G(1,2)=G(1,2) (17)
R(1,2)=R(1,1) (18)
B(1,2)=B(2,2) (19)
本発明の実施の形態は、上述した第1~第3実施形態に限定されない。例えば、本発明は、以下に説明する変形例の態様によっても実施可能である。また、本発明は、第1~第3実施形態及び変形例を適宜に組み合わせた態様で実施されてもよい。
本発明の実施形態において、NIRカットフィルタのパターンは、パターン情報として記述可能なパターンであれば、その具体的形状は限定されない。例えば、NIRカットフィルタに形成されるパターンは、赤外透過部が円形でなくてもよく、また、全ての赤外透過部が必ずしも同じ形状である必要もない。
本発明の実施形態において、可視光成分は、R、G、Bの3成分に限定されない。可視光成分は、例えば、シアン(C)、マゼンタ(M)、イエロー(Y)の3成分が用いられてもよい。また、可視光成分は、3成分で構成される必要はなく、これより多くても少なくてもよい。
図15及び図16は、撮影装置の他の例を示す図である。図15は、いわゆる3板式、すなわちR、G、B各色に対応するセンサが独立した構成の撮影装置400を示す図である。また、図16は、いわゆる積層型のセンサを備えた撮影装置500を示す図である。本発明は、このような構成の撮影装置に対しても適用可能である。
本発明は、その全部又は一部の構成をコンピュータによって実現することが可能である。例えば、映像処理装置100、200及び映像処理部320は、CPU(Central Processing Unit)などの処理装置(プロセッサ)とメモリによって実現可能である。また、本発明は、汎用的なプロセッサによって実現されてもよいし、映像処理専用のプロセッサによって実現されてもよい。
12、311a 赤外透過部
100、200 映像処理装置
110 取得部
120 信号処理部
210 映像データ取得部
220 第1色信号取得部
230 パターン記憶部
240 第2色信号推定部
241 初期値推定部
242 推定値選択部
243 平滑さ評価部
244 第1色信号推定部
245 誤差算出部
246 推定値更新部
250 近赤外信号算出部
260 出力部
300、400、500 撮影装置
310 受光部
312 カラーフィルタ
313 フォトセンサ
320 映像処理部
Claims (8)
- 所定の幾何学的形状を有するパターンに応じた強度の近赤外光を含む映像を表す映像信号を取得する取得手段と、
前記パターンを規定するパターン情報を用いて、前記取得された映像信号に対応する可視光成分を表す色信号と当該映像信号に対応する近赤外光成分を表す近赤外信号とを出力する信号処理手段と
を備える映像処理装置。 - 前記パターン情報は、前記パターンの位置及び形状を表す
請求項1に記載の映像処理装置。 - 前記信号処理手段は、
前記色信号の推定値、前記色信号と前記近赤外信号との強度比の推定値及び前記パターン情報を用いて算出される前記映像信号の推定値と前記取得された映像信号との差分が第1の条件を満たし、かつ前記色信号の推定値と前記強度比の推定値の空間的なばらつきが第2の条件を満たした場合に、当該色信号の推定値と、当該色信号の推定値及び当該強度比の推定値から求められる前記近赤外信号の値とを出力信号とする
請求項1に記載の映像処理装置。 - 近赤外光を前記パターンで透過する透過部を有する光学フィルタを備える受光手段を備える
請求項1に記載の映像処理装置。 - 前記受光手段は、各々が複数の色成分のいずれかに対応する複数のセンサを備え、
前記パターンは、前記複数のセンサの配列と対応関係を有する
請求項4に記載の映像処理装置。 - 所定の幾何学的形状を有するパターンで近赤外光を透過する光学フィルタを有し、当該光学フィルタを透過した近赤外光を含む映像を表す映像信号を生成する受光手段と、
前記パターンを規定するパターン情報を用いて、前記生成された映像信号に対応する可視光成分を表す色信号と当該映像信号に対応する近赤外光成分を表す近赤外信号とを出力する映像処理手段と
を備える撮影装置。 - 所定の幾何学的形状を有するパターンに応じた強度の近赤外光を含む映像を表す映像信号を取得し、
前記パターンを規定するパターン情報を用いて、前記取得された映像信号に対応する可視光成分を表す色信号と当該映像信号に対応する近赤外光成分を表す近赤外信号とを出力する
映像処理方法。 - コンピュータに、
所定の幾何学的形状を有するパターンに応じた強度の近赤外光を含む映像を表す映像信号を取得する処理と、
前記パターンを規定するパターン情報を用いて、前記取得された映像信号に対応する可視光成分を表す色信号と当該映像信号に対応する近赤外光成分を表す近赤外信号とを出力する処理と
を実行させるためのプログラムを記録したコンピュータ読み取り可能なプログラム記録媒体。
Priority Applications (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
JP2017540507A JP6794989B2 (ja) | 2015-09-18 | 2016-09-14 | 映像処理装置、撮影装置、映像処理方法及びプログラム |
US15/757,456 US10440292B2 (en) | 2015-09-18 | 2016-09-14 | Color signal and near-infrared signal generated by using pattern information defining intensity-corresponding pattern |
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
JP2015-184885 | 2015-09-18 | ||
JP2015184885 | 2015-09-18 |
Publications (1)
Publication Number | Publication Date |
---|---|
WO2017047080A1 true WO2017047080A1 (ja) | 2017-03-23 |
Family
ID=58288460
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
PCT/JP2016/004195 WO2017047080A1 (ja) | 2015-09-18 | 2016-09-14 | 映像処理装置、撮影装置、映像処理方法及びプログラム記録媒体 |
Country Status (3)
Country | Link |
---|---|
US (1) | US10440292B2 (ja) |
JP (1) | JP6794989B2 (ja) |
WO (1) | WO2017047080A1 (ja) |
Cited By (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
WO2018179378A1 (ja) * | 2017-03-31 | 2018-10-04 | 日本電気株式会社 | 画像処理装置、画像処理システム、画像処理方法及びプログラム記録媒体 |
Families Citing this family (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP6765494B1 (ja) | 2019-10-31 | 2020-10-07 | パナソニックi−PROセンシングソリューションズ株式会社 | 3板式カメラ |
Citations (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP2007235760A (ja) * | 2006-03-02 | 2007-09-13 | Fujitsu Ltd | 紫外線画像および赤外線画像を撮像する撮像装置および撮像方法 |
JP2012080553A (ja) * | 2011-11-07 | 2012-04-19 | Sony Corp | 半導体装置および撮像装置 |
WO2015133130A1 (ja) * | 2014-03-06 | 2015-09-11 | 日本電気株式会社 | 映像撮影装置、信号分離装置および映像撮影方法 |
Family Cites Families (7)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20010016053A1 (en) * | 1997-10-10 | 2001-08-23 | Monte A. Dickson | Multi-spectral imaging sensor |
US8237121B2 (en) * | 2008-02-07 | 2012-08-07 | Omnivision Technologies, Inc. | Alternating row infrared filter for an image sensor |
JP2011243862A (ja) | 2010-05-20 | 2011-12-01 | Sony Corp | 撮像デバイス及び撮像装置 |
JP5432075B2 (ja) * | 2010-07-06 | 2014-03-05 | パナソニック株式会社 | 撮像装置および色温度算出方法 |
KR101695252B1 (ko) * | 2012-06-07 | 2017-01-13 | 한화테크윈 주식회사 | 멀티 대역 필터 어레이 기반 카메라 시스템 및 그의 영상 처리 방법 |
KR101926489B1 (ko) * | 2013-02-04 | 2018-12-07 | 한화테크윈 주식회사 | 다중대역 필터배열 센서를 이용한 영상융합장치 및 방법 |
US10171757B2 (en) * | 2013-10-23 | 2019-01-01 | Nec Corporation | Image capturing device, image capturing method, coded infrared cut filter, and coded particular color cut filter |
-
2016
- 2016-09-14 WO PCT/JP2016/004195 patent/WO2017047080A1/ja active Application Filing
- 2016-09-14 US US15/757,456 patent/US10440292B2/en active Active
- 2016-09-14 JP JP2017540507A patent/JP6794989B2/ja active Active
Patent Citations (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP2007235760A (ja) * | 2006-03-02 | 2007-09-13 | Fujitsu Ltd | 紫外線画像および赤外線画像を撮像する撮像装置および撮像方法 |
JP2012080553A (ja) * | 2011-11-07 | 2012-04-19 | Sony Corp | 半導体装置および撮像装置 |
WO2015133130A1 (ja) * | 2014-03-06 | 2015-09-11 | 日本電気株式会社 | 映像撮影装置、信号分離装置および映像撮影方法 |
Cited By (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
WO2018179378A1 (ja) * | 2017-03-31 | 2018-10-04 | 日本電気株式会社 | 画像処理装置、画像処理システム、画像処理方法及びプログラム記録媒体 |
JPWO2018179378A1 (ja) * | 2017-03-31 | 2020-01-16 | 日本電気株式会社 | 画像処理装置、画像処理システム、画像処理方法、及びプログラム |
US11140370B2 (en) | 2017-03-31 | 2021-10-05 | Nec Corporation | Image processing device, image processing system, image processing method, and program recording medium for generating a visible image and a near-infrared image |
Also Published As
Publication number | Publication date |
---|---|
JP6794989B2 (ja) | 2020-12-02 |
JPWO2017047080A1 (ja) | 2018-07-05 |
US20180249096A1 (en) | 2018-08-30 |
US10440292B2 (en) | 2019-10-08 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
JP5527448B2 (ja) | 画像入力装置 | |
US7839437B2 (en) | Image pickup apparatus, image processing method, and computer program capable of obtaining high-quality image data by controlling imbalance among sensitivities of light-receiving devices | |
US10136107B2 (en) | Imaging systems with visible light sensitive pixels and infrared light sensitive pixels | |
JP5543616B2 (ja) | 色フィルタアレイ画像反復デノイズ | |
JP6582987B2 (ja) | 映像撮影装置、映像撮影方法、符号型赤外カットフィルタ、および符号型特定色カットフィルタ | |
JP4346634B2 (ja) | 目標物検出装置 | |
US8411176B2 (en) | Image input device | |
CN104756488B (zh) | 信号处理设备以及信号处理方法 | |
WO2016098641A1 (ja) | 撮像装置、撮像方法、およびプログラム | |
JP7143767B2 (ja) | 画像処理方法、画像処理装置および画像処理プログラム | |
JP2015053578A (ja) | カラー撮像装置及びカラー撮像方法 | |
US10863115B2 (en) | Generation of visible and near-infrared images based on estimated incident light spectral characteristics and image capturing device spectral sensitivity characteristics | |
WO2015133130A1 (ja) | 映像撮影装置、信号分離装置および映像撮影方法 | |
JP6794989B2 (ja) | 映像処理装置、撮影装置、映像処理方法及びプログラム | |
KR101923957B1 (ko) | 감도 개선을 위한 영상 처리 장치 및 방법 | |
JP2010276469A (ja) | 画像処理装置及び測距装置の画像処理方法 | |
JP2012010141A (ja) | 画像処理装置 | |
JP2010171950A (ja) | 撮像装置および撮像装置の色補正方法 | |
JP6725060B2 (ja) | 画像処理装置、画像処理システム、画像処理方法、及びプログラム | |
JP5139350B2 (ja) | 画像処理装置、画像処理方法、撮像装置 | |
JP5920144B2 (ja) | 撮像装置および撮像方法 | |
TWI617198B (zh) | 具有透明濾波器像素之成像系統 |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
121 | Ep: the epo has been informed by wipo that ep was designated in this application |
Ref document number: 16845954 Country of ref document: EP Kind code of ref document: A1 |
|
WWE | Wipo information: entry into national phase |
Ref document number: 15757456 Country of ref document: US |
|
ENP | Entry into the national phase |
Ref document number: 2017540507 Country of ref document: JP Kind code of ref document: A |
|
NENP | Non-entry into the national phase |
Ref country code: DE |
|
122 | Ep: pct application non-entry in european phase |
Ref document number: 16845954 Country of ref document: EP Kind code of ref document: A1 |