US20200141804A1 - Method and system for hyperspectral light field imaging - Google Patents

Method and system for hyperspectral light field imaging Download PDF

Info

Publication number
US20200141804A1
US20200141804A1 US16/678,104 US201916678104A US2020141804A1 US 20200141804 A1 US20200141804 A1 US 20200141804A1 US 201916678104 A US201916678104 A US 201916678104A US 2020141804 A1 US2020141804 A1 US 2020141804A1
Authority
US
United States
Prior art keywords
images
histogram
spectral
overlapping
feature descriptor
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
US16/678,104
Other versions
US10641658B1 (en
Inventor
Jingyi Yu
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
ShanghaiTech University
Original Assignee
ShanghaiTech University
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by ShanghaiTech University filed Critical ShanghaiTech University
Assigned to SHANGHAITECH UNIVERSITY reassignment SHANGHAITECH UNIVERSITY ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: YU, JINGYI
Application granted granted Critical
Publication of US10641658B1 publication Critical patent/US10641658B1/en
Publication of US20200141804A1 publication Critical patent/US20200141804A1/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/50Depth or shape recovery
    • G06T7/55Depth or shape recovery from multiple images
    • G06T7/557Depth or shape recovery from multiple images from light fields, e.g. from plenoptic cameras
    • G06K9/0063
    • G06K9/4642
    • G06K9/6212
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N13/00Stereoscopic video systems; Multi-view video systems; Details thereof
    • H04N13/20Image signal generators
    • H04N13/271Image signal generators wherein the generated image signals comprise depth maps or disparity maps
    • GPHYSICS
    • G01MEASURING; TESTING
    • G01JMEASUREMENT OF INTENSITY, VELOCITY, SPECTRAL CONTENT, POLARISATION, PHASE OR PULSE CHARACTERISTICS OF INFRARED, VISIBLE OR ULTRAVIOLET LIGHT; COLORIMETRY; RADIATION PYROMETRY
    • G01J3/00Spectrometry; Spectrophotometry; Monochromators; Measuring colours
    • G01J3/28Investigating the spectrum
    • G01J3/2823Imaging spectrometer
    • G01J2003/2826Multispectral imaging, e.g. filter imaging
    • GPHYSICS
    • G01MEASURING; TESTING
    • G01JMEASUREMENT OF INTENSITY, VELOCITY, SPECTRAL CONTENT, POLARISATION, PHASE OR PULSE CHARACTERISTICS OF INFRARED, VISIBLE OR ULTRAVIOLET LIGHT; COLORIMETRY; RADIATION PYROMETRY
    • G01J3/00Spectrometry; Spectrophotometry; Monochromators; Measuring colours
    • G01J3/28Investigating the spectrum
    • G01J3/2823Imaging spectrometer
    • GPHYSICS
    • G02OPTICS
    • G02BOPTICAL ELEMENTS, SYSTEMS OR APPARATUS
    • G02B27/00Optical systems or apparatus not provided for by any of the groups G02B1/00 - G02B26/00, G02B30/00
    • G02B27/0075Optical systems or apparatus not provided for by any of the groups G02B1/00 - G02B26/00, G02B30/00 with means for altering, e.g. increasing, the depth of field or depth of focus
    • G06K2009/00644
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/10Image acquisition modality
    • G06T2207/10024Color image
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/10Image acquisition modality
    • G06T2207/10032Satellite or aerial image; Remote sensing
    • G06T2207/10036Multispectral image; Hyperspectral image
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/10Image acquisition modality
    • G06T2207/10052Images from lightfield camera
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V20/00Scenes; Scene-specific elements
    • G06V20/10Terrestrial scenes
    • G06V20/194Terrestrial scenes using hyperspectral data, i.e. more or other wavelengths than RGB

Definitions

  • the disclosure relates generally to hyperspectral light field imaging system for generating hyperspectral light field images, and more particularly, to method and system for generating complete hyperspectral data-cubes from captured hyperspectral light field images.
  • Hyperspectral light field (H-LF) imaging is part of a class of techniques commonly referred to as spectral imaging or spectral analysis.
  • the H-LF imaging collects and processes information from across the electromagnetic spectrum.
  • a hyperspectral camera/senor collects information as a set of H-LF images. Each image represents a narrow wavelength range of the electromagnetic spectrum, also known as a spectral band. These images are combined to form a hyperspectral data-cube for processing and analysis.
  • the goal of H-LF imaging is to obtain the spectrum for each pixel in the image of a scene with narrow spectral bands over a continuous spectral range. Every pixel in the image thus can be used to characterize the objects in the scene with great precision and detail.
  • the H-LF images provide much more detailed information about the scene than a normal color camera.
  • the H-LF imaging leads to a vastly improved ability to classify the objects in the scene based on their spectral properties. It can also take advantages of the spatial relationships among the different neighboring spectra, allowing more elaborate spectral-spatial models for a more accurate segmentation and classification of the image.
  • One aspect of the present disclosure is directed to a method for generating hyperspectral data-cubes based on a plurality of hyperspectral light field (H-LF) images.
  • H-LF hyperspectral light field
  • the method may include calculating a magnitude histogram, a direction histogram, and an overlapping histogram of oriented gradient for a plurality of pixels; developing a spectral-invariant feature descriptor by combining the magnitude histogram, the direction histogram, and the overlapping histogram of oriented gradient; obtaining a correspondence cost of the H-LF images based on the spectral-invariable feature descriptor; performing H-LF stereo matching on the H-LF images to obtain a disparity map of a reference view; and generating hyperspectral data-cubes by using the disparity map of the reference view.
  • a bin in the overlapping histogram of oriented gradient may comprise overlapping ranges of directions.
  • Another aspect of the present disclosure is directed to a method for generating hyperspectral data-cubes based on a plurality of hyperspectral light field (H-LF) images.
  • H-LF hyperspectral light field
  • the method may include obtaining an estimated spectra by synthesizing RGB color values from the H-LF images and mapping the RGB color values to a spectral band; obtaining a defocus cost of the H-LF images based on a comparison between estimated spectra and captured spectra of the H-LF images; performing H-LF stereo matching on the H-LF images to obtain a disparity map of a reference view; and generating hyperspectral data-cubes by using the disparity map of the reference view.
  • hyperspectral light field imaging H-LF
  • H-LF hyperspectral light field
  • the system may include a plurality of multi-view hyperspectral data terminals and a data processing unit.
  • the terminals may be configured to capture a plurality of H-LF images, and each image may have a view and a different spectral band.
  • the data processing unit may be configured to calculate a magnitude histogram, a direction histogram, and an overlapping histogram of oriented gradient for a plurality of pixels; develop a spectral-invariant feature descriptor by combining the magnitude histogram, the direction histogram, and the overlapping histogram of oriented gradient; obtain a correspondence cost of the H-LF images based on the spectral-invariable feature descriptor; perform H-LF stereo matching on the H-LF images to obtain a disparity map of a reference view; and generate hyperspectral data-cubes by using the disparity map of the reference view.
  • FIG. 1 is a schematic diagram for a hyperspectral light field imaging (HLFI) system, consistent with exemplary embodiments of the present disclosure.
  • HLFI hyperspectral light field imaging
  • FIGS. 2A and 2B are graphical representations illustrating a data capturing unit of an HLFI system, consistent with exemplary embodiments of the present disclosure.
  • FIG. 3 is a graphical representation illustrating a data processing unit of an HLFI system, consistent with exemplary embodiments of the present disclosure.
  • FIG. 4A is a diagram illustrating a method for H-LF stereo matching, consistent with exemplary embodiments of the present disclosure.
  • FIG. 4B is a flow diagram illustrating a method for estimating a disparity map based on a correspondence cost, consistent with exemplary embodiments of the present disclosure.
  • FIG. 5 is a graphical representation illustrating a structure of a feature descriptor, consistent with exemplary embodiments of the present disclosure.
  • FIG. 6 illustrates a method for obtaining a defocus cost, consistent with exemplary embodiments of the present disclosure.
  • FIG. 7 is a flow diagram illustrating a method for H-LF data-cube reconstruction, consistent with exemplary embodiments of the present disclosure.
  • a hyperspectral light field imaging (HLFI) system including a data capturing unit, a data processing unit and a data displaying unit.
  • the data capturing unit captures a plurality of multi-view H-LF images, which are the sampled H-LF images.
  • the sampled H-LF images are transmitted to the data processing unit.
  • the data processing unit preprocess the sampled H-LF images to obtain rectified and undistorted H-LF images, and performs H-LF stereo matching to obtain a disparity map of a reference view.
  • the data processing unit generates complete hyperspectral data-cubes based on the disparity map.
  • the complete hyperspectral data-cubes are transmitted to the data displaying unit for displaying.
  • the sampled H-LF images are 5 ⁇ 6 H-LF images. Each image is captured at a different view, and samples a different spectrum range with a bandwidth of 10 nm.
  • the total spectra of the 30 H-LF images cover the whole visible spectrum band from 410 nm to 700 nm.
  • the complete H-LF data-cubes include all 5 ⁇ 6 views and 5 ⁇ 6 ⁇ 30 images, where 30 is the number of sampled spectrum bands.
  • the complete H-LF data-cubes cover the spectrum range from 410 to 700 nm with 30 bands each 10 nm wide.
  • FIG. 1 shows an exemplary HLFI system 100 in accordance to an embodiment of the present disclosure.
  • the system 100 may include a number of components, some of which may be optional. In some embodiments, the system 100 may include many more components than those shown in FIG. 1 . However, it is not necessary that all of these components be shown in order to disclose an illustrative embodiment.
  • the system 100 may include a data capturing unit 200 , a data processing unit 300 and a data displaying unit 400 .
  • the data capturing unit 200 may include a plurality of multi-view hyperspectral data terminals 210 , and a camera calibration unit 220 .
  • the data processing unit 300 may include a data preprocessing unit 310 , an H-LF stereo matching unit 320 , and an H-LF data-cube reconstruction unit 330 .
  • the data displaying unit 400 may include an H-LF dynamic refocusing unit 401 and a complete H-LF data-cube unit 402 .
  • the data capture unit 200 may include a plurality of multi-view hyperspectral data terminals 210 , and a camera calibration unit 220 .
  • FIGS. 2A and 2B are graphical representations illustrating the plurality of multi-view hyperspectral data terminals 210 , in accordance to an embodiment of the present disclosure.
  • the terminals 210 may be placed in a rectangular array with 5 rows and 6 columns. Positions of each terminal 210 can be denoted as ⁇ i, j ⁇ , where i stands for the row number, and j stands for the column number.
  • the terminals 210 in each row are equally spaced on an identical plate. All rows may be uniformly fixed on a common limit bracket which ensures that each terminal can have same baselines with its neighboring terminals within the limits of installation accuracy and techniques.
  • a baseline is a distance between adjacent terminals/cameras in pixel, usually measured in unit of mm.
  • Each of the terminals 210 may consist of a monochrome camera and a narrow bandpass optical filter. Each filter may be centered on a specific wavelength with a bandwidth of 10 nm, and the minimum wavelength is 410 nm and the maximum wavelength is 700 nm. All 5 ⁇ 6 filters can sample the whole visible spectrum band from 410 nm to 700 nm, with intervals of 10 nm. In other words, the terminals 210 may capture 5 ⁇ 6 H-LF images. Each image samples a different spectrum range with a bandwidth of 10 nm, and the total spectra of the 30 H-LF images cover the whole visible spectrum band from 410 nm to 700 nm. These 30 H-LF images are sampled H-LF images for generating complete hyperspectral data-cubes in this disclosure. Table 1 lists parameters of hyperspectral data terminal array.
  • Cameras in the multi-view hyperspectral data terminals 210 may include intrinsic, extrinsic and distortion parameters.
  • the intrinsic parameters refer to the parameters to link pixel coordinates of an image point with the corresponding coordinates in a camera reference.
  • the extrinsic parameters may define a location and orientation of the camera reference frame with respect to a known world reference frame.
  • the distortion parameters may be caused due to the limitation of lens production technology and imaging models.
  • the intrinsic, extrinsic and distortion parameters are used to rectify the captured hyperspectral images.
  • the intrinsic, extrinsic and distortion parameters are obtained.
  • a camera calibration method from Zhengyou Zhang Z. Zhang, “A flexible new technique for camera calibration”, IEEE Trans. Pattern Anal. Mach. Intell., 22(11):1330-1334, 2000.
  • Z. Zhang “A flexible new technique for camera calibration”, IEEE Trans. Pattern Anal. Mach. Intell., 22(11):1330-1334, 2000.
  • Zhengyou Zhang's method is applied with the use of Matlab platform.
  • the extrinsic parameters can be obtained by applying Zhengyou Zhang's method to obtain parameters of rotation and translation between each neighboring pair of cameras.
  • a multi-view rectification method for the sampled H-LF images can be developed. Since all cameras' views cover largely the same area, the cameras do not need to be calibrated one by one. Instead, we can calibrate the cameras by guaranteeing a checkerboard in all cameras' views simultaneously when capturing images synchronously.
  • a “checkerboard” is a geometric pattern commonly used in camera calibration.
  • a suited exposure setting is determined to ensure all cameras can capture the contrast of the pattern of the checkerboard simultaneously.
  • the plurality of multi-view hyperspectral data terminals 210 capture H-LF images and send them to the data processing unit 300 via an IP network. Due to the size of the hyperspectral data and the limitation of transmission bandwidth, the data processing unit 300 may not obtain all data simultaneously.
  • software may be developed to ensure that the data of the captured H-LF images from all cameras (5 ⁇ 6) can be transmitted synchronously to the data processing unit 300 .
  • a data buffer queue may be built on each camera, and all the data are transmitted with a time-shared mode, so that the H-LF images are captured synchronously by terminals 210 with an external trigger.
  • the terminals 210 are configured to capture the H-LF images in multiple exposures with different camera exposure settings, so that the terminals 210 can capture more details before overexposure.
  • the data processing unit 300 includes a data preprocessing unit 310 , an H-LF stereo matching unit 320 and an H-LF data-cube reconstruction unit 330 .
  • FIG. 3 is a graphical representation illustrating a data processing unit 300 of an HLFI system in accordance to exemplary embodiments of the present disclosure.
  • the data preprocessing unit 310 may include an exposure effect normalization unit 311 for eliminating effects of different exposure settings, a spectral response elimination unit 312 for eliminating cameras' spectral responses, a distortion removal unit 313 for removing view distortions and an image rectifier 314 for rectifying the hyperspectral images.
  • the data preprocessing unit 310 outputs rectified and undistorted hyperspectral images to the H-LF stereo matching unit 320 .
  • the H-LF stereo matching unit 320 may include a correspondence cost unit 321 for measuring similarity of corresponding pixels, a defocus cost unit 322 for measuring unreliable color/intensity variance, and an energy minimization unit 323 for generating a disparity map of a reference view.
  • the H-LF stereo matching unit 320 outputs a disparity map of a reference view to the H-LF data-cube reconstruction unit 330 .
  • the H-LF data-cube reconstruction unit 330 may include an initial disparity estimation unit 331 for an initial disparity estimation, a pairwise stereo matching unit 332 for obtaining disparity maps between neighboring views, and an image registration unit 333 for generating complete hyperspectral data-cubes.
  • the data processing unit 300 outputs the complete hyperspectral data-cubes and transmits them to the data displaying unit 400 for display.
  • the camera in each multi-view hyperspectral data terminal captures a plurality of images with different exposure settings.
  • the exposure effect normalization unit 311 the effect caused by different exposure settings from each camera is eliminated.
  • the image that has the highest sensitivity and is not overexposed is chosen as an established data of the particular camera.
  • the established data of the camera is used for the rest of data processing.
  • each camera may capture 5 images with 5 different exposure settings.
  • For camera ⁇ i, j ⁇ , its 3 rd image has the highest sensitivity among the 5 images, and the 3 rd image is not overexposed. Then, the 3 rd image is chosen as the established data of camera ⁇ i, j ⁇ for the rest of the data processing.
  • the spectral response elimination unit 312 we use the normalized image data as input for processing to eliminate the effect of camera's spectral response but still make the image depend on the spectrum.
  • the images are warped in accordance to the distortion parameters to obtain images free of distortion.
  • digital filters are used for signal de-noising.
  • all images can be rectified with two rules: (1) performing horizontal alignments for images in the same row; (2) performing vertical alignments for images in the same column.
  • multi-view rectification is an ill-posed problem as opposed to a well-posed problem.
  • a well-posed problem is considered as a mathematical model of a physical phenomenon that has the following properties: 1. a solution exists; 2. the solution is unique; and 3. the solution's behavior changes continuously with initial conditions.
  • a nonlinear optimization method is applied to a sampling pattern which is selected to cover as much spatial position in a field of view as possible. Then a set of alignment parameters of all cameras can be obtained and used in rectifying the images.
  • the data preprocessing unit 310 outputs 5 ⁇ 6 preprocessed H-LF images which are rectified and undistorted.
  • the preprocessed H-LF images cover the spectrum range from 410 to 700 nm with 30 bands each 10 nm wide. Each of the 30 images captures a different view.
  • the rectified and undistorted hyperspectral images are input to the H-LF stereo matching unit 320 .
  • the output of the H-LF stereo matching is an optimal disparity map of a reference view.
  • the reference view is chosen as the image captured by the camera positioned at ⁇ 3 , 4 ⁇ in the terminal array.
  • the H-LF stereo matching unit 320 may include a correspondence cost unit 321 for measuring similarity of corresponding pixels, a defocus cost unit 322 for measuring unreliable color/intensity variance, and an energy minimization unit 323 for generating an optimal disparity map.
  • Stereo matching works by finding corresponding points in rectified images.
  • Approaches to the correspondence problem can be broadly classified into two categories: the intensity-based matching and the feature-based matching techniques.
  • the matching process is applied directly to the intensity profiles of the two images, while in the second, features are first extracted from the images and the matching process is applied to the features.
  • the H-LF stereo matching is quite different from traditional stereo matching due to the color/intensity inconsistency. Images at different spectra have very different appearances, and the tradition stereo matching methods fail to match the correspondence points.
  • To perform an accurate H-LF stereo matching a new type of feature descriptor that is applicable in different spectra is of desire.
  • a feature descriptor is a type of feature representation chosen to stand for a feature in image processing.
  • FIGS. 4A and 4B in accordance to exemplary embodiments of the present disclosure.
  • the preprocessed images are input to the H-LF stereo matching unit 320 .
  • the input consists 30 H-LF images, which covers the visible spectrum range from 410 nm to 700 nm with intervals of 10 nm.
  • both gradient magnitude and gradient direction of each image are obtained.
  • a correspondence cost is obtained for measuring appearance consistency.
  • a new spectrum-aware defocus cost for measuring the unreliable color/intensity variance is obtained.
  • an energy function is constructed by integrating the correspondence cost and the defocus cost are integrated with additional occlusion and smoothness terms.
  • an optimal disparity map of a reference view (positioned at ⁇ 3 , 4 ⁇ ) is obtained. The detailed method for H-LF stereo matching is discussed in the following.
  • a light ray can be mathematically formulated in ray space, and plane parameterization (2PP) is adopted for its simplicity.
  • 2PP plane parameterization
  • each ray is parameterized by its intersections with two parallel planes ⁇ uv and ⁇ st .
  • a light field includes an extra dimension, i.e., spectral dimension, and the 2PP representation can be modified as L(u, v, s, t, ⁇ ) to represent the hyperspectral light field, where (u, v) and (s, t) are used to represent the ray intersections with the aperture and the sensor planes respectively at wavelength A.
  • An image I(s, t, ⁇ i ) on (s, t) responding along a narrow bandpass spectral profile F ⁇ i ( ⁇ ) which is centered at wavelength ⁇ i can be formulated as:
  • I ( s, t , ⁇ i ) ⁇ L ( u, v, s, t , ⁇ ) A ( u, v ) C ( ⁇ ) ⁇ F ⁇ i ( ⁇ )cos 4 ⁇ d ⁇ dudv
  • S( ⁇ i ) is a latent radiance image at spectrum ⁇ i
  • C( ⁇ i ) is the spectral response function
  • I ⁇ ⁇ ( p ) I ⁇ ( p )
  • I _ S p ⁇ ( ⁇ )
  • is a mean intensity and S ( ⁇ ) is an average radiance in the image.
  • S ( ⁇ ) is an average radiance in the image.
  • the correspondence cost aims to measure the similarity of corresponding pixels. As previously discussed, for H-LF stereo matching, we first need to develop a spectral-invariant feature descriptor.
  • adjacent pixels refer to pixels in adjacent H-LF views.
  • M L (p) ⁇ M R (q) and ⁇ L (p) ⁇ R (q) the magnitude M(p) and direction 0 (p) should be approximately the same. i.e., M L (p) ⁇ M R (q) and ⁇ L (p) ⁇ R (q).
  • the pixel p may lie on a foreground and the pixel q may lie on a background.
  • the foreground pixel p and the background pixel q may correspond to an object with different spectral responses. Accordingly, the magnitude measure is no longer consistent, however, the directions should still be similar
  • a new spectral-invariant feature descriptor is developed to measure both edge and non-edge features.
  • the non-edge features can be described by coupling the magnitude and direction histograms whereas the edge features can be described with an extension of histogram of oriented gradient (HoG) which is called Overlapping HoG or O-HoG.
  • HoG histogram of oriented gradient
  • O-HoG Overlapping refers that each bin in O-HoG contains overlapping ranges of directions.
  • a histogram is a graphical representation of the distribution of numerical data. It is an estimate of the probability distribution of a continuous variable (quantitative variable). It is a kind of bar graph.
  • the first step is to “bin” the range of values—that is, divide the entire range of values into a series of intervals—and then count how many values fall into each interval.
  • the spectral-invariant feature descriptor is designed to enclose O-HoG. The detailed method for developing the spectral-invariant feature descriptor can be explained in the following discussion, and is illustrated in FIG. 4B in accordance to exemplary embodiments of the present disclosure.
  • the magnitude and direction histograms are calculated. Given a patch of pixels U(p, w) ⁇ w 2 ⁇ 2 centered at p with size w ⁇ w, weighted votes for bins of the magnitude histogram h i (p, w, K 1 ) and direction histogram h 2 (p, w, K 2 ) can be counted, where K 1 , K 2 are the total level of magnitude and direction bins respectively.
  • b i ( k ) ⁇ ( p , ⁇ ) ⁇ u t ⁇ U ⁇ ( p , ⁇ ) ⁇ ⁇ G ⁇ ( p , u t , ⁇ g ) ⁇ f ⁇ ( u t ) ⁇ j ⁇ [ 0 , K i - 1 ] ⁇ ⁇ b i ( j ) ( 1 )
  • the k-th bin b 3 (k) (p, w) can be computed as:
  • G(p, u t , ⁇ g ) exp( ⁇ p ⁇ u t ⁇ 2 2 /2 ⁇ g 2 ) is a spatial weight kernel, and f(u t ) is a truncation function as:
  • Equations (1)-(3) build completed feature descriptor terms for each pixel. Equations (1) and (2) are two independent parts from two attributes: the edge and non-edge features. Equation (3) serves both Equations (1) and (2) as a function term.
  • h 1 and h 2 represent non-edge features and h 3 represents edge features.
  • H p [D p T ( w 1 ), D p T (w 2 ), D p T ( w 3 )] T with K levels.
  • FIG. 5 is a graphical representation illustrating a structure of an exemplary spectral-invariant feature descriptor, in accordance to the embodiments of the present disclosure.
  • the next step is to compare the spectral-invariant feature descriptor in different views, i.e., measuring similarities in different H-LF images.
  • One commonly adopted similarity measurement algorithm in stereo matching is normalized cross correlation (NCC), and the corresponding correlation coefficient is shown as following:
  • ⁇ ⁇ ( I ) ⁇ u i ⁇ U L u j ⁇ U R ⁇ ( I L ⁇ ( u i ) - I _ L ) ⁇ ( I R ⁇ ( u j ) - I _ R ) ⁇ u i ⁇ U L ⁇ ( I L ⁇ ( u i ) - I _ L ) 2 ⁇ ⁇ u j ⁇ U R ⁇ ( I R ⁇ ( u j ) - I _ R ) 2 ( 4 )
  • ⁇ L , and ⁇ R are the mean values of U L (p, w) and U R (q, w) respectively in the domain I (e.g., intensity).
  • NCC is not directly applicable for matching multi-dimensional features.
  • the spectral-invariant feature descriptor H is multi-dimensional, however, each h (i) in H is independent of any other element h (j) (j ⁇ i) and represents a unique attribute of H (as shown in FIG. 5 ).
  • Equation (4) can be used with appropriate weight for each h (i) , and we can obtain a similarity metric as
  • w i is a similarity weight of h (i) . Since the value of h (i) can reflect the weight of the i-th histogram and the value of h (i) has been normalized to have a range of [ 0 , 1 ], h (i) can be used to substitute for w i . In addition, to suppress noise, the mean values ⁇ tilde over (h) ⁇ (i) are used instead of h (i) as the weights.
  • the similarity metric can adopt a bidirectional weighted normalized cross correlation (BWNCC), and we can get a final similarity metric shown as the following:
  • the forward component weighted by h p (i) represents the similarity between h p (i) and h q (i)
  • the backward components weighted by h q (i) represent the similarity between h q (i) and h p (i) .
  • the correspondence cost can be formulated. Given a hypothesis disparity f(p), the correspondence cost can be formulated by using the spectral-invariant feature descriptor and the similarity metric as:
  • a subset of views are selected for the disparity estimation.
  • Each H-LF image captured at a different positon includes a different LF view. All views can be denoted as ⁇ the reference view (positioned at ⁇ 3 , 4 ⁇ ) is denoted as (s o , t o ), and ⁇ * is a subset of ⁇ which is selected for the disparity estimation.
  • an initial disparity map of the reference view, f* c can be estimated based on the correspondence cost in the subset of selected views by separately treating occluding vs. non-occluding pixels by using an occlusion-aware depth estimation method from Ting-Chun Wang, et al. (T. Wang, et al., “Occlusion-aware Depth Estimation Using Light-field Cameras”, IEEE International Conference on Computer Vision , 2015).
  • a non-occluded pixel refers to a pixel that can be covered by all view rays in the scene if the rays are refocused to the correct depth.
  • an occluded pixel refers to a pixel that some view rays hit an occluder and are not able to cover the pixel.
  • the correspondence cost is to measure appearance consistency, while the defocus cost is to measure the unreliable color/intensity variance.
  • All pixels in the H-LF images are spectral-aware samplings, reflecting the values from different spectra for any single 3D point.
  • a new spectrum-aware defocus cost is developed at step 440 in the defocus cost unit 322 , as illustrated in FIG. 6 , consistent with exemplary embodiments of the present disclosure.
  • the RGB color of a pixel p in a canonical camera can be estimated.
  • a spectral profile of p as P p ( ⁇ ) can be formed by indexing ⁇ (s,t) using I p (s, t) into respective views. Then the spectral profile is used to synthesize the RGB value, as shown in FIG. 6 , block (b).
  • all the pixels reflecting the values of the point in different spectra can compose a curve, which is the spectral response of this 3D point.
  • the 30 spectrum bands cover the visible spectrum range from 410 nm to 700 nm with intervals of 10 nm.
  • the RGB color value can be mapped back to spectra ⁇ r in a CIE 1931 Color Space, as shown in FIG. 6 , block (c), by using a technique of a visible gamut with the RGB's hue from the CIE 1931 Color Space (T. Smith and J. Guild, “The C.I.E. colorimetric standards and their use”, Transactions of the Optical Society , vol. 33, 73.)
  • the CIE 1931 color spaces are the first defined quantitative links between physical pure colors (i.e. wavelengths) in the electromagnetic visible spectrum, and physiological perceived colors in human color vision.
  • the mathematical relationships that define these color spaces are essential tools for color management, important when dealing with color inks, illuminated displays, and recording devices such as digital cameras.
  • FIG. 6 block (c) shows a Gaussian distribution of the captured spectra of the sampled H-LF images. Since a correct disparity hypothesis results in an accurate estimation of the RGB value, the sampled H-LF images should have the captured spectra approximately form a Gaussian distribution centered at ⁇ r , with a probability density function as:
  • the probability density function can be normalized as:
  • a Kullback-Leibler divergence can be measured from P* p ( ⁇ ) to P g ( ⁇ ), as shown in FIG. 6 , block (d).
  • the Kullback-Leibler divergence is a measure of how one probability distribution diverges from a second expected probability distribution.
  • the comparison between the estimated spectra ⁇ r and the captured spectra of the sampled H-LF images indicates the level of focusness.
  • the defocus cost can be computed as:
  • an energy function with a Markov random field (MRF) on a hypothesis disparity f can be constructed as:
  • MRF refers to a Markov random field, Markov network or undirected graphical model. It is a set of random variables having a Markov property described by an undirected graph.
  • the binary term E binary (f) is an energy term for smoothness and occlusion, and is developed by Ting-Chun Wang, et al. (T. Wang, et al., “Occlusion-aware Depth Estimation Using Light-field Cameras”, IEEE International Conference on Computer Vision , 2015).
  • the unary term incorporates both the contributions from the correspondence cost and the defocus cost, and is defined as:
  • E unary ⁇ ( f ) ⁇ p ⁇ ⁇ ⁇ c ⁇ ⁇ C ⁇ ( f ⁇ ( p ) ) - C ⁇ ( f c * ⁇ ( p ) ) ⁇ + ⁇ D ⁇ ( f ⁇ ( p ) ) - D ⁇ ( f d * ⁇ ( p ) ) ⁇
  • the optimal disparity map f ⁇ for the reference view can be obtained.
  • the minimization of the energy function ( 7 ) can be solved by using a graph-cut algorithm. “Graph-cut” is a type of algorithm used to solve a variety of energy minimization problems which employ a max-flow/min-cut optimization.
  • the H-LF stereo matching method has the following advantages: 1. This method can accurately measure the similarity of correspondence points in images with different spectra. 2. It includes a new defocus cost to synthesize the RGB color from the sampled H-LF images and then use the CIE color Gamut to map the estimated hue of color to its spectral band and robustly measure its consistency with the spectra of the sampled H-LF images as the focusness measure.
  • FIG. 7 is a flow diagram illustrating a method for H-LF data-cube reconstruction, in accordance to exemplary embodiments of the present disclosure.
  • the optimal disparity map of the reference view is input to the H-LF data-cube reconstruction unit 330 to generate complete H-LF data-cubes.
  • the disparity map of the reference view is warped to individual H-LF images as an initial disparity estimation, generating a “prior” for each neighboring pair of images.
  • a pairwise stereo matching is performed to generate pairwise disparity maps , , by utilizing the “prior” at the pairwise stereo matching unit 332 .
  • step 704 we can map all pixels p on to qon according and register all images currently on to (reverse implement should use , to eliminate artifacts) at the image registration unit 333 . This process is iterated for all neighboring pairs, and the complete H-LF data-cubes can be obtained.
  • the complete H-LF data-cubes include all 5 ⁇ 6 views, and 5 ⁇ 6 ⁇ 30 images where 30 is the number of sampled spectrum bands.
  • the complete H-LF data-cubes cover the spectrum range from 410 to 700 nm with 30 bands each 10 nm wide.
  • the hyperspectral data-cube reconstruction method has the following advantages: 1. It overcomes the shortcoming of the brute-force approach.
  • the brute-force approach directly warps images to the rest of views by using the disparity map of the reference view.
  • the brute-force approach may cause substantial amount of holes due to occlusion and large baselines in the data capturing unit.
  • This method also fully exploits the properties of the light fields, compared to conducting pairwise stereo matching between all views.
  • the various modules, units, and components described above can be implemented as an Application Specific Integrated Circuit (ASIC); an electronic circuit; a combinational logic circuit; a field programmable gate array (FPGA); a processor (shared, dedicated, or group) that executes code; or other suitable hardware components that provide the described functionality.
  • ASIC Application Specific Integrated Circuit
  • FPGA field programmable gate array
  • the processor can be a microprocessor provided by from Intel, or a mainframe computer provided by IBM.
  • a “computer-readable medium” can be any medium that can contain or store the program for use by or in connection with the instruction execution system, apparatus, or device.
  • the computer readable medium can include, but is not limited to, an electronic, magnetic, optical, electromagnetic, infrared, or semiconductor system, apparatus or device, a portable computer diskette (magnetic), a random access memory (RAM) (magnetic), a read-only memory (ROM) (magnetic), an erasable programmable read-only memory (EPROM) (magnetic), a portable optical disc such a CD, CD-R, CD-RW, DVD, DVD-R, or DVD-RW, or flash memory such as compact flash cards, secured digital cards, USB memory devices, memory sticks, and the like.

Landscapes

  • Engineering & Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Spectroscopy & Molecular Physics (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Theoretical Computer Science (AREA)
  • Multimedia (AREA)
  • Signal Processing (AREA)
  • Image Analysis (AREA)

Abstract

A method for generating hyperspectral data-cubes based on a plurality of hyperspectral light field (H-LF) images is disclosed. Each H-LF image may have a different view and a different spectral band. The method may include calculating a magnitude histogram, a direction histogram, and an overlapping histogram of oriented gradient for a plurality of pixels; developing a spectral-invariant feature descriptor by combining the magnitude histogram, the direction histogram, and the overlapping histogram of oriented gradient; obtaining a correspondence cost of the H-LF images based on the spectral-invariable feature descriptor; performing H-LF stereo matching on the H-LF images to obtain a disparity map of a reference view; and generating hyperspectral data-cubes by using the disparity map of the reference view. A bin in the overlapping histogram of oriented gradient may comprise overlapping ranges of directions.

Description

    CROSS-REFERENCE To RELATED APPLICATION
  • This application is a continuation application of International Patent Application No. PCT/CN2017/087210, filed on Jun. 5, 2017 and entitled “METHOD AND SYSTEM FOR HYPERSPECTRAL LIGHT FIELD IMAGING.” The above-referenced application is incorporated herein by reference in its entirety.
  • TECHNICAL FIELD
  • The disclosure relates generally to hyperspectral light field imaging system for generating hyperspectral light field images, and more particularly, to method and system for generating complete hyperspectral data-cubes from captured hyperspectral light field images.
  • BACKGROUND
  • Hyperspectral light field (H-LF) imaging is part of a class of techniques commonly referred to as spectral imaging or spectral analysis. The H-LF imaging collects and processes information from across the electromagnetic spectrum. A hyperspectral camera/senor collects information as a set of H-LF images. Each image represents a narrow wavelength range of the electromagnetic spectrum, also known as a spectral band. These images are combined to form a hyperspectral data-cube for processing and analysis. The goal of H-LF imaging is to obtain the spectrum for each pixel in the image of a scene with narrow spectral bands over a continuous spectral range. Every pixel in the image thus can be used to characterize the objects in the scene with great precision and detail.
  • The H-LF images provide much more detailed information about the scene than a normal color camera. The H-LF imaging leads to a vastly improved ability to classify the objects in the scene based on their spectral properties. It can also take advantages of the spatial relationships among the different neighboring spectra, allowing more elaborate spectral-spatial models for a more accurate segmentation and classification of the image.
  • In this disclosure, we present a new method for generating complete hyperspectral data-cubes from captured H-LF images.
  • SUMMARY
  • One aspect of the present disclosure is directed to a method for generating hyperspectral data-cubes based on a plurality of hyperspectral light field (H-LF) images. Each H-LF image may have a different view and a different spectral band. The method may include calculating a magnitude histogram, a direction histogram, and an overlapping histogram of oriented gradient for a plurality of pixels; developing a spectral-invariant feature descriptor by combining the magnitude histogram, the direction histogram, and the overlapping histogram of oriented gradient; obtaining a correspondence cost of the H-LF images based on the spectral-invariable feature descriptor; performing H-LF stereo matching on the H-LF images to obtain a disparity map of a reference view; and generating hyperspectral data-cubes by using the disparity map of the reference view. A bin in the overlapping histogram of oriented gradient may comprise overlapping ranges of directions.
  • Another aspect of the present disclosure is directed to a method for generating hyperspectral data-cubes based on a plurality of hyperspectral light field (H-LF) images. Each H-LF image may have a different view and a different spectral band. The method may include obtaining an estimated spectra by synthesizing RGB color values from the H-LF images and mapping the RGB color values to a spectral band; obtaining a defocus cost of the H-LF images based on a comparison between estimated spectra and captured spectra of the H-LF images; performing H-LF stereo matching on the H-LF images to obtain a disparity map of a reference view; and generating hyperspectral data-cubes by using the disparity map of the reference view.
  • Another aspect of the present disclosure is directed to a hyperspectral light field imaging (HLFI) system for capturing hyperspectral light field (H-LF) images to generate hyperspectral data-cubes. The system may include a plurality of multi-view hyperspectral data terminals and a data processing unit. The terminals may be configured to capture a plurality of H-LF images, and each image may have a view and a different spectral band. The data processing unit may be configured to calculate a magnitude histogram, a direction histogram, and an overlapping histogram of oriented gradient for a plurality of pixels; develop a spectral-invariant feature descriptor by combining the magnitude histogram, the direction histogram, and the overlapping histogram of oriented gradient; obtain a correspondence cost of the H-LF images based on the spectral-invariable feature descriptor; perform H-LF stereo matching on the H-LF images to obtain a disparity map of a reference view; and generate hyperspectral data-cubes by using the disparity map of the reference view.
  • It is to be understood that the foregoing general description and the following detailed description are exemplary and explanatory only, and are not restrictive of the invention, as claimed.
  • BRIEF DESCRIPTION OF THE DRAWINGS
  • The accompanying drawings, which constitute a part of this disclosure, illustrate several non-limiting embodiments and, together with the description, serve to explain the disclosed principles.
  • FIG. 1 is a schematic diagram for a hyperspectral light field imaging (HLFI) system, consistent with exemplary embodiments of the present disclosure.
  • FIGS. 2A and 2B are graphical representations illustrating a data capturing unit of an HLFI system, consistent with exemplary embodiments of the present disclosure.
  • FIG. 3 is a graphical representation illustrating a data processing unit of an HLFI system, consistent with exemplary embodiments of the present disclosure.
  • FIG. 4A is a diagram illustrating a method for H-LF stereo matching, consistent with exemplary embodiments of the present disclosure.
  • FIG. 4B is a flow diagram illustrating a method for estimating a disparity map based on a correspondence cost, consistent with exemplary embodiments of the present disclosure.
  • FIG. 5 is a graphical representation illustrating a structure of a feature descriptor, consistent with exemplary embodiments of the present disclosure.
  • FIG. 6 illustrates a method for obtaining a defocus cost, consistent with exemplary embodiments of the present disclosure.
  • FIG. 7 is a flow diagram illustrating a method for H-LF data-cube reconstruction, consistent with exemplary embodiments of the present disclosure.
  • DETAILED DESCRIPTION OF THE EMBODIMENTS
  • Reference will now be made in detail to exemplary embodiments, examples of which are illustrated in the accompanying drawings. The following description refers to the accompanying drawings in which the same numbers in different drawings represent the same or similar elements unless otherwise represented. The implementations set forth in the following description of exemplary embodiments consistent with the present invention do not represent all implementations consistent with the invention. Instead, they are merely examples of systems and methods consistent with aspects related to the invention.
  • In accordance to embodiments of the present disclosure, a hyperspectral light field imaging (HLFI) system including a data capturing unit, a data processing unit and a data displaying unit is provided. The data capturing unit captures a plurality of multi-view H-LF images, which are the sampled H-LF images. The sampled H-LF images are transmitted to the data processing unit. The data processing unit preprocess the sampled H-LF images to obtain rectified and undistorted H-LF images, and performs H-LF stereo matching to obtain a disparity map of a reference view. Then the data processing unit generates complete hyperspectral data-cubes based on the disparity map. The complete hyperspectral data-cubes are transmitted to the data displaying unit for displaying.
  • In some embodiments, the sampled H-LF images are 5×6 H-LF images. Each image is captured at a different view, and samples a different spectrum range with a bandwidth of 10 nm. The total spectra of the 30 H-LF images cover the whole visible spectrum band from 410 nm to 700 nm. In some embodiments, the complete H-LF data-cubes include all 5×6 views and 5×6×30 images, where 30 is the number of sampled spectrum bands. The complete H-LF data-cubes cover the spectrum range from 410 to 700 nm with 30 bands each 10 nm wide.
  • 1. System Overview
  • FIG. 1 shows an exemplary HLFI system 100 in accordance to an embodiment of the present disclosure. The system 100 may include a number of components, some of which may be optional. In some embodiments, the system 100 may include many more components than those shown in FIG. 1. However, it is not necessary that all of these components be shown in order to disclose an illustrative embodiment.
  • As shown in FIG. 1, the system 100 may include a data capturing unit 200, a data processing unit 300 and a data displaying unit 400. The data capturing unit 200 may include a plurality of multi-view hyperspectral data terminals 210, and a camera calibration unit 220. The data processing unit 300 may include a data preprocessing unit 310, an H-LF stereo matching unit 320, and an H-LF data-cube reconstruction unit 330. The data displaying unit 400 may include an H-LF dynamic refocusing unit 401 and a complete H-LF data-cube unit 402.
  • 2. Data Capturing Unit
  • As shown in FIG. 1, the data capture unit 200 may include a plurality of multi-view hyperspectral data terminals 210, and a camera calibration unit 220.
  • 2.1 Hyperspectral Data Terminals
  • FIGS. 2A and 2B are graphical representations illustrating the plurality of multi-view hyperspectral data terminals 210, in accordance to an embodiment of the present disclosure. In some embodiments, the terminals 210 may be placed in a rectangular array with 5 rows and 6 columns. Positions of each terminal 210 can be denoted as {i, j}, where i stands for the row number, and j stands for the column number. As shown in FIG. 2A, the terminals 210 in each row are equally spaced on an identical plate. All rows may be uniformly fixed on a common limit bracket which ensures that each terminal can have same baselines with its neighboring terminals within the limits of installation accuracy and techniques. A baseline is a distance between adjacent terminals/cameras in pixel, usually measured in unit of mm.
  • Each of the terminals 210 may consist of a monochrome camera and a narrow bandpass optical filter. Each filter may be centered on a specific wavelength with a bandwidth of 10 nm, and the minimum wavelength is 410 nm and the maximum wavelength is 700 nm. All 5×6 filters can sample the whole visible spectrum band from 410 nm to 700 nm, with intervals of 10 nm. In other words, the terminals 210 may capture 5×6 H-LF images. Each image samples a different spectrum range with a bandwidth of 10 nm, and the total spectra of the 30 H-LF images cover the whole visible spectrum band from 410 nm to 700 nm. These 30 H-LF images are sampled H-LF images for generating complete hyperspectral data-cubes in this disclosure. Table 1 lists parameters of hyperspectral data terminal array.
  • TABLE 1
    Sensors
    Sensor Type CCD
    Sensor Size
    1/3″ inch
    Pixel Size 3.75 μm × 3.75 μm
    Resolution 1292 × 964
    Frame Rate 30 fps
    Image Data Format Mono8
    Data Interface Gigabit Ethernet (1000M bit/s)
    Synchronous Mode External/Software Trigger
    Lens
    Focal Length 8 mm
    Aperture Range (F-Stop) F1.4-F16C
    Filters
    Range of wavelength 410 nm-700 nm
    Steps 10 nm
    Number of band 30 bands
  • 2.2 Multi-Camera Calibration
  • Cameras in the multi-view hyperspectral data terminals 210 may include intrinsic, extrinsic and distortion parameters. The intrinsic parameters refer to the parameters to link pixel coordinates of an image point with the corresponding coordinates in a camera reference. The extrinsic parameters may define a location and orientation of the camera reference frame with respect to a known world reference frame. The distortion parameters may be caused due to the limitation of lens production technology and imaging models. The intrinsic, extrinsic and distortion parameters are used to rectify the captured hyperspectral images.
  • At the camera calibration unit 220, the intrinsic, extrinsic and distortion parameters are obtained. In some embodiments, a camera calibration method from Zhengyou Zhang (Z. Zhang, “A flexible new technique for camera calibration”, IEEE Trans. Pattern Anal. Mach. Intell., 22(11):1330-1334, 2000.) may be applied to obtain above-mentioned parameters. To obtain the intrinsic and distortion parameters, all cameras are first refocused on a same plane with a similar aperture condition, then Zhengyou Zhang's method is applied with the use of Matlab platform. The extrinsic parameters can be obtained by applying Zhengyou Zhang's method to obtain parameters of rotation and translation between each neighboring pair of cameras.
  • Based on the intrinsic, extrinsic and distortion parameters, a multi-view rectification method for the sampled H-LF images can be developed. Since all cameras' views cover largely the same area, the cameras do not need to be calibrated one by one. Instead, we can calibrate the cameras by guaranteeing a checkerboard in all cameras' views simultaneously when capturing images synchronously. A “checkerboard” is a geometric pattern commonly used in camera calibration. In addition, due to different spectral responses of the filters on the cameras, a suited exposure setting is determined to ensure all cameras can capture the contrast of the pattern of the checkerboard simultaneously.
  • 2.3. H-LF Image Capturing Method
  • The plurality of multi-view hyperspectral data terminals 210 capture H-LF images and send them to the data processing unit 300 via an IP network. Due to the size of the hyperspectral data and the limitation of transmission bandwidth, the data processing unit 300 may not obtain all data simultaneously. In some embodiments, software may be developed to ensure that the data of the captured H-LF images from all cameras (5×6) can be transmitted synchronously to the data processing unit 300. A data buffer queue may be built on each camera, and all the data are transmitted with a time-shared mode, so that the H-LF images are captured synchronously by terminals 210 with an external trigger. In addition, due to different spectral responses of filters on the cameras, for each scene, the terminals 210 are configured to capture the H-LF images in multiple exposures with different camera exposure settings, so that the terminals 210 can capture more details before overexposure.
  • 3. Data Processing Unit
  • As shown in FIG. 1, the data processing unit 300 includes a data preprocessing unit 310, an H-LF stereo matching unit 320 and an H-LF data-cube reconstruction unit 330.
  • FIG. 3 is a graphical representation illustrating a data processing unit 300 of an HLFI system in accordance to exemplary embodiments of the present disclosure. As shown in FIG. 3, the data preprocessing unit 310 may include an exposure effect normalization unit 311 for eliminating effects of different exposure settings, a spectral response elimination unit 312 for eliminating cameras' spectral responses, a distortion removal unit 313 for removing view distortions and an image rectifier 314 for rectifying the hyperspectral images. The data preprocessing unit 310 outputs rectified and undistorted hyperspectral images to the H-LF stereo matching unit 320.
  • The H-LF stereo matching unit 320 may include a correspondence cost unit 321 for measuring similarity of corresponding pixels, a defocus cost unit 322 for measuring unreliable color/intensity variance, and an energy minimization unit 323 for generating a disparity map of a reference view. The H-LF stereo matching unit 320 outputs a disparity map of a reference view to the H-LF data-cube reconstruction unit 330.
  • The H-LF data-cube reconstruction unit 330 may include an initial disparity estimation unit 331 for an initial disparity estimation, a pairwise stereo matching unit 332 for obtaining disparity maps between neighboring views, and an image registration unit 333 for generating complete hyperspectral data-cubes. The data processing unit 300 outputs the complete hyperspectral data-cubes and transmits them to the data displaying unit 400 for display.
  • 3.1 Image Preprocessing
  • During data capturing, the camera in each multi-view hyperspectral data terminal captures a plurality of images with different exposure settings. At the exposure effect normalization unit 311, the effect caused by different exposure settings from each camera is eliminated. Among the plurality of images captured by each camera with different exposure settings, by measuring the quality of images, the image that has the highest sensitivity and is not overexposed is chosen as an established data of the particular camera. The established data of the camera is used for the rest of data processing. For example, each camera may capture 5 images with 5 different exposure settings. For camera {i, j}, its 3rd image has the highest sensitivity among the 5 images, and the 3rd image is not overexposed. Then, the 3rd image is chosen as the established data of camera {i, j} for the rest of the data processing.
  • Then at the spectral response elimination unit 312, we use the normalized image data as input for processing to eliminate the effect of camera's spectral response but still make the image depend on the spectrum.
  • Next, at the distortion removal unit 313, the images are warped in accordance to the distortion parameters to obtain images free of distortion. To improve accuracy, digital filters are used for signal de-noising.
  • Finally, at the image rectifier 314, by using the intrinsic and extrinsic parameters obtained from the camera calibration unit 220, all images can be rectified with two rules: (1) performing horizontal alignments for images in the same row; (2) performing vertical alignments for images in the same column. In computer vision theory, multi-view rectification is an ill-posed problem as opposed to a well-posed problem. A well-posed problem is considered as a mathematical model of a physical phenomenon that has the following properties: 1. a solution exists; 2. the solution is unique; and 3. the solution's behavior changes continuously with initial conditions. To solve the multi-view rectification problem, a nonlinear optimization method is applied to a sampling pattern which is selected to cover as much spatial position in a field of view as possible. Then a set of alignment parameters of all cameras can be obtained and used in rectifying the images.
  • The data preprocessing unit 310 outputs 5×6 preprocessed H-LF images which are rectified and undistorted. The preprocessed H-LF images cover the spectrum range from 410 to 700 nm with 30 bands each 10 nm wide. Each of the 30 images captures a different view.
  • 3.2 H-LF Stereo Matching
  • The rectified and undistorted hyperspectral images are input to the H-LF stereo matching unit 320. The output of the H-LF stereo matching is an optimal disparity map of a reference view. In this disclosure, the reference view is chosen as the image captured by the camera positioned at {3, 4} in the terminal array. As shown in FIG. 3, the H-LF stereo matching unit 320 may include a correspondence cost unit 321 for measuring similarity of corresponding pixels, a defocus cost unit 322 for measuring unreliable color/intensity variance, and an energy minimization unit 323 for generating an optimal disparity map.
  • Stereo matching works by finding corresponding points in rectified images. Approaches to the correspondence problem can be broadly classified into two categories: the intensity-based matching and the feature-based matching techniques. In the first category, the matching process is applied directly to the intensity profiles of the two images, while in the second, features are first extracted from the images and the matching process is applied to the features. The H-LF stereo matching is quite different from traditional stereo matching due to the color/intensity inconsistency. Images at different spectra have very different appearances, and the tradition stereo matching methods fail to match the correspondence points. To perform an accurate H-LF stereo matching, a new type of feature descriptor that is applicable in different spectra is of desire. A feature descriptor is a type of feature representation chosen to stand for a feature in image processing. In this disclosure, a new method for H-LF stereo matching with a new spectral-invariant feature descriptor is presented and shown in FIGS. 4A and 4B in accordance to exemplary embodiments of the present disclosure.
  • As shown in FIG. 4A, at step 410, the preprocessed images are input to the H-LF stereo matching unit 320. The input consists 30 H-LF images, which covers the visible spectrum range from 410 nm to 700 nm with intervals of 10 nm. At step 420, both gradient magnitude and gradient direction of each image are obtained. At step 430, a correspondence cost is obtained for measuring appearance consistency. At step 440, a new spectrum-aware defocus cost for measuring the unreliable color/intensity variance is obtained. At step 450, an energy function is constructed by integrating the correspondence cost and the defocus cost are integrated with additional occlusion and smoothness terms. Finally, at step 460, an optimal disparity map of a reference view (positioned at {3, 4}) is obtained. The detailed method for H-LF stereo matching is discussed in the following.
  • 3.2.1 H-LF Image Formulation
  • A light ray can be mathematically formulated in ray space, and plane parameterization (2PP) is adopted for its simplicity. In 2PP, each ray is parameterized by its intersections with two parallel planes Πuv and Πst. A light field includes an extra dimension, i.e., spectral dimension, and the 2PP representation can be modified as L(u, v, s, t, λ) to represent the hyperspectral light field, where (u, v) and (s, t) are used to represent the ray intersections with the aperture and the sensor planes respectively at wavelength A. An image I(s, t, λi) on (s, t) responding along a narrow bandpass spectral profile Fλ i (λ) which is centered at wavelength λi can be formulated as:

  • I(s, t, λi)=∫∫∫L(u, v, s, t, λ)A(u, v)C(λ)·F λ i (λ)cos4θdλdudv
  • where A(u, v) θ is the aperture function, and θ is an incident angle of the ray, C(λ) is the camera's spectral response function. The spectral profile Fλ i (λ) can be approximated by a Dirac delta function Fλ i and cos4θ can be ignored with a paraxial approximation. Thus, the above formulation can be simplified as:

  • I(s, t, λi)=Ci)∫∫L(u, v, s, t, λi)A(u, v)dudv=Ci)S(s, t, λi)
  • where S(λi) is a latent radiance image at spectrum λi, and C(λi) is the spectral response function.
  • 3.2.2 Preprocessing
  • As previously discussed, at the spectral response elimination unit 312, the effect caused by the cameras' spectral responses are eliminated, and this step can also be presented in a mathematical formulation. Considering a pixel p∈N2, for an image I(p)=C(λ)Sp(λ) as presented in the above equation, it can be normalized as:
  • I ~ ( p ) = I ( p ) I _ = S p ( λ ) S _ ( λ )
  • where Ī is a mean intensity and S(λ) is an average radiance in the image. We can use Ĩ(p) as the input data for data processing, so that the effect from cameras' spectral responses can be eliminated, while the images still depend on the spectrum.
  • In addition, the gradient magnitude and gradient direction of each H-LF image can be calculated, and represented as: the gradient magnitude: M(p)=√xĨ(p)2+∇yĨ(p)2 and the gradient direction: Θ(p)=atan(∇yĨ(p)/∇xĨ(p))). Both the gradient magnitude and the gradient direction are utilized to obtain the correspondence cost.
  • 3.2.3 Correspondence Cost
  • The correspondence cost aims to measure the similarity of corresponding pixels. As previously discussed, for H-LF stereo matching, we first need to develop a spectral-invariant feature descriptor.
  • Considering two pixels p, q and their corresponding adjacent pixels p′, q′, if p, q and p′, q′ lie inside a same object, their relation satisfies:

  • |{tilde over (I)}L(p)−{tilde over (I)}L(p′)|≅|{tilde over (I)}R(q)−{tilde over (I)}R(q′)|
  • Here, adjacent pixels refer to pixels in adjacent H-LF views. This relation indicates the magnitude M(p) and direction 0(p) should be approximately the same. i.e., ML(p)≅MR(q) and ΘL(p)≅R(q). On the contrary, if the pixels lie near an edge, the pixel p may lie on a foreground and the pixel q may lie on a background. The foreground pixel p and the background pixel q may correspond to an object with different spectral responses. Accordingly, the magnitude measure is no longer consistent, however, the directions should still be similar In other words, when pixels p and q lie near an edge, ML(p)≅MR(q) and ΘL(p)≠ΘR(q).
  • Therefore, a new spectral-invariant feature descriptor is developed to measure both edge and non-edge features. Specifically, the non-edge features can be described by coupling the magnitude and direction histograms whereas the edge features can be described with an extension of histogram of oriented gradient (HoG) which is called Overlapping HoG or O-HoG. Overlapping refers that each bin in O-HoG contains overlapping ranges of directions. A histogram is a graphical representation of the distribution of numerical data. It is an estimate of the probability distribution of a continuous variable (quantitative variable). It is a kind of bar graph. To construct a histogram, the first step is to “bin” the range of values—that is, divide the entire range of values into a series of intervals—and then count how many values fall into each interval. As for the HLFI system, any slight change in perspective or spectrum may lead to a misalignment in HoG, whereas O-HoG is much more robust in handling view and spectral variations. Therefore, the spectral-invariant feature descriptor is designed to enclose O-HoG. The detailed method for developing the spectral-invariant feature descriptor can be explained in the following discussion, and is illustrated in FIG. 4B in accordance to exemplary embodiments of the present disclosure.
  • At step 431, the magnitude and direction histograms are calculated. Given a patch of pixels U(p, w)∈
    Figure US20200141804A1-20200507-P00001
    w 2 ×2 centered at p with size w×w, weighted votes for bins of the magnitude histogram hi(p, w, K1) and direction histogram h2(p, w, K2) can be counted, where K1, K2 are the total level of magnitude and direction bins respectively. The k-th bin bi (k)(p, w) of hi(i=1,2; k∈[0, Ki−1)) can be constructed as:
  • b i ( k ) ( p , ω ) = u t U ( p , ω ) G ( p , u t , σ g ) f ( u t ) j [ 0 , K i - 1 ] b i ( j ) ( 1 )
  • Similarly, for the O-HoG term h3(p, w, K3), the k-th bin b3 (k)(p, w) can be computed as:
  • b 3 ( k ) ( p , ω ) = u t U ( p , ω ) G ( p , u t , σ g ) M ( u t ) f ( u t ) j [ 0 , K i - 1 ] b 3 ( j ) ( 2 )
  • where G(p, ut, σg)=exp(−∥p−ut2 2/2σg 2) is a spatial weight kernel, and f(ut) is a truncation function as:
  • f ( u t ) = { 1 Q ( u t ) [ k ( 1 - o ) s , k ( 1 - o ) s + s ) 0 else ( 3 )
  • Here o is the overlapping portion between the neighboring bins and s is the width of bin. In h1, Q(ut)=M(ut), otherwise, Q(ut)=Θ(ut). Equations (1)-(3) build completed feature descriptor terms for each pixel. Equations (1) and (2) are two independent parts from two attributes: the edge and non-edge features. Equation (3) serves both Equations (1) and (2) as a function term.
  • At step 432, all three histograms can be combined by using a weight matrix α=[α1, α2, α3]T. As mentioned above, h1 and h2 represent non-edge features and h3 represents edge features. Since M(p) intrinsically represents the edge strength of p, M(p) can be reused to compute: α12=βexp(−M2(p)/σw) and α3=1−α1−α2 with β∈(0, ½]. The descriptor is formulated as Dp=[α1h1 T, α2h2 T, α3h3 T]T. To further improve robustness, a pyramid structure can be built with a different patch w=[w1, w2, w3]T to obtain the desired spectral-invariant feature descriptor as following:

  • H p =[D p T(w 1), D p T(w2), D p T(w 3)]T with K levels.
  • FIG. 5 is a graphical representation illustrating a structure of an exemplary spectral-invariant feature descriptor, in accordance to the embodiments of the present disclosure.
  • Having obtained the spectral-invariant feature descriptor, the next step is to compare the spectral-invariant feature descriptor in different views, i.e., measuring similarities in different H-LF images. One commonly adopted similarity measurement algorithm in stereo matching is normalized cross correlation (NCC), and the corresponding correlation coefficient is shown as following:
  • ξ ( I ) = u i U L u j U R ( I L ( u i ) - I _ L ) ( I R ( u j ) - I _ R ) u i U L ( I L ( u i ) - I _ L ) 2 u j U R ( I R ( u j ) - I _ R ) 2 ( 4 )
  • where ĪL, and ĪR are the mean values of UL(p, w) and UR(q, w) respectively in the domain I (e.g., intensity). However, NCC is not directly applicable for matching multi-dimensional features. The spectral-invariant feature descriptor H is multi-dimensional, however, each h(i) in H is independent of any other element h(j)(j≠i) and represents a unique attribute of H (as shown in FIG. 5).
  • At step 433, a similarity metric for matching multi-dimensional features can be developed. Equation (4) can be used with appropriate weight for each h(i), and we can obtain a similarity metric as
  • ξ ( H ) = i = 0 K - 1 ω i ξ ( h ( i ) ) .
  • Here, wi is a similarity weight of h(i). Since the value of h(i) can reflect the weight of the i-th histogram and the value of h(i) has been normalized to have a range of [0, 1], h(i) can be used to substitute for wi. In addition, to suppress noise, the mean values {tilde over (h)}(i) are used instead of h(i) as the weights.
  • Moreover, since hp (i) and hq (i) play equally important roles in computing ξ(H), the similarity metric can adopt a bidirectional weighted normalized cross correlation (BWNCC), and we can get a final similarity metric shown as the following:
  • ξ ncc ( h ) = ( i = 0 K - 1 ξ ( h ( i ) ) h _ p ( i ) j = 0 K - 1 ξ ( h ( j ) ) h _ q ( j ) ) 0.5
  • The forward component weighted by h p (i) represents the similarity between hp (i) and hq (i), and the backward components weighted by h q (i) represent the similarity between hq (i) and hp (i).
  • At step 434, the correspondence cost can be formulated. Given a hypothesis disparity f(p), the correspondence cost can be formulated by using the spectral-invariant feature descriptor and the similarity metric as:
  • C ( p , f ( p ) ) = 1 Ω * ( s , t ) Ω * - log ( ξ ncc ( H ) )
  • At step 435, a subset of views are selected for the disparity estimation. Each H-LF image captured at a different positon includes a different LF view. All views can be denoted as Ω the reference view (positioned at {3,4}) is denoted as (so, to), and Ω* is a subset of Ω which is selected for the disparity estimation.
  • Instead of matching p in (so, to) with pixel q in all LF views according to a hypothesis disparity map f(p), we only performing the matching in a selected subset of views Ω* with a coherent appearance. To select a proper subset of views, we first compute the mean gradient magnitude of q in all views, denoted as M(q). Next, we determine if p's gradient magnitude, denoted as M(p) is above or below the mean gradient magnitude of q. If M(p)>M(q), then it implies that P is an edge pixel. Then we only select the views in which pixels q has a higher gradient magnitude than M(q). On the other hand, if M(p)<M(q), it implies p is a non-edge point. Then we only select the views in which pixels q with a lower gradient magnitude than M(q) for the disparity estimation.
  • At step 436, an initial disparity map of the reference view, f*c can be estimated based on the correspondence cost in the subset of selected views by separately treating occluding vs. non-occluding pixels by using an occlusion-aware depth estimation method from Ting-Chun Wang, et al. (T. Wang, et al., “Occlusion-aware Depth Estimation Using Light-field Cameras”, IEEE International Conference on Computer Vision, 2015). A non-occluded pixel refers to a pixel that can be covered by all view rays in the scene if the rays are refocused to the correct depth. On the contrary, an occluded pixel refers to a pixel that some view rays hit an occluder and are not able to cover the pixel.
  • If P is non-occluding, we have f*c(p)=minf{C}. If p is occluding, we can partition Ω* into an occluder vs. an occluded region, denoted as Ω*1 and Ω*2. Then the correspondence cost C1 and C2 for the respective Ω*1 and Ω*2 can be computed by using Ting-Chun Wang's method, and an initial disparity map can be obtained as: f*c(p)=minf{C1, C2} for an occluding pixel. At the end of the step 430, the correspondence cost unit 321 outputs an estimated disparity map based on the correspondence cost.
  • 3.2.4. Defocus Cost
  • The correspondence cost is to measure appearance consistency, while the defocus cost is to measure the unreliable color/intensity variance. All pixels in the H-LF images are spectral-aware samplings, reflecting the values from different spectra for any single 3D point. To address the effect of the spectra, a new spectrum-aware defocus cost is developed at step 440 in the defocus cost unit 322, as illustrated in FIG. 6, consistent with exemplary embodiments of the present disclosure.
  • Given a hypothesis disparity f(p), the RGB color of a pixel p in a canonical camera can be estimated. As shown in FIG. 6, block (a), a spectral profile of p as Pp(λ) can be formed by indexing λ(s,t) using Ip(s, t) into respective views. Then the spectral profile is used to synthesize the RGB value, as shown in FIG. 6, block (b). Given a 3D point, all the pixels reflecting the values of the point in different spectra can compose a curve, which is the spectral response of this 3D point. In some embodiments, a spectral response function of a PTGrey FL3-U3-20E4C-C camera, P3(λ)=[Pr(λ), Pg(λ), Pb(λ)]T, can be used to obtain the RGB value by integrating Pp(s,t)) with Pc(s,t)) over the respective bandwidths. In some embodiments, the 30 spectrum bands cover the visible spectrum range from 410 nm to 700 nm with intervals of 10 nm.
  • Then, the RGB color value can be mapped back to spectra λr in a CIE 1931 Color Space, as shown in FIG. 6, block (c), by using a technique of a visible gamut with the RGB's hue from the CIE 1931 Color Space (T. Smith and J. Guild, “The C.I.E. colorimetric standards and their use”, Transactions of the Optical Society, vol. 33, 73.) The CIE 1931 color spaces are the first defined quantitative links between physical pure colors (i.e. wavelengths) in the electromagnetic visible spectrum, and physiological perceived colors in human color vision. The mathematical relationships that define these color spaces are essential tools for color management, important when dealing with color inks, illuminated displays, and recording devices such as digital cameras.
  • FIG. 6, block (c) shows a Gaussian distribution of the captured spectra of the sampled H-LF images. Since a correct disparity hypothesis results in an accurate estimation of the RGB value, the sampled H-LF images should have the captured spectra approximately form a Gaussian distribution centered at λr, with a probability density function as:

  • P g(λ)=1/σd·exp(−(λ−λr)2/2σd 2)   (5)
  • The probability density function can be normalized as:

  • P p(λ) to P* p(λ)=P p(λ)/Σ(s,t)∈Ω P p(s,t)).
  • In addition, a Kullback-Leibler divergence can be measured from P*p(λ) to Pg(λ), as shown in FIG. 6, block (d). The Kullback-Leibler divergence is a measure of how one probability distribution diverges from a second expected probability distribution. The comparison between the estimated spectra λr and the captured spectra of the sampled H-LF images indicates the level of focusness. The defocus cost can be computed as:
  • D ( p , f ( p ) ) = ( s , t ) Ω P g ( λ ( s , t ) ) log P g ( λ ( s , t ) ) P p * ( λ ( s , t ) ) ( 6 )
  • In some embodiments, the value of σd is selected to guarantee Pg(λ) to have at least 30% of response in bordering of visible spectrum (i.e., 410 nm or 700 nm). For example, if λr=550 nm, we may select σd=96.5. At the end of step 440, the defocus cost unit 322 outputs another estimated disparity map based on the defocus cost: f*d(p)=minf{D}.
  • 3.2.5 Energy Minimization
  • The obtained the estimated disparity maps from both the correspondence cost and the defocus cost are transmitted to the energy minimization unit 323 to obtain an optimal disparity map of the reference view. At step 450, an energy function with a Markov random field (MRF) on a hypothesis disparity f can be constructed as:

  • E(f)=E unary(f)+E binary(f)   (7)
  • Here, MRF refers to a Markov random field, Markov network or undirected graphical model. It is a set of random variables having a Markov property described by an undirected graph. The binary term Ebinary(f) is an energy term for smoothness and occlusion, and is developed by Ting-Chun Wang, et al. (T. Wang, et al., “Occlusion-aware Depth Estimation Using Light-field Cameras”, IEEE International Conference on Computer Vision, 2015). The unary term incorporates both the contributions from the correspondence cost and the defocus cost, and is defined as:
  • E unary ( f ) = p γ c C ( f ( p ) ) - C ( f c * ( p ) ) + D ( f ( p ) ) - D ( f d * ( p ) )
  • where γc adjusts the weights between correspondence and defocus cost.
  • At step 460, by minimize the energy function (7), the optimal disparity map f for the reference view can be obtained. The minimization of the energy function (7) can be solved by using a graph-cut algorithm. “Graph-cut” is a type of algorithm used to solve a variety of energy minimization problems which employ a max-flow/min-cut optimization.
  • The H-LF stereo matching method has the following advantages: 1. This method can accurately measure the similarity of correspondence points in images with different spectra. 2. It includes a new defocus cost to synthesize the RGB color from the sampled H-LF images and then use the CIE color Gamut to map the estimated hue of color to its spectral band and robustly measure its consistency with the spectra of the sampled H-LF images as the focusness measure.
  • 3.3 Hyperspectral Data-Cube Reconstruction
  • FIG. 7 is a flow diagram illustrating a method for H-LF data-cube reconstruction, in accordance to exemplary embodiments of the present disclosure. At step 701, the optimal disparity map of the reference view is input to the H-LF data-cube reconstruction unit 330 to generate complete H-LF data-cubes. At step 702, in the initial disparity estimation unit 331, the disparity map of the reference view is warped to individual H-LF images as an initial disparity estimation, generating a “prior” for each neighboring pair of images. At step 703, for each neighboring pair of images, a pairwise stereo matching is performed to generate pairwise disparity maps
    Figure US20200141804A1-20200507-P00002
    ,
    Figure US20200141804A1-20200507-P00003
    , by utilizing the “prior” at the pairwise stereo matching unit 332. At step 704, we can map all pixels p on
    Figure US20200141804A1-20200507-P00004
    to qon
    Figure US20200141804A1-20200507-P00005
    according
    Figure US20200141804A1-20200507-P00002
    and register all images currently on
    Figure US20200141804A1-20200507-P00005
    to
    Figure US20200141804A1-20200507-P00006
    (reverse implement should use
    Figure US20200141804A1-20200507-P00007
    , to eliminate artifacts) at the image registration unit 333. This process is iterated for all neighboring pairs, and the complete H-LF data-cubes can be obtained. The complete H-LF data-cubes include all 5×6 views, and 5×6×30 images where 30 is the number of sampled spectrum bands. The complete H-LF data-cubes cover the spectrum range from 410 to 700 nm with 30 bands each 10 nm wide.
  • The hyperspectral data-cube reconstruction method has the following advantages: 1. It overcomes the shortcoming of the brute-force approach. The brute-force approach directly warps images to the rest of views by using the disparity map of the reference view. The brute-force approach may cause substantial amount of holes due to occlusion and large baselines in the data capturing unit. 2. This method also fully exploits the properties of the light fields, compared to conducting pairwise stereo matching between all views.
  • The various modules, units, and components described above can be implemented as an Application Specific Integrated Circuit (ASIC); an electronic circuit; a combinational logic circuit; a field programmable gate array (FPGA); a processor (shared, dedicated, or group) that executes code; or other suitable hardware components that provide the described functionality. The processor can be a microprocessor provided by from Intel, or a mainframe computer provided by IBM.
  • Note that one or more of the functions described above can be performed by software or firmware stored in memory and executed by a processor, or stored in program storage and executed by a processor. The software or firmware can also be stored and/or transported within any computer-readable medium for use by or in connection with an instruction execution system, apparatus, or device, such as a computer-based system, processor-containing system, or other system that can fetch the instructions from the instruction execution system, apparatus, or device and execute the instructions. In the context of this document, a “computer-readable medium” can be any medium that can contain or store the program for use by or in connection with the instruction execution system, apparatus, or device. The computer readable medium can include, but is not limited to, an electronic, magnetic, optical, electromagnetic, infrared, or semiconductor system, apparatus or device, a portable computer diskette (magnetic), a random access memory (RAM) (magnetic), a read-only memory (ROM) (magnetic), an erasable programmable read-only memory (EPROM) (magnetic), a portable optical disc such a CD, CD-R, CD-RW, DVD, DVD-R, or DVD-RW, or flash memory such as compact flash cards, secured digital cards, USB memory devices, memory sticks, and the like.
  • The invention described and claimed herein is not to be limited in scope by the specific preferred embodiments disclosed herein, as these embodiments are intended as illustrations of several aspects of the invention. Indeed, various modifications of the invention in addition to those shown and described herein will become apparent to those skilled in the art from the foregoing description. Such modifications are also intended to fall within the scope of the appended claims.

Claims (23)

What is claimed is:
1. A method of generating hyperspectral data-cubes based on a plurality of hyperspectral light field (H-LF) images, wherein each H-LF image comprises a different view and a different spectral band, the method comprising:
calculating a magnitude histogram, a direction histogram, and an overlapping histogram of oriented gradient for a plurality of pixels, wherein a bin in the overlapping histogram of oriented gradient comprises overlapping ranges of directions;
developing a spectral-invariant feature descriptor by combining the magnitude histogram, the direction histogram, and the overlapping histogram of oriented gradient;
obtaining a correspondence cost of the H-LF images based on the spectral-invariable feature descriptor;
performing H-LF stereo matching on the H-LF images to obtain a disparity map of a reference view; and
generating hyperspectral data-cubes by using the disparity map of the reference view.
2. The method of claim 1, wherein the spectral-invariant feature descriptor measures both edge and non-edge features.
3. The method of claim 2, wherein the non-edge feature is described by coupling the magnitude histogram and the direction histogram, and the edge feature is described by the overlapping histogram of oriented gradient.
4. The method of claim 3, further comprising:
combining the magnitude histogram, the direction histogram, and the overlapping histogram of oriented gradient using a weight matrix and a pyramid structure to form the spectral invariant feature descriptor.
5. The method of claim 1, further comprising:
comparing the spectral-invariant feature descriptor to develop a similarity metric using a bidirectional weighted normalized cross correlation (BWNCC).
6. The method of claim 1, further comprising:
obtaining an estimated spectra by synthesizing RGB color values from the H-LF images and mapping the RGB color values to a spectral band; and
obtaining a defocus cost of the H-LF images based on a comparison between estimated spectra and captured spectra of the H-LF images.
7. The method of claim 6, wherein a Kullback-Leibler divergence is used in the comparison between the estimated spectra and the captured spectra of the H-LF images.
8. The method of claim 6, further comprising:
performing H-LF stereo matching on the H-LF images to obtain the disparity map of the reference view based on the correspondence cost and the defocus cost.
9. The method of claim 1, further comprising:
estimating an initial disparity map of the reference view based on the correspondence cost by treating occluding and non-occluding pixels separately.
10. The method of claim 1, further comprising:
capturing the plurality of H-LF images by a plurality of multi-view hyperspectral data terminals placed in a rectangular array comprising a plurality of rows and a plurality of columns, wherein each terminal consists a monochrome camera and a narrow bandpass optical filter.
11. A method of generating hyperspectral data-cubes based on a plurality of hyperspectral light field (H-LF) images, wherein each H-LF image comprises a different view and a different spectral band, the method comprising:
obtaining an estimated spectra by synthesizing RGB color values from the H-LF images and mapping the RGB color values to a spectral band;
obtaining a defocus cost of the H-LF images based on a comparison between estimated spectra and captured spectra of the H-LF images;
performing H-LF stereo matching on the H-LF images to obtain a disparity map of a reference view; and
generating hyperspectral data-cubes by using the disparity map of the reference view.
12. The method of claim 11, wherein a Kullback-Leibler divergence is used in the comparison between the estimated spectra and the captured spectra of the H-LF images.
13. The method of claim 12, further comprising:
performing H-LF stereo matching on the H-LF images to obtain the disparity map of the reference view based on the defocus cost.
14. The method of claim 13, further comprising:
calculating a magnitude histogram, a direction histogram, and an overlapping histogram of oriented gradient for the plurality of pixels, wherein a bin in the overlapping histogram of oriented gradient comprises overlapping ranges of directions;
developing a spectral-invariant feature descriptor by combining the magnitude histogram, the direction histogram, and the overlapping histogram of oriented gradient; and
obtaining a correspondence cost of the H-LF images based on the spectral-invariable feature descriptor.
15. The method of claim 14, wherein the spectral-invariant feature descriptor measures both edge and non-edge features.
16. The method of claim 15, wherein the non-edge feature is described by coupling the magnitude histogram and the direction histogram, and the edge feature is described by the overlapping histogram of oriented gradient.
17. The method of claim 16, further comprising:
combining the magnitude histogram, the direction histogram, and the overlapping histogram of oriented gradient using a weight matrix and a pyramid structure to form the spectral invariant feature descriptor.
18. The method of claim 13, further comprising:
comparing the spectral-invariant feature descriptor to develop a similarity metric using a bidirectional weighted normalized cross correlation (BWNCC).
19. The method of claim 11, further comprising:
estimating an initial disparity map of the reference view based on the correspondence cost by treating occluding and non-occluding pixels separately.
20. The method of claim 11, further comprising:
capturing the plurality of H-LF images by a plurality of multi-view hyperspectral data terminals placed in a rectangular array comprising a plurality of rows and a plurality of columns, wherein each terminal consists a monochrome camera and a narrow bandpass optical filter.
21. A hyperspectral light field imaging (HLFI) system for capturing hyperspectral light field (H-LF) images to generate hyperspectral data-cubes, the system comprising:
a plurality of multi-view hyperspectral data terminals configured to capture a plurality of H-LF images, wherein each H-LF image comprises a different view and a different spectral band; and
a data processing unit configured to:
calculate a magnitude histogram, a direction histogram, and an overlapping histogram of oriented gradient for a plurality of pixels, wherein a bin in the overlapping histogram of oriented gradient comprises overlapping ranges of directions;
develop a spectral-invariant feature descriptor by combining the magnitude histogram, the direction histogram, and the overlapping histogram of oriented gradient;
obtain a correspondence cost of the H-LF images based on the spectral-invariable feature descriptor;
perform H-LF stereo matching on the H-LF images to obtain a disparity map of a reference view; and
generate hyperspectral data-cubes by using the disparity map of the reference view.
22. The system of claim 21, wherein the data processing unit is configured to compare the spectral-invariant feature descriptor to develop a similarity metric using a bidirectional weighted normalized cross correlation (BWNCC).
23. The system of claim 21, wherein the data processing unit is configured to:
obtain an estimated spectra by synthesizing RGB color values from the H-LF images and mapping the RGB color values to a spectral band; and
obtain a defocus cost of the H-LF images based on a comparison between estimated spectra and captured spectra of the H-LF images.
US16/678,104 2017-06-05 2019-11-08 Method and system for hyperspectral light field imaging Active US10641658B1 (en)

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
PCT/CN2017/087210 WO2018223267A1 (en) 2017-06-05 2017-06-05 Method and system for hyperspectral light field imaging

Related Parent Applications (1)

Application Number Title Priority Date Filing Date
PCT/CN2017/087210 Continuation WO2018223267A1 (en) 2017-06-05 2017-06-05 Method and system for hyperspectral light field imaging

Publications (2)

Publication Number Publication Date
US10641658B1 US10641658B1 (en) 2020-05-05
US20200141804A1 true US20200141804A1 (en) 2020-05-07

Family

ID=64566769

Family Applications (1)

Application Number Title Priority Date Filing Date
US16/678,104 Active US10641658B1 (en) 2017-06-05 2019-11-08 Method and system for hyperspectral light field imaging

Country Status (3)

Country Link
US (1) US10641658B1 (en)
CN (1) CN110520768B (en)
WO (1) WO2018223267A1 (en)

Families Citing this family (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US10878590B2 (en) * 2018-05-25 2020-12-29 Microsoft Technology Licensing, Llc Fusing disparity proposals in stereo matching
US11222229B1 (en) 2018-05-31 2022-01-11 The Charles Stark Draper Laboratory, Inc. System and method for multidimensional gradient-based cross-spectral stereo matching
US11532155B1 (en) * 2019-07-09 2022-12-20 ACME Atronomatic, LLC Methods and devices for earth remote sensing using stereoscopic hyperspectral imaging in the visible (VIS) and infrared (IR) bands
CN111709901B (en) * 2020-05-22 2023-04-28 哈尔滨工业大学 FCM cluster matching and Wallis filtering-based no-weight multi/hyperspectral remote sensing image color homogenizing method
CN115728236A (en) * 2022-11-21 2023-03-03 山东大学 Hyperspectral image acquisition and processing system and working method thereof
CN116894794B (en) * 2023-09-11 2023-11-21 长沙超创电子科技有限公司 Quick denoising method for video

Family Cites Families (11)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US6438261B1 (en) * 1998-09-03 2002-08-20 Green Vision Systems Ltd. Method of in-situ focus-fusion multi-layer spectral imaging and analysis of particulate samples
JP5249149B2 (en) * 2009-07-17 2013-07-31 富士フイルム株式会社 Stereoscopic image recording apparatus and method, stereoscopic image output apparatus and method, and stereoscopic image recording and output system
US9191649B2 (en) * 2011-08-12 2015-11-17 Qualcomm Incorporated Systems and methods to capture a stereoscopic image pair
US11510600B2 (en) * 2012-01-04 2022-11-29 The Trustees Of Dartmouth College Method and apparatus for quantitative and depth resolved hyperspectral fluorescence and reflectance imaging for surgical guidance
US9766382B2 (en) * 2012-06-05 2017-09-19 Hypermed Imaging, Inc. Single-sensor hyperspectral imaging device
CN102879098B (en) * 2012-09-19 2014-05-14 南京理工大学 Rapid interference imaging method and rapid interference imaging device for narrow-band hyperspectrum
US10013764B2 (en) * 2014-06-19 2018-07-03 Qualcomm Incorporated Local adaptive histogram equalization
US10511787B2 (en) * 2015-02-12 2019-12-17 Fraunhofer-Gesellschaft Zur Foerderung Der Angewandten Forschung E.V. Light-field camera
CN105424186B (en) * 2015-11-04 2017-06-27 北京航空航天大学 A kind of spectrum calibration bearing calibration of optical field imaging spectrometer
CN105488771B (en) * 2015-12-16 2018-08-28 清华大学 Light field image edit methods and device
CN106022391A (en) * 2016-05-31 2016-10-12 哈尔滨工业大学深圳研究生院 Hyperspectral image characteristic parallel extraction and classification method

Also Published As

Publication number Publication date
US10641658B1 (en) 2020-05-05
CN110520768B (en) 2021-10-22
CN110520768A (en) 2019-11-29
WO2018223267A1 (en) 2018-12-13

Similar Documents

Publication Publication Date Title
US10641658B1 (en) Method and system for hyperspectral light field imaging
US8548257B2 (en) Distinguishing between faces and non-faces
CN109410207B (en) NCC (non-return control) feature-based unmanned aerial vehicle line inspection image transmission line detection method
Tominaga Spectral imaging by a multichannel camera
Finlayson et al. Gamut constrained illuminant estimation
Lin et al. Determining the radiometric response function from a single grayscale image
US8189050B1 (en) Filtering systems and methods for infrared image processing
US8483480B2 (en) Method and system for factoring an illumination image
US9396411B2 (en) Method and system for generating intrinsic images using a single reflectance technique
US10645281B1 (en) Method and system for snapshot multi-spectral light field imaging
KR20190035476A (en) Hyperspectral Imaging Reconstruction Method Using Prism and System Therefor
van Zwanenberg et al. Edge detection techniques for quantifying spatial imaging system performance and image quality
WO2022247840A1 (en) Light source spectrum and multispectral reflectivity image acquisition methods and apparatuses, and electronic device
Aytekin et al. A data set for camera-independent color constancy
CN102543793A (en) Wafer focusing image quality feedback system and method therefor
US9754155B2 (en) Method and system for generating intrinsic images using a single reflectance technique
Koren Measuring camera shannon information capacity with a siemens star image
WO2021195817A1 (en) Method for extracting spectral information of object to be detected
WO2007004864A1 (en) Method and apparatus for visual object recognition
Llanos et al. Simultaneous demosaicing and chromatic aberration correction through spectral reconstruction
Mustaniemi et al. Parallax correction via disparity estimation in a multi-aperture camera
van Zwanenberg et al. Camera system performance derived from natural scenes
Wadhokar et al. SSIM technique for comparison of images
WO2021045140A1 (en) Information processing device, pulse wave measurement system, and pulse wave measurement program
WO2021220444A1 (en) Skin evaluation coefficient learning device, skin evaluation index estimation device, skin evaluation coefficient learning method, skin evaluation index estimation method, focus value acquisition method, and skin smoothness acquisition method

Legal Events

Date Code Title Description
FEPP Fee payment procedure

Free format text: ENTITY STATUS SET TO UNDISCOUNTED (ORIGINAL EVENT CODE: BIG.); ENTITY STATUS OF PATENT OWNER: SMALL ENTITY

FEPP Fee payment procedure

Free format text: ENTITY STATUS SET TO SMALL (ORIGINAL EVENT CODE: SMAL); ENTITY STATUS OF PATENT OWNER: SMALL ENTITY

AS Assignment

Owner name: SHANGHAITECH UNIVERSITY, CHINA

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:YU, JINGYI;REEL/FRAME:051309/0762

Effective date: 20191210

STCF Information on status: patent grant

Free format text: PATENTED CASE

MAFP Maintenance fee payment

Free format text: PAYMENT OF MAINTENANCE FEE, 4TH YR, SMALL ENTITY (ORIGINAL EVENT CODE: M2551); ENTITY STATUS OF PATENT OWNER: SMALL ENTITY

Year of fee payment: 4