US20150003738A1 - Adaptive quality image reconstruction via a compressed sensing framework - Google Patents

Adaptive quality image reconstruction via a compressed sensing framework Download PDF

Info

Publication number
US20150003738A1
US20150003738A1 US14/085,143 US201314085143A US2015003738A1 US 20150003738 A1 US20150003738 A1 US 20150003738A1 US 201314085143 A US201314085143 A US 201314085143A US 2015003738 A1 US2015003738 A1 US 2015003738A1
Authority
US
United States
Prior art keywords
rois
functions
scene
sampling functions
sampling
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
US14/085,143
Other versions
US9025883B2 (en
Inventor
Edgar A. Bernal
Xuejin Wen
Wencheng Wu
Lalit Keshav MESTHA
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Xerox Corp
Original Assignee
Xerox Corp
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Priority claimed from US13/932,791 external-priority patent/US9070218B2/en
Application filed by Xerox Corp filed Critical Xerox Corp
Priority to US14/085,143 priority Critical patent/US9025883B2/en
Assigned to XEROX CORPORATION reassignment XEROX CORPORATION ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: BERNAL, EDGAR A., MESTHA, LALIT KESHAV, WEN, XUEJIN, WU, WENCHENG
Publication of US20150003738A1 publication Critical patent/US20150003738A1/en
Application granted granted Critical
Publication of US9025883B2 publication Critical patent/US9025883B2/en
Expired - Fee Related legal-status Critical Current
Adjusted expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/70Arrangements for image or video recognition or understanding using pattern recognition or machine learning
    • G06V10/77Processing image or video features in feature spaces; using data integration or data reduction, e.g. principal component analysis [PCA] or independent component analysis [ICA] or self-organising maps [SOM]; Blind source separation
    • G06V10/7715Feature extraction, e.g. by transforming the feature space, e.g. multi-dimensional scaling [MDS]; Mappings, e.g. subspace methods
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/21Design or setup of recognition systems or techniques; Extraction of features in feature space; Blind source separation
    • G06F18/213Feature extraction, e.g. by transforming the feature space; Summarisation; Mappings, e.g. subspace methods
    • G06F18/2136Feature extraction, e.g. by transforming the feature space; Summarisation; Mappings, e.g. subspace methods based on sparsity criteria, e.g. with an overcomplete basis
    • G06K9/3233
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T11/002D [Two Dimensional] image generation
    • G06T11/003Reconstruction from projections, e.g. tomography
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/10Image acquisition
    • G06V10/12Details of acquisition arrangements; Constructional details thereof
    • G06V10/14Optical characteristics of the device performing the acquisition or on the illumination arrangements
    • G06V10/147Details of sensors, e.g. sensor lenses
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/20Image preprocessing
    • G06V10/22Image preprocessing by selection of a specific region containing or referencing a pattern; Locating or processing of specific regions to guide the detection or recognition
    • HELECTRICITY
    • H03ELECTRONIC CIRCUITRY
    • H03MCODING; DECODING; CODE CONVERSION IN GENERAL
    • H03M7/00Conversion of a code where information is represented by a given sequence or number of digits to a code where the same, similar or subset of information is represented by a different sequence or number of digits
    • H03M7/30Compression; Expansion; Suppression of unnecessary data, e.g. redundancy reduction
    • H03M7/3059Digital compression and data reduction techniques where the original information is represented by a subset or similar information, e.g. lossy compression
    • H03M7/3062Compressive sampling or sensing
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/60Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using transform coding

Definitions

  • the present invention is directed to systems and methods which use a compressed sensing framework to reconstruct an image of a scene wherein pixel locations associated with identified regions of interest in that scene have a higher image quality when rendered relative to other pixels of the reconstructed image.
  • Compressed sensing is a relatively new area in the signal processing art where one measures a small number of non-adaptive linear combinations of a signal. These measurements are usually much smaller than the number of samples that define the signal. From the small numbers of measurements, the signal is reconstructed by a non-linear process which aims to reduce the overall complexity required by a large variety of measurement systems by introducing signal compression into the measurement process.
  • the theory behind compressed sensing is that sparse signal statistics can be recovered from a small number of measurements. The sparse nature of most signals of interest allows high-fidelity reconstructions to be made using a compressed sensing approach.
  • Compressed sensing can be beneficial because it reduces the number of samples required to spatially and/or temporally reconstruct a given scene thereby enabling the use of inexpensive sensors with reduced spatial and/or temporal resolution in certain applications where complex sensors are otherwise used, while maintaining the quality of the reconstructed image of the scene.
  • Compressed sensing also holds an advantage in fidelity over conventional camera systems due to detector noise issues that may affect measurements due to the limited radiometric efficiency of pixels in two-dimensional sensors.
  • the present application provides a desirable extension to the prior compressed sensing system taught by Bernal et al.
  • What is disclosed is a system and method which reconstructs an N-pixel image of a scene such that Q pixel locations associated with identified regions of interest in a scene have a higher image quality when rendered relative to other pixels in the image.
  • Acquisition and adaptive-quality compression are performed simultaneously by semi-synchronously or asynchronously modulating temporally an ordered set of sampling functions, used to spatially modulate light incoming from the scene.
  • the teachings hereof improve compression efficiency of a compressed sensing framework. Encoding efficiency is improved with respect to traditional compressed sensing techniques.
  • bold Greek letters refer to sets of functions of a given length and to matrices formed by stacking rows, where each row is a function in the set.
  • the present method for reconstructing an N-pixel image of a scene captured using a compressed sensing device involves performing the following.
  • a mask is received which identifies at least one region of interest (ROI) in a scene having Q pixels, where Q ⁇ N.
  • the mask can be a binary image wherein pixels with value ‘1’ (ON pixels) indicate locations associated with the ROI and pixels with value ‘0’ (OFF pixels) indicate locations not associated with the ROI. Consequently, the mask has at least Q ON pixels, and at most N ⁇ Q OFF pixels.
  • the mask can also be multi-level, each level corresponding to one identified ROI, each identified ROI potentially having a different quality of reconstruction assigned to it.
  • the set of sampling functions is then partitioned into K non-overlapping and non-empty subsets of sampling functions ⁇ 1 , . . . , ⁇ M1 ⁇ , ⁇ M1+1 , . . . , ⁇ M1+M2 ⁇ , . . . , ⁇ M ⁇ Mk+1 , . . .
  • M i ⁇ N matrices ⁇ 1 , ⁇ 2 , . . . , ⁇ K .
  • Each matrix has at least Q linearly independent columns and at most N ⁇ Q linearly dependent columns.
  • An index of the linearly independent columns is associated with locations corresponding to the ROIs, and an index of the linearly dependent columns is associated with locations which do not correspond to the ROIs.
  • incoming light is modulated by a spatial light modulator according to a plurality of spatial patterns corresponding to the ordered sampling functions.
  • the light reflected/transmitted off the modulator is focused onto a detector of a compressed sensing device.
  • the detector proceeds to sample sequential measurements of the light focused thereon.
  • the measurements comprise a sequence of projection coefficients corresponding to a scene.
  • x denotes an N-dimensional vector representing an N-pixel sampled version of a scene.
  • a spatial appearance of the scene is reconstructed from the sequence of projection coefficients after M inner products have been sampled.
  • the reconstructed image being such that pixels associated with the ROIs have a higher image quality when rendered relative to other pixels in the image.
  • FIG. 1 illustrates one example embodiment of the present method for reconstructing an image of a scene
  • FIG. 2 shows one example system for performing scene reconstruction in accordance with the teachings hereof
  • FIG. 3 shows the magnitude of the Haar coefficients of an image x used in the simulation of the present method in decreasing order, from largest to smallest;
  • FIG. 4 is a table of results illustrating quality of reconstruction as measured by the mean squared error (MSE) from having reconstructed an image with varying number of samples using the present method, as well as the traditional compressed sensing approach.
  • MSE mean squared error
  • What is disclosed is a system and method which reconstructs an N-pixel image of a scene such that Q pixel locations associated with identified regions of interest in a scene have a higher image quality when rendered relative to other pixels in the image.
  • a “region of interest” is an identified area of a scene intended to be reconstructed using the teachings disclosed herein with a higher quality relative to other areas of that scene. What defines a particular region of interest will largely depend on the application where the present invention finds its uses.
  • the image of the scene from which a region of interest is first identified can be acquired with the compressed sensing device following the traditional compressed sensing framework. Alternatively, it can be acquired with a traditional imaging device having a significantly similar view point of the scene relative to the compressed sensing device.
  • the image of the scene can be received from an existing database of images. A region of interest is then identified from the image of the scene.
  • Regions of interest can be identified by processing an image of the scene to identify using any of: pixel classification, object identification, facial recognition, color, texture, spatial features, spectral information, pattern recognition, motion detection, foreground detection, and a user input.
  • the location, shape or size of the identified region of interest can change over time. Once a region of interest has been identified, changes in the location of the identified region of interest can be identified by tracking techniques. Alternatively, re-identification of the region of interest pixel classification, object identification, facial recognition, color, texture, spatial features, spectral information, pattern recognition, motion detection, foreground detection, and a user input can be performed periodically.
  • a “photodetector” or simply “detector” is a device which measures a magnitude of an intensity of light focused thereon.
  • the photodetector can be a single (diode) detector or a multi-diode detector and may further comprise an analog-to-digital converter and an amplifier.
  • a “Spatial Light Modulator (SLM)” is a device in the compressed sensing device positioned along an optical axis where a camera's focal plane array would typically be located.
  • the SLM is controllable such that it can be configured according to spatial patterns which can be used to modulate incoming light which can then be transmitted or reflected onto a photodetector of a compressed sensing device.
  • a compressed sensing device relies on modulating incoming light from the scene by a spatial light modulator according to a plurality of spatial patterns.
  • Examples of spatial light modulators include, but are not limited to Digital Micromirror Devices, Transmissive Liquid Crystals and Liquid Crystals on Silicon.
  • DMD Digital Micromirror Device
  • MEMS optical micro-electromechanical
  • the microscopic mirrors are electronically controllable and thus modulate incoming light by toggling a reflectivity thereof by individually tilting (or rotating) the mirrors in one direction or another to achieve an ON/OFF state.
  • ON state light is reflected in a desired direction, such as through a lens or onto a photodetector.
  • the OFF state the light is directed elsewhere.
  • the positive (+) state is ON and the negative ( ⁇ ) state is OFF.
  • each DMD pattern is configured to select a defined portion of the incoming light onto a detector.
  • a series of unique patterns are sequentially provided to the DMD and a series of measurements are obtained.
  • Light energy is reflected by the DMD mirrors onto the photo diode or photoreceptor where the photons of the image are converted to an electrical signal.
  • Each signal, produced as a result of each measurement is a function of a specific pattern and of the scene.
  • an image reconstruction algorithm reconstructs the original image from the generated time-series measurement data with knowledge of the temporal sequence of patterns.
  • DMDs are available from vendors in various streams of commerce.
  • LCOS Liquid Crystal on Silicon
  • a reflective “Liquid Crystal on Silicon (LCOS)” refers to a micro-projection or micro-display technology which uses liquid crystals instead of individual mirrors.
  • liquid crystals are applied directly to the surface of a silicon chip coated with an aluminized layer with some type of passivation layer, which is highly reflective.
  • LCOS technology is preferable in many applications because it can produce higher resolution and higher contrast images than standard liquid crystal technologies.
  • a “compressed sensing framework” is a signal processing technique for reconstructing a signal with solutions found by taking advantage of the signal's sparseness or compressibility in some domain, thereby enabling the entire signal to be generated from relatively few measurements.
  • An underdetermined linear system has more unknowns than equations and generally has an infinite number of solutions.
  • constraints are applied. Because many signals are sparse, i.e., they contain many coefficients close to or equal to zero when represented in some domain, the additional constraint of sparsity allows only those solutions with a small number of non-zero coefficients to be considered as feasible. Not all underdetermined systems have a sparse solution. However, if there is a unique sparse representation to that underdetermined linear system then a compressed sensing framework enables a recovery of that solution.
  • Two functions are said to be “largely correlated” if there exists a real number T 1 ′ such that
  • a level of correlation between two functions is indicative of a level of co-linearity: the larger the correlation between two functions, the more collinear they are.
  • T 2 is usually application-dependent and is expressed in terms of N and of the range of values of the functions ⁇ i and ⁇ j in ⁇ and ⁇ , P.
  • N the range of values of the functions ⁇ i and ⁇ j in ⁇ and ⁇ , P.
  • FIG. 1 illustrates one example embodiment of the present method for reconstructing an image of a scene captured using a compressed sensing device with the teachings hereof wherein Q pixel locations associated with identified regions of interest in the scene have a higher image quality when rendered relative to other pixels in the image.
  • Flow processing begins at step 100 and immediately proceeds to step 102 .
  • the mask can be dynamically updated in response to a new region of interest having been identified, a location of a region of interest changing in a scene, or a user input.
  • the mask may be received from a remote device over a network via a wired or wireless pathway, or retrieved from a storage device such as a memory or a hard drive.
  • the mask can be a binary image wherein pixels with value ‘1’ (ON pixels) indicate locations associated with the ROI and pixels with value ‘0’ (OFF pixels) indicate locations not associated with the ROI.
  • the mask can also be multi-level, each level corresponding to one identified ROI.
  • the index m is indicative of an ordering of the function in a sampling sequence. The ordering is such that, given a block length B, partitions of the set ⁇ into non-overlapping and non-empty subsets define sub-matrices ⁇ 1 , ⁇ 2 , . . .
  • ⁇ ⁇ M/B ⁇ all except possibly one of size B ⁇ N (where ⁇ denotes the ceiling operator which maps a real number to a smallest following integer) corresponding to sequences of functions arranged in subsets all except possibly one of length B, ⁇ 1 , . . . , ⁇ B ⁇ , ⁇ B+1 , . . . , ⁇ 2B ⁇ , . . . , ⁇ B ⁇ M/B ⁇ B+1 , . . . , ⁇ M ⁇ , respectively, where B ⁇ M is the block length.
  • Each resulting sub-matrix has at least Q linearly independent columns and at most N ⁇ Q dependent columns. It should be appreciated that the word “possibly” is used here because either all are of size B ⁇ N or all but one are of size B ⁇ N.
  • the index of the linearly independent columns is determined to be associated with the locations corresponding to the identified regions of interest, and the index of the linearly dependent columns is determined to be associated with the locations not corresponding to the identified regions of interest.
  • the value of B determines the difference in the quality of reconstruction of the ROI vs. non-ROI regions, larger values of B corresponding to larger differences.
  • Typical values of B range from 2 to 8 although use of values outside that range may be used.
  • the ordering can be such that, for sequences of sampling functions arranged in subsets all except possibly one of length B, subsets of sampling functions ⁇ 1 , . . . , ⁇ B ⁇ , ⁇ B+1 , . . . , ⁇ 2B ⁇ , ⁇ B ⁇ M/B ⁇ B+1 , . . . , ⁇ M ⁇ are largely incoherent, sampling functions ⁇ i and ⁇ j in a given subset are largely uncorrelated for every i ⁇ j only for pixels within ROIs as identified by the mask, and sampling functions ⁇ i and ⁇ j in a given subset are largely or even completely correlated for pixels outside the ROIs.
  • blocks ⁇ 1 , . . . , ⁇ B ⁇ , ⁇ B+1 , . . . , ⁇ 2B ⁇ , . . . , ⁇ B ⁇ M/B ⁇ B+1 , . . . , ⁇ M ⁇ are largely incoherent, and functions ⁇ i and ⁇ j in a given subset are largely or fully correlated.
  • a similar sampling function design strategy can be implemented in cases when partitions of the set of sampling functions ⁇ into K non-overlapping and non-empty subsets of functions ⁇ 1 , ⁇ 2 , . . . , ⁇ K , each subset having M i elements are used.
  • is a set of M i -block-wise largely correlated functions, where blocks ⁇ 1 , . . . , ⁇ M1 ⁇ , ⁇ M1+1 , . . .
  • a series of measurements comprises a sequence of projection coefficients corresponding to the inner product between that N-pixel sampled version of the scene and each of the sampling functions. An appearance of the scene is reconstructed from the sequence of projection coefficients after M inner products have been sampled, where M ⁇ N, such that pixel locations associated with ROIs in the scene have a higher image quality when rendered relative to other pixels in the image.
  • step 106 configure a spatial light modulator to modulate incoming light according to spatial patterns corresponding to the set of ordered sampling functions.
  • the measurements comprise a sequence of projection coefficients corresponding to the scene.
  • step 110 reconstruct a spatial appearance of an image of the scene from the sequence of projection coefficients after M inner products have been sampled.
  • the reconstruction is such that pixels associated with the localized ROIs in the reconstructed image have a higher image quality, that more faithfully represents the appearance of the sampled scene, when rendered relative to other pixels in the image.
  • This non-uniform quality of the reconstruction is enabled by the sampling scheme described above. Thereafter, in this embodiment, further processing stops.
  • FIG. 2 shows one example system for performing scene reconstruction in accordance with the teachings hereof.
  • incoming light enters the compressed sensing system 200 through an aperture 202 and into a spatial light modulator (SLM) 203 which modulates the incoming light to produce a spatial pattern of light 204 which is focused on to detector 205 .
  • SLM spatial light modulator
  • the detector measures a magnitude of an intensity of the spatial pattern focused thereon.
  • Mask module 208 receives a mask using USB port 209 , and provides the mask to a controller 207 shown comprising a processor (CPU) and a memory.
  • the controller facilitates the configuration of the spatial light modulator to modulate incoming light.
  • the detector outputs sequential measurements 206 which may be provided to USB port 215 .
  • Measurements obtained by the detector are communicated to image reconstruction module 213 wherein a spatial appearance of the scene is reconstructed from the measurements.
  • the measurements and the reconstructed image 214 are communicated to storage device 216 and/or provided as output to workstation 220 .
  • Values, data, measurements, and results of any of the modules and processing units of the system 200 may be obtained or retrieved via communications bus 217 .
  • Shown in communication with the system 200 is a workstation 220 .
  • the workstation is shown comprising a monitor 221 , a keyboard 222 , a mouse 223 , a storage device 224 , and a computer-readable media 225 .
  • the workstation is also placed in communication with one or more remote devices over network 226 using, for example, a network card.
  • a user thereof may change or control the functionality of any of the modules and processing units comprising the system 200 using the workstation.
  • An image of a scene can be displayed on the monitor and corrected and/or cropped.
  • Masks can be generated using the workstation and communicated to the mask module 208 . Measurements and values generated by the system may be displayed on the display device.
  • Intensity values obtained by the detector may be modified by a user of the workstation.
  • the values of the spatial pattern that controls the modulation of incoming light may further be communicated to the workstation and displayed on the monitor.
  • a user can selectively identify regions of interest using, for example, a mouse.
  • the user may further define the ordering of the set of sampling functions using the workstation. Localized regions of interest can be communicated to the mask module by the workstation.
  • An operator of the workstation may modify the results generated by any of the modules or processing units of FIG. 2 as needed and/or re-direct the modified results back to the same or different modules for further processing or re-processing.
  • the workstation has an operating system and other specialized software configured to display a variety of numeric values, text, scroll bars, pull-down menus with user selectable options, and the like, for entering, selecting, or modifying information displayed on the display device.
  • modules may designate one or more components which may, in turn, comprise software and/or hardware designed to perform the intended function.
  • a plurality of modules may collectively perform a single function.
  • Each module may have a specialized processor capable of executing machine-readable program instructions.
  • a module may comprise a single piece of hardware such as an ASIC, electronic circuit, or special purpose processor.
  • a plurality of modules may be executed by either a single special purpose computer system or a plurality of special purpose computer systems in parallel.
  • Modules may include software/hardware which may further comprise an operating system, drivers, controllers, and other apparatuses some or all of which may be connected via a network.
  • Compressed sensing deals with signal recovery from highly incomplete information.
  • can be, for example, an orthonormal transform such as a DCT, wavelet or FFT transform.
  • the number of non-zero entries in s determines the degree of sparseness K of x.
  • a set of binary pseudo-random sampling functions were used as sampling matrix ⁇ of an image x, along with the assumption that each measurement (i.e., the inner product result) was uniformly quantized to a certain number of bits, and that the image was sparse in the Haar wavelet domain; note that assumptions of the image being sparse in other domains are anticipated.
  • FIG. 3 shows the magnitude of the Haar coefficients of an image x used in the simulation of the present method in decreasing order, from largest to smallest. It can be seen that about 80% of the image energy is contained in the largest 500 Haar coefficients, while 99.9% of the image energy is contained in the largest 3000 coefficients.
  • An embodiment of the present system was used to reconstruct a spatial appearance of a scene comprising a 64 ⁇ 64 pixel image of the scene.
  • a binary mask was used to identify an example ROI with dimensions 32 ⁇ 64 pixels.
  • the table of FIG. 4 shows the overall mean squared error as well as the mean squared error between an image reconstructed from various numbers of samples and various block lengths B using the present method. That is, the set of sampling function was partitioned into subsets of the form ⁇ 1 , . . . , ⁇ B ⁇ , ⁇ B+1 , . . . , ⁇ 2B ⁇ , . . . , ⁇ B ⁇ M/B ⁇ B+1 , . . . , ⁇ M ⁇ .

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • General Physics & Mathematics (AREA)
  • Physics & Mathematics (AREA)
  • Multimedia (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Health & Medical Sciences (AREA)
  • General Health & Medical Sciences (AREA)
  • Artificial Intelligence (AREA)
  • Evolutionary Computation (AREA)
  • Databases & Information Systems (AREA)
  • Software Systems (AREA)
  • Medical Informatics (AREA)
  • Computing Systems (AREA)
  • Data Mining & Analysis (AREA)
  • Vascular Medicine (AREA)
  • Signal Processing (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Bioinformatics & Cheminformatics (AREA)
  • Bioinformatics & Computational Biology (AREA)
  • Evolutionary Biology (AREA)
  • General Engineering & Computer Science (AREA)
  • Compression Or Coding Systems Of Tv Signals (AREA)

Abstract

What is disclosed is a system and method which reconstructs an N-pixel image of a scene such that Q pixel locations associated with identified regions of interest in a scene have a higher image quality when rendered relative to other pixels in the image. Acquisition and adaptive-quality compression are performed simultaneously by semi-synchronously or asynchronously temporally modulating an ordered set of sampling functions used to spatially modulate a pattern of light. The teachings hereof improve compression efficiency of a compressed sensing framework while improving encoding efficiency with respect to traditional compressed sensing techniques.

Description

    CROSS REFERENCE TO RELATED APPLICATIONS
  • This patent application is a continuation-in-part of commonly owned and co-pending U.S. patent application Ser. No. 13/932,791, entitled: “Reconstructing An Image Of A Scene Captured Using A Compressed Sensing Device”, by Bernal et al.
  • TECHNICAL FIELD
  • The present invention is directed to systems and methods which use a compressed sensing framework to reconstruct an image of a scene wherein pixel locations associated with identified regions of interest in that scene have a higher image quality when rendered relative to other pixels of the reconstructed image.
  • BACKGROUND
  • Compressed sensing is a relatively new area in the signal processing art where one measures a small number of non-adaptive linear combinations of a signal. These measurements are usually much smaller than the number of samples that define the signal. From the small numbers of measurements, the signal is reconstructed by a non-linear process which aims to reduce the overall complexity required by a large variety of measurement systems by introducing signal compression into the measurement process. Essentially, the theory behind compressed sensing is that sparse signal statistics can be recovered from a small number of measurements. The sparse nature of most signals of interest allows high-fidelity reconstructions to be made using a compressed sensing approach. Compressed sensing can be beneficial because it reduces the number of samples required to spatially and/or temporally reconstruct a given scene thereby enabling the use of inexpensive sensors with reduced spatial and/or temporal resolution in certain applications where complex sensors are otherwise used, while maintaining the quality of the reconstructed image of the scene. Compressed sensing also holds an advantage in fidelity over conventional camera systems due to detector noise issues that may affect measurements due to the limited radiometric efficiency of pixels in two-dimensional sensors. The present application provides a desirable extension to the prior compressed sensing system taught by Bernal et al.
  • BRIEF SUMMARY
  • What is disclosed is a system and method which reconstructs an N-pixel image of a scene such that Q pixel locations associated with identified regions of interest in a scene have a higher image quality when rendered relative to other pixels in the image. Acquisition and adaptive-quality compression are performed simultaneously by semi-synchronously or asynchronously modulating temporally an ordered set of sampling functions, used to spatially modulate light incoming from the scene. The teachings hereof improve compression efficiency of a compressed sensing framework. Encoding efficiency is improved with respect to traditional compressed sensing techniques.
  • In the context of the following discussion, bold Greek letters refer to sets of functions of a given length and to matrices formed by stacking rows, where each row is a function in the set.
  • In one embodiment, the present method for reconstructing an N-pixel image of a scene captured using a compressed sensing device involves performing the following. First, a mask is received which identifies at least one region of interest (ROI) in a scene having Q pixels, where Q<N. The mask can be a binary image wherein pixels with value ‘1’ (ON pixels) indicate locations associated with the ROI and pixels with value ‘0’ (OFF pixels) indicate locations not associated with the ROI. Consequently, the mask has at least Q ON pixels, and at most N−Q OFF pixels. The mask can also be multi-level, each level corresponding to one identified ROI, each identified ROI potentially having a different quality of reconstruction assigned to it. Next, a set of M sampling functions is arranged in an M×N matrix φ={φ1, . . . , φM}, whose mth row vector φmε
    Figure US20150003738A1-20150101-P00001
    N denotes the mth N-dimensional sampling function, and where M<<N. The set of sampling functions is then partitioned into K non-overlapping and non-empty subsets of sampling functions {φ1, . . . , φM1}, {φM1+1, . . . , φM1+M2}, . . . , {φM−Mk+1, . . . , φM} each having Mi elements, where 1≦i≦K and M1+M2+ . . . +Mk=M. Each of the resulting subset of functions are next arranged in Mi×N matrices φ1, φ2, . . . , φK. Each matrix has at least Q linearly independent columns and at most N−Q linearly dependent columns. An index of the linearly independent columns is associated with locations corresponding to the ROIs, and an index of the linearly dependent columns is associated with locations which do not correspond to the ROIs. Thereafter, incoming light is modulated by a spatial light modulator according to a plurality of spatial patterns corresponding to the ordered sampling functions. The light reflected/transmitted off the modulator is focused onto a detector of a compressed sensing device. The detector proceeds to sample sequential measurements of the light focused thereon. The measurements comprise a sequence of projection coefficients corresponding to a scene. Each of the sampled measurements are an inner product result ym=<x,φm>, where x denotes an N-dimensional vector representing an N-pixel sampled version of a scene. Thereafter, a spatial appearance of the scene is reconstructed from the sequence of projection coefficients after M inner products have been sampled. The reconstructed image being such that pixels associated with the ROIs have a higher image quality when rendered relative to other pixels in the image.
  • Features and advantages of the above-described system and method will become readily apparent from the following detailed description and accompanying drawings.
  • BRIEF DESCRIPTION OF THE DRAWINGS
  • The foregoing and other features and advantages of the subject matter disclosed herein will be made apparent from the following detailed description taken in conjunction with the accompanying drawings, in which:
  • FIG. 1 illustrates one example embodiment of the present method for reconstructing an image of a scene;
  • FIG. 2 shows one example system for performing scene reconstruction in accordance with the teachings hereof;
  • FIG. 3 shows the magnitude of the Haar coefficients of an image x used in the simulation of the present method in decreasing order, from largest to smallest; and
  • FIG. 4 is a table of results illustrating quality of reconstruction as measured by the mean squared error (MSE) from having reconstructed an image with varying number of samples using the present method, as well as the traditional compressed sensing approach.
  • DETAILED DESCRIPTION
  • What is disclosed is a system and method which reconstructs an N-pixel image of a scene such that Q pixel locations associated with identified regions of interest in a scene have a higher image quality when rendered relative to other pixels in the image.
  • NON-LIMITING DEFINITIONS
  • A “region of interest” (ROI) is an identified area of a scene intended to be reconstructed using the teachings disclosed herein with a higher quality relative to other areas of that scene. What defines a particular region of interest will largely depend on the application where the present invention finds its uses. The image of the scene from which a region of interest is first identified can be acquired with the compressed sensing device following the traditional compressed sensing framework. Alternatively, it can be acquired with a traditional imaging device having a significantly similar view point of the scene relative to the compressed sensing device. In yet another embodiment, the image of the scene can be received from an existing database of images. A region of interest is then identified from the image of the scene. Regions of interest can be identified by processing an image of the scene to identify using any of: pixel classification, object identification, facial recognition, color, texture, spatial features, spectral information, pattern recognition, motion detection, foreground detection, and a user input. The location, shape or size of the identified region of interest can change over time. Once a region of interest has been identified, changes in the location of the identified region of interest can be identified by tracking techniques. Alternatively, re-identification of the region of interest pixel classification, object identification, facial recognition, color, texture, spatial features, spectral information, pattern recognition, motion detection, foreground detection, and a user input can be performed periodically.
  • A “photodetector” or simply “detector” is a device which measures a magnitude of an intensity of light focused thereon. In various embodiments, the photodetector can be a single (diode) detector or a multi-diode detector and may further comprise an analog-to-digital converter and an amplifier.
  • A “Spatial Light Modulator (SLM)” is a device in the compressed sensing device positioned along an optical axis where a camera's focal plane array would typically be located. The SLM is controllable such that it can be configured according to spatial patterns which can be used to modulate incoming light which can then be transmitted or reflected onto a photodetector of a compressed sensing device. As mentioned earlier, a compressed sensing device relies on modulating incoming light from the scene by a spatial light modulator according to a plurality of spatial patterns. Examples of spatial light modulators include, but are not limited to Digital Micromirror Devices, Transmissive Liquid Crystals and Liquid Crystals on Silicon.
  • “Digital Micromirror Device (DMD)” is an optical micro-electromechanical (MEMS) device which has, on its surface, imaging elements comprising microscopic opto-mechanical mirrors arrayed on a two-dimensional grid. Each mirror in the array is referred to as a DMD pixel. The microscopic mirrors are electronically controllable and thus modulate incoming light by toggling a reflectivity thereof by individually tilting (or rotating) the mirrors in one direction or another to achieve an ON/OFF state. In the ON state, light is reflected in a desired direction, such as through a lens or onto a photodetector. In the OFF state, the light is directed elsewhere. By convention, the positive (+) state is ON and the negative (−) state is OFF. The two states are opposite, i.e., if one element is ‘1’ then the other is ‘0’, and vice versa. As prescribed by compressed sensing theory, each DMD pattern is configured to select a defined portion of the incoming light onto a detector. During image acquisition, a series of unique patterns are sequentially provided to the DMD and a series of measurements are obtained. Light energy is reflected by the DMD mirrors onto the photo diode or photoreceptor where the photons of the image are converted to an electrical signal. Each signal, produced as a result of each measurement, is a function of a specific pattern and of the scene. By rapidly changing the DMD patterns and obtaining measurements therefrom, a time-series signal is obtained. Utilizing a compressed sensing framework, an image reconstruction algorithm reconstructs the original image from the generated time-series measurement data with knowledge of the temporal sequence of patterns. DMDs are available from vendors in various streams of commerce.
  • A “Transmissive Liquid Crystal (TLC)” also referred to a “Liquid Crystal Modulator (LCM)”, is a programmable array of liquid crystal elements. Each liquid crystal element in the array is a pixel. The liquid crystal elements are individually electronically controllable and thus the TLC modulates incoming light by toggling a transparency of each TLC pixel to achieve an ON/OFF state. By convention, in the ON state, the liquid crystal element is transparent so light passes therethrough. In the OFF state, the liquid crystal element is opaque so light does not pass therethrough. TLCs are desirable in many applications because of their fast switching times and a high degree of usability over a broad range of visible to infrared wavelength bands. TLCs are available from vendors in various streams of commerce.
  • A reflective “Liquid Crystal on Silicon (LCOS)” refers to a micro-projection or micro-display technology which uses liquid crystals instead of individual mirrors. In LCOS, liquid crystals are applied directly to the surface of a silicon chip coated with an aluminized layer with some type of passivation layer, which is highly reflective. LCOS technology is preferable in many applications because it can produce higher resolution and higher contrast images than standard liquid crystal technologies.
  • A “compressed sensing framework” is a signal processing technique for reconstructing a signal with solutions found by taking advantage of the signal's sparseness or compressibility in some domain, thereby enabling the entire signal to be generated from relatively few measurements. An underdetermined linear system has more unknowns than equations and generally has an infinite number of solutions. In order to choose a proper solution, constraints are applied. Because many signals are sparse, i.e., they contain many coefficients close to or equal to zero when represented in some domain, the additional constraint of sparsity allows only those solutions with a small number of non-zero coefficients to be considered as feasible. Not all underdetermined systems have a sparse solution. However, if there is a unique sparse representation to that underdetermined linear system then a compressed sensing framework enables a recovery of that solution.
  • “Correlation” between two functions φ, Φε
    Figure US20150003738A1-20150101-P00001
    N is defined as the magnitude or absolute value of their inner product |
    Figure US20150003738A1-20150101-P00002
    φ,Φ
    Figure US20150003738A1-20150101-P00003
    |, where the inner product
    Figure US20150003738A1-20150101-P00002
    φ,Φ
    Figure US20150003738A1-20150101-P00003
    is obtained by performing N element-wise multiplication and adding the N individual results into a single number. Two functions are said to be “largely uncorrelated” if there exists a real number T1 such that |
    Figure US20150003738A1-20150101-P00004
    φ,Φ
    Figure US20150003738A1-20150101-P00005
    ≦T1. The choice for T1 is usually application-dependent and is expressed in terms of N and of the value P that bounds Φ and φ, where P is such that |Φ(k)|≦P and |φ(k)|≦P, for all k. For example, in one application, two functions are considered to be largely uncorrelated if T1=0.1NP2. Two functions are said to be “largely correlated” if there exists a real number T1′ such that |
    Figure US20150003738A1-20150101-P00004
    φ,Φ
    Figure US20150003738A1-20150101-P00005
    ≧T1′. For example, in one application, two functions are considered to be largely uncorrelated if T1′=0.9NP2. In the context of the present disclosure, a level of correlation between two functions is indicative of a level of co-linearity: the larger the correlation between two functions, the more collinear they are.
  • “Coherence” between two sets of functions Φ and φ, where Φ={Φ1, . . . , ΦM1} with Φiε
    Figure US20150003738A1-20150101-P00006
    N for all i, and φ={φ1, . . . , φM2} with φjε
    Figure US20150003738A1-20150101-P00006
    N for all j is defined as: μ(Φ, φ)=√{square root over (N)}max1≦i≦M1, 1≦j≦M2{|
    Figure US20150003738A1-20150101-P00004
    Φij
    Figure US20150003738A1-20150101-P00005
    }. Two sets of functions Φ and φ are said to be “largely incoherent” if there exists a real number T2 such that μ(Φ, φ)≦T2. The choice for T2 is usually application-dependent and is expressed in terms of N and of the range of values of the functions φi and φj in Φ and φ, P. For example, in one application, two sets are considered to be largely incoherent for T2=0.2NP2.
  • Flow Diagram of One Embodiment
  • Reference is now being made to the flow diagram of FIG. 1 which illustrates one example embodiment of the present method for reconstructing an image of a scene captured using a compressed sensing device with the teachings hereof wherein Q pixel locations associated with identified regions of interest in the scene have a higher image quality when rendered relative to other pixels in the image. Flow processing begins at step 100 and immediately proceeds to step 102.
  • At step 102, receive a mask identifying at least one region of interest (ROI) in a scene. The mask can be dynamically updated in response to a new region of interest having been identified, a location of a region of interest changing in a scene, or a user input. The mask may be received from a remote device over a network via a wired or wireless pathway, or retrieved from a storage device such as a memory or a hard drive. The mask can be a binary image wherein pixels with value ‘1’ (ON pixels) indicate locations associated with the ROI and pixels with value ‘0’ (OFF pixels) indicate locations not associated with the ROI. The mask can also be multi-level, each level corresponding to one identified ROI.
  • At step 104, order a set of M sampling functions arranged in an M×N matrix φ={φ1, . . . , φM}, whose mth row vector φmε
    Figure US20150003738A1-20150101-P00006
    N denotes the mth N-dimensional sampling function, where M<<N. The index m is indicative of an ordering of the function in a sampling sequence. The ordering is such that, given a block length B, partitions of the set φ into non-overlapping and non-empty subsets define sub-matrices φ1, φ2, . . . , φ┌M/B┐ all except possibly one of size B×N (where ┌┐ denotes the ceiling operator which maps a real number to a smallest following integer) corresponding to sequences of functions arranged in subsets all except possibly one of length B, {φ1, . . . , φB}, {φB+1, . . . , φ2B}, . . . , {φB┌M/B┐−B+1, . . . , φM}, respectively, where B<M is the block length. Each resulting sub-matrix has at least Q linearly independent columns and at most N−Q dependent columns. It should be appreciated that the word “possibly” is used here because either all are of size B×N or all but one are of size B×N.
  • Other partitions of the original set into
  • M B
  • subsets, all except one of length B may be utilized. For example, the partition {φ1, . . . , φB}, {φB+1, . . . , φ2B}, . . . , {φB┌M/B┐−B+1, . . . , φM} is valid. Partitions of the original set into
  • M B
  • subsets, all except possibly one of length B, (where └┘ denotes a floor operation which maps a real number to a largest previous integer) may be used. It should be appreciated that the word “possibly” is used here because either all are of length B or all but one are of length B.
  • It will be appreciated by someone skilled in the art that other partitions into different numbers of subsets each with a possibly different number of elements also serve the purpose of the methods introduced in the present disclosure. For example, and more generally (i.e., independently of a block length B), partitions of the set of sampling functions φ into K non-overlapping and non-empty subsets of functions φ1, φ2, . . . , φK, each subset having Mi elements, where 1≦i≦K and M1+M2+ . . . +Mk=M may also be used. The corresponding matrices of size Mi×N, where Mi<<N, each can have at least Q linearly independent columns and at most N−Q linearly dependent columns.
  • Since the reconstructed image and the sampling functions have the same number of pixels N, there is a one to one correspondence between the indices in the sampling functions and spatial locations in the image. From this correspondence, the index of the linearly independent columns is determined to be associated with the locations corresponding to the identified regions of interest, and the index of the linearly dependent columns is determined to be associated with the locations not corresponding to the identified regions of interest.
  • The value of B determines the difference in the quality of reconstruction of the ROI vs. non-ROI regions, larger values of B corresponding to larger differences. Typical values of B range from 2 to 8 although use of values outside that range may be used. In the multiple ROI case, each ROI may have an associated B value relative to the common non-ROI regions; regions with larger B values will be reconstructed at higher quality relative to regions with smaller B values. For example, if there are two ROIs of sizes Q1 and Q2 pixels respectively, and each associated with block lengths B1 and B2, then let B be the least common multiple between B1 and B2. Without loss of generality, assume B2>B1 and B2=kB1 for some integer k so that B=B2. Each sub-matrix φ1, φ2, . . . , φ┌M/B┐ with B rows and N columns will have: a sub-matrix with B rows and at least Q2 columns with full rank (e.g., rank B=B2), an index of the linearly independent columns in said sub-matrix being associated with locations corresponding to ROI number 2; a sub-matrix with B rows and at least Q1 columns with rank B1=B/k, an index of the columns in said sub-matrix being associated with locations corresponding to ROI number 1; and a sub-matrix with B rows and at most N−Q1−Q2 columns with rank 1, an index of the linearly dependent columns in said sub-matrix being associated with locations which do not correspond to either ROI. In the case where partitions of the set of sampling functions φ into K non-overlapping and non-empty subsets of functions φ1, φ2, . . . , φK, each subset having Mi elements, larger values of Mi will lead to larger differences in the quality of reconstruction of ROI vs. non-ROI regions.
  • Alternatively, the ordering can be such that, for sequences of sampling functions arranged in subsets all except possibly one of length B, subsets of sampling functions {φ1, . . . , φB}, {φB+1, . . . , φ2B}, {φB┌M/B┐−B+1, . . . , φM} are largely incoherent, sampling functions φi and φj in a given subset are largely uncorrelated for every i≠j only for pixels within ROIs as identified by the mask, and sampling functions φi and φj in a given subset are largely or even completely correlated for pixels outside the ROIs. In one embodiment, the sampling functions are obtained by adaptively modifying a set of largely uncorrelated sampling functions φ={φ1 . . . , φM} (where φ is extracted, for example, from a random matrix with general independent rows or columns, or from a sub-Gaussian matrix,) in such a way that φii for 1≦i≦M at locations associated with the ROIs and φik for 1≦i≦M and some fixed kε[B┌i/B┐−B+1,B┌i/B┐]∩[1,M], (where ┌┐ denotes the ceiling operator which maps a real number to a smallest following integer) at locations not associated with the ROIs. In the general case where partitions of the set of sampling functions φ into K non-overlapping and non-empty subsets of functions φ1, φ2, . . . , φK, each subset having Mi elements, where 1≦i≦K and M1+M2+ . . . +Mk=M are used, the sampling functions can be obtained by adaptively modifying a set of largely uncorrelated sampling functions φ={φ1 . . . , φM} in such a way that φii for 1≦i≦M at locations associated with the ROIs; for locations not associated with the ROIs, φik for 1≦i≦M, where k is such that when M1+ . . . +Mk−1+1≦i≦M1+ . . . +Mk, kε[M1+ . . . +Mk−1+1, M1+ . . . +Mk].
  • In an alternative embodiment, the sampling functions are obtained by adaptively combining a set of largely uncorrelated sampling functions φ={φ1 . . . , φM} and a set of B-block-wise largely correlated functions α={α1 . . . , αM}. This is to say that blocks {α1, . . . , αB}, {αB+1, . . . , α2B}, . . . , {αB┌M/B┐−B+1, . . . , αM} are largely incoherent, and functions αi and αj in a given subset are largely or fully correlated. The combination is performed in such a way that φii for 1≦i≦M at locations associated with the ROIs and φii, for 1≦i≦M and at locations not associated with the ROIs. A similar sampling function design strategy can be implemented in cases when partitions of the set of sampling functions φ into K non-overlapping and non-empty subsets of functions φ1, φ2, . . . , φK, each subset having Mi elements are used. In this case, α is a set of Mi-block-wise largely correlated functions, where blocks {α1, . . . , αM1}, {αM1+1, . . . , αM1+M2}, . . . , {αM1−Mk+1, . . . , αM} are largely incoherent, and functions αi and αj in a given subset are largely or possibly completely correlated. From these sets of functions, φii for 1≦i≦M at locations associated with the ROIs as before; for locations not associated with the ROIs, φii for 1≦i≦M.
  • There are several approaches to obtaining the set of sampling functions φ with the desired characteristics. In one embodiment, a set of largely uncorrelated sampling functions φ={φ1 . . . , φM} extracted, for example, from a random matrix with general independent rows or from a sub-Gaussian matrix, are adaptively modified such that φii for 1≦i≦M at locations associated with the ROIs and φik for 1≦i≦M and a fixed kε[B┌i/B┐−B+1,B┌i/B┐]∩[1,M], (where ┌┐ denotes the ceiling operator which maps a real number to a smallest following integer) at locations not associated with the ROIs. In cases when partitions of the set of sampling functions φ into K non-overlapping and non-empty subsets of functions φ1, φ2, . . . , φK, each subset having Mi elements are used, φii for 1≦i≦M at locations associated with the ROIs; for locations not associated with the ROIs, φik for 1≦i≦M, where k is such that when M1+ . . . +Mk−1+1≦i≦M1+ . . . +Mk, kε[M1+ . . . +Mk−1+1, M1+ . . . +Mk].
  • Measurements are obtained of the scene using a compressed sensing device which comprises, at least in part, a spatial light modulator configured according to a plurality of spatial patterns corresponding to the set of sampling functions. Each pattern focuses a portion of incoming light onto a detector which samples sequential measurements of light focused thereon. Each of the measurements is an inner product result ym=<x,φm>, where x denotes an N-dimensional vector representing the N-pixel sampled version of the scene. A series of measurements comprises a sequence of projection coefficients corresponding to the inner product between that N-pixel sampled version of the scene and each of the sampling functions. An appearance of the scene is reconstructed from the sequence of projection coefficients after M inner products have been sampled, where M<<N, such that pixel locations associated with ROIs in the scene have a higher image quality when rendered relative to other pixels in the image.
  • The reason that this method is advantageous, particularly in the context of a DMD-based single pixel camera, has to do with the way a DMD module switching occurs. The switching capabilities of the array are limited so that instead of switching all individual micromirrors simultaneously, a group of micromirrors (e.g., a column, row or more generally, a cluster) has to be switched before the next group can be switched. Since single-pixel camera measurements cannot be taken until all micromirrors in the array have switched to their intended position, this sequential switching introduces undesirable delays and, consequently, reduces the effective sampling rate of the camera. In the present approach, since only a fraction of the micromirrors is switched at every cycle, the effect of the sequential switching mode on the sampling rate is ameliorated by a factor approximately equal to the ratio of the number of non-ROI pixels to the number of ROI pixels. Note that performing measurements with the ordered sampling functions obtained in the manner described above, achieves reconstruction of an ROI with increased quality relative to the rest of the image; additionally, and in the context of a DMD-based single pixel camera it also enables faster sampling rates due to the way a DMD module switching occurs, as stated above.
  • At step 106, configure a spatial light modulator to modulate incoming light according to spatial patterns corresponding to the set of ordered sampling functions.
  • At step 108, use a detector to sequentially sample the focused spatial pattern of light to obtain measurements of the scene, each comprising an inner product result ym=<x,φm>, where x denotes an N-dimensional vector representing the N-pixel sampled version of the scene (i.e., the vectorized matrix representation of an image). The measurements comprise a sequence of projection coefficients corresponding to the scene.
  • At step 110, reconstruct a spatial appearance of an image of the scene from the sequence of projection coefficients after M inner products have been sampled. The reconstruction is such that pixels associated with the localized ROIs in the reconstructed image have a higher image quality, that more faithfully represents the appearance of the sampled scene, when rendered relative to other pixels in the image. This non-uniform quality of the reconstruction is enabled by the sampling scheme described above. Thereafter, in this embodiment, further processing stops.
  • It should be appreciated that the flow diagrams hereof are illustrative. One or more of the operations illustrated in the flow diagrams may be performed in a differing order. Other operations may be added, modified, enhanced, or consolidated. Variations thereof are intended to fall within the scope of the appended claims.
  • Example System Architecture
  • Reference is now being made to FIG. 2, which shows one example system for performing scene reconstruction in accordance with the teachings hereof.
  • In the system of FIG. 2, incoming light (collectively at 201) enters the compressed sensing system 200 through an aperture 202 and into a spatial light modulator (SLM) 203 which modulates the incoming light to produce a spatial pattern of light 204 which is focused on to detector 205. The detector measures a magnitude of an intensity of the spatial pattern focused thereon. Mask module 208 receives a mask using USB port 209, and provides the mask to a controller 207 shown comprising a processor (CPU) and a memory. The controller facilitates the configuration of the spatial light modulator to modulate incoming light. The detector outputs sequential measurements 206 which may be provided to USB port 215. Measurements obtained by the detector are communicated to image reconstruction module 213 wherein a spatial appearance of the scene is reconstructed from the measurements. The measurements and the reconstructed image 214 are communicated to storage device 216 and/or provided as output to workstation 220. Values, data, measurements, and results of any of the modules and processing units of the system 200 may be obtained or retrieved via communications bus 217.
  • Shown in communication with the system 200 is a workstation 220. The workstation is shown comprising a monitor 221, a keyboard 222, a mouse 223, a storage device 224, and a computer-readable media 225. The workstation is also placed in communication with one or more remote devices over network 226 using, for example, a network card. A user thereof may change or control the functionality of any of the modules and processing units comprising the system 200 using the workstation. An image of a scene can be displayed on the monitor and corrected and/or cropped. Masks can be generated using the workstation and communicated to the mask module 208. Measurements and values generated by the system may be displayed on the display device. Intensity values obtained by the detector may be modified by a user of the workstation. The values of the spatial pattern that controls the modulation of incoming light may further be communicated to the workstation and displayed on the monitor. A user can selectively identify regions of interest using, for example, a mouse. The user may further define the ordering of the set of sampling functions using the workstation. Localized regions of interest can be communicated to the mask module by the workstation. An operator of the workstation may modify the results generated by any of the modules or processing units of FIG. 2 as needed and/or re-direct the modified results back to the same or different modules for further processing or re-processing. It should be appreciated that the workstation has an operating system and other specialized software configured to display a variety of numeric values, text, scroll bars, pull-down menus with user selectable options, and the like, for entering, selecting, or modifying information displayed on the display device.
  • Various modules may designate one or more components which may, in turn, comprise software and/or hardware designed to perform the intended function. A plurality of modules may collectively perform a single function. Each module may have a specialized processor capable of executing machine-readable program instructions. A module may comprise a single piece of hardware such as an ASIC, electronic circuit, or special purpose processor. A plurality of modules may be executed by either a single special purpose computer system or a plurality of special purpose computer systems in parallel. Modules may include software/hardware which may further comprise an operating system, drivers, controllers, and other apparatuses some or all of which may be connected via a network.
  • Implementation Details
  • Compressed sensing deals with signal recovery from highly incomplete information. A cornerstone of compressed sensing is that an N-dimensional sparse vector x[] can be recovered from a small number M where M<<N of linear measurements ym=<x,φm>, m=1, 2, . . . , M, under a certain set of assumptions and by solving a convex optimization problem. In matrix form, y=φx where x=Ψs, with s being a sparse vector and Ψ defining a transformation to a domain where x is sparse; Ψ can be, for example, an orthonormal transform such as a DCT, wavelet or FFT transform. The number of non-zero entries in s determines the degree of sparseness K of x. For testing purposes, a set of binary pseudo-random sampling functions were used as sampling matrix φ of an image x, along with the assumption that each measurement (i.e., the inner product result) was uniformly quantized to a certain number of bits, and that the image was sparse in the Haar wavelet domain; note that assumptions of the image being sparse in other domains are anticipated. FIG. 3 shows the magnitude of the Haar coefficients of an image x used in the simulation of the present method in decreasing order, from largest to smallest. It can be seen that about 80% of the image energy is contained in the largest 500 Haar coefficients, while 99.9% of the image energy is contained in the largest 3000 coefficients.
  • The test image was sampled according to the teachings herein, and each of the measurements or inner products, quantized or digitized to a fixed bit depth. From the quantized set of measurements, we posed the recovery problem as an optimization x=argmin{∥x∥1} subject to y=φx where, ∥∥1 denotes the l1 norm. It should be appreciated that this is only one of multiple ways to solve an inverse problem. Other approaches include exploiting the assumption that the gradient of the image is sparse, as well as building over-complete dictionaries in which the representation target image is sparse. The method described herein is equally adaptable to these techniques as well, as they rely on the use of sequences of sampling functions to perform the sensing.
  • An embodiment of the present system was used to reconstruct a spatial appearance of a scene comprising a 64×64 pixel image of the scene. A binary mask was used to identify an example ROI with dimensions 32×64 pixels. The table of FIG. 4 shows the overall mean squared error as well as the mean squared error between an image reconstructed from various numbers of samples and various block lengths B using the present method. That is, the set of sampling function was partitioned into subsets of the form {φ1, . . . , φB}, {φB+1, . . . , φ2B}, . . . , {φB┌M/B┐−B+1, . . . , φM}.
  • It can be seen by an examination of the results thereof that, as the block length increases, the quality of the reconstructed image as measured by the mean square error decreases; however, the quality of the reconstructed ROI suffers little for the range of B values tested, and, it is consistently better or on par relative to that yielded by the traditional compressed sensing approach. These results showcase the efficiency of the present method in preserving the quality of the targeted region of interest as the lengths of the blocks increase; we refer to this as improved encoding efficiency, since the quality of the reconstructed ROI is better than that yielded by traditional compressed sensing methods for a given number of measurements. Note that, in the context of a single pixel camera, use of the present method would result in increased switching frequencies: the larger the value of B, the faster the subsequent samples can be acquired.
  • One or more aspects of the teachings disclosed herein are intended to be incorporated in an article of manufacture. The above-disclosed features and functions or alternatives thereof, may be combined into other systems or applications. Presently unforeseen or unanticipated alternatives, modifications, variations, or improvements may become apparent and/or subsequently made by those skilled in the art and, further, may be desirably combined into other different systems or applications. Changes to the above-described embodiments may be made without departing from the spirit and scope of the invention. The teachings of any printed publications including patents and patent applications, are each separately hereby incorporated by reference in their entirety.

Claims (25)

What is claimed is:
1. A method for reconstructing an N-pixel image of a scene captured using a compressed sensing device, the method comprising:
receiving a mask identifying at least one region of interest (ROI) of size Q pixels in the scene;
arranging a set of M sampling functions in an M×N matrix φ={φ1, . . . , φM}, whose mth row vector φmε
Figure US20150003738A1-20150101-P00001
N denotes the mth N-dimensional sampling function, and where M<<N;
partitioning said set of sampling functions into K non-overlapping and non-empty subsets of functions, each subset having Mi elements, where 1≦i≦K and M1+M2+ . . . +Mk=M;
arranging each resulting subset of functions in Mi×N matrices φ1, φ2, . . . , φK, each matrix having at least Q linearly independent columns and at most N−Q linearly dependent columns, an index of said linearly independent columns being associated with locations corresponding to said ROIs, and an index of said linearly dependent columns being associated with locations which do not correspond to said ROIs;
modulating, by a spatial light modulator, incoming light according to a plurality of spatial patterns corresponding to said ordered sampling functions, said modulated light being focused onto a detector of a compressed sensing device;
sampling, by said detector, sequential measurements of said modulated light focused thereon, each of said sampled measurements being an inner product result ym=<x,φm>, where x denotes an N-dimension vector representing the N-pixel sampled version of said scene, said measurements comprising a sequence of projection coefficients corresponding to said scene; and
reconstructing a spatial appearance of said scene from said sequence of projection coefficients after M inner products have been sampled, said reconstructed image being such that pixels associated with said ROIs have a higher image quality when rendered relative to other pixels in said image.
2. The method of claim 1, wherein said partition is performed according to a block size B determining one of two possible partitions, φ1, φ2, . . . , φ┌M/B┐ and φ1, φ2, . . . , φ┌M/B┐ so that K equals one of ┌M/B┐ and ┌M/B┐, wherein each subset in the partition except possibly one has B elements.
3. The method of claim 2, wherein said partition according to a block size B is one of: {φ1, . . . , φB}, {φB+1, . . . , φ2B}, . . . , {φB┌M/B┐−B+1, . . . , φM}, and {φ1, . . . , φB}, {φB+1, . . . , φ2B}, . . . , {φB┌M/B┐−2B+1, . . . , φM}, where M>2B.
4. The method of claim 1, wherein subsets of sampling functions φ1, φ2, . . . , φK are largely incoherent, sampling functions φi and φj in a given subset are largely uncorrelated for every i≠j only for indices associated with said ROIs, and sampling functions φi and φj in a given subset are largely correlated for indices not associated with said ROIs.
5. The method of claim 4, wherein said sampling functions are obtained by adaptively modifying a set of largely uncorrelated sampling functions φ={φ1, . . . , φM} in such a way that φii for 1≦i≦M at locations associated with said ROIs, and φik for 1≦i≦M where k is such that when M1+ . . . +Mk−1+1≦i≦M1+ . . . +Mk, kε[M1+ . . . +Mk−1+1, M1+ . . . +Mk] at locations not associated with said ROIs.
6. The method of claim 2, wherein said sampling functions are obtained by adaptively modifying a set of largely uncorrelated sampling functions φ={φ1 . . . , φM} in such a way that φii for 1≦i≦M at locations associated with said ROIs and φik for 1≦i≦M and kε[B┌i/B┐−B+1,B┌i/B┐]∩[1,M], where ┌┐ denotes the ceiling operator which maps a real number to a smallest following integer at locations not associated with said ROIs.
7. The method of claim 2, wherein said sampling functions are obtained by adaptively combining a set of largely uncorrelated sampling functions φ={φ1 . . . , φM} and a set of B-block-wise largely correlated functions α={α1 . . . , αM} in such a way that φii for 1≦i≦M at indices associated with said ROIs, and φii for 1≦i≦M at indices not associated with said ROIs.
8. The method of claim 1, wherein said sampling functions are generated based on an output of a random number generator, wherein generated values of sampling functions associated with said ROIs update at a different rate than those not associated with said ROIs, said output of said random number generator being passed through a deterministic function in order to obtain samples from different distributions.
9. The method of claim 1, wherein said spatial light modulator comprises any of: a digital micromirror device, a transmissive liquid crystal, and reflective liquid crystal on silicon.
10. The method of claim 1, further comprising processing an image of said scene to identify said ROIs using any of: pixel classification, object identification, facial recognition, color, texture, spatial features, spectral information, pattern recognition, motion detection, foreground detection, and a user input
11. The method of claim 1, further comprising processing an image of said scene to generate said mask wherein pixels associated with said ROI as being active and pixels outside said ROI as being inactive.
12. The method of claim 1, wherein said mask is updated in real-time in response to a location of any of said ROIs changing over time.
13. The method of claim 2, wherein different ROIs are associated with different subset lengths B.
14. A compressed sensing system for reconstructing an N-pixel image of a scene captured using a compressed sensing device, the system comprising:
a mask identifying at least one Q-pixel region of interest (ROI) in a scene;
a spatial light modulator for modulating incoming light according to spatial patterns corresponding to a set of M sampling functions arranged in an M×N matrix φ={φ1, . . . , φM}, whose mth row vector φmε
Figure US20150003738A1-20150101-P00001
N denotes the mth N-dimensional sampling function, and where M<<N, said set of sampling functions being partitioned into K non-overlapping and non-empty subsets of functions, each subset having Mi elements, where 1≦i≦K and M1+M2+ . . . +Mk=M, each resulting subset of functions being arranged in Mi×N matrices φ1, φ2, . . . , φK, each matrix having at least Q linearly independent columns and at most N−Q linearly dependent columns, an index of said linearly independent columns being associated with locations corresponding to said ROIs, and an index of said linearly dependent columns being associated with locations which do not correspond to said ROIs;
a detector of a compressed sensing device, said detector sampling sequential measurements of said modulated light focused thereon, each of said sampled measurements being an inner product result ym=<x,φm>, where x denotes an N-dimension vector representing the N-pixel sampled version of a scene, said measurements comprising a sequence of projection coefficients corresponding to said scene; and
a processor executing machine readable program instructions for reconstructing a spatial appearance of said scene from said sequence of projection coefficients after M inner products have been sampled, said reconstructed image being such that pixels associated with said ROIs have a higher image quality when rendered relative to other pixels in said image.
15. The system of claim 14, wherein said partition is performed according to a block size B determining one of two possible partitions, φ1, φ2, . . . , φ┌M/B┐ and φ1, φ2, . . . , φ┌M/B┐ so that K equals one of ┌M/B┐ and ┌M/B┐, wherein each subset in the partition except possibly one has B elements.
16. The system of claim 15 wherein said partition according to a block size B is one of: {φ1, . . . , φB}, {φB+1, . . . , φ2B}, . . . , {φB┌M/B┐−B+1, . . . , φM}, and {φ1, . . . , φB}, {φB+1, . . . , φ2B}, . . . , {φB┌M/B┐−2B+1, . . . , φM}, where M>2B.
17. The system of claim 14, wherein subsets of sampling functions φ1, φ2, . . . , φK are largely incoherent, sampling functions φi and φj in a given subset are largely uncorrelated for every i≠j only for indices associated with said ROIs, and sampling functions φi and φj in a given subset are largely correlated for indices not associated with said ROIs.
18. The system of claim 17, wherein said sampling functions are obtained by adaptively modifying a set of largely uncorrelated sampling functions φ={φ1 . . . , φM} in such a way that φii for 1≦i≦M at locations associated with said ROIs, and φik for 1≦i≦M where k is such that when M1+ . . . +Mk−1+1≦i≦M1+ . . . +Mk, kε[M1+ . . . +Mk−1+1, M1+ . . . +Mk] at locations not associated with said ROIs.
19. The system of claim 15, wherein said sampling functions are obtained by adaptively modifying a set of largely uncorrelated sampling functions φ={φ1 . . . , φM} in such a way that φii for 1≦i≦M at locations associated with said ROIs and φik for 1≦i≦M and kε[B┌i/B┐−B+1,B┌i/B┐]∩[1,M], where ┌┐ denotes the ceiling operator which maps a real number to a smallest following integer at locations not associated with said ROIs.
20. The system of claim 15, wherein said sampling functions are obtained by adaptively combining a set of largely uncorrelated sampling functions φ={φ1 . . . , φM} and a set of B-block-wise largely correlated functions α={α1 . . . , αM} in such a way that φii for 1≦i≦M at indices associated with said ROIs, and φii for 1≦i≦M at indices not associated with said ROIs.
21. The system of claim 14, wherein said sampling functions are generated based on an output of a random number generator, wherein generated values of sampling functions associated with said ROIs update at a different rate than those not associated with said ROIs, said output of said random number generator being passed through a deterministic function in order to obtain samples from different distributions.
22. The system of claim 14, wherein said spatial light modulator comprises any of: a digital micromirror device, a transmissive liquid crystal, and reflective liquid crystal on silicon.
23. The system of claim 14, further comprising processing an image of said scene to identify said ROIs using any of: pixel classification, object identification, facial recognition, color, texture, spatial features, spectral information, pattern recognition, motion detection, foreground detection, and a user input
24. The system of claim 14, further comprising processing an image of said scene to generate said mask wherein pixels associated with said ROI as being active and pixels outside said ROI as being inactive.
25. The system of claim 14, wherein different ROIs are associated with different subset lengths B.
US14/085,143 2013-07-01 2013-11-20 Adaptive quality image reconstruction via a compressed sensing framework Expired - Fee Related US9025883B2 (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
US14/085,143 US9025883B2 (en) 2013-07-01 2013-11-20 Adaptive quality image reconstruction via a compressed sensing framework

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
US13/932,791 US9070218B2 (en) 2013-07-01 2013-07-01 Reconstructing an image of a scene captured using a compressed sensing device
US14/085,143 US9025883B2 (en) 2013-07-01 2013-11-20 Adaptive quality image reconstruction via a compressed sensing framework

Related Parent Applications (1)

Application Number Title Priority Date Filing Date
US13/932,791 Continuation-In-Part US9070218B2 (en) 2013-07-01 2013-07-01 Reconstructing an image of a scene captured using a compressed sensing device

Publications (2)

Publication Number Publication Date
US20150003738A1 true US20150003738A1 (en) 2015-01-01
US9025883B2 US9025883B2 (en) 2015-05-05

Family

ID=52115662

Family Applications (1)

Application Number Title Priority Date Filing Date
US14/085,143 Expired - Fee Related US9025883B2 (en) 2013-07-01 2013-11-20 Adaptive quality image reconstruction via a compressed sensing framework

Country Status (1)

Country Link
US (1) US9025883B2 (en)

Cited By (10)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN104574323A (en) * 2015-02-03 2015-04-29 中国人民解放军国防科学技术大学 Quick imaging method on basis of hierarchical models and guide priori for single-pixel cameras
CN104992456A (en) * 2015-06-18 2015-10-21 南京理工大学 Multi-scale matrix coding method
CN106708457A (en) * 2016-12-22 2017-05-24 合肥工业大学 FPGA (Field Programmable Gate Array) processing module and method used for DMD (Digital Micromirror Device) dynamic area selection
US10091440B1 (en) * 2014-05-05 2018-10-02 Lockheed Martin Corporation System and method for providing compressive infrared imaging
CN108765255A (en) * 2018-05-31 2018-11-06 东南大学 Angular quantification index modulation image watermark System and method for based on compressed sensing technology
US10250831B2 (en) * 2015-07-01 2019-04-02 Massachusetts Institute Of Technology Method and apparatus for on-chip per-pixel pseudo-random time coded exposure
US10863127B1 (en) 2018-08-21 2020-12-08 Perceive Corporation Compressive sensing based image capture using multi-lens array
CN112381746A (en) * 2020-11-16 2021-02-19 中南民族大学 Image compressed sensing reconstruction system and method for learning sampling and grouping
US11195061B2 (en) * 2017-09-12 2021-12-07 Panasonic Intellectual Property Management Co., Ltd. Image generation apparatus and method for generating image
EP4207605A1 (en) * 2021-12-29 2023-07-05 Commissariat à l'énergie atomique et aux énergies alternatives Method for compressed sensing and reconstruction of a spectrally sparse signal

Families Citing this family (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US10235608B2 (en) 2015-12-22 2019-03-19 The Nielsen Company (Us), Llc Image quality assessment using adaptive non-overlapping mean estimation
CN107796788B (en) * 2016-08-29 2021-02-23 南京理工大学 Sensing matrix measurement method based on variational Bayesian expectation maximization algorithm
CN106815876B (en) * 2016-12-30 2019-08-02 清华大学 Image sparse characterizes the combined optimization training method of more dictionary learnings
US11877079B2 (en) 2020-12-22 2024-01-16 Samsung Electronics Co., Ltd. Time-resolving computational image sensor architecture for time-of-flight, high-dynamic-range, and high-speed imaging

Citations (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20140152986A1 (en) * 2004-03-06 2014-06-05 Michael Trainer Methods and apparatus for determining characteristics of particles

Patent Citations (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20140152986A1 (en) * 2004-03-06 2014-06-05 Michael Trainer Methods and apparatus for determining characteristics of particles

Cited By (17)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US10091440B1 (en) * 2014-05-05 2018-10-02 Lockheed Martin Corporation System and method for providing compressive infrared imaging
CN104574323A (en) * 2015-02-03 2015-04-29 中国人民解放军国防科学技术大学 Quick imaging method on basis of hierarchical models and guide priori for single-pixel cameras
CN104992456A (en) * 2015-06-18 2015-10-21 南京理工大学 Multi-scale matrix coding method
US11050963B2 (en) 2015-07-01 2021-06-29 Massachusetts Institute Of Technology Method and apparatus for on-chip per-pixel pseudo-random time coded exposure
US10250831B2 (en) * 2015-07-01 2019-04-02 Massachusetts Institute Of Technology Method and apparatus for on-chip per-pixel pseudo-random time coded exposure
US10616520B2 (en) 2015-07-01 2020-04-07 Massachusetts Institute Of Technology Method and apparatus for on-chip per-pixel pseudo-random time coded exposure
CN106708457A (en) * 2016-12-22 2017-05-24 合肥工业大学 FPGA (Field Programmable Gate Array) processing module and method used for DMD (Digital Micromirror Device) dynamic area selection
US11195061B2 (en) * 2017-09-12 2021-12-07 Panasonic Intellectual Property Management Co., Ltd. Image generation apparatus and method for generating image
CN108765255A (en) * 2018-05-31 2018-11-06 东南大学 Angular quantification index modulation image watermark System and method for based on compressed sensing technology
US10887537B1 (en) * 2018-08-21 2021-01-05 Perceive Corporation Compressive sensing based image capture using dynamic masking
US10937196B1 (en) 2018-08-21 2021-03-02 Perceive Corporation Compressive sensing based image capture device
US10885674B1 (en) 2018-08-21 2021-01-05 Perceive Corporation Training network for compressive sensing based image processing
US11094090B1 (en) 2018-08-21 2021-08-17 Perceive Corporation Compressive sensing based image capture using diffractive mask
US10863127B1 (en) 2018-08-21 2020-12-08 Perceive Corporation Compressive sensing based image capture using multi-lens array
US11244477B1 (en) 2018-08-21 2022-02-08 Perceive Corporation Compressive sensing based image processing
CN112381746A (en) * 2020-11-16 2021-02-19 中南民族大学 Image compressed sensing reconstruction system and method for learning sampling and grouping
EP4207605A1 (en) * 2021-12-29 2023-07-05 Commissariat à l'énergie atomique et aux énergies alternatives Method for compressed sensing and reconstruction of a spectrally sparse signal

Also Published As

Publication number Publication date
US9025883B2 (en) 2015-05-05

Similar Documents

Publication Publication Date Title
US9025883B2 (en) Adaptive quality image reconstruction via a compressed sensing framework
US9412185B2 (en) Reconstructing an image of a scene captured using a compressed sensing device
US10861143B2 (en) Method and apparatus for reconstructing hyperspectral image using artificial intelligence
Baraniuk et al. Compressive video sensing: Algorithms, architectures, and applications
Liang Punching holes in light: recent progress in single-shot coded-aperture optical imaging
Cao et al. Computational snapshot multispectral cameras: Toward dynamic capture of the spectral world
US8970740B2 (en) Overlap patterns and image stitching for multiple-detector compressive-sensing camera
US9081731B2 (en) Efficient transforms and efficient row generation for Kronecker products of Hadamard matrices
Tan et al. Compressive hyperspectral imaging via approximate message passing
Patel et al. Sparse representations and compressive sensing for imaging and vision
CN107306333B (en) High-speed single-pixel imaging method
US10302491B2 (en) Imaging method and apparatus
Yu et al. Compressive microscopic imaging with “positive–negative” light modulation
US9277139B2 (en) Generating modulation patterns for the acquisition of multiscale information in received signals
US20150116563A1 (en) Adaptive Sensing of a Programmable Modulator System
US9344736B2 (en) Systems and methods for compressive sense imaging
Yu et al. Single-photon compressive imaging with some performance benefits over raster scanning
Llull et al. Temporal compressive sensing for video
Kuusela Single-pixel camera
Güngör et al. A matrix-free reconstruction method for compressive focal plane array imaging
Yuan et al. Lensless compressive imaging
Xu et al. A compressive hyperspectral video imaging system using a single-pixel detector
Rueda et al. Dual-ARM VIS/NIR compressive spectral imager
US20140211000A1 (en) Sensing Signals with Affine-Harmonically Related Rows of Kronecker-Product Matrices
Edeler et al. Multi image super resolution using compressed sensing

Legal Events

Date Code Title Description
AS Assignment

Owner name: XEROX CORPORATION, CONNECTICUT

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:BERNAL, EDGAR A.;WEN, XUEJIN;WU, WENCHENG;AND OTHERS;REEL/FRAME:031640/0480

Effective date: 20131119

FEPP Fee payment procedure

Free format text: PAYOR NUMBER ASSIGNED (ORIGINAL EVENT CODE: ASPN); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY

STCF Information on status: patent grant

Free format text: PATENTED CASE

FEPP Fee payment procedure

Free format text: MAINTENANCE FEE REMINDER MAILED (ORIGINAL EVENT CODE: REM.); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY

LAPS Lapse for failure to pay maintenance fees

Free format text: PATENT EXPIRED FOR FAILURE TO PAY MAINTENANCE FEES (ORIGINAL EVENT CODE: EXP.); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY

STCH Information on status: patent discontinuation

Free format text: PATENT EXPIRED DUE TO NONPAYMENT OF MAINTENANCE FEES UNDER 37 CFR 1.362

FP Lapsed due to failure to pay maintenance fee

Effective date: 20190505