CN115131233A - Intensity background light suppression method for computational imaging - Google Patents
Intensity background light suppression method for computational imaging Download PDFInfo
- Publication number
- CN115131233A CN115131233A CN202210661148.4A CN202210661148A CN115131233A CN 115131233 A CN115131233 A CN 115131233A CN 202210661148 A CN202210661148 A CN 202210661148A CN 115131233 A CN115131233 A CN 115131233A
- Authority
- CN
- China
- Prior art keywords
- polynomial
- background light
- intensity
- diffraction
- spot
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Granted
Links
- 238000000034 method Methods 0.000 title claims abstract description 47
- 238000003384 imaging method Methods 0.000 title claims abstract description 23
- 230000001629 suppression Effects 0.000 title claims abstract description 16
- 238000011084 recovery Methods 0.000 claims description 15
- 238000004364 calculation method Methods 0.000 claims description 11
- 239000011159 matrix material Substances 0.000 claims description 4
- 238000012545 processing Methods 0.000 claims description 3
- 230000015572 biosynthetic process Effects 0.000 claims 1
- 238000002591 computed tomography Methods 0.000 claims 1
- 230000003287 optical effect Effects 0.000 abstract description 4
- 238000010586 diagram Methods 0.000 description 3
- 238000009826 distribution Methods 0.000 description 3
- 230000008030 elimination Effects 0.000 description 3
- 238000003379 elimination reaction Methods 0.000 description 3
- 238000005259 measurement Methods 0.000 description 3
- 238000012546 transfer Methods 0.000 description 3
- 230000004075 alteration Effects 0.000 description 2
- 238000001514 detection method Methods 0.000 description 2
- 230000000694 effects Effects 0.000 description 2
- 238000007792 addition Methods 0.000 description 1
- 238000003705 background correction Methods 0.000 description 1
- 230000009286 beneficial effect Effects 0.000 description 1
- 238000007796 conventional method Methods 0.000 description 1
- 230000007547 defect Effects 0.000 description 1
- 230000001419 dependent effect Effects 0.000 description 1
- 238000003745 diagnosis Methods 0.000 description 1
- 238000012986 modification Methods 0.000 description 1
- 230000004048 modification Effects 0.000 description 1
- 238000012634 optical imaging Methods 0.000 description 1
- 238000001454 recorded image Methods 0.000 description 1
- 238000011160 research Methods 0.000 description 1
- 230000000452 restraining effect Effects 0.000 description 1
- 238000010869 super-resolution microscopy Methods 0.000 description 1
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T5/00—Image enhancement or restoration
- G06T5/90—Dynamic range modification of images or parts thereof
- G06T5/94—Dynamic range modification of images or parts thereof based on local image properties, e.g. for local contrast enhancement
-
- G—PHYSICS
- G01—MEASURING; TESTING
- G01S—RADIO DIRECTION-FINDING; RADIO NAVIGATION; DETERMINING DISTANCE OR VELOCITY BY USE OF RADIO WAVES; LOCATING OR PRESENCE-DETECTING BY USE OF THE REFLECTION OR RERADIATION OF RADIO WAVES; ANALOGOUS ARRANGEMENTS USING OTHER WAVES
- G01S17/00—Systems using the reflection or reradiation of electromagnetic waves other than radio waves, e.g. lidar systems
- G01S17/88—Lidar systems specially adapted for specific applications
- G01S17/89—Lidar systems specially adapted for specific applications for mapping or imaging
Landscapes
- Physics & Mathematics (AREA)
- Engineering & Computer Science (AREA)
- General Physics & Mathematics (AREA)
- Computer Networks & Wireless Communication (AREA)
- Electromagnetism (AREA)
- Radar, Positioning & Navigation (AREA)
- Remote Sensing (AREA)
- Theoretical Computer Science (AREA)
- Testing Of Optical Devices Or Fibers (AREA)
- Photometry And Measurement Of Optical Pulse Characteristics (AREA)
Abstract
The invention discloses an intensity background light suppression method for computational imaging, which comprises the steps of firstly, simulating an actual optical system through mathematical modeling to obtain the size of a diffraction spot area with a high signal-to-noise ratio, and further establishing a support domain for establishing a background light model; and secondly, re-segmenting the polynomial according to the background light support domain, fitting and representing the background light outside the background light support domain to obtain a coefficient, combining the unsegmented polynomial and the coefficient to fit and represent the background light of the full-spot image, and removing the background light from the acquired intensity. According to the invention, the support domain outside the light spot is obtained through mathematical modeling, and the suppression and removal of the background light are realized based on the polynomial, so that the image acquisition process of the computed imaging is simplified, the application range of the computed imaging is expanded, and the accuracy of wavefront reconstruction is improved.
Description
Technical Field
The invention relates to the technical field of optical measurement, in particular to an intensity background light suppression method for computational imaging.
Background
Non-interference phase recovery is a widely used computational imaging technique, which greatly simplifies the hardware requirements of the conventional optical imaging system and reduces the influence of inherent aberration. Iterative phase recovery, one of the methods of non-interferometric PR, allows reconstruction of the required field from a series of acquired images.
For example, chinese patent publication No. CN105403508A discloses a non-interference phase imaging method based on a synthetic phase transfer function, which includes first shooting a set of light intensity images at different depths where objects are exponentially distributed along an optical axis direction at intervals with a camera, then solving an initial phase map according to the synthetic phase transfer function assuming that the objects are weakly absorbing and weakly phased, and finally substituting the initial phase map into an iterative compensation algorithm of the synthetic phase transfer function to solve an accurate phase map.
Chinese patent publication No. CN112629678A discloses a fast phase recovery method for general shape non-diffraction iterative computation, which decomposes the wavefront to be measured using a numerical orthogonal polynomial pattern, then computes a diffraction basis function for each numerical orthogonal polynomial based on fast fourier transform, and then iteratively solves a coefficient gradient between diffraction surfaces by using matrix operation, thereby realizing the wavefront detection of a high-speed general shape.
The high flexibility and high versatility make iterative PR an easy-to-access and easy-to-deploy tool, particularly suitable for research and diagnosis of wavefront sensing, super-resolution microscopy, and quantitative phase imaging. Iterative PR uses forward and backward fourier transforms to establish the relationship between the inverted field and the measured intensity field, typically involving estimating complex distributions from the corresponding intensity distributions under the constraints of experimental model parameters. Iterative PR is an inverse problem in optics and may be non-unique. The multi-image PR preferably eliminates local minimum lag.
The performance of PR methods depends fundamentally on the quality of the recorded intensity data and the accuracy of the match between the numerical PR model and the actual experimental system. Relaxed support constraints can also affect reconstruction accuracy. Acquired intensity images with low background and noise levels are a necessary condition for a high precision iterative PR method. In the PR model, the dominant noise sources are camera noise, including dark shot noise, read noise, and photon shot noise. Dark shot noise subject to poisson statistics is independent of signal level, but dependent on the temperature of the sensor; the reading noise is a fixed noise of the camera; photon shot noise follows a poisson distribution as intensity varies. Furthermore, the influence of the background is not negligible. Ambient light in an optical plant can cause background non-uniformity in the recorded image.
Currently, researchers pay more attention to noise removal and neglect background light suppression, and the quality of the background light suppression directly affects the accuracy of phase recovery wavefront detection.
Disclosure of Invention
Aiming at the defects of the prior art, the invention provides an intensity background light suppression method for computational imaging, which obtains a support domain outside a light spot through mathematical modeling, realizes background light suppression and removal based on a polynomial, simplifies the acquisition process of a computational imaging image, expands the application range of computational imaging and improves the precision of wavefront reconstruction.
A method for restraining background light with intensity for calculating imaging is applied to a scene of collecting a light intensity diagram, wherein a beam expander, a flat plate to be tested, a convergent lens and an image sensor are sequentially arranged on an emergent light path of a laser, and the method comprises the following steps:
s1: moving the image sensor, and acquiring n defocused diffraction spots containing the wavefront error of the flat plate to be detected at different defocused distances;
s2: respectively setting the focal length s of the convergent lens, the caliber D and the out-of-focus position z of each out-of-focus diffraction spot j Setting the initial complex amplitude of the measuring surface as a full matrix;
s3: setting a pixel size d of an image sensor n And the number of pixels M 0 ×N 0 For each defocusing diffraction spot, obtaining effective defocusing diffraction spotsCentral position (x) of j ,y j ) And off-focus diffraction spots with dimension of M multiplied by M;
s4: based on scalar diffraction theory, calculating the initial complex amplitude diffraction of the measuring surface to each out-of-focus surface according to the parameters of the imaging system to obtain the estimated light spot of each out-of-focus surfaceFor estimating light spotsBinarization is carried out to obtain a binary mask M j ;
S5: for each effectively defocused diffraction spotThe background light is roughly removed, namely the average value of redundant light intensity is calculated by the formula
Wherein A is j Is the average of the redundant light intensities;
s6 effective defocusing diffraction spot processed by the S5Reconstructing the wavefront to be measured by using a phase recovery algorithm;
s7, calculating the corresponding defocused diffraction light spots based on the wave front to be measured reconstructed in S6, and calculating to obtain the corresponding mask plate M 'by using the method in S4' j Using the mask plate M' j Calculating to obtain N by polynomial processing z Polynomial of item number
S8: calculating the obtained mask M 'by using S7' j Extracting redundant light intensity, wherein the formula is as follows:
fitting the redundant light intensity by using the numerical polynomial S7 to obtain a fitting coefficient, wherein the formula is as follows:
s9: calculating the background light intensity S j First, before calculation using S8Term coefficient combined with zernike polynomial Z i (u, v) superimposing, and superimposing the rest coefficients with a numerical polynomial;
s10: removing background light intensity S j The formula is as follows:
s11: and reconstructing the wavefront to be measured by using a phase recovery algorithm.
Further, in step S3, the effective defocus diffraction spot is calculated based on the centroid methodCentral position (x) of j ,y j ) And an out-of-focus diffraction spot of dimension mxm.
wherein σ 1 Is an empirical value and j is the index of the jth out-of-focus diffracted spot.
In step S5, the average value of the redundant light intensity is calculated by the following formula:
in the formula, n and m represent the order of the polynomial.
In steps S6 and S11, the phase recovery algorithm is the GS algorithm.
In step S7, the mask plate M 'is used' j To polynomial positionPhysical calculation to obtain N z Polynomial of term valueThe formula of (1) is:
wherein Z is i (u, v) is a Zernike polynomial applied to calculate a numerical polynomial, and i is an index of the polynomial.
In step S8, the method of calculating the coefficient is a least square method.
In step S9, the background light intensity S is calculated j The formula of (1) is:
in the formula, c i And fitting the redundant light intensity for the numerical polynomial to obtain a coefficient.
Compared with the prior art, the invention has the following beneficial effects:
according to the invention, background light of diffraction spots is removed through polynomial fitting, and firstly, a non-uniform background can be corrected by using a numerical polynomial; the other is to predict the background of the light intensity effective part by using some polynomials according to the fact that the wavefront of the natural light is smooth. During image reconstruction, a background correction method may be iteratively performed to enhance the effect of removing the background.
Drawings
FIG. 1 is a schematic diagram of an apparatus for calculating an intensity background light suppression method for imaging according to an embodiment of the present invention;
FIG. 2 is a flow chart of an intensity backlight suppression method for computed imaging according to the present invention;
FIG. 3 is a diagram illustrating the recovery results of the method and the mean value elimination method according to the present invention.
Detailed Description
The invention will be described in further detail below with reference to the drawings and examples, which are intended to facilitate the understanding of the invention without limiting it in any way.
As shown in fig. 1, as an example of collecting a light intensity pattern, a beam expander 2, a flat plate 3 to be measured, a converging lens 4, and an image sensor 5 are sequentially arranged on an outgoing light path of a laser 1, and the image sensor 5 is fixed on a precision guide rail 6.
As shown in fig. 2, an intensity background light suppression method for computed imaging includes the following steps:
s1: and moving the image sensor, and acquiring n defocused diffraction spots containing the wavefront error of the flat plate to be detected at different defocused distances.
S2: respectively setting the focal length s, the caliber D and the defocusing position z of each defocusing diffraction spot of the convergent lens k The initial complex amplitude of the measurement plane is set to be a full matrix.
S3: setting image sensor pixel size d n And the number of pixels M 0 ×N 0 For each defocusing diffraction spot, obtaining effective defocusing diffraction spotsCentral position (x) of j ,y j ) And an out-of-focus diffraction spot of dimension mxm.
S4: based on scalar diffraction theory, calculating the initial complex amplitude diffraction of the measuring surface to each out-of-focus surface according to the parameters of the imaging system to obtain the estimated light spot of each out-of-focus surfaceBinarizing the estimated light spot to obtain a binary mask
Wherein σ 1 Is an empirical value, j is the index of the jth out-of-focus diffraction spot;
s5: for each effective separationFocal diffraction spotRough removal of background light, i.e. average of redundant light intensity
Wherein the average value of the redundant light intensities is calculated by the following formula:
s6, reconstructing the wavefront to be detected by using the defocused diffraction spots processed by the S5 and a phase recovery algorithm;
s7, calculating corresponding defocused diffraction light spots based on the wave front to be measured reconstructed in S6, and calculating corresponding mask plate M 'by using the method in S4' j Using the mask plate M' j Calculating the polynomial to obtain N z Polynomial of term value
Wherein, Zernike polynomial Z i (u, v) is applied to the arithmetic numerical polynomial and i is the index of the polynomial.
S8: mask M 'prepared from S7' j The redundant light intensity is extracted and,
fitting the redundant light intensity by using the numerical polynomial in the S6 to obtain a coefficient;
s9: background light intensity calculation first uses S8 before calculationThe term coefficients are combined with polynomial superposition, and the rest coefficients are superposed with numerical polynomial
S10: the background light is removed
S11: and reconstructing the wavefront to be measured by using a phase recovery algorithm.
The center position calculation method and the effective spot area determination described in consideration of the calculation speed and the calculation accuracy S3 are calculated based on the centroid method.
The phase recovery algorithm described in S6 and S11 is the GS algorithm, taking into account the speed and accuracy requirements of the wavefront reconstruction.
In order to consider the correspondence between the polynomial and the aberration and the orthogonality of the polynomial, the polynomials in S7 and S9 are zernike polynomials.
The method of calculating coefficients described in S8 is a least square method in consideration of the polynomial fitting accuracy.
A specific example of the method of the present invention is given below to illustrate the technical effect of the method of the present invention.
Here, s is 1079 for focal length selection . 41mm,z 1 ,z 2 ,z 3 =[-10,-15,-20]mm, bore D22 . 9mm, total number of iterations N of wavefront measurement 5000, number N of Zernike polynomials z =50,The effective spot number is 1024 × 1024.
Three defocusing diffraction images are collected to carry out wave front phase and amplitude reconstruction, non-uniform background light is added in the three defocusing diffraction spots, the selected diffraction calculation model is a Fresnel diffraction model, and FIG. 3 is a recovery result graph of the method and the average value elimination method. Wherein (a) and (d) are the recovered phase and amplitude of the method of the present invention, (b) and (e) are the recovered phase and amplitude of the mean value elimination method, and (c) and (f) are the phase and amplitude of the real map. It can be seen from the figure that the method proposed herein is more efficient than the conventional method.
The embodiments described above are intended to illustrate the technical solutions and advantages of the present invention, and it should be understood that the above-mentioned embodiments are only specific embodiments of the present invention, and are not intended to limit the present invention, and any modifications, additions and equivalents made within the scope of the principles of the present invention should be included in the scope of the present invention.
Claims (8)
1. The utility model provides an intensity background light suppression method for calculation formation of image, is applied to the scene that the light intensity map was gathered, has arranged beam expander, the flat board that awaits measuring, convergent lens, image sensor in proper order on the emergent light path of laser instrument, its characterized in that includes following steps:
s1: moving the image sensor, and acquiring n defocused diffraction spots containing the wavefront error of the flat plate to be detected at different defocused distances;
s2: respectively setting the focal length s, the caliber D and the defocusing position z of each defocusing diffraction spot of the convergent lens j Setting the initial complex amplitude of the measuring surface as a full matrix;
s3: setting a pixel size d of an image sensor n And the number of pixels M 0 ×N 0 For each defocusing diffraction spot, obtaining effective defocusing diffraction spotsCentral position (x) of j ,y j ) And defocusing diffraction spots with dimension of M multiplied by M;
s4: based on scalar diffraction theory, according to the parameters of the imaging system, the initial complex amplitude diffraction of the measuring surface is calculated to each defocusing surface to obtain the estimated light spot of each defocusing surfaceFor estimating light spotBinarization is carried out to obtain a binary mask M j ;
S5: for each effectively defocused diffraction spotThe background light is roughly removed, namely the average value of redundant light intensity is calculated by the formula
Wherein, A j Is the average of the redundant light intensities;
s6 effective defocusing diffraction spot processed by the S5Reconstructing the wavefront to be detected by using a phase recovery algorithm;
s7, calculating corresponding defocused diffraction light spots based on the wave front to be detected reconstructed in S6, and calculating to obtain corresponding mask plate M 'by using the method in S4' j Using the mask plate M' j Processing the polynomial to obtain N z Polynomial of term value
S8: mask M 'obtained by calculation through S7' j Extracting redundant light intensity, wherein the formula is as follows:
fitting the redundant light intensity by using the numerical polynomial S7 to obtain a coefficient, wherein the formula is as follows:
s9: calculating the background light intensity S j First, before calculation using S8The coefficient of the neck in combination with Zernike polynomials Z i (u, v) superimposing, and superimposing the rest coefficients with a numerical polynomial;
s10: removing background light intensity S j The formula is as follows:
s11: and reconstructing the wavefront to be detected by using a phase recovery algorithm.
5. The method of claim 1, wherein the phase-recovery algorithm is GS algorithm in steps S6 and S11.
6. The intensity background light suppression method for computed tomography according to claim 1, wherein in step S7, the mask plate M 'is used' j Processing the polynomial to obtain N z Polynomial of item numberThe formula of (1) is:
wherein Z is i (u, v) is a Zernike polynomial applied to calculate a numerical polynomial, and i is an index of the polynomial.
7. The method for suppressing background light with intensity used for computed imaging according to claim 6, wherein in step S8, the method for computing the coefficient is least squares method.
8. Intensity backlight suppression method for computed imaging according to claim 6The method is characterized in that in step S9, the background light intensity S is calculated j The formula of (1) is:
in the formula, c i And fitting the redundant light intensity for the numerical polynomial to obtain a coefficient.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202210661148.4A CN115131233B (en) | 2022-06-13 | 2022-06-13 | Intensity background light inhibition method for computational imaging |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202210661148.4A CN115131233B (en) | 2022-06-13 | 2022-06-13 | Intensity background light inhibition method for computational imaging |
Publications (2)
Publication Number | Publication Date |
---|---|
CN115131233A true CN115131233A (en) | 2022-09-30 |
CN115131233B CN115131233B (en) | 2024-06-25 |
Family
ID=83378452
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN202210661148.4A Active CN115131233B (en) | 2022-06-13 | 2022-06-13 | Intensity background light inhibition method for computational imaging |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN115131233B (en) |
Citations (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20170168285A1 (en) * | 2015-12-14 | 2017-06-15 | The Regents Of The University Of California | Systems and methods for image reconstruction |
CN113188671A (en) * | 2021-04-27 | 2021-07-30 | 浙江大学 | Wavefront detection method based on cross iteration automatic position correction |
US20220003633A1 (en) * | 2019-07-18 | 2022-01-06 | Zhejiang University | Device and method for detecting wavefront error by modal-based optimization phase retrieval using extended nijboer-zernike theory |
-
2022
- 2022-06-13 CN CN202210661148.4A patent/CN115131233B/en active Active
Patent Citations (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20170168285A1 (en) * | 2015-12-14 | 2017-06-15 | The Regents Of The University Of California | Systems and methods for image reconstruction |
US20220003633A1 (en) * | 2019-07-18 | 2022-01-06 | Zhejiang University | Device and method for detecting wavefront error by modal-based optimization phase retrieval using extended nijboer-zernike theory |
CN113188671A (en) * | 2021-04-27 | 2021-07-30 | 浙江大学 | Wavefront detection method based on cross iteration automatic position correction |
Non-Patent Citations (2)
Title |
---|
"Hybrid phase retrieval from a single defocused image", 《SPIE》, 5 August 2015 (2015-08-05) * |
李斐;饶长辉;: "基于离焦光栅的最小二乘波前复原算法", 中国激光, no. 11, 10 November 2010 (2010-11-10) * |
Also Published As
Publication number | Publication date |
---|---|
CN115131233B (en) | 2024-06-25 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US9448160B2 (en) | Method and apparatus for providing image data for constructing an image of a region of a target object | |
AU2002362137B2 (en) | Systems and methods for wavefront measurement | |
WO2004052189A1 (en) | Optical wavefront analyzer | |
EP2890975B1 (en) | Estimation of background radiation from ptychography | |
CN108254295B (en) | Method and device for positioning and representing spherical particles | |
CN112629677B (en) | Rapid large-dynamic-range wavefront detection device and detection method based on mode recovery | |
CN112903121B (en) | Wavefront detection method based on cross iterative automatic focusing | |
CN113188671B (en) | Wavefront detection method based on cross iteration automatic position correction | |
CN114241072B (en) | Laminated imaging reconstruction method and system | |
CN113393472B (en) | Zoning composite phase unwrapping method based on digital holographic microscopic imaging | |
CN112629678B (en) | Rapid phase recovery method for general shape diffraction-free iterative computation | |
CN113298700A (en) | High-resolution image reconstruction method in scattering scene | |
CN115131233A (en) | Intensity background light suppression method for computational imaging | |
WO2024087406A1 (en) | Digital holographic wrapped phase distortion compensation method based on deep learning | |
KR101166961B1 (en) | Wavefront aberration retrieval method by 3D beam measurement | |
JP2021051038A (en) | Aberration estimation method, aberration estimation device, program, and recording medium | |
CN113432731B (en) | Compensation method in grating transverse shearing interference wavefront reconstruction process | |
CN113405676B (en) | Correction method for micro-vibration influence in phase difference wavefront detection of space telescope | |
CN113189101B (en) | Lens-free imaging method with negative feedback adjustment | |
CN115014545A (en) | Phase recovery wavefront detection method without support domain constraint | |
CN112857750B (en) | Expanded target wavefront detection method based on edge enhancement | |
Banerjee | Recent advances in digital holography: referenceless holography using transport of intensity and its applications | |
CN112258628A (en) | Imaging system and method based on grating phase contrast | |
Holeva | Range estimation from camera blur by regularized adaptive identification | |
AU2004201109A1 (en) | Method and apparatus for producing a phase image of an object |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
GR01 | Patent grant |