CN110415201A - Single exposure super-resolution imaging method and device based on structure light and deep learning - Google Patents

Single exposure super-resolution imaging method and device based on structure light and deep learning Download PDF

Info

Publication number
CN110415201A
CN110415201A CN201910697301.7A CN201910697301A CN110415201A CN 110415201 A CN110415201 A CN 110415201A CN 201910697301 A CN201910697301 A CN 201910697301A CN 110415201 A CN110415201 A CN 110415201A
Authority
CN
China
Prior art keywords
model
neural network
predetermined depth
training
resolution image
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN201910697301.7A
Other languages
Chinese (zh)
Other versions
CN110415201B (en
Inventor
边丽蘅
李道钰
张军
曹先彬
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Beijing University of Technology
Beijing Institute of Technology BIT
Original Assignee
Beijing University of Technology
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Beijing University of Technology filed Critical Beijing University of Technology
Priority to CN201910697301.7A priority Critical patent/CN110415201B/en
Publication of CN110415201A publication Critical patent/CN110415201A/en
Application granted granted Critical
Publication of CN110415201B publication Critical patent/CN110415201B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T5/00Image enhancement or restoration
    • G06T5/50Image enhancement or restoration by the use of more than one image, e.g. averaging, subtraction
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/10Image acquisition modality
    • G06T2207/10052Images from lightfield camera
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/20Special algorithmic details
    • G06T2207/20081Training; Learning
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/20Special algorithmic details
    • G06T2207/20084Artificial neural networks [ANN]

Abstract

Single exposure super-resolution imaging method and device based on structure light and deep learning that the invention discloses a kind of, wherein generate composite construction optical illumination coding this method comprises: being overlapped to multiple and different traditional lightings coding;It acquires target scene to be processed and encodes the low-resolution image generated under single exposure in composite construction optical illumination;Low-resolution image is inputted default imaging model to handle, exports the corresponding super-resolution image of target scene to be processed.This method has the support of deep learning theory and structure light coding illumination theory, and implementation is simple, it is only necessary to which the quick reconstruction of super resolution image can be realized in single exposure.

Description

Single exposure super-resolution imaging method and device based on structure light and deep learning
Technical field
The present invention relates to the super-resolution imaging technical fields calculated in camera shooting, in particular to a kind of to be based on structure light and depth Single exposure super-resolution imaging method and device of degree study.
Background technique
The resolution ratio of optical microscopy imaging is limited by optical system diffraction characteristic, and in recent years, researcher is using now There is imaging system to restore to pay a large amount of effort in terms of the image of higher resolution.2000, University of California San Francisco point Doctor M.Gustafsson in school proposes Structured Illumination microscopy (Structured Illumination Microscopy, SIM).SIM irradiates object by structure light coding lighting source (usually sinusoidal coding), in Fourier's sky Between frequency domain, Structured Illumination encoded frequency spectrum and sample preimage frequency spectrum convolution, the optical delivery of micro-imaging process will be located at this Information except function (Optical Transfer Function, OTF) is transferred within the scope of Observable, passes through special algorithm The low-frequency information in high-frequency information and OTF outside reduction splicing OTF range, to go back original high resolution image.SIM imaging is differentiated Rate reaches as high as twice of former Optical Resolution of Imaging System.2005, Gustafsson was on the basis of SIM it is further proposed that saturation Structured Illumination microscopy (Saturated Structure Illumination Microscopy, SSIM), SSIM utilize base Sample is irradiated in the nonlinear organization light of fluorescent molecule excitation state saturation effect, illumination light contains the higher hamonic wave point of spatial frequency Amount, therefore the full resolution pricture for being higher than former more times of Optical Resolution of Imaging System can be extracted.
Specifically, SIM is raw by special algorithm by the low resolution coded image for acquiring the illumination of multiple structure light codings At a full resolution pricture, traditional SIM Image Restoration Algorithm has direct method, Bayes' assessment and Gerchberg-Saxton Algorithm (G-S algorithm) etc..
Since the image Fourier space frequency spectrum encoded through SIN function is considered as low resolution image frequency spectrum and two The combination of the Sideband Spectrum, direct method irradiate sample by the sinusoidal coding structure light of multiple and different phases of coding mode of the same race, The low resolution coded image of multiple groups is acquired, obtains its Fourier space frequency spectrum through two-dimensional Fourier transform, simultaneous equations solve OTF model With the frequency spectrum outside OTF range in enclosing, and it is combined splicing and can secures satisfactory grades and distinguish image spectrum, then obtained through two-dimentional inverse Fourier transform To high-definition picture.Direct method calculating process is simple, but needs to acquire multiple low resolution coded images to restore a Zhang Gaofen Distinguish image, and poor robustness, the noise introduced to actual imaging process is more sensitive.
Bayes' assessment is based on Bayesian statistical model, it is known that the high-definition picture and low resolution code pattern of scene The statistical law of picture, and by high-definition picture to the create-rule of low resolution coded image, according to Bayesian statistics mould Type calculates the statistical model of the corresponding high-definition picture of low resolution coded image, can pass through the modes such as the Posterior Mean estimation technique Restore high-definition picture.But Bayes' assessment is also required to acquire multiple low resolution coded images to restore a High-Resolution Map Picture.
G-S algorithm using it is low resolution coded image and its Fourier space frequency spectrum establish target full resolution pricture airspace and Frequency-domain constraint, iteration update target image and its Fourier space frequency spectrum, until the error arrival between iteration result is set twice Determine threshold value.Compared to direct method and Bayes' assessment, low resolution coded image needed for G-S algorithm is less, and robustness is stronger, but by In needing iteration to update, so real-time is poor.
In recent years, with the development of deep neural network, researcher starts to carry out the super of image with deep learning theory Research is differentiated, Hong Kong Chinese University Tang Xiaoou professor seminar in 2016 carries out oversubscription to single image using convolutional neural networks Distinguish recovery.Then, researcher proposes using deeper convolutional neural networks to improve resolution ratio.U.S. Twitter in 2017 Company proposes to utilize the super-resolution for generating confrontation network progress single image.It is however generally that true full resolution pricture institute It is higher than individual low resolution image information contained amount containing information content, so only restoring high-resolution by a untreated low resolution image Original image is ill-conditioning problem, and deep learning theory carries out super-resolution recovery based on the statistical law of image, and there is no at this This is solved the problems, such as in matter, so being only not considered as true high-resolution by the full resolution pricture that deep neural network study restores Image.
Summary of the invention
The present invention is directed to solve at least some of the technical problems in related technologies.
For this purpose, exposing super-resolution based on structure light and the single of deep learning an object of the present invention is to provide a kind of Imaging method, this method have the support of deep learning theory and structure light coding illumination theory, and implementation is simple, it is only necessary to single It is secondary to expose the quick reconstruction that super resolution image can be realized.
It is another object of the present invention to propose a kind of single exposure super-resolution based on structure light and deep learning at As device.
In order to achieve the above objectives, one aspect of the present invention embodiment proposes a kind of single exposure based on structure light and deep learning Light super-resolution imaging method, comprising:
S1 is overlapped multiple and different traditional lightings coding and generates composite construction optical illumination coding;
S2 acquires target scene to be processed and encodes the low resolution generated under single exposure in the composite construction optical illumination Image;
The low-resolution image is inputted default imaging model and handled, exports the target scene to be processed by S3 Corresponding super-resolution image.
Single exposure super-resolution imaging method based on structure light and deep learning of the embodiment of the present invention, according to structure light The Fourier space spectral characteristic of coded illumination image, analysis Structured Illumination in Fourier space frequency domain are encoded to image It influences, to establish composite illuminating coding;Based on deep learning theory, building is suitable for what structure light coding illumination image restored Deep neural network model;It is exposed by single composite construction pumped FIR laser, carries out super resolution image in conjunction with deep neural network It quickly rebuilds, this method has the support of deep learning theory and structure light coding illumination theory, and implementation is simple, it is only necessary to single It is secondary to expose the quick reconstruction that super resolution image can be realized.
In addition, single exposure super-resolution imaging side according to the above embodiment of the present invention based on structure light and deep learning Method can also have following additional technical characteristic:
Further, in one embodiment of the invention, before step S2, further includes:
Acquisition Training scene encodes the low-resolution image training generated under single exposure in the composite construction optical illumination Sample;
The low-resolution image training sample is trained in predetermined depth neural network model;
If the error amount between training result and the high-definition picture sample of the Training scene is less than preset threshold, The default imaging model is set by the trained predetermined depth neural network model;
If the error amount between the training result and the high-definition picture sample of the Training scene is more than or equal to pre- If threshold value, then the parameter of the predetermined depth neural network is updated according to error amount and the low-resolution image is trained into sample Originally updated predetermined depth neural network model is input to be trained again.
Further, in one embodiment of the invention, the multiple different traditional lighting codings include: SIN function Coding and/or Nonlinear fluorescence coding and/or random coded;
It is superimposed to obtain the composite construction optical illumination coding by the airspace for encoding the multiple traditional lighting, it is described multiple The Fourier space frequency spectrum for closing Structured Illumination coding is the superposition of the multiple different traditional lighting coding frequency spectrums.
Further, in one embodiment of the invention, the predetermined depth neural network model is convolutional Neural net The low-resolution image training sample of acquisition is inputted the predetermined depth neural network model and is trained by network model;
It is output high-definition picture by the output after the predetermined depth neural network model, high-resolution will be exported The high-definition picture sample of image and the Training scene compares to obtain error amount;
If error amount is less than preset threshold, set described default for the trained predetermined depth neural network model Imaging model;
If error amount is more than or equal to preset threshold, the parameter of the predetermined depth neural network is updated according to error amount And the low-resolution image training sample is input to the updated predetermined depth neural network model and is trained again.
Further, in one embodiment of the invention, the predetermined depth neural network model is the generation pair Anti- network model;
Generation confrontation network model includes generating model and discrimination model, and the generation model is by the low resolution Image training sample as input, export for export high-definition picture, the generations model will export high-definition picture and The high-definition picture sample of the Training scene compares generation and generates model error value;
The discrimination model carries out the high-definition picture sample for exporting high-definition picture and the Training scene pair Than generating discrimination model error amount;
The generation model error value and the generation model error value weighting summation are generated into training pattern error amount;
If the training pattern error amount is less than preset threshold, the trained predetermined depth neural network model is set It is set to the default imaging model;
If the training pattern error amount is more than or equal to preset threshold, according to training pattern error amount update The low-resolution image training sample is simultaneously input to the updated predetermined depth by the parameter of predetermined depth neural network Neural network model is trained again.
Further, in one embodiment of the invention, pass through low-resolution image described in collection optical system and institute State low-resolution image training sample.
In order to achieve the above objectives, another aspect of the present invention embodiment proposes a kind of list based on structure light and deep learning Expose super-resolution imaging device, comprising:
Constructing module generates composite construction optical illumination coding for being overlapped to multiple and different traditional lightings coding;
First acquisition module encodes under single exposure for acquiring target scene to be processed in the composite construction optical illumination The low-resolution image of generation;
Output module is handled for the low-resolution image to be inputted default imaging model, and output is described wait locate Manage the corresponding super-resolution image of high target scene.
Single exposure super-resolution imaging device based on structure light and deep learning of the embodiment of the present invention, according to structure light The Fourier space spectral characteristic of coded illumination image, analysis Structured Illumination in Fourier space frequency domain are encoded to image It influences, to establish composite illuminating coding;Based on deep learning theory, building is suitable for what structure light coding illumination image restored Deep neural network model;It is exposed by single composite construction pumped FIR laser, carries out super resolution image in conjunction with deep neural network It quickly rebuilds, this method has the support of deep learning theory and structure light coding illumination theory, and implementation is simple, it is only necessary to single It is secondary to expose the quick reconstruction that super resolution image can be realized.
In addition, single exposure super-resolution imaging dress according to the above embodiment of the present invention based on structure light and deep learning Following additional technical characteristic can also be had by setting:
Further, in one embodiment of the invention, further includes: the second acquisition module, training module, generation module And iteration module;
Second acquisition module encodes single exposure in the composite construction optical illumination for acquiring Training scene sample The low-resolution image training sample of lower generation;
The training module, for by the low-resolution image training sample in predetermined depth neural network model into Row training;
The generation module, if for the error between training result and the high-definition picture sample of the Training scene Value is less than preset threshold, then sets the default imaging model for the trained predetermined depth neural network model;
The iteration module, if between the training result and the high-definition picture sample of the Training scene Error amount is more than or equal to preset threshold, then the parameter of the predetermined depth neural network is updated according to error amount and will be described low Image in different resolution training sample is input to updated predetermined depth neural network model and is trained again.
Further, in one embodiment of the invention, the predetermined depth neural network model is convolutional Neural net The low-resolution image training sample of acquisition is inputted the predetermined depth neural network model and is trained by network model;
It is output high-definition picture by the output after the predetermined depth neural network model, high-resolution will be exported The high-definition picture sample of image and the Training scene compares to obtain error amount;
If error amount is less than preset threshold, set described default for the trained predetermined depth neural network model Imaging model;
If error amount is more than or equal to preset threshold, the parameter of the predetermined depth neural network is updated according to error amount And the low-resolution image training sample is input to the updated predetermined depth neural network model and is trained again.
Further, in one embodiment of the invention, the predetermined depth neural network model is the generation pair Anti- network model;
Generation confrontation network model includes generating model and discrimination model, and the generation model is by the low resolution Image training sample as input, export for export high-definition picture, the generations model will export high-definition picture and The high-definition picture sample of the Training scene compares generation and generates model error value;
The discrimination model carries out the high-definition picture sample for exporting high-definition picture and the Training scene pair Than generating discrimination model error amount;
The generation model error value and the generation model error value weighting summation are generated into training pattern error amount;
If the training pattern error amount is less than preset threshold, the trained predetermined depth neural network model is set It is set to the default imaging model;
If the training pattern error amount is more than or equal to preset threshold, according to training pattern error amount update The low-resolution image training sample is simultaneously input to the updated predetermined depth by the parameter of predetermined depth neural network Neural network model is trained again.
The additional aspect of the present invention and advantage will be set forth in part in the description, and will partially become from the following description Obviously, or practice through the invention is recognized.
Detailed description of the invention
Above-mentioned and/or additional aspect and advantage of the invention will become from the following description of the accompanying drawings of embodiments Obviously and it is readily appreciated that, in which:
Fig. 1 is single exposure super-resolution imaging side based on structure light and deep learning according to one embodiment of the invention Method flow chart;
Fig. 2 is the generation composite construction optical illumination coding schematic diagram according to one embodiment of the invention;
Fig. 3 is the low resolution coded image schematic diagram according to one embodiment of the invention;
Fig. 4 is the convolutional neural networks model training schematic diagram according to one embodiment of the invention;
Fig. 5 is that network model training schematic diagram is fought according to the generation of one embodiment of the invention;
Fig. 6 is building and the Irnaging procedures figure of the neural network according to one embodiment of the invention;
Fig. 7 is the Structured Illumination imaging optical path figure according to one embodiment of the invention;
Fig. 8 is single exposure super-resolution imaging dress based on structure light and deep learning according to one embodiment of the invention Set structural schematic diagram.
Specific embodiment
The embodiment of the present invention is described below in detail, examples of the embodiments are shown in the accompanying drawings, wherein from beginning to end Same or similar label indicates same or similar element or element with the same or similar functions.Below with reference to attached The embodiment of figure description is exemplary, it is intended to is used to explain the present invention, and is not considered as limiting the invention.
The single exposure based on structure light and deep learning for describing to propose according to embodiments of the present invention with reference to the accompanying drawings is super Resolution imaging method and device.
The single exposure based on structure light and deep learning proposed according to embodiments of the present invention is described with reference to the accompanying drawings first Super-resolution imaging method.
Fig. 1 is single exposure super-resolution imaging side based on structure light and deep learning according to one embodiment of the invention Method flow chart.
As shown in Figure 1, should single exposure super-resolution imaging method based on structure light and deep learning the following steps are included:
Step S1 is overlapped multiple and different traditional lightings coding and generates composite construction optical illumination coding.
Specifically, the spectral characteristic of the influence according to Structured Illumination to imaging process and Structured Illumination coding, to more A different traditional lighting coding, which is overlapped, generates composite construction optical illumination coding.
It should be noted that the composite construction optical illumination coding that step S1 is generated generates default imaging model with following training The composite construction optical illumination coding of Shi Caiyong is identical.
Further, multiple and different traditional lightings coding includes but is not limited to: SIN function coding and/or non-linear glimmering Pumped FIR laser and/or random coded.
It is superimposed to obtain composite construction optical illumination coding, composite construction optical illumination by the airspace for encoding multiple traditional lightings The Fourier space frequency spectrum of coding is the superposition that multiple and different traditional lightings encode frequency spectrum.
It is understood that multiple and different traditional lightings is encoded superposition, so that containing in the image of acquisition The Fourier space spectrum information lost under traditional lighting.
As shown in Fig. 2, Structured Illumination imaging after generating coded illumination structure light irradiation sample by being imaged, it is traditional Structure light coding mode includes but is not limited to SIN function coding, Nonlinear fluorescence coding and random coded, following sinusoidal coding For, the airspace expression formula of the sinusoidal illumination coding of two dimension are as follows:
Wherein,For wave vector, φ1For phase.
Transformed to Fourier space frequency domain:
If composite illuminating is encoded to the superposition of three kinds of conventional sinusoidals coding:
The Fourier space frequency spectrum of composite illuminating coding are as follows:
If the Fourier space frequency spectrum of sample patterns isThe Fourier space being then imaged through composite construction optical illumination Frequency spectrum are as follows:
Due toHigh-frequency information comprising original image sample except system OTF range, so can basisIt is extensive Multiple super-resolution image.
As shown in figure 3, obtaining low resolution using composite structure light coded illumination Training scene through collection optical system and compiling Code image.The training dataset of neural network can be generated using the full resolution pricture sample of a collection of Training scene according to this step.
Step S2 acquires target scene to be processed and encodes the low resolution generated under single exposure in composite construction optical illumination Image.
The composite coding structure light single exposure that processing target scene utilizes step S1 to generate is treated, is adopted by optical system Collect corresponding low-resolution image.
Further, before step S2 further include:
Acquisition Training scene encodes the low-resolution image training sample generated under single exposure in composite construction optical illumination This;
Low-resolution image training sample is trained in predetermined depth neural network model;
It, will instruction if the error amount between training result and the high-definition picture sample of Training scene is less than preset threshold Experienced predetermined depth neural network model is set as default imaging model;
If the error amount between training result and the high-definition picture sample of Training scene is more than or equal to preset threshold, The parameter of predetermined depth neural network is updated according to error amount and is input to low-resolution image training sample updated Predetermined depth neural network model is trained again.
Specifically, wherein target scene and Training scene can be true object or sample image.To target scene Or Training scene is directly acquired to obtain the high-definition picture sample of target scene or Training scene.
It is understood that after obtaining the error amount between training result and the high-definition picture sample of Training scene, By the error back propagation using network output and high-definition picture, iteration updates model parameter.It is set when error reaches Its convergence is thought when threshold value.
Predetermined depth neural network model is including but not limited to convolutional neural networks, generation confrontation network and based on the above mind Mutation or progressions model through network, predetermined depth neural network single exposure under composite construction optical illumination coding acquire low Resolution image sample is compared as network inputs by the high-definition picture sample with Training scene, and iteration updates net Network model parameter to network output and the error of high-definition picture reaches given threshold.
Convolutional neural networks each convolution kernel in network hidden layer carries out convolution algorithm to preceding layer, and convolution kernel output is made It inputs for next stage to carry out corresponding calculation processing.
Further, predetermined depth neural network model is convolutional neural networks model, by the low-resolution image of acquisition Training sample input predetermined depth neural network model is trained;
It is output high-definition picture by the output after predetermined depth neural network model, high-definition picture will be exported It compares to obtain error amount with the high-definition picture sample of Training scene;
If error amount is less than preset threshold, default imaging mould is set by trained predetermined depth neural network model Type;
If more than or equal to preset threshold, the parameter of predetermined depth neural network is updated according to error amount and will for error amount Low-resolution image training sample is input to updated predetermined depth neural network model and is trained again.
Further, predetermined depth neural network model makes a living into confrontation network model;
Generating confrontation network model includes generating model and discrimination model, generates model for low-resolution image training sample It as input, exports to export high-definition picture, generates the high-resolution that model will export high-definition picture and Training scene Rate image pattern compares generation and generates model error value;
The high-definition picture sample for exporting high-definition picture and Training scene is compared generation and sentenced by discrimination model Other model error value;
Model error value will be generated and generate model error value weighting summation and generate training pattern error amount;
If training pattern error amount is less than preset threshold, set default for trained predetermined depth neural network model Imaging model;
If training pattern error amount is more than or equal to preset threshold, predetermined depth nerve is updated according to training pattern error amount Low-resolution image training sample is simultaneously input to updated predetermined depth neural network model and instructed again by the parameter of network Practice.
It is understood that the error amount of discrimination model can be only considered as a kind of implementation, for example, differentiating mould Type is compared the high-definition picture sample of output high-definition picture and Training scene, judges the similarity of the two, In When exporting high-definition picture and the similar or the same high-definition picture sample of Training scene, then by trained predetermined depth mind Default imaging model is set as through network model.
As the mode of alternatively possible realization, in order to improve model imaging resolution ratio, comprehensively consider generation confrontation Model and the respective error amount of discrimination model are generated in network model, and low-resolution image training sample is inputted and generates model In, output high-definition picture is obtained, the high-definition picture sample for exporting high-definition picture and Training scene is carried out pair It is by discrimination model that the high resolution graphics for exporting high-definition picture and Training scene is decent than obtaining generating model error value Originally it compares and generates discrimination model error amount, comprehensively considered according to weight and generate model error value and discrimination model error amount, To generate model error value and discrimination model error amount respectively with carry out being added summation again after corresponding multiplied by weight, obtain One training pattern error amount compares this training pattern error amount with preset threshold.
Specifically, instruction low-resolution image training sample being trained in predetermined depth neural network model Practicing result is output high-definition picture.Network in predetermined depth neural network model is different, and specific training process is not Together.Eventually by the update iteration to predetermined depth neural network model, predetermined depth neural network model is made not meet convergence Condition sets it to default imaging model.
Low-resolution image is inputted default imaging model and handled, it is corresponding to export target scene to be processed by step S3 Super-resolution image.
After obtaining default imaging model by training predetermined depth neural network, low-resolution image is inputted default Imaging model is handled, and the corresponding super-resolution image of target scene to be processed is exported.
As shown in Figure 4 and Figure 5, the deep neural network that can carry out super resolution image reconstruction includes but is not limited to convolutional Neural Network and generation confrontation network.The core of convolutional neural networks is convolutional calculation, and convolution kernel is generally two-dimensional matrix, each convolution It checks preceding layer output characteristic pattern sliding to be multiplied, multiplied result is added, and convolution results can be through nonlinear activation unit (including but not It is limited to line rectification unit, hyperbolic tangent function, Sigmoid function etc.) operation input next stage hidden layer or output layer.It generates Fighting network includes to generate model and discrimination model, generates model for low resolution coded image as input and generates High-Resolution Map Picture, discrimination model export true according to the difference judgement generation model generated between model output and true high-definition picture Puppet, generates model and discrimination model alternating iteration updates, until discrimination model can not identify true high-definition picture and generate Model exports image.
As shown in fig. 6, generating the training dataset of deep neural network according to composite construction pumped FIR laser, convolutional Neural is built Network generates confrontation network even depth neural network model, using above-mentioned training dataset training neural network parameter, obtains Network model.After forming network, the low resolution coded image of composite structure light single exposure can be inputted depth nerve net Network, network output are high-definition picture.
The method of the embodiment of the present invention is using deep learning model as realization rate, while according to quilt under Structured Illumination The Fourier space spectral characteristic for acquiring image establishes composite construction pumped FIR laser, realizes Structured Illumination super-resolution image It quickly rebuilds, and builds system and carry out Proof-Of Principle.As shown in fig. 7, composite coding light structures light is produced by spatial light modulator It is raw, it is incident upon through spectroscope and is observed sample plane, sample reflection light is received by sensor and acquired.
The single exposure super-resolution imaging method based on structure light and deep learning proposed according to embodiments of the present invention, root According to the Fourier space spectral characteristic of structure light coding illumination image, analysis Structured Illumination in Fourier space frequency domain is encoded Influence to image, to establish composite illuminating coding;Based on deep learning theory, building is suitable for structure light coding illumination figure As the deep neural network model restored;It is exposed by single composite construction pumped FIR laser, carries out oversubscription in conjunction with deep neural network Distinguish the quick reconstruction of image, this method has the support of deep learning theory and structure light coding illumination theory, implementation letter It is single, it is only necessary to which that the quick reconstruction of super resolution image can be realized in single exposure.
The single exposure based on structure light and deep learning proposed according to embodiments of the present invention referring next to attached drawing description is super Resolution imaging devices.
Fig. 8 is single exposure super-resolution imaging dress based on structure light and deep learning according to one embodiment of the invention Set structural schematic diagram.
As shown in figure 8, should include: constructing module based on single exposure super-resolution imaging device of structure light and deep learning 100, the first acquisition module 200 and output module 300.
Constructing module 100 generates composite construction optical illumination coding for being overlapped to multiple and different traditional lightings coding.
First acquisition module 200 encodes under single exposure for acquiring target scene to be processed in composite construction optical illumination The low-resolution image of generation.
Output module 300 handles for low-resolution image to be inputted default imaging model, exports target to be processed The corresponding super-resolution image of scene.
Further, in one embodiment of the invention, further includes: the second acquisition module, training module, generation module And iteration module;
Second acquisition module, low point generated for acquiring Training scene in the case where composite construction optical illumination encodes single exposure Resolution image training sample;
Training module, for low-resolution image training sample to be trained in predetermined depth neural network model;
Generation module, if being less than for the error amount between training result and the high-definition picture sample of Training scene pre- If threshold value, then default imaging model is set by trained predetermined depth neural network model;
Iteration module, if be greater than for the error amount between training result and the high-definition picture sample of Training scene etc. In preset threshold, then the parameter of predetermined depth neural network and low-resolution image training sample is defeated is updated according to error amount Enter to updated predetermined depth neural network model and is trained again.
Further, in one embodiment of the invention, predetermined depth neural network model is convolutional neural networks mould The low-resolution image training sample input predetermined depth neural network model of acquisition is trained by type;
It is output high-definition picture by the output after predetermined depth neural network model, high-definition picture will be exported It compares to obtain error amount with the high-definition picture sample of Training scene;
If error amount is less than preset threshold, default imaging mould is set by trained predetermined depth neural network model Type;
If more than or equal to preset threshold, the parameter of predetermined depth neural network is updated according to error amount and will for error amount Low-resolution image training sample is input to updated predetermined depth neural network model and is trained again.
Further, in one embodiment of the invention, predetermined depth neural network model makes a living into confrontation network mould Type;
Generating confrontation network model includes generating model and discrimination model, generates model for low-resolution image training sample It as input, exports to export high-definition picture, generates the high-resolution that model will export high-definition picture and Training scene Rate image pattern compares generation and generates model error value;
The high-definition picture sample for exporting high-definition picture and Training scene is compared generation and sentenced by discrimination model Other model error value;
Model error value will be generated and generate model error value weighting summation and generate training pattern error amount;
If training pattern error amount is less than preset threshold, set default for trained predetermined depth neural network model Imaging model;
If training pattern error amount is more than or equal to preset threshold, predetermined depth nerve is updated according to training pattern error amount Low-resolution image training sample is simultaneously input to updated predetermined depth neural network model and instructed again by the parameter of network Practice.
It should be noted that aforementioned to single exposure super-resolution imaging embodiment of the method based on structure light and deep learning Explanation be also applied for the device of the embodiment, details are not described herein again.
The single exposure super-resolution imaging based on structure light and deep learning proposed according to embodiments of the present invention is up to root According to the Fourier space spectral characteristic of structure light coding illumination image, analysis Structured Illumination in Fourier space frequency domain is encoded Influence to image, to establish composite illuminating coding;Based on deep learning theory, building is suitable for structure light coding illumination figure As the deep neural network model restored;It is exposed by single composite construction pumped FIR laser, carries out oversubscription in conjunction with deep neural network Distinguish the quick reconstruction of image, this method has the support of deep learning theory and structure light coding illumination theory, implementation letter It is single, it is only necessary to which that the quick reconstruction of super resolution image can be realized in single exposure.
In addition, term " first ", " second " are used for descriptive purposes only and cannot be understood as indicating or suggesting relative importance Or implicitly indicate the quantity of indicated technical characteristic.Define " first " as a result, the feature of " second " can be expressed or Implicitly include at least one this feature.In the description of the present invention, the meaning of " plurality " is at least two, such as two, three It is a etc., unless otherwise specifically defined.
In the description of this specification, reference term " one embodiment ", " some embodiments ", " example ", " specifically show The description of example " or " some examples " etc. means specific features, structure, material or spy described in conjunction with this embodiment or example Point is included at least one embodiment or example of the invention.In the present specification, schematic expression of the above terms are not It must be directed to identical embodiment or example.Moreover, particular features, structures, materials, or characteristics described can be in office It can be combined in any suitable manner in one or more embodiment or examples.In addition, without conflicting with each other, the skill of this field Art personnel can tie the feature of different embodiments or examples described in this specification and different embodiments or examples It closes and combines.
Although the embodiments of the present invention has been shown and described above, it is to be understood that above-described embodiment is example Property, it is not considered as limiting the invention, those skilled in the art within the scope of the invention can be to above-mentioned Embodiment is changed, modifies, replacement and variant.

Claims (10)

1. a kind of single exposure super-resolution imaging method based on structure light and deep learning, which is characterized in that including following step It is rapid:
S1 is overlapped multiple and different traditional lightings coding and generates composite construction optical illumination coding;
S2 acquires target scene to be processed and encodes the low resolution figure generated under single exposure in the composite construction optical illumination Picture;
The low-resolution image is inputted default imaging model and handled, it is corresponding to export the target scene to be processed by S3 Super-resolution image.
2. the method according to claim 1, wherein before step S2, further includes:
Acquisition Training scene encodes the low-resolution image training sample generated under single exposure in the composite construction optical illumination;
The low-resolution image training sample is trained in predetermined depth neural network model;
It, will instruction if the error amount between training result and the high-definition picture sample of the Training scene is less than preset threshold The experienced predetermined depth neural network model is set as the default imaging model;
If the error amount between the training result and the high-definition picture sample of the Training scene is more than or equal to default threshold Value then updates the parameter of the predetermined depth neural network and low-resolution image training sample is defeated according to error amount Enter to updated predetermined depth neural network model and is trained again.
3. the method according to claim 1, wherein
The multiple difference traditional lighting coding includes: SIN function coding and/or Nonlinear fluorescence coding, and/or is compiled at random Code;
It is superimposed to obtain the composite construction optical illumination coding, the composite junction by the airspace for encoding the multiple traditional lighting The Fourier space frequency spectrum of structure optical illumination coding is the superposition that the multiple different traditional lightings encode frequency spectrum.
4. according to the method described in claim 2, it is characterized in that,
The predetermined depth neural network model is convolutional neural networks model, by the low-resolution image training sample of acquisition This input predetermined depth neural network model is trained;
It is output high-definition picture by the output after the predetermined depth neural network model, high-definition picture will be exported It compares to obtain error amount with the high-definition picture sample of the Training scene;
If error amount is less than preset threshold, the default imaging is set by the trained predetermined depth neural network model Model;
If more than or equal to preset threshold, the parameter of the predetermined depth neural network is updated according to error amount and will for error amount The low-resolution image training sample is input to the updated predetermined depth neural network model and is trained again.
5. according to the method described in claim 2, it is characterized in that,
The predetermined depth neural network model is that the generation fights network model;
Generation confrontation network model includes generating model and discrimination model, and the generation model is by the low-resolution image Training sample as input, export for export high-definition picture, the generations model will output high-definition picture with it is described The high-definition picture sample of Training scene compares generation and generates model error value;
The high-definition picture sample for exporting high-definition picture and the Training scene is compared life by the discrimination model At discrimination model error amount;
The generation model error value and the generation model error value weighting summation are generated into training pattern error amount;
If the training pattern error amount is less than preset threshold, set the trained predetermined depth neural network model to The default imaging model;
If the training pattern error amount is more than or equal to preset threshold, updated according to the training pattern error amount described default The low-resolution image training sample is simultaneously input to the updated predetermined depth nerve by the parameter of deep neural network Network model is trained again.
6. the method according to claim 1, wherein
Pass through low-resolution image described in collection optical system and the low-resolution image training sample.
7. a kind of single exposure super-resolution imaging device based on structure light and deep learning characterized by comprising
Constructing module generates composite construction optical illumination coding for being overlapped to multiple and different traditional lightings coding;
First acquisition module is generated for acquiring target scene to be processed in the case where the composite construction optical illumination encodes single exposure Low-resolution image;
Output module handles for the low-resolution image to be inputted default imaging model, exports the mesh to be processed Mark the corresponding super-resolution image of scene.
8. device according to claim 7, which is characterized in that further include: the second acquisition module, generates mould at training module Block and iteration module;
Second acquisition module is generated for acquiring Training scene in the case where the composite construction optical illumination encodes single exposure Low-resolution image training sample;
The training module, for instructing the low-resolution image training sample in predetermined depth neural network model Practice;
The generation module, if the error amount between training result and the high-definition picture sample of the Training scene is small In preset threshold, then the default imaging model is set by the trained predetermined depth neural network model;
The iteration module, if for the error between the training result and the high-definition picture sample of the Training scene Value is more than or equal to preset threshold, then the parameter of the predetermined depth neural network is updated according to error amount and by the low resolution Rate image training sample is input to updated predetermined depth neural network model and is trained again.
9. device according to claim 8, which is characterized in that
The predetermined depth neural network model is convolutional neural networks model, by the low-resolution image training sample of acquisition This input predetermined depth neural network model is trained;
It is output high-definition picture by the output after the predetermined depth neural network model, high-definition picture will be exported It compares to obtain error amount with the high-definition picture sample of the Training scene;
If error amount is less than preset threshold, the default imaging is set by the trained predetermined depth neural network model Model;
If more than or equal to preset threshold, the parameter of the predetermined depth neural network is updated according to error amount and will for error amount The low-resolution image training sample is input to the updated predetermined depth neural network model and is trained again.
10. device according to claim 8, which is characterized in that
The predetermined depth neural network model is that the generation fights network model;
Generation confrontation network model includes generating model and discrimination model, and the generation model is by the low-resolution image Training sample as input, export for export high-definition picture, the generations model will output high-definition picture with it is described The high-definition picture sample of Training scene compares generation and generates model error value;
The high-definition picture sample for exporting high-definition picture and the Training scene is compared life by the discrimination model At discrimination model error amount;
The generation model error value and the generation model error value weighting summation are generated into training pattern error amount;
If the training pattern error amount is less than preset threshold, set the trained predetermined depth neural network model to The default imaging model;
If the training pattern error amount is more than or equal to preset threshold, updated according to the training pattern error amount described default The low-resolution image training sample is simultaneously input to the updated predetermined depth nerve by the parameter of deep neural network Network model is trained again.
CN201910697301.7A 2019-07-30 2019-07-30 Single-exposure super-resolution imaging method and device based on structured light and deep learning Active CN110415201B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201910697301.7A CN110415201B (en) 2019-07-30 2019-07-30 Single-exposure super-resolution imaging method and device based on structured light and deep learning

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201910697301.7A CN110415201B (en) 2019-07-30 2019-07-30 Single-exposure super-resolution imaging method and device based on structured light and deep learning

Publications (2)

Publication Number Publication Date
CN110415201A true CN110415201A (en) 2019-11-05
CN110415201B CN110415201B (en) 2022-03-22

Family

ID=68364263

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201910697301.7A Active CN110415201B (en) 2019-07-30 2019-07-30 Single-exposure super-resolution imaging method and device based on structured light and deep learning

Country Status (1)

Country Link
CN (1) CN110415201B (en)

Cited By (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN110796603A (en) * 2019-11-25 2020-02-14 曲阜师范大学 High-resolution photoacoustic imaging method for deep tumor neovascularization
CN111292240A (en) * 2020-01-23 2020-06-16 上海交通大学 Magnetic resonance super-resolution imaging method based on imaging model and machine learning
CN112233034A (en) * 2020-10-20 2021-01-15 常州机电职业技术学院 Deep learning phase-shift interference fringe pattern blind denoising method for interferometry
CN113947565A (en) * 2021-09-03 2022-01-18 中国科学院西安光学精密机械研究所 Structured light illumination super-resolution imaging gene detection method based on deep learning
CN115144075A (en) * 2022-06-30 2022-10-04 北京理工大学 High-speed spectral imaging method and device
CN116757926A (en) * 2023-05-22 2023-09-15 华南师范大学 Super-resolution SIM-FRET imaging method and system based on self-supervision learning image denoising

Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN104197861A (en) * 2014-08-25 2014-12-10 深圳大学 Three-dimensional digital imaging method based on structured light gray level vector
US20160182889A1 (en) * 2014-12-19 2016-06-23 Datalogic ADC, Inc. Depth camera system using coded structured light
CN107967697A (en) * 2017-11-23 2018-04-27 四川大学 Method for three-dimensional measurement and system based on colored random binary coding structured illumination
CN108734660A (en) * 2018-05-25 2018-11-02 上海通途半导体科技有限公司 A kind of image super-resolution rebuilding method and device based on deep learning
US20180347967A1 (en) * 2017-06-01 2018-12-06 RGBDsense Information Technology Ltd. Method and apparatus for generating a random coding pattern for coding structured light
CN109685716A (en) * 2018-12-14 2019-04-26 大连海事大学 A kind of image super-resolution rebuilding method of the generation confrontation network based on Gauss encoder feedback

Patent Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN104197861A (en) * 2014-08-25 2014-12-10 深圳大学 Three-dimensional digital imaging method based on structured light gray level vector
US20160182889A1 (en) * 2014-12-19 2016-06-23 Datalogic ADC, Inc. Depth camera system using coded structured light
US20180347967A1 (en) * 2017-06-01 2018-12-06 RGBDsense Information Technology Ltd. Method and apparatus for generating a random coding pattern for coding structured light
CN107967697A (en) * 2017-11-23 2018-04-27 四川大学 Method for three-dimensional measurement and system based on colored random binary coding structured illumination
CN108734660A (en) * 2018-05-25 2018-11-02 上海通途半导体科技有限公司 A kind of image super-resolution rebuilding method and device based on deep learning
CN109685716A (en) * 2018-12-14 2019-04-26 大连海事大学 A kind of image super-resolution rebuilding method of the generation confrontation network based on Gauss encoder feedback

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
刘阳阳: "基于深度神经网络的RGB-D图像识别算法研究", 《中国优秀硕士学位论文全文数据库信息科技辑》 *

Cited By (11)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN110796603A (en) * 2019-11-25 2020-02-14 曲阜师范大学 High-resolution photoacoustic imaging method for deep tumor neovascularization
CN111292240A (en) * 2020-01-23 2020-06-16 上海交通大学 Magnetic resonance super-resolution imaging method based on imaging model and machine learning
CN111292240B (en) * 2020-01-23 2022-01-07 上海交通大学 Magnetic resonance super-resolution imaging method based on imaging model and machine learning
CN112233034A (en) * 2020-10-20 2021-01-15 常州机电职业技术学院 Deep learning phase-shift interference fringe pattern blind denoising method for interferometry
CN112233034B (en) * 2020-10-20 2022-05-13 常州机电职业技术学院 Deep learning phase-shift interference fringe pattern blind denoising method for interferometry
CN113947565A (en) * 2021-09-03 2022-01-18 中国科学院西安光学精密机械研究所 Structured light illumination super-resolution imaging gene detection method based on deep learning
CN113947565B (en) * 2021-09-03 2023-04-18 中国科学院西安光学精密机械研究所 Structured light illumination super-resolution imaging gene detection method based on deep learning
CN115144075A (en) * 2022-06-30 2022-10-04 北京理工大学 High-speed spectral imaging method and device
CN115144075B (en) * 2022-06-30 2023-08-29 北京理工大学 High-speed spectrum imaging method and device
CN116757926A (en) * 2023-05-22 2023-09-15 华南师范大学 Super-resolution SIM-FRET imaging method and system based on self-supervision learning image denoising
CN116757926B (en) * 2023-05-22 2024-04-05 华南师范大学 Super-resolution SIM-FRET imaging method and system based on self-supervision learning image denoising

Also Published As

Publication number Publication date
CN110415201B (en) 2022-03-22

Similar Documents

Publication Publication Date Title
CN110415201A (en) Single exposure super-resolution imaging method and device based on structure light and deep learning
CN107154023A (en) Face super-resolution reconstruction method based on generation confrontation network and sub-pix convolution
CN106897986B (en) A kind of visible images based on multiscale analysis and far infrared image interfusion method
EP3832359A1 (en) Method and device for imaging of lensless hyperspectral image
CN104835130A (en) Multi-exposure image fusion method
CN102063712B (en) Multi-exposure image fusion method based on sub-band structure
CN103247036A (en) Multiple-exposure image fusion method and device
Bhandari et al. A new beta differential evolution algorithm for edge preserved colored satellite image enhancement
Monnier et al. The 2014 interferometric imaging beauty contest
Ignatov et al. Learned smartphone ISP on mobile GPUs with deep learning, mobile AI & AIM 2022 challenge: report
CN114638767A (en) Laparoscope image smoke removal method based on generation of countermeasure network
CN116563693A (en) Underwater image color restoration method based on lightweight attention mechanism
CN112163998A (en) Single-image super-resolution analysis method matched with natural degradation conditions
Chen et al. GCAM: lightweight image inpainting via group convolution and attention mechanism
Guo et al. Haze removal for single image: A comprehensive review
Li et al. Effective data-driven technology for efficient vision-based outdoor industrial systems
CN105491371A (en) Tone mapping image quality evaluation method based on gradient magnitude similarity
Hassan et al. Image quality measurement-based comparative analysis of illumination compensation methods for face image normalization
Anand et al. Architectural distortion detection in mammogram using contourlet transform and texture features
CN116433518A (en) Fire image smoke removing method based on improved Cycle-Dehaze neural network
CN115311430A (en) Training method and system of human body reconstruction model and computer equipment
CN110352387A (en) For the system and method by more depth sparse phase restoration and reconstruction holographic lensless images
CN114266713A (en) NonshadowGAN-based unmanned aerial vehicle railway fastener image shadow removing method and system
CN114648604A (en) Image rendering method, electronic device, storage medium and program product
Xue et al. Saraniqa: Self-attention restorative adversarial network for no-reference image quality assessment

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant