CN114240797A - OCT image denoising method, device, equipment and medium - Google Patents
OCT image denoising method, device, equipment and medium Download PDFInfo
- Publication number
- CN114240797A CN114240797A CN202111579290.6A CN202111579290A CN114240797A CN 114240797 A CN114240797 A CN 114240797A CN 202111579290 A CN202111579290 A CN 202111579290A CN 114240797 A CN114240797 A CN 114240797A
- Authority
- CN
- China
- Prior art keywords
- image
- feature
- oct image
- attention
- layer
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Granted
Links
- 238000000034 method Methods 0.000 title claims abstract description 57
- 230000004927 fusion Effects 0.000 claims abstract description 54
- 238000012549 training Methods 0.000 claims abstract description 24
- 238000012545 processing Methods 0.000 claims abstract description 22
- 230000007246 mechanism Effects 0.000 claims abstract description 16
- 238000012360 testing method Methods 0.000 claims abstract description 12
- 238000012795 verification Methods 0.000 claims abstract description 11
- 238000004590 computer program Methods 0.000 claims description 12
- 230000006870 function Effects 0.000 claims description 10
- 238000007781 pre-processing Methods 0.000 claims description 8
- 238000001514 detection method Methods 0.000 claims description 6
- 238000001914 filtration Methods 0.000 claims description 6
- 238000010276 construction Methods 0.000 claims description 5
- 238000000605 extraction Methods 0.000 claims description 5
- 230000007704 transition Effects 0.000 claims description 5
- 208000037170 Delayed Emergence from Anesthesia Diseases 0.000 claims description 4
- 150000001875 compounds Chemical class 0.000 claims description 3
- 238000013507 mapping Methods 0.000 claims description 3
- 230000009191 jumping Effects 0.000 claims description 2
- 230000000750 progressive effect Effects 0.000 abstract description 6
- 238000004422 calculation algorithm Methods 0.000 description 20
- 230000000694 effects Effects 0.000 description 14
- 238000010586 diagram Methods 0.000 description 10
- 238000002474 experimental method Methods 0.000 description 7
- 230000008569 process Effects 0.000 description 6
- 238000013135 deep learning Methods 0.000 description 4
- 230000009467 reduction Effects 0.000 description 4
- 230000002146 bilateral effect Effects 0.000 description 3
- 238000000354 decomposition reaction Methods 0.000 description 3
- 238000003384 imaging method Methods 0.000 description 3
- 230000001965 increasing effect Effects 0.000 description 3
- 210000001525 retina Anatomy 0.000 description 3
- 238000012937 correction Methods 0.000 description 2
- 230000002708 enhancing effect Effects 0.000 description 2
- 238000011156 evaluation Methods 0.000 description 2
- 230000006872 improvement Effects 0.000 description 2
- 230000001360 synchronised effect Effects 0.000 description 2
- ORILYTVJVMAKLC-UHFFFAOYSA-N Adamantane Natural products C1C(C2)CC3CC1CC2C3 ORILYTVJVMAKLC-UHFFFAOYSA-N 0.000 description 1
- 230000004913 activation Effects 0.000 description 1
- 230000003044 adaptive effect Effects 0.000 description 1
- 239000000654 additive Substances 0.000 description 1
- 230000000996 additive effect Effects 0.000 description 1
- 230000009286 beneficial effect Effects 0.000 description 1
- 238000004364 calculation method Methods 0.000 description 1
- 230000001413 cellular effect Effects 0.000 description 1
- 238000007796 conventional method Methods 0.000 description 1
- 230000007547 defect Effects 0.000 description 1
- 230000007812 deficiency Effects 0.000 description 1
- 238000013461 design Methods 0.000 description 1
- 206010012601 diabetes mellitus Diseases 0.000 description 1
- 230000008034 disappearance Effects 0.000 description 1
- 239000000945 filler Substances 0.000 description 1
- 230000003116 impacting effect Effects 0.000 description 1
- 239000004973 liquid crystal related substance Substances 0.000 description 1
- 238000013178 mathematical model Methods 0.000 description 1
- 238000012986 modification Methods 0.000 description 1
- 230000004048 modification Effects 0.000 description 1
- 230000002085 persistent effect Effects 0.000 description 1
- 238000011160 research Methods 0.000 description 1
- 230000004044 response Effects 0.000 description 1
- 230000011218 segmentation Effects 0.000 description 1
- 238000004088 simulation Methods 0.000 description 1
- 230000003595 spectral effect Effects 0.000 description 1
- 230000003068 static effect Effects 0.000 description 1
- 238000003325 tomography Methods 0.000 description 1
- 238000012546 transfer Methods 0.000 description 1
- 238000004800 variational method Methods 0.000 description 1
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T5/00—Image enhancement or restoration
- G06T5/70—Denoising; Smoothing
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/04—Architecture, e.g. interconnection topology
- G06N3/045—Combinations of networks
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/08—Learning methods
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/10—Image acquisition modality
- G06T2207/10072—Tomographic images
- G06T2207/10101—Optical tomography; Optical coherence tomography [OCT]
-
- Y—GENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
- Y02—TECHNOLOGIES OR APPLICATIONS FOR MITIGATION OR ADAPTATION AGAINST CLIMATE CHANGE
- Y02D—CLIMATE CHANGE MITIGATION TECHNOLOGIES IN INFORMATION AND COMMUNICATION TECHNOLOGIES [ICT], I.E. INFORMATION AND COMMUNICATION TECHNOLOGIES AIMING AT THE REDUCTION OF THEIR OWN ENERGY USE
- Y02D10/00—Energy efficient computing, e.g. low power processors, power management or thermal management
Landscapes
- Engineering & Computer Science (AREA)
- Physics & Mathematics (AREA)
- Theoretical Computer Science (AREA)
- General Physics & Mathematics (AREA)
- General Health & Medical Sciences (AREA)
- Molecular Biology (AREA)
- Biophysics (AREA)
- Computational Linguistics (AREA)
- Data Mining & Analysis (AREA)
- Evolutionary Computation (AREA)
- Artificial Intelligence (AREA)
- Biomedical Technology (AREA)
- Computing Systems (AREA)
- General Engineering & Computer Science (AREA)
- Life Sciences & Earth Sciences (AREA)
- Mathematical Physics (AREA)
- Software Systems (AREA)
- Health & Medical Sciences (AREA)
- Image Analysis (AREA)
- Image Processing (AREA)
Abstract
The invention discloses an OCT image denoising method, a device, equipment and a medium, wherein the method comprises the following steps: obtaining an OCT image to form a sample data set, wherein the sample data set is divided into a training set and a verification set; constructing a feature fusion attention intensive network for OCT image denoising, wherein the main structure of the feature fusion attention intensive network comprises a sparse block, a feature enhancement block, an attention mechanism and a reconstruction block, and the feature fusion attention intensive network is used for denoising the OCT image; inputting the sample data set into the feature fusion attention intensive network for training and testing until convergence; and inputting the image to be detected into a trained feature fusion attention-intensive network for processing to obtain a clear OCT image after denoising processing. The invention provides a network for progressive feature fusion and attention intensive based on the original deep convolutional network, and OCT image noise can be effectively removed based on the network.
Description
Technical Field
The invention relates to the technical field of image processing, in particular to an OCT image denoising method, device, equipment and medium.
Background
Coherence Tomography (OCT) has been widely used for ophthalmic diagnostics as an effective diagnostic tool. However, the OCT imaging process is inevitably disturbed by sample-based speckle and detector noise, which degrades image quality, thereby severely impacting subsequent processing and analysis of the OCT image. Speckle noise is multiplicative, as distinguished from gaussian noise, and is related to the microstructure of tissue. The noise that is common at present is mainly 4 types: speckle noise, gaussian noise, poisson noise, and impulse noise, wherein speckle noise is the most common and representative noise. The invention mainly aims at the denoising algorithm of speckle noise to study. The image denoising is an essential step in various image processing algorithms, is helpful for further distinguishing and explaining the image content, and has important significance in the field of image processing.
Image denoising is to remove noise from the disturbed image and restore the disturbed image to a clean image. It is a challenging task to design a suitable denoising method for the types of noise present in the image. Conventional image denoising methods are generally classified into 2 types: although the methods of time domain denoising and frequency domain denoising can better remove noise, the methods can cause the phenomenon that edges are excessively smoothed, thereby causing the loss of detail information. For the denoising method of the time domain, the bilateral filtering algorithm can keep a better filtering effect. The bilateral filter is a nonlinear filter, and can replace the gray value of each pixel point with the weighted average of the gray values of the adjacent pixel points. For the frequency domain denoising method, in 2014, researchers proposed a Variational Modal Decomposition (VMD) algorithm improved based on an Empirical Mode Decomposition (EMD), which mainly seeks an optimal value of each intrinsic modal component by a variational method to overcome the defect of the EMD. In addition, VMD is also extended to two-dimensional range, i.e., two-dimensional variational modal decomposition (2D-VMD) methods. The traditional method can obtain a certain denoising effect, but the effect is not ideal when processing speckle noise images, and image details are easy to lose. In response to this deficiency, a great number of researchers have proposed denoising using an image prior model. Buads et al propose a denoising method of Non-Local mean NLM (Non-Local Means). The method makes full use of redundant information in the image, and can greatly save the detailed characteristics of the image while denoising. The BM3D (Block Method of 3-Dimension) algorithm proposed by Dabov et al fuses ideas of a non-local denoising Method and a wavelet transform domain denoising Method, and achieves a good denoising effect. The PID (progressive Image denoising) algorithm proposed by Knaus et al takes Image denoising as a simple physical process, and can obtain a good denoising effect under the condition of no artifact by gradually reducing the noise. Gu et al propose a weighted nuclear norm minimization denoising method WNNM (weighted norm minimization method) by combining a non-local idea and a low-rank approximation theory. Although the denoising methods can achieve better denoising effect, the denoising performance of the methods is greatly reduced when the noise density is increased.
Some conventional methods involve filtering, such as bilateral filters and wavelet filters. Can be used to suppress speckle patterns, but at the same time is particularly prone to loss of image detail, especially if the image features are not of the same size. Non-local mean (NLM) and block matching 3D (BM3D) was originally designed to remove additive white gaussian noise and has been successfully applied to various imaging techniques. Still more advanced digital noise identification algorithms have been used to noise identify OCT images. However, these methods may cause blurring of the contours and loss of detail. Sparse representation [11] has become a modern denoising tool, but ignores the inherent features of spectral signals.
At present, some researches prove that the image recognition rate can be greatly improved by using a deep learning network to denoise the OCT image, including RED-Net and a persistent memory network (MemNet). In these works, a mathematical model of the noise is built and the noise is randomly multiplied in the OCT image to simulate a labeled noisy image, thereby avoiding the labor intensive work of efficiently generating a training data set. This method has limitations because the noise is simulated in the training data, and because the noise in the OCT image is complex, it is difficult to obtain a perfect simulation result. Thus, using true noisy data may be more effective in training the data set. In contrast to noise modeling, a noise map is obtained by subtracting a noise image from a sharp image. However, in an OCT scanning system, moving the sample will cause distortion of the OCT image. Therefore, an additional alignment work between B-scan images is very necessary, and the quality of scanning largely affects the denoising performance. Although FF-OCT can easily obtain multiple B-scans in good registration, it has the disadvantage that the imaging speed of the sensor frame rate can be limited.
Disclosure of Invention
In order to solve the technical problems, the invention provides an OCT image denoising method, device, equipment and medium, and provides a network (DnFFA) for progressive feature fusion and attention intensive based on the original deep convolutional network, and the network (DnFFA) is used for speckle noise removal. Firstly, local features of an image are extracted and a series of preprocessing is carried out, and the preprocessing strategy is to remove a background by enhancing contrast and eliminating noise. And then, performing feature fusion on the network by adopting a progressive method, extracting global network features of noise, then introducing an attention mechanism to further improve the network, and finally, fusing output feature maps of all dense blocks and inputting the fused output feature maps to a reconstruction output layer to obtain the denoised OCT image.
In order to achieve the purpose, the technical scheme of the invention is as follows:
an OCT image denoising method comprises the following steps:
obtaining an OCT image to form a sample data set, wherein the sample data set is divided into a training set and a verification set;
constructing a feature fusion attention intensive network for OCT image denoising, wherein the main structure of the feature fusion attention intensive network comprises a sparse block, a feature enhancement block, an attention mechanism and a reconstruction block, and the feature fusion attention intensive network is used for denoising the OCT image;
inputting the sample data set into the feature fusion attention intensive network for training and testing until convergence;
and inputting the image to be detected into a trained feature fusion attention-intensive network for processing to obtain a clear OCT image after denoising processing.
Preferably, the method further comprises the following steps: and preprocessing the sample data set and the image to be detected.
Preferably, the preprocessing includes graying, clipping, and filtering processing.
Preferably, the sparse block is used for learning the noise distribution in the image after each residual error network inputs a new fusion feature;
the characteristic enhancement block is used for extracting an image containing noise in a shallow layer;
the attention mechanism is used for summarizing and evaluating the importance of information of each channel between layers, extracting an attention heat map of a space when summarizing and evaluating the importance of the same position of multiple channels, and finally multiplying the attention heat map back to the original channel;
and the reconstruction block is used for fusing the global features output by the previous layer and keeping the sizes of the input and the output consistent.
Preferably, the feature fusion attention-intensive network module is divided into fifteen layers, wherein the first layer is Conv + ReLU, the convolution kernel of the convolution layer has the size of 3 × 3, and the number of the convolution kernels is 128, and the convolution kernel is used for performing shallow feature extraction on an input image containing noise; the second layer is Conv, the convolution kernel size is 3 x 3, the number is 48, and the second layer is used for adjusting the channel size of the feature map; third to twelfth layers: the image fusion system comprises 4 Dense Block modules, four residual blocks and a Transition layer between every two modules, wherein in the Dense Block modules, the sizes of convolution kernels of all residual structures are 3 × 3, the numbers of the convolution kernels are 64 respectively, each residual network can learn the noise distribution in an image after new fusion characteristics are input, and the sizes of the convolution kernels in the Transition layer are 1 × 1, and the numbers of the convolution kernels are 48; the thirteenth layer is a Concat layer and is used for fusing the feature modules; the fourteenth layer is an attention mechanism layer, the feature graph generated by the dense connecting blocks is sent to an attention module to generate a new feature graph, the new feature graph is added to the feature graph generated by the jumping connection of the encoder end, then the feature graph is subjected to deconvolution by 3 x 3 and then sent to the dense connecting blocks, the operation is repeated, and finally the high-resolution feature graph generated by the network is converted into the feature graph with the channel number of 2 through convolution mapping of 1 x 1; the fifteenth layer is a reconstruction output layer of a single convolution kernel, and is used for fusing the global features of the previous layer and keeping the size of the output consistent with that of the input.
Preferably, the loss function expression of the feature fusion attention-intensive network is as follows:
in the formula (I), the compound is shown in the specification,residual image input for noise, yiFor input noisy images, xiAn image without any foreign matter, (y)i-xi) The result is a standard residual image, N being the number of input samples.
Preferably, the weight value of the feature fusion attention-intensive network is updated as follows:
wherein W is a convolution kernel; l is the second convolution layer; b is the number of iterations; α is a learning rate.
An OCT image denoising apparatus, comprising: an acquisition module, a construction module, a model acquisition module and a detection module, wherein,
the acquisition module is used for acquiring an OCT image to form a sample data set, and the sample data set is divided into a training set and a verification set;
the construction module is used for constructing a feature fusion attention intensive network for OCT image denoising, and the main structure of the feature fusion attention intensive network comprises a sparse block, a feature enhancement block, an attention mechanism and a reconstruction block and is used for denoising the OCT image;
the model acquisition module is used for inputting the sample data set into the feature fusion attention intensive network for training and testing until convergence;
and the detection module is used for inputting the image to be detected into the trained feature fusion attention-intensive network for processing to obtain the clear OCT image after denoising processing.
A computer device, comprising: a memory for storing a computer program; a processor for implementing an OCT image denoising method as described in any one of the above when executing the computer program.
A readable storage medium having stored thereon a computer program which, when executed by a processor, implements an OCT image denoising method as recited in any one of the above.
Based on the technical scheme, the invention has the beneficial effects that:
1) according to the method, densely connected dense blocks are arranged in the network, each layer takes the previous feature map as input, and the shallow convolution feature map is connected with the short lines of the deep convolution feature map extracted from each dense block in sequence by combining a progressive idea to form a residual block, so that the network can better predict noise distribution;
2) the invention adopts the DenseNet with the attention mechanism to focus attention on key features and inhibit other irrelevant features in the feature extraction process;
3) the dense short circuit structure designed by the invention can effectively reduce the computation complexity of the network, reduce a large number of network parameters and shorten the computation time of the algorithm.
Drawings
FIG. 1 is a flow diagram of an OCT image denoising method in one embodiment;
FIG. 2 is a block diagram of a feature fusion attention-intensive network architecture;
fig. 3 is a test artwork for an embodiment with a noise level of 15, 25, and 40;
fig. 4 is a graph of the denoising result when the noise level is 15, 25, and 40 in one embodiment;
FIG. 5 is a schematic structural diagram of an OCT image denoising apparatus in an embodiment;
FIG. 6 is a diagram illustrating an internal structure of a computer device in one embodiment.
Detailed Description
The technical solution in the embodiments of the present invention will be clearly and completely described below with reference to the accompanying drawings in the embodiments of the present invention.
As shown in fig. 1, the present embodiment provides an OCT image denoising method, which includes the following specific steps:
and step 110, acquiring the OCT image to form a sample data set, wherein the sample data set is divided into a training set and a verification set.
In this embodiment, 512 OCT images of 224 × 224 pixels are selected from the sample data set. In order to make the network training converge faster, the images used in the experiment are uniformly cut into 512 × 451 (width × height), namely, the middle effective retina area is taken, 412 images are cut into 200000 sub image blocks of 50 × 50 to be used as a training set, the remaining 100 images are used as a verification set, and the robustness and the practicability of the network are verified by carrying out multiple experiments. The experimental verification sets are all selected from noise image data sets constructed by SD-OCT diabetic retina images of different patients.
Step 120, constructing a feature fusion attention intensive network for OCT image denoising; the main structure of the feature fusion attention-intensive network comprises a sparse block, a feature enhancement block, an attention mechanism and a reconstruction block, and is used for denoising the OCT image.
In this embodiment, the feature fusion attention-intensive network (DnFFA) mainly utilizes four modules, as shown in fig. 2: a sparse block (DB), a Feature Enhancement Block (FEB), an Attention Mechanism (AM) and a Reconstruction Block (RB) for image denoising. The linear correction unit ReLU is adopted by each convolution-laminated activation function, and the correction unit enables the network to be sparse, so that the dependency relationship among parameters is reduced, and the problem of disappearance of network gradients is avoided.
The sparse block (DB) learns the noise distribution in the image after the new fusion feature is input in each residual network, wherein the dense block structure connects the layers by short-circuiting lines, the output of each layer will be used as the input of the next layer, and the connection is expressed by the following formula:
Ml=Xl([M0,M1,...,ML-1])
in the formula: mlRepresents the output characteristic diagram of the l-th layer, [ M ]0,M1,...,ML-1]The output characteristic diagrams of the 0 to l-1 layers are merged into channels, namely the channels are directly stacked without other operations on the channels. XlIt means that the merged feature maps are sequentially input into the bottleneck layer. Because the output of each layer will be used as the input of the next layer, all the inputs are superposed, and the extracted feature maps of all the previous layers are fused, the input channels of the next layers will be increased by a lot. By designing a convolution layer with convolution kernel of 3 multiplied by 3 in the bottleneck layer, the number of characteristic graphs is reduced, the calculated amount of the network can be reduced, and the gradient transfer efficiency is improved;
the Feature Enhancement Block (FEB) mainly aims at shallow extraction of images containing noise; the Attention Mechanism (AM) is to perform summary evaluation on the importance of information of each channel among layers to improve the proportion of the important information, and when the importance of the same position of multiple channels is performed the summary evaluation, an attention heat map of a space is extracted and finally multiplied back to the original channel. Therefore, the network performance can be effectively improved on the basis of not increasing the network parameter quantity; the Reconstruction Block (RB) fuses the global features output by the previous layer and keeps the size of the input and the output consistent.
The DnFFA network modules are mainly divided into fifteen layers. The first layer was Conv + ReLU, the convolution kernel size of this convolutional layer was 3 x 3, and the number was 128. The main function is to perform shallow feature extraction on an input image containing noise. The second layer is Conv, the convolution kernel size is 3 x 3, and the number is 48. The method has the main function of adjusting the channel size of the feature diagram, and is convenient for removing redundant feature diagrams in the Dense Block module, thereby reducing the calculation burden of the network. The most core part of the network appears in the third to twelfth layers, and the part mainly comprises 4 Dense Block modules, four residual blocks, and a Transition layer and a ReLU + Conv layer between each module. In the sense Block module, the convolution kernels of all residual structures have a size of 3 × 3, and the number of the convolution kernels is 64. Each residual network can learn the noise distribution in the image after the new fusion features are input. The convolution kernels in the TransitionLayer are all 1 x 1 in size and 48 in number, so that the network parameters are further compressed, and the channel features are fused to form a new feature. The thirteenth layer is a Concat layer, and the layer mainly fuses feature modules, namely the feature modules are fused with the first and second sense Block modules and the third sense Block module to strengthen the global features of the learning network. And the fourteenth layer is an attention mechanism layer, the feature map generated by the dense connecting block is sent to an attention module to generate a new feature map, the new feature map is added with the feature map generated by the skip connection at the encoder end, then the feature map is subjected to deconvolution by 3 x 3 and then sent to the dense connecting block, the process is repeated, finally the high-resolution feature map generated by the network is converted into the feature map with the channel number of 2 through convolution mapping of 1 x 1, the layer has the main function of extracting noise information hidden in a complex background, and the module is very effective for noise images, namely real noise images and blind noise. The fifteenth layer is a reconstructed output layer of a single convolution kernel, which fuses the global features of the previous layer and keeps the output size consistent with the input size, so that the predicted noise can be more easily extracted from the image containing noise through residual error learning, and the output of the network is a purer image.
Step 130, inputting the sample data set into the feature fusion attention intensive network for training and testing until convergence.
In this embodiment, these noise images are input into a designed feature fusion attention-intensive network during training, and parameters are adjusted by back propagation through a loss function, so that the network finally reaches a convergence state. And in the testing stage, the noise-containing image is input into the converged network, and the corresponding prediction de-noising image can be directly output. The loss function expression used is as follows:
in the formula (I), the compound is shown in the specification,residual image input for noise, yiFor input noisy images, xiAn image without any foreign matter, (y)i-xi) The result is a standard residual image. N is the number of samples input. The network model is trained in a batch processing mode, namely N noisy images are input each time, the N noisy images are correspondingly output through a feature fusion attention-intensive network, then a loss function L is calculated, parameters are optimized by using an adam (adaptive motion estimation) method, and then the weight value is updated as follows:
in the formula: w is a convolution kernel; l is the second convolution layer; b is the number of iterations; α is learning rate, and is initially set to 10-3With training, it will fall to 10-4. Through continuous iteration of the training process, the error between the estimated residual error and the standard residual error is reduced, so that the predicted de-noised image is closer to the original image, and a better de-noising effect is obtained.
And 140, inputting the image to be detected into the trained feature fusion attention-intensive network for processing to obtain the OCT image after denoising processing.
In one embodiment, in order to achieve the purpose of removing noise and simultaneously maximally preserving original image features, a preprocessing method based on a dense network with progressive feature fusion attention is provided, and the preprocessing method specifically comprises the following steps:
step 111: firstly, reading an image with speckle noise, carrying out gray processing on the image to obtain f (x, y), and uniformly setting the window size of the image to be 224 × 224 pixels;
step 112: the image data set obtained in step 111 is decomposed into high frequency detail layers y by filtering using a low pass filterhighAnd a low frequency background layer ylowThe formula is as follows:
y=yhigh+ylow
some pixels with strong gray changes, namely pixels which are not processed yet, are reserved in the high-frequency layer. The information contained in the low-frequency layer of the processed image is the same, and the pixel points are all 0;
step 113: the feature fusion attention intensive network is fully pre-trained, and edge profile features such as edges, corners and the like can be efficiently extracted by the low-layer convolution layer of the network;
step 114: after the step 113, the weight of the high layer of the network can be changed by using less data for training, so that the network learns more abstract characteristics;
step 115: as can be seen from the above steps, noise is basically present in the high frequency layer, and the high frequency layer is input to the residual network of the image.
Experiment of
The experiment is carried out in a Linux system, a Deep Learning framework of Deep Learning Toolbox is used for network training, the hardware configuration is an Intel Xeon CPU, the memory is 16G, and the display card is 11GB NVIDIATeslaV100 PCIe GPU. OCT image data is from a common dataset: the AI-filler 2018for automatic segmentation of the relevant images data set contains a lot of speckle noise in 85 OCT images of the retina cube (1024 × 512 × 128), and in the original OCT images, there are many areas with pure noise. In the experiment, the larger the PSNR value is, the higher the similarity between the denoised image and the original image is, and the better the denoising effect is. The PSNR is calculated as:
in the formula: m and N represent predicted values and true values; j and k represent all the pixels in the image. H and W respectively represent the height and width of the image, and the value of n is 8.
In order to evaluate the denoising effect more objectively and fairly, five comparison algorithms of WNNM, BM3D, NLM and DnCNN and DnFFA of the invention are selected for the experiment, FIG. 3 is a test original image, a verification set of the experiment is from an OCT image, and the denoising performance of each algorithm is compared from the subjective and objective aspects when Gaussian white noise levels of 15, 25 and 40 are respectively added in the image verification set. Wherein in the image denoising field, BM3D is considered as the best denoising algorithm, and DnCNN is an advanced denoising algorithm in the deep learning field. The denoising result is shown in fig. 4, it can be seen that the WNNM algorithm is very limited in the aspect of capturing the features of the image structure, the BM3D algorithm edge is well preserved, although the noise points are reduced, a lot of clear noise points can be seen, but the number of noise points of BM3D is obviously much smaller than that of wavelet denoising, the DnCNN algorithm has a good effect on white gaussian noise, and NLM can retain more detailed information, but the edge is still not clear enough. The experimental results are shown in the following table, with bold font representing the best values of the test results.
As can be seen from table 1, in the case of noise levels of 15, 25, and 40, the noise reduction effect of DnFFA is the best, and compared with the best noise reduction algorithm BM3D considered in the image noise reduction field, the noise level PSNR of DnFFA is all over about 0.31 on average, and compared with other algorithms, the noise reduction effect of DnFFA has the best performance in PSNR value.
TABLE 1 average PSNR (dB) of the denoise results of OCT test plots under 5 algorithms
The OCT image quality reference index characteristic shows that the image quality is better when the PSNR is larger, and the table 1 shows that the denoising result of the DnFFA algorithm can obtain higher PSNR values for 5 noises with different levels, and the PSNR increases more and more along with the enhancement of the noise level, so that the improvement effect is better and better. The experimental result shows that the comparison of the 5 algorithms shows that the DnFFA network structure has better denoising effect to a certain extent.
In one embodiment, as shown in fig. 5, an OCT image denoising apparatus 200 includes: an acquisition module 210, a construction module 220, a model acquisition module 230, and a detection module 240, wherein,
the acquiring module 210 is configured to acquire an OCT image forming sample data set, where the sample data set is divided into a training set and a verification set;
the constructing module 220 is configured to construct a feature fusion attention-intensive network for OCT image denoising, where a main structure of the feature fusion attention-intensive network includes a sparse block, a feature enhancing block, an attention mechanism, and a reconstruction block, and is configured to denoise an OCT image;
the model obtaining module 230 is configured to input a sample data set into the feature fusion attention-intensive network for training and testing until convergence;
the detection module 240 is configured to input the image to be detected into the trained feature fusion attention-intensive network for processing, so as to obtain a clear OCT image after denoising processing.
The apparatuses or modules illustrated in the above embodiments may be implemented by a computer chip or an entity, or by a product with certain functions. One typical implementation device is a computer. In particular, the computer may be, for example, a personal computer, a laptop computer, a cellular telephone, a camera phone, a smartphone, a personal digital assistant, a media player, a navigation device, an email device, a game console, a tablet computer, a wearable device, or a combination of any of these devices.
In one embodiment, a computer device is provided, which may be a terminal, and its internal structure diagram may be as shown in fig. 6. The computer device includes a processor, a memory, a network interface, a display screen, and an input device connected by a system bus. Wherein the processor of the computer device is configured to provide computing and control capabilities. The memory of the computer device comprises a nonvolatile storage medium and an internal memory. The non-volatile storage medium stores an operating system and a computer program. The internal memory provides an environment for the operation of an operating system and computer programs in the non-volatile storage medium. The network interface of the computer device is used for communicating with an external terminal through a network connection. The computer program is executed by a processor to implement an OCT image denoising method. The display screen of the computer equipment can be a liquid crystal display screen or an electronic ink display screen, and the input device of the computer equipment can be a touch layer covered on the display screen, a key, a track ball or a touch pad arranged on the shell of the computer equipment, an external keyboard, a touch pad or a mouse and the like.
Those skilled in the art will appreciate that the architecture shown in fig. 6 is merely a block diagram of some of the structures associated with the disclosed aspects and is not intended to limit the computing devices to which the disclosed aspects apply, as particular computing devices may include more or less components than those shown, or may combine certain components, or have a different arrangement of components.
In one embodiment, a readable storage medium is provided, on which a computer program is stored, which when executed by a processor implements the OCT image denoising method as described in any one of the above.
It will be understood by those skilled in the art that all or part of the processes of the OCT image denoising method according to the above embodiments may be implemented by a computer program, which may be stored in a non-volatile computer-readable storage medium, and the computer program may include the processes of the above embodiments of the methods when executed. Any reference to memory, storage, database, or other medium used in the embodiments provided herein may include non-volatile and/or volatile memory, among others. Non-volatile memory can include read-only memory (ROM), Programmable ROM (PROM), Electrically Programmable ROM (EPROM), Electrically Erasable Programmable ROM (EEPROM), or flash memory. Volatile memory can include Random Access Memory (RAM) or external cache memory. By way of illustration and not limitation, RAM is available in a variety of forms such as Static RAM (SRAM), Dynamic RAM (DRAM), Synchronous DRAM (SDRAM), Double Data Rate SDRAM (DDRSDRAM), Enhanced SDRAM (ESDRAM), Synchronous Link DRAM (SLDRAM), Rambus Direct RAM (RDRAM), direct bus dynamic RAM (DRDRAM), and memory bus dynamic RAM (RDRAM).
The above description is only the preferred embodiment of the OCT image denoising method, apparatus, device and medium disclosed in the present invention, and is not intended to limit the scope of protection of the embodiments of the present specification. Any modification, equivalent replacement, improvement and the like made within the spirit and principle of the embodiments of the present disclosure should be included in the protection scope of the embodiments of the present disclosure.
Claims (10)
1. An OCT image denoising method is characterized by comprising the following steps:
obtaining an OCT image to form a sample data set, wherein the sample data set is divided into a training set and a verification set;
constructing a feature fusion attention intensive network for OCT image denoising, wherein the main structure of the feature fusion attention intensive network comprises a sparse block, a feature enhancement block, an attention mechanism and a reconstruction block, and the feature fusion attention intensive network is used for denoising the OCT image;
inputting the sample data set into the feature fusion attention intensive network for training and testing until convergence;
and inputting the image to be detected into a trained feature fusion attention-intensive network for processing to obtain a clear OCT image after denoising processing.
2. The OCT image denoising method of claim 1, further comprising: and preprocessing the sample data set and the image to be detected.
3. The method of claim 2, wherein the preprocessing comprises graying, cropping, and filtering.
4. The OCT image denoising method of claim 1, wherein the sparse block is configured to learn noise distribution in the image after inputting the new fusion features for each residual network;
the characteristic enhancement block is used for extracting an image containing noise in a shallow layer;
the attention mechanism is used for summarizing and evaluating the importance of information of each channel between layers, extracting an attention heat map of a space when summarizing and evaluating the importance of the same position of multiple channels, and finally multiplying the attention heat map back to the original channel;
and the reconstruction block is used for fusing the global features output by the previous layer and keeping the sizes of the input and the output consistent.
5. The OCT image denoising method of claim 4, wherein the feature fusion attention-intensive network module is divided into fifteen layers, wherein the first layer is Conv + ReLU, the convolution kernel size of the convolution layer is 3 x 3, and the number of the convolution kernels is 128, and the shallow feature extraction is performed on the input image containing noise; the second layer is Conv, the convolution kernel size is 3 x 3, the number is 48, and the second layer is used for adjusting the channel size of the feature map; third to twelfth layers: the image fusion system comprises 4 Dense Block modules, four residual blocks and a Transition layer between every two modules, wherein in the Dense Block modules, the sizes of convolution kernels of all residual structures are 3 × 3, the numbers of the convolution kernels are 64 respectively, each residual network can learn the noise distribution in an image after new fusion characteristics are input, and the sizes of the convolution kernels in the Transition layer are 1 × 1, and the numbers of the convolution kernels are 48; the thirteenth layer is a Concat layer and is used for fusing the feature modules; the fourteenth layer is an attention mechanism layer, the feature graph generated by the dense connecting blocks is sent to an attention module to generate a new feature graph, the new feature graph is added to the feature graph generated by the jumping connection of the encoder end, then the feature graph is subjected to deconvolution by 3 x 3 and then sent to the dense connecting blocks, the operation is repeated, and finally the high-resolution feature graph generated by the network is converted into the feature graph with the channel number of 2 through convolution mapping of 1 x 1; the fifteenth layer is a reconstruction output layer of a single convolution kernel, and is used for fusing the global features of the previous layer and keeping the size of the output consistent with that of the input.
6. The OCT image denoising method of claim 1, wherein the loss function expression of the feature fusion attention-intensive network is as follows:
8. An OCT image denoising apparatus, comprising: an acquisition module, a construction module, a model acquisition module and a detection module, wherein,
the acquisition module is used for acquiring an OCT image to form a sample data set, and the sample data set is divided into a training set and a verification set;
the construction module is used for constructing a feature fusion attention intensive network for OCT image denoising, and the main structure of the feature fusion attention intensive network comprises a sparse block, a feature enhancement block, an attention mechanism and a reconstruction block and is used for denoising the OCT image;
the model acquisition module is used for inputting the sample data set into the feature fusion attention intensive network for training and testing until convergence;
and the detection module is used for inputting the image to be detected into the trained feature fusion attention-intensive network for processing to obtain the clear OCT image after denoising processing.
9. A computer device, comprising: a memory for storing a computer program; a processor for implementing an OCT image denoising method as claimed in any one of claims 1 to 7 when executing the computer program.
10. A readable storage medium, characterized in that the readable storage medium has stored thereon a computer program which, when being executed by a processor, implements an OCT image denoising method according to any one of claims 1 to 7.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202111579290.6A CN114240797B (en) | 2021-12-22 | 2021-12-22 | OCT image denoising method, device, equipment and medium |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202111579290.6A CN114240797B (en) | 2021-12-22 | 2021-12-22 | OCT image denoising method, device, equipment and medium |
Publications (2)
Publication Number | Publication Date |
---|---|
CN114240797A true CN114240797A (en) | 2022-03-25 |
CN114240797B CN114240797B (en) | 2023-04-18 |
Family
ID=80761218
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN202111579290.6A Active CN114240797B (en) | 2021-12-22 | 2021-12-22 | OCT image denoising method, device, equipment and medium |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN114240797B (en) |
Cited By (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN114675118A (en) * | 2022-05-30 | 2022-06-28 | 广东电网有限责任公司佛山供电局 | Transformer winding abnormality detection method, device, equipment and storage medium |
CN114936974A (en) * | 2022-05-12 | 2022-08-23 | 中山大学中山眼科中心 | Semi-supervised OCT image denoising method and device based on attention mechanism |
CN114972130A (en) * | 2022-08-02 | 2022-08-30 | 深圳精智达技术股份有限公司 | Training method, device and training equipment for denoising neural network |
CN116797493A (en) * | 2023-08-02 | 2023-09-22 | 北京中科闻歌科技股份有限公司 | Image denoising processing system |
Citations (7)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20170213321A1 (en) * | 2016-01-22 | 2017-07-27 | Siemens Healthcare Gmbh | Deep Unfolding Algorithm For Efficient Image Denoising Under Varying Noise Conditions |
CN110675330A (en) * | 2019-08-12 | 2020-01-10 | 广东石油化工学院 | Image rain removing method of encoding-decoding network based on channel level attention mechanism |
CN111242862A (en) * | 2020-01-09 | 2020-06-05 | 西安理工大学 | Multi-scale fusion parallel dense residual convolution neural network image denoising method |
CN111292259A (en) * | 2020-01-14 | 2020-06-16 | 西安交通大学 | Deep learning image denoising method integrating multi-scale and attention mechanism |
US20210174149A1 (en) * | 2018-11-20 | 2021-06-10 | Xidian University | Feature fusion and dense connection-based method for infrared plane object detection |
CN113191983A (en) * | 2021-05-18 | 2021-07-30 | 陕西师范大学 | Image denoising method and device based on deep learning attention mechanism |
CN113610719A (en) * | 2021-07-19 | 2021-11-05 | 河南大学 | Attention and dense connection residual block convolution kernel neural network image denoising method |
-
2021
- 2021-12-22 CN CN202111579290.6A patent/CN114240797B/en active Active
Patent Citations (7)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20170213321A1 (en) * | 2016-01-22 | 2017-07-27 | Siemens Healthcare Gmbh | Deep Unfolding Algorithm For Efficient Image Denoising Under Varying Noise Conditions |
US20210174149A1 (en) * | 2018-11-20 | 2021-06-10 | Xidian University | Feature fusion and dense connection-based method for infrared plane object detection |
CN110675330A (en) * | 2019-08-12 | 2020-01-10 | 广东石油化工学院 | Image rain removing method of encoding-decoding network based on channel level attention mechanism |
CN111242862A (en) * | 2020-01-09 | 2020-06-05 | 西安理工大学 | Multi-scale fusion parallel dense residual convolution neural network image denoising method |
CN111292259A (en) * | 2020-01-14 | 2020-06-16 | 西安交通大学 | Deep learning image denoising method integrating multi-scale and attention mechanism |
CN113191983A (en) * | 2021-05-18 | 2021-07-30 | 陕西师范大学 | Image denoising method and device based on deep learning attention mechanism |
CN113610719A (en) * | 2021-07-19 | 2021-11-05 | 河南大学 | Attention and dense connection residual block convolution kernel neural network image denoising method |
Cited By (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN114936974A (en) * | 2022-05-12 | 2022-08-23 | 中山大学中山眼科中心 | Semi-supervised OCT image denoising method and device based on attention mechanism |
CN114675118A (en) * | 2022-05-30 | 2022-06-28 | 广东电网有限责任公司佛山供电局 | Transformer winding abnormality detection method, device, equipment and storage medium |
CN114972130A (en) * | 2022-08-02 | 2022-08-30 | 深圳精智达技术股份有限公司 | Training method, device and training equipment for denoising neural network |
CN114972130B (en) * | 2022-08-02 | 2022-11-18 | 深圳精智达技术股份有限公司 | Training method, device and training equipment for denoising neural network |
CN116797493A (en) * | 2023-08-02 | 2023-09-22 | 北京中科闻歌科技股份有限公司 | Image denoising processing system |
CN116797493B (en) * | 2023-08-02 | 2024-01-26 | 北京中科闻歌科技股份有限公司 | Image denoising processing system |
Also Published As
Publication number | Publication date |
---|---|
CN114240797B (en) | 2023-04-18 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN114240797B (en) | OCT image denoising method, device, equipment and medium | |
CN108053417B (en) | lung segmentation device of 3D U-Net network based on mixed rough segmentation characteristics | |
CN110706174B (en) | Image enhancement method, terminal equipment and storage medium | |
CN105761216B (en) | A kind of image denoising processing method and processing device | |
CN109377457B (en) | Pill coating image processing method and device, computer equipment and storage medium | |
CN108830791B (en) | Image super-resolution method based on self sample and sparse representation | |
CN113362338B (en) | Rail segmentation method, device, computer equipment and rail segmentation processing system | |
Zou et al. | 3D filtering by block matching and convolutional neural network for image denoising | |
CN114897741A (en) | Image blind deblurring method based on depth residual Fourier transform | |
CN114331913B (en) | Motion blurred image restoration method based on residual attention block | |
CN104036472A (en) | Method and device for enhancing quality of 3D image | |
CN115205136A (en) | Image rain removing method based on Fourier prior | |
CN113379647B (en) | Multi-feature image restoration method for optimizing PSF estimation | |
Ma et al. | Edge-guided cnn for denoising images from portable ultrasound devices | |
Talavera-Martínez et al. | An encoder-decoder CNN for hair removal in dermoscopic images | |
CN111340699B (en) | Magnetic resonance image denoising method and device based on non-local prior and sparse representation | |
Tian et al. | Retinal fundus image superresolution generated by optical coherence tomography based on a realistic mixed attention GAN | |
Bania et al. | Adaptive trimmed median filter for impulse noise detection and removal with an application to mammogram images | |
Saoji et al. | Speckle and rician noise removal from medical images and Ultrasound images | |
Zhang et al. | Deep residual network based medical image reconstruction | |
Gupta et al. | A new computational approach for edge-preserving image decomposition | |
CN115526785A (en) | OCT image denoising method based on FCFDS improved NLM | |
Si et al. | TFR: Texture Defect Detection with Fourier Transform using Normal Reconstructed Template of Simple Autoencoder | |
Oroumchian | ECEI Spectrogram Denoising and Feature Extraction Using a Multi-Wavelet Convolutional Neural Network Model | |
Shang et al. | A non-convex low-rank image decomposition model via unsupervised network |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
GR01 | Patent grant | ||
GR01 | Patent grant |