CN114842104A - Capsule endoscope image super-resolution reconstruction method based on multi-scale residual errors - Google Patents

Capsule endoscope image super-resolution reconstruction method based on multi-scale residual errors Download PDF

Info

Publication number
CN114842104A
CN114842104A CN202210539075.1A CN202210539075A CN114842104A CN 114842104 A CN114842104 A CN 114842104A CN 202210539075 A CN202210539075 A CN 202210539075A CN 114842104 A CN114842104 A CN 114842104A
Authority
CN
China
Prior art keywords
resolution
module
feature
convolution
super
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN202210539075.1A
Other languages
Chinese (zh)
Inventor
黄胜
陈贤龙
廖星
曹维俊
牟星宇
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Chongqing University of Post and Telecommunications
Original Assignee
Chongqing University of Post and Telecommunications
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Chongqing University of Post and Telecommunications filed Critical Chongqing University of Post and Telecommunications
Priority to CN202210539075.1A priority Critical patent/CN114842104A/en
Publication of CN114842104A publication Critical patent/CN114842104A/en
Pending legal-status Critical Current

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T11/002D [Two Dimensional] image generation
    • G06T11/003Reconstruction from projections, e.g. tomography
    • G06T11/005Specific pre-processing for tomographic reconstruction, e.g. calibration, source positioning, rebinning, scatter correction, retrospective gating
    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61BDIAGNOSIS; SURGERY; IDENTIFICATION
    • A61B1/00Instruments for performing medical examinations of the interior of cavities or tubes of the body by visual or photographical inspection, e.g. endoscopes; Illuminating arrangements therefor
    • A61B1/04Instruments for performing medical examinations of the interior of cavities or tubes of the body by visual or photographical inspection, e.g. endoscopes; Illuminating arrangements therefor combined with photographic or television appliances
    • A61B1/041Capsule endoscopes for imaging
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/045Combinations of networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T3/00Geometric image transformation in the plane of the image
    • G06T3/40Scaling the whole image or part thereof
    • G06T3/4053Super resolution, i.e. output image resolution higher than sensor resolution
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/70Arrangements for image or video recognition or understanding using pattern recognition or machine learning
    • G06V10/77Processing image or video features in feature spaces; using data integration or data reduction, e.g. principal component analysis [PCA] or independent component analysis [ICA] or self-organising maps [SOM]; Blind source separation
    • G06V10/80Fusion, i.e. combining data from various sources at the sensor level, preprocessing level, feature extraction level or classification level
    • G06V10/806Fusion, i.e. combining data from various sources at the sensor level, preprocessing level, feature extraction level or classification level of extracted features
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/70Arrangements for image or video recognition or understanding using pattern recognition or machine learning
    • G06V10/82Arrangements for image or video recognition or understanding using pattern recognition or machine learning using neural networks

Abstract

The invention provides a super-resolution reconstruction method based on multi-scale residual error for super-resolution reconstruction of capsule endoscope images, which aims at the problem that clear images with structure, detail and other characteristics cannot be obtained due to low camera resolution and inspection environment limitation in Capsule Endoscope (CE) detection, wherein a super-resolution reconstruction network mainly comprises two parts, one part is a characteristic extraction part consisting of a shallow characteristic extraction module, a multi-scale residual error characteristic extraction Module (MRCB) and a hierarchical characteristic processing module (HFCA), used for extracting and utilizing information under different scales from the input low-resolution image (LR), and endowing corresponding weight to the characteristic information while avoiding information loss, for subsequent upsampled reconstruction, and an image reconstruction module consisting of a sub-pixel layer upsampling method and a convolutional layer for generating a final super-resolution image (SR).

Description

Capsule endoscope image super-resolution reconstruction method based on multi-scale residual errors
Technical Field
The invention relates to the field of deep learning and computer vision, in particular to a capsule endoscope image super-resolution reconstruction method based on multi-scale residual errors.
Background
Capsule Endoscopy (CE) is a modality used for non-invasive examination of the digestive tract, and is currently the first choice for diagnosis of digestive tract diseases, particularly small bowel diseases. Capsule endoscopy is a capsule-sized electronic device that is swallowed by a patient, primarily containing a CMOS camera, and typically produces about 6 million images in a single capsule endoscopy, for which 6 million images a physician typically uses a specific gaze pattern to observe pathology-related visual cues to diagnose pathology from CE images. In recent years, the rapid development of deep learning realizes the automatic and objective analysis of endoscope images, including depth estimation, polyp detection and feature description, which reduces the workload of doctors to a certain extent and improves the working efficiency. However, due to poor illumination, limited camera performance, and complex environments in the intestinal tract, images captured by CE tend to be Low Resolution (LR) which poses a limitation of capsule endoscopy. Research shows that low-resolution images have great influence on the diagnosis method, and the enhanced image quality may have better effect on the aspects of lesion detection, region segmentation, pathological analysis and the like. Therefore, there is a need for a method of improving the resolution of a capsule endoscope for subjective and objective analysis.
In order to obtain high-resolution endoscopic images, the most direct method is to improve hardware devices and add optical elements, but the cost required for the method is high, and the method is limited by the space size of the capsule endoscope device, and the method is not reasonable. To address this problem, the computer vision community has been studying a series of algorithms known as super resolution, which use signal processing techniques to reconstruct high resolution images from low resolution images. Using super-resolution techniques to overcome the fundamental resolution limit to achieve better image quality is a very challenging reconstruction process.
Super-resolution reconstruction is a super-resolution reconstruction (SR) technique that enables a low-resolution image captured under the influence of imaging condition limitations, a transmission medium, and the like to be reconstructed into a high-resolution image having richer texture details, greater pixel density, and higher reliability. According to the number of input images, the image super-resolution reconstruction can be divided into single-image super-resolution reconstruction and multi-image super-resolution reconstruction. In practical studies, single image super resolution reconstruction (SISR) is the hot spot of research.
Single image super resolution reconstruction (SISR) has two large categories in common, one based on tradition and the other based on deep learning. At present, with the rapid development of deep learning and the excellent performance of the deep learning in the aspect of image processing, the SR technology has achieved good success in the field of computer vision, and has been widely applied in the fields of medical imaging, security monitoring, satellite remote sensing, video recovery, and the like, and the development speed and the application achievement thereof are far superior to those of the traditional image super-resolution reconstruction method.
Due to the wide variety of medical images and the difficulty of acquiring medical image datasets. At present, super-resolution reconstruction on medical images mainly aims at medical images such as MRI, PET, CT, retina and the like, only a few researchers carry out super-resolution technology research on capsule endoscope images, and the research effect is not ideal.
The invention discloses a single-image super-resolution reconstruction based on deep learning, which has the key problem of how to mine and utilize an input LR image. The resolution of the reconstructed capsule endoscopic image is greater than the input capsule endoscopic image.
Disclosure of Invention
Aiming at the defects in the prior art, the invention provides a capsule endoscope image super-resolution reconstruction method based on multi-scale residual errors, so that the super-resolution (SR) reconstruction of a capsule endoscope image is realized on the basis, and a CE image with clear characteristics such as structure and texture is generated.
In order to achieve the purpose, the invention adopts the following solution: a super-resolution reconstruction method for a capsule endoscope image comprises the following specific scheme and steps:
1. acquiring a capsule endoscope image of a patient to be diagnosed, and acquiring a pre-trained super-resolution reconstruction network, wherein the super-resolution reconstruction network (MRCN) consists of two parts, one part is a feature extraction part consisting of a shallow feature extraction module, a multi-scale residual feature extraction Module (MRCB) and a hierarchical feature processing module (HFCA), and the other part is an image reconstruction module consisting of a convolution layer formed by a sub-pixel layer up-sampling method and a convolution kernel with the size of 3x 3;
2. inputting a capsule endoscope image to be reconstructed into a shallow feature extraction module to perform shallow feature extraction to obtain a primary feature map;
3. sequentially passing the primary feature map through n MRCB modules to generate and obtain a secondary feature map;
4. extracting the hierarchical feature maps generated in the MRCB modules, inputting the two-level feature maps and the hierarchical feature maps generated in the MRCB modules into an HFCA module, fusing the two-level feature maps and the hierarchical feature maps by using the HFCA module, enhancing important features in the images, and outputting to obtain a three-level feature map;
5. sending the three-level feature map into a convolution layer with 3x3 as a convolution kernel for further feature extraction to generate a four-level feature map;
6. and inputting the four-level characteristic diagram into an image reconstruction module, and carrying out amplification reconstruction to obtain a super-resolution capsule endoscope image, wherein the resolution of the super-resolution capsule endoscope image is greater than that of the capsule endoscope image to be reconstructed.
The shallow feature extraction module in the above step is a convolution layer with a convolution kernel size of 3 × 3.
The MRCB module adopts the residual blocks as basic structural units, each residual block is composed of two convolution layers and a ReLU activating function, and the residual blocks can be represented by the following mathematical model:
B 1 =φ 211 (B)))+B
B 2 =φ 423 (B)))+B
wherein B denotes information of the input residual block, phi 1 (. and phi) 2 (. The) shows a convolution operation with 3x3 as the convolution kernel, phi 3 (. and phi) 4 (. The) shows the convolution operation with 5x5 as the convolution kernel, σ 1 (. and σ) 2 (. The) shows an information activation operation with ReLU as the activation function, B 1 And B 2 Shown are the outputs of the residual blocks with 3x3 and 5x5 as convolution kernels, respectively;
the MRCB module MRCB is composed of 8 parallel branches, and can be specifically represented by the following mathematical model:
Y 1 =B 11 (X)
Y 2 =B 21 (X)
Y 3 =F 3 (concat(B 12 (F 1 (concat([Y 1 ,Y 2 ]))),B 22 (F 2 (concat([Y 1 ,Y 2 ]))),
B 13 (X),B 23 (X),X))
Y 4 =B 14 (Y 3 )
Y 5 =B 24 (Y 3 )
Y 6 =F 6 (concat(B 15 (F 4 (concat([Y 4 ,Y 5 ]))),B 25 (F 5 (concat([Y 4 ,Y 5 ]))),
B 16 (Y 3 ),B 26 (Y 3 ),B 17 (X),B 27 (X),Y 3 ))
Y=F 7 (Y 6 )+X
wherein X denotes information input to the MRCB module, B 11 (·)、B 12 (·)、B 13 (·)、B 14 (·)、B 15 (·)、B 16 (. and B) 17 (. 2) shows a convolution kernel residual block, B, of 3x3 21 (·)、B 22 (·)、B 23 (·)、B 24 (·)、B 25 (·)、B 26 (. and B) 27 Shown is a residual block convolved with 5x5, concat (a) shows a feature fusion operation, F 1 (·)、F 2 (·)、F 3 (·)、F 4 (·)、F 5 (. and F) 6 (. The) shows a convolution operation with 1x1 as the convolution kernel, F 7 (. cndot.) represents a convolution operation with 3x3 as the convolution kernel, and Y represents the final output of a single multi-scale feature extraction module.
The HFCA module is provided with a shuffling module and a mean value and standard deviation based enhanced attention module (COAM); cascading the secondary characteristic diagram and the hierarchical characteristic diagrams generated in all MRCB modules, inputting the cascaded characteristic diagrams into a shuffling module, and performing channel compression through convolution of 1x1 to obtain a fusion characteristic diagram; and inputting the fused feature map into a COAM module to obtain the three-level feature map.
The shuffling module groups the channel number of the input feature information according to the required group number to obtain the channel number in each group, the output channel number at the moment can be represented as a result of multiplying the channel number of each group by the group number, then Reshape operation is carried out on the channel number at the moment to convert the channel number into a matrix form, then transposition operation is carried out on the channel number, and after the transposition is finished, the channel number is flattened and grouped and output according to the initial group number.
The COAM module respectively performs mean pooling and standard deviation pooling on input information in a coordinate-like mode according to horizontal and vertical directions, cascades the pooled information, uses shuffling operation to enable information among channels to be fully fused, reduces the dimension of the fused information by using convolution with 1x1 as a convolution kernel, sends the reduced-dimension information to a nonlinear activation layer for activation operation, and then calls a function h-swish to process the activated features to obtain new features; dividing the new features along the horizontal and vertical directions to obtain two independent features, then adopting a convolution layer with 1x1 as a convolution kernel to carry out dimension increasing on the new features, and respectively adopting Sigmoid activation functions to activate the feature information after dimension increasing to obtain weight features in the horizontal and vertical directions; and finally, multiplying the two weighted characteristics by the input characteristic information to obtain the three-level characteristic diagram.
Further, in order to train the network better, while enabling the network to learn richer features, the data set used for the network training proposed by the present invention is composed of two parts, one is a picture of Kvasir that is not secondarily contaminated and destroyed, and the other is a picture of ETIS-lairb poly DB provided in MICCAI2015 tournament regarding Polyp symptoms in endoscopy. Specifically, there were 1000 images of the capsule endoscope covering different symptoms for training of the images, 20 images for verification, 26 images for testing, each image having a resolution between 1080P and 2K. Low resolution pictures throughout the experiment were generated using Bicubic (BI) and small scale (BD) degradation models. Finally, all SR results are evaluated for PSNR and SSIM on the Y channel of the YCbCr space on the output image.
During the training process, using RGB images as input, all training images are randomly rotated 90 °, 180 °, 270 °, and then horizontally flipped for effective data enhancement. Loss function of the network is L 1 Loss function, initial learning rate of 1x10 -4 The optimizer is Adam optimizer (and parameter β is set) 1 =0.9,β 2 =0.999,ε=10 -8 ). Each round of training is iterated 1000 times, 16 low-resolution patches with the size of 48x48 are randomly selected as input in each training batch, and the learning rate is halved every 200 rounds for 1000 rounds of training.
Due to the adoption of the technical scheme, the invention has the following advantages:
1. the multi-scale feature extraction module is adopted in feature extraction of the input image, combines the residual error thought and fully considers the influence of the external features of the network, namely the width of the network and the depth of the network, on the network performance, extracts feature streams of the image under different scales, and excavates and utilizes feature information of the input image;
2. in the process of processing the characteristic information among the layers, the method of sampling the channel shuffling enables the characteristic information among the layers to be fully fused, and the problem of information loss caused by insufficient fusion is effectively solved;
3. after the feature information between layers is subjected to primary processing, an enhanced attention mechanism based on a mean value and a standard deviation is used, the information between channels is captured in a coordinate mode, and corresponding weights are given to important feature information, so that the final super-resolution (SR) image has better texture and other feature details, and the quality of the final reconstructed image is improved.
Drawings
In order to make the object, technical scheme and beneficial effect of the invention more clear, the invention provides the following drawings for explanation:
FIG. 1 is a schematic view of the overall frame structure proposed by the present invention;
FIG. 2 is a schematic structural diagram of a multi-scale feature extraction module proposed by the present invention;
FIG. 3 is a schematic diagram of a hierarchical feature processing module according to the present invention;
FIG. 4 is a diagram of a channel shuffle used in the present invention;
detailed description of the preferred embodiments
The technical solution in the embodiments of the present invention will be clearly and completely described below with reference to the accompanying drawings in the embodiments of the present invention.
The invention provides a capsule endoscope image super-resolution reconstruction method based on multi-scale residual errors, which specifically comprises the following steps as shown in figure 1:
step 1, obtaining a capsule endoscope image of a patient to be diagnosed, and obtaining a pre-trained super-resolution reconstruction network, wherein the super-resolution reconstruction network (MRCN) consists of two parts, one part is a feature extraction part consisting of a shallow feature extraction module, a multi-scale residual feature extraction Module (MRCB) and a hierarchical feature processing module (HFCA), and the other part is an image reconstruction module consisting of a convolution layer formed by a sub-pixel layer up-sampling method and a convolution kernel with the size of 3x 3;
step 2, inputting the capsule endoscope image to be reconstructed into a shallow feature extraction module for shallow feature extraction to obtain a primary feature map, wherein the shallow feature extraction module is composed of a convolution layer with a convolution kernel size of 3x 3;
step 3, sequentially passing the primary feature map through n MRCB modules to generate and obtain a secondary feature map;
fig. 2 shows a schematic structural diagram of a single MRCB module, where the MRCB module uses a basic structural unit with a residual block as a structure, each residual block is composed of two convolution layers and an activation function, RU3 and RU5 in the structural diagram refer to residual blocks with convolution kernels of 3x3 and 5x5, respectively, and a specific structure of the residual block can be shown by the following formula:
B 1 =φ 211 (B)))+B
B 2 =φ 423 (B)))+B
wherein B denotes information of the input residual block, phi 1 (. and phi) 2 (. The) shows a convolution operation with 3x3 as the convolution kernel, phi 3 (. and phi) 4 (. The) shows the convolution operation with 5x5 as the convolution kernel, σ 1 (. and σ) 2 (. represents the information activation operation with Relu as the activation function, B 1 And B 2 Shown are the outputs of the residual blocks with 3x3 and 5x5 as convolution kernels, respectively.
The operation flow of a single multi-scale feature extraction Module (MRCB) can be decomposed correspondingly, and the flow before the first conv1x1 in FIG. 2 can be represented by the following several formulas
Y 1 =B 11 (X)
Y 2 =B 21 (X)
Y 3 =F 3 (concat(B 12 (F 1 (concat([Y 1 ,Y 2 ]))),B 22 (F 2 (concat([Y 1 ,Y 2 ]))),
B 13 (X),B 23 (X),X))
Wherein X represents information input to the MRCB module, B 11 (·)、B 12 (. and B) 13 (. denotes) the operation of RU3 in FIG. 2, B 21 (·)、B 22 (. and B) 23 (. denotes the operation of RU5 in FIG. 2, F 1 (·)、F 2 (. and F) 3 (. The) shows a convolution operation with 1x1 as the convolution kernel, Y 1 And Y 2 Shown are the outputs of the first RU3 and first RU5 blocks, Y, respectively, at the top left of FIG. 2 3 Shown is the information output after the first conv1x1 in fig. 2.
The remaining portion of fig. 2 can be represented by 4 equations as follows:
Y 4 =B 14 (Y 3 )
Y 5 =B 24 (Y 3 )
Y 6 =F 6 (concat(B 15 (F 4 (concat([Y 4 ,Y 5 ]))),B 25 (F 5 (concat([Y 4 ,Y 5 ]))),
B 16 (Y 3 ),B 26 (Y 3 ),B 17 (X),B 27 (X),Y 3 ))
Y=F 7 (Y 6 )+X
wherein X represents information input to the MRCB module, B 14 (·)、B 15 (·)、B 16 (. and B) 17 (. to) also shows RU3 operation in FIG. 2, B 24 (·)、B 25 (·)、B 26 (. and B) 27 (. expression) is RU5 operation in FIG. 2, F 4 (·)、F 5 (. and F) 6 (. The) shows a convolution operation with 1x1 as the convolution kernel, F 7 (. The) shows a convolution operation with 3x3 as the convolution kernel, Y 4 And Y 5 Shown are the outputs, Y, of the first RU3 and first RU5 blocks, respectively, of the middle portion of FIG. 2 6 Shown is the output of the second conv1x1 block in fig. 2. Y represents the final output of a single multi-MRCB module.
Step 4, extracting the hierarchical feature maps generated in the MRCB modules, inputting the secondary feature maps and the hierarchical feature maps generated in the MRCB modules into a hierarchical feature processing module (HFCA), as shown in fig. 3, fusing the secondary feature maps and the hierarchical feature maps by using the HFCA module, enhancing important features in the image, and outputting to obtain a tertiary feature map;
the HFCA module is provided with a shuffling module and a mean value and standard deviation based enhanced attention module (COAM); cascading the secondary characteristic diagram and the hierarchical characteristic diagrams generated in all MRCB modules, inputting the cascaded characteristic diagrams into a shuffling module, and performing channel compression through convolution of 1x1 to obtain a fusion characteristic diagram; and inputting the fused feature map into a COAM module to obtain the three-level feature map.
The shuffle module groups the number of channels of the input feature information according to the number of groups required to obtain the number of channels in each group, and the number of output channels at this time may be represented as a result of multiplying the number of channels of each group by the number of groups, then performs Reshape operation on the number of channels at this time to convert the number of channels into a matrix form, and then performs transpose operation on the number of channels, and after transpose is completed, flattens the number of channels and performs packet output according to the number of groups at first, and the structure of the shuffle module is shown in fig. 4.
The COAM module respectively performs mean pooling and standard deviation pooling on input information in a coordinate-like mode according to horizontal and vertical directions, cascades the pooled information, uses shuffling operation to enable information among channels to be fully fused, reduces the dimension of the fused information by using convolution with 1x1 as a convolution kernel, sends the reduced-dimension information to a nonlinear activation layer for activation operation, and then calls a function h-swish to process the activated features to obtain new features; dividing the new features along the horizontal and vertical directions to obtain two independent features, then adopting a convolution layer with 1x1 as a convolution kernel to carry out dimension increasing on the new features, and respectively adopting Sigmoid activation functions to activate the feature information after dimension increasing to obtain weight features in the horizontal and vertical directions; and finally, multiplying the two weighted characteristics by the input characteristic information to obtain the three-level characteristic diagram.
Step 5, sending the three-level feature graph into a convolution layer with 3x3 as a convolution kernel for further feature extraction to generate a four-level feature graph;
and 6, inputting the four-level characteristic diagram into an image reconstruction module, and carrying out amplification reconstruction to obtain a super-resolution capsule endoscope image, wherein the resolution of the super-resolution capsule endoscope image is greater than that of the capsule endoscope image to be reconstructed.
Further, in order to train the network better, so that the network can learn richer features, the data set used in the network training proposed by the present invention is composed of two parts, one is a picture of Kvasir that is not secondarily contaminated and destroyed, and the other is a picture of ETIS-lairb poly DB provided in MICCAI2015 tournament regarding Polyp symptoms in endoscopy. Specifically, there were 1000 images of the capsule endoscope covering different symptoms for training of the images, 20 images for verification, 26 images for testing, each image having a resolution between 1080P and 2K. Low resolution pictures throughout the experiment were generated using Bicubic (BI) and small scale (BD) degradation models. Finally, the peak signal-to-noise ratio (PSNR) and Structural Similarity (SSIM) evaluations were performed on all SR results on the Y channel of the YCbCr space on the output image.
The whole parameters in the network are set as follows, the convolution kernel of the basic block in the network is formed by combining two sizes, namely 3x3 and 5x5, the number of channels is 64, and the step length is 1. In the convolution layer for dimension reduction, the convolution kernels are all 1 × 1. The number of multi-scale residual error feature extraction modules in the network is 8, the number of groups subjected to the shuffling operation before entering the attention mechanism in the hierarchical feature fusion layer is 8, and the number of groups subjected to the shuffling operation in the attention mechanism is 2.
During the training process, using RGB images as input, all training images are randomly rotated 90 °, 180 °, 270 °, and then horizontally flipped for effective data enhancement. Loss function of the network is L 1 Loss function, super-resolution optimization having multiple loss functions, e.g. L 2 、L 1 Perceptual and hostile losses. In the present invention, the document follows EDSR in the field of natural images and uses L 1 The loss function is optimized. Given N LR-HR pairs as the training set, 1000 image pairs are used hereinAs training set, L 1 The loss function can be represented by the following functional form:
Figure BDA0003649449750000091
where θ represents the set of parameters and i represents the ith pair of training image blocks. The initial learning rate in training is 1x10 -4 The optimizer is Adam optimizer (and parameter β is set) 1 =0.9,β 2 =0.999,ε=10 -8 ). And (2) carrying out 1000 times of iteration in each training round, randomly selecting 16 low-resolution patches with the size of 48x48 as input in each training batch, reducing the learning rate by half in each 200 training rounds for 1000 training rounds, and keeping the model with the highest PSNR in a verification set in the training process as a final training result.

Claims (4)

1. The method for reconstructing the super-resolution image of the capsule endoscope based on the multi-scale residual error is characterized by being capable of correspondingly processing an image acquired by the capsule endoscope to obtain the super-resolution image of the capsule endoscope with clearer characteristics such as structure, texture and the like, and specifically comprises the following steps:
step 1, obtaining a capsule endoscope image of a patient to be diagnosed, and obtaining a pre-trained super-resolution reconstruction network, wherein the super-resolution reconstruction network consists of two parts, one part is a feature extraction part consisting of a shallow feature extraction module, a multi-scale residual feature extraction module and a hierarchical feature processing module, and the other part is an image reconstruction module consisting of a convolution layer formed by a sub-pixel layer up-sampling method and a convolution kernel with the size of 3x 3;
step 2, inputting the capsule endoscope image to be reconstructed into a shallow feature extraction module for shallow feature extraction to obtain a primary feature map, wherein the shallow feature extraction module is a convolution layer with a convolution kernel size of 3x 3;
step 3, sequentially passing the primary feature map through n multi-scale residual error feature extraction modules to generate and obtain a secondary feature map;
step 4, extracting the hierarchical feature map generated in each multi-scale residual error feature extraction module, inputting the secondary feature map and the hierarchical feature map generated in each multi-scale residual error feature extraction module into a hierarchical feature processing module, fusing the secondary feature map and the hierarchical feature map by using the hierarchical feature processing module, enhancing important features in the image, and outputting to obtain a tertiary feature map;
step 5, sending the three-level feature graph into a convolution layer formed by taking 3x3 as a convolution kernel, and performing further feature extraction to generate a four-level feature graph;
and 6, inputting the four-level characteristic diagram into an image reconstruction module, and carrying out amplification reconstruction to obtain a super-resolution capsule endoscope image, wherein the resolution of the super-resolution capsule endoscope image is greater than that of the capsule endoscope image to be reconstructed.
2. The super-resolution capsule endoscopic image reconstruction method according to claim 1, wherein: the multi-scale residual error feature extraction module adopts a residual error block as a basic structural unit, each residual error block consists of two convolution layers and a ReLU activation function, and the residual error block can be represented by the following mathematical model:
B 1 =φ 211 (B)))+B
B 2 =φ 423 (B)))+B
wherein B denotes information of the input residual block, phi 1 (. and phi) 2 (. The) shows a convolution operation with 3x3 as the convolution kernel, phi 3 (. and phi) 4 (. The) shows the convolution operation with 5x5 as the convolution kernel, σ 1 (. and σ) 2 (. The) shows an information activation operation with ReLU as the activation function, B 1 And B 2 Shown are the outputs of the residual blocks with 3x3 and 5x5 as convolution kernels, respectively;
the multi-scale residual error feature extraction module consists of 8 parallel branches, and can be specifically represented by the following mathematical model:
Y 1 =B 11 (X)
Y 2 =B 21 (X)
Y 3 =F 3 (concat(B 12 (F 1 (concat([Y 1 ,Y 2 ]))),B 22 (F 2 (concat([Y 1 ,Y 2 ]))),
B 13 (X),B 23 (X),X))
Y 4 =B 14 (Y 3 )
Y 5 =B 24 (Y 3 )
Y 6 =F 6 (concat(B 15 (F 4 (concat([Y 4 ,Y 5 ]))),B 25 (F 5 (concat([Y 4 ,Y 5 ]))),
B 16 (Y 3 ),B 26 (Y 3 ),B 17 (X),B 27 (X),Y 3 ))
Y=F 7 (Y 6 )+X
wherein X represents the information input to the multi-scale residual feature extraction module, B 11 (·)、B 12 (·)、B 13 (·)、B 14 (·)、B 15 (·)、B 16 (. and B) 17 (. 2) shows a convolution kernel residual block, B, of 3x3 21 (·)、B 22 (·)、B 23 (·)、B 24 (·)、B 25 (·)、B 26 (. and B) 27 Shown is a residual block convolved with 5x5, concat (a) shows a feature fusion operation, F 1 (·)、F 2 (·)、F 3 (·)、F 4 (·)、F 5 (. and F) 6 (. The) shows a convolution operation with 1x1 as the convolution kernel, F 7 Shown is the convolution operation with 3x3 as the convolution kernel, and Y is the final output of a single multi-scale feature extraction module.
3. The super-resolution capsule endoscopic image reconstruction method according to claim 1, wherein: the hierarchical feature processing module is provided with a shuffling module and an enhanced attention module based on a mean value and a standard deviation; the second-level feature map and the hierarchical feature maps generated in the multi-scale residual error feature extraction modules are cascaded and then input into a shuffling module, and then channel compression is carried out through convolution of 1x1 to obtain a fusion feature map; and inputting the fused feature map into an enhanced attention module based on the mean value and the standard deviation to obtain the three-level feature map.
4. The super-resolution capsule endoscopic image reconstruction method according to claim 4, wherein: the enhanced attention module based on the mean value and the standard deviation respectively performs mean value pooling and standard deviation pooling on input information in a coordinate-like mode according to the horizontal direction and the vertical direction, after pooling is performed, information is cascaded and then is subjected to shuffling operation, information among channels is fully fused, dimension reduction is performed on the fused information by convolution with 1x1 as a convolution kernel, the dimension reduced information is sent to a nonlinear activation layer for activation operation, and then a function h-swish is called to process the activated features to obtain new features; dividing the new features along the horizontal and vertical directions to obtain two independent features, then adopting a convolution layer with 1x1 as a convolution kernel to carry out dimension increasing on the new features, and respectively adopting Sigmoid activation functions to activate the feature information after dimension increasing to obtain weight features in the horizontal and vertical directions; and finally, multiplying the two weighted characteristics by the input characteristic information to obtain the three-level characteristic diagram.
CN202210539075.1A 2022-05-18 2022-05-18 Capsule endoscope image super-resolution reconstruction method based on multi-scale residual errors Pending CN114842104A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202210539075.1A CN114842104A (en) 2022-05-18 2022-05-18 Capsule endoscope image super-resolution reconstruction method based on multi-scale residual errors

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202210539075.1A CN114842104A (en) 2022-05-18 2022-05-18 Capsule endoscope image super-resolution reconstruction method based on multi-scale residual errors

Publications (1)

Publication Number Publication Date
CN114842104A true CN114842104A (en) 2022-08-02

Family

ID=82569436

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202210539075.1A Pending CN114842104A (en) 2022-05-18 2022-05-18 Capsule endoscope image super-resolution reconstruction method based on multi-scale residual errors

Country Status (1)

Country Link
CN (1) CN114842104A (en)

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN116523759A (en) * 2023-07-04 2023-08-01 江西财经大学 Image super-resolution reconstruction method and system based on frequency decomposition and restarting mechanism

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN116523759A (en) * 2023-07-04 2023-08-01 江西财经大学 Image super-resolution reconstruction method and system based on frequency decomposition and restarting mechanism
CN116523759B (en) * 2023-07-04 2023-09-05 江西财经大学 Image super-resolution reconstruction method and system based on frequency decomposition and restarting mechanism

Similar Documents

Publication Publication Date Title
CN110570353B (en) Super-resolution reconstruction method for generating single image of countermeasure network by dense connection
CN111145170B (en) Medical image segmentation method based on deep learning
CN107492071A (en) Medical image processing method and equipment
CN110232653A (en) The quick light-duty intensive residual error network of super-resolution rebuilding
An et al. TR-MISR: Multiimage super-resolution based on feature fusion with transformers
WO2020211530A1 (en) Model training method and apparatus for detection on fundus image, method and apparatus for detection on fundus image, computer device, and medium
CN110570351A (en) Image super-resolution reconstruction method based on convolution sparse coding
CN114187296B (en) Capsule endoscope image focus segmentation method, server and system
CN116664397B (en) TransSR-Net structured image super-resolution reconstruction method
CN116309648A (en) Medical image segmentation model construction method based on multi-attention fusion
CN111260639A (en) Multi-view information-collaborative breast benign and malignant tumor classification method
Jiang et al. CT image super resolution based on improved SRGAN
CN111696042B (en) Image super-resolution reconstruction method based on sample learning
Zhou et al. High dynamic range imaging with context-aware transformer
CN115375711A (en) Image segmentation method of global context attention network based on multi-scale fusion
Qiu et al. Residual dense attention networks for COVID-19 computed tomography images super-resolution
CN114842104A (en) Capsule endoscope image super-resolution reconstruction method based on multi-scale residual errors
CN113781489B (en) Polyp image semantic segmentation method and device
CN114187181A (en) Double-path lung CT image super-resolution method based on residual information refining
Yang et al. Lesion classification of wireless capsule endoscopy images
CN116797541A (en) Transformer-based lung CT image super-resolution reconstruction method
CN116468887A (en) Method for segmenting colon polyp with universality
CN112598581B (en) Training method and image generation method of RDN super-resolution network
CN114842029A (en) Convolutional neural network polyp segmentation method fusing channel and spatial attention
Tian et al. Retinal fundus image superresolution generated by optical coherence tomography based on a realistic mixed attention GAN

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination