CN116310329A - Skin lesion image segmentation method based on lightweight multi-scale UNet - Google Patents
Skin lesion image segmentation method based on lightweight multi-scale UNet Download PDFInfo
- Publication number
- CN116310329A CN116310329A CN202310194557.2A CN202310194557A CN116310329A CN 116310329 A CN116310329 A CN 116310329A CN 202310194557 A CN202310194557 A CN 202310194557A CN 116310329 A CN116310329 A CN 116310329A
- Authority
- CN
- China
- Prior art keywords
- convolution
- module
- scale
- depth
- unet
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Pending
Links
- 206010040882 skin lesion Diseases 0.000 title claims abstract description 37
- 231100000444 skin lesion Toxicity 0.000 title claims abstract description 37
- 238000000034 method Methods 0.000 title claims abstract description 31
- 238000003709 image segmentation Methods 0.000 title claims abstract description 16
- 238000011176 pooling Methods 0.000 claims abstract description 16
- 230000011218 segmentation Effects 0.000 claims abstract description 11
- 238000007781 pre-processing Methods 0.000 claims abstract description 8
- 238000012549 training Methods 0.000 claims abstract description 8
- 230000004913 activation Effects 0.000 claims description 18
- 238000010606 normalization Methods 0.000 claims description 18
- 230000004927 fusion Effects 0.000 claims description 14
- 230000010339 dilation Effects 0.000 claims description 4
- 238000012360 testing method Methods 0.000 claims description 3
- 238000011426 transformation method Methods 0.000 claims description 3
- 238000012795 verification Methods 0.000 claims description 3
- 230000009191 jumping Effects 0.000 abstract 1
- 230000006870 function Effects 0.000 description 17
- 238000004364 calculation method Methods 0.000 description 3
- 238000013527 convolutional neural network Methods 0.000 description 3
- 238000010586 diagram Methods 0.000 description 3
- 238000003745 diagnosis Methods 0.000 description 2
- 230000000694 effects Effects 0.000 description 2
- 239000011796 hollow space material Substances 0.000 description 2
- 230000008569 process Effects 0.000 description 2
- 238000013528 artificial neural network Methods 0.000 description 1
- 238000013135 deep learning Methods 0.000 description 1
- 230000007547 defect Effects 0.000 description 1
- 238000013461 design Methods 0.000 description 1
- 238000001514 detection method Methods 0.000 description 1
- 238000003708 edge detection Methods 0.000 description 1
- 238000011478 gradient descent method Methods 0.000 description 1
- 230000003902 lesion Effects 0.000 description 1
- 238000010801 machine learning Methods 0.000 description 1
- 238000012545 processing Methods 0.000 description 1
- 238000011160 research Methods 0.000 description 1
- 238000006467 substitution reaction Methods 0.000 description 1
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V10/00—Arrangements for image or video recognition or understanding
- G06V10/20—Image preprocessing
- G06V10/26—Segmentation of patterns in the image field; Cutting or merging of image elements to establish the pattern region, e.g. clustering-based techniques; Detection of occlusion
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/08—Learning methods
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
- G06T7/0002—Inspection of images, e.g. flaw detection
- G06T7/0012—Biomedical image inspection
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V10/00—Arrangements for image or video recognition or understanding
- G06V10/70—Arrangements for image or video recognition or understanding using pattern recognition or machine learning
- G06V10/82—Arrangements for image or video recognition or understanding using pattern recognition or machine learning using neural networks
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/20—Special algorithmic details
- G06T2207/20081—Training; Learning
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/20—Special algorithmic details
- G06T2207/20084—Artificial neural networks [ANN]
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/30—Subject of image; Context of image processing
- G06T2207/30004—Biomedical image processing
- G06T2207/30088—Skin; Dermal
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V2201/00—Indexing scheme relating to image or video recognition or understanding
- G06V2201/03—Recognition of patterns in medical or anatomical images
Landscapes
- Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- Physics & Mathematics (AREA)
- General Physics & Mathematics (AREA)
- General Health & Medical Sciences (AREA)
- Evolutionary Computation (AREA)
- Health & Medical Sciences (AREA)
- Multimedia (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Computing Systems (AREA)
- Artificial Intelligence (AREA)
- Software Systems (AREA)
- Medical Informatics (AREA)
- Nuclear Medicine, Radiotherapy & Molecular Imaging (AREA)
- Computational Linguistics (AREA)
- Radiology & Medical Imaging (AREA)
- Quality & Reliability (AREA)
- Life Sciences & Earth Sciences (AREA)
- Biomedical Technology (AREA)
- Biophysics (AREA)
- Databases & Information Systems (AREA)
- Data Mining & Analysis (AREA)
- Molecular Biology (AREA)
- General Engineering & Computer Science (AREA)
- Mathematical Physics (AREA)
- Image Processing (AREA)
- Image Analysis (AREA)
- Measuring And Recording Apparatus For Diagnosis (AREA)
Abstract
Description
技术领域technical field
本发明属于医学图像处理技术领域,具体为一种基于轻量多尺度UNet的皮肤病变图像分割方法。The invention belongs to the technical field of medical image processing, in particular to a skin lesion image segmentation method based on lightweight multi-scale UNet.
背景技术Background technique
皮肤病变图像分割是皮肤诊断研究的一个重要部分。传统的方法是在皮肤镜下制作高分辨率的受损皮肤图像,然后由专业医生进行诊断。然而,由于皮损区域的大小和形状各不相同,人工判断费时费力,且含有主观成分,增加了诊断的难度。早期的计算机辅助医学图像分割方法通常依赖于边缘检测、模板匹配技术和传统的机器学习技术。这些方法在一定程度上取得了良好的效果,但它们往往需要对原始图像进行复杂的预处理,这就需要有经验的工程师来设计特征提取器并选择合适的分类器进行分类。所以,这种方法的泛化能力较弱,难以实现复杂的多分类任务。Image segmentation of skin lesions is an important part of skin diagnosis research. The traditional method is to make high-resolution images of damaged skin under a dermatoscope, which is then diagnosed by professional doctors. However, because the size and shape of the lesion area vary, manual judgment is time-consuming and laborious, and contains subjective components, which increases the difficulty of diagnosis. Early computer-aided medical image segmentation methods usually relied on edge detection, template matching techniques, and traditional machine learning techniques. These methods have achieved good results to a certain extent, but they often require complex preprocessing of the original image, which requires experienced engineers to design feature extractors and select appropriate classifiers for classification. Therefore, the generalization ability of this method is weak, and it is difficult to achieve complex multi-classification tasks.
随着大数据时代的到来和计算机硬件的巨大进步,深度学习技术,特别是卷积神经网络,在图像分类和检测等许多任务中取得了比传统方法更好的效果。近年来,许多研究人员都专注于开发高精度的分割方法。其中,UNet是一个基于全卷积网络的语义分割网络,适用于医学图像分割,该网络修改和扩展了全卷积神经网络结构,在使用少量数据进行训练时获得了准确的分割结果,因此广受关注。此后许多神经网络在UNet基础上发展而来,如UNet++、LinkNet、3DUNet、ResUNet等,虽然改进后的模型在分割效果上有所提升,但是模型往往参数过多、运算量大,不利于实现图像的快速分割。With the advent of the era of big data and great advances in computer hardware, deep learning techniques, especially convolutional neural networks, have achieved better results than traditional methods in many tasks such as image classification and detection. In recent years, many researchers have focused on developing high-precision segmentation methods. Among them, UNet is a semantic segmentation network based on a fully convolutional network, which is suitable for medical image segmentation. attention. Since then, many neural networks have been developed on the basis of UNet, such as UNet++, LinkNet, 3DUNet, ResUNet, etc. Although the improved model has improved the segmentation effect, the model often has too many parameters and a large amount of calculation, which is not conducive to the realization of image quality. quick split.
发明内容Contents of the invention
为了解决现有技术中的上述技术缺陷,本发明提出了一种基于轻量多尺度UNet的皮肤病变图像分割方法。In order to solve the above-mentioned technical defects in the prior art, the present invention proposes a skin lesion image segmentation method based on lightweight multi-scale UNet.
实现本发明目的的技术方案为:一种基于轻量多尺度UNet的皮肤病变图像分割方法,包括以下步骤:The technical solution for realizing the object of the present invention is: a skin lesion image segmentation method based on lightweight multi-scale UNet, comprising the following steps:
步骤1、获取皮肤病变图像并进行预处理;Step 1. Obtain the skin lesion image and perform preprocessing;
步骤2、建立轻量多尺度UNet网络结构即LMUNet:以原始的UNet模型为基础,使用多尺度倒置残差模块替代UNet编码路径中原有的卷积模块,并在编码路径和解码路径之间加入非对称空洞空间金字塔池化模块,同时减少每层的通道数,并将UNet原有的跳跃连接修改为通道相加;Step 2. Establish a lightweight multi-scale UNet network structure, namely LMUNet: based on the original UNet model, use a multi-scale inverted residual module to replace the original convolution module in the UNet encoding path, and add Asymmetric empty space pyramid pooling module, while reducing the number of channels in each layer, and modifying UNet's original skip connection to channel addition;
步骤3、利用预处理的皮肤病变图像对LMUNet网络进行训练;Step 3, utilize the preprocessed skin lesion image to train the LMUNet network;
步骤4、将待分割的皮肤病变图像输入训练好的LMUNet网络获取分割结果。Step 4. Input the image of the skin lesion to be segmented into the trained LMUNet network to obtain the segmentation result.
优选地,步骤1对皮肤病变图像进行预处理具体包括:Preferably, the preprocessing of the skin lesion image in step 1 specifically includes:
统一图像尺寸、数据增强和数据集划分,其中,统一图像尺寸是将通过缩放或裁剪的方式将获取的皮肤病变图像尺寸进行统一;数据增强采用旋转和翻转图像几何变换方法以及直方图均衡中的一种或多种;数据集划分是指将数据增强操作后的图像划分为训练集、验证集和测试集。Unified image size, data enhancement and data set division, wherein the unified image size is to unify the acquired skin lesion image size by scaling or cropping; data enhancement adopts rotation and flip image geometric transformation method and histogram equalization One or more; Data set division refers to dividing the image after the data enhancement operation into a training set, a verification set and a test set.
优选地,步骤2所述的建立轻量多尺度UNet网络结构即LMUNet,具体如下:Preferably, the establishment of a lightweight multi-scale UNet network structure, namely LMUNet, as described in step 2, is as follows:
LMUNet网络结构包括编码路径、解码路径和多尺度信息融合模块;LMUNet network structure includes encoding path, decoding path and multi-scale information fusion module;
编码路径由多尺度倒置残差模块和2×2最大化池化模块组成,负责对输入图像进行特征提取并降低图像尺寸,减少冗余的参数量;解码路径由反卷积层和标准卷积层组成,负责恢复特征图信息;多尺度信息融合模块位于编码路径和解码路径之间,用于融合不同尺度下的特征,丰富上下文信息。The encoding path consists of a multi-scale inverted residual module and a 2×2 maximization pooling module, which is responsible for extracting features from the input image and reducing the size of the image, reducing the amount of redundant parameters; the decoding path consists of a deconvolution layer and a standard convolution Layers are responsible for restoring feature map information; the multi-scale information fusion module is located between the encoding path and the decoding path, and is used to fuse features at different scales and enrich context information.
优选地,所述编码路径包括多尺度倒置残差模块和2×2最大化池化模块;其中,多尺度倒置残差模块负责提取在多尺度下图像的特征信息,而2×2最大化池化模块可以对图像进行压缩,提取图像的主要特征并且减少网络冗余的参数量。Preferably, the encoding path includes a multi-scale inversion residual module and a 2×2 maximization pooling module; wherein, the multi-scale inversion residual module is responsible for extracting feature information of images at multiple scales, and the 2×2 maximization pooling module The module can compress the image, extract the main features of the image and reduce the amount of network redundant parameters.
所述多尺度倒置残差模块包括第一3×3深度卷积模块、第二3×3深度卷积模块、3×3深度空洞卷积模块以及1×1点卷积模块,第一3×3深度卷积模块包括3×3深度卷积、批归一化层和ReLU6激活函数,输入信息一次经过3×3深度卷积、批归一化层和ReLU6激活函数,得到的结果与输入信息进行通道拼接,将拼接后的信息分别输入第二3×3深度卷积模块、3×3深度空洞卷积模块,所述第二3×3深度卷积模块包括3×3深度卷积、批归一化层和ReLU6激活函数;所述3×3深度空洞卷积模块包括空洞率为2的3×3深度空洞卷积、批归一化层和ReLU6激活函数;The multi-scale inversion residual module includes a first 3×3 depth convolution module, a second 3×3 depth convolution module, a 3×3 depth hole convolution module and a 1×1 point convolution module, the first 3× The 3-depth convolution module includes 3×3 depth convolution, batch normalization layer and ReLU6 activation function. The input information passes through 3×3 depth convolution, batch normalization layer and ReLU6 activation function once, and the obtained result is consistent with the input information Perform channel splicing, and input the spliced information into the second 3×3 depth convolution module and the 3×3 depth hole convolution module, the second 3×3 depth convolution module includes 3×3 depth convolution, batch A normalization layer and a ReLU6 activation function; the 3×3 depth hole convolution module includes a 3×3 depth hole convolution with a hole rate of 2, a batch normalization layer and a ReLU6 activation function;
将第二3×3深度卷积模块、3×3深度空洞卷积模块得到的结果进行相加后输入1×1点卷积模块,所述1×1点卷积模块包括3×3深度卷积、批归一化层和ReLU6激活函数,经3×3深度卷积、批归一化层和ReLU6激活函数,将得到的结果与开始时的输入信息进行通道相加,获得不同尺度下的特征信息。The results obtained by the second 3×3 depth convolution module and the 3×3 depth hole convolution module are added and then input into the 1×1 point convolution module. The 1×1 point convolution module includes a 3×3 depth convolution module Product, batch normalization layer and ReLU6 activation function, after 3×3 depth convolution, batch normalization layer and ReLU6 activation function, the result obtained is added to the input information at the beginning to obtain different scales characteristic information.
优选地,所述多尺度信息融合模块由非对称空洞空间金字塔池化模块组成,包括1×1点卷积和3个并联的分支,3个并联的分支分别为:Preferably, the multi-scale information fusion module is composed of an asymmetric hollow space pyramid pooling module, including 1×1 point convolution and 3 parallel branches, and the 3 parallel branches are:
分支一:由3×1深度非对称卷积和1×3深度非对称卷积组成;Branch 1: It consists of 3×1 depth asymmetric convolution and 1×3 depth asymmetric convolution;
分支二:由空洞率为2的3×1深度非对称空洞卷积和1×3深度非对称空洞卷积组成;Branch 2: It consists of a 3×1 depth asymmetric atrous convolution with a dilation rate of 2 and a 1×3 depth asymmetric atrous convolution;
分支三:由空洞率为3的3×1深度非对称空洞卷积和1×3深度非对称空洞卷积组成;Branch 3: It consists of a 3×1 depth asymmetric hole convolution and a 1×3 depth asymmetric hole convolution with a hole rate of 3;
输入信息经过1×1点卷积、批归一化层和ReLU6激活函数,将通道数降为原来的一半,经过3个并联的分支,将得到的结果与输入分支的信息进行通道拼接,最后经过1×1点卷积进行特征信息融合,获得融合后的多尺度下的特征信息;The input information goes through 1×1 point convolution, batch normalization layer and ReLU6 activation function, the number of channels is reduced to half of the original, and after three parallel branches, the obtained result and the information of the input branch are channel spliced, and finally After 1×1 point convolution for feature information fusion, the fused multi-scale feature information is obtained;
优选地,步骤3利用预处理的皮肤病变图像对LMUNet网络进行训练,使用的损失函数为交叉熵函数,函数形式为:Preferably, step 3 utilizes the preprocessed skin lesion image to train the LMUNet network, and the loss function used is a cross-entropy function, and the function form is:
式中H(P,Q代表交叉熵,P(xi)代表真实概率分布,Q(xi)代表预测概率分布。In the formula, H(P, Q represents cross entropy, P( xi ) represents the real probability distribution, and Q( xi ) represents the predicted probability distribution.
本发明与现有技术相比,其显著优点为:本发明基于UNet模型,通过引入多尺度倒置残差模块、非对称空洞空间金字塔池化模块和减少通道数量来对UNet模型进行改进。实验结果表明,相较于其他分割模型,本发明提出的LMUNet模型性能优越且十分轻量,只需很少的运算量便可以实现更优的分割效果。Compared with the prior art, the present invention has the following significant advantages: the present invention is based on the UNet model, and improves the UNet model by introducing a multi-scale inverted residual module, an asymmetric empty space pyramid pooling module, and reducing the number of channels. Experimental results show that compared with other segmentation models, the LMUNet model proposed by the present invention has superior performance and is very lightweight, and can achieve a better segmentation effect with only a small amount of calculation.
本发明的其他特征和优点将在随后的说明书中阐述,并且,部分的从说明书中变得显而易见,或者通过实施本发明而了解。本发明的目的和其他优点可通过在所写的说明书、权利要求书、以及附图中所特别指出的结构来实现和获得。Additional features and advantages of the invention will be set forth in the description which follows, and in part will be apparent from the description, or may be learned by practice of the invention. The objectives and other advantages of the invention may be realized and attained by the structure particularly pointed out in the written description and claims hereof as well as the appended drawings.
附图说明Description of drawings
附图仅用于示出具体实施例的目的,而并不认为是对本发明的限制,在整个附图中,相同的参考符号表示相同的部件。The drawings are for the purpose of illustrating specific embodiments only and are not to be considered as limitations of the invention, and like reference numerals refer to like parts throughout the drawings.
图1是本发明中LMUNet网络的整体结构图。Fig. 1 is the overall structural diagram of LMUNet network among the present invention.
图2是本发明中多尺度倒置残差模块的结构图。Fig. 2 is a structural diagram of the multi-scale inverted residual module in the present invention.
图3是本发明中多尺度信息融合模块模块结构图。Fig. 3 is a block diagram of the multi-scale information fusion module in the present invention.
图4为本发明所述方法的流程示意图。Fig. 4 is a schematic flow chart of the method of the present invention.
具体实施方式Detailed ways
容易理解,依据本发明的技术方案,在不变更本发明的实质精神的情况下,本领域的一般技术人员可以想象出本发明的多种实施方式。因此,以下具体实施方式和附图仅是对本发明的技术方案的示例性说明,而不应当视为本发明的全部或者视为对本发明技术方案的限制或限定。相反,提供这些实施例的目的是为了使本领域的技术人员更透彻地理解本发明。下面结合附图来具体描述本发明的优选实施例,其中,附图构成本申请一部分,并与本发明的实施例一起用于阐释本发明的创新构思。It is easy to understand that, according to the technical solution of the present invention, those skilled in the art can imagine various implementations of the present invention without changing the essence and spirit of the present invention. Therefore, the following specific embodiments and drawings are only exemplary descriptions of the technical solution of the present invention, and should not be regarded as the entirety of the present invention or as a limitation or limitation on the technical solution of the present invention. Rather, these embodiments are provided to enable those skilled in the art to more thoroughly understand the present invention. Preferred embodiments of the present invention will be specifically described below in conjunction with the accompanying drawings, wherein the accompanying drawings constitute a part of the application and are used together with the embodiments of the present invention to explain the innovative concept of the present invention.
本发明构思为,如图1~图4所示,一种基于轻量多尺度UNet的皮肤病变图像分割方法,所述方法的具体步骤如下:The concept of the present invention is, as shown in Figures 1 to 4, a skin lesion image segmentation method based on lightweight multi-scale UNet, the specific steps of the method are as follows:
步骤1、获取皮肤病变图像并进行预处理;Step 1. Obtain the skin lesion image and perform preprocessing;
步骤2、建立轻量多尺度UNet网络结构即LMUNet:以原始的UNet模型为基础,使用多尺度倒置残差模块替代UNet编码路径中原有的卷积模块,并在编码路径和解码路径之间加入非对称空洞空间金字塔池化模块,同时减少每层的通道数,并将UNet原有的跳跃连接修改为通道相加;Step 2. Establish a lightweight multi-scale UNet network structure, namely LMUNet: based on the original UNet model, use a multi-scale inverted residual module to replace the original convolution module in the UNet encoding path, and add Asymmetric empty space pyramid pooling module, while reducing the number of channels in each layer, and modifying UNet's original skip connection to channel addition;
步骤3、利用预处理的皮肤病变图像对LMUNet网络进行训练;Step 3, utilize the preprocessed skin lesion image to train the LMUNet network;
步骤4、将待分割的皮肤病变图像输入训练好的LMUNet网络获取分割结果。Step 4. Input the image of the skin lesion to be segmented into the trained LMUNet network to obtain the segmentation result.
进一步的实施例中,获取皮肤病变图像并进行预处理的具体方法为:统一图像尺寸、数据增强和数据集划分,其中统一图像尺寸是将通过缩放或裁剪的方式将获取的皮肤病变图像尺寸进行统一;数据增强采用旋转和翻转图像等几何变换方法以及直方图均衡等图像增强方法中的一种或多种;数据集划分将数据增强操作后的图像划分为训练集、验证集和测试集。In a further embodiment, the specific method of acquiring and preprocessing the skin lesion image is: unifying the image size, data enhancement, and data set division, wherein the unifying image size is to scale or crop the acquired skin lesion image size. Unification; data enhancement adopts one or more of geometric transformation methods such as rotating and flipping images, and image enhancement methods such as histogram equalization; data set division divides the image after data enhancement operation into training set, verification set and test set.
进一步的实施例中,如图1所示,轻量多尺度UNet网络结构即LMUNet,包括编码路径、解码路径和多尺度信息融合模块。编码路径由多尺度倒置残差模块和2×2最大化池化模块组成,其中最大池化模块一共进行4次下采样;解码路径由3×3标准卷积块和2×2反卷积组成,其中反卷积层一共进行4次上采样;多尺度信息融合模块位于编码路径和解码路径之间,用于融合不同尺度下的特征,丰富上下文信息。In a further embodiment, as shown in FIG. 1 , the lightweight multi-scale UNet network structure, namely LMUNet, includes an encoding path, a decoding path, and a multi-scale information fusion module. The encoding path consists of a multi-scale inverted residual module and a 2×2 max pooling module, where the max pooling module performs a total of 4 downsampling; the decoding path consists of a 3×3 standard convolution block and a 2×2 deconvolution , where the deconvolution layer performs a total of 4 times of upsampling; the multi-scale information fusion module is located between the encoding path and the decoding path, and is used to fuse features at different scales and enrich context information.
具体地,所述多尺度倒置残差模块结构如图2所示,具体为多尺度倒置残差模块结构由3×3深度卷积、3×3深度空洞卷积、1×1点卷积、批归一化层和ReLU6激活函数组成,该模块的输入信息首先经过3×3深度卷积、批归一化层和ReLU6激活函数,得到的结果与输入信息进行通道拼接,然后将经过2个并联的分支:Specifically, the multi-scale inverted residual module structure is shown in Figure 2, specifically, the multi-scale inverted residual module structure consists of 3×3 depth convolution, 3×3 depth hole convolution, 1×1 point convolution, Batch normalization layer and ReLU6 activation function. The input information of this module first passes through 3×3 depth convolution, batch normalization layer and ReLU6 activation function. Parallel branches:
分支一:由3×3深度卷积组成;Branch 1: It consists of 3×3 depth convolution;
分支二:由空洞率为2的3×3深度空洞卷积组成;Branch 2: Consists of 3×3 deep hole convolution with a hole rate of 2;
将2个分支得到的结果进行相加然后经过3×3深度卷积、批归一化层和ReLU6激活函数,最后将得到的结果与开始时的输入信息进行通道相加。利用该模块可以获得不同尺度下的特征信息。Add the results obtained by the two branches and then go through 3×3 depth convolution, batch normalization layer and ReLU6 activation function, and finally add the obtained results to the input information at the beginning. Using this module, feature information at different scales can be obtained.
进一步的,所述多尺度信息融合模块结构如图3所示,具体为多尺度信息融合模块由非对称空洞空间金字塔池化模块组成,包括3个并联的分支:Further, the structure of the multi-scale information fusion module is shown in Figure 3, specifically, the multi-scale information fusion module is composed of an asymmetric hollow space pyramid pooling module, including three parallel branches:
分支一:由3×1深度非对称卷积和1×3深度非对称卷积组成;Branch 1: It consists of 3×1 depth asymmetric convolution and 1×3 depth asymmetric convolution;
分支二:由空洞率为2的3×1深度非对称空洞卷积和1×3深度非对称空洞卷积组成;Branch 2: It consists of a 3×1 depth asymmetric atrous convolution with a dilation rate of 2 and a 1×3 depth asymmetric atrous convolution;
分支三:由空洞率为3的3×1深度非对称空洞卷积和1×3深度非对称空洞卷积组成;Branch 3: It consists of a 3×1 depth asymmetric hole convolution and a 1×3 depth asymmetric hole convolution with a hole rate of 3;
输入信息首先经过1×1点卷积、批归一化层和ReLU6激活函数,将通道数降为原来的一半,然后经过3个并联的分支,将3个分支得到的结果与输入分支的信息进行通道拼接,最后经过1×1点卷积进行特征信息融合。利用该模块可以融合多尺度下的特征信息,增强模型能力。The input information first passes through 1×1 point convolution, batch normalization layer and ReLU6 activation function, reducing the number of channels to half of the original, and then passes through 3 parallel branches, and combines the results obtained by the 3 branches with the information of the input branch Carry out channel splicing, and finally perform feature information fusion through 1×1 point convolution. This module can be used to fuse feature information at multiple scales to enhance model capabilities.
训练过程中,网络模型都在Pytorch平台上实现,版本为Pytorch1.11.0。训练过程中的超参数如下:学习率设置为0.01,训练迭代次数设置为100,每次训练的样本数为4。优化器使用随机梯度下降法更新网络参数,加快收敛速度。使用的损失函数是交叉熵函数,函数形式为:During the training process, the network models are implemented on the Pytorch platform, and the version is Pytorch1.11.0. The hyperparameters during the training process are as follows: the learning rate is set to 0.01, the number of training iterations is set to 100, and the number of samples for each training is 4. The optimizer uses the stochastic gradient descent method to update the network parameters to speed up the convergence. The loss function used is the cross entropy function, and the function form is:
式中H(P,Q代表交叉熵,P(xi)代表真实概率分布,Q(xi)代表预测概率分布。In the formula, H(P, Q represents cross entropy, P( xi ) represents the real probability distribution, and Q( xi ) represents the predicted probability distribution.
表1本发明方法和部分卷积神经网络方法的分割结果对比Table 1 The comparison of segmentation results between the method of the present invention and the partial convolutional neural network method
由表1的分割结果可知,相比于表1中其他卷积神经网络,本发明性能优越且十分轻量,减少了参数量和计算量,并提高了皮肤病变图像分割的准确率和速度。From the segmentation results in Table 1, it can be seen that compared with other convolutional neural networks in Table 1, the present invention has superior performance and is very lightweight, reduces the amount of parameters and calculations, and improves the accuracy and speed of skin lesion image segmentation.
以上所述,仅为本发明较佳的具体实施方式,但本发明的保护范围并不局限于此,The above is only a preferred embodiment of the present invention, but the protection scope of the present invention is not limited thereto.
任何熟悉本技术领域的技术人员在本发明揭露的技术范围内,可轻易想到的变化或替换,都应涵盖在本发明的保护范围之内。Any changes or substitutions that can be easily conceived by any person skilled in the art within the technical scope disclosed in the present invention shall fall within the protection scope of the present invention.
应当理解,为了精简本发明并帮助本领域的技术人员理解本发明的各个方面,在上面对本发明的示例性实施例的描述中,本发明的各个特征有时在单个实施例中进行描述,或者参照单个图进行描述。但是,不应将本发明解释成示例性实施例中包括的特征均为本专利权利要求的必要技术特征。It should be appreciated that in the foregoing description of exemplary embodiments of the invention, in order to streamline the present disclosure and to assist those skilled in the art in understanding its various aspects, various features of the invention are sometimes described in the context of a single embodiment, or with reference to A single graph is described. However, the present invention should not be interpreted that the features included in the exemplary embodiments are all essential technical features of the patent claims.
应当理解,可以对本发明的一个实施例的设备中包括的模块、单元、组件等进行自适应性地改变以把它们设置在与该实施例不同的设备中。可以把实施例的设备包括的不同模块、单元或组件组合成一个模块、单元或组件,也可以把它们分成多个子模块、子单元或子组件。It should be understood that the modules, units, components, etc. included in the device of one embodiment of the present invention can be adaptively changed so as to be arranged in a device different from that of the embodiment. Different modules, units or components included in the device of the embodiment can be combined into one module, unit or component, or they can be divided into multiple sub-modules, sub-units or sub-components.
Claims (6)
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202310194557.2A CN116310329A (en) | 2023-03-03 | 2023-03-03 | Skin lesion image segmentation method based on lightweight multi-scale UNet |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202310194557.2A CN116310329A (en) | 2023-03-03 | 2023-03-03 | Skin lesion image segmentation method based on lightweight multi-scale UNet |
Publications (1)
Publication Number | Publication Date |
---|---|
CN116310329A true CN116310329A (en) | 2023-06-23 |
Family
ID=86789919
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN202310194557.2A Pending CN116310329A (en) | 2023-03-03 | 2023-03-03 | Skin lesion image segmentation method based on lightweight multi-scale UNet |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN116310329A (en) |
Cited By (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN116612122A (en) * | 2023-07-20 | 2023-08-18 | 湖南快乐阳光互动娱乐传媒有限公司 | Image significance region detection method and device, storage medium and electronic equipment |
CN117351202A (en) * | 2023-09-28 | 2024-01-05 | 河北翔拓航空科技有限公司 | Image segmentation method for lightweight multi-scale UNet network |
CN118608545A (en) * | 2024-07-08 | 2024-09-06 | 济南大学 | A dual-path skin lesion image segmentation method based on lightweight multi-scale fusion |
CN118967704A (en) * | 2024-08-06 | 2024-11-15 | 电子科技大学 | Deep learning-based semantic segmentation method for sentinel lymph node ultrasound images |
-
2023
- 2023-03-03 CN CN202310194557.2A patent/CN116310329A/en active Pending
Cited By (7)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN116612122A (en) * | 2023-07-20 | 2023-08-18 | 湖南快乐阳光互动娱乐传媒有限公司 | Image significance region detection method and device, storage medium and electronic equipment |
CN116612122B (en) * | 2023-07-20 | 2023-10-10 | 湖南快乐阳光互动娱乐传媒有限公司 | Image significance region detection method and device, storage medium and electronic equipment |
CN117351202A (en) * | 2023-09-28 | 2024-01-05 | 河北翔拓航空科技有限公司 | Image segmentation method for lightweight multi-scale UNet network |
CN117351202B (en) * | 2023-09-28 | 2024-04-30 | 河北翔拓航空科技有限公司 | Image segmentation method based on lightweight multi-scale UNet network |
CN118608545A (en) * | 2024-07-08 | 2024-09-06 | 济南大学 | A dual-path skin lesion image segmentation method based on lightweight multi-scale fusion |
CN118608545B (en) * | 2024-07-08 | 2025-01-24 | 济南大学 | A dual-path skin lesion image segmentation method based on lightweight multi-scale fusion |
CN118967704A (en) * | 2024-08-06 | 2024-11-15 | 电子科技大学 | Deep learning-based semantic segmentation method for sentinel lymph node ultrasound images |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN111369563B (en) | A Semantic Segmentation Method Based on Pyramid Atrous Convolutional Network | |
CN115205300B (en) | Fundus blood vessel image segmentation method and system based on cavity convolution and semantic fusion | |
CN111242288B (en) | A Multi-scale Parallel Deep Neural Network Model Construction Method for Lesion Image Segmentation | |
CN108062753B (en) | Unsupervised domain self-adaptive brain tumor semantic segmentation method based on deep counterstudy | |
CN107016681B (en) | Brain MRI tumor segmentation method based on full convolution network | |
CN116310329A (en) | Skin lesion image segmentation method based on lightweight multi-scale UNet | |
CN109410219A (en) | A kind of image partition method, device and computer readable storage medium based on pyramid fusion study | |
CN107506761A (en) | Brain image dividing method and system based on notable inquiry learning convolutional neural networks | |
CN114004811B (en) | Image segmentation method and system based on multi-scale residual codec network | |
CN116309648A (en) | A medical image segmentation model construction method based on multi-attention fusion | |
CN105701508A (en) | Global-local optimization model based on multistage convolution neural network and significant detection algorithm | |
CN113192076B (en) | MRI Brain Tumor Image Segmentation Using Combined Classification Prediction and Multiscale Feature Extraction | |
CN111860528B (en) | Image segmentation model based on improved U-Net network and training method | |
CN113205524A (en) | Blood vessel image segmentation method, device and equipment based on U-Net | |
CN112001928A (en) | Retinal blood vessel segmentation method and system | |
CN112270366A (en) | Micro target detection method based on self-adaptive multi-feature fusion | |
CN115908241A (en) | Retinal vessel segmentation method based on fusion of UNet and Transformer | |
CN111986101A (en) | Cerebrovascular map construction method | |
CN114972254A (en) | A cervical cell image segmentation method based on convolutional neural network | |
CN109948575A (en) | Segmentation Method of Eyeball Region in Ultrasound Image | |
CN116740119A (en) | Tobacco leaf image active contour segmentation method based on deep learning | |
CN116958537A (en) | A pulmonary nodule segmentation method based on U-Net model | |
CN117173595A (en) | Unmanned aerial vehicle aerial image target detection method based on improved YOLOv7 | |
CN110992309B (en) | Fundus image segmentation method based on deep information transfer network | |
CN117853730A (en) | U-shaped full convolution medical image segmentation network based on convolution kernel attention mechanism |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination |