CN114266794B - Pathological section image cancer region segmentation system based on full convolution neural network - Google Patents

Pathological section image cancer region segmentation system based on full convolution neural network Download PDF

Info

Publication number
CN114266794B
CN114266794B CN202210183367.6A CN202210183367A CN114266794B CN 114266794 B CN114266794 B CN 114266794B CN 202210183367 A CN202210183367 A CN 202210183367A CN 114266794 B CN114266794 B CN 114266794B
Authority
CN
China
Prior art keywords
image
segmentation
cancer
module
pathological slice
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN202210183367.6A
Other languages
Chinese (zh)
Other versions
CN114266794A (en
Inventor
唐杰
肖鸿昭
宋弘健
李清华
胡俊承
王丽萍
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
South China University of Technology SCUT
Guilin Medical University
Original Assignee
South China University of Technology SCUT
Guilin Medical University
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by South China University of Technology SCUT, Guilin Medical University filed Critical South China University of Technology SCUT
Priority to CN202210183367.6A priority Critical patent/CN114266794B/en
Publication of CN114266794A publication Critical patent/CN114266794A/en
Application granted granted Critical
Publication of CN114266794B publication Critical patent/CN114266794B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Landscapes

  • Image Analysis (AREA)

Abstract

本发明公开了一种基于全卷积神经网络的病理切片图像癌症区域分割系统,该系统的实现包括:提取病理切片图像中组织区域的掩码,去除空白背景;将组织区域与病理切片图像中标注的癌症区域结合,对全视野的病理切片图像切割获得训练样本数据;对样本数据进行扩充;构建以Resnet50为编码器的Unet分割网络,将第一级编码器的卷积单元替换成组合卷积单元,提取输入图像中不同尺度的信息,在解码器引入特征融合模块,充分利用每一级解码器输出的信息;分割网络在经过数据增强的数据集上训练并调优;采用网格处理算法对整张病理切片图像进行预测,识别其中的癌症区域。本发明在特征提取阶段引入多尺度信息,提高了对癌症区域的分割精度。

Figure 202210183367

The invention discloses a pathological slice image cancer area segmentation system based on a full convolutional neural network. The realization of the system includes: extracting the mask of the tissue area in the pathological slice image, removing the blank background; Combine the labeled cancer areas, cut the full-field pathological slice images to obtain training sample data; expand the sample data; build a Unet segmentation network with Resnet50 as the encoder, and replace the convolution unit of the first-level encoder with a combined volume The product unit extracts information of different scales in the input image, and introduces a feature fusion module in the decoder to make full use of the information output by each stage of the decoder; the segmentation network is trained and tuned on the data-enhanced dataset; grid processing is adopted The algorithm makes predictions on the entire pathological slide image and identifies areas of cancer within it. The invention introduces multi-scale information in the feature extraction stage, and improves the segmentation accuracy of the cancer area.

Figure 202210183367

Description

基于全卷积神经网络的病理切片图像癌症区域分割系统Cancer region segmentation system in pathological slice images based on fully convolutional neural network

技术领域technical field

本发明涉及图像识别技术领域,具体涉及一种基于全卷积神经网络的病理切片图像癌症区域分割系统。The invention relates to the technical field of image recognition, in particular to a pathological slice image cancer region segmentation system based on a full convolutional neural network.

背景技术Background technique

在神经网络流行之前,研究人员就已开始采用图像处理的技术对病理切片图像进行辅助诊断,主要是从图像的统计特征、纹理特征和形态学特征等方面对图像进行特征提取,判断图像中是否包含某些结构和特征并将它们分离出来。尽管传统的图像处理方法已经可以对医生的诊断起到辅助作用,但普遍存在精度低,效果不稳定等问题。Before the popularity of neural networks, researchers have begun to use image processing technology to assist in the diagnosis of pathological slice images, mainly from the statistical features, texture features and morphological features of the image to extract features from the image to determine whether the image is in the image. Include certain structures and features and isolate them. Although traditional image processing methods can already play an auxiliary role in the diagnosis of doctors, there are generally problems such as low precision and unstable effect.

而采用神经网络对病理切片图像进行识别,大部分研究也只停留在判断图像中是否有癌症,而没有具体分割出具有癌症的区域,对病理医生的依赖还是十分重的,对病理医生只能起到轻微的辅助作用。However, using neural networks to identify pathological slice images, most of the research only stops at judging whether there is cancer in the image, and does not specifically segment the area with cancer. The reliance on pathologists is still very heavy, and pathologists can only play a minor auxiliary role.

发明内容SUMMARY OF THE INVENTION

为了克服现有技术存在的缺陷与不足,本发明提供一种基于全卷积神经网络的病理切片图像癌症区域分割系统,本发明在网络结构中引入组合卷积,充分提取输入图像中不同尺度的特征信息,相较于没有引入不同尺度信息的网络具有更加优秀的分割效果。In order to overcome the defects and deficiencies of the prior art, the present invention provides a pathological slice image cancer region segmentation system based on a full convolutional neural network. The present invention introduces combined convolution into the network structure to fully extract different scales in the input image. Compared with the network that does not introduce different scale information, it has a better segmentation effect.

为了达到上述目的,本发明采用以下技术方案:In order to achieve the above object, the present invention adopts the following technical solutions:

一种基于全卷积神经网络的病理切片图像癌症区域分割系统,包括:组织区域提取模块、样本数据构建模块、样本数据扩充模块、分割网络构建模块、网络训练模块、癌症区域识别模块;A pathological slice image cancer region segmentation system based on a full convolutional neural network, comprising: a tissue region extraction module, a sample data construction module, a sample data expansion module, a segmentation network construction module, a network training module, and a cancer region identification module;

所述组织区域提取模块用于提取病理切片图像中的组织区域,得到组织区域的掩码,去除空白背景;The tissue region extraction module is used to extract the tissue region in the pathological slice image, obtain the mask of the tissue region, and remove the blank background;

所述样本数据构建模块用于将提取的组织区域与病理切片图像中标注的癌症区域结合,对全视野的病理切片图像进行切割,获得训练用的样本数据;The sample data building module is used to combine the extracted tissue area with the cancer area marked in the pathological section image, and cut the full-field pathological section image to obtain sample data for training;

所述样本数据扩充模块用于采用图像增强对样本数据进行扩充;The sample data expansion module is used to expand the sample data by using image enhancement;

所述分割网络构建模块用于构建以Resnet50为编码器的Unet分割网络,将编码器Resnet50的第一级编码器的卷积单元替换成组合卷积单元,所述组合卷积单元设有多种不同大小卷积核的卷积,提取输入图像中不同尺度的信息,并在解码器部分引入特征融合模块,充分利用每一级解码器输出的信息;The segmentation network building module is used to construct a Unet segmentation network with Resnet50 as the encoder, and replace the convolutional unit of the first-stage encoder of the encoder Resnet50 with a combined convolutional unit, and the combined convolutional unit is provided with a variety of The convolution of different size convolution kernels extracts the information of different scales in the input image, and introduces a feature fusion module in the decoder part to make full use of the information output by each stage of the decoder;

将每级解码器输出的特征图进行上采样操作后经过参数不同的卷积,并在通道上拼接得到整体特征图TAfter the feature map output by each level of decoder is subjected to the upsampling operation, the convolution with different parameters is performed, and the overall feature map T is obtained by splicing on the channel;

整体特征图T分别经过一个最大池化层和一个平均池化层得到两个向量V 1 V 2 ,两个向量V 1 V 2 通过一个共享参数的多层感知机后相加得到一个向量CThe overall feature map T passes through a maximum pooling layer and an average pooling layer to obtain two vectors V 1 and V 2 , and the two vectors V 1 and V 2 are added through a shared parameter multilayer perceptron to obtain a vector C ;

将Sigmoid函数应用于向量C并和整体特征图T相乘得到包含通道注意力信息的特征图T’The Sigmoid function is applied to the vector C and multiplied with the overall feature map T to obtain a feature map T' containing channel attention information;

将整体特征图T和特征图T’相加之后经过输出卷积模块得到最终的分割结果;After adding the overall feature map T and the feature map T' , the final segmentation result is obtained through the output convolution module;

所述网络训练模块用于将所述分割网络在经过数据增强的数据集上训练并调优;The network training module is used for training and tuning the segmentation network on the data-enhanced dataset;

所述癌症区域识别模块用于采用训练好的模型并基于网格处理算法对整张病理切片图像进行预测,识别其中的癌症区域。The cancer region identification module is used for using the trained model and based on the grid processing algorithm to predict the entire pathological slice image, and identify the cancer region therein.

作为优选的技术方案,所述组织区域提取模块用于提取病理切片图像中的组织区域,具体步骤包括:As a preferred technical solution, the tissue region extraction module is used to extract tissue regions in pathological slice images, and the specific steps include:

将原病理切片图像从RGB颜色空间转化至HSV颜色空间;Convert the original pathological slice image from RGB color space to HSV color space;

采用大津算法得到转换后的HSV图像的S分量的分割阈值;The segmentation threshold of the S component of the converted HSV image is obtained by using the Otsu algorithm;

基于分割阈值对图像进行二值化处理,得到组织区域的掩码。The image is binarized based on the segmentation threshold to obtain the mask of the tissue area.

作为优选的技术方案,所述对全视野的病理切片图像进行切割,获得训练用的样本数据,具体步骤包括:As a preferred technical solution, the pathological slice images of the full field of view are cut to obtain sample data for training, and the specific steps include:

将所述组织区域的掩码与原始标注的癌症区域掩码按位相与,得到整张病理切片图像组织区域的最终标注掩码;The mask of the tissue region and the mask of the originally marked cancer region are phased together to obtain the final marked mask of the tissue region of the entire pathological slice image;

将原病理切片和最终标注掩码同时进行切割获得图像块及其标注。The original pathological slice and the final annotation mask are simultaneously cut to obtain image blocks and their annotations.

作为优选的技术方案,所述样本数据扩充模块用于采用图像增强对样本数据进行扩充,具体步骤包括:As a preferred technical solution, the sample data expansion module is used to expand the sample data by using image enhancement, and the specific steps include:

将病理切片图像本身及其标注图像进行图像变换,所述图像变换包括水平翻转、随机角度旋转、颜色变化、噪声干扰和模糊中的任意一种或多种。Image transformation is performed on the pathological slice image itself and its annotated image, and the image transformation includes any one or more of horizontal flipping, random angle rotation, color change, noise interference and blurring.

作为优选的技术方案,所述组合卷积单元采用4种不同大小的卷积,包括3*3,5*5,7*7和1*1的卷积核,输入图像分别经过这4种卷积,之后将4个卷积结果在通道上拼接,作为最后整个组合卷积的输出。As a preferred technical solution, the combined convolution unit adopts 4 convolutions of different sizes, including 3*3, 5*5, 7*7 and 1*1 convolution kernels, and the input image passes through these 4 convolutions respectively. product, and then concatenate the 4 convolution results on the channel as the output of the final combined convolution.

作为优选的技术方案,所述输出卷积模块包括一个3*3的卷积和Sigmoid激活函数。As a preferred technical solution, the output convolution module includes a 3*3 convolution and a sigmoid activation function.

作为优选的技术方案,所述多层感知机包括一层1*1卷积层、ReLu激活函数和另一层1*1卷积层,用于压缩信息,生成通道上的注意力结果。As a preferred technical solution, the multilayer perceptron includes a 1*1 convolutional layer, a ReLu activation function and another 1*1 convolutional layer for compressing information and generating on-channel attention results.

作为优选的技术方案,所述癌症区域识别模块用于采用训练好的模型并基于网格处理算法对整张病理切片图像进行预测,识别其中的癌症区域,具体步骤包括:As a preferred technical solution, the cancer region identification module is used to predict the entire pathological slice image based on the grid processing algorithm using the trained model and identify the cancer region. The specific steps include:

采用网格算法将整张病理切片图像分块形成网格,每个网格作为一个预测样本;The whole pathological slice image is divided into grids by grid algorithm, and each grid is used as a prediction sample;

采用滑动窗口每次取网格中的图像块输入分割网络获得分割结果;Use a sliding window to take the image blocks in the grid each time and input the segmentation network to obtain the segmentation result;

将每个滑动窗口的分割结果进行拼接,得到整张病理切片图像的分割结果。The segmentation results of each sliding window are spliced to obtain the segmentation results of the entire pathological slice image.

本发明与现有技术相比,具有如下优点和有益效果:Compared with the prior art, the present invention has the following advantages and beneficial effects:

(1)本发明在网络结构中引入组合卷积,充分提取输入图像中不同尺度的特征信息,相较于没有引入不同尺度信息的网络具有更加优秀的分割效果,引入不同尺度信息提高了对癌症区域的分割精度。(1) The present invention introduces combined convolution in the network structure to fully extract the feature information of different scales in the input image. Compared with the network without introducing different scale information, it has a better segmentation effect. Segmentation accuracy of the region.

(2)本发明在生成分割结果时,分别提取了5级解码器的输出并采用特征融合模块进行特征融合,这样的处理可以充分利用低级特征图中的空间位置等信息和高级特征图中的高级语义信息,使分割的结果更加精确。(2) When generating the segmentation result, the present invention extracts the output of the 5-level decoder respectively and uses the feature fusion module to perform feature fusion. Such processing can make full use of the information such as the spatial position in the low-level feature map and the information in the high-level feature map. Advanced semantic information makes the segmentation results more accurate.

附图说明Description of drawings

图1为本发明基于全卷积神经网络的病理切片图像癌症区域分割系统的实现流程示意图;Fig. 1 is the realization flow schematic diagram of the pathological slice image cancer region segmentation system based on the fully convolutional neural network of the present invention;

图2为本发明的全卷积网络结构示意图;2 is a schematic diagram of a fully convolutional network structure of the present invention;

图3为本发明的网格处理算法的实现过程示意图3 is a schematic diagram of the implementation process of the grid processing algorithm of the present invention

图4(a)为本发明用于预测的整张病理切片图像的原图示意图;Figure 4(a) is a schematic diagram of the original image of the entire pathological slice image used for prediction in the present invention;

图4(b)为本发明用于预测的整张病理切片图像的标注掩码示意图;Figure 4(b) is a schematic diagram of the labeling mask of the entire pathological slice image used for prediction in the present invention;

图4(c)为本发明对整张病理切片图像的分割结果示意图。FIG. 4( c ) is a schematic diagram of the segmentation result of the entire pathological slice image according to the present invention.

具体实施方式Detailed ways

为了使本发明的目的、技术方案及优点更加清楚明白,以下结合附图及实施例,对本发明进行进一步详细说明。应当理解,此处所描述的具体实施例仅仅用以解释本发明,并不用于限定本发明。In order to make the objectives, technical solutions and advantages of the present invention clearer, the present invention will be further described in detail below with reference to the accompanying drawings and embodiments. It should be understood that the specific embodiments described herein are only used to explain the present invention, but not to limit the present invention.

实施例Example

本实施例提供一种基于全卷积神经网络的病理切片图像癌症区域分割系统,包括:组织区域提取模块、样本数据构建模块、样本数据扩充模块、分割网络构建模块、网络训练模块、癌症区域识别模块;This embodiment provides a pathological slice image cancer region segmentation system based on a fully convolutional neural network, including: a tissue region extraction module, a sample data construction module, a sample data expansion module, a segmentation network construction module, a network training module, and a cancer region identification module. module;

在本实施例中,组织区域提取模块用于提取病理切片图像中的组织区域,得到组织区域的掩码,去除空白背景;In this embodiment, the tissue region extraction module is used to extract the tissue region in the pathological slice image, obtain the mask of the tissue region, and remove the blank background;

在本实施例中,样本数据构建模块用于将提取的组织区域与病理切片图像中标注的癌症区域结合,对全视野的病理切片图像进行切割,获得训练用的样本数据;In this embodiment, the sample data building module is used to combine the extracted tissue area with the cancer area marked in the pathological slice image, and cut the full-field pathological slice image to obtain sample data for training;

在本实施例中,样本数据扩充模块用于采用图像增强对样本数据进行扩充;In this embodiment, the sample data expansion module is used to expand the sample data by using image enhancement;

在本实施例中,分割网络构建模块用于构建以Resnet50为编码器的Unet分割网络,将编码器Resnet50的第一级编码器的卷积单元替换成组合卷积单元,所述组合卷积单元设有多种不同大小卷积核的卷积,提取输入图像中不同尺度的信息,并在解码器部分引入特征融合模块,充分利用每一级解码器输出的信息;In this embodiment, the segmentation network building module is used to construct the Unet segmentation network with Resnet50 as the encoder, and the convolutional unit of the first-stage encoder of the encoder Resnet50 is replaced by a combined convolutional unit, the combined convolutional unit There are multiple convolution kernels of different sizes, extracting information of different scales in the input image, and introducing a feature fusion module in the decoder part to make full use of the information output by each level of decoder;

将每级解码器输出的特征图进行上采样操作后经过参数不同的卷积,并在通道上拼接得到整体特征图TAfter the feature map output by each level of decoder is subjected to the upsampling operation, the convolution with different parameters is performed, and the overall feature map T is obtained by splicing on the channel;

整体特征图T分别经过一个最大池化层和一个平均池化层得到两个向量V 1 V 2 ,两个向量V 1 V 2 通过一个共享参数的多层感知机后相加得到一个向量CThe overall feature map T passes through a maximum pooling layer and an average pooling layer to obtain two vectors V 1 and V 2 , and the two vectors V 1 and V 2 are added through a shared parameter multilayer perceptron to obtain a vector C ;

将Sigmoid函数应用于向量C并和整体特征图T相乘得到包含通道注意力信息的特征图T’The Sigmoid function is applied to the vector C and multiplied with the overall feature map T to obtain a feature map T' containing channel attention information;

将整体特征图T和特征图T’相加之后经过输出卷积模块得到最终的分割结果;After adding the overall feature map T and the feature map T' , the final segmentation result is obtained through the output convolution module;

在本实施例中,网络训练模块用于将所述分割网络在经过数据增强的数据集上训练并调优;In this embodiment, the network training module is used to train and optimize the segmentation network on the data-enhanced dataset;

在本实施例中,癌症区域识别模块用于采用训练好的模型并基于网格处理算法对整张病理切片图像进行预测,识别其中的癌症区域。In this embodiment, the cancer region identification module is used for using the trained model and based on the grid processing algorithm to predict the entire pathological slice image, and identify the cancer region therein.

如图1所示,在本实施例基于全卷积神经网络的病理切片图像癌症区域分割系统的具体实现方法中,具体实施时采用的是Grand Challenge 比赛Camelyon16上公开的乳腺癌数据集,该数据集包含了111张带标注的包含癌症的乳腺癌病理切片图像和159张正常的病理切片图像,测试集包含129张病理切片图像(其中只有48张具有肿瘤区域的标注),包括下述步骤:As shown in FIG. 1 , in the specific implementation method of the pathological slice image cancer region segmentation system based on the fully convolutional neural network in this embodiment, the breast cancer data set disclosed in the Grand Challenge competition Camelyon16 is used in the specific implementation. The set contains 111 annotated breast cancer pathological section images containing cancer and 159 normal pathological section images, and the test set contains 129 pathological section images (of which only 48 have annotations of tumor regions), including the following steps:

S1、使用大津算法和颜色空间转换,提取病理切片图像中的组织区域,去除空白背景;本实施例的具体过程如下:S1, use Otsu algorithm and color space conversion, extract the tissue area in the pathological slice image, remove blank background; The concrete process of this embodiment is as follows:

S11、将原病理切片图像从RGB颜色空间转化至HSV颜色空间;S11. Convert the original pathological slice image from the RGB color space to the HSV color space;

S12、应用大津算法得到转换后的HSV图像的S分量的分割阈值;S12, applying the Otsu algorithm to obtain the segmentation threshold of the S component of the converted HSV image;

S13、利用S12步骤中的到的阈值对图像进行二值化处理,得到组织区域的掩码,掩码中,属于组织区域的值为255,不属于组织的区域的值为0。S13. Binarize the image by using the threshold obtained in step S12 to obtain a mask of the tissue area. In the mask, the value of the area belonging to the tissue is 255, and the value of the area not belonging to the tissue is 0.

S2、将提取的组织区域与病理切片图像中标注的癌症区域结合,对全视野的病理切片图像进行切割,获得训练用的样本数据,具体处理过程如下:S2. Combine the extracted tissue area with the cancer area marked in the pathological section image, and cut the full-field pathological section image to obtain sample data for training. The specific processing process is as follows:

S21、在原始标注的癌症区域掩码中,癌症区域的值为1,其他区域的值为0。为了使组织区域的掩码不对标注的掩码值产生影响,将S1步骤中获得的组织区域掩码与原始标注的癌症区域掩码按位相与,便可得到整张病理切片图像组织区域的标注掩码,记为Mask,Mask中不为0的区域表示该区域既是组织又包含癌症,其他区域的掩码值均为0;S21. In the original labeled cancer region mask, the value of the cancer region is 1, and the value of other regions is 0. In order to make the mask of the tissue area not affect the marked mask value, the tissue area mask obtained in step S1 and the original marked cancer area mask are bitwise ANDed, and then the labeling of the tissue area of the entire pathological slice image can be obtained. Mask, denoted as Mask, an area not 0 in Mask indicates that the area is both tissue and cancer, and the mask value of other areas is 0;

S22、将原病理切片图像和步骤S21中得到的Mask同时进行切割获得小的图像块及其标注,图像块的大小记为W,W的取值可以是1024,512,256等,具体实施时为了提高效率、减少计算复杂度,采用的W为1024,但对切割后的图像块进行下采样,使其大小为256,这样既可以保证图像块中包含足够的信息又加快计算速度;S22. Cut the original pathological slice image and the Mask obtained in step S21 at the same time to obtain small image blocks and their labels. The size of the image blocks is denoted as W, and the value of W can be 1024, 512, 256, etc. Efficiency and reducing computational complexity, the W used is 1024, but the cut image block is downsampled to make its size 256, which can ensure that the image block contains enough information and speed up the calculation;

S3、采用图像增强技术对样本数据进行扩充,以提高模型的泛化性能,具体实施过程如下:S3. Use image enhancement technology to expand the sample data to improve the generalization performance of the model. The specific implementation process is as follows:

根据病理切片图像没有固定方向的特点,当图像输入网络进行训练时,将其本身及标注同时以一定概率应用水平翻转、随机角度旋转、颜色变化、噪声干扰和模糊等图像变换,从而达到扩充样本数据的目的。According to the fact that pathological slice images do not have a fixed direction, when the image is input to the network for training, image transformations such as horizontal flipping, random angle rotation, color change, noise interference and blurring are applied to the image itself and the label at the same time with a certain probability, so as to expand the sample. the purpose of the data.

S4、构建引入不同尺度信息的分割网络,并训练及调优,本实施例的具体过程如下:S4, constructing a segmentation network that introduces information of different scales, and training and tuning. The specific process of this embodiment is as follows:

S41、构建以Resnet50为编码器的Unet分割网络,其结构可以概括为编码器和解码器相连的结构,解码器和编码器各有5级,每级编码器由Resnet50的各个Stage(共5个)构成,每级解码器由一个卷积核大小为2*2,步长为2的反卷积和两层3*3卷积构成,实现图像分辨率的恢复;S41. Construct a Unet segmentation network with Resnet50 as the encoder. Its structure can be summarized as a structure in which the encoder and the decoder are connected. The decoder and the encoder each have 5 levels. ), each stage of the decoder consists of a deconvolution with a convolution kernel size of 2*2 and a stride of 2 and two layers of 3*3 convolutions to restore image resolution;

S42、将步骤S41构建的分割网络中的编码器Resnet50的第一级编码器的卷积单元(即编码器Resnet50的第一个残差模块)替换成组合卷积单元,充分提取输入图像中不同尺度的信息,对提高分割精度有增益;本实施例采用的组合卷积由4种不同大小的卷积组成,如分别为3*3,5*5,7*7和1*1的卷积核,不同大小的卷积核的组合并没有限制,输入图像分别经过这4种卷积,之后将4个卷积结果在通道上拼接,作为最后整个组合卷积的输出;S42. Replace the convolution unit of the first-stage encoder of the encoder Resnet50 in the segmentation network constructed in step S41 (that is, the first residual module of the encoder Resnet50) with a combined convolution unit to fully extract the difference in the input image. The scale information has a gain in improving the segmentation accuracy; the combined convolution used in this embodiment consists of 4 convolutions of different sizes, such as 3*3, 5*5, 7*7 and 1*1 convolutions respectively. Kernel, the combination of convolution kernels of different sizes is not limited. The input image undergoes these 4 convolutions respectively, and then the 4 convolution results are spliced on the channel as the output of the final combined convolution;

S43、在步骤S42构建的分割网络的解码器中引入特征融合模块,充分利用5级解码器的信息,使分割效果更佳,引入特征融合模块的过程如下:S43. Introduce a feature fusion module into the decoder of the segmentation network constructed in step S42, and make full use of the information of the 5-level decoder to make the segmentation effect better. The process of introducing the feature fusion module is as follows:

将5级解码器输出的特征图都进行上采样操作,使这些特征图的大小与最终的目标输出大小一致,记为

Figure DEST_PATH_IMAGE001
;The feature maps output by the 5-level decoder are all up-sampled, so that the size of these feature maps is consistent with the final target output size, denoted as
Figure DEST_PATH_IMAGE001
;

将上述获得的

Figure 243511DEST_PATH_IMAGE001
5个特征图都经过参数不同的1*1卷积,主要作用是减少特征图的通道数量,可以起到降低计算量的作用,所获得的特征图记为
Figure DEST_PATH_IMAGE002
;obtained above
Figure 243511DEST_PATH_IMAGE001
The 5 feature maps are all subjected to 1*1 convolution with different parameters. The main function is to reduce the number of channels in the feature map, which can reduce the amount of calculation. The obtained feature map is recorded as
Figure DEST_PATH_IMAGE002
;

Figure 285286DEST_PATH_IMAGE002
5个特征图在通道上拼接获得一个整体特征图T;Will
Figure 285286DEST_PATH_IMAGE002
5 feature maps are spliced on the channel to obtain an overall feature map T ;

将整体特征图T分别经过一个最大池化层和一个平均池化层得到两个向量V 1 V 2 ,将两个向量V 1 V 2 通过一个共享参数的多层感知机后相加得到一个向量C,所述多层感知机是由一层1*1卷积层、ReLu激活函数和另一层1*1卷积层构成的三层结构,用于压缩信息,生成通道上的注意力结果;Pass the overall feature map T through a maximum pooling layer and an average pooling layer to obtain two vectors V 1 and V 2 , and add the two vectors V 1 and V 2 through a shared parameter multi-layer perceptron to obtain A vector C , the multi-layer perceptron is a three-layer structure composed of a 1*1 convolutional layer, a ReLu activation function and another 1*1 convolutional layer, which is used to compress information and generate attention on the channel force result;

将Sigmoid函数应用于向量C并和整体特征图T相乘得到包含通道注意力信息的特征图T’The Sigmoid function is applied to the vector C and multiplied with the overall feature map T to obtain a feature map T' containing channel attention information;

将整体特征图T和特征图T’相加之后经过一个由3*3卷积和Sigmoid激活函数构成的输出卷积模块得到最终的分割结果。After adding the overall feature map T and the feature map T' , an output convolution module composed of a 3*3 convolution and a sigmoid activation function is used to obtain the final segmentation result.

S44、如图2所示,图中各级编码器分别对应原始Resnet50中的各个残差模块,将改进的分割网络在经过数据增强的数据集上训练并调优,具体实施时采用如下方案:S44. As shown in Figure 2, the encoders at all levels in the figure correspond to each residual module in the original Resnet50 respectively, and the improved segmentation network is trained and optimized on the data-enhanced data set. The specific implementation adopts the following scheme:

使用Adam优化器进行模型参数的更新,并使用指数变换的方式进行学习率的更新,在训练过程中监控验证集mIoU指标,每次mIoU指标得到提升,则保存对应的模型参数,用于之后的测试和预测。Use the Adam optimizer to update the model parameters, and use the exponential transformation method to update the learning rate. During the training process, monitor the mIoU index of the validation set. Every time the mIoU index is improved, the corresponding model parameters are saved for later use. Test and predict.

训练模型的测试结果如下表1所示,表中mIoU指标和Dice指标都是分割任务评价的常用指标。The test results of the training model are shown in Table 1 below. The mIoU index and the Dice index in the table are common indicators for segmentation task evaluation.

表1 模型测试结果表Table 1 Model test result table

Figure DEST_PATH_IMAGE003
Figure DEST_PATH_IMAGE003

S5、采用训练好的模型并采用网格处理算法对整张病理切片图像进行预测,识别其中的癌症区域,本实施例的具体过程如下:S5, adopt the trained model and adopt the grid processing algorithm to predict the entire pathological slice image, and identify the cancer area therein, and the specific process of the present embodiment is as follows:

S51、如图3所示,采用网格算法将整张病理切片图像分成一个个小块,形成网格,每个小格作为一个预测样本,具体步骤如下:S51. As shown in Figure 3, the entire pathological slice image is divided into small blocks by using a grid algorithm to form a grid, and each small grid is used as a prediction sample. The specific steps are as follows:

假设滑动窗口的大小是W,每个滑动窗口的重叠区域大小为O,则滑动窗口的移动步长S=W-O;Assuming that the size of the sliding window is W, and the size of the overlapping area of each sliding window is O, then the moving step size of the sliding window is S=W-O;

根据步长S计算原病理切片图像长X和宽Y两个方向上可获得的图像块数量nx和ny,其计算公式分别为:Calculate the number of image blocks n x and ny that can be obtained in the two directions of length X and width Y of the original pathological slice image according to the step size S, and the calculation formulas are:

Figure DEST_PATH_IMAGE004
Figure DEST_PATH_IMAGE004

Figure DEST_PATH_IMAGE005
Figure DEST_PATH_IMAGE005

按照nx和ny对整张病理切片图像的长X和宽Y进行划分,并将坐标点保存在X a Y a 中,X a =[X 1 ,X 2 ,…,X n ],Y a =[Y 1 ,Y 2 ,…,Y n ];Divide the length X and width Y of the entire pathological slice image according to n x and ny , and save the coordinate points in X a and Y a , X a =[ X 1 , X 2 ,…, X n ], Y a =[ Y 1 , Y 2 ,…, Y n ];

根据每个窗口的起始点坐标(X i ,Y i )和滑动窗口的大小,获取切片图像中的小图像块,网格中的每个小格的左上角顶点坐标为切割小图像块时索引的坐标,若图像块的大小超出了切片图像的边界(右边界和下边界),则将图像块的起始点往左移或者往上移,使超出切片图像边界的图像块恰好与切片图像的边界吻合。According to the starting point coordinates ( X i , Y i ) of each window and the size of the sliding window, the small image blocks in the sliced image are obtained. The coordinates of the upper left corner of each small cell in the grid are the indices when the small image blocks are cut. If the size of the image block exceeds the boundary of the sliced image (right and lower boundaries), move the starting point of the image block to the left or up, so that the image block beyond the boundary of the sliced image is exactly the same as the sliced image. The borders match.

S52、采用滑动窗口每次取网格中的一个小块输入网络获得分割结果;S52, adopting the sliding window to take a small block in the grid each time to input the network to obtain the segmentation result;

S53、如图4(c)所示,并结合图4(a)、图4(b)所示,将每个滑动窗口的分割结果进行拼接,得到整张病理切片图像的分割结果,整张病理切片图像的分割结果。S53. As shown in Figure 4(c), combined with Figures 4(a) and 4(b), splicing the segmentation results of each sliding window to obtain the segmentation results of the entire pathological slice image. Segmentation results of pathological section images.

上述实施例为本发明较佳的实施方式,但本发明的实施方式并不受上述实施例的限制,其他的任何未背离本发明的精神实质与原理下所作的改变、修饰、替代、组合、简化,均应为等效的置换方式,都包含在本发明的保护范围之内。The above-mentioned embodiments are preferred embodiments of the present invention, but the embodiments of the present invention are not limited by the above-mentioned embodiments, and any other changes, modifications, substitutions, combinations, The simplification should be equivalent replacement manners, which are all included in the protection scope of the present invention.

Claims (7)

1.一种基于全卷积神经网络的病理切片图像癌症区域分割系统,其特征在于,包括:组织区域提取模块、样本数据构建模块、样本数据扩充模块、分割网络构建模块、网络训练模块、癌症区域识别模块;1. a pathological slice image cancer region segmentation system based on full convolutional neural network, is characterized in that, comprises: tissue region extraction module, sample data building module, sample data expansion module, segmentation network building module, network training module, cancer area identification module; 所述组织区域提取模块用于提取病理切片图像中的组织区域,得到组织区域的掩码,去除空白背景;The tissue region extraction module is used to extract the tissue region in the pathological slice image, obtain the mask of the tissue region, and remove the blank background; 所述样本数据构建模块用于将提取的组织区域与病理切片图像中标注的癌症区域结合,对全视野的病理切片图像进行切割,获得训练用的样本数据;The sample data building module is used to combine the extracted tissue area with the cancer area marked in the pathological section image, and cut the full-field pathological section image to obtain sample data for training; 将所述组织区域的掩码与原始标注的癌症区域掩码按位相与,得到整张病理切片图像组织区域的最终标注掩码;The mask of the tissue region and the mask of the originally marked cancer region are bitwise added to obtain the final marked mask of the tissue region of the entire pathological slice image; 将原病理切片和最终标注掩码同时进行切割获得图像块及其标注;Cut the original pathological slice and the final annotation mask at the same time to obtain image blocks and their annotations; 所述样本数据扩充模块用于采用图像增强对样本数据进行扩充;The sample data expansion module is used to expand the sample data by using image enhancement; 所述分割网络构建模块用于构建以Resnet50为编码器的Unet分割网络,将编码器Resnet50的第一级编码器的卷积单元替换成组合卷积单元,所述组合卷积单元设有多种不同大小卷积核的卷积,提取输入图像中不同尺度的信息,并在解码器部分引入特征融合模块,充分利用每一级解码器输出的信息;The segmentation network building module is used to construct a Unet segmentation network with Resnet50 as the encoder, and replace the convolutional unit of the first-stage encoder of the encoder Resnet50 with a combined convolutional unit, and the combined convolutional unit is provided with a variety of The convolution of different size convolution kernels extracts the information of different scales in the input image, and introduces a feature fusion module in the decoder part to make full use of the information output by each stage of the decoder; 将每级解码器输出的特征图进行上采样操作后经过参数不同的卷积,并在通道上拼接得到整体特征图TAfter the feature map output by each level of decoder is subjected to the upsampling operation, the convolution with different parameters is performed, and the overall feature map T is obtained by splicing on the channel; 整体特征图T分别经过一个最大池化层和一个平均池化层得到两个向量V 1 V 2 ,两个向量V 1 V 2 通过一个共享参数的多层感知机后相加得到一个向量CThe overall feature map T passes through a maximum pooling layer and an average pooling layer to obtain two vectors V 1 and V 2 , and the two vectors V 1 and V 2 are added through a shared parameter multilayer perceptron to obtain a vector C ; 将Sigmoid函数应用于向量C并和整体特征图T相乘得到包含通道注意力信息的特征图T’The Sigmoid function is applied to the vector C and multiplied with the overall feature map T to obtain a feature map T' containing channel attention information; 将整体特征图T和特征图T’相加之后经过输出卷积模块得到最终的分割结果;After adding the overall feature map T and the feature map T' , the final segmentation result is obtained through the output convolution module; 所述网络训练模块用于将所述分割网络在经过数据增强的数据集上训练并调优;The network training module is used for training and tuning the segmentation network on the data-enhanced dataset; 所述癌症区域识别模块用于采用训练好的模型并基于网格处理算法对整张病理切片图像进行预测,识别其中的癌症区域。The cancer region identification module is used for using the trained model and based on the grid processing algorithm to predict the entire pathological slice image, and identify the cancer region therein. 2.根据权利要求1所述的基于全卷积神经网络的病理切片图像癌症区域分割系统,其特征在于,所述组织区域提取模块用于提取病理切片图像中的组织区域,具体步骤包括:2. The pathological slice image cancer region segmentation system based on a full convolutional neural network according to claim 1, wherein the tissue region extraction module is used to extract the tissue region in the pathological slice image, and the concrete steps include: 将原病理切片图像从RGB颜色空间转化至HSV颜色空间;Convert the original pathological slice image from RGB color space to HSV color space; 采用大津算法得到转换后的HSV图像的S分量的分割阈值;The segmentation threshold of the S component of the converted HSV image is obtained by using the Otsu algorithm; 基于分割阈值对图像进行二值化处理,得到组织区域的掩码。The image is binarized based on the segmentation threshold to obtain the mask of the tissue area. 3.根据权利要求1所述的基于全卷积神经网络的病理切片图像癌症区域分割系统,其特征在于,所述样本数据扩充模块用于采用图像增强对样本数据进行扩充,具体步骤包括:3. the pathological slice image cancer region segmentation system based on full convolutional neural network according to claim 1, is characterized in that, described sample data expansion module is used for adopting image enhancement to expand sample data, and concrete steps comprise: 将病理切片图像本身及其标注图像进行图像变换,所述图像变换包括水平翻转、随机角度旋转、颜色变化、噪声干扰和模糊中的任意一种或多种。Image transformation is performed on the pathological slice image itself and its annotated image, and the image transformation includes any one or more of horizontal flipping, random angle rotation, color change, noise interference and blurring. 4.根据权利要求1所述的基于全卷积神经网络的病理切片图像癌症区域分割系统,其特征在于,所述组合卷积单元采用4种不同大小的卷积,包括3*3,5*5,7*7和1*1的卷积核,输入图像分别经过这4种卷积,之后将4个卷积结果在通道上拼接,作为最后整个组合卷积的输出。4. The pathological slice image cancer region segmentation system based on a fully convolutional neural network according to claim 1, wherein the combined convolution unit adopts 4 convolutions of different sizes, including 3*3, 5* For the convolution kernels of 5, 7*7 and 1*1, the input image undergoes these 4 convolutions respectively, and then the 4 convolution results are spliced on the channel as the output of the final combined convolution. 5.根据权利要求1所述的基于全卷积神经网络的病理切片图像癌症区域分割系统,其特征在于,所述输出卷积模块包括一个3*3的卷积和Sigmoid激活函数。5. The full convolutional neural network-based pathological slice image cancer region segmentation system according to claim 1, wherein the output convolution module comprises a 3*3 convolution and a sigmoid activation function. 6.根据权利要求1所述的基于全卷积神经网络的病理切片图像癌症区域分割系统,其特征在于,所述多层感知机包括一层1*1卷积层、ReLu激活函数和另一层1*1卷积层,用于压缩信息,生成通道上的注意力结果。6. The pathological slice image cancer region segmentation system based on a fully convolutional neural network according to claim 1, wherein the multilayer perceptron comprises a layer of 1*1 convolution layer, a ReLu activation function and another Layer 1*1 convolutional layer for compressing information and generating on-channel attention results. 7.根据权利要求1所述的基于全卷积神经网络的病理切片图像癌症区域分割系统,其特征在于,所述癌症区域识别模块用于采用训练好的模型并基于网格处理算法对整张病理切片图像进行预测,识别其中的癌症区域,具体步骤包括:7. The pathological slice image cancer region segmentation system based on a fully convolutional neural network according to claim 1, wherein the cancer region identification module is used for adopting a trained model and based on a grid processing algorithm. The pathological section image is predicted to identify the cancer area, and the specific steps include: 采用网格算法将整张病理切片图像分块形成网格,每个网格作为一个预测样本;The whole pathological slice image is divided into grids by grid algorithm, and each grid is used as a prediction sample; 采用滑动窗口每次取网格中的图像块输入分割网络获得分割结果;Use a sliding window to take the image blocks in the grid each time and input the segmentation network to obtain the segmentation result; 将每个滑动窗口的分割结果进行拼接,得到整张病理切片图像的分割结果。The segmentation results of each sliding window are spliced to obtain the segmentation results of the entire pathological slice image.
CN202210183367.6A 2022-02-28 2022-02-28 Pathological section image cancer region segmentation system based on full convolution neural network Active CN114266794B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202210183367.6A CN114266794B (en) 2022-02-28 2022-02-28 Pathological section image cancer region segmentation system based on full convolution neural network

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202210183367.6A CN114266794B (en) 2022-02-28 2022-02-28 Pathological section image cancer region segmentation system based on full convolution neural network

Publications (2)

Publication Number Publication Date
CN114266794A CN114266794A (en) 2022-04-01
CN114266794B true CN114266794B (en) 2022-06-10

Family

ID=80833663

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202210183367.6A Active CN114266794B (en) 2022-02-28 2022-02-28 Pathological section image cancer region segmentation system based on full convolution neural network

Country Status (1)

Country Link
CN (1) CN114266794B (en)

Families Citing this family (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN114549520B (en) * 2022-04-08 2024-05-07 北京端点医药研究开发有限公司 Retina pathological section analysis system based on full convolution attention enhancement network
CN114862763B (en) * 2022-04-13 2024-06-21 华南理工大学 EFFICIENTNET-based gastric cancer pathological section image segmentation prediction method
CN114820502B (en) * 2022-04-21 2023-10-24 济宁医学院附属医院 Coloring detection method for protein kinase CK2 in intestinal mucosa tissue
CN115294126B (en) * 2022-10-08 2022-12-16 南京诺源医疗器械有限公司 Cancer cell intelligent identification method for pathological image
CN115631161A (en) * 2022-10-20 2023-01-20 武汉市昱呈星医疗科技有限责任公司 Method and system for processing pathological section image of gastric cancer
CN115690134A (en) * 2022-11-15 2023-02-03 杭州智汇明壹医疗科技有限公司 Segmentation device and method for IHC nuclear expression pathological image cancer region
CN117952969B (en) * 2024-03-26 2024-06-21 济南大学 Endometrial cancer analysis method and system based on selective attention

Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN110322435A (en) * 2019-01-20 2019-10-11 北京工业大学 A kind of gastric cancer pathological image cancerous region dividing method based on deep learning
CN112419286A (en) * 2020-11-27 2021-02-26 苏州斯玛维科技有限公司 Dermoscopy image segmentation method and segmentation device
CN112435246A (en) * 2020-11-30 2021-03-02 武汉楚精灵医疗科技有限公司 Artificial intelligent diagnosis method for gastric cancer under narrow-band imaging amplification gastroscope

Family Cites Families (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN111340828A (en) * 2020-01-10 2020-06-26 南京航空航天大学 Brain Glioma Segmentation Based on Cascaded Convolutional Neural Networks
CN111784671B (en) * 2020-06-30 2022-07-05 天津大学 Pathological image lesion area detection method based on multi-scale deep learning
CN111968127B (en) * 2020-07-06 2021-08-27 中国科学院计算技术研究所 Cancer focus area identification method and system based on full-section pathological image

Patent Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN110322435A (en) * 2019-01-20 2019-10-11 北京工业大学 A kind of gastric cancer pathological image cancerous region dividing method based on deep learning
CN112419286A (en) * 2020-11-27 2021-02-26 苏州斯玛维科技有限公司 Dermoscopy image segmentation method and segmentation device
CN112435246A (en) * 2020-11-30 2021-03-02 武汉楚精灵医疗科技有限公司 Artificial intelligent diagnosis method for gastric cancer under narrow-band imaging amplification gastroscope

Non-Patent Citations (2)

* Cited by examiner, † Cited by third party
Title
Automatic Polyp Segmentation using U-Net-ResNet50;Saruar Alam1 et al.;《arXiv:2012.15247 v1》;20201230;第1-3页 *
Selective Kernel Networks;Xiang Li et al.;《arXiv:1903.06586v2》;20190318;第1-12页 *

Also Published As

Publication number Publication date
CN114266794A (en) 2022-04-01

Similar Documents

Publication Publication Date Title
CN114266794B (en) Pathological section image cancer region segmentation system based on full convolution neural network
CN111369581B (en) Image processing method, device, equipment and storage medium
CN111145181B (en) Skeleton CT image three-dimensional segmentation method based on multi-view separation convolutional neural network
CN110889852B (en) Liver segmentation method based on residual error-attention deep neural network
CN111325750B (en) Medical image segmentation method based on multi-scale fusion U-shaped chain neural network
CN111898432B (en) Pedestrian detection system and method based on improved YOLOv3 algorithm
CN111882560A (en) A lung parenchyma CT image segmentation method based on weighted fully convolutional neural network
CN113192076B (en) MRI Brain Tumor Image Segmentation Using Combined Classification Prediction and Multiscale Feature Extraction
CN113223005A (en) Thyroid nodule automatic segmentation and grading intelligent system
CN116596846A (en) Image segmentation method, image segmentation model construction method, device and medium
CN116524189A (en) High-resolution remote sensing image semantic segmentation method based on coding and decoding indexing edge characterization
CN114565770A (en) Image segmentation method and system based on edge-assisted computation and mask attention
CN116468732A (en) Lung CT image segmentation method and imaging method based on deep learning
CN113096071A (en) CHS-Net network-based medical image segmentation method
CN112990041B (en) A method for extracting buildings from remote sensing images based on improved U-net
CN118072090A (en) Dermatological image detection method based on U2-Net and ResNeXt-50 models
CN110895815A (en) A chest X-ray pneumothorax segmentation method based on deep learning
CN118196113A (en) A liver and tumor segmentation method based on SNAU-Net
CN116703945A (en) Method, system, medium and terminal for tumor segmentation of breast ultrasound images
CN115424017A (en) Building internal and external contour segmentation method, device and storage medium
CN116993661A (en) Clinical diagnosis method for potential cancerous polyps based on feature fusion and attention mechanism
CN114821067B (en) Pathological image segmentation method based on point annotation data
CN114708591B (en) Chinese character detection method in document images based on single-word connection
CN116542924A (en) Method, device and storage medium for detecting prostate lesion area
CN116030256A (en) Small object segmentation method, small object segmentation system, device and medium

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant