CN116524315A - A Method for Recognition and Segmentation of Lung Cancer Pathological Tissue Slices Based on Mask R-CNN - Google Patents

A Method for Recognition and Segmentation of Lung Cancer Pathological Tissue Slices Based on Mask R-CNN Download PDF

Info

Publication number
CN116524315A
CN116524315A CN202210051013.6A CN202210051013A CN116524315A CN 116524315 A CN116524315 A CN 116524315A CN 202210051013 A CN202210051013 A CN 202210051013A CN 116524315 A CN116524315 A CN 116524315A
Authority
CN
China
Prior art keywords
mask
pathological tissue
lung cancer
segmentation
cnn
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN202210051013.6A
Other languages
Chinese (zh)
Inventor
张飒飒
王昭
甄军晖
田遴博
续玉新
杨易
王韬
迟庆金
赵峰榕
金桂蕾
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Shandong University
Original Assignee
Shandong University
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Shandong University filed Critical Shandong University
Priority to CN202210051013.6A priority Critical patent/CN116524315A/en
Publication of CN116524315A publication Critical patent/CN116524315A/en
Pending legal-status Critical Current

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/0002Inspection of images, e.g. flaw detection
    • G06T7/0012Biomedical image inspection
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/08Learning methods
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/10Segmentation; Edge detection
    • G06T7/11Region-based segmentation
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/10Segmentation; Edge detection
    • G06T7/136Segmentation; Edge detection involving thresholding
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/60Analysis of geometric attributes
    • G06T7/62Analysis of geometric attributes of area, perimeter, diameter or volume
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/10Image acquisition modality
    • G06T2207/10072Tomographic images
    • G06T2207/10081Computed x-ray tomography [CT]
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/30Subject of image; Context of image processing
    • G06T2207/30004Biomedical image processing
    • G06T2207/30061Lung
    • G06T2207/30064Lung nodule
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/30Subject of image; Context of image processing
    • G06T2207/30004Biomedical image processing
    • G06T2207/30096Tumor; Lesion
    • YGENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
    • Y02TECHNOLOGIES OR APPLICATIONS FOR MITIGATION OR ADAPTATION AGAINST CLIMATE CHANGE
    • Y02ATECHNOLOGIES FOR ADAPTATION TO CLIMATE CHANGE
    • Y02A90/00Technologies having an indirect contribution to adaptation to climate change
    • Y02A90/10Information and communication technologies [ICT] supporting adaptation to climate change, e.g. for weather forecasting or climate simulation

Landscapes

  • Engineering & Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • Theoretical Computer Science (AREA)
  • General Physics & Mathematics (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • General Health & Medical Sciences (AREA)
  • Health & Medical Sciences (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Computational Linguistics (AREA)
  • Radiology & Medical Imaging (AREA)
  • Nuclear Medicine, Radiotherapy & Molecular Imaging (AREA)
  • Geometry (AREA)
  • Medical Informatics (AREA)
  • Artificial Intelligence (AREA)
  • Biomedical Technology (AREA)
  • Biophysics (AREA)
  • Quality & Reliability (AREA)
  • Data Mining & Analysis (AREA)
  • Evolutionary Computation (AREA)
  • Molecular Biology (AREA)
  • Computing Systems (AREA)
  • General Engineering & Computer Science (AREA)
  • Mathematical Physics (AREA)
  • Software Systems (AREA)
  • Image Analysis (AREA)

Abstract

本发明属于医学影像处理技术领域,具体涉及一种肺癌病理组织切片病症识别及分割方法。一种基于Mask R‑CNN的肺癌病理组织切片识别及分割方法,包括以下步骤:S1,获取患者肺癌病理组织切片扫描图像,并进行预处理;S2,将所述预处理后的扫描图像输入到预先训练好的病症分类及分割模型中,判断切片病症类别,并得到病变区域分割的可视化类激活图;所述的病症分类及分割模型为改进的Mask R‑CNN神经网络;S3,根据获取的可视化类激活图,计算病变区域占全局病理组织切片的面积比例。本发明具有分类准确度高,区域分割平滑,定量计算准确,能够多指标分析图像,辅助医生更迅速、便捷、准确地进行病理研判的优点。

The invention belongs to the technical field of medical image processing, and in particular relates to a disease identification and segmentation method for pathological tissue slices of lung cancer. A method for identifying and segmenting lung cancer pathological tissue slices based on Mask R-CNN, comprising the following steps: S1, obtaining scanned images of patient lung cancer pathological tissue slices, and performing preprocessing; S2, inputting the preprocessed scanned images into In the pre-trained disease classification and segmentation model, the slice disease category is judged, and the visualization class activation map of lesion area segmentation is obtained; the disease classification and segmentation model is an improved Mask R-CNN neural network; S3, according to the acquired Visualize the class activation map and calculate the proportion of the lesion area to the global pathological tissue slice. The invention has the advantages of high classification accuracy, smooth region segmentation, accurate quantitative calculation, multi-indicator image analysis, and assisting doctors to conduct pathological research and judgment more quickly, conveniently and accurately.

Description

一种基于Mask R-CNN的肺癌病理组织切片识别及分割方法A Method for Recognition and Segmentation of Lung Cancer Pathological Tissue Slices Based on Mask R-CNN

技术领域technical field

本发明属于医学影像处理技术领域,具体涉及一种肺癌病理组织切片病症识别及分割方法The invention belongs to the technical field of medical image processing, and in particular relates to a disease identification and segmentation method for pathological tissue slices of lung cancer

背景技术Background technique

肺癌是中国癌症死亡人数最多的一种病,严重威胁着中国居民的身体健康。组织病理学检查是准确性最可靠的诊断依据,医生通过H&E染色的病理组织切片进行活检,是诊断肺癌并辨别其类型及严重程度的“金标准”。随着计算机识别技术的发展,数字病理技术将现阶段数字成像系统与传统光学成像装置相互结合,为医生的诊断分析过程提供更高分辨率、更高清晰度、更稳定的成像条件。Lung cancer is the disease with the largest number of cancer deaths in China, which seriously threatens the health of Chinese residents. Histopathological examination is the most reliable basis for diagnosis. Doctors conduct biopsy through H&E stained pathological tissue sections, which is the "gold standard" for diagnosing lung cancer and distinguishing its type and severity. With the development of computer recognition technology, digital pathology technology combines the current digital imaging system with traditional optical imaging devices to provide doctors with higher resolution, higher definition, and more stable imaging conditions for the diagnosis and analysis process.

医学图像处理是当今世界范围内的热门研究领域之一,计算机辅助诊断的出现,减轻了医生的工作负担,但传统的“专家”系统仍然需要对病变特征进行手工提取,其系统开发周期长,开发成本高。随着深度学习的发展,图像处理领域有了长足的进步。Kaggle2017年举办的肺肿瘤结节(CT图像数据集)识别比赛中,平均召回率(AR)达到了89.7%。何克磊设计了一种端到端的基于原型学习的多实例深度卷积网络,实现了对肺癌病理细胞图像的弱标记环境滤噪识别。现有研究工作集中于肿瘤分类与分割,可解释性差。然而医学以其特有的医学道德伦理,在现阶段令机器学习代替人工进行结论性诊断依然为时过早。因此,增加辅助诊断系统面向医生的多维度评价指标,设计贴合病理的解释性功能,有利于医生进行更精准便捷的诊断参考。Medical image processing is one of the hot research fields in the world today. The emergence of computer-aided diagnosis has reduced the workload of doctors, but the traditional "expert" system still needs to manually extract lesion features, and its system development cycle is long. Development costs are high. With the development of deep learning, the field of image processing has made great progress. In the recognition competition of lung tumor nodules (CT image data set) held by Kaggle in 2017, the average recall rate (AR) reached 89.7%. He Kelei designed an end-to-end multi-instance deep convolutional network based on prototype learning, which realized the weakly labeled environmental noise filtering recognition of lung cancer pathological cell images. Existing research work focuses on tumor classification and segmentation, which has poor interpretability. However, with its unique medical ethics, it is still too early for machine learning to replace manual conclusive diagnosis at this stage. Therefore, adding multi-dimensional evaluation indicators for doctors in the auxiliary diagnosis system and designing explanatory functions that fit pathology will help doctors to provide more accurate and convenient diagnostic references.

常规的组织切片病理变化评价,多采用4级法(即轻微、轻度、中度、重度)。这是一种经典的组织病变评价方法,也是当前的主流。但随着全切片扫描和量化分析观念的普及,定量分析组织病变逐渐流行。有时,定量分析后获取数据,能够较准确地反映病变的实际程度和病变范围。此外,获取的这些数据也更方便对组间差异进行统计学分析。面积测量是定量分析时的一个重要指标。目前已有的测量软件需要人工标识病变区域才能计算面积,本发明将定性分析与定量测量功能融合,打造智慧医疗辅助系统。For routine evaluation of pathological changes in tissue sections, a 4-level method (ie mild, mild, moderate, and severe) is often used. This is a classic method for evaluating tissue lesions, and it is also the current mainstream. However, with the popularization of whole-slice scanning and quantitative analysis concepts, quantitative analysis of tissue lesions has gradually become popular. Sometimes, data obtained after quantitative analysis can more accurately reflect the actual extent and extent of the lesion. In addition, the obtained data are also more convenient for statistical analysis of differences between groups. Area measurement is an important indicator in quantitative analysis. The existing measurement software needs to manually mark the lesion area to calculate the area. The present invention integrates qualitative analysis and quantitative measurement functions to create a smart medical assistance system.

近年来,卷积神经网络是应用于图像处理领域最热门的方法之一,把卷积神经网络与病理图像深度融合,并结合病理特性设计特定功能与评价指标,是未来值得研究的方向。In recent years, convolutional neural network is one of the most popular methods applied in the field of image processing. Deeply integrating convolutional neural network with pathological images and designing specific functions and evaluation indicators in combination with pathological characteristics is a direction worth studying in the future.

发明内容Contents of the invention

本发明的目的是针对上述现有方法的不足提供一种基于深度学习的肺癌病理组织切片识别、分割与定量计算于一体的方法,将图像识别中的Mask R-CNN模型分阶段应用于图像中并结合定量计算的回归算法,实现对肺癌病理组织切片的病症分类及病变区域定位,并计算占全局病历组织切片的面积。本发明具有分类准确度高,区域分割平滑,定量计算准确,能够多指标分析图像,辅助医生更迅速、便捷、准确地进行病理研判的优点。The purpose of the present invention is to provide a method based on deep learning for the identification, segmentation and quantitative calculation of lung cancer pathological tissue slices in order to address the shortcomings of the above existing methods, and apply the Mask R-CNN model in image recognition to images in stages Combined with the regression algorithm of quantitative calculation, the disease classification and lesion area positioning of lung cancer pathological tissue slices are realized, and the area occupied by the global medical record tissue slices is calculated. The invention has the advantages of high classification accuracy, smooth region segmentation, accurate quantitative calculation, multi-indicator image analysis, and assisting doctors to conduct pathological research and judgment more quickly, conveniently and accurately.

为了解决上述技术问题,本发明采用的技术方案是:一种基于Mask R-CNN的肺癌病理组织切片识别及分割方法,包括以下步骤:In order to solve the above-mentioned technical problems, the technical solution adopted in the present invention is: a method for identifying and segmenting lung cancer pathological tissue slices based on Mask R-CNN, comprising the following steps:

S1,获取患者肺癌病理组织切片扫描图像,并进行预处理;S1, obtaining the scanned image of the patient's lung cancer pathological tissue slice, and performing preprocessing;

S2,将所述预处理后的扫描图像输入到预先训练好的病症分类及分割模型中,判断切片病症类别,并得到病变区域分割的可视化类激活图;所述的病症分类及分割模型为改进的Mask R-CNN神经网络;S2, inputting the pre-processed scanned image into the pre-trained disease classification and segmentation model, judging the slice disease category, and obtaining the visualization class activation map of lesion area segmentation; the disease classification and segmentation model is improved Mask R-CNN neural network;

S3,根据获取的可视化类激活图,计算病变区域占全局病理组织切片的面积比例。S3. Calculate the area ratio of the lesion area to the global pathological tissue slice according to the obtained visualization class activation map.

进一步地,所述S1中,所述预处理的具体方法为:Further, in the S1, the specific method of the pretreatment is:

对患者肺癌病理组织切片扫描图像进行20×倍放大,并从TIFP格式转换为jpeg格式。The scanned images of lung cancer pathological tissue slices were magnified by 20×, and converted from TIFP format to jpeg format.

进一步地,所述改进的Mask R-CNN神经网络,具体包括:Further, the improved Mask R-CNN neural network specifically includes:

特征提取网络,所述特征提取网络包括改进后的残差网络ResNet;所述改进后的残差网络ResNet,在最后的分类层前增加全连接层及dropout层;A feature extraction network, the feature extraction network includes an improved residual network ResNet; the improved residual network ResNet adds a fully connected layer and a dropout layer before the last classification layer;

FPN网络,在所述特征提取网络中加入FPN网络,对提取特征进行多尺度融合;FPN network, add FPN network in described feature extraction network, carry out multi-scale fusion to extraction feature;

RPN网络,用于对FPN融合后的特征进行目标区域生成,并将其分数值最高的设定数量个候选区域输入Mask R-CNN网络;The RPN network is used to generate the target area for the features after FPN fusion, and input the set number of candidate areas with the highest score value into the Mask R-CNN network;

Mask R-CNN网络,对输入的候选区域进行病症的分类,及病变区域的分割,产生背景和病变区域的分割Mask。The Mask R-CNN network classifies the input candidate regions for disease classification and the segmentation of the lesion area, and generates the segmentation Mask of the background and lesion area.

进一步地,所述S3中,病变区域占全局病理组织切片的面积比例计算方法包括:Further, in said S3, the method for calculating the area ratio of the lesion area to the global pathological tissue slice includes:

(1),对获取的可视化类激活图像进行高斯模糊处理,设置病变区域的灰度为0,背景区域的灰度为255;(1), Gaussian blur processing is performed on the obtained visualization activation image, the gray level of the lesion area is set to 0, and the gray level of the background area is 255;

(2),遍历图像像素,对病变区域像素进行计数后,计算病变区域面积及占比。(2) Traverse the image pixels, count the pixels of the lesion area, and calculate the area and proportion of the lesion area.

进一步地,所述高斯模糊处理利用正态分布计算图像中每个像素的变换,二维空间正态分布方程为:Further, the Gaussian blur processing uses a normal distribution to calculate the transformation of each pixel in the image, and the two-dimensional space normal distribution equation is:

其中,(u,v)是图像像素点二维坐标,r是模糊半径,σ是正态分布的标准偏差。Among them, (u, v) is the two-dimensional coordinates of image pixels, r is the blur radius, and σ is the standard deviation of the normal distribution.

进一步地,所述病变识别模型的训练数据采用以下方法获取:Further, the training data of the lesion recognition model is obtained by the following method:

(1)对全扫描病理组织切片图像进行20×倍放大后进行切分,并从TIFP格式转换为jpeg格式。(1) Segment the full-scan pathological tissue slice image after 20× magnification, and convert from TIFP format to jpeg format.

(2)把所有图像按照分类说明分为正常、肺腺癌、肺鳞癌和肺小细胞癌四类;(2) According to the classification instructions, all images were divided into four categories: normal, lung adenocarcinoma, lung squamous cell carcinoma and lung small cell carcinoma;

(3)对病变区域,应用labelme软件进行手工分割;(3) For the lesion area, apply labelme software to carry out manual segmentation;

(4)将所有数据按比例8:1:1分割成训练集、验证集和测试集。(4) Divide all data into training set, verification set and test set according to the ratio of 8:1:1.

本发明提供的肺癌病理组织切片识别及分割方法,基于改进的Mask R-CNN神经网络,对于提高肺癌的病症诊断准确率具有重要参考意义,其有益的效果体现在以下几个方面:The lung cancer pathological tissue section identification and segmentation method provided by the present invention is based on the improved Mask R-CNN neural network, which has important reference significance for improving the accuracy of lung cancer diagnosis, and its beneficial effects are reflected in the following aspects:

本发明从医学影像数据库中患者的肿瘤病理组织切片图像信息,通过改进后的Mask R-CNN神经网络应用于肺癌病变分割及病症识别中,实现自动学习得到的分类、分割结果,并得到归一化后的病变区域图像及其对应的二值化掩码图。病变特征提取网络提取相关几何形态学特征参数,作为后续定量计算病变面积及比例的参考依据,辅助病理医生提高肺癌病症识别的检出效率并提高肿瘤分化的评估准确率。另外本发明极大减少了临床医师的阅片时间,减缓人工资源紧张的压力。In the present invention, the image information of tumor pathological tissue slices of patients in the medical image database is applied to lung cancer lesion segmentation and disease identification through the improved Mask R-CNN neural network, and the classification and segmentation results obtained by automatic learning are realized, and normalized The transformed lesion area image and its corresponding binarized mask image. The lesion feature extraction network extracts relevant geometric and morphological feature parameters as a reference for subsequent quantitative calculation of lesion area and proportion, assisting pathologists to improve the detection efficiency of lung cancer disease identification and the accuracy of tumor differentiation assessment. In addition, the present invention greatly reduces the reading time of clinicians and relieves the pressure of human resource shortage.

附图说明Description of drawings

图1为本发明的系统结构示意图;Fig. 1 is a schematic diagram of the system structure of the present invention;

图2为改进泛化能力的ResNet结构示意图;Figure 2 is a schematic diagram of the ResNet structure with improved generalization capabilities;

图3为用于病症识别与病变区域分割的神经网络结构示意图;Fig. 3 is a schematic diagram of a neural network structure for disease identification and lesion region segmentation;

图4为计算二值化病变区域面积比例流程图。Fig. 4 is a flow chart for calculating the area ratio of the binarized lesion area.

具体实施方式Detailed ways

为了便于理解本发明,下面结合附图和具体实施例,对本发明进行更详细的说明。附图中给出了本发明的较佳的实施例。但是,本发明可以以许多不同的形式来实现,并不限于本说明书所描述的实施例。相反地,提供这些实施例的目的是使对本发明公开内容的理解更加透彻全面。In order to facilitate the understanding of the present invention, the present invention will be described in more detail below in conjunction with the accompanying drawings and specific embodiments. Preferred embodiments of the invention are shown in the accompanying drawings. However, the present invention can be implemented in many different forms and is not limited to the embodiments described in this specification. On the contrary, these embodiments are provided to make the understanding of the present disclosure more thorough and comprehensive.

本实施例提供的一种基于Mask R-CNN的肺癌病理组织切片病症识别及分割方法,流程如图1所示,包括以下具体步骤:A method for identification and segmentation of pathological tissue slices of lung cancer based on Mask R-CNN provided in this embodiment, the process is shown in Figure 1, including the following specific steps:

一、构建数据集1. Build a data set

从医学影像数据库中获取患者肺癌病理组织切片扫描图像,并对获取的病理组织切片扫描图像进行预处理并进行标注,获得预处理后的图像。The scanned image of the patient's lung cancer pathological tissue slice is obtained from the medical image database, and the acquired pathological tissue slice scanned image is preprocessed and marked to obtain the preprocessed image.

其中,预处理主要包括有以下步骤:Among them, preprocessing mainly includes the following steps:

(1)对全扫描病理组织切片图像进行20倍放大后进行切分,并从TIFP格式转换为jpeg格式。(1) Segment the full-scan pathological tissue slice image after 20 times magnification, and convert from TIFP format to jpeg format.

(2)把所有图像按照文件说明分为正常、肺腺癌、肺鳞癌和肺小细胞癌四类。(2) Divide all images into four categories: normal, lung adenocarcinoma, lung squamous cell carcinoma, and lung small cell carcinoma according to the document description.

(3)对病变区域,病理科医生应用labelme软件进行手工分割。得到的标注信息以json文件保存(3) Pathologists use labelme software to manually segment the lesion area. The obtained annotation information is saved in a json file

(4)将所有数据按比例8:1:1分割训练集、验证集和测试集,用于训练、测试模型。(4) Divide all data into training set, verification set and test set in proportion of 8:1:1 for training and testing models.

标注包括:(1)应用含有标注信息的json文件生成每幅图像的病变区域二值掩码图;(2)标注肺癌分类信息。The labeling includes: (1) applying the json file containing the labeling information to generate a binary mask map of the lesion area of each image; (2) labeling the classification information of lung cancer.

二、病症分类及分割模型的构建及训练2. Construction and training of disease classification and segmentation models

1、特征提取网1. Feature extraction network

特征提取网络选取改进后的残差网络ResNet,削减卷积层数,同时在最后的分类层前增加全连接层及dropout层,提高神经网络的泛化能力,同时在特征提取网络中加入FPN网络,对提取特征进行多尺度融合。The feature extraction network selects the improved residual network ResNet, reduces the number of convolutional layers, and adds a fully connected layer and a dropout layer before the final classification layer to improve the generalization ability of the neural network. At the same time, FPN network is added to the feature extraction network , to perform multi-scale fusion of extracted features.

如图2所示的特征提取卷积网络,其结构为改进的ResNet网络。网络的第一卷积部分由一个卷积层+BatchNorm层+Relu激活层+最大池化层,其中卷积核大小为7x7,最大池化层的卷积核步长为2。网络的第二卷积部分包括3个残差区块ResidualBlock。每个残差区块包括1个1x1的卷积层+1个3x3的卷积层+3个BatchNorm层+3个Relu激活层,对于每个残差区块的第一层特征图,都要对其进行上卷积,保证提取出来的特征图尺寸与第二层卷积特征图尺寸一致。The feature extraction convolutional network shown in Figure 2 is an improved ResNet network. The first convolution part of the network consists of a convolution layer + BatchNorm layer + Relu activation layer + maximum pooling layer, where the convolution kernel size is 7x7, and the convolution kernel step size of the maximum pooling layer is 2. The second convolutional part of the network consists of 3 residual blocks ResidualBlock. Each residual block includes 1 1x1 convolutional layer + 1 3x3 convolutional layer + 3 BatchNorm layers + 3 Relu activation layers. For the first layer feature map of each residual block, you must Perform upconvolution to ensure that the extracted feature map size is consistent with the size of the second layer convolution feature map.

将FPN融合后的特征送入RPN网络中进行目标区域生成,并将其分数值最高的候选区域(数量作为超参数自行设置)输入Mask R-CNN网络利用边框回归操作实现候选框位置精修,得到最终的目标框。The FPN fused features are sent to the RPN network to generate the target area, and the candidate area with the highest score value (the number is set as a hyperparameter by itself) is input into the Mask R-CNN network and the frame regression operation is used to refine the position of the candidate frame. Get the final target box.

2.基于Mask R-CNN的分类分割神经网络,实现病变的分类和病变区域的分割,如图3所示。2. Based on the classification and segmentation neural network of Mask R-CNN, the classification of lesions and the segmentation of lesion areas are realized, as shown in Figure 3.

3.将上述构建的病理组织图像训练集输入到Mask R-CNN神经网络进行训练,经过验证和测试,得到病症分类及分割模型。3. Input the pathological tissue image training set constructed above into the Mask R-CNN neural network for training, after verification and testing, a disease classification and segmentation model is obtained.

三、肺癌病理组织切片病症识别及分割3. Disease identification and segmentation of pathological tissue slices of lung cancer

1、将获取的患者的癌病理组织切片扫描图像,进行20倍放大后进行切分,并从TIFP格式转换为jpeg格式。1. The scanned image of the obtained patient's cancer pathological tissue slice is magnified by 20 times, then segmented, and converted from TIFP format to jpeg format.

2、将上述图像输入病症分类及分割模型2. Input the above image into the disease classification and segmentation model

首先由改进的ResNet网络进行初步卷积,提取图像抽象特征;其次,FPN特征图金字塔网络对多层抽象特征图进行多尺度融合;接着,将FPN融合后的特征送入RPN网络中进行目标区域生成,利用RoI Align在全图特征上摘取每个RoI对应的特征,再通过全连接层进行分类。与全连接层分类平行进行的是分割任务——RoI Align使用双线性插值法:First, the improved ResNet network performs preliminary convolution to extract image abstract features; secondly, the FPN feature map pyramid network performs multi-scale fusion of multi-layer abstract feature maps; then, the FPN fused features are sent to the RPN network for target area Generate, use RoI Align to extract the features corresponding to each RoI on the full image features, and then classify through the fully connected layer. Parallel to fully connected layer classification is the segmentation task - RoI Align uses bilinear interpolation:

其中Q11=(x1,y1)、Q12=(x11,y2)、Q21=(x2,y1)、Q22=(x2,y2)为用于插值定位的四个点。Among them, Q 11 = (x 1 , y 1 ), Q 12 = (x 11 , y 2 ), Q 21 = (x 2 , y 1 ), Q 22 = (x 2 , y 2 ) are used for interpolation positioning four points.

病变区域检测目标结果精炼:获取每个目标推荐区域得分最高的class得分和推荐区域的坐标,删除掉得分最高为背景的推荐区域,剔除掉其中最高得分达不到阈值的推荐区域,对同一类别的候选框进行非极大值抑制NMS,对NMS后的框索引剔除-1占位符,获取前n,最后返回每个框(y1,x1,y2,x2,Class_ID,Score)信息。Lesion area detection target result refinement: Obtain the class score and coordinates of the recommended area with the highest score for each target recommended area, delete the recommended area with the highest score as the background, and eliminate the recommended area whose highest score does not reach the threshold. For the same category Non-maximum value suppression NMS is performed on the candidate frame of NMS, the -1 placeholder is removed from the frame index after NMS, the top n is obtained, and the information of each frame (y1, x1, y2, x2, Class_ID, Score) is returned at last.

病变区域图像的分割Mask生成:获取到目标推荐区域作为输入送入FCN网络输出一个2层的Mask,每层代表不同的类,以log输出并用进行阈值进行二值化,产生背景和病变区域的分割Mask。Segmentation Mask generation of the lesion area image: The target recommended area is obtained as input and sent to the FCN network to output a 2-layer Mask. Each layer represents a different class, and the log output is used to perform binarization with a threshold to generate the background and lesion area. Split Mask.

3.病变区域面积比例的计算3. Calculation of the proportion of lesion area

(1)对上述获得的“背景和病变区域的分割Mask”图像进行高斯模糊处理,设置病变区域的灰度为0,背景区域的灰度为255。(1) Perform Gaussian blur processing on the image of "segmentation mask of background and lesion area" obtained above, set the gray level of the lesion area to 0, and set the gray level of the background area to 255.

高斯模糊是一种图像模糊滤波器,它用正态分布计算图像中每个像素的变换。Gaussian blur is an image blur filter that computes the transformation of each pixel in an image using a normal distribution.

二维空间正态分布方程为:The two-dimensional space normal distribution equation is:

其中(u,v)是图像像素点二维坐标,r是模糊半径,σ是正态分布的标准偏差。在二维图像中,符合该公式的从中心开始呈正态分布的等高线同心圆,与图像原始对应像素做卷积,卷积后的每个像素值都是周围相邻像素值的加权平均。Where (u,v) is the two-dimensional coordinates of image pixels, r is the blur radius, and σ is the standard deviation of the normal distribution. In a two-dimensional image, the contour line concentric circles that conform to the formula and are normally distributed from the center are convolved with the original corresponding pixels of the image, and each pixel value after convolution is the weight of the surrounding adjacent pixel values average.

(2)遍历图像像素,对病变区域像素进行计数后,计算病变区域面积及占比。(2) Traverse the image pixels, count the pixels of the lesion area, and calculate the area and proportion of the lesion area.

灰度图像经二值化处理,处于病变区域的像素值为0,背景区域的像素值为255。病变像素点计数变量count初始设为0,利用循环语句,遍历图像全部像素,对每个像素点进行灰度值判断,流程如图4所示,具体为:The grayscale image was binarized, the pixel value in the lesion area was 0, and the pixel value in the background area was 255. The lesion pixel count variable count is initially set to 0. Use a loop statement to traverse all pixels in the image and judge the gray value of each pixel. The process is shown in Figure 4, specifically:

设对于(h_x,w_x)处的像素点:Suppose for the pixel at (h_x,w_x):

当像素值value(h_x,w_x)为0,count=count+1;否则不计数。When the pixel value value(h_x, w_x) is 0, count=count+1; otherwise, it does not count.

最后求得病变比例proportion:Finally, the lesion ratio ratio is obtained:

proportion=count/pic_shapeproportion=count/pic_shape

其中pic_shape为图像尺寸。Where pic_shape is the image size.

Claims (7)

1. A Mask R-CNN-based lung cancer pathological tissue section identification and segmentation method is characterized by comprising the following steps of: the method comprises the following steps:
s1, acquiring a lung cancer pathological tissue section scanning image of a patient, and preprocessing;
s2, inputting the preprocessed scanning image into a pre-trained disease classification and segmentation model, judging the type of the slice disease, and obtaining a visual activation diagram for lesion region segmentation; the disease classification and segmentation model is an improved Mask R-CNN neural network;
s3, calculating the area proportion of the lesion area to the global pathological tissue section according to the acquired visual activation diagram.
2. The Mask R-CNN-based lung cancer pathological tissue section identification and segmentation method according to claim 1, wherein the method comprises the following steps: in the step S1, the specific method for preprocessing is as follows:
the scanned image of lung cancer pathological tissue section of patient is magnified 20 times and converted into jpeg format.
3. The Mask R-CNN-based lung cancer pathological tissue section identification and segmentation method according to claim 1, wherein the method comprises the following steps: the improved Mask R-CNN neural network specifically comprises:
a feature extraction network comprising an improved residual network, res net; the improved residual error network ResNet is added with a full connection layer and a dropout layer before the last classification layer;
the FPN network is added into the feature extraction network, and multiscale fusion is carried out on the extracted features;
the RPN network is used for generating target areas of the features after the FPN fusion and inputting a set number of candidate areas with the highest score values into the Mask R-CNN network;
and (3) classifying the input candidate areas by using a Mask R-CNN network, and dividing the lesion areas to generate a segmentation Mask of the background and the lesion areas.
4. The Mask R-CNN-based lung cancer pathological tissue section recognition and segmentation method according to claim 3, wherein: in the step S3, the area ratio calculation method of the lesion area to the global pathological tissue section comprises the following steps:
(1) Performing Gaussian blur processing on the obtained visual activation image, setting the gray level of a lesion area to be 0 and setting the gray level of a background area to be 255;
(2) And traversing the image pixels, counting the pixels of the lesion area, and calculating the area and the duty ratio of the lesion area.
5. The Mask R-CNN-based lung cancer pathological tissue section identification and segmentation method according to claim 4, wherein the method comprises the following steps of: the Gaussian blur processing calculates the transformation of each pixel in the image by using normal distribution, and a two-dimensional space normal distribution equation is as follows:
where (u, v) is the two-dimensional coordinates of the image pixel point, r is the blur radius, and σ is the standard deviation of the normal distribution.
6. The Mask R-CNN-based lung cancer pathological tissue section recognition and segmentation method according to any one of claims 1 to 5, wherein: training data of the lesion recognition model is obtained by adopting the following method:
(1) The full-scanning pathological tissue slice image is segmented after 20 times magnification, and is converted from a TIFP format to a jpeg format;
(2) Classifying all images into four types of normal lung adenocarcinoma, lung squamous carcinoma and lung small cell carcinoma according to classification;
(3) Manually segmenting a lesion area by using labelme software;
(4) All data are scaled 8:1:1 into a training set, a validation set and a test set.
7. The Mask R-CNN-based lung cancer pathological tissue section identification and segmentation method according to claim 6, wherein the method comprises the following steps: the preprocessed image also needs to be marked, including: (1) Generating a binary mask map of a lesion area of each image by applying a json file containing labeling information; and (2) marking lung cancer classification information.
CN202210051013.6A 2022-01-17 2022-01-17 A Method for Recognition and Segmentation of Lung Cancer Pathological Tissue Slices Based on Mask R-CNN Pending CN116524315A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202210051013.6A CN116524315A (en) 2022-01-17 2022-01-17 A Method for Recognition and Segmentation of Lung Cancer Pathological Tissue Slices Based on Mask R-CNN

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202210051013.6A CN116524315A (en) 2022-01-17 2022-01-17 A Method for Recognition and Segmentation of Lung Cancer Pathological Tissue Slices Based on Mask R-CNN

Publications (1)

Publication Number Publication Date
CN116524315A true CN116524315A (en) 2023-08-01

Family

ID=87403368

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202210051013.6A Pending CN116524315A (en) 2022-01-17 2022-01-17 A Method for Recognition and Segmentation of Lung Cancer Pathological Tissue Slices Based on Mask R-CNN

Country Status (1)

Country Link
CN (1) CN116524315A (en)

Cited By (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN116883397A (en) * 2023-09-06 2023-10-13 佳木斯大学 Automatic lean method and system applied to anatomic pathology
CN118154975A (en) * 2024-03-27 2024-06-07 广州市中西医结合医院 Tumor pathological diagnosis image classification method based on big data
CN118299044A (en) * 2024-04-25 2024-07-05 山东大学 Self-explanatory medical image diagnosis system based on class activation diagram

Cited By (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN116883397A (en) * 2023-09-06 2023-10-13 佳木斯大学 Automatic lean method and system applied to anatomic pathology
CN116883397B (en) * 2023-09-06 2023-12-08 佳木斯大学 Automatic lean method and system applied to anatomic pathology
CN118154975A (en) * 2024-03-27 2024-06-07 广州市中西医结合医院 Tumor pathological diagnosis image classification method based on big data
CN118154975B (en) * 2024-03-27 2024-10-01 广州市中西医结合医院 Tumor pathological diagnosis image classification method based on big data
CN118299044A (en) * 2024-04-25 2024-07-05 山东大学 Self-explanatory medical image diagnosis system based on class activation diagram

Similar Documents

Publication Publication Date Title
CN112529894B (en) Thyroid nodule diagnosis method based on deep learning network
CN108464840B (en) Automatic detection method and system for breast lumps
CN111028206A (en) Prostate cancer automatic detection and classification system based on deep learning
CN104751178B (en) Lung neoplasm detection means and method based on shape template matching combining classification device
CN112001218B (en) A three-dimensional particle category detection method and system based on convolutional neural network
CN113223005B (en) Thyroid nodule automatic segmentation and grading intelligent system
CN116524315A (en) A Method for Recognition and Segmentation of Lung Cancer Pathological Tissue Slices Based on Mask R-CNN
CN109300113B (en) Pulmonary nodule auxiliary detection system and method based on improved convex hull method
CN107798679A (en) Breast molybdenum target image breast area is split and tufa formation method
CN101403743B (en) Automatic separating method for X type overlapping and adhering chromosome
CN113240623B (en) Pavement disease detection method and device
CN107480677A (en) The method and device of area-of-interest in a kind of identification three-dimensional CT image
CN108062749B (en) Recognition method, device and electronic device for levator hiatus
CN112270667B (en) TI-RADS-based integrated deep learning multi-tag identification method
CN105913432A (en) Aorta extracting method and aorta extracting device based on CT sequence image
CN112263217B (en) A Lesion Area Detection Method Based on Improved Convolutional Neural Network in Pathological Images of Non-melanoma Skin Cancer
CN112365973B (en) Pulmonary nodule auxiliary diagnosis system based on countermeasure network and fast R-CNN
CN110555835A (en) brain slice image region division method and device
CN114549446A (en) A detection method for cylinder liner defect marks based on deep learning
CN114004813A (en) A method and device for clinical target identification in cervical cancer radiotherapy
CN113658151A (en) Mammary gland lesion magnetic resonance image classification method and device and readable storage medium
CN116758336A (en) Medical image intelligent analysis system based on artificial intelligence
CN116433976A (en) Image processing method, device, equipment and storage medium
CN113379691B (en) A deep learning segmentation method of breast lesions based on prior guidance
CN114693622A (en) Plaque erosion automatic detection system based on artificial intelligence

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination