CN111047589B - Attention-enhanced brain tumor auxiliary intelligent detection and identification method - Google Patents
Attention-enhanced brain tumor auxiliary intelligent detection and identification method Download PDFInfo
- Publication number
- CN111047589B CN111047589B CN201911393654.4A CN201911393654A CN111047589B CN 111047589 B CN111047589 B CN 111047589B CN 201911393654 A CN201911393654 A CN 201911393654A CN 111047589 B CN111047589 B CN 111047589B
- Authority
- CN
- China
- Prior art keywords
- model
- classification
- segmentation
- convolution
- result
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Active
Links
- 238000000034 method Methods 0.000 title claims abstract description 32
- 208000003174 Brain Neoplasms Diseases 0.000 title claims abstract description 26
- 238000001514 detection method Methods 0.000 title claims abstract description 11
- 230000011218 segmentation Effects 0.000 claims abstract description 43
- 238000012549 training Methods 0.000 claims abstract description 34
- 206010028980 Neoplasm Diseases 0.000 claims description 37
- 238000003745 diagnosis Methods 0.000 claims description 17
- 210000004556 brain Anatomy 0.000 claims description 16
- 238000005070 sampling Methods 0.000 claims description 13
- 238000003709 image segmentation Methods 0.000 claims description 11
- 238000011176 pooling Methods 0.000 claims description 11
- 238000013145 classification model Methods 0.000 claims description 8
- 230000008569 process Effects 0.000 claims description 8
- 230000003902 lesion Effects 0.000 claims description 7
- 206010003571 Astrocytoma Diseases 0.000 claims description 6
- 206010030113 Oedema Diseases 0.000 claims description 6
- 201000010133 Oligodendroglioma Diseases 0.000 claims description 6
- 230000000694 effects Effects 0.000 claims description 6
- 230000017074 necrotic cell death Effects 0.000 claims description 6
- 238000004364 calculation method Methods 0.000 claims description 5
- 238000002372 labelling Methods 0.000 claims description 5
- 238000010827 pathological analysis Methods 0.000 claims description 5
- 238000012545 processing Methods 0.000 claims description 5
- 238000013528 artificial neural network Methods 0.000 claims description 4
- 201000007983 brain glioma Diseases 0.000 claims description 4
- 238000011156 evaluation Methods 0.000 claims description 4
- 206010073128 Anaplastic oligodendroglioma Diseases 0.000 claims description 3
- 230000004913 activation Effects 0.000 claims description 3
- 206010002224 anaplastic astrocytoma Diseases 0.000 claims description 3
- 208000005017 glioblastoma Diseases 0.000 claims description 3
- 238000003672 processing method Methods 0.000 claims description 3
- 238000007747 plating Methods 0.000 claims 1
- 238000013461 design Methods 0.000 abstract description 4
- 230000007246 mechanism Effects 0.000 abstract description 2
- 230000003631 expected effect Effects 0.000 abstract 1
- 238000005259 measurement Methods 0.000 abstract 1
- 230000006870 function Effects 0.000 description 21
- 238000002595 magnetic resonance imaging Methods 0.000 description 20
- 206010018338 Glioma Diseases 0.000 description 9
- 238000007917 intracranial administration Methods 0.000 description 6
- 208000032612 Glial tumor Diseases 0.000 description 4
- 201000011510 cancer Diseases 0.000 description 4
- 238000013135 deep learning Methods 0.000 description 3
- 201000010099 disease Diseases 0.000 description 3
- 208000037265 diseases, disorders, signs and symptoms Diseases 0.000 description 3
- 238000005516 engineering process Methods 0.000 description 3
- 230000001575 pathological effect Effects 0.000 description 3
- 238000011160 research Methods 0.000 description 3
- 238000002059 diagnostic imaging Methods 0.000 description 2
- 238000003384 imaging method Methods 0.000 description 2
- 230000036210 malignancy Effects 0.000 description 2
- 238000004393 prognosis Methods 0.000 description 2
- 238000002271 resection Methods 0.000 description 2
- 238000012360 testing method Methods 0.000 description 2
- 206010064571 Gene mutation Diseases 0.000 description 1
- 208000005234 Granulosa Cell Tumor Diseases 0.000 description 1
- UFHFLCQGNIYNRP-UHFFFAOYSA-N Hydrogen Chemical compound [H][H] UFHFLCQGNIYNRP-UHFFFAOYSA-N 0.000 description 1
- 230000002159 abnormal effect Effects 0.000 description 1
- 230000009471 action Effects 0.000 description 1
- 238000004458 analytical method Methods 0.000 description 1
- 238000004195 computer-aided diagnosis Methods 0.000 description 1
- 238000013527 convolutional neural network Methods 0.000 description 1
- 238000011161 development Methods 0.000 description 1
- 230000018109 developmental process Effects 0.000 description 1
- 238000011207 functional examination Methods 0.000 description 1
- 238000002599 functional magnetic resonance imaging Methods 0.000 description 1
- 230000004927 fusion Effects 0.000 description 1
- 210000004602 germ cell Anatomy 0.000 description 1
- 239000003292 glue Substances 0.000 description 1
- 230000003394 haemopoietic effect Effects 0.000 description 1
- 208000029824 high grade glioma Diseases 0.000 description 1
- 229910052739 hydrogen Inorganic materials 0.000 description 1
- 239000001257 hydrogen Substances 0.000 description 1
- 125000004435 hydrogen atom Chemical group [H]* 0.000 description 1
- 230000003993 interaction Effects 0.000 description 1
- 208000030173 low grade glioma Diseases 0.000 description 1
- 238000010801 machine learning Methods 0.000 description 1
- 201000011614 malignant glioma Diseases 0.000 description 1
- 210000002418 meninge Anatomy 0.000 description 1
- 206010061289 metastatic neoplasm Diseases 0.000 description 1
- 206010061311 nervous system neoplasm Diseases 0.000 description 1
- 238000002610 neuroimaging Methods 0.000 description 1
- 238000010606 normalization Methods 0.000 description 1
- 238000005457 optimization Methods 0.000 description 1
- 210000000056 organ Anatomy 0.000 description 1
- 210000000578 peripheral nerve Anatomy 0.000 description 1
- 230000002980 postoperative effect Effects 0.000 description 1
- 238000007781 pre-processing Methods 0.000 description 1
- 239000013589 supplement Substances 0.000 description 1
- 230000001629 suppression Effects 0.000 description 1
- 238000001356 surgical procedure Methods 0.000 description 1
- XLYOFNOQVPJJNP-UHFFFAOYSA-N water Substances O XLYOFNOQVPJJNP-UHFFFAOYSA-N 0.000 description 1
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
- G06T7/0002—Inspection of images, e.g. flaw detection
- G06T7/0012—Biomedical image inspection
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/04—Architecture, e.g. interconnection topology
- G06N3/045—Combinations of networks
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/08—Learning methods
- G06N3/084—Backpropagation, e.g. using gradient descent
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
- G06T7/10—Segmentation; Edge detection
- G06T7/11—Region-based segmentation
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/10—Image acquisition modality
- G06T2207/10072—Tomographic images
- G06T2207/10088—Magnetic resonance imaging [MRI]
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/30—Subject of image; Context of image processing
- G06T2207/30004—Biomedical image processing
- G06T2207/30016—Brain
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/30—Subject of image; Context of image processing
- G06T2207/30004—Biomedical image processing
- G06T2207/30096—Tumor; Lesion
Landscapes
- Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- Physics & Mathematics (AREA)
- General Physics & Mathematics (AREA)
- General Health & Medical Sciences (AREA)
- Health & Medical Sciences (AREA)
- Computing Systems (AREA)
- Mathematical Physics (AREA)
- Data Mining & Analysis (AREA)
- Evolutionary Computation (AREA)
- Biophysics (AREA)
- Molecular Biology (AREA)
- Biomedical Technology (AREA)
- General Engineering & Computer Science (AREA)
- Artificial Intelligence (AREA)
- Computational Linguistics (AREA)
- Software Systems (AREA)
- Life Sciences & Earth Sciences (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Medical Informatics (AREA)
- Nuclear Medicine, Radiotherapy & Molecular Imaging (AREA)
- Radiology & Medical Imaging (AREA)
- Quality & Reliability (AREA)
- Magnetic Resonance Imaging Apparatus (AREA)
Abstract
Description
技术领域technical field
本发明涉及图像处理领域,尤其涉及医学影像领域和计算机辅助诊断领域一种注意力增强的脑肿瘤辅助智能检测识别方法。The invention relates to the field of image processing, in particular to an attention-enhanced brain tumor-assisted intelligent detection and identification method in the field of medical imaging and computer-aided diagnosis.
背景技术Background technique
生长于颅内的肿瘤统称为脑瘤,是指发生于颅腔内的神经系统肿瘤,包括起源于神经上皮、外周神经、脑膜和生殖细胞的肿瘤,淋巴和造血组织肿瘤,蝶鞍区的颅咽管瘤与颗粒细胞瘤,以及转移性肿瘤。肿瘤发生自脑实质的称为原发性颅内肿瘤,由身体其它脏器组织的恶性肿瘤转移至颅内的称为继发性颅内肿瘤。颅内肿瘤可发生于任何年龄,以20-50岁为最多见。近年来随着神经影像学技术和功能性检查技术的发展,辅助检查已成为诊断颅内肿瘤的主要手段。Tumors that grow in the brain are collectively referred to as brain tumors, which refer to tumors of the nervous system that occur in the cranial cavity, including tumors originating from neuroepithelial, peripheral nerve, meninges, and germ cells, lymphoid and hematopoietic tissue tumors, and craniopharynx in the sella region. Angiomas and granulosa cell tumors, and metastatic tumors. Tumors that arise from the brain parenchyma are called primary intracranial tumors, and those that metastasize to the brain from malignant tumors of other organs and tissues of the body are called secondary intracranial tumors. Intracranial tumors can occur at any age, with 20-50 years being the most common. In recent years, with the development of neuroimaging technology and functional examination technology, auxiliary examination has become the main means of diagnosing intracranial tumors.
脑胶质瘤是最常见的颅内原发恶性肿瘤,占75%以上。脑胶质瘤分为局限性胶质瘤和弥散性胶质瘤,根据肿瘤恶性程度可分为WHO I-IV级,随级别增加,恶性程度增高。脑胶质瘤可分为低级别胶质瘤(LGG,WHO I-II级)和高级别胶质瘤(HGG,WHO III-IV级),根据基因突变、染色体改变等不同可分为不同亚型,不同级别、不同基因突变的脑胶质瘤的治疗方式和预后存在差异。因此如果能够在进行手术治疗前对肿瘤区域、肿瘤级别进行准确的分割与判断,将有助于指导治疗方案和手术切除区域的选择,对提高患者治疗效果、改善预后有着重要价值。Glioma is the most common primary intracranial malignant tumor, accounting for more than 75%. Brain gliomas are divided into localized gliomas and diffuse gliomas, which can be classified into WHO grades I-IV according to the degree of tumor malignancy. Brain gliomas can be divided into low-grade gliomas (LGG, WHO grades I-II) and high-grade gliomas (HGG, WHO grades III-IV). There are differences in the treatment methods and prognosis of gliomas with different grades and gene mutations. Therefore, if the tumor area and tumor grade can be accurately segmented and judged before surgical treatment, it will help to guide the treatment plan and the selection of surgical resection area, which is of great value to improve the treatment effect and prognosis of patients.
磁共振成像(MRI)是利用强外磁场内人体的氢原子核即氢质子,在特定射频脉冲作用下产生磁共振现象,所进行的一种医学成像技术。MRI是颅内各种疾病的主要影像检查技术,并可作为一些疾病的首选检查方法,也是CT检查的重要补充。MRI检查具有组织分辨率高、多序列、多参数、多方位和多种fMRI检查等优势,能够更敏感地发现病变并显示病变特征,从而有利于疾病的早期检出和准确诊断。常见的用于脑胶质瘤诊断的MRI图像包括轴状面(Axis)、矢状面(Sagittal)和冠状面(Coronal)三个平面和T1、T1增强、T2、T2水抑制四个模态的图像,临床上常结合以上三个平面和四个模态的信息来综合确定肿瘤的位置、范围和级别。但由于脑肿瘤外观和形状的多变,多模态MRI图像中脑肿瘤的分隔是医学图像处理中最具挑战性和难度的人物之一,而基于脑MRI等无创检查对脑胶质瘤进行分级、甚至于是基因型的判断和分类是临床上非常关注的研究方向。Magnetic Resonance Imaging (MRI) is a medical imaging technology that uses the hydrogen nuclei of the human body, namely hydrogen protons, in a strong external magnetic field to generate magnetic resonance phenomena under the action of specific radio frequency pulses. MRI is the main imaging technique for various intracranial diseases, and it can be used as the first-choice examination method for some diseases, and it is also an important supplement to CT examination. MRI examination has the advantages of high tissue resolution, multi-sequence, multi-parameter, multi-directional and various fMRI examinations. Common MRI images used for the diagnosis of glioma include three planes of axial, sagittal and coronal planes and four modalities of T1, T1 enhancement, T2 and T2 water suppression. In clinical practice, the information of the above three planes and four modalities is often combined to comprehensively determine the location, extent and grade of the tumor. However, due to the changeable appearance and shape of brain tumors, the segmentation of brain tumors in multimodal MRI images is one of the most challenging and difficult figures in medical image processing. Grading, and even the judgment and classification of genotypes, are the research directions of great concern in clinical practice.
局限性胶质瘤多见于儿童,成人相对较少,多数患者经手术全切可治愈,恶性程度较低不作为本专利的研究重点,本专利通过以基于3D U-Net的卷积神经网络对MRI图像进行分割,并通过分割任务作为注意力增强机制对弥散性脑胶质瘤进行辅助检测,对肿瘤进行分级和分类。Localized gliomas are more common in children and relatively rare in adults. Most patients can be cured by complete resection. The low degree of malignancy is not the research focus of this patent. MRI images were segmented, and a segmentation task was used as an attention-enhancing mechanism to aid in the detection of diffuse gliomas to grade and classify tumors.
目前的深度学习在脑MRI图像上的应用,尤其是其它相关专利,集中于脑肿瘤分割领域,对脑肿瘤的分类、分级等与诊断、治疗直接相关的领域的涉及较少,而这是临床上更为关注并且目前人眼在无创的影像检查难以做到的。目前在脑肿瘤分割任务上得到较好效果的方法大多采用U-Net作为基础框架,在此基础上进行改进,本申请技术方案也以U-Net作为基础,使用三维模型对三维图像进行操作,并在原始U-Net的基础上参照目前机器学习和深度学习在计算机视觉领域的最新进展进行优化以达到更优效果,同时与其它工作相比更多的关注于肿瘤分级、分类诊断层面,通过分类任务作为注意力增强方式,关注脑MRI影像中的异常区域实现对肿瘤的诊断任务。The current application of deep learning on brain MRI images, especially other related patents, focuses on the field of brain tumor segmentation, and the classification and grading of brain tumors are less involved in the fields directly related to diagnosis and treatment, which are clinical It is more concerned about and currently the human eye is difficult to achieve in non-invasive imaging examinations. At present, most of the methods that have achieved good results in brain tumor segmentation tasks use U-Net as the basic framework, and make improvements on this basis. On the basis of the original U-Net, it is optimized with reference to the latest advances in machine learning and deep learning in the field of computer vision to achieve better results. At the same time, compared with other work, it focuses more on tumor grading, classification and diagnosis. As an attention enhancement method, the classification task focuses on abnormal regions in brain MRI images to realize the task of diagnosing tumors.
发明内容SUMMARY OF THE INVENTION
目前将U-Net应用到脑肿瘤MRI图像处理上的多是进行图像分割任务,这些应用只是分割出肿瘤的所在区域,没有进一步使用分割之后的图像得到有价值的医学层面的信息,而这是我们在与临床医生沟通的过程中他们更为关注、需要我们着重研究的问题。At present, most of the applications of U-Net to brain tumor MRI image processing are image segmentation tasks. These applications only segment the region where the tumor is located, and do not further use the segmented images to obtain valuable medical information. In the process of communicating with clinicians, they are more concerned about the issues that we need to focus on.
为达到上述目的,本发明采用了下列技术方案:To achieve the above object, the present invention has adopted the following technical solutions:
一种注意力增强的脑肿瘤辅助智能检测识别方法,其特征在于:包括;An attention-enhanced brain tumor-assisted intelligent detection and identification method, characterized by: comprising;
步骤一:建立基于U-Net的适用于脑部MRI图像中脑胶质瘤病灶区域的分割与诊断的多任务神经网络的三维卷积网络模型;Step 1: Establish a three-dimensional convolutional network model of a multi-task neural network based on U-Net, which is suitable for segmentation and diagnosis of glioma lesions in brain MRI images;
步骤二:多任务联合训练目标;Step 2: Multi-task joint training target;
步骤三:衡量多任务的损失,优化结果;Step 3: Measure the loss of multi-tasking and optimize the results;
步骤四:模型训练与结果评估和输出。Step 4: Model training and result evaluation and output.
所述建立基于U-Net的适用于脑部MRI图像中脑胶质瘤病灶区域的分割与诊断的多任务神经网络的三维卷积网络模型步骤包括:The steps of establishing a U-Net-based 3D convolutional network model of a multi-task neural network suitable for segmentation and diagnosis of glioma lesions in brain MRI images include:
基于原始U-Net网络,使用三维卷积处理方法,搭建基于肿瘤区域作为模型的注意力区域的三维模型,模型的架构包含一个下采样的数据通路和一个上采样的数据通路,下采样路径上每一层包含两个3×3×3的卷积层,并且每一个卷积层均采用了dropout的方式以防止过拟合,使用ReLU的激活函数。在两个卷积层后,使用一个步长为2、大小为3×3×3的最大池化层进行下采样操作;上采样数据通路的两层之间使用解卷积的方式进行上采样操作,并将上采样后的特征与左侧下采样层的特征进行拼接,再将拼接后的特征进行两次与左侧上采样层相同的卷积操作,上采样通路最后得到融合了深层与浅层信息的特征之后,进行两次卷积核大小为3×3×3的卷积,得到经过卷积的特征后,模型分为两支:一支继续进行一次卷积操作,输出通道与语义分割结果中的分类相同,然后经过softmax计算,得到包含背景、水肿、肿瘤实质、坏死和增强核的输出结果;另一支进行一次卷积操作后,使用全局平均池化(Global Average Pooling),然后紧跟两个全连接层,第二个全连接层的输出与病理诊断分类数相同;Based on the original U-Net network, a three-dimensional convolution processing method is used to build a three-dimensional model of the attention region based on the tumor area as a model. The architecture of the model includes a down-sampling data path and an up-sampling data path. Each layer contains two 3×3×3 convolutional layers, and each convolutional layer adopts a dropout method to prevent overfitting, using the activation function of ReLU. After two convolutional layers, a max pooling layer with stride 2 and size 3×3×3 is used for downsampling; deconvolution is used for upsampling between the two layers of the upsampling datapath operation, splicing the upsampled features with the features of the left downsampling layer, and then performing the same convolution operation as the left upsampling layer twice on the spliced features. After the features of the shallow information, two convolutions with the convolution kernel size of 3 × 3 × 3 are performed, and after the convolutional features are obtained, the model is divided into two branches: one continues to perform a convolution operation, and the output channel is the same as the one. The classification in the semantic segmentation results is the same, and then after softmax calculation, the output results including background, edema, tumor parenchyma, necrosis and enhancement kernel are obtained; the other branch performs a convolution operation and uses Global Average Pooling (Global Average Pooling) , followed by two fully connected layers, the output of the second fully connected layer is the same as the number of pathological diagnosis classifications;
之后遍历所有从病例导入的脑部MRI图像,统计平均值和方差信息,留作训练和预测过程中的标准化操作使用,接收输入的T1、T1增强、T2、T2-Flair四个模态的脑MRI图像序列,每个模态作为一个通道,然后将该扫描序列下的所有切片图像与分割结果标注遮罩分别进行拼接形成三维图片和标注序列,再将这两个序列与该序列对应的诊断结果进行绑定作为一个样例,对所有图像进行处理。Then traverse all the brain MRI images imported from the cases, count the mean and variance information, and reserve it for the standardization operation in the training and prediction process, and receive the input of the four modalities of the brain: T1, T1 enhanced, T2, and T2-Flair. MRI image sequence, each modality is used as a channel, then all slice images under the scanning sequence and the segmentation result labeling mask are spliced respectively to form a 3D picture and labeling sequence, and then these two sequences are combined with the corresponding diagnosis of the sequence. The result is bound as a sample, and all images are processed.
所述多任务联合训练目标步骤包括:The multi-task joint training target steps include:
在全卷积模型上,融合浅层与深层信息后增加分类分支,使用单一任务分割和分类模型,同时得到语义分割与分类结果,使得脑肿瘤的分割任务和分类任务能够同时执行并且由于共用浅层特征;On the fully convolutional model, a classification branch is added after merging shallow and deep information, and a single task segmentation and classification model is used to obtain semantic segmentation and classification results at the same time, so that the segmentation task and classification task of brain tumors can be performed at the same time. layer features;
得到融合了深层与浅层信息的特征之后,进行两次卷积核大小为3×3×3的卷积,得到经过卷积的特征后,模型分为两支:一支继续进行一次卷积操作,输出通道与语义分割结果中的分类相同,然后经过softmax计算,得到包含背景、水肿、肿瘤实质、坏死和增强核的输出结果;另一支进行一次卷积操作后,使用全局平均池化(Global Average Pooling),然后紧跟两个全连接层,第二个全连接层的输出与病理诊断分类数相同,包含以下类别:少突胶质细胞瘤、间变期少突胶质细胞瘤、星形胶质细胞瘤、间变期星形胶质细胞瘤、胶质母细胞瘤。After obtaining the features that integrate the deep and shallow information, perform two convolutions with a convolution kernel size of 3 × 3 × 3. After obtaining the convolved features, the model is divided into two branches: one continues to perform a convolution Operation, the output channel is the same as the classification in the semantic segmentation result, and then after softmax calculation, the output results including background, edema, tumor parenchyma, necrosis and enhancement kernel are obtained; after another convolution operation, the global average pooling is used (Global Average Pooling), followed by two fully connected layers, the output of the second fully connected layer is the same as the number of pathological diagnosis classifications, including the following categories: oligodendroglioma, anaplastic oligodendroglioma , astrocytoma, anaplastic astrocytoma, glioblastoma.
所述衡量多任务的损失,优化结果步骤使用多任务联合的损失函数,对分割结果与分类结果进行衡量,优化分割与分类结果,其中:In the step of measuring the loss of multiple tasks, the optimization result step uses the loss function of the multi-task combination to measure the segmentation result and the classification result, and optimize the segmentation and classification results, wherein:
图像分割模型的损失函数采用Dice损失函数;The loss function of the image segmentation model adopts the Dice loss function;
肿瘤分类模块的损失函数选择交叉熵函数。The loss function of the tumor classification module selects the cross-entropy function.
所述模型训练与结果评估和输出步骤包括:The model training and result evaluation and output steps include:
模型训练步骤,设置只允许图像分割模型的Dice loss或者肿瘤分类模型的交叉熵损失函数进行反向传播进行迭代训练,将图像分割模块的loss值与肿瘤分类模型的loss值以一定的比例组合为之后进行反向传播进行迭代训练;The model training step is set to allow only the Dice loss of the image segmentation model or the cross-entropy loss function of the tumor classification model to perform back-propagation for iterative training, and combine the loss value of the image segmentation module with the loss value of the tumor classification model in a certain proportion as Then perform backpropagation for iterative training;
在训练效果收敛且得到较为理想的结果之后,在训练集上进行效果的评估;After the training effect converges and a relatively ideal result is obtained, the effect is evaluated on the training set;
对经过评估的模型输入新的病例图像,并将检测识别结果输出。A new case image is input to the evaluated model, and the detection and recognition results are output.
本发明相对于现有技术的优点在于:The advantages of the present invention relative to the prior art are:
本设计方案充分利用分割的信息,针对分割后图像中隐含的医学信息,扩展分类任务模块,从而在脑肿瘤的分类、分级等方面发挥作为医生的辅助医疗系统在诊断过程中提供建议、提高医疗机构对与相关疾病的诊断能力、反哺医学领域对脑肿瘤的病理研究的作用。This design scheme makes full use of the segmentation information, expands the classification task module for the medical information hidden in the segmented images, so as to play a role as a doctor's auxiliary medical system in the classification and grading of brain tumors, and provide suggestions and improve the diagnosis process. The ability of medical institutions to diagnose related diseases, and the role of feeding back the pathological research of brain tumors in the medical field.
并且目前将U-Net应用到脑肿瘤MRI图像处理上的多是进行图像分割任务,这些应用只是分割出肿瘤的所在区域,没有进一步使用分割之后的图像得到有价值的医学层面的信息,我们通过对图像的处理与分析,同时得到了病灶区域的分割与诊断结果。And at present, most of the applications of U-Net to brain tumor MRI image processing are image segmentation tasks. These applications only segment the region where the tumor is located, and do not further use the segmented images to obtain valuable medical-level information. The images are processed and analyzed, and the segmentation and diagnosis results of the lesion area are obtained at the same time.
附图说明Description of drawings
图1脑肿瘤辅助检测识别系统设计框架;Figure 1 Design framework of brain tumor auxiliary detection and recognition system;
图2脑肿瘤分割与分类诊断任务的多任务学习模型;Figure 2. Multi-task learning model for brain tumor segmentation and classification diagnosis tasks;
具体实施方式Detailed ways
参见说明书附图1-2,本发明提出一种将医学图像与深度学习、计算机视觉方法结合,使用计算机视觉的分析方法,对三维脑核磁共振图像进行分析和处理,在脑核磁图像上进行胶质瘤病灶区域的分割和基于图像进行的分类诊断任务。针对医学图像数据集数据量小、类别不平衡严重、现有方法集中于进行病灶区域分割而忽略了分类诊断任务的问题,提出一种基于改进的3D U-Net卷积神经网络,增加分类诊断分支,通过多任务联合训练的方式,同时获得分割与分类结果。图1是本发明提出的算法设计流程,首先对MRI图像以及与其对应的人工标注的分割结果和从病理信息中获取的分类诊断信息进行预处理,得到包含四个模态的三维图像序列,将处理后的图像按比例分为训练集和测试集,在训练集上进行训练,优化模型在分割与分类两个任务上的表现,最后在训练效果收敛且得到较为理想的结果之后,在训练集上进行效果的评估。Referring to the accompanying drawings 1-2 of the description, the present invention proposes a method that combines medical images with deep learning and computer vision methods, and uses an analysis method of computer vision to analyze and process three-dimensional brain MRI images, and perform glue on the brain MRI images. Segmentation and image-based classification and diagnosis tasks of plasmoma lesion regions. Aiming at the problems of small amount of data in medical image datasets, serious category imbalance, and existing methods focus on segmentation of lesion areas while ignoring the task of classification and diagnosis, an improved 3D U-Net convolutional neural network is proposed to increase classification and diagnosis. Branch, through the multi-task joint training, the segmentation and classification results are obtained at the same time. Fig. 1 is the algorithm design flow proposed by the present invention. First, preprocess the MRI image and its corresponding manually labeled segmentation results and the classification and diagnosis information obtained from the pathological information to obtain a three-dimensional image sequence including four modalities. The processed images are divided into a training set and a test set proportionally, and the training is carried out on the training set to optimize the performance of the model on the two tasks of segmentation and classification. Evaluate the effect above.
再建立模型前,我们使用的脑MRI影像数据直接来自于医院的病历系统,已经完成去噪、亮度对比度调节等增强可见性的图像预处理操作,并人工对图片数据进行了标注,对肿瘤的不同部分,包括水肿、肿瘤实质、增强核、坏死四个部分进行了标注,并从病例的病历诊断和术后病理信息中得到了肿瘤类别、分期等诊断信息。Before establishing the model, the brain MRI image data we used came directly from the hospital's medical record system. Image preprocessing operations such as denoising, brightness and contrast adjustment to enhance visibility had been completed, and the image data had been manually annotated. Different parts, including edema, tumor parenchyma, enhanced nucleus, and necrosis, were marked, and diagnostic information such as tumor type and stage were obtained from the case's medical record diagnosis and postoperative pathological information.
遍历所有MRI图像,统计平均值和方差等信息,留作训练和预测过程中的标准化操作使用。此外,将同一扫描序列的同一切片位置的四个模态图像堆叠,每个模态作为一个通道,然后将该扫描序列下的所有切片图像与分割结果标注遮罩(mask)分别进行拼接形成三维图片和标注序列,再将这两个序列与该序列对应的诊断结果进行绑定作为一个样例。Traverse all MRI images, statistics such as mean and variance, and reserve it for normalization during training and prediction. In addition, four modal images of the same slice position in the same scanning sequence are stacked, and each modality is used as a channel, and then all slice images under the scanning sequence and the segmentation result labeling mask are spliced respectively to form a three-dimensional Picture and label sequence, and then bind these two sequences with the diagnostic result corresponding to the sequence as an example.
将所有样例进行处理后,按确定比例分为训练集和测试集以备后用。After all samples are processed, they are divided into training set and test set according to a certain proportion for later use.
之后基于原始U-Net网络,使用三维卷积处理方法,搭建三维模型。Then, based on the original U-Net network, a three-dimensional model is built using the three-dimensional convolution processing method.
如图2所示是本发明中使用的模型的架构,包含左侧一个下采样的数据通路和右侧的一个上采样的数据通路。左侧的下采样路径上每一层包含两个3×3×3的卷积层,并且每一个卷积层均采用了dropout的方式以防止过拟合,使用ReLU的激活函数。在两个卷积层后,使用一个步长为2、大小为3×3×3的最大池化层进行下采样操作;右侧数据通路的两层之间使用解卷积的方式进行上采样操作,并将上采样后的特征与左侧下采样层的特征进行拼接,再将拼接后的特征进行两次与左侧上采样层相同的卷积操作。Figure 2 shows the architecture of the model used in the present invention, including a down-sampled data path on the left and an up-sampled data path on the right. Each layer on the downsampling path on the left contains two 3×3×3 convolutional layers, and each convolutional layer adopts a dropout method to prevent overfitting, using the activation function of ReLU. After two convolutional layers, a max pooling layer with stride 2 and size 3×3×3 is used for downsampling; deconvolution is used for upsampling between the two layers of the right data path operation, splicing the upsampled features with the features of the left downsampling layer, and then performing the same convolution operation twice with the left upsampling layer on the spliced features.
下采样层通过不断降低特征的分辨率,提高最终得到的最深层特征中每一个像素的感受野,并得到特征更为抽象的高层表示,即为深层信息,对于图片分类和图片中像素的类别判断具有更高的表示能力,但是由于分辨率的损失使得像素级分类的准确度和得到的分类结果的分辨率都大幅降低。中间的跨层连接即为浅层信息,通过与右侧的上采样路径特征融合,解决了结果与输入相比分辨率下降的问题,使得融合之后的结果得到了更好的表达结果。The downsampling layer improves the receptive field of each pixel in the final deepest feature by continuously reducing the resolution of the feature, and obtains a more abstract high-level representation of the feature, which is the deep information. For image classification and the category of pixels in the image The judgment has higher representation ability, but due to the loss of resolution, the accuracy of pixel-level classification and the resolution of the obtained classification results are greatly reduced. The cross-layer connection in the middle is the shallow layer information. By merging with the up-sampling path feature on the right, the problem that the resolution of the result is reduced compared with the input is solved, and the result after fusion is better expressed.
从上述的上采样通路最后得到融合了深层与浅层信息的特征之后,进行两次卷积核大小为3×3×3的卷积,得到经过卷积的特征后,模型分为两支:一支继续进行一次卷积操作,输出通道与语义分割结果中的分类相同,然后经过softmax计算,得到包含背景、水肿、肿瘤实质、坏死和增强核的输出结果;另一支进行一次卷积操作后,使用全局平均池化(Global Average Pooling),然后紧跟两个全连接层,第二个全连接层的输出与病理诊断分类数相同,包含以下类别:少突胶质细胞瘤、间变期少突胶质细胞瘤、星形胶质细胞瘤、间变期星形胶质细胞瘤、胶质母细胞瘤。From the above-mentioned upsampling path, after finally obtaining the features that integrate the deep and shallow information, perform two convolutions with a convolution kernel size of 3 × 3 × 3. After the convolutional features are obtained, the model is divided into two branches: One continues to perform a convolution operation, and the output channel is the same as the classification in the semantic segmentation result, and then after softmax calculation, the output results including background, edema, tumor parenchyma, necrosis and enhancement kernel are obtained; the other branch performs a convolution operation. Then, use Global Average Pooling, followed by two fully connected layers, the output of the second fully connected layer is the same as the number of pathological diagnosis classifications, including the following categories: oligodendroglioma, anaplastic oligodendroglioma, astrocytoma, anaplastic astrocytoma, glioblastoma.
上述过程可视为,基于肿瘤区域作为模型的注意力区域,结合背景对脑肿瘤类别进行精准识别的过程。上述过程使用多任务学习方法,使得脑肿瘤的分割任务和分类任务能够同时执行并且由于共用浅层特征,使其二者能够在相互的作用下共同提升。The above process can be regarded as a process of accurately identifying the brain tumor category based on the tumor region as the attention region of the model combined with the background. The above process uses a multi-task learning method, so that the segmentation task and the classification task of brain tumors can be performed at the same time, and due to the shared shallow features, the two can be jointly improved under the interaction.
所用多任务学习方法通过下述损失函数衡量并使用下述模型训练方法进行模型性能的优化。The multi-task learning method used is measured by the following loss function and the model performance is optimized using the following model training method.
1.损失函数1. Loss function
1)脑肿瘤分割任务模型的损失函数选择:1) Loss function selection of brain tumor segmentation task model:
医学图像分割中的一个挑战是数据中的类别不平衡问题,例如在脑肿瘤MRI图像中,要分割的目标对象所占整个数据的比例特别小,导致严重的类别失衡。在这种情况下,使用传统的分类交叉熵损失函数会妨碍训练,而Dice损失函数可以有效应对类别失衡问题,因此本发明采用此损失函数作为分割模型的损失函数,具体表示如下:One of the challenges in medical image segmentation is the problem of class imbalance in the data. For example, in brain tumor MRI images, the target object to be segmented occupies a particularly small proportion of the entire data, resulting in severe class imbalance. In this case, using the traditional categorical cross-entropy loss function will hinder training, while the Dice loss function can effectively deal with the class imbalance problem, so the present invention adopts this loss function as the loss function of the segmentation model, which is specifically expressed as follows:
其中u为网络的分割结果输出,v为标签的分割,i是训练块中的像素个数where u is the segmentation result output of the network, v is the label segmentation, and i is the number of pixels in the training block
2)脑肿瘤分类任务的损失函数选择:2) Loss function selection for brain tumor classification task:
该问题为一个分类问题,采用分类问题广泛使用的交叉熵函数作为该模块的损失函数,具体表示如下:This problem is a classification problem, and the cross-entropy function widely used in classification problems is used as the loss function of this module, which is specifically expressed as follows:
其中N为样本总数,K为类别总数,yi,j为标签值,pi,j为预测值where N is the total number of samples, K is the total number of categories, y i, j are the label values, and p i, j are the predicted values
2.模型训练方法如下:2. The model training method is as follows:
1)设置只允许图像分割模型的Dice loss或者肿瘤分类模型的交叉熵损失函数进行反向传播进行迭代训练,即总体损失函数表示为:1) The setting only allows the Dice loss of the image segmentation model or the cross entropy loss function of the tumor classification model to perform back-propagation for iterative training, that is, the overall loss function is expressed as:
2)将图像分割模块的loss值与肿瘤分类模型的loss值以一定的比例组合为之后进行反向传播进行迭代训练,即总体损失函数表示为:2) The loss value of the image segmentation module and the loss value of the tumor classification model are combined in a certain proportion, and then back-propagation is performed for iterative training, that is, the overall loss function is expressed as:
Loss2=LossDice+αLosscross Loss 2 = Loss Dice + αLoss cross
其中α是可调整的比例系数。where α is an adjustable scaling factor.
Claims (3)
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201911393654.4A CN111047589B (en) | 2019-12-30 | 2019-12-30 | Attention-enhanced brain tumor auxiliary intelligent detection and identification method |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201911393654.4A CN111047589B (en) | 2019-12-30 | 2019-12-30 | Attention-enhanced brain tumor auxiliary intelligent detection and identification method |
Publications (2)
Publication Number | Publication Date |
---|---|
CN111047589A CN111047589A (en) | 2020-04-21 |
CN111047589B true CN111047589B (en) | 2022-07-26 |
Family
ID=70241643
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN201911393654.4A Active CN111047589B (en) | 2019-12-30 | 2019-12-30 | Attention-enhanced brain tumor auxiliary intelligent detection and identification method |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN111047589B (en) |
Families Citing this family (14)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN111667458B (en) * | 2020-04-30 | 2023-09-01 | 杭州深睿博联科技有限公司 | Early acute cerebral infarction detection method and device in flat scanning CT |
CN111968127B (en) * | 2020-07-06 | 2021-08-27 | 中国科学院计算技术研究所 | Cancer focus area identification method and system based on full-section pathological image |
CN112085113B (en) * | 2020-09-14 | 2021-05-04 | 四川大学华西医院 | Severe tumor image recognition system and method |
CN112733873A (en) * | 2020-09-23 | 2021-04-30 | 浙江大学山东工业技术研究院 | Chromosome karyotype graph classification method and device based on deep learning |
CN112766333B (en) * | 2021-01-08 | 2022-09-23 | 广东中科天机医疗装备有限公司 | Medical image processing model training method, medical image processing method and device |
CN112927240B (en) * | 2021-03-08 | 2022-04-05 | 重庆邮电大学 | A CT Image Segmentation Method Based on Improved AU-Net Network |
CN113112465B (en) * | 2021-03-31 | 2022-10-18 | 上海深至信息科技有限公司 | System and method for generating carotid intima-media segmentation model |
CN113223014B (en) * | 2021-05-08 | 2023-04-28 | 中国科学院自动化研究所 | Brain image analysis system, method and equipment based on data enhancement |
CN113223704B (en) * | 2021-05-20 | 2022-07-26 | 吉林大学 | Auxiliary diagnosis method for computed tomography aortic aneurysm based on deep learning |
CN113516671B (en) * | 2021-08-06 | 2022-07-01 | 重庆邮电大学 | An image segmentation method of infant brain tissue based on U-net and attention mechanism |
CN114947807B (en) * | 2022-05-06 | 2024-09-17 | 天津大学 | Multitasking prediction method for brain attack classification and meningioma classification |
CN115222007B (en) * | 2022-05-31 | 2023-06-20 | 复旦大学 | Improved particle swarm parameter optimization method for glioma multi-task integrated network |
CN116645381B (en) * | 2023-06-26 | 2024-11-08 | 海南大学 | Brain tumor MRI image segmentation method, system, electronic device and storage medium |
CN117726624B (en) * | 2024-02-07 | 2024-05-28 | 北京长木谷医疗科技股份有限公司 | Method and device for intelligently identifying and evaluating adenoid lesions in real time under video stream |
Citations (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN109087318A (en) * | 2018-07-26 | 2018-12-25 | 东北大学 | A kind of MRI brain tumor image partition method based on optimization U-net network model |
CN109191476A (en) * | 2018-09-10 | 2019-01-11 | 重庆邮电大学 | The automatic segmentation of Biomedical Image based on U-net network structure |
CN109754404A (en) * | 2019-01-02 | 2019-05-14 | 清华大学深圳研究生院 | A kind of lesion segmentation approach end to end based on more attention mechanism |
CN110120033A (en) * | 2019-04-12 | 2019-08-13 | 天津大学 | Based on improved U-Net neural network three-dimensional brain tumor image partition method |
CN110298844A (en) * | 2019-06-17 | 2019-10-01 | 艾瑞迈迪科技石家庄有限公司 | X-ray contrastographic picture blood vessel segmentation and recognition methods and device |
Family Cites Families (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US10878219B2 (en) * | 2016-07-21 | 2020-12-29 | Siemens Healthcare Gmbh | Method and system for artificial intelligence based medical image segmentation |
-
2019
- 2019-12-30 CN CN201911393654.4A patent/CN111047589B/en active Active
Patent Citations (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN109087318A (en) * | 2018-07-26 | 2018-12-25 | 东北大学 | A kind of MRI brain tumor image partition method based on optimization U-net network model |
CN109191476A (en) * | 2018-09-10 | 2019-01-11 | 重庆邮电大学 | The automatic segmentation of Biomedical Image based on U-net network structure |
CN109754404A (en) * | 2019-01-02 | 2019-05-14 | 清华大学深圳研究生院 | A kind of lesion segmentation approach end to end based on more attention mechanism |
CN110120033A (en) * | 2019-04-12 | 2019-08-13 | 天津大学 | Based on improved U-Net neural network three-dimensional brain tumor image partition method |
CN110298844A (en) * | 2019-06-17 | 2019-10-01 | 艾瑞迈迪科技石家庄有限公司 | X-ray contrastographic picture blood vessel segmentation and recognition methods and device |
Non-Patent Citations (1)
Title |
---|
A Multi-Task Convolutional Neural Network for Renal Tumor Segmentation and Classification Using Multi-Phasic CT Images;Tan Pan 等;《2019 IEEE International Conference on Image Processing》;20190826;全文 * |
Also Published As
Publication number | Publication date |
---|---|
CN111047589A (en) | 2020-04-21 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN111047589B (en) | Attention-enhanced brain tumor auxiliary intelligent detection and identification method | |
Liu et al. | Region-to-boundary deep learning model with multi-scale feature fusion for medical image segmentation | |
Qayyum et al. | Automatic segmentation of tumors and affected organs in the abdomen using a 3D hybrid model for computed tomography imaging | |
CN109035263A (en) | Brain tumor image automatic segmentation method based on convolutional neural networks | |
Nizamani et al. | Advance brain tumor segmentation using feature fusion methods with deep U-Net model with CNN for MRI data | |
JP2023540910A (en) | Connected Machine Learning Model with Collaborative Training for Lesion Detection | |
CN111598864B (en) | Liver cell cancer differentiation evaluation method based on multi-modal image contribution fusion | |
Reda et al. | A novel adcs-based cnn classification system for precise diagnosis of prostate cancer | |
CN106096636A (en) | A kind of Advancement Type mild cognition impairment recognition methods based on neuroimaging | |
Zhan et al. | LR-cGAN: Latent representation based conditional generative adversarial network for multi-modality MRI synthesis | |
Wu et al. | MR brain segmentation based on DE-ResUnet combining texture features and background knowledge | |
Mohan et al. | Medical imaging with intelligent systems: a review | |
Wu et al. | Transformer-based 3D U-Net for pulmonary vessel segmentation and artery-vein separation from CT images | |
Randhawa et al. | Deep learning for liver tumour classification: enhanced loss function | |
Zhang et al. | Dual parallel net: A novel deep learning model for rectal tumor segmentation via CNN and transformer with Gaussian Mixture prior | |
Zhou | Modality-level cross-connection and attentional feature fusion based deep neural network for multi-modal brain tumor segmentation | |
Qayyum et al. | Automatic segmentation using a hybrid dense network integrated with an 3D-atrous spatial pyramid pooling module for computed tomography (CT) imaging | |
Yee et al. | 3D hemisphere-based convolutional neural network for whole-brain MRI segmentation | |
Ren et al. | Prostate segmentation in MRI using transformer encoder and decoder framework | |
CN115147404B (en) | Intracranial aneurysm segmentation method by fusing dual features with MRA image | |
Li et al. | Deep attention super-resolution of brain magnetic resonance images acquired under clinical protocols | |
Pallawi et al. | Study of Alzheimer’s disease brain impairment and methods for its early diagnosis: a comprehensive survey | |
Nguyen-Tat et al. | Enhancing brain tumor segmentation in MRI images: A hybrid approach using UNet, attention mechanisms, and transformers | |
Zhou et al. | M-DDC: MRI based demyelinative diseases classification with U-Net segmentation and convolutional network | |
Wang et al. | A calibrated SVM based on weighted smooth GL1/2 for Alzheimer’s disease prediction |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
GR01 | Patent grant | ||
GR01 | Patent grant |