WO2021184817A1 - Method for segmenting liver and focus thereof in medical image - Google Patents

Method for segmenting liver and focus thereof in medical image Download PDF

Info

Publication number
WO2021184817A1
WO2021184817A1 PCT/CN2020/131402 CN2020131402W WO2021184817A1 WO 2021184817 A1 WO2021184817 A1 WO 2021184817A1 CN 2020131402 W CN2020131402 W CN 2020131402W WO 2021184817 A1 WO2021184817 A1 WO 2021184817A1
Authority
WO
WIPO (PCT)
Prior art keywords
training
liver
loss
data
network
Prior art date
Application number
PCT/CN2020/131402
Other languages
French (fr)
Chinese (zh)
Inventor
奚雪峰
郑志华
程成
崔志明
胡伏原
付保川
Original Assignee
苏州科技大学
苏州金比特信息科技有限公司
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by 苏州科技大学, 苏州金比特信息科技有限公司 filed Critical 苏州科技大学
Publication of WO2021184817A1 publication Critical patent/WO2021184817A1/en

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/10Segmentation; Edge detection
    • G06T7/11Region-based segmentation
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/045Combinations of networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T5/00Image enhancement or restoration
    • G06T5/20Image enhancement or restoration using local operators
    • G06T5/30Erosion or dilatation, e.g. thinning
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T5/00Image enhancement or restoration
    • G06T5/70Denoising; Smoothing
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/10Image acquisition modality
    • G06T2207/10072Tomographic images
    • G06T2207/10081Computed x-ray tomography [CT]
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/20Special algorithmic details
    • G06T2207/20081Training; Learning
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/30Subject of image; Context of image processing
    • G06T2207/30004Biomedical image processing
    • G06T2207/30056Liver; Hepatic

Definitions

  • the invention relates to a method for segmenting liver and its lesions in medical images.
  • liver disease is one of the parts with high morbidity and mortality in the world.
  • liver disease occurs in the early stage, and the focus can be located in time, the focus can be controlled and defended, and the metastasis of the focus can be avoided.
  • the treatment of liver disease is of great significance.
  • the appearance of CT images has greatly improved the diagnosis level of doctors, but it is necessary to have a deep professional background and doctors with rich clinical experience to locate the lesion, and it is very time-consuming to diagnose the patient's disease.
  • segmentation algorithms based on region, threshold segmentation and machine learning have emerged.
  • the research on image semantic segmentation has made great progress. Medical image segmentation can accurately understand the location and size of the lesion, but it is accurate The degree needs to be improved.
  • U-Net is an encoding and decoding network. It performs feature extraction first, then upsampling and restoring. The features with the same number of channels at different scales are spliced, and the feature information of different scales is merged by jump connection. It can be trained with less data. The excellent model was later widely used in the segmentation of super-large images and medical images. On the basis of U-Net, there are several problems that need to be solved.
  • U-Net generally adopts a five-layer structure. Simple data can be solved at a shallow level, and complex data can be optimized by deepening the network. Multi-deep networks are the most suitable and not solved; each network The importance of the layers is not clear, and the depth of the network is not pointed out; just through the short connections of each layer, the deep and shallow features cannot be effectively integrated. U-Net++ modified to directly forward the high-resolution feature mapping from the encoder to the decoder network, effectively overcoming the semantic gap of the codec.
  • the purpose of the present invention is to overcome the shortcomings of traditional neural networks used for medical image segmentation in terms of network depth, importance of different depths, and rationality of jump connections, and provide a method for segmenting liver and its lesions in medical images.
  • a method for segmenting the liver and its lesions in medical images, the characteristics are:
  • the abdominal CT image data is filtered and integrated and preprocessed, and divided into multiple data sets for different purposes, and then a new neural network is built, and the small image data is used for initial training;
  • the prediction results of the models trained with the DL, GDL, and TL loss functions are respectively used.
  • the prediction results of the above three loss models are added and the average is formed to form a fusion feature.
  • the network is modified and the three loss models are merged into a single network for training prediction.
  • the above-mentioned method for segmenting the liver and its lesions in a medical image specifically includes the following steps:
  • the training set size is divided into 224 ⁇ 224 and 512 ⁇ 512;
  • the full path plus the ResNet structure form a codec structure.
  • the dense jump connection is replaced by a 1 ⁇ 1 convolutional layer on the basis of DenseNet.
  • the liver outputs The information becomes the input and convolution of the lesion, and the output of other layers of the liver is short-connected to the input of the corresponding depth of the lesion;
  • Reduce to 224 ⁇ 224 data through network training apply effective weight distribution to subsequent model training, train the adjusted picture for 40-60 rounds, with 12-16 slices in each round, rotate the picture during the training process , Zoom in, zoom out, and combine with random probability;
  • the network structure and weight distribution are retained, the original image is combined with the probability of rotation, scaling, flipping, and stretching, and the new learning rate is used for secondary training;
  • step c) uses the original image and a new data enhancement method for secondary training, the original image is 512 ⁇ 512 size, and the image is rotated , Zoom, flip, and stretch operations, combined with random probability, using exponential decay learning rate, adjust each round of decay size adjustment changes, the equation is as follows:
  • the decayed learning rate decayed_learning_rate multiply the initial learning rate set first by the base decay speed decay_steps, the decay rate is set to 0.8 ⁇ 0.9, decay every global_step step, global_step is the current iteration round number, that is, how many The round can iterate through all the sample data, the initial learning rate is set to 1e-3 ⁇ 3e-3, and the original image training is set to 1e-4 ⁇ 3e-4, the result is that the learning rate is based on the base 0.8 ⁇ 0.9 per round The number of steps is attenuated.
  • step d) different medical evaluation results are obtained by adjusting the combination of loss functions.
  • the loss functions are specifically DL, GDL, TL, and the three used are
  • the loss function is the following formula.
  • the loss functions applicable to the liver and the lesion are selected respectively.
  • DL is used to evaluate the similarity between the predicted set and the true set, and is used in the case of unbalanced samples.
  • the expression is as follows:
  • the quantitative calculation of the denominator adopts the method of element square and then summation, where k and t represent the elements of the prediction area and the true value area respectively, and ij represents the traversal of the elements; it is a set similarity measurement function, usually used to calculate the two samples Similarity, the range is [0,1], the coefficient in the numerator, due to the repeated calculation of the common element between k and t in the denominator, and finally the loss value is obtained by multiplying the double points of each category by the sum of the squares of the respective elements ;
  • GDL Generalized Dice loss: When the liver lesion has multiple segmented areas, there is one Dice for each category, and GDL integrates multiple categories and uses one index for quantitative calculation. The formula is as follows:
  • k ij is the true value of category i in the jth pixel, and t ij is the corresponding predicted probability value; compared to DL, there is more weight wi as each category, wi is used to maintain the relationship between the lesion area and the DL coefficient Balance
  • ⁇ and ⁇ respectively control the proportion of false positives and false negatives.
  • the TL coefficient is the DL coefficient.
  • the present invention has significant advantages and beneficial effects, which are specifically embodied in the following aspects:
  • the respective networks of the liver and the lesion use the codec network, and the transition area of the liver and the lesion segmentation is designed to better connect the resolution gap between the codecs.
  • the lesion only receives information from the liver, further narrowing the correct range , So that the network can reduce parameters and time to learn context information, and at the same time can accelerate the network convergence; the original resolution of the picture from the input to the final extraction of the smallest feature map using 16 or 32 times sampling, which can not only reduce the network reasoning time but also Used for denser feature extraction; in addition, remove U-Net++'s Droopout and maximum pooling operations to collect more low-level feature information;
  • 3In terms of loss function compare the performance of multiple Loss, select the optimal function for liver and lesions, and add weighted loss functions to networks of different depths, which can improve the discriminative ability of classifiers of different depth networks and effectively overcome gradients. Disappear the problem, and provide additional regularization; in addition, compare the single optimal loss function model and the loss model based on the combination of weight and similarity for in-depth supervision, select the output of the last residual block, and the loss of other layers is 0.3 The weight is added to the optimizer, and then the output results are weighted and averaged as the final loss. Joint decision-making at each level can effectively avoid the problem of multiple models consuming a lot of resources and time, and at the same time, it can absorb the advantages of each model to alleviate Over-segmentation and under-segmentation issues;
  • the method of the present invention can perform end-to-end training tests, recognize liver and lesions at the same time with high precision and high speed, effectively help doctors recognize CT images, greatly reduce the time and energy consumed by doctors, reduce the probability of misdiagnosis, and have better The practical application value.
  • Figure 1 Schematic diagram of the network structure of the present invention
  • FIG. 1 Schematic diagram of the processing flow of the present invention
  • Figure 3 A schematic diagram of data enhancement and serialization of the present invention
  • Figure 4 Part of the network liver segmentation diagram of the present invention
  • the method for segmenting the liver and its lesions in the medical image of the present invention firstly, data screening and integration;
  • the training set size is divided into 224 ⁇ 224 and 512 ⁇ 512;
  • the full path plus the ResNet structure form a codec structure.
  • the dense jump connection is replaced by a 1 ⁇ 1 convolutional layer on the basis of DenseNet.
  • the liver outputs The information becomes the input and convolution of the lesion, and the output of other layers of the liver is short-connected to the input of the corresponding depth of the lesion;
  • Reduce to 224 ⁇ 224 data through network training apply effective weight distribution to subsequent model training, train the adjusted picture for 40-60 rounds, with 12-16 slices in each round, rotate the picture during the training process , Zoom in, zoom out, and combine with random probability;
  • the network structure and weight distribution are retained, the original image is combined with the probability of rotation, scaling, flipping, and stretching, and the new learning rate is used for secondary training;
  • the original image is 512 ⁇ 512.
  • the image is rotated, zoomed, flipped, and stretched, combined with random probability, and the exponential decay learning rate is used to adjust each round.
  • the attenuation is adjusted and changed, and the equation is as follows:
  • the decayed learning rate decayed_learning_rate multiply the initial learning rate set first by the base decay speed decay_steps, the decay rate is set to 0.8 ⁇ 0.9, decay every global_step step, global_step is the current iteration round number, that is, how many The round can iterate through all the sample data, the initial learning rate is set to 1e-3 ⁇ 3e-3, and the original image training is set to 1e-4 ⁇ 3e-4, the result is that the learning rate is based on the base 0.8 ⁇ 0.9 per round The number of steps is attenuated.
  • the loss functions are specifically DL, GDL, and TL.
  • the three loss functions used are as follows, and the loss functions applicable to the liver and the lesion are selected respectively.
  • DL Generalized Dice loss
  • DL is used to evaluate the similarity between the predicted set and the real set. When the sample is unbalanced, the expression is as follows:
  • the quantitative calculation of the denominator adopts the method of element square and then summation, where k and t represent the elements of the prediction area and the true value area respectively, and ij represents the traversal of the elements; it is a set similarity measurement function, usually used to calculate the two samples Similarity, the range is [0,1], the coefficient in the numerator, due to the repeated calculation of the common element between k and t in the denominator, and finally the loss value is obtained by multiplying the double points of each category by the sum of the squares of the respective elements ;
  • GDL When the liver lesion has multiple segmented areas, there is one Dice for each category, and GDL integrates multiple categories and uses one index for quantitative calculation. The formula is as follows:
  • k ij is the true value of category i in the jth pixel, and t ij is the corresponding predicted probability value; compared to DL, there is more weight wi as each category, wi is used to maintain the relationship between the lesion area and the DL coefficient Balance
  • ⁇ and ⁇ respectively control the proportion of false positives and false negatives
  • the TL coefficient is the DL coefficient.
  • Figure 1 The jump connection is similar to the original U-Net network, C and N use convolution, and R and D use the residual network structure.
  • the DenseNet structure is used between each, and the output of other layers of the liver is connected to the input corresponding to the depth of the lesion.
  • loss function the performance of multiple Loss is compared, and the optimal function for the liver and the lesion is selected.
  • the original image of the data set is a serialized single-channel grayscale image of the abdomen, and the size of the original image and the label are both 512 ⁇ 512.
  • the label is divided into two foreground and one background. 0 represents the background, 1 represents the liver, and 2 represents the lesion.
  • the data set contains a training set of 131 patient sequences.
  • the training data set is eliminated without liver slices, and then shuffled into 19,000 to 20,000 3d slices.
  • the 3d slice is the current slice and the two preceding and following slices as the overall input. Select one of them
  • the 17,000 to 18,000 slices of the sample were used as the training set, the remaining 1800 to 1900 slices were used as the validation set, and there were 70 patient sequences used for testing.
  • Table 1 shows the comparison between the current mainstream semantic segmentation method and the method of the present invention, and the horizontal is the evaluation index.
  • Table 1 in liver segmentation, except that the accuracy is not as good as the joint decision model of multiple loss functions, other indicators are better than all the previous networks. Due to the computer's hard disk read and write mechanism, multiple models and joint decision-making will greatly reduce the computer's computing speed. Adding parameters to a single model for in-depth supervision can achieve the combined effect. The speed and computing resource utilization are significantly better than those of multiple models.
  • Table 2 The segmentation results of the lesion are shown in Table 2:
  • Table 2 The bottom of Table 2 represents different losses and their combinations, and the horizontal is the evaluation index. Table 2 shows that the combination of weighting and similarity-based loss is invalid, and weighting can even reduce network performance. Compared with the results of liver segmentation, the performance of DL and GDL in lesion segmentation is better than that of TL and GTL, respectively. Therefore, using DL for in-depth supervision of lesion segmentation is better than each loss and joint decision-making results.
  • the first row of Figure 4 represents the real image of the label, and the longitudinal direction represents the training iteration effect of different networks.
  • the last row represents the training output effect of the network. Compared with other network structures, the effect is the best.
  • the method for segmenting the liver and its lesions in the medical image of the present invention proposes an end-to-end codec network for segmenting the liver and its lesions.
  • the 1*1 convolution kernel is used as the core unit of dense jump connections.
  • the neural unit integrates multi-scale features, and spreads information with similar semantics that is easier to be processed by the optimizer without introducing too many parameters;
  • the ResNet structure is used to strengthen the backbone network, and the overlap operation is used to replace the addition operation to ensure the depth and width of the network. Designing the transition zone between the liver and the lesion, limiting the lesion segmentation to the liver, saving computing resources, and the effect is better than using the network to segment the lesion alone.
  • the loss functions that are most suitable for liver and lesion segmentation are selected for in-depth supervision. It is better to use different loss functions for in-depth supervision than a single optimal loss function to meet the actual needs of doctors in diagnosis.

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Computational Linguistics (AREA)
  • General Health & Medical Sciences (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Artificial Intelligence (AREA)
  • Biomedical Technology (AREA)
  • Biophysics (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Data Mining & Analysis (AREA)
  • Evolutionary Computation (AREA)
  • Health & Medical Sciences (AREA)
  • Molecular Biology (AREA)
  • Computing Systems (AREA)
  • General Engineering & Computer Science (AREA)
  • Mathematical Physics (AREA)
  • Software Systems (AREA)
  • Image Analysis (AREA)
  • Apparatus For Radiation Diagnosis (AREA)

Abstract

The present invention relates to a method for segmenting a liver and a focus thereof in a medical image. The method comprises: firstly carrying out the screening and integration preprocessing of abdominal CT image data, dividing the abdominal CT image data into a plurality of data sets with different purposes, then building a new neural network, and carrying out initial training by employing small image data; then, storing the trained model, carrying out secondary training by using an original image and a new data enhancement mode, carrying out expansion and corrosion processing on a predicted image, and evaluating same by using a medical evaluation index; respectively by means of model prediction results trained by DL, GDL and TL loss functions, adding and averaging the prediction results of three loss models to form a fusion feature; and finally, modifying a network, wherein the three loss models are fused in a single network for training prediction. An end-to-end training test can be carried out, the liver and a focus can be identified at the same time with high precision and high speed, doctors are effectively helped to identify CT images, thereby reducing time and energy consumed by doctors, and reducing the probability of misdiagnosis.

Description

一种医学图像中肝脏及其病灶分割的方法A method for segmenting liver and its lesions in medical images 技术领域Technical field
本发明涉及一种医学图像中肝脏及其病灶分割的方法。The invention relates to a method for segmenting liver and its lesions in medical images.
背景技术Background technique
目前肝脏部位的疾病是世界上发病率和致死率均较高的部位之一,但若肝部疾病发生在早期,并且能够及时定位病灶,对病灶加以控制与防御,可避免病灶的转移,对肝病的治疗具有重大的意义。CT图像的出现极大提高了医生的诊断水平,但需要具有深厚的专业背景,并且需要具有丰富的临床经验的医生才能定位病灶,诊断出患者的病症非常耗时。随着计算机视觉技术的迅速发展,出现了基于区域、阈值分割和机器学习等分割算法,图像语义分割方面的研究取得了长足的进步,医疗图像分割能精确的了解病灶的位置、大小,但是精确度还有待提高。At present, liver disease is one of the parts with high morbidity and mortality in the world. However, if liver disease occurs in the early stage, and the focus can be located in time, the focus can be controlled and defended, and the metastasis of the focus can be avoided. The treatment of liver disease is of great significance. The appearance of CT images has greatly improved the diagnosis level of doctors, but it is necessary to have a deep professional background and doctors with rich clinical experience to locate the lesion, and it is very time-consuming to diagnose the patient's disease. With the rapid development of computer vision technology, segmentation algorithms based on region, threshold segmentation and machine learning have emerged. The research on image semantic segmentation has made great progress. Medical image segmentation can accurately understand the location and size of the lesion, but it is accurate The degree needs to be improved.
FCN利用反卷积使得被卷积缩小的特征图回到原图,把分割做成像素级别分类,可以处理任意大小的图片。但是,直接将特征进行上采样,使得深层和浅层的信息进行不对等结合,丢失关键性特征信息。U-Net是一个编解码网络,先进行特征提取,再进行上采样还原,将不同尺度通道数相同的特征进行拼接,跳转连接融合不同尺度的特征信息,可以用较少的数据训练出较优模型,后来被广泛用在超大图和医疗图像分割。在U-Net基础上有几点问题需要解决,U-Net普遍采用五层结构,简单数据浅层就能解决,复杂数据加深网络也能得到优化,多深网络最适合没有解决;网络每一层的重要性没有明确,需要多深的网络没有指出;仅仅通过各层的短连接,不能有效融合深浅层特征。U-Net++修改了直接将高分辨的特征 映射从编码器快速转发到解码器网络,有效克服编解码器的语义鸿沟问题。FCN uses deconvolution to make the feature map reduced by convolution return to the original image, and divide the segmentation into pixel-level classification, which can process images of any size. However, by directly up-sampling the features, the deep and shallow information will be unequally combined, and the key feature information will be lost. U-Net is an encoding and decoding network. It performs feature extraction first, then upsampling and restoring. The features with the same number of channels at different scales are spliced, and the feature information of different scales is merged by jump connection. It can be trained with less data. The excellent model was later widely used in the segmentation of super-large images and medical images. On the basis of U-Net, there are several problems that need to be solved. U-Net generally adopts a five-layer structure. Simple data can be solved at a shallow level, and complex data can be optimized by deepening the network. Multi-deep networks are the most suitable and not solved; each network The importance of the layers is not clear, and the depth of the network is not pointed out; just through the short connections of each layer, the deep and shallow features cannot be effectively integrated. U-Net++ modified to directly forward the high-resolution feature mapping from the encoder to the decoder network, effectively overcoming the semantic gap of the codec.
发明内容Summary of the invention
本发明的目的是克服传统用于医学图像分割的神经网络在网络深度、不同深度的重要性和跳转连接的合理性等方面的不足,提供一种医学图像中肝脏及其病灶分割的方法。The purpose of the present invention is to overcome the shortcomings of traditional neural networks used for medical image segmentation in terms of network depth, importance of different depths, and rationality of jump connections, and provide a method for segmenting liver and its lesions in medical images.
本发明的目的通过以下技术方案来实现:The purpose of the present invention is achieved through the following technical solutions:
一种医学图像中肝脏及其病灶分割的方法,特点是:A method for segmenting the liver and its lesions in medical images, the characteristics are:
首先将腹部CT图像数据进行筛选及整合预处理,并划分成多个不同用处的数据集,然后搭建新的神经网络,使用小图数据进行初始训练;Firstly, the abdominal CT image data is filtered and integrated and preprocessed, and divided into multiple data sets for different purposes, and then a new neural network is built, and the small image data is used for initial training;
之后,保存训练好的模型,使用原图和新的数据增强方式进行二次训练,对预测的图片进行膨胀和腐蚀处理,用医学评价指标进行评测;After that, save the trained model, use the original image and the new data enhancement method for secondary training, perform expansion and corrosion processing on the predicted image, and evaluate it with medical evaluation indicators;
分别用DL、GDL、TL损失函数训练的模型预测结果,将上述三个损失模型的预测结果相加取平均值形成融合特征,最后修改网络,把三个损失模型融合在单个网络进行训练预测。The prediction results of the models trained with the DL, GDL, and TL loss functions are respectively used. The prediction results of the above three loss models are added and the average is formed to form a fusion feature. Finally, the network is modified and the three loss models are merged into a single network for training prediction.
进一步地,上述的一种医学图像中肝脏及其病灶分割的方法,其中,具体包括以下步骤:Further, the above-mentioned method for segmenting the liver and its lesions in a medical image specifically includes the following steps:
a)首先,数据筛选及整合;a) First, data screening and integration;
将训练的数据集剔除掉没有肝脏的切片,然后打乱成19000~20000张3d切片,3d切片是当前切片和其前后两张切片作为整体输入,选取其中的17000~18000张切片作为训练集,剩下的1800~1900张切片作为验证集,70个病人序列用于测试,其中训练集大小分为224×224和512×512;Remove the slices without liver from the training data set, and then shuffle them into 19,000 to 20,000 3d slices. The 3d slice is the current slice and the two preceding and following slices as the overall input, and 17,000 to 18,000 slices are selected as the training set. The remaining 1800-1900 slices are used as the verification set, and 70 patient sequences are used for testing. The training set size is divided into 224×224 and 512×512;
b)然后,搭建新的神经网络,使用小图数据进行初始训练;b) Then, build a new neural network and use the small image data for initial training;
把Unet中U型路径设置为主路径,全路径加上ResNet结构形成编解码结构,密集跳转连接在DenseNet基础上换成1×1的卷积层,在肝脏和病灶的过渡地带,肝脏输出的信息成为病灶的输入和卷积,肝脏其他层的 输出短连接至病灶对应深度的输入;Set the U-shaped path in Unet as the main path. The full path plus the ResNet structure form a codec structure. The dense jump connection is replaced by a 1×1 convolutional layer on the basis of DenseNet. In the transition zone between the liver and the lesion, the liver outputs The information becomes the input and convolution of the lesion, and the output of other layers of the liver is short-connected to the input of the corresponding depth of the lesion;
通过网络训练缩小成224×224的数据,把有效的权重分布应用在后续模型训练上,对调整后的图片训练40~60轮,每轮12~16份切片,在训练过程中将图片进行旋转、放大、缩小,并以随机概率组合;Reduce to 224×224 data through network training, apply effective weight distribution to subsequent model training, train the adjusted picture for 40-60 rounds, with 12-16 slices in each round, rotate the picture during the training process , Zoom in, zoom out, and combine with random probability;
c)继而,使用原图和新的数据增强方式进行二次训练;c) Then, use the original image and the new data enhancement method for secondary training;
在缩小的图片数据上训练模型之后,保留网络结构和权重分布,将原图以旋转、缩放、翻转、拉伸的概率组合,使用新的学习率再进行二次训练;After training the model on the reduced image data, the network structure and weight distribution are retained, the original image is combined with the probability of rotation, scaling, flipping, and stretching, and the new learning rate is used for secondary training;
d)最后,通过调整损失函数组合方式得到不同的医学评价结果;d) Finally, different medical evaluation results are obtained by adjusting the combination of loss functions;
通过单个最优损失函数模型和基于权重及相似性相结合的损失模型做不同层的监督信号,得到不同的评价结果。Through a single optimal loss function model and a loss model based on a combination of weights and similarities, different levels of supervision signals are used to obtain different evaluation results.
进一步地,上述的一种医学图像中肝脏及其病灶分割的方法,其中,步骤c)使用原图和新的数据增强方式进行二次训练,原图即是512×512大小,对图片进行旋转、缩放、翻转、拉伸操作,以随机概率组合,采用指数衰减学习率,调整每轮衰减大小调节变化,方程如下:Further, the above-mentioned method for segmenting the liver and its lesions in medical images, wherein step c) uses the original image and a new data enhancement method for secondary training, the original image is 512×512 size, and the image is rotated , Zoom, flip, and stretch operations, combined with random probability, using exponential decay learning rate, adjust each round of decay size adjustment changes, the equation is as follows:
Figure PCTCN2020131402-appb-000001
Figure PCTCN2020131402-appb-000001
上述公式,衰减后的学习率decayed_learning_rate,由先设定的初始学习率learning_rate乘以基数衰减速度decay_steps,衰减速度设定为0.8~0.9,每global_step步进行衰减,global_step是当前迭代轮数,即多少轮可迭代完所有的样本数据,初始学习率设定为1e-3~3e-3,在原图训练时设定为1e-4~3e-4,结果是学习率以基数0.8~0.9每一轮的步数进行衰减。The above formula, the decayed learning rate decayed_learning_rate, multiply the initial learning rate set first by the base decay speed decay_steps, the decay rate is set to 0.8~0.9, decay every global_step step, global_step is the current iteration round number, that is, how many The round can iterate through all the sample data, the initial learning rate is set to 1e-3 ~ 3e-3, and the original image training is set to 1e-4 ~ 3e-4, the result is that the learning rate is based on the base 0.8 ~ 0.9 per round The number of steps is attenuated.
进一步地,上述的一种医学图像中肝脏及其病灶分割的方法,其中,步骤d)通过调整损失函数组合方式得到不同的医学评价结果,损失函数具体为DL、GDL、TL,采用的三个损失函数如下公式,分别选取对于肝 脏和病灶适用的损失函数,DL用于评估预测集合和真实集合的相似度,用于样本不均衡的情况,表达式如下:Further, in the above-mentioned method for segmenting liver and its lesions in medical images, in step d) different medical evaluation results are obtained by adjusting the combination of loss functions. The loss functions are specifically DL, GDL, TL, and the three used are The loss function is the following formula. The loss functions applicable to the liver and the lesion are selected respectively. DL is used to evaluate the similarity between the predicted set and the true set, and is used in the case of unbalanced samples. The expression is as follows:
Figure PCTCN2020131402-appb-000002
Figure PCTCN2020131402-appb-000002
分母的量化计算采取元素平方再求和的方法,其中k和t分别代表预测区域和真值区域元素,ij代表遍历其中元素;是一种集合相似度度量函数,通常用于计算两个样本的相似度,范围为[0,1],分子中的系数,因分母存在重复计算k和t之间的共同元素的原因,最终由各类别的2倍点乘除以各自元素的平方和得到损失值;The quantitative calculation of the denominator adopts the method of element square and then summation, where k and t represent the elements of the prediction area and the true value area respectively, and ij represents the traversal of the elements; it is a set similarity measurement function, usually used to calculate the two samples Similarity, the range is [0,1], the coefficient in the numerator, due to the repeated calculation of the common element between k and t in the denominator, and finally the loss value is obtained by multiplying the double points of each category by the sum of the squares of the respective elements ;
GDL(Generalized Dice loss):当肝脏病灶有多个分割区域时,针对每一类有一个Dice,而GDL将多个类别进行整合,采用一个指标进行量化计算,公式如下:GDL (Generalized Dice loss): When the liver lesion has multiple segmented areas, there is one Dice for each category, and GDL integrates multiple categories and uses one index for quantitative calculation. The formula is as follows:
Figure PCTCN2020131402-appb-000003
Figure PCTCN2020131402-appb-000003
其中k ij为类别i在第j个像素的真实值,t ij为相应的预测概率值;相比于DL,多了作为每个类别的权重wi,wi用于维系病灶区域和DL系数之间的平衡; Among them, k ij is the true value of category i in the jth pixel, and t ij is the corresponding predicted probability value; compared to DL, there is more weight wi as each category, wi is used to maintain the relationship between the lesion area and the DL coefficient Balance
TL(Tversky loss)公式如下:The TL (Tversky loss) formula is as follows:
Figure PCTCN2020131402-appb-000004
Figure PCTCN2020131402-appb-000004
其中k ij为类别i在第j个像素的真实值,t ij为相应的预测概率值; Where k ij is the true value of category i at the jth pixel, and t ij is the corresponding predicted probability value;
α和β分别控制假阳性和假阴性的比重。α and β respectively control the proportion of false positives and false negatives.
进一步地,上述的一种医学图像中肝脏及其病灶分割的方法,其中,当α=β=0.5,TL系数就是DL系数。Further, in the above-mentioned method for segmenting the liver and its lesions in a medical image, when α=β=0.5, the TL coefficient is the DL coefficient.
本发明与现有技术相比具有显著的优点和有益效果,具体体现在以下 方面:Compared with the prior art, the present invention has significant advantages and beneficial effects, which are specifically embodied in the following aspects:
①通过数据预处理,去除无效的肝脏图片;将图像去噪提高对比度更易于网络分割模糊边缘;使用序列化的3d图像做融合分割,可以保留上下文的语义信息;将合成的3d图做不同组合的数据增强强化了数据集鲁棒性和防止过拟合;①Through data preprocessing, remove invalid liver images; denoise the image to improve the contrast and make it easier to segment fuzzy edges on the network; use serialized 3d images for fusion segmentation, which can retain the semantic information of the context; combine the synthesized 3d images in different combinations The data enhancement strengthens the robustness of the data set and prevents overfitting;
②肝脏和病灶各自的网络都采用编解码网络,设计了肝脏和病灶分割的过度区更好的连接编解码器之间的分辨率鸿沟,同时病灶只接收来自肝脏的信息,把正确范围进一步缩小,使得网络能够减少参数和时间去学习上下文信息,同时能加速网络收敛;原始分辨率的图片从输入开始到最终提取的最小特征图采用16或32倍采样,这样不仅可以减少网络推理时间还可以用于更密集的特征提取;此外,去掉U-Net++的Droopout和最大池化操作,收集更多的底层特征信息;②The respective networks of the liver and the lesion use the codec network, and the transition area of the liver and the lesion segmentation is designed to better connect the resolution gap between the codecs. At the same time, the lesion only receives information from the liver, further narrowing the correct range , So that the network can reduce parameters and time to learn context information, and at the same time can accelerate the network convergence; the original resolution of the picture from the input to the final extraction of the smallest feature map using 16 or 32 times sampling, which can not only reduce the network reasoning time but also Used for denser feature extraction; in addition, remove U-Net++'s Droopout and maximum pooling operations to collect more low-level feature information;
③在损失函数方面,对比多个Loss的性能,选取对于肝脏和病灶最优的函数,在不同深度的网络添加带权重的损失函数,可以提高不同深度网络的分类器判别能力,能有效克服梯度消失问题,并提供额外的正则化;此外,对比了单个最优损失函数模型和基于权重及相似性相结合的损失模型做深度监督,选取最后一个残差块的输出,其它层损失以0.3的权重加到优化器里,再以输出结果进行加权求和取平均作为最终损失,通过各个层级的联合决策能够有效避免多个模型消耗大量资源和时间的问题,同时能够吸取各个模型的优点,缓解过分割和欠分割的问题;③In terms of loss function, compare the performance of multiple Loss, select the optimal function for liver and lesions, and add weighted loss functions to networks of different depths, which can improve the discriminative ability of classifiers of different depth networks and effectively overcome gradients. Disappear the problem, and provide additional regularization; in addition, compare the single optimal loss function model and the loss model based on the combination of weight and similarity for in-depth supervision, select the output of the last residual block, and the loss of other layers is 0.3 The weight is added to the optimizer, and then the output results are weighted and averaged as the final loss. Joint decision-making at each level can effectively avoid the problem of multiple models consuming a lot of resources and time, and at the same time, it can absorb the advantages of each model to alleviate Over-segmentation and under-segmentation issues;
④本发明方法能进行端到端的训练测试,以高精度和高速度来同时识别肝脏和病灶,有效帮助医生识别CT图像,大大减少医生所消耗的时间和精力,减少误诊的概率,具有较好的实际应用价值。④The method of the present invention can perform end-to-end training tests, recognize liver and lesions at the same time with high precision and high speed, effectively help doctors recognize CT images, greatly reduce the time and energy consumed by doctors, reduce the probability of misdiagnosis, and have better The practical application value.
本发明的其他特征和优点将在随后的说明书阐述,并且,部分地从说明书中变得显而易见,或者通过实施本发明具体实施方式了解。本发明的目的和其他优点可通过在所写的说明书、权利要求书、以及附图中所特别 指出的结构来实现和获得。Other features and advantages of the present invention will be described in the following description, and partly become obvious from the description, or understood by implementing specific embodiments of the present invention. The purpose and other advantages of the present invention can be realized and obtained by the structures specifically pointed out in the written specification, claims, and drawings.
附图说明Description of the drawings
为了更清楚地说明本发明实施例的技术方案,下面将对实施例中所需要使用的附图作简单地介绍,应当理解,以下附图仅示出了本发明的某些实施例,因此不应被看作是对范围的限定,对于本领域普通技术人员来讲,在不付出创造性劳动的前提下,还可以根据这些附图获得其他相关的附图。In order to explain the technical solutions of the embodiments of the present invention more clearly, the following will briefly introduce the drawings that need to be used in the embodiments. It should be understood that the following drawings only show some embodiments of the present invention, and therefore do not It should be regarded as a limitation of the scope. For those of ordinary skill in the art, other related drawings can be obtained based on these drawings without creative work.
图1:本发明的网络结构示意图;Figure 1: Schematic diagram of the network structure of the present invention;
图2:本发明处理流程示意图;Figure 2: Schematic diagram of the processing flow of the present invention;
图3:本发明的数据增强及序列化的示意图;Figure 3: A schematic diagram of data enhancement and serialization of the present invention;
图4:本发明的部分网络肝脏分割图;Figure 4: Part of the network liver segmentation diagram of the present invention;
具体实施方式Detailed ways
下面将结合本发明实施例中附图,对本发明实施例中的技术方案进行清楚、完整地描述,显然,所描述的实施例仅仅是本发明一部分实施例,而不是全部的实施例。通常在此处附图中描述和示出的本发明实施例的组件可以以各种不同的配置来布置和设计。因此,以下对在附图中提供的本发明的实施例的详细描述并非旨在限制要求保护的本发明的范围,而是仅仅表示本发明的选定实施例。基于本发明的实施例,本领域技术人员在没有做出创造性劳动的前提下所获得的所有其他实施例,都属于本发明保护的范围。The technical solutions in the embodiments of the present invention will be clearly and completely described below in conjunction with the accompanying drawings in the embodiments of the present invention. Obviously, the described embodiments are only a part of the embodiments of the present invention, rather than all the embodiments. The components of the embodiments of the present invention generally described and shown in the drawings herein may be arranged and designed in various different configurations. Therefore, the following detailed description of the embodiments of the present invention provided in the accompanying drawings is not intended to limit the scope of the claimed invention, but merely represents selected embodiments of the present invention. Based on the embodiments of the present invention, all other embodiments obtained by those skilled in the art without creative work shall fall within the protection scope of the present invention.
应注意到:相似的标号和字母在下面的附图中表示类似项,因此,一旦某一项在一个附图中被定义,则在随后的附图中不需要对其进行进一步定义和解释。同时,在本发明的描述中,方位术语和次序术语等仅用于区分描述,而不能理解为指示或暗示相对重要性。It should be noted that similar reference numerals and letters indicate similar items in the following figures. Therefore, once a certain item is defined in one figure, it does not need to be further defined and explained in subsequent figures. At the same time, in the description of the present invention, orientation terms and sequence terms are only used for distinguishing description, and cannot be understood as indicating or implying relative importance.
本发明医学图像中肝脏及其病灶分割的方法,首先,数据筛选及整合;The method for segmenting the liver and its lesions in the medical image of the present invention, firstly, data screening and integration;
将训练的数据集剔除掉没有肝脏的切片,然后打乱成19000~20000张3d切片,3d切片是当前切片和其前后两张切片作为整体输入,选取其中的17000~18000张切片作为训练集,剩下的1800~1900张切片作为验证集,70个病人序列用于测试,其中训练集大小分为224×224和512×512;Remove the slices without liver from the training data set, and then shuffle them into 19,000 to 20,000 3d slices. The 3d slice is the current slice and the two preceding and following slices as the overall input, and 17,000 to 18,000 slices are selected as the training set. The remaining 1800-1900 slices are used as the verification set, and 70 patient sequences are used for testing. The training set size is divided into 224×224 and 512×512;
然后,搭建新的神经网络,使用小图数据进行初始训练;Then, build a new neural network and use the small image data for initial training;
把Unet中U型路径设置为主路径,全路径加上ResNet结构形成编解码结构,密集跳转连接在DenseNet基础上换成1×1的卷积层,在肝脏和病灶的过渡地带,肝脏输出的信息成为病灶的输入和卷积,肝脏其他层的输出短连接至病灶对应深度的输入;Set the U-shaped path in Unet as the main path. The full path plus the ResNet structure form a codec structure. The dense jump connection is replaced by a 1×1 convolutional layer on the basis of DenseNet. In the transition zone between the liver and the lesion, the liver outputs The information becomes the input and convolution of the lesion, and the output of other layers of the liver is short-connected to the input of the corresponding depth of the lesion;
通过网络训练缩小成224×224的数据,把有效的权重分布应用在后续模型训练上,对调整后的图片训练40~60轮,每轮12~16份切片,在训练过程中将图片进行旋转、放大、缩小,并以随机概率组合;Reduce to 224×224 data through network training, apply effective weight distribution to subsequent model training, train the adjusted picture for 40-60 rounds, with 12-16 slices in each round, rotate the picture during the training process , Zoom in, zoom out, and combine with random probability;
继而,使用原图和新的数据增强方式进行二次训练;Then, use the original image and the new data enhancement method for secondary training;
在缩小的图片数据上训练模型之后,保留网络结构和权重分布,将原图以旋转、缩放、翻转、拉伸的概率组合,使用新的学习率再进行二次训练;After training the model on the reduced image data, the network structure and weight distribution are retained, the original image is combined with the probability of rotation, scaling, flipping, and stretching, and the new learning rate is used for secondary training;
使用原图和新的数据增强方式进行二次训练,原图即是512×512大小,对图片进行旋转、缩放、翻转、拉伸操作,以随机概率组合,采用指数衰减学习率,调整每轮衰减大小调节变化,方程如下:Use the original image and the new data enhancement method for secondary training. The original image is 512×512. The image is rotated, zoomed, flipped, and stretched, combined with random probability, and the exponential decay learning rate is used to adjust each round. The attenuation is adjusted and changed, and the equation is as follows:
Figure PCTCN2020131402-appb-000005
Figure PCTCN2020131402-appb-000005
上述公式,衰减后的学习率decayed_learning_rate,由先设定的初始学习率learning_rate乘以基数衰减速度decay_steps,衰减速度设定为0.8~0.9,每global_step步进行衰减,global_step是当前迭代轮数,即多少轮可迭代完所有的样本数据,初始学习率设定为1e-3~3e-3,在原图训练时设定为1e-4~3e-4,结果是学习率以基数0.8~0.9每一轮的步数进行衰减。The above formula, the decayed learning rate decayed_learning_rate, multiply the initial learning rate set first by the base decay speed decay_steps, the decay rate is set to 0.8~0.9, decay every global_step step, global_step is the current iteration round number, that is, how many The round can iterate through all the sample data, the initial learning rate is set to 1e-3 ~ 3e-3, and the original image training is set to 1e-4 ~ 3e-4, the result is that the learning rate is based on the base 0.8 ~ 0.9 per round The number of steps is attenuated.
最后,通过调整损失函数组合方式得到不同的医学评价结果;Finally, different medical evaluation results can be obtained by adjusting the combination of loss functions;
损失函数具体为DL、GDL、TL,采用的三个损失函数如下公式,分别选取对于肝脏和病灶适用的损失函数,DL(Generalized Dice loss)用于评估预测集合和真实集合的相似度,用于样本不均衡的情况,表达式如下:The loss functions are specifically DL, GDL, and TL. The three loss functions used are as follows, and the loss functions applicable to the liver and the lesion are selected respectively. DL (Generalized Dice loss) is used to evaluate the similarity between the predicted set and the real set. When the sample is unbalanced, the expression is as follows:
Figure PCTCN2020131402-appb-000006
Figure PCTCN2020131402-appb-000006
分母的量化计算采取元素平方再求和的方法,其中k和t分别代表预测区域和真值区域元素,ij代表遍历其中元素;是一种集合相似度度量函数,通常用于计算两个样本的相似度,范围为[0,1],分子中的系数,因分母存在重复计算k和t之间的共同元素的原因,最终由各类别的2倍点乘除以各自元素的平方和得到损失值;The quantitative calculation of the denominator adopts the method of element square and then summation, where k and t represent the elements of the prediction area and the true value area respectively, and ij represents the traversal of the elements; it is a set similarity measurement function, usually used to calculate the two samples Similarity, the range is [0,1], the coefficient in the numerator, due to the repeated calculation of the common element between k and t in the denominator, and finally the loss value is obtained by multiplying the double points of each category by the sum of the squares of the respective elements ;
GDL:当肝脏病灶有多个分割区域时,针对每一类有一个Dice,而GDL将多个类别进行整合,采用一个指标进行量化计算,公式如下:GDL: When the liver lesion has multiple segmented areas, there is one Dice for each category, and GDL integrates multiple categories and uses one index for quantitative calculation. The formula is as follows:
Figure PCTCN2020131402-appb-000007
Figure PCTCN2020131402-appb-000007
其中k ij为类别i在第j个像素的真实值,t ij为相应的预测概率值;相比于DL,多了作为每个类别的权重wi,wi用于维系病灶区域和DL系数之间的平衡; Among them, k ij is the true value of category i in the jth pixel, and t ij is the corresponding predicted probability value; compared to DL, there is more weight wi as each category, wi is used to maintain the relationship between the lesion area and the DL coefficient Balance
TL(Tversky loss)公式如下:The TL (Tversky loss) formula is as follows:
Figure PCTCN2020131402-appb-000008
Figure PCTCN2020131402-appb-000008
其中k ij为类别i在第j个像素的真实值,t ij为相应的预测概率值; Where k ij is the true value of category i at the jth pixel, and t ij is the corresponding predicted probability value;
α和β分别控制假阳性和假阴性的比重;α and β respectively control the proportion of false positives and false negatives;
当α=β=0.5,TL系数即是DL系数。When α=β=0.5, the TL coefficient is the DL coefficient.
首先进行数据预处理,去掉没有肝脏的CT图像,将剩余的图像整合序列化,以3D数据流的形式做数据增强,通过合成新的样本来提高准确性,然后使用相关评价指标评价,同时将预测的图像进行膨胀和腐蚀后处理得到预测标签;First, perform data preprocessing, remove the CT images without liver, integrate and serialize the remaining images, and do data enhancement in the form of 3D data streams. The accuracy is improved by synthesizing new samples, and then the relevant evaluation indicators are used for evaluation. The predicted image is post-processed after expansion and corrosion to obtain the predicted label;
图1跳转连接类似原始U-Net网络,C和N采用卷积,R和D采用残差网络结构。在圆弧方向,各自之间分别采用DenseNet结构,肝脏其他层的输出短连接至病灶对应深度的输入,在损失函数方面,对比多个Loss的性能,选取对于肝脏和病灶最优的函数。Figure 1 The jump connection is similar to the original U-Net network, C and N use convolution, and R and D use the residual network structure. In the arc direction, the DenseNet structure is used between each, and the output of other layers of the liver is connected to the input corresponding to the depth of the lesion. In terms of loss function, the performance of multiple Loss is compared, and the optimal function for the liver and the lesion is selected.
具体步骤如下:Specific steps are as follows:
a)首先,数据预处理;a) First, data preprocessing;
选择肝脏肿瘤分割挑战(LiTS)官网上数据集,由PatrickChrist主办。数据集原图是序列化的腹部单通道灰度图,原图和标签的大小都是512×512。标签是分为两个前景和一个背景,0代表背景,1代表肝脏,2代表病灶,把标签重新编排成三通道的图:背景、肝脏和病灶,相应的地方用1代表,其他都为0。数据集包含131个病人序列的训练集,把训练的数据集剔除掉没有肝脏的切片,然后打乱成19000~20000张3d切片,3d切片是当前切片和前后两张切片作为整体输入,选取其中的17000~18000张切片作为训练集,剩下的1800~1900张切片作为验证集,还有70个病人序列用作测试。Select the liver tumor segmentation challenge (LiTS) official website data set, hosted by PatrickChrist. The original image of the data set is a serialized single-channel grayscale image of the abdomen, and the size of the original image and the label are both 512×512. The label is divided into two foreground and one background. 0 represents the background, 1 represents the liver, and 2 represents the lesion. Reorganize the labels into a three-channel image: background, liver, and lesions. The corresponding places are represented by 1 and the others are 0. . The data set contains a training set of 131 patient sequences. The training data set is eliminated without liver slices, and then shuffled into 19,000 to 20,000 3d slices. The 3d slice is the current slice and the two preceding and following slices as the overall input. Select one of them The 17,000 to 18,000 slices of the sample were used as the training set, the remaining 1800 to 1900 slices were used as the validation set, and there were 70 patient sequences used for testing.
b)设置训练参数,进行初始训练;b) Set training parameters and perform initial training;
为方便把有效的权重分布应用在模型训练上,初始把图片大小缩小成224×224,对调整后的图片训练40~60轮,每轮12~16份切片,最后在原图上微调20~40轮达到模型最优,具体处理流程,如图2所示。最后进行图片旋转、放大、缩小等操作,并以一定的概率组合,目的在于数据增强,数据增强的直观效果如图3所示。表一表示肝脏分割相关网络性能:To facilitate the application of effective weight distribution to model training, initially reduce the size of the image to 224×224, train the adjusted image for 40-60 rounds, with 12-16 slices per round, and finally fine-tune 20-40 on the original image. The round reaches the optimal model, and the specific processing flow is shown in Figure 2. Finally, perform operations such as image rotation, enlargement, and reduction, and combine them with a certain probability for the purpose of data enhancement. The intuitive effect of data enhancement is shown in Figure 3. Table 1 shows the network performance related to liver segmentation:
表一 肝脏分割评价指标Table 1 Evaluation Index of Liver Segmentation
Figure PCTCN2020131402-appb-000009
Figure PCTCN2020131402-appb-000009
表一下方表示目前主流的语义分割方法和本发明方法的比较,横向是评价指标。从表一可以看出,在肝脏分割上,除精度不及多个损失函数联合决策模型,其他指标均优于前面所有网络。由于计算机硬盘读写机制,多个模型取联合决策会大大降低计算机的运算速度,单个模型添加参数做深度监督,达到联合的效果,速度和运算资源利用要明显优于多模型。病灶的分割结果见表二:The bottom of Table 1 shows the comparison between the current mainstream semantic segmentation method and the method of the present invention, and the horizontal is the evaluation index. As can be seen from Table 1, in liver segmentation, except that the accuracy is not as good as the joint decision model of multiple loss functions, other indicators are better than all the previous networks. Due to the computer's hard disk read and write mechanism, multiple models and joint decision-making will greatly reduce the computer's computing speed. Adding parameters to a single model for in-depth supervision can achieve the combined effect. The speed and computing resource utilization are significantly better than those of multiple models. The segmentation results of the lesion are shown in Table 2:
表二 网络结构参数Table 2 Network structure parameters
Figure PCTCN2020131402-appb-000010
Figure PCTCN2020131402-appb-000010
表二下面代表不同损失及其组合,横向是评价指标。表二可以看出,加权和基于相似性相结合的损失是无效的,加权甚至会降低网络性能。与肝脏分割结果相比,病灶分割中DL和GDL性能分别优于TL和GTL。于是使用DL做病灶分割的深度监督,效果优于各个损失以及联合决策结果。The bottom of Table 2 represents different losses and their combinations, and the horizontal is the evaluation index. Table 2 shows that the combination of weighting and similarity-based loss is invalid, and weighting can even reduce network performance. Compared with the results of liver segmentation, the performance of DL and GDL in lesion segmentation is better than that of TL and GTL, respectively. Therefore, using DL for in-depth supervision of lesion segmentation is better than each loss and joint decision-making results.
图4第一行表示标签真实图像,纵向表示不同网络的训练迭代效果, 最后一行为网络的训练输出效果,相比其他网络结构,效果最好。The first row of Figure 4 represents the real image of the label, and the longitudinal direction represents the training iteration effect of different networks. The last row represents the training output effect of the network. Compared with other network structures, the effect is the best.
使用双层编解码半圆型网络,通过密集的跳转连接,把深层和浅层的语义信息结合起来,更易于给优化器处理;设计肝脏和病灶分割的过渡区,使得肝脏分割的结果有效的传递给病灶分割,大大节约了分割原图的时间;选取互补的损失函数来组合进行深度监督,可以有效的接受反传播时的梯度信号,获得更多的正则化效果。在基于加权和基于相似性中选取了最适合肝脏及其病灶分割的损失函数,用最优损失做深度监督,同时使用两者一起做深度监督。最后在肝脏分割上,除了在精度上低于多个损失函数做联合决策的模型,其他评价指标包括病灶所有的评价指标均超过多模型融合结果。Using a two-layer codec semicircular network, through dense jump connections, the deep and shallow semantic information are combined, which is easier for the optimizer to process; the transition zone between liver and lesion segmentation is designed to make the results of liver segmentation effective Passing to lesion segmentation greatly saves the time to segment the original image; selecting complementary loss functions to combine for in-depth supervision can effectively accept the gradient signal during back propagation and obtain more regularization effects. In weight-based and similarity-based, the most suitable loss function for liver and its lesion segmentation is selected, and the optimal loss is used for in-depth supervision, and both are used together for in-depth supervision. Finally, in liver segmentation, except for the model that is lower in accuracy than multiple loss functions for joint decision-making, other evaluation indicators, including lesions, all evaluation indicators exceed the multi-model fusion result.
综上所述,本发明医学图像中肝脏及其病灶分割的方法,提出端到端的肝脏及其病灶分割的编解码网络,1*1的卷积核作为密集跳转连接的核心单元,多个神经单元融合多尺度特征,传播语义相近的信息更易被优化器处理,同时不引入过多参数;使用ResNet结构加强主干网络,用叠操作取代加操作,保证网络的深度和宽度。设计肝脏和病灶的过渡区,把病灶分割限定在肝脏里,节约计算资源,效果优于单独用网络来分割病灶。基于加权策略和基于相似性的损失模型,选取分别最适合肝脏和病灶分割的损失函数,用做深度监督,用不同的损失函数比单个最优损失函数做深度监督要好,满足医生诊断实际需求。In summary, the method for segmenting the liver and its lesions in the medical image of the present invention proposes an end-to-end codec network for segmenting the liver and its lesions. The 1*1 convolution kernel is used as the core unit of dense jump connections. The neural unit integrates multi-scale features, and spreads information with similar semantics that is easier to be processed by the optimizer without introducing too many parameters; the ResNet structure is used to strengthen the backbone network, and the overlap operation is used to replace the addition operation to ensure the depth and width of the network. Designing the transition zone between the liver and the lesion, limiting the lesion segmentation to the liver, saving computing resources, and the effect is better than using the network to segment the lesion alone. Based on the weighting strategy and the similarity-based loss model, the loss functions that are most suitable for liver and lesion segmentation are selected for in-depth supervision. It is better to use different loss functions for in-depth supervision than a single optimal loss function to meet the actual needs of doctors in diagnosis.
以上所述仅为本发明的优选实施例而已,并不用于限制本发明,对于本领域的技术人员来说,本发明可以有各种更改和变化。凡在本发明的精神和原则之内,所作的任何修改、等同替换、改进等,均应包含在本发明的保护范围之内。应注意到:相似的标号和字母在下面的附图中表示类似项,因此,一旦某一项在一个附图中被定义,则在随后的附图中不需要对其进行进一步定义和解释。The above are only preferred embodiments of the present invention and are not used to limit the present invention. For those skilled in the art, the present invention can have various modifications and changes. Any modification, equivalent replacement, improvement, etc. made within the spirit and principle of the present invention should be included in the protection scope of the present invention. It should be noted that similar reference numerals and letters indicate similar items in the following figures. Therefore, once a certain item is defined in one figure, it does not need to be further defined and explained in subsequent figures.
上述仅为本发明的具体实施方式,但本发明的保护范围并不局限于此,任何熟悉本技术领域的技术人员在本发明揭露的技术范围内,可轻易想到变化或替换,都应涵盖在本发明的保护范围之内。因此,本发明的保护范围应以权利要求所述的保护范围为准。The above are only specific embodiments of the present invention, but the protection scope of the present invention is not limited thereto. Any person skilled in the art can easily conceive of changes or substitutions within the technical scope disclosed by the present invention, which shall be covered Within the protection scope of the present invention. Therefore, the protection scope of the present invention should be subject to the protection scope described in the claims.
需要说明的是,在本文中,诸如第一和第二等之类的关系术语仅仅用来将一个实体或者操作与另一个实体或操作区分开来,而不一定要求或者暗示这些实体或操作之间存在任何这种实际的关系或者顺序。而且,术语“包括”、“包含”或者其任何其他变体意在涵盖非排他性的包含,从而使得包括一系列要素的过程、方法、物品或者设备不仅包括那些要素,而且还包括没有明确列出的其他要素,或者是还包括为这种过程、方法、物品或者设备所固有的要素。在没有更多限制的情况下,由语句“包括一个……”限定的要素,并不排除在包括所述要素的过程、方法、物品或者设备中还存在另外的相同要素。It should be noted that in this article, relational terms such as first and second are only used to distinguish one entity or operation from another entity or operation, and do not necessarily require or imply one of these entities or operations. There is any such actual relationship or order between. Moreover, the terms "include", "include" or any other variants thereof are intended to cover non-exclusive inclusion, so that a process, method, article or device including a series of elements not only includes those elements, but also includes those that are not explicitly listed Other elements of, or also include elements inherent to this process, method, article or equipment. If there are no more restrictions, the element defined by the sentence "including a..." does not exclude the existence of other identical elements in the process, method, article, or equipment that includes the element.

Claims (5)

  1. 一种医学图像中肝脏及其病灶分割的方法,其特征在于:A method for segmenting liver and its lesions in medical images, which is characterized in:
    首先将腹部CT图像数据进行筛选及整合预处理,并划分成多个不同用处的数据集,然后搭建新的神经网络,使用小图数据进行初始训练;Firstly, the abdominal CT image data is filtered and integrated and preprocessed, and divided into multiple data sets for different purposes, and then a new neural network is built, and the small image data is used for initial training;
    之后,保存训练好的模型,使用原图和新的数据增强方式进行二次训练,对预测的图片进行膨胀和腐蚀处理,用医学评价指标进行评测;After that, save the trained model, use the original image and the new data enhancement method for secondary training, perform expansion and corrosion processing on the predicted image, and evaluate it with medical evaluation indicators;
    分别用DL、GDL、TL损失函数训练的模型预测结果,将上述三个损失模型的预测结果相加取平均值形成融合特征,最后修改网络,把三个损失模型融合在单个网络进行训练预测。The prediction results of the models trained with the DL, GDL, and TL loss functions are respectively used. The prediction results of the above three loss models are added and the average is formed to form a fusion feature. Finally, the network is modified and the three loss models are merged into a single network for training prediction.
  2. 根据权利要求1所述的一种医学图像中肝脏及其病灶分割的方法,其特征在于:具体包括以下步骤:The method for segmenting the liver and its lesions in a medical image according to claim 1, characterized in that it specifically comprises the following steps:
    a)首先,数据筛选及整合;a) First, data screening and integration;
    将训练的数据集剔除掉没有肝脏的切片,然后打乱成19000~20000张3d切片,3d切片是当前切片和其前后两张切片作为整体输入,选取其中的17000~18000张切片作为训练集,剩下的1800~1900张切片作为验证集,70个病人序列用于测试,其中训练集大小分为224×224和512×512;Remove the slices without liver from the training data set, and then shuffle them into 19,000 to 20,000 3d slices. The 3d slice is the current slice and the two preceding and following slices as the overall input, and 17,000 to 18,000 slices are selected as the training set. The remaining 1800-1900 slices are used as the verification set, and 70 patient sequences are used for testing. The training set size is divided into 224×224 and 512×512;
    b)然后,搭建新的神经网络,使用小图数据进行初始训练;b) Then, build a new neural network and use the small image data for initial training;
    把Unet中U型路径设置为主路径,全路径加上ResNet结构形成编解码结构,密集跳转连接在DenseNet基础上换成1×1的卷积层,在肝脏和病灶的过渡地带,肝脏输出的信息成为病灶的输入和卷积,肝脏其他层的输出短连接至病灶对应深度的输入;Set the U-shaped path in Unet as the main path. The full path plus the ResNet structure form a codec structure. The dense jump connection is replaced by a 1×1 convolutional layer on the basis of DenseNet. In the transition zone between the liver and the lesion, the liver outputs The information becomes the input and convolution of the lesion, and the output of other layers of the liver is short-connected to the input of the corresponding depth of the lesion;
    通过网络训练缩小成224×224的数据,把有效的权重分布应用在后续模型训练上,对调整后的图片训练40~60轮,每轮12~16份切片,在训练过程中将图片进行旋转、放大、缩小,并以随机概率组合;Reduce to 224×224 data through network training, apply effective weight distribution to subsequent model training, train the adjusted picture for 40-60 rounds, with 12-16 slices in each round, rotate the picture during the training process , Zoom in, zoom out, and combine with random probability;
    c)继而,使用原图和新的数据增强方式进行二次训练;c) Then, use the original image and the new data enhancement method for secondary training;
    在缩小的图片数据上训练模型之后,保留网络结构和权重分布,将原图以旋转、缩放、翻转、拉伸的概率组合,使用新的学习率再进行二次训练;After training the model on the reduced image data, the network structure and weight distribution are retained, the original image is combined with the probability of rotation, scaling, flipping, and stretching, and the new learning rate is used for secondary training;
    d)最后,通过调整损失函数组合方式得到不同的医学评价结果;d) Finally, different medical evaluation results are obtained by adjusting the combination of loss functions;
    通过单个最优损失函数模型和基于权重及相似性相结合的损失模型做不同层的监督信号,得到不同的评价结果。Through a single optimal loss function model and a loss model based on a combination of weights and similarities, different levels of supervision signals are used to obtain different evaluation results.
  3. 根据权利要求2所述的一种医学图像中肝脏及其病灶分割的方法,其特征在于:步骤c)使用原图和新的数据增强方式进行二次训练,原图即是512×512大小,对图片进行旋转、缩放、翻转、拉伸操作,以随机概率组合,采用指数衰减学习率,调整每轮衰减大小调节变化,方程如下:The method for segmenting the liver and its lesions in a medical image according to claim 2, characterized in that: step c) uses the original image and the new data enhancement method for secondary training, the original image is 512×512 size, Rotate, zoom, flip, and stretch the picture, combine them with random probability, and use the exponential decay learning rate to adjust each round of attenuation. The equation is as follows:
    Figure PCTCN2020131402-appb-100001
    Figure PCTCN2020131402-appb-100001
    上述公式,衰减后的学习率decayed_learning_rate,由先设定的初始学习率learning_rate乘以基数衰减速度decay_steps,衰减速度设定为0.8~0.9,每global_step步进行衰减,global_step是当前迭代轮数,即多少轮可迭代完所有的样本数据,初始学习率设定为1e-3~3e-3,在原图训练时设定为1e-4~3e-4,结果是学习率以基数0.8~0.9每一轮的步数进行衰减。The above formula, the decayed learning rate decayed_learning_rate, multiply the initial learning rate set first by the base decay speed decay_steps, the decay rate is set to 0.8~0.9, decay every global_step step, global_step is the current iteration round number, that is, how many The round can iterate through all the sample data, the initial learning rate is set to 1e-3 ~ 3e-3, and the original image training is set to 1e-4 ~ 3e-4, the result is that the learning rate is based on the base 0.8 ~ 0.9 per round The number of steps is attenuated.
  4. 根据权利要求2所述的一种医学图像中肝脏及其病灶分割的方法,其特征在于:步骤d)通过调整损失函数组合方式得到不同的医学评价结果,损失函数具体为DL、GDL、TL,采用的三个损失函数如下公式,分别选取对于肝脏和病灶适用的损失函数,DL用于评估预测集合和真实集合的相似度,用于样本不均衡的情况,表达式如下:The method for segmenting liver and its lesions in medical images according to claim 2, characterized in that: step d) obtain different medical evaluation results by adjusting the combination of loss functions, the loss functions are specifically DL, GDL, TL, The three loss functions used are as follows, and the loss functions applicable to the liver and the lesion are selected respectively. DL is used to evaluate the similarity between the predicted set and the true set, and is used in the case of unbalanced samples. The expression is as follows:
    Figure PCTCN2020131402-appb-100002
    Figure PCTCN2020131402-appb-100002
    分母的量化计算采取元素平方再求和的方法,其中k和t分别代表预测区域和真值区域元素,ij代表遍历其中元素;是一种集合相似度度量函数,通常用于计算两个样本的相似度,范围为[0,1],分子中的系数,因分母存在重复计算k和t之间的共同元素的原因,最终由各类别的2倍点乘除以各自元素的平方和得到损失值;The quantitative calculation of the denominator adopts the method of element square and then summation, where k and t represent the elements of the prediction area and the true value area respectively, and ij represents the traversal of the elements; it is a set similarity measurement function, usually used to calculate the two samples Similarity, the range is [0,1], the coefficient in the numerator, due to the repeated calculation of the common element between k and t in the denominator, and finally the loss value is obtained by multiplying the double points of each category by the sum of the squares of the respective elements ;
    GDL:当肝脏病灶有多个分割区域时,针对每一类有一个Dice,而GDL将多个类别进行整合,采用一个指标进行量化计算,公式如下:GDL: When the liver lesion has multiple segmented areas, there is one Dice for each category, and GDL integrates multiple categories and uses one index for quantitative calculation. The formula is as follows:
    Figure PCTCN2020131402-appb-100003
    Figure PCTCN2020131402-appb-100003
    其中k ij为类别i在第j个像素的真实值,t ij为相应的预测概率值;相比于DL,多了作为每个类别的权重wi,wi用于维系病灶区域和DL系数之间的平衡; Among them, k ij is the true value of category i in the jth pixel, and t ij is the corresponding predicted probability value; compared to DL, there is more weight wi as each category, wi is used to maintain the relationship between the lesion area and the DL coefficient Balance
    TL公式如下:The TL formula is as follows:
    Figure PCTCN2020131402-appb-100004
    Figure PCTCN2020131402-appb-100004
    其中k ij为类别i在第j个像素的真实值,t ij为相应的预测概率值; Where k ij is the true value of category i at the jth pixel, and t ij is the corresponding predicted probability value;
    α和β分别控制假阳性和假阴性的比重。α and β respectively control the proportion of false positives and false negatives.
  5. 根据权利要求4所述的一种医学图像中肝脏及其病灶分割的方法,其特征在于:当α=β=0.5,TL系数即是DL系数。The method for segmenting the liver and its lesions in a medical image according to claim 4, wherein when α=β=0.5, the TL coefficient is the DL coefficient.
PCT/CN2020/131402 2020-03-16 2020-11-25 Method for segmenting liver and focus thereof in medical image WO2021184817A1 (en)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
CN202010180378.XA CN111402268B (en) 2020-03-16 2020-03-16 Liver in medical image and focus segmentation method thereof
CN202010180378.X 2020-03-16

Publications (1)

Publication Number Publication Date
WO2021184817A1 true WO2021184817A1 (en) 2021-09-23

Family

ID=71428820

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/CN2020/131402 WO2021184817A1 (en) 2020-03-16 2020-11-25 Method for segmenting liver and focus thereof in medical image

Country Status (2)

Country Link
CN (1) CN111402268B (en)
WO (1) WO2021184817A1 (en)

Cited By (26)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN113902752A (en) * 2021-12-06 2022-01-07 湖南大学 Medical image segmentation method and device and computer readable storage medium
CN114049339A (en) * 2021-11-22 2022-02-15 江苏科技大学 Fetal cerebellum ultrasonic image segmentation method based on convolutional neural network
CN114067092A (en) * 2022-01-17 2022-02-18 山东药品食品职业学院 Fatty liver B-mode ultrasound image classification method based on DenseNet and lightGBM
CN114155195A (en) * 2021-11-01 2022-03-08 中南大学湘雅医院 Brain tumor segmentation quality evaluation method, device and medium based on deep learning
CN114358144A (en) * 2021-12-16 2022-04-15 西南交通大学 Image segmentation quality evaluation method
CN114359194A (en) * 2021-12-27 2022-04-15 浙江大学 Multi-mode stroke infarct area image processing method based on improved U-Net network
CN114511581A (en) * 2022-04-20 2022-05-17 四川大学华西医院 Multi-task multi-resolution collaborative esophageal cancer lesion segmentation method and device
CN114511523A (en) * 2022-01-21 2022-05-17 中山大学 Gastric cancer molecular subtype classification method and device based on self-supervision learning
CN114511850A (en) * 2021-12-30 2022-05-17 广西慧云信息技术有限公司 Method for identifying image of fruit size and granule of sunshine rose grape
CN114511728A (en) * 2021-12-24 2022-05-17 之江实验室 Electronic endoscope esophageal focus intelligent detection model establishing method
CN114693689A (en) * 2022-03-02 2022-07-01 西北工业大学 Adaptive neural network segmentation model construction method for medical image
CN114764811A (en) * 2022-03-14 2022-07-19 什维新智医疗科技(上海)有限公司 Dynamic ultrasonic video-based real-time lesion area segmentation device
CN114818838A (en) * 2022-06-30 2022-07-29 中国科学院国家空间科学中心 Low signal-to-noise ratio moving point target detection method based on pixel time domain distribution learning
CN114897922A (en) * 2022-04-03 2022-08-12 西北工业大学 Histopathology image segmentation method based on deep reinforcement learning
CN114937171A (en) * 2022-05-11 2022-08-23 复旦大学 Alzheimer's classification system based on deep learning
CN115019110A (en) * 2022-07-13 2022-09-06 北京深睿博联科技有限责任公司 Focus identification method and device based on chest image
CN115661144A (en) * 2022-12-15 2023-01-31 湖南工商大学 Self-adaptive medical image segmentation method based on deformable U-Net
CN115880262A (en) * 2022-12-20 2023-03-31 桂林电子科技大学 Weakly supervised pathological image tissue segmentation method based on online noise suppression strategy
CN115937423A (en) * 2022-12-13 2023-04-07 西安电子科技大学 Three-dimensional intelligent reconstruction method for liver tumor medical image
CN116563285A (en) * 2023-07-10 2023-08-08 邦世科技(南京)有限公司 Focus characteristic identifying and dividing method and system based on full neural network
CN116630971A (en) * 2023-07-24 2023-08-22 安徽大学 Wheat scab spore segmentation method based on CRF_Resunate++ network
CN116777793A (en) * 2023-07-31 2023-09-19 深圳扬奇医芯智能科技有限公司 Medical image prompt denoising technology based on logarithmic guide contrast learning
CN116894820A (en) * 2023-07-13 2023-10-17 国药(武汉)精准医疗科技有限公司 Pigment skin disease classification detection method, device, equipment and storage medium
CN117409019A (en) * 2023-09-15 2024-01-16 哈尔滨工业大学(深圳)(哈尔滨工业大学深圳科技创新研究院) Multi-mode brain tumor image segmentation method and system based on ensemble learning
CN117524427A (en) * 2024-01-05 2024-02-06 莱凯医疗器械(北京)有限公司 Intelligent medical image analysis method
CN117830332A (en) * 2024-01-09 2024-04-05 四川大学 Medical image segmentation method based on weak supervision

Families Citing this family (19)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN111402268B (en) * 2020-03-16 2023-05-23 苏州科技大学 Liver in medical image and focus segmentation method thereof
CN111950595A (en) * 2020-07-14 2020-11-17 十堰市太和医院(湖北医药学院附属医院) Liver focus image processing method, system, storage medium, program, and terminal
CN111860840B (en) * 2020-07-28 2023-10-17 上海联影医疗科技股份有限公司 Deep learning model training method, device, computer equipment and storage medium
CN112070685B (en) * 2020-08-10 2023-08-18 武汉大学 Method for predicting dynamic soft tissue movement of HIFU treatment system
US11694329B2 (en) 2020-10-30 2023-07-04 International Business Machines Corporation Logistic model to determine 3D z-wise lesion connectivity
US11688517B2 (en) 2020-10-30 2023-06-27 Guerbet Multiple operating point false positive removal for lesion identification
US11688063B2 (en) 2020-10-30 2023-06-27 Guerbet Ensemble machine learning model architecture for lesion detection
US11749401B2 (en) 2020-10-30 2023-09-05 Guerbet Seed relabeling for seed-based segmentation of a medical image
US11436724B2 (en) 2020-10-30 2022-09-06 International Business Machines Corporation Lesion detection artificial intelligence pipeline computing system
US11587236B2 (en) 2020-10-30 2023-02-21 International Business Machines Corporation Refining lesion contours with combined active contour and inpainting
CN112819032B (en) * 2021-01-11 2023-10-27 平安科技(深圳)有限公司 Multi-model-based slice feature classification method, device, equipment and medium
CN113052857A (en) * 2021-03-22 2021-06-29 山西三友和智慧信息技术股份有限公司 Lung lesion image segmentation method based on CovSegNet
CN113159147B (en) * 2021-04-08 2023-09-26 平安科技(深圳)有限公司 Image recognition method and device based on neural network and electronic equipment
CN113112475B (en) * 2021-04-13 2023-04-18 五邑大学 Traditional Chinese medicine ear five-organ region segmentation method and device based on machine learning
CN113240014B (en) * 2021-05-18 2022-05-31 长春理工大学 Application method of class II segmentation loss function in achieving class II segmentation of intervertebral disc tissue image
CN113378984B (en) * 2021-07-05 2023-05-02 国药(武汉)医学实验室有限公司 Medical image classification method, system, terminal and storage medium
CN113902761B (en) * 2021-11-02 2024-04-16 大连理工大学 Knowledge distillation-based unsupervised segmentation method for lung disease focus
CN114066871B (en) * 2021-11-19 2024-06-21 江苏科技大学 Method for training new coronal pneumonia focus area segmentation model
CN114511599B (en) * 2022-01-20 2022-09-20 推想医疗科技股份有限公司 Model training method and device, medical image registration method and device

Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN108492286A (en) * 2018-03-13 2018-09-04 成都大学 A kind of medical image cutting method based on the U-shaped convolutional neural networks of binary channel
CN110120033A (en) * 2019-04-12 2019-08-13 天津大学 Based on improved U-Net neural network three-dimensional brain tumor image partition method
CN110853038A (en) * 2019-10-15 2020-02-28 哈尔滨工程大学 DN-U-net network method for liver tumor CT image segmentation technology
US20200074637A1 (en) * 2018-08-28 2020-03-05 International Business Machines Corporation 3d segmentation with exponential logarithmic loss for highly unbalanced object sizes
CN111402268A (en) * 2020-03-16 2020-07-10 苏州科技大学 Method for segmenting liver and focus thereof in medical image

Family Cites Families (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN108937934A (en) * 2018-05-07 2018-12-07 武汉科技大学 A kind of detection of Brain MRI hippocampus and dividing method based on deep learning
WO2020028382A1 (en) * 2018-07-30 2020-02-06 Memorial Sloan Kettering Cancer Center Multi-modal, multi-resolution deep learning neural networks for segmentation, outcomes prediction and longitudinal response monitoring to immunotherapy and radiotherapy
CN109410220B (en) * 2018-10-16 2019-12-24 腾讯科技(深圳)有限公司 Image segmentation method and device, computer equipment and storage medium
CN109614921B (en) * 2018-12-07 2022-09-30 安徽大学 Cell segmentation method based on semi-supervised learning of confrontation generation network
CN110047078B (en) * 2019-04-18 2021-11-09 北京市商汤科技开发有限公司 Image processing method and device, electronic equipment and storage medium
CN110232721B (en) * 2019-05-16 2023-04-07 福建自贸试验区厦门片区Manteia数据科技有限公司 Training method and device for automatic sketching model of organs at risk
CN110689543A (en) * 2019-09-19 2020-01-14 天津大学 Improved convolutional neural network brain tumor image segmentation method based on attention mechanism

Patent Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN108492286A (en) * 2018-03-13 2018-09-04 成都大学 A kind of medical image cutting method based on the U-shaped convolutional neural networks of binary channel
US20200074637A1 (en) * 2018-08-28 2020-03-05 International Business Machines Corporation 3d segmentation with exponential logarithmic loss for highly unbalanced object sizes
CN110120033A (en) * 2019-04-12 2019-08-13 天津大学 Based on improved U-Net neural network three-dimensional brain tumor image partition method
CN110853038A (en) * 2019-10-15 2020-02-28 哈尔滨工程大学 DN-U-net network method for liver tumor CT image segmentation technology
CN111402268A (en) * 2020-03-16 2020-07-10 苏州科技大学 Method for segmenting liver and focus thereof in medical image

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
ABRAHAM NABILA; KHAN NAIMUL MEFRAZ: "A Novel Focal Tversky Loss Function With Improved Attention U-Net for Lesion Segmentation", 2019 IEEE 16TH INTERNATIONAL SYMPOSIUM ON BIOMEDICAL IMAGING (ISBI 2019), 8 April 2019 (2019-04-08), pages 683 - 687, XP033576451, DOI: 10.1109/ISBI.2019.8759329 *

Cited By (44)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN114155195A (en) * 2021-11-01 2022-03-08 中南大学湘雅医院 Brain tumor segmentation quality evaluation method, device and medium based on deep learning
CN114155195B (en) * 2021-11-01 2023-04-07 中南大学湘雅医院 Brain tumor segmentation quality evaluation method, device and medium based on deep learning
CN114049339A (en) * 2021-11-22 2022-02-15 江苏科技大学 Fetal cerebellum ultrasonic image segmentation method based on convolutional neural network
CN114049339B (en) * 2021-11-22 2023-05-12 江苏科技大学 Fetal cerebellum ultrasonic image segmentation method based on convolutional neural network
CN113902752A (en) * 2021-12-06 2022-01-07 湖南大学 Medical image segmentation method and device and computer readable storage medium
CN113902752B (en) * 2021-12-06 2022-02-22 湖南大学 Medical image segmentation method and device and computer readable storage medium
CN114358144B (en) * 2021-12-16 2023-09-26 西南交通大学 Image segmentation quality assessment method
CN114358144A (en) * 2021-12-16 2022-04-15 西南交通大学 Image segmentation quality evaluation method
CN114511728A (en) * 2021-12-24 2022-05-17 之江实验室 Electronic endoscope esophageal focus intelligent detection model establishing method
CN114511728B (en) * 2021-12-24 2024-05-14 之江实验室 Method for establishing intelligent detection model of esophageal lesion of electronic endoscope
CN114359194A (en) * 2021-12-27 2022-04-15 浙江大学 Multi-mode stroke infarct area image processing method based on improved U-Net network
CN114511850B (en) * 2021-12-30 2024-05-14 广西慧云信息技术有限公司 Method for identifying size particle image of sunlight rose grape fruit
CN114511850A (en) * 2021-12-30 2022-05-17 广西慧云信息技术有限公司 Method for identifying image of fruit size and granule of sunshine rose grape
CN114067092A (en) * 2022-01-17 2022-02-18 山东药品食品职业学院 Fatty liver B-mode ultrasound image classification method based on DenseNet and lightGBM
CN114067092B (en) * 2022-01-17 2022-04-19 山东药品食品职业学院 Fatty liver B-mode ultrasound image classification method based on DenseNet and lightGBM
CN114511523A (en) * 2022-01-21 2022-05-17 中山大学 Gastric cancer molecular subtype classification method and device based on self-supervision learning
CN114511523B (en) * 2022-01-21 2024-05-31 中山大学 Gastric cancer molecular subtype classification method and device based on self-supervision learning
CN114693689B (en) * 2022-03-02 2024-03-15 西北工业大学 Self-adaptive neural network segmentation model construction method for medical image
CN114693689A (en) * 2022-03-02 2022-07-01 西北工业大学 Adaptive neural network segmentation model construction method for medical image
CN114764811A (en) * 2022-03-14 2022-07-19 什维新智医疗科技(上海)有限公司 Dynamic ultrasonic video-based real-time lesion area segmentation device
CN114897922A (en) * 2022-04-03 2022-08-12 西北工业大学 Histopathology image segmentation method based on deep reinforcement learning
CN114897922B (en) * 2022-04-03 2024-04-26 西北工业大学 Tissue pathology image segmentation method based on deep reinforcement learning
CN114511581A (en) * 2022-04-20 2022-05-17 四川大学华西医院 Multi-task multi-resolution collaborative esophageal cancer lesion segmentation method and device
CN114937171A (en) * 2022-05-11 2022-08-23 复旦大学 Alzheimer's classification system based on deep learning
CN114937171B (en) * 2022-05-11 2023-06-09 复旦大学 Deep learning-based Alzheimer's classification system
CN114818838A (en) * 2022-06-30 2022-07-29 中国科学院国家空间科学中心 Low signal-to-noise ratio moving point target detection method based on pixel time domain distribution learning
CN115019110A (en) * 2022-07-13 2022-09-06 北京深睿博联科技有限责任公司 Focus identification method and device based on chest image
CN115937423B (en) * 2022-12-13 2023-08-15 西安电子科技大学 Three-dimensional intelligent reconstruction method for liver tumor medical image
CN115937423A (en) * 2022-12-13 2023-04-07 西安电子科技大学 Three-dimensional intelligent reconstruction method for liver tumor medical image
CN115661144A (en) * 2022-12-15 2023-01-31 湖南工商大学 Self-adaptive medical image segmentation method based on deformable U-Net
US11935279B1 (en) 2022-12-20 2024-03-19 Guilin University Of Electronic Technology Weakly supervised pathological image tissue segmentation method based on online noise suppression strategy
CN115880262B (en) * 2022-12-20 2023-09-05 桂林电子科技大学 Weak supervision pathological image tissue segmentation method based on online noise suppression strategy
CN115880262A (en) * 2022-12-20 2023-03-31 桂林电子科技大学 Weakly supervised pathological image tissue segmentation method based on online noise suppression strategy
CN116563285A (en) * 2023-07-10 2023-08-08 邦世科技(南京)有限公司 Focus characteristic identifying and dividing method and system based on full neural network
CN116563285B (en) * 2023-07-10 2023-09-19 邦世科技(南京)有限公司 Focus characteristic identifying and dividing method and system based on full neural network
CN116894820B (en) * 2023-07-13 2024-04-19 国药(武汉)精准医疗科技有限公司 Pigment skin disease classification detection method, device, equipment and storage medium
CN116894820A (en) * 2023-07-13 2023-10-17 国药(武汉)精准医疗科技有限公司 Pigment skin disease classification detection method, device, equipment and storage medium
CN116630971A (en) * 2023-07-24 2023-08-22 安徽大学 Wheat scab spore segmentation method based on CRF_Resunate++ network
CN116630971B (en) * 2023-07-24 2023-09-29 安徽大学 Wheat scab spore segmentation method based on CRF_Resunate++ network
CN116777793A (en) * 2023-07-31 2023-09-19 深圳扬奇医芯智能科技有限公司 Medical image prompt denoising technology based on logarithmic guide contrast learning
CN117409019A (en) * 2023-09-15 2024-01-16 哈尔滨工业大学(深圳)(哈尔滨工业大学深圳科技创新研究院) Multi-mode brain tumor image segmentation method and system based on ensemble learning
CN117524427A (en) * 2024-01-05 2024-02-06 莱凯医疗器械(北京)有限公司 Intelligent medical image analysis method
CN117524427B (en) * 2024-01-05 2024-04-02 莱凯医疗器械(北京)有限公司 Intelligent medical image analysis method
CN117830332A (en) * 2024-01-09 2024-04-05 四川大学 Medical image segmentation method based on weak supervision

Also Published As

Publication number Publication date
CN111402268A (en) 2020-07-10
CN111402268B (en) 2023-05-23

Similar Documents

Publication Publication Date Title
WO2021184817A1 (en) Method for segmenting liver and focus thereof in medical image
CN110992382B (en) Fundus image optic cup optic disc segmentation method and system for assisting glaucoma screening
US11580646B2 (en) Medical image segmentation method based on U-Net
Fuhrman et al. A review of explainable and interpretable AI with applications in COVID‐19 imaging
CN110428432B (en) Deep neural network algorithm for automatically segmenting colon gland image
CN107341812B (en) A kind of sequence Lung neoplasm image partition method based on super-pixel and Density Clustering
Tian et al. Multi-path convolutional neural network in fundus segmentation of blood vessels
Guo et al. Msanet: multiscale aggregation network integrating spatial and channel information for lung nodule detection
CN115375711A (en) Image segmentation method of global context attention network based on multi-scale fusion
CN115512110A (en) Medical image tumor segmentation method related to cross-modal attention mechanism
CN117036288A (en) Tumor subtype diagnosis method for full-slice pathological image
Sun et al. COVID-19 CT image segmentation method based on swin transformer
CN117611599B (en) Blood vessel segmentation method and system integrating centre line diagram and contrast enhancement network
Peng et al. Cov-TransNet: Dual branch fusion network with transformer for COVID-19 infection segmentation
Nan et al. Automatic fine-grained glomerular lesion recognition in kidney pathology
Pan et al. A multilevel remote relational modeling network for accurate segmentation of fundus blood vessels
Yang et al. NAUNet: lightweight retinal vessel segmentation network with nested connections and efficient attention
Gao et al. A semi-supervised learning framework for micropapillary adenocarcinoma detection
Xu et al. Big Model and Small Model: Remote modeling and local information extraction module for medical image segmentation
CN111768420A (en) Cell image segmentation model
CN116188455A (en) Breast cancer cell image segmentation method based on U-Net
Yan et al. Two and multiple categorization of breast pathological images by transfer learning
Ni et al. FSE-Net: Rethinking the up-sampling operation in encoder-decoder structure for retinal vessel segmentation
CN115131628A (en) Mammary gland image classification method and equipment based on typing auxiliary information
Wu et al. Mscan: Multi-scale channel attention for fundus retinal vessel segmentation

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 20925775

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: DE

122 Ep: pct application non-entry in european phase

Ref document number: 20925775

Country of ref document: EP

Kind code of ref document: A1