CN116206109B - A liver tumor segmentation method based on cascade network - Google Patents

A liver tumor segmentation method based on cascade network Download PDF

Info

Publication number
CN116206109B
CN116206109B CN202310146446.4A CN202310146446A CN116206109B CN 116206109 B CN116206109 B CN 116206109B CN 202310146446 A CN202310146446 A CN 202310146446A CN 116206109 B CN116206109 B CN 116206109B
Authority
CN
China
Prior art keywords
input
layer
output
module
residual
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN202310146446.4A
Other languages
Chinese (zh)
Other versions
CN116206109A (en
Inventor
李为坤
张文辉
林镇源
蒋小莲
李佳玮
陈晓婕
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Guilin University of Electronic Technology
Original Assignee
Guilin University of Electronic Technology
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Guilin University of Electronic Technology filed Critical Guilin University of Electronic Technology
Priority to CN202310146446.4A priority Critical patent/CN116206109B/en
Publication of CN116206109A publication Critical patent/CN116206109A/en
Application granted granted Critical
Publication of CN116206109B publication Critical patent/CN116206109B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/20Image preprocessing
    • G06V10/26Segmentation of patterns in the image field; Cutting or merging of image elements to establish the pattern region, e.g. clustering-based techniques; Detection of occlusion
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/08Learning methods
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/20Image preprocessing
    • G06V10/25Determination of region of interest [ROI] or a volume of interest [VOI]
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/40Extraction of image or video features
    • G06V10/44Local feature extraction by analysis of parts of the pattern, e.g. by detecting edges, contours, loops, corners, strokes or intersections; Connectivity analysis, e.g. of connected components
    • G06V10/443Local feature extraction by analysis of parts of the pattern, e.g. by detecting edges, contours, loops, corners, strokes or intersections; Connectivity analysis, e.g. of connected components by matching or filtering
    • G06V10/449Biologically inspired filters, e.g. difference of Gaussians [DoG] or Gabor filters
    • G06V10/451Biologically inspired filters, e.g. difference of Gaussians [DoG] or Gabor filters with interaction between the filter responses, e.g. cortical complex cells
    • G06V10/454Integrating the filters into a hierarchical structure, e.g. convolutional neural networks [CNN]
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/40Extraction of image or video features
    • G06V10/52Scale-space analysis, e.g. wavelet analysis
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/70Arrangements for image or video recognition or understanding using pattern recognition or machine learning
    • G06V10/77Processing image or video features in feature spaces; using data integration or data reduction, e.g. principal component analysis [PCA] or independent component analysis [ICA] or self-organising maps [SOM]; Blind source separation
    • G06V10/774Generating sets of training patterns; Bootstrap methods, e.g. bagging or boosting
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/70Arrangements for image or video recognition or understanding using pattern recognition or machine learning
    • G06V10/82Arrangements for image or video recognition or understanding using pattern recognition or machine learning using neural networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V2201/00Indexing scheme relating to image or video recognition or understanding
    • G06V2201/03Recognition of patterns in medical or anatomical images

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Multimedia (AREA)
  • Evolutionary Computation (AREA)
  • Artificial Intelligence (AREA)
  • Health & Medical Sciences (AREA)
  • General Health & Medical Sciences (AREA)
  • Software Systems (AREA)
  • Computing Systems (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Molecular Biology (AREA)
  • Biomedical Technology (AREA)
  • Databases & Information Systems (AREA)
  • Medical Informatics (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Biodiversity & Conservation Biology (AREA)
  • Biophysics (AREA)
  • Computational Linguistics (AREA)
  • Data Mining & Analysis (AREA)
  • General Engineering & Computer Science (AREA)
  • Mathematical Physics (AREA)
  • Apparatus For Radiation Diagnosis (AREA)
  • Image Analysis (AREA)

Abstract

本发明公开一种基于级联网络的肝脏肿瘤分割方法,采取肝脏分割网络和肝肿瘤分割网络级联的方式,肝脏分割网络先从CT图像中分割肝脏,提取出肿瘤感兴趣区域,再将肿瘤感兴趣区域作为肝肿瘤分割网络的输入,进一步实现对肝脏肿瘤的精准分割,这解决了因肿瘤在整幅CT图像中占比小而造成的数据不平衡问题。肝脏分割网络和肝肿瘤分割网络均以残差网络作为骨架,残差网络极大提高了可以有效训练的网络的深度,加速训练网络的收敛,减少模型退化,从而有效避免了随着网络层数加深梯度消失的问题,解决了因网络过深而出现的梯度弥散问题。

The invention discloses a liver tumor segmentation method based on a cascade network, which adopts a cascade method of liver segmentation network and liver tumor segmentation network. The liver segmentation network first segments the liver from CT images, extracts the tumor area of interest, and then divides the tumor into The region of interest is used as the input of the liver tumor segmentation network to further achieve accurate segmentation of liver tumors, which solves the data imbalance problem caused by the small proportion of tumors in the entire CT image. Both the liver segmentation network and the liver tumor segmentation network use the residual network as the skeleton. The residual network greatly improves the depth of the network that can be effectively trained, accelerates the convergence of the training network, and reduces model degradation, thus effectively avoiding the problem of increasing the number of network layers. It deepens the problem of vanishing gradients and solves the problem of gradient dispersion that occurs when the network is too deep.

Description

Liver tumor segmentation method based on cascade network
Technical Field
The invention relates to the technical field of image segmentation, in particular to a liver tumor segmentation method based on a cascade network.
Background
Liver cancer is the most common and most fatal tumor in the world, and seriously threatens the life health of people. According to the national cancer center data, the incidence rate of liver cancer in China is 5 th in all malignant tumors, and the death rate is 2 nd higher. With the development of science and technology, the computer application technology and the medical informatization level are greatly developed, and the medical image facilities are more perfect. The computed tomography imaging (computed tomography, CT) has the characteristics of quick scanning time, high image resolution and the like, and is a diagnosis mode commonly adopted by liver lesions at present. Because the liver tumor in the CT image is usually characterized by low contrast, fuzzy boundary, unfixed size, shape, position and quantity, and the like, the current clinical liver tumor segmentation needs manual intervention, namely, a doctor with abundant experience marks the outline manually, which is time-consuming and labor-consuming, is difficult to effectively adapt to the complexity and diversity of the liver tumor, has poor segmentation accuracy and lower efficiency on the target, and cannot realize the automatic segmentation of the tumor region; and the tumor area of the liver CT image of the same patient can cause different results under the mark of different doctors, and the experience and skill of the doctors are seriously depended. Therefore, the research on the accurate and efficient automatic liver tumor segmentation method has important significance for clinical diagnosis and treatment of liver cancer.
In recent years, deep learning techniques have been rapidly developed and widely used in the field of medical image segmentation. Ronneeberger et al propose a U-shaped convolutional network (Unet) that introduces a jumping connection into the convolutional network for the first time, which achieves end-to-end semantic segmentation of images through encoding-decoding operations, an encoder downsamples extracted features to capture contextual information of the images, and a decoder upsamples extracted features to precisely locate segmented regions. Li et al propose a bottleneck-supervised Unet model (bottleneck supervised Unet, BS-Unet) which is a hybrid tightly-coupled structure that is partitioned by making full use of information between layers of the network. Schlemper et al incorporate the attention mechanism into the Unet network and propose an A-Unet (attention Unet) model that automatically learns regional features related to segmentation tasks and suppresses irrelevant features. Although these networks are widely used in the field of liver tumor segmentation, there are still problems of inaccurate boundary segmentation, difficult detection of small tumors, difficult tumor segmentation due to unbalanced data types, and the like.
Disclosure of Invention
The invention aims to solve the problems of difficult segmentation and inaccurate boundary segmentation in the existing liver tumor segmentation and provides a liver tumor segmentation method based on a cascade network.
In order to solve the problems, the invention is realized by the following technical scheme:
a liver tumor segmentation method based on cascade network comprises the following steps:
firstly, constructing a liver tumor segmentation model based on a cascade network; the liver tumor segmentation model based on the cascade network consists of a liver segmentation network, a liver tumor segmentation network and a characteristic addition layer; the input of the liver segmentation network is used as the input of a liver tumor segmentation model based on a cascade network, the input and the output of the liver segmentation network are simultaneously connected with the input of the feature addition layer, the output of the feature addition layer is connected with the input of the liver tumor segmentation network, and the output of the liver tumor segmentation network is used as the output of the liver tumor segmentation model based on the cascade network;
then, constructing a training sample set by utilizing CT images which have been segmented into liver tumors in advance, and performing deep learning training on the constructed liver tumor segmentation model based on the cascade network by utilizing the training sample set to obtain a trained liver tumor segmentation model based on the cascade network;
and finally, sending the CT image which is currently required to be segmented into a trained liver tumor segmentation model based on a cascade network, and obtaining the CT image of segmented liver tumor.
The liver segmentation network and the liver tumor segmentation network are both separable convolution residual segmentation networks based on the mixed depth; the separable convolution residual error segmentation network based on the mixed depth consists of 1 input layer, 2 convolution layers, 5 first residual error modules, 4 second residual error modules, 4 pooling modules, 4 up-sampling modules and 1 output layer; the input of the input layer is used as the input of the separable convolution residual segmentation network based on the mixed depth; the output of the input layer is connected with the input of a first residual module through a first convolution layer, the output of the first residual module is connected with the input of a first pooling module and the first input of a fourth upsampling module, the output of the first pooling module is connected with the input of a second first residual module, the output of the second first residual module is connected with the input of a second pooling module and the first input of a third upsampling module, the output of the second pooling module is connected with the input of a third first residual module, the output of the third first residual module is connected with the input of a third pooling module and the first input of a second upsampling module, the output of the third pooling module is connected with the input of a fourth first residual module, the output of the fourth pooling module is connected with the input of a fourth pooling module and the first input of the first upsampling module, and the output of the fourth pooling module is connected with the input of a fifth first residual module; the output of the fifth first residual module is connected with the second input of the first up-sampling module, the output of the first up-sampling module is connected with the input of the first second residual module, the output of the first second residual module is connected with the second input of the second up-sampling module, the output of the second up-sampling module is connected with the second input of the third up-sampling module, the output of the third up-sampling module is connected with the input of the third second residual module, the output of the third up-sampling module is connected with the second input of the fourth up-sampling module, the output of the fourth up-sampling module is connected with the input of the fourth third residual module, and the output of the fourth second residual module is connected with the input of the output layer through the second convolution layer; the output of the output layer is output as a separable convolutional residual segmentation network based on the blend depth.
The first residual module of the liver segmentation network is different from the first residual module of the liver tumor segmentation network. The first residual error module of the liver segmentation network consists of 2 mixed depth separable convolution layers, 2 convolution layers and 1 characteristic addition layer; the input of the first mixed depth separable convolution layer is used as the input of a first residual error module of the liver segmentation network, the output of the first mixed depth separable convolution layer is connected with the input of the second mixed depth separable convolution layer, and the output of the second mixed depth separable convolution layer is connected with the input of the first convolution layer; the input of the second convolution layer is connected with the input of the first mixed depth separable convolution layer; the outputs of the first convolution layer and the second convolution layer are simultaneously connected with the input of the feature addition layer, and the output of the feature addition layer is used as the output of a first residual error module of the liver segmentation network. The first residual error module of the liver tumor segmentation network consists of 2 mixed depth separable convolution layers, 2 convolution layers, 1 coordinate attention mechanism layer and 1 characteristic addition layer; the input of the first mixed depth separable convolution layer is used as the input of a first residual error module of the liver segmentation network, the output of the first mixed depth separable convolution layer is connected with the input of the second mixed depth separable convolution layer, the output of the second mixed depth separable convolution layer is connected with the input of the coordinate attention mechanism layer, and the output of the coordinate attention mechanism layer is connected with the input of the first convolution layer; the input of the second convolution layer is connected with the input of the first mixed depth separable convolution layer; the outputs of the first convolution layer and the second convolution layer are simultaneously connected with the input of the feature addition layer, and the output of the feature addition layer is used as the output of a first residual error module of the liver segmentation network.
The second residual error module consists of 3 convolution layers and 1 characteristic addition layer; the input of the first convolution layer is used as the input of the second residual error module, and the output of the first convolution layer is connected with the input of the second convolution layer; the input of the third convolution layer is connected with the input of the first convolution layer; the outputs of the second convolution layer and the third convolution layer are simultaneously connected with the input of the characteristic addition layer; the output of the feature addition layer serves as the output of the second residual block.
The pooling module consists of 1 maximum pooling layer, 1 convolution layer and 1 splicing layer; the input of the maximum pooling layer and the output of the convolution layer are jointly used as the input of the pooling module, the output of the maximum pooling layer and the output of the convolution layer are jointly connected with the input of the splicing layer, and the output of the splicing layer is used as the output of the pooling module.
The up-sampling module consists of 1 bilinear interpolation layer and 1 splicing layer; the input of the bilinear interpolation layer is used as the first input of the up-sampling module, the output of the bilinear interpolation layer is connected with one input of the splicing layer, the other input of the splicing layer is used as the second input of the up-sampling module, and the output of the splicing layer is used as the output of the up-sampling module.
Compared with the prior art, the invention has the following characteristics:
1. the liver segmentation network and the liver tumor segmentation network are adopted to carry out cascading, the liver segmentation network segments the liver from the CT image firstly, a tumor region of interest is extracted, then the tumor region of interest is used as the input of the liver tumor segmentation network, the accurate segmentation of the liver tumor is further realized, and the problem of unbalanced data caused by small proportion of the tumor in the whole CT image is solved;
2. the liver segmentation network and the liver tumor segmentation network both take a residual network as a framework, the residual network greatly improves the depth of the network which can be effectively trained, accelerates the convergence of the training network, and reduces the model degradation, thereby effectively avoiding the problem of gradient disappearance along with the deepening of the network layers and solving the problem of gradient dispersion caused by the over-deep network;
3. the mixed depth separable convolution is operated on different channels by using convolution kernels with different sizes, and the multi-scale convolution kernels are fused into a single convolution operation, so that characteristic modes with different resolutions are captured, and edge details and deeper small target characteristics are extracted; through strengthening the receptive field of the segmented network feature map and fully utilizing the channel and space structure information, pixel level detail and space information can be captured better, so that the segmentation performance of the network on medical images is improved.
4. The coordinate attention mechanism can capture cross-channel information so that the model can more accurately locate and identify the lesion area.
Drawings
Fig. 1 is a schematic diagram of a liver tumor segmentation model based on a cascade network.
Fig. 2 is a schematic diagram of a separable convolutional residual partitioning network based on hybrid depth (CMDCRA-UNet).
Fig. 3 is a schematic diagram of a first Residual block (a) of a liver segmentation network and (b) of a liver tumor segmentation network.
Fig. 4 is a schematic diagram of a second Residual block 2.
Fig. 5 is a schematic diagram of a pooling module (Pool).
Fig. 6 is a schematic diagram of an Up sampling module (Up Sample).
Detailed Description
The present invention will be further described in detail with reference to specific examples in order to make the objects, technical solutions and advantages of the present invention more apparent.
A liver tumor segmentation method based on cascade network, firstly constructing a liver tumor segmentation model based on cascade network; then constructing a training sample set by utilizing CT images which have been segmented into liver tumors in advance, and performing deep learning training on the constructed liver tumor segmentation model based on the cascade network by utilizing the training sample set to obtain a trained liver tumor segmentation model based on the cascade network; and finally, sending the CT image which is currently required to be segmented into a trained liver tumor segmentation model based on a cascade network, and obtaining the CT image of segmented liver tumor.
Although the liver tumor density is different from the normal liver tissue density, the liver tumor density is similar to the tissue density of other organs in the abdomen, so that the ideal effect is difficult to obtain by directly dividing the tumor by using a network, and the liver interested region can ensure that the original CT image only keeps the liver region, thereby effectively avoiding the interference of other organs in the abdomen on the division of the liver tumor. Therefore, the liver segmentation network is firstly utilized to extract the liver region in the CT image, and then the liver tumor segmentation network is utilized to extract the tumor region in the liver region. That is, the liver tumor segmentation model based on the cascade network constructed by the invention consists of a liver segmentation network, a liver tumor segmentation network and a characteristic addition layer. As shown in fig. 1. The input of the liver segmentation network is used as the input of the liver tumor segmentation model based on the cascade network, the input and the output of the liver segmentation network are simultaneously connected with the input of the feature addition layer, the output of the feature addition layer is connected with the input of the liver tumor segmentation network, and the output of the liver tumor segmentation network is used as the output of the liver tumor segmentation model based on the cascade network. The liver segmentation network segments the liver from the original CT image, extracts a tumor region of interest, and inputs the tumor region of interest and the tumor region of interest as the liver tumor segmentation network, so that the liver tumor segmentation network further realizes accurate segmentation of liver tumors.
The liver segmentation network and the tumor segmentation network are both separable convolution residual segmentation networks based on the mixed depth. Based on the mixed depth separable convolution residual error segmentation network, the whole adopts an encoding-decoding architecture, and the residual error network is used as a framework. The separable convolution residual segmentation network based on the mixed depth consists of 1 input layer, 2 convolution layers, 5 first residual modules, 4 second residual modules, 4 pooling modules, 4 up-sampling modules and 1 output layer. As shown in fig. 2. The input of the input layer serves as the input to the separable convolutional residual segmentation network based on the blend depth. The output of the input layer is connected with the input of a first residual module through a first convolution layer, the output of the first residual module is connected with the input of a first pooling module and the first input of a fourth upsampling module, the output of the first pooling module is connected with the input of a second first residual module, the output of the second first residual module is connected with the input of a second pooling module and the first input of a third upsampling module, the output of the second pooling module is connected with the input of a third first residual module, the output of the third first residual module is connected with the input of a third pooling module and the first input of a second upsampling module, the output of the third pooling module is connected with the input of a fourth first residual module, the output of the fourth pooling module is connected with the input of a fourth pooling module and the first input of the first upsampling module, and the output of the fourth pooling module is connected with the input of a fifth first residual module. The output of the fifth first residual module is connected with the second input of the first up-sampling module, the output of the first up-sampling module is connected with the input of the first second residual module, the output of the first second residual module is connected with the second input of the second up-sampling module, the output of the second up-sampling module is connected with the second input of the third up-sampling module, the output of the third up-sampling module is connected with the input of the third second residual module, the output of the third second residual module is connected with the second input of the fourth up-sampling module, the output of the fourth up-sampling module is connected with the input of the fourth third residual module, and the output of the fourth second residual module is connected with the input of the output layer through the second convolution layer. The output of the output layer is output as a separable convolutional residual segmentation network based on the blend depth. In a separable convolution residual segmentation network based on mixed depth, firstly, an input image (an original CT image or a tumor region of interest) is convolved and dimensionality-increased by using 3×3, a coder adopts a first residual module to carry out convolution operation and extract feature images of different layers in cooperation with pooling operation, then a decoder uses a second residual module to carry out convolution operation and cooperate with up-sampling operation, the information of downsampling deletion is complemented by fusing corresponding coding layer features, and finally, pixel-level classification is carried out by using 1×1 convolution to segment a liver region or a tumor region.
In a separable convolutional residual segmentation network based on hybrid depth, a first residual module is used to extract features on the encoding path and obtain context information. In the present invention, the first residual module of the liver segmentation network and the first residual module of the liver tumor segmentation network are slightly different.
In the liver segmentation network, a first residual module of the mixed depth separable convolution residual segmentation network consists of 2 mixed depth separable convolution layers, 2 convolution layers and 1 feature addition layer. As shown in fig. 3 (a). The input of the first mixed depth separable convolution layer is used as the input of a first residual error module of the liver segmentation network, the output of the first mixed depth separable convolution layer is connected with the input of the second mixed depth separable convolution layer, and the output of the second mixed depth separable convolution layer is connected with the input of the first convolution layer; the input of the second convolution layer is connected with the input of the first mixed depth separable convolution layer; the outputs of the first convolution layer and the second convolution layer are simultaneously connected with the input of the feature addition layer, and the output of the feature addition layer is used as the output of a first residual error module of the liver segmentation network. In a first residual error module of the liver segmentation network, an input feature map is subjected to two-time mixed depth separable convolution, then 1×1 convolution is performed, and feature addition is performed on the feature map after the 1×1 convolution and the feature map after the 1×1 convolution which is originally input as an output of the first residual error module.
In a liver tumor segmentation network, a first residual module of the mixed depth separable convolution residual segmentation network consists of 2 mixed depth separable convolution layers, 2 convolution layers, 1 coordinate attention mechanism layer and 1 feature addition layer. As shown in fig. 3 (b). The input of the first mixed depth separable convolution layer is used as the input of a first residual error module of the liver segmentation network, the output of the first mixed depth separable convolution layer is connected with the input of the second mixed depth separable convolution layer, the output of the second mixed depth separable convolution layer is connected with the input of the coordinate attention mechanism layer, and the output of the coordinate attention mechanism layer is connected with the input of the first convolution layer; the input of the second convolution layer is connected with the input of the first mixed depth separable convolution layer; the outputs of the first convolution layer and the second convolution layer are simultaneously connected with the input of the feature addition layer, and the output of the feature addition layer is used as the output of a first residual error module of the liver segmentation network. In a first residual error module of the liver tumor segmentation network, an input feature map is subjected to two-time mixed depth separable convolution, then is subjected to a coordinate attention mechanism, is subjected to 1×1 convolution, and is subjected to feature addition with a feature map obtained by carrying out 1×1 convolution on the feature map obtained after the 1×1 convolution and the feature map which is originally input, so as to be used as the output of the first residual error module.
Because the convolution receptive field range in the residual error network is limited, the image features of the high-resolution liver edge and the tumor extracted by the network are insufficient, the method adds the mixed depth separable convolution into the first residual error module, groups the channels by the mixed depth separable convolution, and convolves by using convolution kernels with different sizes, thereby obtaining the mixed receptive field and capturing the high-resolution features and the low-resolution features. In the mixed depth separable convolution, channels of an input image are uniformly divided into 4 groups, convolution kernels of {3×3,5×5,7×7,9×9} are used for convolution, and finally, four feature maps after convolution are spliced.
In a separable convolution residual segmentation network based on mixed depth, a second residual module is used for precisely positioning a liver region and a tumor region on a decoding path. In the present invention, the second residual module of the liver segmentation network is identical to the second residual module of the liver tumor segmentation network. The second residual block consists of 3 convolutional layers and 1 feature addition layer. As shown in fig. 4. The input of the first convolution layer is used as the input of the second residual error module, and the output of the first convolution layer is connected with the input of the second convolution layer; the input of the third convolution layer is connected with the input of the first convolution layer; the outputs of the second convolution layer and the third convolution layer are simultaneously connected with the input of the characteristic addition layer; the output of the feature addition layer serves as the output of the second residual block. In the second residual module, the input feature map is subjected to 3×3 convolution operation twice, and then feature addition is performed on the feature map subjected to 1×1 convolution with the feature map which is initially input, so as to serve as the output of the second residual module.
In a separable convolutional residual segmentation network based on mixed depth, the pooling module consists of 1 max pooling layer, 1 convolutional layer and 1 splice layer. As shown in fig. 5. The input of the maximum pooling layer and the output of the convolution layer are jointly used as the input of the pooling module, the output of the maximum pooling layer and the output of the convolution layer are jointly connected with the input of the splicing layer, and the output of the splicing layer is used as the output of the pooling module. In the pooling module, the size of the feature map is reduced by adopting 3×3 convolution with the maximum pooling and the step length of 2, and the pooled feature map and the convolved feature map are spliced, so that the receptive field is enlarged.
In a separable convolution residual segmentation network based on mixed depth, an up-sampling module consists of 1 bilinear interpolation layer and 1 splicing layer. As shown in fig. 6. The input of the bilinear interpolation layer is used as the first input of the up-sampling module, the output of the bilinear interpolation layer is connected with one input of the splicing layer, the other input of the splicing layer is used as the second input of the up-sampling module, and the output of the splicing layer is used as the output of the up-sampling module. In the up-sampling module, the image size is expanded by bilinear interpolation, and the expanded feature image is spliced with the corresponding feature image in the coding path, so that a better feature reconstruction effect is achieved.
It should be noted that, although the examples described above are illustrative, this is not a limitation of the present invention, and thus the present invention is not limited to the above-described specific embodiments. Other embodiments, which are apparent to those skilled in the art from consideration of the specification and practice of the invention disclosed herein, are considered to be within the scope of the invention as claimed.

Claims (5)

1.一种基于级联网络的肝脏肿瘤分割方法,其特征是,包括步骤如下:1. A liver tumor segmentation method based on cascade network, which is characterized by including the following steps: 首先,构建基于级联网络的肝脏肿瘤分割模型;该基于级联网络的肝脏肿瘤分割模型由肝脏分割网络、肝肿瘤分割网络、以及特征相加层组成;肝脏分割网络的输入作为基于级联网络的肝脏肿瘤分割模型的输入,肝脏分割网络的输入和输出同时连接特征相加层的输入,特征相加层的输出连接肝肿瘤分割网络的输入,肝肿瘤分割网络的输出作为基于级联网络的肝脏肿瘤分割模型的输出;First, a liver tumor segmentation model based on the cascade network is constructed; the liver tumor segmentation model based on the cascade network is composed of a liver segmentation network, a liver tumor segmentation network, and a feature addition layer; the input of the liver segmentation network is as a cascade network-based liver tumor segmentation model. As the input of the liver tumor segmentation model, the input and output of the liver segmentation network are simultaneously connected to the input of the feature addition layer, the output of the feature addition layer is connected to the input of the liver tumor segmentation network, and the output of the liver tumor segmentation network is used as a cascade network-based Output of liver tumor segmentation model; 肝脏分割网络和肝肿瘤分割网络均为基于混合深度可分离卷积残差分割网络;该基于混合深度可分离卷积残差分割网络由1个输入层、2个卷积层、5个第一残差模块、4个第二残差模块、4个池化模块、4个上采样模块和1个输出层组成;The liver segmentation network and the liver tumor segmentation network are both based on a hybrid depth separable convolution residual segmentation network; the hybrid depth separable convolution residual segmentation network consists of 1 input layer, 2 convolution layers, 5 first It consists of a residual module, 4 second residual modules, 4 pooling modules, 4 upsampling modules and 1 output layer; 输入层的输入作为基于混合深度可分离卷积残差分割网络的输入;输入层的输出经由第一个卷积层连接第一个第一残差模块的输入,第一残差模块的输出连接第一个池化模块的输入和第四上采样模块的第一输入,第一个池化模块的输出连接第二个第一残差模块的输入,第二个第一残差模块的输出连接第二个池化模块的输入和第三上采样模块的第一输入,第二个池化模块的输出连接第三个第一残差模块的输入,第三个第一残差模块的输出连接第三个池化模块的输入和第二上采样模块的第一输入,第三个池化模块的输出连接第四个第一残差模块的输入,第四个第一残差模块的输出连接第四个池化模块的输入和第一上采样模块的第一输入,第四个池化模块的输出连接第五个第一残差模块的输入;第五个第一残差模块的输出连接第一个上采样模块的第二输入,第一个上采样模块的输出连接第一个第二残差模块的输入,第一个第二残差模块的输出连接第二个上采样模块的第二输入,第二个上采样模块的输出连接第二个第二残差模块的输入,第二个第二残差模块的输出连接第三个上采样模块的第二输入,第三个上采样模块的输出连接第三个第二残差模块的输入,第三个第二残差模块的输出连接第四个上采样模块的第二输入,第四个上采样模块的输出连接第四个第三残差模块的输入,第四个第二残差模块的输出经由第二个卷积层连接输出层的输入;输出层的输出作为基于混合深度可分离卷积残差分割网络的输出;The input of the input layer is used as the input of the hybrid depth-based separable convolutional residual segmentation network; the output of the input layer is connected to the input of the first first residual module through the first convolutional layer, and the output of the first residual module is connected The input of the first pooling module and the first input of the fourth upsampling module, the output of the first pooling module is connected to the input of the second first residual module, and the output of the second first residual module is connected The input of the second pooling module is connected to the first input of the third upsampling module. The output of the second pooling module is connected to the input of the third first residual module. The output of the third first residual module is connected. The input of the third pooling module is connected to the first input of the second upsampling module. The output of the third pooling module is connected to the input of the fourth first residual module. The output of the fourth first residual module is connected. The input of the fourth pooling module and the first input of the first upsampling module, the output of the fourth pooling module is connected to the input of the fifth first residual module; the output of the fifth first residual module is connected The second input of the first upsampling module, the output of the first upsampling module is connected to the input of the first second residual module, and the output of the first second residual module is connected to the second upsampling module. Two inputs, the output of the second upsampling module is connected to the input of the second second residual module, the output of the second second residual module is connected to the second input of the third upsampling module, and the third upsampling module The output of the module is connected to the input of the third second residual module, the output of the third second residual module is connected to the second input of the fourth upsampling module, and the output of the fourth upsampling module is connected to the fourth The input of the three residual modules, the output of the fourth second residual module is connected to the input of the output layer via the second convolution layer; the output of the output layer is used as the output of the hybrid depth separable convolutional residual segmentation network; 然后,利用事先已经分割出肝脏肿瘤的CT图像构建训练样本集,并利用训练样本集对所构建的基于级联网络的肝脏肿瘤分割模型进行深度学习训练,得到训练好的基于级联网络的肝脏肿瘤分割模型;Then, a training sample set is constructed using CT images that have segmented liver tumors in advance, and the training sample set is used to perform deep learning training on the constructed liver tumor segmentation model based on the cascade network, and the trained liver tumor segmentation model based on the cascade network is obtained. Tumor segmentation model; 最后,将当前需要进行肝脏肿瘤分割的CT图像送到训练好的基于级联网络的肝脏肿瘤分割模型中,得到已分割出肝脏肿瘤的CT图像。Finally, the CT images that currently need to be segmented for liver tumors are sent to the trained liver tumor segmentation model based on the cascade network to obtain CT images with segmented liver tumors. 2.根据权利要求1所述的一种基于级联网络的肝脏肿瘤分割方法,其特征是,肝脏分割网络的第一残差模块和肝肿瘤分割网络的第一残差模块不同;2. A liver tumor segmentation method based on cascade network according to claim 1, characterized in that the first residual module of the liver segmentation network and the first residual module of the liver tumor segmentation network are different; 肝脏分割网络的第一残差模块由2个混合深度可分离卷积层、2个卷积层和1个特征相加层组成;第一个混合深度可分离卷积层的输入作为肝脏分割网络的第一残差模块的输入,第一个混合深度可分离卷积层的输出连接第二个混合深度可分离卷积层的输入,第二个混合深度可分离卷积层的输出连接第一个卷积层的输入;第二个卷积层的输入连接第一个混合深度可分离卷积层的输入;第一个卷积层和第二个卷积层的输出同时连接特征相加层的输入,特征相加层的输出作为肝脏分割网络的第一残差模块的输出;The first residual module of the liver segmentation network consists of 2 hybrid depth-separable convolutional layers, 2 convolutional layers and 1 feature addition layer; the input of the first hybrid depth-separable convolutional layer is used as the liver segmentation network The input of the first residual module, the output of the first hybrid depth-separable convolution layer is connected to the input of the second hybrid depth-separable convolution layer, and the output of the second hybrid depth-separable convolution layer is connected to the first The input of the first convolutional layer; the input of the second convolutional layer is connected to the input of the first mixed depth separable convolutional layer; the output of the first convolutional layer and the second convolutional layer are simultaneously connected to the feature addition layer The input of the feature addition layer is used as the output of the first residual module of the liver segmentation network; 肝肿瘤分割网络的第一残差模块由2个混合深度可分离卷积层、2个卷积层、1个坐标注意力机制层和1个特征相加层组成;第一个混合深度可分离卷积层的输入作为肝脏分割网络的第一残差模块的输入,第一个混合深度可分离卷积层的输出连接第二个混合深度可分离卷积层的输入,第二个混合深度可分离卷积层的输出连接坐标注意力机制层的输入,坐标注意力机制层的输出连接第一个卷积层的输入;第二个卷积层的输入连接第一个混合深度可分离卷积层的输入;第一个卷积层和第二个卷积层的输出同时连接特征相加层的输入,特征相加层的输出作为肝脏分割网络的第一残差模块的输出。The first residual module of the liver tumor segmentation network consists of 2 hybrid depth-separable convolutional layers, 2 convolutional layers, 1 coordinate attention mechanism layer and 1 feature addition layer; the first hybrid depth-separable The input of the convolutional layer serves as the input of the first residual module of the liver segmentation network. The output of the first hybrid depth-separable convolutional layer is connected to the input of the second hybrid depth-separable convolutional layer. The second hybrid depth-separable convolutional layer is The output of the separation convolution layer is connected to the input of the coordinate attention mechanism layer, the output of the coordinate attention mechanism layer is connected to the input of the first convolution layer; the input of the second convolution layer is connected to the first hybrid depth separable convolution The input of the layer; the output of the first convolutional layer and the second convolutional layer are simultaneously connected to the input of the feature addition layer, and the output of the feature addition layer is used as the output of the first residual module of the liver segmentation network. 3.根据权利要求1所述的一种基于级联网络的肝脏肿瘤分割方法,其特征是,第二残差模块由3个卷积层和1个特征相加层组成;第一个卷积层的输入作为第二残差模块的输入,第一个卷积层的输出连接第二个卷积层的输入;第三个卷积层的输入连接第一个卷积层的输入;第二个卷积层和第三个卷积层的输出同时连接特征相加层的输入;特征相加层的输出作为第二残差模块的输出。3. A liver tumor segmentation method based on cascade network according to claim 1, characterized in that the second residual module is composed of 3 convolution layers and 1 feature addition layer; the first convolution The input of the layer is used as the input of the second residual module, the output of the first convolutional layer is connected to the input of the second convolutional layer; the input of the third convolutional layer is connected to the input of the first convolutional layer; the second The outputs of the first convolutional layer and the third convolutional layer are simultaneously connected to the input of the feature addition layer; the output of the feature addition layer is used as the output of the second residual module. 4.根据权利要求1所述的一种基于级联网络的肝脏肿瘤分割方法,其特征是,池化模块由1个最大池化层、1个卷积层和1个拼接层组成;最大池化层和卷积层的输入共同作为池化模块的输入,最大池化层和卷积层的输出共同连接拼接层的输入,拼接层的输出作为池化模块的输出。4. A liver tumor segmentation method based on cascade network according to claim 1, characterized in that the pooling module consists of a maximum pooling layer, a convolution layer and a splicing layer; the maximum pooling layer The inputs of the convolution layer and the convolution layer are jointly used as the input of the pooling module, the outputs of the max pooling layer and the convolution layer are jointly connected to the input of the concatenation layer, and the output of the concatenation layer is used as the output of the pooling module. 5.根据权利要求1所述的一种基于级联网络的肝脏肿瘤分割方法,其特征是,上采样模块由1个双线性插值层和1个拼接层组成;双线性插值层的输入作为上采样模块的第一输入,双线性插值层的输出连接拼接层的一个输入,拼接层的另一个输入作为上采样模块的第二输入,拼接层的输出作为上采样模块的输出。5. A liver tumor segmentation method based on cascade network according to claim 1, characterized in that the upsampling module consists of a bilinear interpolation layer and a splicing layer; the input of the bilinear interpolation layer As the first input of the upsampling module, the output of the bilinear interpolation layer is connected to one input of the splicing layer, the other input of the splicing layer serves as the second input of the upsampling module, and the output of the splicing layer serves as the output of the upsampling module.
CN202310146446.4A 2023-02-21 2023-02-21 A liver tumor segmentation method based on cascade network Active CN116206109B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202310146446.4A CN116206109B (en) 2023-02-21 2023-02-21 A liver tumor segmentation method based on cascade network

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202310146446.4A CN116206109B (en) 2023-02-21 2023-02-21 A liver tumor segmentation method based on cascade network

Publications (2)

Publication Number Publication Date
CN116206109A CN116206109A (en) 2023-06-02
CN116206109B true CN116206109B (en) 2023-11-07

Family

ID=86515543

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202310146446.4A Active CN116206109B (en) 2023-02-21 2023-02-21 A liver tumor segmentation method based on cascade network

Country Status (1)

Country Link
CN (1) CN116206109B (en)

Citations (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN111179237A (en) * 2019-12-23 2020-05-19 北京理工大学 A kind of liver and liver tumor image segmentation method and device
CN112686905A (en) * 2020-12-22 2021-04-20 天津大学 Lightweight brain tumor segmentation method based on depth separable convolution
CN112861727A (en) * 2021-02-09 2021-05-28 北京工业大学 Real-time semantic segmentation method based on mixed depth separable convolution
CN114240962A (en) * 2021-11-23 2022-03-25 湖南科技大学 CT image liver tumor region automatic segmentation method based on deep learning
CN114693933A (en) * 2022-04-07 2022-07-01 天津大学 Medical image segmentation device based on generative adversarial network and multi-scale feature fusion
CN115131369A (en) * 2022-06-15 2022-09-30 南京邮电大学 CT image liver tumor segmentation method based on 3DA-U-Nets framework
CN115661538A (en) * 2022-11-02 2023-01-31 南通大学 An image classification method based on multi-granularity Transformer

Family Cites Families (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN110084095B (en) * 2019-03-12 2022-03-25 浙江大华技术股份有限公司 Lane line detection method, lane line detection apparatus, and computer storage medium
KR102215757B1 (en) * 2019-05-14 2021-02-15 경희대학교 산학협력단 Method, apparatus and computer program for image segmentation
EP4172315A1 (en) * 2020-06-25 2023-05-03 Arrive Pte Ltd. System and methods for characterizing cells and microenvironments

Patent Citations (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN111179237A (en) * 2019-12-23 2020-05-19 北京理工大学 A kind of liver and liver tumor image segmentation method and device
CN112686905A (en) * 2020-12-22 2021-04-20 天津大学 Lightweight brain tumor segmentation method based on depth separable convolution
CN112861727A (en) * 2021-02-09 2021-05-28 北京工业大学 Real-time semantic segmentation method based on mixed depth separable convolution
CN114240962A (en) * 2021-11-23 2022-03-25 湖南科技大学 CT image liver tumor region automatic segmentation method based on deep learning
CN114693933A (en) * 2022-04-07 2022-07-01 天津大学 Medical image segmentation device based on generative adversarial network and multi-scale feature fusion
CN115131369A (en) * 2022-06-15 2022-09-30 南京邮电大学 CT image liver tumor segmentation method based on 3DA-U-Nets framework
CN115661538A (en) * 2022-11-02 2023-01-31 南通大学 An image classification method based on multi-granularity Transformer

Non-Patent Citations (4)

* Cited by examiner, † Cited by third party
Title
《MTU: A multi-tasking U-net with hybrid convolutional learning and attention modules for cancer classification and gland Segmentation in Colon Histopathological Images》;Manju Dabass等;《Computers in Biology and Medicine》;全文 *
《Multilevel depth-wise context attention network with atrous mechanism for segmentation of COVID19 affected regions》;Abdul Qayyum等;《Neural Computing and Applications》;全文 *
《基于全局深度分离卷积残差网络的高效人脸识别算法》;陈希彤 等;《武汉工程大学学报》;全文 *
《多通道深度可分离卷积模型实时识别复杂背景下甜菜与杂草》;孙俊 等;《农业工程学报》;第35卷(第12期);全文 *

Also Published As

Publication number Publication date
CN116206109A (en) 2023-06-02

Similar Documents

Publication Publication Date Title
CN109584252B (en) Lung lobe segment segmentation method and device of CT image based on deep learning
CN111369565B (en) Digital pathological image segmentation and classification method based on graph convolution network
CN111127482A (en) CT image lung trachea segmentation method and system based on deep learning
CN111275712B (en) Residual semantic network training method oriented to large-scale image data
CN114092439A (en) Multi-organ instance segmentation method and system
CN112465754B (en) 3D medical image segmentation method and device based on layered perception fusion and storage medium
CN110310289A (en) Lung tissue image segmentation method based on deep learning
CN113658201B (en) Deep learning colorectal cancer polyp segmentation device based on enhanced multi-scale features
CN108961274A (en) Automatic H/N tumors dividing method in a kind of MRI image
CN114037714A (en) 3D MR and TRUS image segmentation method for prostate system puncture
CN112991365A (en) Coronary artery segmentation method, system and storage medium
CN114399519B (en) MR image 3D semantic segmentation method and system based on multi-modal fusion
Zhou et al. H-Net: A dual-decoder enhanced FCNN for automated biomedical image diagnosis
CN115471470A (en) Esophageal cancer CT image segmentation method
CN112150449A (en) Cerebral apoplexy focus segmentation method and system
CN116206109B (en) A liver tumor segmentation method based on cascade network
CN115294023A (en) Liver tumor automatic segmentation method and device
CN116664590B (en) Automatic segmentation method and device based on dynamic contrast-enhanced magnetic resonance images
CN111667488B (en) A medical image segmentation method based on multi-angle U-Net
CN113643310A (en) Context polymerization-based MRI image hepatic vessel segmentation method
CN117314931A (en) Liver tumor image segmentation method based on improved V-Net network
CN111341438A (en) Image processing apparatus, electronic device, and medium
CN116091458A (en) A Pancreas Image Segmentation Method Based on Complementary Attention
Zhao et al. Liver segemtation in CT image with no-edge-cuting UNet
CN113205454A (en) Segmentation model establishing and segmenting method and device based on multi-scale feature extraction

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant