CN116206109B - Liver tumor segmentation method based on cascade network - Google Patents

Liver tumor segmentation method based on cascade network Download PDF

Info

Publication number
CN116206109B
CN116206109B CN202310146446.4A CN202310146446A CN116206109B CN 116206109 B CN116206109 B CN 116206109B CN 202310146446 A CN202310146446 A CN 202310146446A CN 116206109 B CN116206109 B CN 116206109B
Authority
CN
China
Prior art keywords
input
output
layer
module
network
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN202310146446.4A
Other languages
Chinese (zh)
Other versions
CN116206109A (en
Inventor
李为坤
张文辉
林镇源
蒋小莲
李佳玮
陈晓婕
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Guilin University of Electronic Technology
Original Assignee
Guilin University of Electronic Technology
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Guilin University of Electronic Technology filed Critical Guilin University of Electronic Technology
Priority to CN202310146446.4A priority Critical patent/CN116206109B/en
Publication of CN116206109A publication Critical patent/CN116206109A/en
Application granted granted Critical
Publication of CN116206109B publication Critical patent/CN116206109B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/20Image preprocessing
    • G06V10/26Segmentation of patterns in the image field; Cutting or merging of image elements to establish the pattern region, e.g. clustering-based techniques; Detection of occlusion
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/08Learning methods
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/20Image preprocessing
    • G06V10/25Determination of region of interest [ROI] or a volume of interest [VOI]
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/40Extraction of image or video features
    • G06V10/44Local feature extraction by analysis of parts of the pattern, e.g. by detecting edges, contours, loops, corners, strokes or intersections; Connectivity analysis, e.g. of connected components
    • G06V10/443Local feature extraction by analysis of parts of the pattern, e.g. by detecting edges, contours, loops, corners, strokes or intersections; Connectivity analysis, e.g. of connected components by matching or filtering
    • G06V10/449Biologically inspired filters, e.g. difference of Gaussians [DoG] or Gabor filters
    • G06V10/451Biologically inspired filters, e.g. difference of Gaussians [DoG] or Gabor filters with interaction between the filter responses, e.g. cortical complex cells
    • G06V10/454Integrating the filters into a hierarchical structure, e.g. convolutional neural networks [CNN]
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/40Extraction of image or video features
    • G06V10/52Scale-space analysis, e.g. wavelet analysis
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/70Arrangements for image or video recognition or understanding using pattern recognition or machine learning
    • G06V10/77Processing image or video features in feature spaces; using data integration or data reduction, e.g. principal component analysis [PCA] or independent component analysis [ICA] or self-organising maps [SOM]; Blind source separation
    • G06V10/774Generating sets of training patterns; Bootstrap methods, e.g. bagging or boosting
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/70Arrangements for image or video recognition or understanding using pattern recognition or machine learning
    • G06V10/82Arrangements for image or video recognition or understanding using pattern recognition or machine learning using neural networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V2201/00Indexing scheme relating to image or video recognition or understanding
    • G06V2201/03Recognition of patterns in medical or anatomical images

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Multimedia (AREA)
  • Evolutionary Computation (AREA)
  • Artificial Intelligence (AREA)
  • Health & Medical Sciences (AREA)
  • General Health & Medical Sciences (AREA)
  • Software Systems (AREA)
  • Computing Systems (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Molecular Biology (AREA)
  • Biomedical Technology (AREA)
  • Databases & Information Systems (AREA)
  • Medical Informatics (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Biodiversity & Conservation Biology (AREA)
  • Biophysics (AREA)
  • Computational Linguistics (AREA)
  • Data Mining & Analysis (AREA)
  • General Engineering & Computer Science (AREA)
  • Mathematical Physics (AREA)
  • Apparatus For Radiation Diagnosis (AREA)
  • Image Analysis (AREA)

Abstract

The invention discloses a liver tumor segmentation method based on a cascade network, which adopts a mode of cascade connection of a liver segmentation network and a liver tumor segmentation network, wherein the liver segmentation network segments liver from a CT image to extract a tumor region of interest, and then the tumor region of interest is used as input of the liver tumor segmentation network to further realize accurate segmentation of liver tumors, so that the problem of unbalanced data caused by small occupation ratio of tumors in the whole CT image is solved. The liver segmentation network and the liver tumor segmentation network both take a residual network as a framework, the residual network greatly improves the depth of the network which can be effectively trained, accelerates the convergence of the training network, and reduces the model degradation, thereby effectively avoiding the problem of gradient disappearance along with the deepening of the network layers and solving the problem of gradient dispersion caused by the over-deep network.

Description

Liver tumor segmentation method based on cascade network
Technical Field
The invention relates to the technical field of image segmentation, in particular to a liver tumor segmentation method based on a cascade network.
Background
Liver cancer is the most common and most fatal tumor in the world, and seriously threatens the life health of people. According to the national cancer center data, the incidence rate of liver cancer in China is 5 th in all malignant tumors, and the death rate is 2 nd higher. With the development of science and technology, the computer application technology and the medical informatization level are greatly developed, and the medical image facilities are more perfect. The computed tomography imaging (computed tomography, CT) has the characteristics of quick scanning time, high image resolution and the like, and is a diagnosis mode commonly adopted by liver lesions at present. Because the liver tumor in the CT image is usually characterized by low contrast, fuzzy boundary, unfixed size, shape, position and quantity, and the like, the current clinical liver tumor segmentation needs manual intervention, namely, a doctor with abundant experience marks the outline manually, which is time-consuming and labor-consuming, is difficult to effectively adapt to the complexity and diversity of the liver tumor, has poor segmentation accuracy and lower efficiency on the target, and cannot realize the automatic segmentation of the tumor region; and the tumor area of the liver CT image of the same patient can cause different results under the mark of different doctors, and the experience and skill of the doctors are seriously depended. Therefore, the research on the accurate and efficient automatic liver tumor segmentation method has important significance for clinical diagnosis and treatment of liver cancer.
In recent years, deep learning techniques have been rapidly developed and widely used in the field of medical image segmentation. Ronneeberger et al propose a U-shaped convolutional network (Unet) that introduces a jumping connection into the convolutional network for the first time, which achieves end-to-end semantic segmentation of images through encoding-decoding operations, an encoder downsamples extracted features to capture contextual information of the images, and a decoder upsamples extracted features to precisely locate segmented regions. Li et al propose a bottleneck-supervised Unet model (bottleneck supervised Unet, BS-Unet) which is a hybrid tightly-coupled structure that is partitioned by making full use of information between layers of the network. Schlemper et al incorporate the attention mechanism into the Unet network and propose an A-Unet (attention Unet) model that automatically learns regional features related to segmentation tasks and suppresses irrelevant features. Although these networks are widely used in the field of liver tumor segmentation, there are still problems of inaccurate boundary segmentation, difficult detection of small tumors, difficult tumor segmentation due to unbalanced data types, and the like.
Disclosure of Invention
The invention aims to solve the problems of difficult segmentation and inaccurate boundary segmentation in the existing liver tumor segmentation and provides a liver tumor segmentation method based on a cascade network.
In order to solve the problems, the invention is realized by the following technical scheme:
a liver tumor segmentation method based on cascade network comprises the following steps:
firstly, constructing a liver tumor segmentation model based on a cascade network; the liver tumor segmentation model based on the cascade network consists of a liver segmentation network, a liver tumor segmentation network and a characteristic addition layer; the input of the liver segmentation network is used as the input of a liver tumor segmentation model based on a cascade network, the input and the output of the liver segmentation network are simultaneously connected with the input of the feature addition layer, the output of the feature addition layer is connected with the input of the liver tumor segmentation network, and the output of the liver tumor segmentation network is used as the output of the liver tumor segmentation model based on the cascade network;
then, constructing a training sample set by utilizing CT images which have been segmented into liver tumors in advance, and performing deep learning training on the constructed liver tumor segmentation model based on the cascade network by utilizing the training sample set to obtain a trained liver tumor segmentation model based on the cascade network;
and finally, sending the CT image which is currently required to be segmented into a trained liver tumor segmentation model based on a cascade network, and obtaining the CT image of segmented liver tumor.
The liver segmentation network and the liver tumor segmentation network are both separable convolution residual segmentation networks based on the mixed depth; the separable convolution residual error segmentation network based on the mixed depth consists of 1 input layer, 2 convolution layers, 5 first residual error modules, 4 second residual error modules, 4 pooling modules, 4 up-sampling modules and 1 output layer; the input of the input layer is used as the input of the separable convolution residual segmentation network based on the mixed depth; the output of the input layer is connected with the input of a first residual module through a first convolution layer, the output of the first residual module is connected with the input of a first pooling module and the first input of a fourth upsampling module, the output of the first pooling module is connected with the input of a second first residual module, the output of the second first residual module is connected with the input of a second pooling module and the first input of a third upsampling module, the output of the second pooling module is connected with the input of a third first residual module, the output of the third first residual module is connected with the input of a third pooling module and the first input of a second upsampling module, the output of the third pooling module is connected with the input of a fourth first residual module, the output of the fourth pooling module is connected with the input of a fourth pooling module and the first input of the first upsampling module, and the output of the fourth pooling module is connected with the input of a fifth first residual module; the output of the fifth first residual module is connected with the second input of the first up-sampling module, the output of the first up-sampling module is connected with the input of the first second residual module, the output of the first second residual module is connected with the second input of the second up-sampling module, the output of the second up-sampling module is connected with the second input of the third up-sampling module, the output of the third up-sampling module is connected with the input of the third second residual module, the output of the third up-sampling module is connected with the second input of the fourth up-sampling module, the output of the fourth up-sampling module is connected with the input of the fourth third residual module, and the output of the fourth second residual module is connected with the input of the output layer through the second convolution layer; the output of the output layer is output as a separable convolutional residual segmentation network based on the blend depth.
The first residual module of the liver segmentation network is different from the first residual module of the liver tumor segmentation network. The first residual error module of the liver segmentation network consists of 2 mixed depth separable convolution layers, 2 convolution layers and 1 characteristic addition layer; the input of the first mixed depth separable convolution layer is used as the input of a first residual error module of the liver segmentation network, the output of the first mixed depth separable convolution layer is connected with the input of the second mixed depth separable convolution layer, and the output of the second mixed depth separable convolution layer is connected with the input of the first convolution layer; the input of the second convolution layer is connected with the input of the first mixed depth separable convolution layer; the outputs of the first convolution layer and the second convolution layer are simultaneously connected with the input of the feature addition layer, and the output of the feature addition layer is used as the output of a first residual error module of the liver segmentation network. The first residual error module of the liver tumor segmentation network consists of 2 mixed depth separable convolution layers, 2 convolution layers, 1 coordinate attention mechanism layer and 1 characteristic addition layer; the input of the first mixed depth separable convolution layer is used as the input of a first residual error module of the liver segmentation network, the output of the first mixed depth separable convolution layer is connected with the input of the second mixed depth separable convolution layer, the output of the second mixed depth separable convolution layer is connected with the input of the coordinate attention mechanism layer, and the output of the coordinate attention mechanism layer is connected with the input of the first convolution layer; the input of the second convolution layer is connected with the input of the first mixed depth separable convolution layer; the outputs of the first convolution layer and the second convolution layer are simultaneously connected with the input of the feature addition layer, and the output of the feature addition layer is used as the output of a first residual error module of the liver segmentation network.
The second residual error module consists of 3 convolution layers and 1 characteristic addition layer; the input of the first convolution layer is used as the input of the second residual error module, and the output of the first convolution layer is connected with the input of the second convolution layer; the input of the third convolution layer is connected with the input of the first convolution layer; the outputs of the second convolution layer and the third convolution layer are simultaneously connected with the input of the characteristic addition layer; the output of the feature addition layer serves as the output of the second residual block.
The pooling module consists of 1 maximum pooling layer, 1 convolution layer and 1 splicing layer; the input of the maximum pooling layer and the output of the convolution layer are jointly used as the input of the pooling module, the output of the maximum pooling layer and the output of the convolution layer are jointly connected with the input of the splicing layer, and the output of the splicing layer is used as the output of the pooling module.
The up-sampling module consists of 1 bilinear interpolation layer and 1 splicing layer; the input of the bilinear interpolation layer is used as the first input of the up-sampling module, the output of the bilinear interpolation layer is connected with one input of the splicing layer, the other input of the splicing layer is used as the second input of the up-sampling module, and the output of the splicing layer is used as the output of the up-sampling module.
Compared with the prior art, the invention has the following characteristics:
1. the liver segmentation network and the liver tumor segmentation network are adopted to carry out cascading, the liver segmentation network segments the liver from the CT image firstly, a tumor region of interest is extracted, then the tumor region of interest is used as the input of the liver tumor segmentation network, the accurate segmentation of the liver tumor is further realized, and the problem of unbalanced data caused by small proportion of the tumor in the whole CT image is solved;
2. the liver segmentation network and the liver tumor segmentation network both take a residual network as a framework, the residual network greatly improves the depth of the network which can be effectively trained, accelerates the convergence of the training network, and reduces the model degradation, thereby effectively avoiding the problem of gradient disappearance along with the deepening of the network layers and solving the problem of gradient dispersion caused by the over-deep network;
3. the mixed depth separable convolution is operated on different channels by using convolution kernels with different sizes, and the multi-scale convolution kernels are fused into a single convolution operation, so that characteristic modes with different resolutions are captured, and edge details and deeper small target characteristics are extracted; through strengthening the receptive field of the segmented network feature map and fully utilizing the channel and space structure information, pixel level detail and space information can be captured better, so that the segmentation performance of the network on medical images is improved.
4. The coordinate attention mechanism can capture cross-channel information so that the model can more accurately locate and identify the lesion area.
Drawings
Fig. 1 is a schematic diagram of a liver tumor segmentation model based on a cascade network.
Fig. 2 is a schematic diagram of a separable convolutional residual partitioning network based on hybrid depth (CMDCRA-UNet).
Fig. 3 is a schematic diagram of a first Residual block (a) of a liver segmentation network and (b) of a liver tumor segmentation network.
Fig. 4 is a schematic diagram of a second Residual block 2.
Fig. 5 is a schematic diagram of a pooling module (Pool).
Fig. 6 is a schematic diagram of an Up sampling module (Up Sample).
Detailed Description
The present invention will be further described in detail with reference to specific examples in order to make the objects, technical solutions and advantages of the present invention more apparent.
A liver tumor segmentation method based on cascade network, firstly constructing a liver tumor segmentation model based on cascade network; then constructing a training sample set by utilizing CT images which have been segmented into liver tumors in advance, and performing deep learning training on the constructed liver tumor segmentation model based on the cascade network by utilizing the training sample set to obtain a trained liver tumor segmentation model based on the cascade network; and finally, sending the CT image which is currently required to be segmented into a trained liver tumor segmentation model based on a cascade network, and obtaining the CT image of segmented liver tumor.
Although the liver tumor density is different from the normal liver tissue density, the liver tumor density is similar to the tissue density of other organs in the abdomen, so that the ideal effect is difficult to obtain by directly dividing the tumor by using a network, and the liver interested region can ensure that the original CT image only keeps the liver region, thereby effectively avoiding the interference of other organs in the abdomen on the division of the liver tumor. Therefore, the liver segmentation network is firstly utilized to extract the liver region in the CT image, and then the liver tumor segmentation network is utilized to extract the tumor region in the liver region. That is, the liver tumor segmentation model based on the cascade network constructed by the invention consists of a liver segmentation network, a liver tumor segmentation network and a characteristic addition layer. As shown in fig. 1. The input of the liver segmentation network is used as the input of the liver tumor segmentation model based on the cascade network, the input and the output of the liver segmentation network are simultaneously connected with the input of the feature addition layer, the output of the feature addition layer is connected with the input of the liver tumor segmentation network, and the output of the liver tumor segmentation network is used as the output of the liver tumor segmentation model based on the cascade network. The liver segmentation network segments the liver from the original CT image, extracts a tumor region of interest, and inputs the tumor region of interest and the tumor region of interest as the liver tumor segmentation network, so that the liver tumor segmentation network further realizes accurate segmentation of liver tumors.
The liver segmentation network and the tumor segmentation network are both separable convolution residual segmentation networks based on the mixed depth. Based on the mixed depth separable convolution residual error segmentation network, the whole adopts an encoding-decoding architecture, and the residual error network is used as a framework. The separable convolution residual segmentation network based on the mixed depth consists of 1 input layer, 2 convolution layers, 5 first residual modules, 4 second residual modules, 4 pooling modules, 4 up-sampling modules and 1 output layer. As shown in fig. 2. The input of the input layer serves as the input to the separable convolutional residual segmentation network based on the blend depth. The output of the input layer is connected with the input of a first residual module through a first convolution layer, the output of the first residual module is connected with the input of a first pooling module and the first input of a fourth upsampling module, the output of the first pooling module is connected with the input of a second first residual module, the output of the second first residual module is connected with the input of a second pooling module and the first input of a third upsampling module, the output of the second pooling module is connected with the input of a third first residual module, the output of the third first residual module is connected with the input of a third pooling module and the first input of a second upsampling module, the output of the third pooling module is connected with the input of a fourth first residual module, the output of the fourth pooling module is connected with the input of a fourth pooling module and the first input of the first upsampling module, and the output of the fourth pooling module is connected with the input of a fifth first residual module. The output of the fifth first residual module is connected with the second input of the first up-sampling module, the output of the first up-sampling module is connected with the input of the first second residual module, the output of the first second residual module is connected with the second input of the second up-sampling module, the output of the second up-sampling module is connected with the second input of the third up-sampling module, the output of the third up-sampling module is connected with the input of the third second residual module, the output of the third second residual module is connected with the second input of the fourth up-sampling module, the output of the fourth up-sampling module is connected with the input of the fourth third residual module, and the output of the fourth second residual module is connected with the input of the output layer through the second convolution layer. The output of the output layer is output as a separable convolutional residual segmentation network based on the blend depth. In a separable convolution residual segmentation network based on mixed depth, firstly, an input image (an original CT image or a tumor region of interest) is convolved and dimensionality-increased by using 3×3, a coder adopts a first residual module to carry out convolution operation and extract feature images of different layers in cooperation with pooling operation, then a decoder uses a second residual module to carry out convolution operation and cooperate with up-sampling operation, the information of downsampling deletion is complemented by fusing corresponding coding layer features, and finally, pixel-level classification is carried out by using 1×1 convolution to segment a liver region or a tumor region.
In a separable convolutional residual segmentation network based on hybrid depth, a first residual module is used to extract features on the encoding path and obtain context information. In the present invention, the first residual module of the liver segmentation network and the first residual module of the liver tumor segmentation network are slightly different.
In the liver segmentation network, a first residual module of the mixed depth separable convolution residual segmentation network consists of 2 mixed depth separable convolution layers, 2 convolution layers and 1 feature addition layer. As shown in fig. 3 (a). The input of the first mixed depth separable convolution layer is used as the input of a first residual error module of the liver segmentation network, the output of the first mixed depth separable convolution layer is connected with the input of the second mixed depth separable convolution layer, and the output of the second mixed depth separable convolution layer is connected with the input of the first convolution layer; the input of the second convolution layer is connected with the input of the first mixed depth separable convolution layer; the outputs of the first convolution layer and the second convolution layer are simultaneously connected with the input of the feature addition layer, and the output of the feature addition layer is used as the output of a first residual error module of the liver segmentation network. In a first residual error module of the liver segmentation network, an input feature map is subjected to two-time mixed depth separable convolution, then 1×1 convolution is performed, and feature addition is performed on the feature map after the 1×1 convolution and the feature map after the 1×1 convolution which is originally input as an output of the first residual error module.
In a liver tumor segmentation network, a first residual module of the mixed depth separable convolution residual segmentation network consists of 2 mixed depth separable convolution layers, 2 convolution layers, 1 coordinate attention mechanism layer and 1 feature addition layer. As shown in fig. 3 (b). The input of the first mixed depth separable convolution layer is used as the input of a first residual error module of the liver segmentation network, the output of the first mixed depth separable convolution layer is connected with the input of the second mixed depth separable convolution layer, the output of the second mixed depth separable convolution layer is connected with the input of the coordinate attention mechanism layer, and the output of the coordinate attention mechanism layer is connected with the input of the first convolution layer; the input of the second convolution layer is connected with the input of the first mixed depth separable convolution layer; the outputs of the first convolution layer and the second convolution layer are simultaneously connected with the input of the feature addition layer, and the output of the feature addition layer is used as the output of a first residual error module of the liver segmentation network. In a first residual error module of the liver tumor segmentation network, an input feature map is subjected to two-time mixed depth separable convolution, then is subjected to a coordinate attention mechanism, is subjected to 1×1 convolution, and is subjected to feature addition with a feature map obtained by carrying out 1×1 convolution on the feature map obtained after the 1×1 convolution and the feature map which is originally input, so as to be used as the output of the first residual error module.
Because the convolution receptive field range in the residual error network is limited, the image features of the high-resolution liver edge and the tumor extracted by the network are insufficient, the method adds the mixed depth separable convolution into the first residual error module, groups the channels by the mixed depth separable convolution, and convolves by using convolution kernels with different sizes, thereby obtaining the mixed receptive field and capturing the high-resolution features and the low-resolution features. In the mixed depth separable convolution, channels of an input image are uniformly divided into 4 groups, convolution kernels of {3×3,5×5,7×7,9×9} are used for convolution, and finally, four feature maps after convolution are spliced.
In a separable convolution residual segmentation network based on mixed depth, a second residual module is used for precisely positioning a liver region and a tumor region on a decoding path. In the present invention, the second residual module of the liver segmentation network is identical to the second residual module of the liver tumor segmentation network. The second residual block consists of 3 convolutional layers and 1 feature addition layer. As shown in fig. 4. The input of the first convolution layer is used as the input of the second residual error module, and the output of the first convolution layer is connected with the input of the second convolution layer; the input of the third convolution layer is connected with the input of the first convolution layer; the outputs of the second convolution layer and the third convolution layer are simultaneously connected with the input of the characteristic addition layer; the output of the feature addition layer serves as the output of the second residual block. In the second residual module, the input feature map is subjected to 3×3 convolution operation twice, and then feature addition is performed on the feature map subjected to 1×1 convolution with the feature map which is initially input, so as to serve as the output of the second residual module.
In a separable convolutional residual segmentation network based on mixed depth, the pooling module consists of 1 max pooling layer, 1 convolutional layer and 1 splice layer. As shown in fig. 5. The input of the maximum pooling layer and the output of the convolution layer are jointly used as the input of the pooling module, the output of the maximum pooling layer and the output of the convolution layer are jointly connected with the input of the splicing layer, and the output of the splicing layer is used as the output of the pooling module. In the pooling module, the size of the feature map is reduced by adopting 3×3 convolution with the maximum pooling and the step length of 2, and the pooled feature map and the convolved feature map are spliced, so that the receptive field is enlarged.
In a separable convolution residual segmentation network based on mixed depth, an up-sampling module consists of 1 bilinear interpolation layer and 1 splicing layer. As shown in fig. 6. The input of the bilinear interpolation layer is used as the first input of the up-sampling module, the output of the bilinear interpolation layer is connected with one input of the splicing layer, the other input of the splicing layer is used as the second input of the up-sampling module, and the output of the splicing layer is used as the output of the up-sampling module. In the up-sampling module, the image size is expanded by bilinear interpolation, and the expanded feature image is spliced with the corresponding feature image in the coding path, so that a better feature reconstruction effect is achieved.
It should be noted that, although the examples described above are illustrative, this is not a limitation of the present invention, and thus the present invention is not limited to the above-described specific embodiments. Other embodiments, which are apparent to those skilled in the art from consideration of the specification and practice of the invention disclosed herein, are considered to be within the scope of the invention as claimed.

Claims (5)

1. The liver tumor segmentation method based on the cascade network is characterized by comprising the following steps of:
firstly, constructing a liver tumor segmentation model based on a cascade network; the liver tumor segmentation model based on the cascade network consists of a liver segmentation network, a liver tumor segmentation network and a characteristic addition layer; the input of the liver segmentation network is used as the input of a liver tumor segmentation model based on a cascade network, the input and the output of the liver segmentation network are simultaneously connected with the input of the feature addition layer, the output of the feature addition layer is connected with the input of the liver tumor segmentation network, and the output of the liver tumor segmentation network is used as the output of the liver tumor segmentation model based on the cascade network;
the liver segmentation network and the liver tumor segmentation network are both separable convolution residual segmentation networks based on the mixed depth; the separable convolution residual error segmentation network based on the mixed depth consists of 1 input layer, 2 convolution layers, 5 first residual error modules, 4 second residual error modules, 4 pooling modules, 4 up-sampling modules and 1 output layer;
the input of the input layer is used as the input of the separable convolution residual segmentation network based on the mixed depth; the output of the input layer is connected with the input of a first residual module through a first convolution layer, the output of the first residual module is connected with the input of a first pooling module and the first input of a fourth upsampling module, the output of the first pooling module is connected with the input of a second first residual module, the output of the second first residual module is connected with the input of a second pooling module and the first input of a third upsampling module, the output of the second pooling module is connected with the input of a third first residual module, the output of the third first residual module is connected with the input of a third pooling module and the first input of a second upsampling module, the output of the third pooling module is connected with the input of a fourth first residual module, the output of the fourth pooling module is connected with the input of a fourth pooling module and the first input of the first upsampling module, and the output of the fourth pooling module is connected with the input of a fifth first residual module; the output of the fifth first residual module is connected with the second input of the first up-sampling module, the output of the first up-sampling module is connected with the input of the first second residual module, the output of the first second residual module is connected with the second input of the second up-sampling module, the output of the second up-sampling module is connected with the second input of the third up-sampling module, the output of the third up-sampling module is connected with the input of the third second residual module, the output of the third up-sampling module is connected with the second input of the fourth up-sampling module, the output of the fourth up-sampling module is connected with the input of the fourth third residual module, and the output of the fourth second residual module is connected with the input of the output layer through the second convolution layer; the output of the output layer is used as the output of a separable convolution residual segmentation network based on the mixed depth;
then, constructing a training sample set by utilizing CT images which have been segmented into liver tumors in advance, and performing deep learning training on the constructed liver tumor segmentation model based on the cascade network by utilizing the training sample set to obtain a trained liver tumor segmentation model based on the cascade network;
and finally, sending the CT image which is currently required to be segmented into a trained liver tumor segmentation model based on a cascade network, and obtaining the CT image of segmented liver tumor.
2. The cascade network-based liver tumor segmentation method according to claim 1, wherein a first residual module of the liver segmentation network is different from a first residual module of the liver tumor segmentation network;
the first residual error module of the liver segmentation network consists of 2 mixed depth separable convolution layers, 2 convolution layers and 1 characteristic addition layer; the input of the first mixed depth separable convolution layer is used as the input of a first residual error module of the liver segmentation network, the output of the first mixed depth separable convolution layer is connected with the input of the second mixed depth separable convolution layer, and the output of the second mixed depth separable convolution layer is connected with the input of the first convolution layer; the input of the second convolution layer is connected with the input of the first mixed depth separable convolution layer; the output of the first convolution layer and the output of the second convolution layer are simultaneously connected with the input of the characteristic addition layer, and the output of the characteristic addition layer is used as the output of a first residual error module of the liver segmentation network;
the first residual error module of the liver tumor segmentation network consists of 2 mixed depth separable convolution layers, 2 convolution layers, 1 coordinate attention mechanism layer and 1 characteristic addition layer; the input of the first mixed depth separable convolution layer is used as the input of a first residual error module of the liver segmentation network, the output of the first mixed depth separable convolution layer is connected with the input of the second mixed depth separable convolution layer, the output of the second mixed depth separable convolution layer is connected with the input of the coordinate attention mechanism layer, and the output of the coordinate attention mechanism layer is connected with the input of the first convolution layer; the input of the second convolution layer is connected with the input of the first mixed depth separable convolution layer; the outputs of the first convolution layer and the second convolution layer are simultaneously connected with the input of the feature addition layer, and the output of the feature addition layer is used as the output of a first residual error module of the liver segmentation network.
3. The cascade network-based liver tumor segmentation method of claim 1, wherein the second residual module consists of 3 convolution layers and 1 feature addition layer; the input of the first convolution layer is used as the input of the second residual error module, and the output of the first convolution layer is connected with the input of the second convolution layer; the input of the third convolution layer is connected with the input of the first convolution layer; the outputs of the second convolution layer and the third convolution layer are simultaneously connected with the input of the characteristic addition layer; the output of the feature addition layer serves as the output of the second residual block.
4. The liver tumor segmentation method based on cascade network as claimed in claim 1, wherein the pooling module is composed of 1 maximum pooling layer, 1 convolution layer and 1 splicing layer; the input of the maximum pooling layer and the output of the convolution layer are jointly used as the input of the pooling module, the output of the maximum pooling layer and the output of the convolution layer are jointly connected with the input of the splicing layer, and the output of the splicing layer is used as the output of the pooling module.
5. The liver tumor segmentation method based on cascade network as claimed in claim 1, wherein the up-sampling module is composed of 1 bilinear interpolation layer and 1 splicing layer; the input of the bilinear interpolation layer is used as the first input of the up-sampling module, the output of the bilinear interpolation layer is connected with one input of the splicing layer, the other input of the splicing layer is used as the second input of the up-sampling module, and the output of the splicing layer is used as the output of the up-sampling module.
CN202310146446.4A 2023-02-21 2023-02-21 Liver tumor segmentation method based on cascade network Active CN116206109B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202310146446.4A CN116206109B (en) 2023-02-21 2023-02-21 Liver tumor segmentation method based on cascade network

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202310146446.4A CN116206109B (en) 2023-02-21 2023-02-21 Liver tumor segmentation method based on cascade network

Publications (2)

Publication Number Publication Date
CN116206109A CN116206109A (en) 2023-06-02
CN116206109B true CN116206109B (en) 2023-11-07

Family

ID=86515543

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202310146446.4A Active CN116206109B (en) 2023-02-21 2023-02-21 Liver tumor segmentation method based on cascade network

Country Status (1)

Country Link
CN (1) CN116206109B (en)

Citations (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN111179237A (en) * 2019-12-23 2020-05-19 北京理工大学 Image segmentation method and device for liver and liver tumor
CN112686905A (en) * 2020-12-22 2021-04-20 天津大学 Lightweight brain tumor segmentation method based on depth separable convolution
CN112861727A (en) * 2021-02-09 2021-05-28 北京工业大学 Real-time semantic segmentation method based on mixed depth separable convolution
CN114240962A (en) * 2021-11-23 2022-03-25 湖南科技大学 CT image liver tumor region automatic segmentation method based on deep learning
CN114693933A (en) * 2022-04-07 2022-07-01 天津大学 Medical image segmentation device based on generation of confrontation network and multi-scale feature fusion
CN115131369A (en) * 2022-06-15 2022-09-30 南京邮电大学 CT image liver tumor segmentation method based on 3DA-U-Nets framework
CN115661538A (en) * 2022-11-02 2023-01-31 南通大学 Image classification method based on multi-granularity Transformer

Family Cites Families (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN110084095B (en) * 2019-03-12 2022-03-25 浙江大华技术股份有限公司 Lane line detection method, lane line detection apparatus, and computer storage medium
KR102215757B1 (en) * 2019-05-14 2021-02-15 경희대학교 산학협력단 Method, apparatus and computer program for image segmentation
WO2021263038A1 (en) * 2020-06-25 2021-12-30 Arrive PTE Ltd. System and methods for characterizing cells and microenvironments

Patent Citations (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN111179237A (en) * 2019-12-23 2020-05-19 北京理工大学 Image segmentation method and device for liver and liver tumor
CN112686905A (en) * 2020-12-22 2021-04-20 天津大学 Lightweight brain tumor segmentation method based on depth separable convolution
CN112861727A (en) * 2021-02-09 2021-05-28 北京工业大学 Real-time semantic segmentation method based on mixed depth separable convolution
CN114240962A (en) * 2021-11-23 2022-03-25 湖南科技大学 CT image liver tumor region automatic segmentation method based on deep learning
CN114693933A (en) * 2022-04-07 2022-07-01 天津大学 Medical image segmentation device based on generation of confrontation network and multi-scale feature fusion
CN115131369A (en) * 2022-06-15 2022-09-30 南京邮电大学 CT image liver tumor segmentation method based on 3DA-U-Nets framework
CN115661538A (en) * 2022-11-02 2023-01-31 南通大学 Image classification method based on multi-granularity Transformer

Non-Patent Citations (4)

* Cited by examiner, † Cited by third party
Title
《MTU: A multi-tasking U-net with hybrid convolutional learning and attention modules for cancer classification and gland Segmentation in Colon Histopathological Images》;Manju Dabass等;《Computers in Biology and Medicine》;全文 *
《Multilevel depth-wise context attention network with atrous mechanism for segmentation of COVID19 affected regions》;Abdul Qayyum等;《Neural Computing and Applications》;全文 *
《基于全局深度分离卷积残差网络的高效人脸识别算法》;陈希彤 等;《武汉工程大学学报》;全文 *
《多通道深度可分离卷积模型实时识别复杂背景下甜菜与杂草》;孙俊 等;《农业工程学报》;第35卷(第12期);全文 *

Also Published As

Publication number Publication date
CN116206109A (en) 2023-06-02

Similar Documents

Publication Publication Date Title
CN113012172B (en) AS-UNet-based medical image segmentation method and system
CN109615636B (en) Blood vessel tree construction method and device in lung lobe segment segmentation of CT (computed tomography) image
CN113674253B (en) Automatic segmentation method for rectal cancer CT image based on U-transducer
CN111369565B (en) Digital pathological image segmentation and classification method based on graph convolution network
WO2023071531A1 (en) Liver ct automatic segmentation method based on deep shape learning
CN111127482A (en) CT image lung trachea segmentation method and system based on deep learning
CN114092439A (en) Multi-organ instance segmentation method and system
CN111047605B (en) Construction method and segmentation method of vertebra CT segmentation network model
CN112446892A (en) Cell nucleus segmentation method based on attention learning
CN111275712B (en) Residual semantic network training method oriented to large-scale image data
CN114119977A (en) Graph convolution-based Transformer gastric cancer canceration region image segmentation method
CN112465754B (en) 3D medical image segmentation method and device based on layered perception fusion and storage medium
CN111429473A (en) Chest film lung field segmentation model establishment and segmentation method based on multi-scale feature fusion
CN108305253A (en) A kind of pathology full slice diagnostic method based on more multiplying power deep learnings
CN114037714A (en) 3D MR and TRUS image segmentation method for prostate system puncture
CN113888466A (en) Pulmonary nodule image detection method and system based on CT image
CN112734755A (en) Lung lobe segmentation method based on 3D full convolution neural network and multitask learning
CN113436173A (en) Abdomen multi-organ segmentation modeling and segmentation method and system based on edge perception
CN112150449A (en) Cerebral apoplexy focus segmentation method and system
CN111091575A (en) Medical image segmentation method based on reinforcement learning method
CN116664590B (en) Automatic segmentation method and device based on dynamic contrast enhancement magnetic resonance image
CN113643310A (en) Context polymerization-based MRI image hepatic vessel segmentation method
CN116206109B (en) Liver tumor segmentation method based on cascade network
CN117523204A (en) Liver tumor image segmentation method and device oriented to medical scene and readable storage medium
CN115731387A (en) Multi-type pulmonary nodule accurate segmentation method and model based on mixed Transformer

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant