CN117078941B - Cardiac MRI segmentation method based on context cascade attention - Google Patents
Cardiac MRI segmentation method based on context cascade attention Download PDFInfo
- Publication number
- CN117078941B CN117078941B CN202311231297.8A CN202311231297A CN117078941B CN 117078941 B CN117078941 B CN 117078941B CN 202311231297 A CN202311231297 A CN 202311231297A CN 117078941 B CN117078941 B CN 117078941B
- Authority
- CN
- China
- Prior art keywords
- feature map
- layer
- convolution layer
- convolution
- branch
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Active
Links
- 230000011218 segmentation Effects 0.000 title claims abstract description 44
- 238000000034 method Methods 0.000 title claims abstract description 23
- 238000013184 cardiac magnetic resonance imaging Methods 0.000 title claims description 33
- 238000010586 diagram Methods 0.000 claims description 91
- 230000006870 function Effects 0.000 claims description 84
- 238000011176 pooling Methods 0.000 claims description 61
- 230000004913 activation Effects 0.000 claims description 36
- 238000012549 training Methods 0.000 claims description 14
- 238000013507 mapping Methods 0.000 claims description 10
- 238000002595 magnetic resonance imaging Methods 0.000 claims description 9
- 238000012952 Resampling Methods 0.000 claims description 8
- 238000012856 packing Methods 0.000 claims description 7
- 238000012360 testing method Methods 0.000 claims description 6
- 238000012795 verification Methods 0.000 claims description 6
- 238000012545 processing Methods 0.000 claims description 4
- ORILYTVJVMAKLC-UHFFFAOYSA-N Adamantane Natural products C1C(C2)CC3CC1CC2C3 ORILYTVJVMAKLC-UHFFFAOYSA-N 0.000 claims description 3
- FGUUSXIOTUKUDN-IBGZPJMESA-N C1(=CC=CC=C1)N1C2=C(NC([C@H](C1)NC=1OC(=NN=1)C1=CC=CC=C1)=O)C=CC=C2 Chemical compound C1(=CC=CC=C1)N1C2=C(NC([C@H](C1)NC=1OC(=NN=1)C1=CC=CC=C1)=O)C=CC=C2 FGUUSXIOTUKUDN-IBGZPJMESA-N 0.000 claims description 3
- 238000010606 normalization Methods 0.000 claims description 3
- 238000007781 pre-processing Methods 0.000 claims description 3
- 230000000295 complement effect Effects 0.000 abstract description 2
- 238000003709 image segmentation Methods 0.000 abstract description 2
- 230000000747 cardiac effect Effects 0.000 description 5
- 238000013135 deep learning Methods 0.000 description 3
- GNFTZDOKVXKIBK-UHFFFAOYSA-N 3-(2-methoxyethoxy)benzohydrazide Chemical compound COCCOC1=CC=CC(C(=O)NN)=C1 GNFTZDOKVXKIBK-UHFFFAOYSA-N 0.000 description 2
- 238000012986 modification Methods 0.000 description 2
- 230000004048 modification Effects 0.000 description 2
- 238000011160 research Methods 0.000 description 2
- 230000009286 beneficial effect Effects 0.000 description 1
- 238000013527 convolutional neural network Methods 0.000 description 1
- 230000007812 deficiency Effects 0.000 description 1
- 230000000694 effects Effects 0.000 description 1
- 210000002837 heart atrium Anatomy 0.000 description 1
- 208000019622 heart disease Diseases 0.000 description 1
- 230000006872 improvement Effects 0.000 description 1
- 210000004165 myocardium Anatomy 0.000 description 1
- 230000008569 process Effects 0.000 description 1
- 238000000638 solvent extraction Methods 0.000 description 1
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V10/00—Arrangements for image or video recognition or understanding
- G06V10/20—Image preprocessing
- G06V10/26—Segmentation of patterns in the image field; Cutting or merging of image elements to establish the pattern region, e.g. clustering-based techniques; Detection of occlusion
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/04—Architecture, e.g. interconnection topology
- G06N3/045—Combinations of networks
- G06N3/0455—Auto-encoder networks; Encoder-decoder networks
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/04—Architecture, e.g. interconnection topology
- G06N3/0464—Convolutional networks [CNN, ConvNet]
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V10/00—Arrangements for image or video recognition or understanding
- G06V10/40—Extraction of image or video features
- G06V10/44—Local feature extraction by analysis of parts of the pattern, e.g. by detecting edges, contours, loops, corners, strokes or intersections; Connectivity analysis, e.g. of connected components
- G06V10/443—Local feature extraction by analysis of parts of the pattern, e.g. by detecting edges, contours, loops, corners, strokes or intersections; Connectivity analysis, e.g. of connected components by matching or filtering
- G06V10/449—Biologically inspired filters, e.g. difference of Gaussians [DoG] or Gabor filters
- G06V10/451—Biologically inspired filters, e.g. difference of Gaussians [DoG] or Gabor filters with interaction between the filter responses, e.g. cortical complex cells
- G06V10/454—Integrating the filters into a hierarchical structure, e.g. convolutional neural networks [CNN]
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V10/00—Arrangements for image or video recognition or understanding
- G06V10/70—Arrangements for image or video recognition or understanding using pattern recognition or machine learning
- G06V10/768—Arrangements for image or video recognition or understanding using pattern recognition or machine learning using context analysis, e.g. recognition aided by known co-occurring patterns
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V10/00—Arrangements for image or video recognition or understanding
- G06V10/70—Arrangements for image or video recognition or understanding using pattern recognition or machine learning
- G06V10/77—Processing image or video features in feature spaces; using data integration or data reduction, e.g. principal component analysis [PCA] or independent component analysis [ICA] or self-organising maps [SOM]; Blind source separation
- G06V10/80—Fusion, i.e. combining data from various sources at the sensor level, preprocessing level, feature extraction level or classification level
- G06V10/806—Fusion, i.e. combining data from various sources at the sensor level, preprocessing level, feature extraction level or classification level of extracted features
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V10/00—Arrangements for image or video recognition or understanding
- G06V10/70—Arrangements for image or video recognition or understanding using pattern recognition or machine learning
- G06V10/82—Arrangements for image or video recognition or understanding using pattern recognition or machine learning using neural networks
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V2201/00—Indexing scheme relating to image or video recognition or understanding
- G06V2201/03—Recognition of patterns in medical or anatomical images
-
- Y—GENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
- Y02—TECHNOLOGIES OR APPLICATIONS FOR MITIGATION OR ADAPTATION AGAINST CLIMATE CHANGE
- Y02T—CLIMATE CHANGE MITIGATION TECHNOLOGIES RELATED TO TRANSPORTATION
- Y02T10/00—Road transport of goods or passengers
- Y02T10/10—Internal combustion engine [ICE] based vehicles
- Y02T10/40—Engine management systems
Landscapes
- Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- Physics & Mathematics (AREA)
- General Physics & Mathematics (AREA)
- Evolutionary Computation (AREA)
- Artificial Intelligence (AREA)
- Health & Medical Sciences (AREA)
- General Health & Medical Sciences (AREA)
- Multimedia (AREA)
- Software Systems (AREA)
- Computing Systems (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Molecular Biology (AREA)
- Biomedical Technology (AREA)
- Databases & Information Systems (AREA)
- Medical Informatics (AREA)
- Life Sciences & Earth Sciences (AREA)
- Biophysics (AREA)
- Computational Linguistics (AREA)
- Data Mining & Analysis (AREA)
- General Engineering & Computer Science (AREA)
- Mathematical Physics (AREA)
- Biodiversity & Conservation Biology (AREA)
- Image Processing (AREA)
Abstract
A heart MRI segmentation method based on context cascade attention relates to the technical field of medical image segmentation, and a residual error initial module is introduced into an encoder to better learn effective characteristic representation, so that the performance and generalization capability of a model are improved. The features of each layer are refined by concatenation operations with the complementary information of the different layers in the decoder, while the context information of each layer is explored by preserving local information and compressing global information using a context attention module, and salient features are highlighted by suppressing unimportant information areas, thereby improving the accuracy of segmentation.
Description
Technical Field
The invention relates to the technical field of medical image segmentation, in particular to a cardiac MRI segmentation method based on context cascade attention.
Background
The accurate segmentation of cardiac MRI is of great importance in the field of medical image processing, aimed at accurately extracting cardiac structures, such as ventricles, atria, cardiac muscles, etc., from cardiac MR images, to assist doctors in accurately diagnosing and treating cardiac diseases. At present, research methods for segmenting cardiac MR images are mainly divided into a traditional method and a deep learning method, wherein the traditional method is difficult to process complex cardiac structures in cardiac MRI segmentation, so that the segmentation effect is poor, and the deep learning method can better capture complex features in cardiac MR images. U-Net is used as a representative deep learning system architecture, an encoder-decoder structure is adopted, jump connection is introduced, the problem of information loss of a traditional convolutional neural network in segmentation tasks is solved, and important breakthroughs are brought for research and application of cardiac MRI segmentation. However, feature abstraction capability is insufficient and there is a lack of problem of learning context between pixels, which results in complexity of segmentation.
Disclosure of Invention
In order to overcome the deficiencies of the above techniques, the present invention provides a method of context-cascade attention-based cardiac MRI segmentation that better captures and represents complex data features.
The technical scheme adopted for overcoming the technical problems is as follows:
a method of cardiac MRI segmentation based on contextual cascade attention, comprising the steps of:
a) Collecting cardiac MRI data of n subjects to obtain MRI data sets s, s= { s 1 ,s 2 ,...,s i ,...,s n },s i Cardiac MRI data for the i-th subject, i e {1,2,., n };
b) Preprocessing the MRI data set s to obtain a preprocessed data set F, F= { F 1 ,F 2 ,...,F i ,...,F n },F i The i-th preprocessed two-dimensional image data;
c) Dividing the preprocessed data set F into a training set, a testing set and a verification set;
d) Establishing a segmentation network model formed by an encoder and a decoder;
e) The preprocessed two-dimensional image data F i Input into encoder of dividing network model, and output to obtain characteristic diagramFeature map->Feature map->Feature map->
f) Map the characteristic mapFeature map->Feature map->Feature map->Input into decoder of segmentation network model, output to obtain predictive segmentation image>
g) Training the segmentation network model to obtain an optimized segmentation network model.
Further, in step a), cardiac MRI data of n subjects containing LV, RV, MYO structures are collected from ACDC2017 public data, resulting in MRI dataset s.
Preferably, n=100 in step a).
Further, step b) comprises the steps of:
b-1) cardiac MRI data s for the ith subject i Resampling is carried out on the z-axis one by one, wherein the resampling is that the pixel pitch in the x-axis direction is 1.5, and the pixel pitch in the y-axis direction is 1.5;
b-2) resampling cardiac MRI data s i Performing 2D center clipping operation with clipping size 224×224 to obtain clipped data F i ' cutting the cut data F i ' normalization processing is carried out to obtain preprocessed two-dimensional image data F i 。
Preferably, in step c), the preprocessed data set F is divided into a training set, a testing set and a verification set according to a ratio of 7:2:1.
Further, step e) comprises the steps of:
e-1) an encoder for dividing a network model is composed of a first residual initial module, a first maximum pooling layer, a second residual initial module, a second maximum pooling layer, a third residual initial module, a third maximum pooling layer and a fourth residual initial module;
e-2) the first residual error initial module is composed of a first branch, a second branch, a third branch, a fourth branch and a fifth branch, wherein the first branch is composed of a convolution layer, the second branch is sequentially composed of an average pooling layer and a convolution layer, the third branch is sequentially composed of a convolution layer and a BN layer, the fourth branch is sequentially composed of the first convolution layer, the second convolution layer and the BN layer, the fifth branch is sequentially composed of the first convolution layer, the second convolution layer, the first BN layer, the third convolution layer and the second BN layer, and the preprocessed two-dimensional image data F is obtained by i Input into the first branch, output to obtain a feature mapThe preprocessed two-dimensional image data F i Input into the second branch, output the obtained feature map +.>Will be pretreatedTwo-dimensional image data F i Input into the third branch, output the obtained feature map +.>The preprocessed two-dimensional image data F i Input into the fourth branch, output the obtained feature map +.>The preprocessed two-dimensional image data F i Input into the fifth branch, output the obtained feature map +.>Feature map +.>Feature mapFeature map->Feature map->Performing splicing operation to obtain characteristic diagram->Feature map +.>And feature mapThe addition operation results in a feature map->e-3) mapping the features->Input into the first maximum pooling layer, and output to obtain characteristic diagram +.>e-4) a second residual error initial module is composed of a first branch, a second branch, a third branch, a fourth branch and a fifth branch, wherein the first branch is composed of a convolution layer, the second branch is sequentially composed of an average pooling layer and a convolution layer, the third branch is sequentially composed of a convolution layer and a BN layer, the fourth branch is sequentially composed of a first convolution layer, a second convolution layer and a BN layer, and the fifth branch is sequentially composed of the first convolution layer, the second convolution layer, the first BN layer, the third convolution layer and the second BN layer, so that the characteristic diagram is obtained >Input into the first branch, output the obtained feature map +.>Feature map +.>Input into the second branch, output the obtained feature map +.>Map the characteristic mapInput into the third branch, output the obtained feature map +.>Feature map +.>Input into the fourth branch, output the obtained feature map +.>Feature map +.>Input into the fifth branch, output the obtained feature map +.>Feature map +.>Feature map->Feature map->Feature map->Performing splicing operation to obtain characteristic diagram->Feature map +.>And (4) feature map>The addition operation results in a feature map->e-5) mapping the characteristic pattern->Inputting into the second maximum pooling layer, outputting to obtain characteristic diagram +.>e-6) a third residual error initial module is composed of a first branch, a second branch, a third branch, a fourth branch and a fifth branch, wherein the first branch is composed of a convolution layer, the second branch is sequentially composed of an average pooling layer and a convolution layer, and the third branch is sequentially composed of convolution layersThe fourth branch is sequentially composed of a first convolution layer, a second convolution layer and a BN layer, the fifth branch is sequentially composed of the first convolution layer, the second convolution layer, the first BN layer, the third convolution layer and the second BN layer, and the characteristic diagram is expressed by->Input into the first branch, output the obtained feature map +.>Feature map +. >Input into the second branch, output the obtained feature map +.>Feature map +.>Input into the third branch, output the obtained feature map +.>Feature map +.>Input into the fourth branch, output the obtained feature map +.>Feature map +.>Input into the fifth branch, output the obtained feature map +.>Feature map +.>Feature map->Feature map->Feature map->Performing splicing operation to obtain characteristic diagram->Feature map +.>And (4) feature map>The addition operation results in a feature map->e-7) characterizing diagrams->Inputting into the third maximum pooling layer, outputting to obtain characteristic diagram +.>e-8) a fourth residual error initial module is composed of a first branch, a second branch, a third branch, a fourth branch and a fifth branch, wherein the first branch is composed of a convolution layer, the second branch is sequentially composed of an average pooling layer and a convolution layer, the third branch is sequentially composed of a convolution layer and a BN layer, the fourth branch is sequentially composed of a first convolution layer, a second convolution layer and a BN layer, and the fifth branch is sequentially composed of the first convolution layer, the second convolution layer, the first BN layer, the third convolution layer and the second BN layer, so that the characteristic diagram is obtained>Input into the first branch, output the obtained feature map +.>Feature map +.>Input into the second branch, output the obtained feature map +.>Feature map +. >Input into the third branch, output the obtained feature map +.>Feature map +.>Input into the fourth branch, output the obtained feature map +.>Feature map +.>Input into the fifth branch, output the obtained feature map +.>Map the characteristic mapFeature map->Feature map->Feature map->Performing splicing operationObtaining a characteristic map->Map the characteristic mapAnd (4) feature map>The addition operation results in a feature map->Preferably, in step e-2), the convolution kernel size of the first branch is 1×1, the convolution kernel size of the second branch is 1×1, the convolution kernel size of the third branch is 1×01, the convolution kernel size of the first convolution layer of the fourth branch is 1×11, the convolution kernel size of the second convolution layer is 5×25, the padding is 2, the convolution kernel size of the first convolution layer of the fifth branch is 1×31, the convolution kernel size of the second convolution layer is 3×43, and the convolution kernel size of the third convolution layer is 3×53; in the step e-4), the convolution kernel size of the convolution layer of the first branch is 1×61, the convolution kernel size of the convolution layer of the second branch is 1×71, the convolution kernel size of the convolution layer of the third branch is 1×81, the convolution kernel size of the first convolution layer of the fourth branch is 1×91, the convolution kernel size of the second convolution layer is 5×5, the packing is 2, the convolution kernel size of the first convolution layer of the fifth branch is 1×01, the convolution kernel size of the second convolution layer is 3×13, and the convolution kernel size of the third convolution layer is 3×23; in the step e-6), the convolution kernel size of the convolution layer of the first branch is 1×31, the convolution kernel size of the convolution layer of the second branch is 1×41, the convolution kernel size of the convolution layer of the third branch is 1×51, the convolution kernel size of the first convolution layer of the fourth branch is 1×1, the convolution kernel size of the second convolution layer is 5×5, the packing is 2, the convolution kernel size of the first convolution layer of the fifth branch is 1×1, the convolution kernel size of the second convolution layer is 3×3, and the convolution kernel size of the third convolution layer is 3×3; step e-8) wherein the convolution kernel size of the convolution layer of the first branch is 1×1, the convolution kernel size of the convolution layer of the second branch is 1×1, and the convolution of the third branch The convolution kernel size of the layers is 1×1, the convolution kernel size of the first convolution layer of the fourth branch is 1×1, the convolution kernel size of the second convolution layer is 5×5, the padding is 2, the convolution kernel size of the first convolution layer of the fifth branch is 1×1, the convolution kernel size of the second convolution layer is 3×3, and the convolution kernel size of the third convolution layer is 3×3; the convolution kernel size of the first maximum pooling layer in step e-3) is 2 x 2; the convolution kernel size of the second largest pooling layer in step e-5) is 2 x 2; the convolution kernel size of the third largest pooling layer in step e-7) is 2 x 2.
Further, step f) comprises the steps of:
f-1) the decoder of the split network model is composed of Conv-block1, a first attention gating module, a first upsampling layer, a first context attention module, a first convolution layer, conv-block2, a second attention gating module, a second upsampling layer, a second context attention module, a second convolution layer, conv-block3, a third attention gating module, a third upsampling layer, a third context attention module, a third convolution layer, conv-block4, a fourth attention gating module, a fourth context attention module and a fourth convolution layer;
f-2) Conv-block4 is composed of a convolution layer, a BatchNorm layer and a ReLU activation function in sequence, and is characterized in that Input into Conv-block4, and output to obtain characteristic diagram +.>The fourth attention gating module consists of a first convolution layer, a first BN layer, a second convolution layer, a second BN layer, a ReLU activation function, a third convolution layer, a third BN layer and a sigmoid function, and is used for decoding a feature map->Sequentially inputting into a first convolution layer and a first BN layer of a fourth attention gate control module, and outputting to obtain a characteristic diagramFeature map +.>Sequentially inputting into a second convolution layer and a second BN layer of a fourth attention gating module, and outputting to obtain a characteristic diagram ++>Feature map +.>And (4) feature map>Adding to obtain a feature map->Feature map +.>Sequentially inputting into a ReLU activation function, a third convolution layer, a third BN layer and a sigmoid function of a fourth attention gating module, and outputting to obtain a feature map +.>Feature map +.>And (4) feature map>Multiplication to obtain a feature map->The fourth context attention module consists of a first convolution layer, a second convolution layer, a global maximum pooling layer, a third convolution layer, a fourth convolution layer and a sigmoid function, and is used for adding a characteristic diagram->Sequentially conveyingThe first and second convolution layers of the fourth context attention module are input to output and obtain a characteristic diagram +.>Feature map +.>Input into the global maximum pooling layer of the fourth contextual attention module, output the resulting feature map +. >Feature map +.>Sequentially inputting into a third convolution layer and a fourth convolution layer of a fourth context attention module, and outputting to obtain a characteristic diagram ++>Feature map +.>And (4) feature map>Multiplication to obtain a feature map->Map the characteristic mapAnd (4) feature map>Adding to obtain a feature map->Feature map +.>Inputting into the sigmoid function of the fourth context attention module, outputting the obtained feature map +.>Feature map +.>Inputting into a fourth convolution layer to obtain a feature mapf-3) Conv-block3 is composed of a convolution layer, a BatchNorm layer and a ReLU activation function in sequence, and a characteristic diagram is +.>Input into Conv-block3, and output to obtain characteristic diagram +.>Feature map +.>Input into the third upsampling layer, output to get the feature map +.>The third attention gating module consists of a first convolution layer, a first BN layer, a second convolution layer, a second BN layer, a ReLU activation function, a third convolution layer, a third BN layer and a sigmoid function, and is used for decoding a characteristic diagram->Sequentially inputting into a first convolution layer and a first BN layer of a third attention gate control module, and outputting to obtain a characteristic diagram +.>Feature map +.>Sequentially input to a third attention gateIn the second convolution layer and the second BN layer of the module, a characteristic diagram is output and obtained>Feature map +.>And feature mapAdding to obtain a feature map- >Feature map +.>Sequentially inputting into a ReLU activation function, a third convolution layer, a third BN layer and a sigmoid function of a third attention gating module, and outputting to obtain a feature map +.>Feature map +.>And (4) feature map>Multiplication to obtain a feature map->Feature map +.>And (4) feature map>Cascading get feature map->Third context attention module is composed of first convolution layer, second convolution layer and third convolution layerThe two convolution layers, the global maximum pooling layer, the third convolution layer, the fourth convolution layer and the sigmoid function constitute a characteristic diagram +.>Sequentially inputting into the first convolution layer and the second convolution layer of the third context attention module, and outputting to obtain a characteristic diagram ++>Feature map +.>Input into the global maximum pooling layer of the third contextual attention module, output the resulting feature map +.>Feature map +.>Sequentially inputting into a third convolution layer and a fourth convolution layer of a third context attention module, and outputting to obtain a characteristic diagram ++>Feature map +.>And (4) feature map>Multiplication to obtain a feature map->Feature map +.>And (4) feature map>Adding to obtain a feature map->Feature map +.>Inputting into the sigmoid function of the third context attention module, outputting the obtained feature map +.>Feature map +.>Inputting into the third convolution layer to obtain a feature map +.>f-4) Conv-block2 is composed of a convolution layer, a BatchNorm layer and a ReLU activation function in sequence, and a characteristic diagram is +. >Input into Conv-block2, and output to obtain feature map +.>Feature map +.>Input into the second upsampling layer, output to get the feature map +.>The second attention gating module consists of a first convolution layer, a first BN layer, a second convolution layer, a second BN layer, a ReLU activation function, a third convolution layer, a third BN layer and a sigmoid function, and is used for decoding a characteristic diagram->Sequentially inputting into a first convolution layer and a first BN layer of a second attention gating module, and outputting to obtain a characteristic diagram ++>Feature map +.>Sequentially inputting into a second convolution layer and a second BN layer of a second attention gate control module, and outputting to obtain a characteristic diagram +.>Feature map +.>And (4) feature map>Adding to obtain a feature map->Feature map +.>Sequentially inputting into a ReLU activation function, a third convolution layer, a third BN layer and a sigmoid function of the second attention gating module, and outputting to obtain a feature map +.>Map the characteristic mapAnd (4) feature map>Multiplication to obtain a feature map->Feature map +.>And (4) feature map>Cascading get feature map->The second context attention module consists of a first convolution layer, a second convolution layer, a global maximum pooling layer, a third convolution layer, a fourth convolution layer and a sigmoid function, and is used for adding a characteristic diagram->Sequentially inputting into a first convolution layer and a second convolution layer of a second context attention module, and outputting to obtain a characteristic diagram ++ >Feature map +.>Input into the global maximization layer of the second contextual attention module, output the resulting feature map +.>Feature map +.>Sequentially inputting into a third convolution layer and a fourth convolution layer of the second context attention module, and outputting to obtain a characteristic diagram ++>Feature map +.>And (4) feature map>Multiplication to obtain a feature map->Feature map +.>And (4) feature map>Adding to obtain a feature map->Feature map +.>Inputting into the sigmoid function of the second context attention module, outputting the obtained feature map +.>Feature map +.>Inputting the first convolution layer to obtain a characteristic diagram +.>f-5) Conv-block1 is composed of a convolution layer, a BatchNorm layer and a ReLU activation function in sequence, and a characteristic diagram is +.>Input into Conv-block1, and output to obtain characteristic diagram +.>Feature map +.>Input into the first upsampling layer, output to get the feature map +.>The first attention gating module consists of a first convolution layer, a first BN layer, a second convolution layer, a second BN layer, a ReLU activation function, a third convolution layer, a third BN layer and a sigmoid function, and is used for decoding a characteristic diagram->Sequentially inputting into a first convolution layer and a first BN layer of a first attention gate control module, and outputting to obtain a characteristic diagram ++>Feature map +.>Sequentially inputting into a second convolution layer and a second BN layer of the first attention gating module, and outputting to obtain a characteristic diagram ++ >Map the characteristic mapAnd (4) feature map>Adding to obtain a feature map->Feature map +.>Sequentially inputting into a ReLU activation function, a third convolution layer, a third BN layer and a sigmoid function of the first attention gating module, and outputting to obtain a feature map +.>Feature map +.>And (4) feature map>Multiplication to obtain a feature map->Feature map +.>And (4) feature map>Cascading get feature map->The first context attention module consists of a first convolution layer, a second convolution layer, a global maximum pooling layer, a third convolution layer, a fourth convolution layer and a sigmoid function, and is used for adding a characteristic diagram->Sequentially inputting into a first convolution layer and a second convolution layer of a first context attention module, and outputting to obtain a characteristic diagram ++>Feature map +.>Input into the global maximum pooling layer of the first contextual attention module, output the resulting feature map +.>Feature map +.>Sequentially inputting into a third convolution layer and a fourth convolution layer of the first context attention module, and outputting to obtain a characteristic diagram ++>Feature map +.>And (4) feature map>Multiplication to obtain a feature map->Feature map +.>And (4) feature map>Adding to obtain a feature map->Feature map +.>Inputting into the sigmoid function of the first context attention module, outputting the obtained feature map +.>Feature map +.>Inputting into the first convolution layer to obtain a feature map +. >f-6) mapping of the characteristics->Feature map->Feature map->Feature map->Residual addition is carried out to obtain a characteristic diagram->
Preferably, the convolution kernel size of the convolution layer of Conv-block4 in step f-2) is 3×3 and padding is 1; the convolution kernel sizes of the first convolution layer, the second convolution layer and the third convolution layer of the fourth attention gating module are all 1 multiplied by 1; the convolution kernel sizes of the first convolution layer, the second convolution layer and the third convolution layer of the fourth context attention module are 3×3, the padding is 1, and the convolution kernel size of the fourth convolution layer of the fourth context attention module is 1×1; the convolution kernel size of the fourth convolution layer of the decoder is 1×1; the convolution kernel size of the convolution layer of Conv-block3 in step f-3) is 3×3, and padding is 1; the convolution kernel sizes of the first convolution layer, the second convolution layer and the third convolution layer of the third attention gating module are all 1 multiplied by 1; the convolution kernel size of the third upsampling layer is 2×2; the convolution kernel sizes of the first convolution layer, the second convolution layer and the third convolution layer of the third context attention module are 3×3, the padding is 1, and the convolution kernel size of the fourth convolution layer of the third context attention module is 1×1; the convolution kernel size of the third convolution layer of the decoder is 1×1; f-4) the convolution kernel size of the convolution layer of Conv-block2 is 3×3, and padding is 1; the convolution kernel sizes of the first convolution layer, the second convolution layer and the third convolution layer of the second attention gating module are all 1 multiplied by 1; the convolution kernel size of the second upsampling layer is 2 x 2; the convolution kernel sizes of the first convolution layer, the second convolution layer and the third convolution layer of the second context attention module are all 3×3, the padding is 1, and the convolution kernel size of the fourth convolution layer of the second context attention module is 1×1; the convolution kernel size of the second convolution layer of the decoder is 1×1; f-5) the convolution kernel size of the convolution layer of Conv-block1 is 3×3, and padding is 1; the convolution kernel sizes of the first convolution layer, the second convolution layer and the third convolution layer of the first attention gating module are all 1 multiplied by 1; the convolution kernel size of the first upsampling layer is 2 x 2; the convolution kernel sizes of the first convolution layer, the second convolution layer and the third convolution layer of the first context attention module are all 3×3, the padding is 1, and the convolution kernel size of the fourth convolution layer of the first context attention module is 1×1; the convolution kernel size of the first convolution layer of the decoder is 1 x 1. Further, in step g), the Dice loss and the cross entropy loss are summed to obtain total loss, an Adam optimizer is used for training the segmentation network model by using the total loss to obtain an optimized segmentation network model, the batch size during training is set to be 32, the iteration period is set to be 200, and the learning rate is set to be 0.001.
The beneficial effects of the invention are as follows: and a residual error initial module is introduced into the encoder to better learn effective characteristic representation, so that the performance and generalization capability of the model are improved. The features of each layer are refined by concatenation operations with the complementary information of the different layers in the decoder, while the context information of each layer is explored by preserving local information and compressing global information using a context attention module, and salient features are highlighted by suppressing unimportant information areas, thereby improving the accuracy of segmentation.
Drawings
FIG. 1 is a block diagram of a split network model of the present invention;
FIG. 2 is a block diagram of a residual initialization module of the present invention;
FIG. 3 is a block diagram of an attention gating module of the present invention;
fig. 4 is a block diagram of a contextual attention module of the present invention.
Detailed Description
The invention is further described with reference to fig. 1 to 4.
A method of cardiac MRI segmentation based on contextual cascade attention, comprising the steps of:
a) Collecting cardiac MRI data of n subjects to obtain MRI data sets s, s= { s 1 ,s 2 ,...,s i ,...,s n },s i Cardiac MRI data for the ith subject, i e {1,2,..n }.
b) Preprocessing the MRI data set s to obtain a preprocessed data set F, F= { F 1 ,F 2 ,...,F i ,...,F n },F i Is the i-th preprocessed two-dimensional image data.
c) The preprocessed data set F is divided into a training set, a testing set and a verification set.
d) A partitioning network model is built consisting of an encoder and a decoder.
e) The preprocessed two-dimensional image data F i Input into encoder of dividing network model, and output to obtain characteristic diagramFeature map->Feature map->Feature map->
f) Map the characteristic mapFeature map->Feature map->Feature map->Input into decoder of segmentation network model, output to obtain predictive segmentation image>
g) Training the segmentation network model to obtain an optimized segmentation network model.
Complex data features are better captured and represented in the encoder using a residual initialization module, and features of multi-scale and multi-resolution spatial representation are learned in the decoder through multi-layer contextual concatenation operations.
In one embodiment of the invention, from ACDC2017 common data in step a), cardiac MRI data of n subjects containing LV, RV, MYO structures are collected, resulting in MRI dataset s. In this embodiment, preferably, n=100 in step a).
In one embodiment of the invention, step b) comprises the steps of:
b-1) cardiac MRI data s for the ith subject i The resampling is performed slice by slice along the z-axis with a pixel pitch of 1.5 in the x-axis direction and a pixel pitch of 1.5 in the y-axis direction.
b-2) resampling cardiac MRI data s i Performing 2D center clipping operation with clipping size 224×224 to obtain clipped data F i ' in order to ensure the consistency of the data, the cut data F i ' normalization processing is carried out to obtain preprocessed two-dimensional image data F i 。
In one embodiment of the invention, the preprocessed data set F is divided into a training set, a test set and a verification set according to the ratio of 7:2:1 in step c).
In one embodiment of the invention, step e) comprises the steps of:
the e-1) encoder for dividing the network model is composed of a first residual initial module, a first maximum pooling layer, a second residual initial module, a second maximum pooling layer, a third residual initial module, a third maximum pooling layer and a fourth residual initial module.
e-2) the first residual error initial module is composed of a first branch, a second branch, a third branch, a fourth branch and a fifth branch, wherein the first branch is composed of a convolution layer, the second branch is sequentially composed of an average pooling layer and a convolution layer, the third branch is sequentially composed of a convolution layer and a BN layer, the fourth branch is sequentially composed of the first convolution layer, the second convolution layer and the BN layer, the fifth branch is sequentially composed of the first convolution layer, the second convolution layer, the first BN layer, the third convolution layer and the second BN layer, and the preprocessed two-dimensional image data F is obtained by i Input into the first branch, output to obtain a feature mapThe preprocessed two-dimensional image data F i Input into the second branch, output the obtained feature map +.>The preprocessed two-dimensional image data F i Input into the third branch, output the obtained feature map +.>The preprocessed two-dimensional image data F i Input into the fourth branch, output the obtained feature map +.>The preprocessed two-dimensional image data F i Input into the fifth branch, output the obtained feature map +.>Feature map +.>Feature mapFeature map->Feature map->Performing splicing operation to obtain characteristic diagram->Feature map +.>And feature mapThe addition operation results in a feature map->e-3) mapping the features->Input into the first maximum pooling layer, and output to obtain characteristic diagram +.>e-4) a second residual error initial module is composed of a first branch, a second branch, a third branch, a fourth branch and a fifth branch, wherein the first branch is composed of a convolution layer, the second branch is sequentially composed of an average pooling layer and a convolution layer, the third branch is sequentially composed of a convolution layer and a BN layer, the fourth branch is sequentially composed of a first convolution layer, a second convolution layer and a BN layer, and the fifth branch is sequentially composed of the first convolution layer, the second convolution layer, the first BN layer, the third convolution layer and the second BN layer, so that the characteristic diagram is obtained >Input into the first branch, output the obtained feature map +.>Feature map +.>Input into the second branch, output the obtained feature map +.>Map the characteristic mapInput into the third branch, output the obtained feature map +.>Feature map +.>Input into the fourth branch, output the obtained feature map +.>Feature map +.>Input into the fifth branch, output the obtained feature map +.>Feature map +.>Feature map->Feature map->Feature map->Performing splicing operation to obtain characteristic diagram->Feature map +.>And (4) feature map>The addition operation results in a feature map->e-5) mapping the characteristic pattern->Inputting into the second maximum pooling layer, outputting to obtain characteristic diagram +.>e-6) a third residual error initial module is composed of a first branch, a second branch, a third branch, a fourth branch and a fifth branch, wherein the first branch is composed of a convolution layer, the second branch is sequentially composed of an average pooling layer and a convolution layer, the third branch is sequentially composed of a convolution layer and a BN layer, the fourth branch is sequentially composed of a first convolution layer, a second convolution layer and a BN layer, and the fifth branch is sequentially composed of the first convolution layer, the second convolution layer, the first BN layer, the third convolution layer and the second BN layer, so that the characteristic diagram is obtained>Input into the first branch, output the obtained feature map +.>Feature map +. >Input into the second branch, output the obtained feature map +.>Feature map +.>Input into the third branch, output the obtained feature map +.>Feature map +.>Input into the fourth branch, output the obtained feature map +.>Feature map +.>Input into the fifth branch, output the obtained feature map +.>Feature map +.>Feature map->Feature map->Feature map->Performing splicing operation to obtain characteristic diagram->Feature map +.>And (4) feature map>The addition operation results in a feature map->e-7) characterizing diagrams->Inputting into the third maximum pooling layer, outputting to obtain characteristic diagram +.>e-8) a fourth residual error initial module is composed of a first branch, a second branch, a third branch, a fourth branch and a fifth branch, wherein the first branch is composed of a convolution layer, the second branch is sequentially composed of an average pooling layer and a convolution layer, the third branch is sequentially composed of a convolution layer and a BN layer, the fourth branch is sequentially composed of a first convolution layer, a second convolution layer and a BN layer, and the fifth branch is sequentially composed of a first convolution layer, a second convolution layer and a third convolution layerA BN layer, a third convolution layer, a second BN layer, and a feature map +.>Input into the first branch, output the obtained feature map +.>Feature map +.>Input into the second branch, output the obtained feature map +.>Feature map +. >Input into the third branch, output the obtained feature map +.>Feature map +.>Input into the fourth branch, output the obtained feature map +.>Feature map +.>Input into the fifth branch, output the obtained feature map +.>Feature map +.>Feature map->Feature map->Feature map->Performing splicing operation to obtain characteristic diagram->Feature map +.>And (4) feature map>The addition operation results in a feature map->In this embodiment, preferably, in step e-2), the convolution kernel size of the convolution layer of the first branch is 1×1, the convolution kernel size of the convolution layer of the second branch is 1×1, the convolution kernel size of the convolution layer of the third branch is 1×01, the convolution kernel size of the first convolution layer of the fourth branch is 1×11, the convolution kernel size of the second convolution layer is 5×25, the padding is 2, the convolution kernel size of the first convolution layer of the fifth branch is 1×31, the convolution kernel size of the second convolution layer is 3×43, and the convolution kernel size of the third convolution layer is 3×53; in the step e-4), the convolution kernel size of the convolution layer of the first branch is 1×61, the convolution kernel size of the convolution layer of the second branch is 1×71, the convolution kernel size of the convolution layer of the third branch is 1×81, the convolution kernel size of the first convolution layer of the fourth branch is 1×91, the convolution kernel size of the second convolution layer is 5×5, the packing is 2, the convolution kernel size of the first convolution layer of the fifth branch is 1×1, the convolution kernel size of the second convolution layer is 3×3, and the convolution kernel size of the third convolution layer is 3×3; the convolution kernel size of the convolution layer of the first branch in step e-6) is 1×1, the convolution kernel size of the convolution layer of the second branch is 1×1, the convolution kernel size of the convolution layer of the third branch is 1×1, the convolution kernel size of the first convolution layer of the fourth branch is 1×1, the convolution kernel size of the second convolution layer is 1×1 The size of the convolution kernel of the first convolution layer of the fifth branch is 1 multiplied by 1, the convolution kernel of the second convolution layer is 3 multiplied by 03, and the convolution kernel of the third convolution layer is 3 multiplied by 13, which are 5 multiplied by 5, and padding is 2; in the step e-8), the convolution kernel size of the convolution layer of the first branch is 1×21, the convolution kernel size of the convolution layer of the second branch is 1×31, the convolution kernel size of the convolution layer of the third branch is 1×41, the convolution kernel size of the first convolution layer of the fourth branch is 1×1, the convolution kernel size of the second convolution layer is 5×5, the padding is 2, the convolution kernel size of the first convolution layer of the fifth branch is 1×1, the convolution kernel size of the second convolution layer is 3×3, and the convolution kernel size of the third convolution layer is 3×3; the convolution kernel size of the first maximum pooling layer in step e-3) is 2 x 2; the convolution kernel size of the second largest pooling layer in step e-5) is 2 x 2; the convolution kernel size of the third largest pooling layer in step e-7) is 2 x 2. In one embodiment of the invention, step f) comprises the steps of:
f-1) the decoder of the split network model is composed of Conv-block1, a first attention gating module, a first upsampling layer, a first context attention module, a first convolution layer, conv-block2, a second attention gating module, a second upsampling layer, a second context attention module, a second convolution layer, conv-block3, a third attention gating module, a third upsampling layer, a third context attention module, a third convolution layer, conv-block4, a fourth attention gating module, a fourth context attention module, and a fourth convolution layer.
f-2) Conv-block4 is composed of a convolution layer, a BatchNorm layer and a ReLU activation function in sequence, and is characterized in thatInput into Conv-block4, and output to obtain characteristic diagram +.>The fourth attention gating module consists of a first convolution layer, a first BN layer, a second convolution layer, a second BN layer, a ReLU activation function, a third convolution layer, a third BN layer and a sigmoid function, and is used for decoding a feature map->Sequentially inputting into a first convolution layer and a first BN layer of a fourth attention gate control module, and outputting to obtain a characteristic diagramFeature map +.>Sequentially inputting into a second convolution layer and a second BN layer of a fourth attention gating module, and outputting to obtain a characteristic diagram ++>Feature map +.>And (4) feature map>Adding to obtain a feature map->Feature map +.>Sequentially inputting into a ReLU activation function, a third convolution layer, a third BN layer and a sigmoid function of a fourth attention gating module, and outputting to obtain a feature map +.>Feature map +.>And (4) feature map>Multiplication to obtain a feature map->Fourth context noteThe force intention module consists of a first convolution layer, a second convolution layer, a global maximum pooling layer, a third convolution layer, a fourth convolution layer and a sigmoid function, and is used for designing a characteristic diagram->Sequentially inputting into a first convolution layer and a second convolution layer of a fourth context attention module, and outputting to obtain a characteristic diagram +. >Feature map +.>Input into the global maximum pooling layer of the fourth contextual attention module, output the resulting feature map +.>Feature map +.>Sequentially inputting into a third convolution layer and a fourth convolution layer of a fourth context attention module, and outputting to obtain a characteristic diagram ++>Feature map +.>And (4) feature map>Multiplication to obtain a feature map->Map the characteristic mapAnd (4) feature map>Adding to obtain a feature map->Feature map +.>Inputting into the sigmoid function of the fourth context attention module, outputting the obtained feature map +.>Feature map +.>Inputting into a fourth convolution layer to obtain a feature mapf-3) Conv-block3 is composed of a convolution layer, a BatchNorm layer and a ReLU activation function in sequence, and a characteristic diagram is +.>Input into Conv-block3, and output to obtain characteristic diagram +.>Feature map +.>Input into the third upsampling layer, output to get the feature map +.>The third attention gating module consists of a first convolution layer, a first BN layer, a second convolution layer, a second BN layer, a ReLU activation function, a third convolution layer, a third BN layer and a sigmoid function, and is used for decoding a characteristic diagram->Sequentially inputting into a first convolution layer and a first BN layer of a third attention gate control module, and outputting to obtain a characteristic diagram +.>Feature map +.>Sequentially inputting into a second convolution layer and a second BN layer of a third attention gate control module, and outputting to obtain a characteristic diagram +. >Feature map +.>And feature mapAdding to obtain a feature map->Feature map +.>Sequentially inputting into a ReLU activation function, a third convolution layer, a third BN layer and a sigmoid function of a third attention gating module, and outputting to obtain a feature map +.>Feature map +.>And (4) feature map>Multiplication to obtain a feature map->Feature map +.>And (4) feature map>Cascading get feature map->The third context attention module consists of a first convolution layer, a second convolution layer, a global maximum pooling layer, a third convolution layer, a fourth convolution layer and a sigmoid function, and is used for adding a characteristic diagram->Sequentially inputting into the first convolution layer and the second convolution layer of the third context attention module, and outputting to obtain a characteristic diagram ++>Feature map +.>Input into the global maximum pooling layer of the third contextual attention module, output the resulting feature map +.>Feature map +.>Sequentially inputting into a third convolution layer and a fourth convolution layer of a third context attention module, and outputting to obtain a characteristic diagram ++>Feature map +.>And (4) feature map>Multiplication to obtain a feature mapFeature map +.>And (4) feature map>Adding to obtain a feature map->Feature map +.>Inputting into the sigmoid function of the third context attention module, outputting the obtained feature map +.>Feature map +.>Inputting into the third convolution layer to obtain a feature map +. >f-4) Conv-block2 is composed of a convolution layer, a BatchNorm layer and a ReLU activation function in sequence, and a characteristic diagram is +.>Input into Conv-block2, and output to obtain feature map +.>Feature map +.>Input into the second upsampling layer, output to get the feature map +.>The second attention gating module consists of a first convolution layer, a first BN layer, a second convolution layer, a second BN layer, a ReLU activation function, a third convolution layer, a third BN layer and a sigmoid function, and is to be specificSyndrome/pattern of->Sequentially inputting into a first convolution layer and a first BN layer of a second attention gating module, and outputting to obtain a characteristic diagram ++>Feature map +.>Sequentially inputting into a second convolution layer and a second BN layer of a second attention gate control module, and outputting to obtain a characteristic diagram +.>Feature map +.>And (4) feature map>Adding to obtain a feature map->Feature map +.>Sequentially inputting into a ReLU activation function, a third convolution layer, a third BN layer and a sigmoid function of the second attention gating module, and outputting to obtain a feature map +.>Feature map +.>And (4) feature map>Multiplication to obtain a feature map->Feature map +.>And (4) feature map>Cascading get feature map->The second context attention module consists of a first convolution layer, a second convolution layer, a global maximum pooling layer, a third convolution layer, a fourth convolution layer and a sigmoid function, and is used for adding a characteristic diagram- >Sequentially inputting into a first convolution layer and a second convolution layer of a second context attention module, and outputting to obtain a characteristic diagram ++>Feature map +.>Input into the global maximization layer of the second contextual attention module, output the resulting feature map +.>Feature map +.>Sequentially inputting into a third convolution layer and a fourth convolution layer of the second context attention module, and outputting to obtain a characteristic diagram ++>Feature map +.>And (4) feature map>Multiplication to obtain a feature mapFeature map +.>And (4) feature map>Adding to obtain a feature map->Feature map +.>Inputting into the sigmoid function of the second context attention module, outputting the obtained feature map +.>Feature map +.>Inputting the first convolution layer to obtain a characteristic diagram +.>f-5) Conv-block1 is composed of a convolution layer, a BatchNorm layer and a ReLU activation function in sequence, and a characteristic diagram is +.>Input into Conv-block1, and output to obtain characteristic diagram +.>Feature map +.>Input into the first upsampling layer, output to get the feature map +.>The first attention gating module consists of a first convolution layer, a first BN layer, a second convolution layer, a second BN layer, a ReLU activation function, a third convolution layer, a third BN layer and a sigmoid function, and is used for decoding a characteristic diagram->Sequentially inputting into a first convolution layer and a first BN layer of a first attention gate control module, and outputting to obtain a characteristic diagram ++ >Feature map +.>Sequentially inputting into a second convolution layer and a second BN layer of the first attention gating module, and outputting to obtain a characteristic diagram ++>Feature map +.>And (4) feature map>Adding to obtain a feature map->Feature map +.>Sequentially inputting into a ReLU activation function, a third convolution layer, a third BN layer and a sigmoid function of the first attention gating module, and outputting to obtain a feature map +.>Map the characteristic mapAnd (4) feature map>Multiplication to obtain a feature map->Feature map +.>And (4) feature map>Cascading get feature map->The first context attention module consists of a first convolution layer, a second convolution layer, a global maximum pooling layer, a third convolution layer, a fourth convolution layer and a sigmoid function, and is used for adding a characteristic diagram->Sequentially inputting into a first convolution layer and a second convolution layer of a first context attention module, and outputting to obtain a characteristic diagram ++>Feature map +.>Input into the global maximum pooling layer of the first contextual attention module, output the resulting feature map +.>Feature map +.>Sequentially inputting into a third convolution layer and a fourth convolution layer of the first context attention module, and outputting to obtain a characteristic diagram ++>Feature map +.>And (4) feature map>Multiplication to obtain a feature map->Feature map +.>And (4) feature map>Adding to obtain a feature map- >Feature map +.>Inputting into the sigmoid function of the first context attention module, outputting the obtained feature map +.>Feature map +.>Inputting into the first convolution layer to obtain a feature map +.>f-6) mapping of the characteristics->Feature map->Feature map->Feature map->Residual addition is carried out to obtain a characteristic diagram->
In this embodiment, it is preferred that the convolution kernel size of the convolutions layer of Conv-block4 in step f-2) is 3×3 and padding is 1; the convolution kernel sizes of the first convolution layer, the second convolution layer and the third convolution layer of the fourth attention gating module are all 1 multiplied by 1; the convolution kernel sizes of the first convolution layer, the second convolution layer and the third convolution layer of the fourth context attention module are 3×3, the padding is 1, and the convolution kernel size of the fourth convolution layer of the fourth context attention module is 1×1; the convolution kernel size of the fourth convolution layer of the decoder is 1×1; the convolution kernel size of the convolution layer of Conv-block3 in step f-3) is 3×3, and padding is 1; the convolution kernel sizes of the first convolution layer, the second convolution layer and the third convolution layer of the third attention gating module are all 1 multiplied by 1; the convolution kernel size of the third upsampling layer is 2×2; the convolution kernel sizes of the first convolution layer, the second convolution layer and the third convolution layer of the third context attention module are 3×3, the padding is 1, and the convolution kernel size of the fourth convolution layer of the third context attention module is 1×1; the convolution kernel size of the third convolution layer of the decoder is 1×1; f-4) the convolution kernel size of the convolution layer of Conv-block2 is 3×3, and padding is 1; the convolution kernel sizes of the first convolution layer, the second convolution layer and the third convolution layer of the second attention gating module are all 1 multiplied by 1; the convolution kernel size of the second upsampling layer is 2 x 2; the convolution kernel sizes of the first convolution layer, the second convolution layer and the third convolution layer of the second context attention module are all 3×3, the padding is 1, and the convolution kernel size of the fourth convolution layer of the second context attention module is 1×1; the convolution kernel size of the second convolution layer of the decoder is 1×1; f-5) the convolution kernel size of the convolution layer of Conv-block1 is 3×3, and padding is 1; the convolution kernel sizes of the first convolution layer, the second convolution layer and the third convolution layer of the first attention gating module are all 1 multiplied by 1; the convolution kernel size of the first upsampling layer is 2 x 2; the convolution kernel sizes of the first convolution layer, the second convolution layer and the third convolution layer of the first context attention module are all 3×3, the padding is 1, and the convolution kernel size of the fourth convolution layer of the first context attention module is 1×1; the convolution kernel size of the first convolution layer of the decoder is 1 x 1.
In one embodiment of the present invention, in step g), the Dice loss and the cross entropy loss are summed to obtain a total loss, and the Adam optimizer is used to train the segmentation network model by using the total loss to obtain an optimized segmentation network model, wherein the batch size during training is set to 32, the iteration cycle is set to 200, and the learning rate is set to 0.001.
Finally, it should be noted that: the foregoing description is only a preferred embodiment of the present invention, and the present invention is not limited thereto, but it is to be understood that modifications and equivalents of some of the technical features described in the foregoing embodiments may be made by those skilled in the art, although the present invention has been described in detail with reference to the foregoing embodiments. Any modification, equivalent replacement, improvement, etc. made within the spirit and principle of the present invention should be included in the protection scope of the present invention.
Claims (9)
1. A method of cardiac MRI segmentation based on contextual cascade attention, comprising the steps of:
a) Collecting cardiac MRI data of n subjects to obtain MRI data sets s, s= { s 1 ,s 2 ,...,s i ,...,s n },s i Cardiac MRI data for the i-th subject, i e {1,2,., n };
b) Preprocessing the MRI data set s to obtain a preprocessed data set F, F= { F 1 ,F 2 ,...,F i ,...,F n },F i The i-th preprocessed two-dimensional image data;
c) Dividing the preprocessed data set F into a training set, a testing set and a verification set;
d) Establishing a segmentation network model formed by an encoder and a decoder;
e) The preprocessed two-dimensional image data F i Input into encoder of dividing network model, and output to obtain characteristic diagramFeature map->Feature map->Feature map->
f) Map the characteristic mapFeature map->Feature map->Feature map->Input into decoder of segmentation network model, output to obtain predictive segmentation image>
g) Training a segmentation network model to obtain an optimized segmentation network model;
step e) comprises the steps of:
e-1) an encoder for dividing a network model is composed of a first residual initial module, a first maximum pooling layer, a second residual initial module, a second maximum pooling layer, a third residual initial module, a third maximum pooling layer and a fourth residual initial module;
e-2) the first residual error initial module is composed of a first branch, a second branch, a third branch, a fourth branch and a fifth branch, wherein the first branch is composed of a convolution layer, the second branch is sequentially composed of an average pooling layer and a convolution layer, the third branch is sequentially composed of a convolution layer and a BN layer, the fourth branch is sequentially composed of the first convolution layer, the second convolution layer and the BN layer, the fifth branch is sequentially composed of the first convolution layer, the second convolution layer, the first BN layer, the third convolution layer and the second BN layer, and the preprocessed two-dimensional image data F is obtained by i Input into the first branch, output to obtain a feature mapThe preprocessed two-dimensional image data F i Input into the second branch, output the obtained feature map +.>The preprocessed two-dimensional image data F i Input into the third branch, output to obtain a feature mapThe preprocessed two-dimensional image data F i Input into the fourth branch, output the obtained feature map +.>The preprocessed two-dimensional image data F i Input into the fifth branch, output the obtained feature map +.>Feature map +.>Feature map->Feature map->Feature map->Performing splicing operation to obtain characteristic diagram->Feature map +.>And (4) feature map>The addition operation results in a feature map->
e-3) mapping the featuresInput into the first maximum pooling layer, and output to obtain characteristic diagram +.>
e-4) a second residual error initial module is composed of a first branch, a second branch, a third branch, a fourth branch and a fifth branch, wherein the first branch is composed of a convolution layer, the second branch is sequentially composed of an average pooling layer and a convolution layer, the third branch is sequentially composed of a convolution layer and a BN layer, the fourth branch is sequentially composed of a first convolution layer, a second convolution layer and a BN layer, the fifth branch is sequentially composed of the first convolution layer, the second convolution layer, the first BN layer, the third convolution layer and the second BN layer, and the characteristic diagram is formed by Input into the first branchOutputting the obtained characteristic diagram->Feature map +.>Input into the second branch, output the obtained feature map +.>Feature map +.>Input into the third branch, output the obtained feature map +.>Feature map +.>Input into the fourth branch, output the obtained feature map +.>Feature map +.>Input into the fifth branch, output the obtained feature map +.>Feature map +.>Feature mapFeature map->Feature map->Performing splicing operation to obtain characteristic diagram->Feature map +.>And feature mapThe addition operation results in a feature map->
e-5) mapping the featuresInputting into the second maximum pooling layer, outputting to obtain characteristic diagram +.>
e-6) a third residual error initial module is composed of a first branch, a second branch, a third branch, a fourth branch and a fifth branch, wherein the first branch is composed of a convolution layer, the second branch is sequentially composed of an average pooling layer and a convolution layer, the third branch is sequentially composed of a convolution layer and a BN layer, the fourth branch is sequentially composed of a first convolution layer, a second convolution layer and a BN layer, the fifth branch is sequentially composed of the first convolution layer, the second convolution layer, the first BN layer, the third convolution layer and the second BN layer, and the characteristic diagram is formed byInput into the first branch, output the obtained feature map +.>Feature map +. >Input into the second branch, output the obtained feature map +.>Feature map +.>Input into the third branch, output the obtained feature map +.>Feature map +.>Input into the fourth branch, output the obtained feature map +.>Feature map +.>Input into the fifth branch, output the obtained feature map +.>Feature map +.>Feature mapFeature map->Feature map->Performing splicing operation to obtain characteristic diagram->Feature map +.>And feature mapThe addition operation results in a feature map->
e-7) mapping the featuresInputting into the third maximum pooling layer, outputting to obtain characteristic diagram +.>
e-8) a fourth residual error initial module is composed of a first branch, a second branch, a third branch, a fourth branch and a fifth branch, wherein the first branch is composed of a convolution layer, the second branch is sequentially composed of an average pooling layer and a convolution layer, the third branch is sequentially composed of a convolution layer and a BN layer, the fourth branch is sequentially composed of a first convolution layer, a second convolution layer and a BN layer, the fifth branch is sequentially composed of the first convolution layer, the second convolution layer, the first BN layer, the third convolution layer and the second BN layer, and the characteristic diagram is formed byInput into the first branch, output the obtained feature map +.>Feature map +.>Input into the second branch, output the obtained feature map +.>Feature map +. >Input into the third branch, output the obtained feature map +.>Feature map +.>Input into the fourth branch, output the obtained feature map +.>Feature map +.>Input into the fifth branch, output the obtained feature map +.>Feature map +.>Feature mapFeature map->Feature map->Performing splicingThe join operation gets a feature map->Feature map +.>And feature mapThe addition operation results in a feature map->
2. The contextual cascade attention-based cardiac MRI segmentation method according to claim 1, wherein: from ACDC2017 public data in step a), cardiac MRI data of n subjects containing LV, RV, MYO structures were collected, resulting in MRI dataset s.
3. The contextual cascade attention-based cardiac MRI segmentation method according to claim 1, wherein: n=100 in step a).
4. The method of contextual cascade attention-based cardiac MRI segmentation according to claim 1, wherein step b) comprises the steps of:
b-1) cardiac MRI data s for the ith subject i Resampling is carried out on the z-axis one by one, wherein the resampling is that the pixel pitch in the x-axis direction is 1.5, and the pixel pitch in the y-axis direction is 1.5;
b-2) resampling cardiac MRI data s i Performing 2D center clipping operation with clipping size 224×224 to obtain clipped data F i ' cutting the cut data F i ' normalization processing is carried out to obtain preprocessed two-dimensional image data F i 。
5. The contextual cascade attention-based cardiac MRI segmentation method according to claim 1, wherein: in the step c), the preprocessed data set F is divided into a training set, a testing set and a verification set according to the proportion of 7:2:1.
6. The contextual cascade attention-based cardiac MRI segmentation method according to claim 1, wherein: in the step e-2), the convolution kernel size of the convolution layer of the first branch is 1×1, the convolution kernel size of the convolution layer of the second branch is 1×1, the convolution kernel size of the convolution layer of the third branch is 1×1, the convolution kernel size of the first convolution layer of the fourth branch is 1×1, the convolution kernel size of the second convolution layer is 5×5, the packing is 2, the convolution kernel size of the first convolution layer of the fifth branch is 1×1, the convolution kernel size of the second convolution layer is 3×3, and the convolution kernel size of the third convolution layer is 3×3; in the step e-4), the convolution kernel size of the convolution layer of the first branch is 1×1, the convolution kernel size of the convolution layer of the second branch is 1×1, the convolution kernel size of the convolution layer of the third branch is 1×1, the convolution kernel size of the first convolution layer of the fourth branch is 1×1, the convolution kernel size of the second convolution layer is 5×5, the packing is 2, the convolution kernel size of the first convolution layer of the fifth branch is 1×1, the convolution kernel size of the second convolution layer is 3×3, and the convolution kernel size of the third convolution layer is 3×3; in the step e-6), the convolution kernel size of the convolution layer of the first branch is 1×1, the convolution kernel size of the convolution layer of the second branch is 1×1, the convolution kernel size of the convolution layer of the third branch is 1×1, the convolution kernel size of the first convolution layer of the fourth branch is 1×1, the convolution kernel size of the second convolution layer is 5×5, the packing is 2, the convolution kernel size of the first convolution layer of the fifth branch is 1×1, the convolution kernel size of the second convolution layer is 3×3, and the convolution kernel size of the third convolution layer is 3×3; in the step e-8), the convolution kernel size of the convolution layer of the first branch is 1×1, the convolution kernel size of the convolution layer of the second branch is 1×1, the convolution kernel size of the convolution layer of the third branch is 1×1, the convolution kernel size of the first convolution layer of the fourth branch is 1×1, the convolution kernel size of the second convolution layer is 5×5, the packing is 2, the convolution kernel size of the first convolution layer of the fifth branch is 1×1, the convolution kernel size of the second convolution layer is 3×3, and the convolution kernel size of the third convolution layer is 3×3; the convolution kernel size of the first maximum pooling layer in step e-3) is 2 x 2; the convolution kernel size of the second largest pooling layer in step e-5) is 2 x 2; the convolution kernel size of the third largest pooling layer in step e-7) is 2 x 2.
7. The method of contextual cascade attention-based cardiac MRI segmentation according to claim 1, wherein step f) comprises the steps of:
f-1) the decoder of the split network model is composed of Conv-block1, a first attention gating module, a first upsampling layer, a first context attention module, a first convolution layer, conv-block2, a second attention gating module, a second upsampling layer, a second context attention module, a second convolution layer, conv-block3, a third attention gating module, a third upsampling layer, a third context attention module, a third convolution layer, conv-block4, a fourth attention gating module, a fourth context attention module and a fourth convolution layer;
f-2) Conv-block4 is composed of a convolution layer, a BatchNorm layer and a ReLU activation function in sequence, and is characterized in thatInput into Conv-block4, and output to obtain characteristic diagram +.>The fourth attention gating module consists of a first convolution layer, a first BN layer, a second convolution layer, a second BN layer, a ReLU activation function, a third convolution layer, a third BN layer and a sigmoid function, and maps the characteristicsSequentially inputting into a first convolution layer and a first BN layer of a fourth attention gating module, and outputting to obtain a characteristic diagram ++ >Feature map +.>Sequentially inputting into a second convolution layer and a second BN layer of the fourth attention gate module, and outputting to obtain a feature mapFeature map +.>And (4) feature map>Adding to obtain a feature map->Feature map +.>Sequentially inputting into a ReLU activation function, a third convolution layer, a third BN layer and a sigmoid function of a fourth attention gating module, and outputting to obtain a feature map +.>Feature map +.>And (4) feature map>Multiplication to obtain a feature map->The fourth context attention module consists of a first convolution layer, a second convolution layer, a global maximum pooling layer, a third convolution layer, a fourth convolution layer and a sigmoid function, and is used for adding a characteristic diagram->Sequentially inputting the images into a first convolution layer and a second convolution layer of a fourth context attention module, and outputting to obtain a characteristic diagramFeature map +.>Input into the global maximum pooling layer of the fourth context attention module, and output to obtain a feature mapFeature map +.>Sequentially inputting into a third convolution layer and a fourth convolution layer of a fourth context attention module, and outputting to obtain a characteristic diagram ++>Feature map +.>And (4) feature map>Multiplication to obtain a feature map->Feature map +.>And (4) feature map>Adding to obtain a feature map->Feature map +.>Inputting into the sigmoid function of the fourth context attention module, outputting the obtained feature map +. >Feature map +.>Inputting into a fourth convolution layer to obtain a feature map +.>
f-3) Conv-block3 is composed of a convolution layer, a BatchNorm layer and a ReLU activation function in sequence, and is characterized in thatInput into Conv-block3, and output to obtain characteristic diagram +.>Feature map +.>Input into the third upsampling layer, output to get the feature map +.>The third attention gating module consists of a first convolution layer, a first BN layer, a second convolution layer, a second BN layer, a ReLU activation function, a third convolution layer, a third BN layer and a sigmoid function, and is used for decoding a characteristic diagram->Sequentially input to a third attention gating moduleIn the first convolution layer and the first BN layer, the obtained characteristic diagram is output>Feature map +.>Sequentially inputting into a second convolution layer and a second BN layer of a third attention gate control module, and outputting to obtain a characteristic diagram +.>Feature map +.>And feature mapAdding to obtain a feature map->Feature map +.>Sequentially inputting into a ReLU activation function, a third convolution layer, a third BN layer and a sigmoid function of a third attention gating module, and outputting to obtain a feature map +.>Feature map +.>And (4) feature map>Multiplication to obtain a feature map->Feature map +.>And (4) feature map>Cascading get feature map->The third context attention module consists of a first convolution layer, a second convolution layer, a global maximum pooling layer, a third convolution layer, a fourth convolution layer and a sigmoid function, and is used for adding a characteristic diagram- >Sequentially inputting into the first convolution layer and the second convolution layer of the third context attention module, and outputting to obtain a characteristic diagram ++>Feature map +.>Input into the global maximum pooling layer of the third contextual attention module, output the resulting feature map +.>Feature map +.>Sequentially inputting into a third convolution layer and a fourth convolution layer of a third context attention module, and outputting to obtain a characteristic diagram ++>Feature map +.>And (4) feature map>Multiplication to obtain a feature map->Feature map +.>And (4) feature map>Adding to obtain a feature map->Feature map +.>Inputting into the sigmoid function of the third context attention module, outputting the obtained feature map +.>Feature map +.>Inputting into the third convolution layer to obtain a feature map +.>
f-4) Conv-block2 is composed of a convolution layer, a BatchNorm layer and a ReLU activation function in sequence, and is characterized byInput into Conv-block2, and output to obtain feature map +.>Feature map +.>Input into the second upsampling layer, output to get the feature map +.>The second attention gating module consists of a first convolution layer, a first BN layer, a second convolution layer, a second BN layer, a ReLU activation function, a third convolution layer, a third BN layer and a sigmoid function, and is used for decoding a characteristic diagram->Sequentially inputting into a first convolution layer and a first BN layer of a second attention gating module, and outputting to obtain a characteristic diagram ++ >Feature map +.>Sequentially inputting into a second convolution layer and a second BN layer of a second attention gate control module, and outputting to obtain a characteristic diagram +.>Feature map +.>And feature mapAdding to obtain a feature map->Feature map +.>In turnInputting into a ReLU activation function, a third convolution layer, a third BN layer and a sigmoid function of the second attention gating module, and outputting to obtain a feature map +.>Feature map +.>And feature mapMultiplication to obtain a feature map->Feature map +.>And (4) feature map>Cascading get feature map->The second context attention module consists of a first convolution layer, a second convolution layer, a global maximum pooling layer, a third convolution layer, a fourth convolution layer and a sigmoid function, and is used for adding a characteristic diagram->Sequentially inputting into a first convolution layer and a second convolution layer of a second context attention module, and outputting to obtain a characteristic diagram ++>Feature map +.>Global maximum input to the second contextual attention moduleIn the pooling layer, the characteristic diagram is obtained by outputting>Feature map +.>Sequentially inputting into a third convolution layer and a fourth convolution layer of the second context attention module, and outputting to obtain a characteristic diagram ++>Feature map +.>And (4) feature map>Multiplication to obtain a feature map->Feature map +.>And (4) feature map>Adding to obtain a feature map- >Feature map +.>Inputting into the sigmoid function of the second context attention module, outputting the obtained feature map +.>Feature map +.>Inputting the first convolution layer to obtain a characteristic diagram +.>
f-5) Conv-block1 is composed of a convolution layer, a BatchNorm layer and a ReLU activation function in sequence, and is characterized in thatInput into Conv-block1, and output to obtain characteristic diagram +.>Feature map +.>Input into the first upsampling layer, output to get the feature map +.>The first attention gating module consists of a first convolution layer, a first BN layer, a second convolution layer, a second BN layer, a ReLU activation function, a third convolution layer, a third BN layer and a sigmoid function, and is used for decoding a characteristic diagram->Sequentially inputting into a first convolution layer and a first BN layer of a first attention gate control module, and outputting to obtain a characteristic diagram ++>Feature map +.>Sequentially inputting into a second convolution layer and a second BN layer of the first attention gating module, and outputting to obtain a characteristic diagram ++>Feature map +.>And feature mapAdding to obtain a feature map->Feature map +.>Sequentially inputting into a ReLU activation function, a third convolution layer, a third BN layer and a sigmoid function of the first attention gating module, and outputting to obtain a feature map +.>Feature map +.>And (4) feature map>Multiplication to obtain a feature map->Feature map +. >And (4) feature map>Cascading get feature map->First contextual attention modelThe block is composed of a first convolution layer, a second convolution layer, a global maximum pooling layer, a third convolution layer, a fourth convolution layer and a sigmoid function, and the feature map is->Sequentially inputting into a first convolution layer and a second convolution layer of a first context attention module, and outputting to obtain a characteristic diagram ++>Feature map +.>Input into the global maximum pooling layer of the first contextual attention module, output the resulting feature map +.>Feature map +.>Sequentially inputting into a third convolution layer and a fourth convolution layer of the first context attention module, and outputting to obtain a characteristic diagram ++>Feature map +.>And (4) feature map>Multiplication to obtain a feature mapFeature map +.>And (4) feature map>Adding to obtain a feature map->Feature map +.>Inputting into the sigmoid function of the first context attention module, outputting the obtained feature map +.>Feature map +.>Inputting into the first convolution layer to obtain a feature map +.>
f-6) mapping the featuresFeature map->Feature map->Feature map->Residual addition is carried out to obtain a characteristic diagram->
8. The contextual cascade attention-based cardiac MRI segmentation method according to claim 7, wherein: the convolution kernel size of the convolution layer of Conv-block4 in step f-2) is 3×3, and padding is 1; the convolution kernel sizes of the first convolution layer, the second convolution layer and the third convolution layer of the fourth attention gating module are all 1 multiplied by 1; the convolution kernel sizes of the first convolution layer, the second convolution layer and the third convolution layer of the fourth context attention module are 3×3, the padding is 1, and the convolution kernel size of the fourth convolution layer of the fourth context attention module is 1×1; the convolution kernel size of the fourth convolution layer of the decoder is 1×1; the convolution kernel size of the convolution layer of Conv-block3 in step f-3) is 3×3, and padding is 1; the convolution kernel sizes of the first convolution layer, the second convolution layer and the third convolution layer of the third attention gating module are all 1 multiplied by 1; the convolution kernel size of the third upsampling layer is 2×2; the convolution kernel sizes of the first convolution layer, the second convolution layer and the third convolution layer of the third context attention module are 3×3, the padding is 1, and the convolution kernel size of the fourth convolution layer of the third context attention module is 1×1; the convolution kernel size of the third convolution layer of the decoder is 1×1; f-4) the convolution kernel size of the convolution layer of Conv-block2 is 3×3, and padding is 1; the convolution kernel sizes of the first convolution layer, the second convolution layer and the third convolution layer of the second attention gating module are all 1 multiplied by 1; the convolution kernel size of the second upsampling layer is 2 x 2; the convolution kernel sizes of the first convolution layer, the second convolution layer and the third convolution layer of the second context attention module are all 3×3, the padding is 1, and the convolution kernel size of the fourth convolution layer of the second context attention module is 1×1; the convolution kernel size of the second convolution layer of the decoder is 1×1; f-5) the convolution kernel size of the convolution layer of Conv-block1 is 3×3, and padding is 1; the convolution kernel sizes of the first convolution layer, the second convolution layer and the third convolution layer of the first attention gating module are all 1 multiplied by 1; the convolution kernel size of the first upsampling layer is 2 x 2; the convolution kernel sizes of the first convolution layer, the second convolution layer and the third convolution layer of the first context attention module are all 3×3, the padding is 1, and the convolution kernel size of the fourth convolution layer of the first context attention module is 1×1; the convolution kernel size of the first convolution layer of the decoder is 1 x 1.
9. The contextual cascade attention-based cardiac MRI segmentation method according to claim 1, wherein: and g) summing the Dice loss and the cross entropy loss to obtain total loss, training the segmentation network model by using the Adam optimizer to obtain an optimized segmentation network model, wherein the batch size during training is set to be 32, the iteration period is set to be 200, and the learning rate is set to be 0.001.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202311231297.8A CN117078941B (en) | 2023-09-22 | 2023-09-22 | Cardiac MRI segmentation method based on context cascade attention |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202311231297.8A CN117078941B (en) | 2023-09-22 | 2023-09-22 | Cardiac MRI segmentation method based on context cascade attention |
Publications (2)
Publication Number | Publication Date |
---|---|
CN117078941A CN117078941A (en) | 2023-11-17 |
CN117078941B true CN117078941B (en) | 2024-03-01 |
Family
ID=88719650
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN202311231297.8A Active CN117078941B (en) | 2023-09-22 | 2023-09-22 | Cardiac MRI segmentation method based on context cascade attention |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN117078941B (en) |
Families Citing this family (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN117649523B (en) * | 2023-11-28 | 2024-07-09 | 齐鲁工业大学(山东省科学院) | Variable heart MRI segmentation method based on LUnetr model |
CN117593274B (en) * | 2023-11-30 | 2024-06-04 | 齐鲁工业大学(山东省科学院) | Cardiac MRI segmentation method based on shared channel attention mechanism |
Citations (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
WO2021104056A1 (en) * | 2019-11-27 | 2021-06-03 | 中国科学院深圳先进技术研究院 | Automatic tumor segmentation system and method, and electronic device |
CN116563265A (en) * | 2023-05-23 | 2023-08-08 | 山东省人工智能研究院 | Cardiac MRI (magnetic resonance imaging) segmentation method based on multi-scale attention and self-adaptive feature fusion |
CN116612131A (en) * | 2023-05-22 | 2023-08-18 | 山东省人工智能研究院 | Cardiac MRI structure segmentation method based on ADC-UNet model |
-
2023
- 2023-09-22 CN CN202311231297.8A patent/CN117078941B/en active Active
Patent Citations (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
WO2021104056A1 (en) * | 2019-11-27 | 2021-06-03 | 中国科学院深圳先进技术研究院 | Automatic tumor segmentation system and method, and electronic device |
CN116612131A (en) * | 2023-05-22 | 2023-08-18 | 山东省人工智能研究院 | Cardiac MRI structure segmentation method based on ADC-UNet model |
CN116563265A (en) * | 2023-05-23 | 2023-08-08 | 山东省人工智能研究院 | Cardiac MRI (magnetic resonance imaging) segmentation method based on multi-scale attention and self-adaptive feature fusion |
Also Published As
Publication number | Publication date |
---|---|
CN117078941A (en) | 2023-11-17 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN113077471B (en) | Medical image segmentation method based on U-shaped network | |
CN111145170B (en) | Medical image segmentation method based on deep learning | |
CN117078941B (en) | Cardiac MRI segmentation method based on context cascade attention | |
CN111242288B (en) | Multi-scale parallel deep neural network model construction method for lesion image segmentation | |
CN109949309A (en) | A kind of CT image for liver dividing method based on deep learning | |
WO2024104035A1 (en) | Long short-term memory self-attention model-based three-dimensional medical image segmentation method and system | |
CN114119515B (en) | Brain tumor detection method based on attention mechanism and MRI multi-mode fusion | |
CN112712528B (en) | Intestinal tract focus segmentation method combining multi-scale U-shaped residual error encoder and integral reverse attention mechanism | |
Ma et al. | ATFE-Net: Axial Transformer and Feature Enhancement-based CNN for ultrasound breast mass segmentation | |
CN113192076B (en) | MRI brain tumor image segmentation method combining classification prediction and multi-scale feature extraction | |
CN111161271A (en) | Ultrasonic image segmentation method | |
CN115526829A (en) | Honeycomb lung focus segmentation method and network based on ViT and context feature fusion | |
CN114596318A (en) | Breast cancer magnetic resonance imaging focus segmentation method based on Transformer | |
CN111626296B (en) | Medical image segmentation system and method based on deep neural network and terminal | |
CN116228785A (en) | Pneumonia CT image segmentation method based on improved Unet network | |
CN117333497A (en) | Mask supervision strategy-based three-dimensional medical image segmentation method for efficient modeling | |
CN116823850A (en) | Cardiac MRI segmentation method and system based on U-Net and transducer fusion improvement | |
CN118279319A (en) | Medical image segmentation method based on global attention and multi-scale features | |
CN114565601A (en) | Improved liver CT image segmentation algorithm based on DeepLabV3+ | |
CN117351030A (en) | Medical image segmentation method based on Swin transducer and CNN parallel network | |
CN118037791A (en) | Construction method and application of multi-mode three-dimensional medical image segmentation registration model | |
CN113487560A (en) | Brain tumor segmentation method and device based on spatial feature attention mechanism | |
CN113313699A (en) | X-ray chest disease classification and positioning method based on weak supervised learning and electronic equipment | |
CN117853861A (en) | Training method of crop disease recognition model and crop disease recognition method | |
CN115330600A (en) | Lung CT image super-resolution method based on improved SRGAN |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
GR01 | Patent grant | ||
GR01 | Patent grant |