CN113744210A - Heart segmentation method based on multi-scale attention U-net network - Google Patents
Heart segmentation method based on multi-scale attention U-net network Download PDFInfo
- Publication number
- CN113744210A CN113744210A CN202110964471.4A CN202110964471A CN113744210A CN 113744210 A CN113744210 A CN 113744210A CN 202110964471 A CN202110964471 A CN 202110964471A CN 113744210 A CN113744210 A CN 113744210A
- Authority
- CN
- China
- Prior art keywords
- training
- data
- segmentation
- heart
- network
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Withdrawn
Links
- 230000011218 segmentation Effects 0.000 title claims abstract description 31
- 238000000034 method Methods 0.000 title claims description 11
- 238000007781 pre-processing Methods 0.000 claims abstract description 6
- 238000012545 processing Methods 0.000 claims abstract description 5
- 238000012549 training Methods 0.000 claims description 33
- 238000012360 testing method Methods 0.000 claims description 14
- 238000012795 verification Methods 0.000 claims description 5
- 238000013135 deep learning Methods 0.000 claims description 3
- ORILYTVJVMAKLC-UHFFFAOYSA-N Adamantane Natural products C1C(C2)CC3CC1CC2C3 ORILYTVJVMAKLC-UHFFFAOYSA-N 0.000 claims description 2
- 238000002474 experimental method Methods 0.000 claims description 2
- 230000006870 function Effects 0.000 claims description 2
- 230000003287 optical effect Effects 0.000 claims 1
- 230000007246 mechanism Effects 0.000 abstract description 7
- 238000010606 normalization Methods 0.000 abstract description 2
- 238000011425 standardization method Methods 0.000 abstract 1
- 230000000747 cardiac effect Effects 0.000 description 6
- 238000003709 image segmentation Methods 0.000 description 4
- 230000006872 improvement Effects 0.000 description 3
- 238000011160 research Methods 0.000 description 3
- 208000024172 Cardiovascular disease Diseases 0.000 description 2
- 241000282414 Homo sapiens Species 0.000 description 2
- 238000000638 solvent extraction Methods 0.000 description 2
- 206010063746 Accidental death Diseases 0.000 description 1
- 241001436679 Adama Species 0.000 description 1
- 238000004458 analytical method Methods 0.000 description 1
- 238000004364 calculation method Methods 0.000 description 1
- 208000029078 coronary artery disease Diseases 0.000 description 1
- 210000004351 coronary vessel Anatomy 0.000 description 1
- 230000007547 defect Effects 0.000 description 1
- 238000011161 development Methods 0.000 description 1
- 230000018109 developmental process Effects 0.000 description 1
- 238000010586 diagram Methods 0.000 description 1
- 201000010099 disease Diseases 0.000 description 1
- 208000037265 diseases, disorders, signs and symptoms Diseases 0.000 description 1
- 238000011156 evaluation Methods 0.000 description 1
- 208000019622 heart disease Diseases 0.000 description 1
- 230000002265 prevention Effects 0.000 description 1
- 230000008569 process Effects 0.000 description 1
- 230000000306 recurrent effect Effects 0.000 description 1
- 230000009466 transformation Effects 0.000 description 1
- 238000013519 translation Methods 0.000 description 1
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
- G06T7/0002—Inspection of images, e.g. flaw detection
- G06T7/0012—Biomedical image inspection
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/04—Architecture, e.g. interconnection topology
- G06N3/045—Combinations of networks
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/08—Learning methods
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T3/00—Geometric image transformations in the plane of the image
- G06T3/40—Scaling of whole images or parts thereof, e.g. expanding or contracting
- G06T3/4046—Scaling of whole images or parts thereof, e.g. expanding or contracting using neural networks
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T3/00—Geometric image transformations in the plane of the image
- G06T3/60—Rotation of whole images or parts thereof
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T5/00—Image enhancement or restoration
- G06T5/90—Dynamic range modification of images or parts thereof
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
- G06T7/10—Segmentation; Edge detection
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/10—Image acquisition modality
- G06T2207/10072—Tomographic images
- G06T2207/10081—Computed x-ray tomography [CT]
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/20—Special algorithmic details
- G06T2207/20081—Training; Learning
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/20—Special algorithmic details
- G06T2207/20084—Artificial neural networks [ANN]
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/20—Special algorithmic details
- G06T2207/20112—Image segmentation details
- G06T2207/20132—Image cropping
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/30—Subject of image; Context of image processing
- G06T2207/30004—Biomedical image processing
- G06T2207/30048—Heart; Cardiac
Landscapes
- Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- Physics & Mathematics (AREA)
- General Physics & Mathematics (AREA)
- Evolutionary Computation (AREA)
- Health & Medical Sciences (AREA)
- Artificial Intelligence (AREA)
- General Health & Medical Sciences (AREA)
- Computing Systems (AREA)
- Life Sciences & Earth Sciences (AREA)
- Computational Linguistics (AREA)
- Molecular Biology (AREA)
- Biophysics (AREA)
- General Engineering & Computer Science (AREA)
- Biomedical Technology (AREA)
- Mathematical Physics (AREA)
- Software Systems (AREA)
- Data Mining & Analysis (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Medical Informatics (AREA)
- Nuclear Medicine, Radiotherapy & Molecular Imaging (AREA)
- Radiology & Medical Imaging (AREA)
- Quality & Reliability (AREA)
- Image Analysis (AREA)
- Apparatus For Radiation Diagnosis (AREA)
Abstract
Aiming at the problem that the accuracy of a classical U-shaped segmentation network in heart substructure segmentation is not high, the invention provides a multi-scale attention U-Net network for heart segmentation. On the basis of a classic U-Net structure, an algorithm firstly cuts a CT heart image for preprocessing, and normalization processing is carried out by reducing input of background pixels and adopting a Z-score standardization method to eliminate difference of pixel gray distribution ranges; then, an attention mechanism is introduced to fully utilize shallow information. Adding a space attention mechanism in a backbone network, and adding a channel attention mechanism in a jump connection, so that the network can fully utilize a shallow convolution layer to extract information, retain useful information and remove redundant information; and simultaneously introducing the inclusion modules of convolution kernels with different scales, and simultaneously extracting and fusing feature information with different scales to realize accurate segmentation. The invention has good application prospect in the automatic heart segmentation.
Description
Technical Field
The invention belongs to the technical field of medical image processing, and particularly relates to a method for cardiac substructure segmentation.
Background
According to the heart disease and stroke statistical report of the American Heart Association (AHA) in 2019, it is pointed out that about 1055000 cases are expected to develop coronary heart disease in 2019 in the united states, including 720000 new and 335000 recurrent coronary artery cases, and this figure is also increasing year by year. At present, cardiovascular diseases are one of the main causes of human non-accidental death, the morbidity is high, and no fixed morbidity rule exists. Now, cardiovascular diseases gradually develop to young people and threaten the healthy life of human beings all the time. Accurate calculation, modeling and analysis of the entire cardiac structure is critical for research and application in the medical field for effective treatment and prevention of these diseases. At present, the cardiac substructure is generally segmented manually by doctors or experts according to the existing medical knowledge, medical conditions and clinical knowledge, the method is time-consuming and labor-consuming and has strong subjectivity, and the segmentation result is different from person to person. With the advent of large-scale labeled data and the development of computers, the realization of automatic segmentation of the heart using deep learning algorithms has become a hot spot of current research. Among them, Olaf Ronneberger proposed a U-Net model for medical image segmentation based on the Full Convolutional Network (FCN) model in 2015. Both U-Net and FCN have classical encoding-decoding topologies, but U-Net has a symmetric network structure and hopping connections, and the results of U-Net are superior to FCN in segmentation of cardiac images. Aiming at the problem that the heart image segmentation precision is difficult to improve, the improvement research performed by researchers on the basis of U-Net can be roughly divided into two types: improvement study based on 2D U-Net framework, improvement study based on 3D U-Net framework. Although the 2D network partitioning is less computationally and storage demanding, the 2D network partitioning typically discards spatial information between slices. 3D network segmentation can maximize the information between slice sequences. In summary, aiming at improving the segmentation accuracy, the invention provides a 3D U heart segmentation network AU-Net with attention mechanism, which is mainly used for solving the problem of low segmentation accuracy of the segmented heart substructure.
Disclosure of Invention
The invention aims to solve the problem of low precision of heart substructure segmentation, and provides an automatic method for accurately segmenting the heart substructure.
The invention is realized by the following technical scheme: a heart segmentation method based on a multi-scale attention U-net network. Firstly, cutting and scaling preprocessing are carried out on an image so as to reduce training parameters and cover global information; then 30000 times of training is carried out on the AMU-Net network of the invention, and the trained parameters are stored; and finally, segmenting the test data set by using the trained weight to finally obtain a segmentation result graph.
(1) Data preprocessing: image preprocessing first re-encodes the label data of the CT images of 10 training volume data in the MM-WHS2017 dataset to make them suitable for multi-classification tasks. The data is then randomly cropped, the image is cropped to 256 x 16 size, and light data enhancement techniques are applied to the data and label. On a random basis, the data is rotated between-15 and +15 degrees and scaled in between by 0.9-1.1. This ensures slight robustness and variability of network training.
(2) A training stage: in the training stage, the AMU-Net network provided by the invention is subjected to parameter training. And (3) learning model parameters of the training set by using a Tensorflow deep learning framework, setting the batch processing size to be 4 and setting the training iteration number to be 30000 by adopting an Adam optimizer. Selecting 50% of data in the training set as a verification set, training the data by using a cross entropy loss function, and selecting a weight coefficient when the verification set is minimally lost as a final training weight in a training stage in an experiment to provide a weight parameter for a subsequent testing stage.
The main innovation of the AU-Net network provided by the invention is as follows:
a. introducing a multi-scale inclusion module: by introducing the inclusion modules of convolution kernels with different scales, as shown in fig. 1, feature information with different scales is extracted and fused, global information and local information are better combined, and more features are obtained.
b. An attention mechanism is introduced: by introducing an attention mechanism, non-information areas are filtered, areas containing useful information are emphasized, the resolution of characteristic images is enhanced, the expression capacity of detail characteristics is improved, and more detail characteristics can be obtained.
(3) And (3) a testing stage: randomly selecting 10% of training volume data in the MM-WHS2017 data set as a test set. Firstly, a CT image with the size of 256 multiplied by 16 is directly input into a testing stage, and the final weight parameters obtained in a training stage are used for carrying out heart substructure segmentation on the testing image, so that a segmented heart substructure result graph is finally obtained.
The invention provides an improved AMU-Net heart image segmentation algorithm aiming at the defects that the heart image segmentation precision of a U-shaped convolution network is not high and the boundary of each region is fuzzy. By introducing a multi-scale inclusion module, feature information of different scales is extracted and fused. And an attention mechanism is introduced, the reusability of the features is increased, the shallow features are fused with the corresponding high-level features, and finally a trainable end-to-end segmentation algorithm is formed. Compared with the classic U-Net segmentation algorithm, the algorithm has a finer structure, can effectively solve the problems of over-segmentation and under-segmentation of heart segmentation, and has higher accuracy of segmentation results.
Drawings
Fig. 1 is a schematic diagram of a multi-scale inclusion structure.
Fig. 2 is a flow chart of a network training method for cardiac segmentation.
Detailed Description
To verify the cardiac substructure segmentation performance of the present invention, we selected the MM-WHS2017 dataset for training and testing.
The method comprises the steps of preprocessing CT image data, using Spyder software, and performing image normalization processing by image rotation, translation transformation and contrast enhancement.
Step two, training the AMU-Net network in Spyder software, wherein the batch _ size is 4, the learning _ rate is 0.001, adopting an Adama optimizer, using L2 regularization to prevent overfitting, and setting the regularization coefficient to be 0.0005. 60000 epochs are trained, the training set and the verification set are divided by the ratio of 1:1, the two processes are discontinuously trained, network parameters are adjusted until the network converges, and the training is finished.
And step three, testing the AMU-Net network by adopting a test set of the MM-WHS2017 data set. To evaluate the segmentation results, 2 common evaluation criteria, similarity coefficient (Dice) and Jaccard index, were used, as in table 1.
TABLE 1 comparison with other people's work (Dice index)
Experimental results show that the algorithm structure of the invention is finer, the problems of over-segmentation and under-segmentation of heart segmentation can be effectively solved, the segmentation result precision is higher, and the completeness and the accuracy of the heart segmentation can be ensured.
Claims (1)
1. The invention discloses a heart segmentation method based on a multi-scale attention U-net network, which comprises the following steps of:
(1) data preprocessing: firstly, carrying out label data recoding on CT images of 10 training volume data in an MM-WHS2017 data set to enable the CT images to be suitable for a multi-classification task; then randomly cutting the data, cutting the image into 256 multiplied by 16 size, and applying optical data enhancement technology on the data and the label; on a random basis, the data is rotated between-15 and +15 degrees and scaled in the range of 0.9-1.1 therebetween; this ensures slight robustness and variability of network training;
(2) a training stage: in the training stage, parameter training is carried out on the AMU-Net network provided by the invention; learning model parameters of a training set by using a Tensorflow deep learning framework, setting the batch processing size to be 4 and the training iteration number to be 30000 by using an Adam optimizer; selecting 50% of data in a training set as a verification set, training the data by using a cross entropy loss function, and selecting a weight coefficient when the verification set is minimally lost as a final training weight in a training stage in an experiment to provide a weight parameter for a subsequent testing stage;
(3) and (3) a testing stage: randomly selecting 10% of training volume data in the MM-WHS2017 data set as a test set; firstly, a CT image with the size of 256 multiplied by 16 is directly input into a testing stage, and the final weight parameters obtained in a training stage are used for carrying out heart substructure segmentation on the testing image, so that a segmented heart substructure result graph is finally obtained.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202110964471.4A CN113744210A (en) | 2021-08-22 | 2021-08-22 | Heart segmentation method based on multi-scale attention U-net network |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202110964471.4A CN113744210A (en) | 2021-08-22 | 2021-08-22 | Heart segmentation method based on multi-scale attention U-net network |
Publications (1)
Publication Number | Publication Date |
---|---|
CN113744210A true CN113744210A (en) | 2021-12-03 |
Family
ID=78732134
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN202110964471.4A Withdrawn CN113744210A (en) | 2021-08-22 | 2021-08-22 | Heart segmentation method based on multi-scale attention U-net network |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN113744210A (en) |
Cited By (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN114066913A (en) * | 2022-01-12 | 2022-02-18 | 广东工业大学 | Heart image segmentation method and system |
-
2021
- 2021-08-22 CN CN202110964471.4A patent/CN113744210A/en not_active Withdrawn
Cited By (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN114066913A (en) * | 2022-01-12 | 2022-02-18 | 广东工业大学 | Heart image segmentation method and system |
CN114066913B (en) * | 2022-01-12 | 2022-04-22 | 广东工业大学 | Heart image segmentation method and system |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN107154043B (en) | Pulmonary nodule false positive sample inhibition method based on 3DCNN | |
Tang et al. | A multi-stage framework with context information fusion structure for skin lesion segmentation | |
CN111882560B (en) | Lung parenchyma CT image segmentation method based on weighted full convolution neural network | |
CN111612008A (en) | Image segmentation method based on convolution network | |
CN115170582A (en) | Liver image segmentation method based on multi-scale feature fusion and grid attention mechanism | |
CN113344933B (en) | Glandular cell segmentation method based on multi-level feature fusion network | |
CN111784713A (en) | Attention mechanism-introduced U-shaped heart segmentation method | |
Zhang et al. | A novel denoising method for CT images based on U-net and multi-attention | |
CN113192076B (en) | MRI brain tumor image segmentation method combining classification prediction and multi-scale feature extraction | |
Zhang et al. | A novel denoising method for low-dose CT images based on transformer and CNN | |
CN114693935A (en) | Medical image segmentation method based on automatic data augmentation | |
Wang et al. | EAR-U-Net: EfficientNet and attention-based residual U-Net for automatic liver segmentation in CT | |
CN113744209A (en) | Heart segmentation method based on multi-scale residual U-net network | |
CN114092450A (en) | Real-time image segmentation method, system and device based on gastroscopy video | |
Priyadharshini et al. | A Logical Design of Robust Methodology to Detect and Classify Melanoma Disease using Hybrid Deep Learning Principles | |
Mai et al. | Few-shot transfer learning for hereditary retinal diseases recognition | |
CN115471512A (en) | Medical image segmentation method based on self-supervision contrast learning | |
CN118430790A (en) | Mammary tumor BI-RADS grading method based on multi-modal-diagram neural network | |
CN114066883A (en) | Liver tumor segmentation method based on feature selection and residual fusion | |
CN113744210A (en) | Heart segmentation method based on multi-scale attention U-net network | |
CN113822858A (en) | Heart image segmentation method based on residual attention unit | |
CN116894943A (en) | Double-constraint camouflage target detection method and system | |
CN114373080B (en) | Hyperspectral classification method of lightweight hybrid convolution model based on global reasoning | |
CN111798455A (en) | Thyroid nodule real-time segmentation method based on full convolution dense cavity network | |
Zhou et al. | Subcellular structure segmentation from cryo-electron tomograms via machine learning |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
WW01 | Invention patent application withdrawn after publication | ||
WW01 | Invention patent application withdrawn after publication |
Application publication date: 20211203 |