CN113744210A - Heart segmentation method based on multi-scale attention U-net network - Google Patents

Heart segmentation method based on multi-scale attention U-net network Download PDF

Info

Publication number
CN113744210A
CN113744210A CN202110964471.4A CN202110964471A CN113744210A CN 113744210 A CN113744210 A CN 113744210A CN 202110964471 A CN202110964471 A CN 202110964471A CN 113744210 A CN113744210 A CN 113744210A
Authority
CN
China
Prior art keywords
training
data
segmentation
heart
network
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Withdrawn
Application number
CN202110964471.4A
Other languages
Chinese (zh)
Inventor
崔晓娟
白鑫昊
杨铁军
李磊
樊超
巩跃洪
苗建雨
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Henan University of Technology
Original Assignee
Henan University of Technology
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Henan University of Technology filed Critical Henan University of Technology
Priority to CN202110964471.4A priority Critical patent/CN113744210A/en
Publication of CN113744210A publication Critical patent/CN113744210A/en
Withdrawn legal-status Critical Current

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/0002Inspection of images, e.g. flaw detection
    • G06T7/0012Biomedical image inspection
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/045Combinations of networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/08Learning methods
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T3/00Geometric image transformations in the plane of the image
    • G06T3/40Scaling of whole images or parts thereof, e.g. expanding or contracting
    • G06T3/4046Scaling of whole images or parts thereof, e.g. expanding or contracting using neural networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T3/00Geometric image transformations in the plane of the image
    • G06T3/60Rotation of whole images or parts thereof
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T5/00Image enhancement or restoration
    • G06T5/90Dynamic range modification of images or parts thereof
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/10Segmentation; Edge detection
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/10Image acquisition modality
    • G06T2207/10072Tomographic images
    • G06T2207/10081Computed x-ray tomography [CT]
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/20Special algorithmic details
    • G06T2207/20081Training; Learning
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/20Special algorithmic details
    • G06T2207/20084Artificial neural networks [ANN]
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/20Special algorithmic details
    • G06T2207/20112Image segmentation details
    • G06T2207/20132Image cropping
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/30Subject of image; Context of image processing
    • G06T2207/30004Biomedical image processing
    • G06T2207/30048Heart; Cardiac

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Evolutionary Computation (AREA)
  • Health & Medical Sciences (AREA)
  • Artificial Intelligence (AREA)
  • General Health & Medical Sciences (AREA)
  • Computing Systems (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Computational Linguistics (AREA)
  • Molecular Biology (AREA)
  • Biophysics (AREA)
  • General Engineering & Computer Science (AREA)
  • Biomedical Technology (AREA)
  • Mathematical Physics (AREA)
  • Software Systems (AREA)
  • Data Mining & Analysis (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Medical Informatics (AREA)
  • Nuclear Medicine, Radiotherapy & Molecular Imaging (AREA)
  • Radiology & Medical Imaging (AREA)
  • Quality & Reliability (AREA)
  • Image Analysis (AREA)
  • Apparatus For Radiation Diagnosis (AREA)

Abstract

Aiming at the problem that the accuracy of a classical U-shaped segmentation network in heart substructure segmentation is not high, the invention provides a multi-scale attention U-Net network for heart segmentation. On the basis of a classic U-Net structure, an algorithm firstly cuts a CT heart image for preprocessing, and normalization processing is carried out by reducing input of background pixels and adopting a Z-score standardization method to eliminate difference of pixel gray distribution ranges; then, an attention mechanism is introduced to fully utilize shallow information. Adding a space attention mechanism in a backbone network, and adding a channel attention mechanism in a jump connection, so that the network can fully utilize a shallow convolution layer to extract information, retain useful information and remove redundant information; and simultaneously introducing the inclusion modules of convolution kernels with different scales, and simultaneously extracting and fusing feature information with different scales to realize accurate segmentation. The invention has good application prospect in the automatic heart segmentation.

Description

Heart segmentation method based on multi-scale attention U-net network
Technical Field
The invention belongs to the technical field of medical image processing, and particularly relates to a method for cardiac substructure segmentation.
Background
According to the heart disease and stroke statistical report of the American Heart Association (AHA) in 2019, it is pointed out that about 1055000 cases are expected to develop coronary heart disease in 2019 in the united states, including 720000 new and 335000 recurrent coronary artery cases, and this figure is also increasing year by year. At present, cardiovascular diseases are one of the main causes of human non-accidental death, the morbidity is high, and no fixed morbidity rule exists. Now, cardiovascular diseases gradually develop to young people and threaten the healthy life of human beings all the time. Accurate calculation, modeling and analysis of the entire cardiac structure is critical for research and application in the medical field for effective treatment and prevention of these diseases. At present, the cardiac substructure is generally segmented manually by doctors or experts according to the existing medical knowledge, medical conditions and clinical knowledge, the method is time-consuming and labor-consuming and has strong subjectivity, and the segmentation result is different from person to person. With the advent of large-scale labeled data and the development of computers, the realization of automatic segmentation of the heart using deep learning algorithms has become a hot spot of current research. Among them, Olaf Ronneberger proposed a U-Net model for medical image segmentation based on the Full Convolutional Network (FCN) model in 2015. Both U-Net and FCN have classical encoding-decoding topologies, but U-Net has a symmetric network structure and hopping connections, and the results of U-Net are superior to FCN in segmentation of cardiac images. Aiming at the problem that the heart image segmentation precision is difficult to improve, the improvement research performed by researchers on the basis of U-Net can be roughly divided into two types: improvement study based on 2D U-Net framework, improvement study based on 3D U-Net framework. Although the 2D network partitioning is less computationally and storage demanding, the 2D network partitioning typically discards spatial information between slices. 3D network segmentation can maximize the information between slice sequences. In summary, aiming at improving the segmentation accuracy, the invention provides a 3D U heart segmentation network AU-Net with attention mechanism, which is mainly used for solving the problem of low segmentation accuracy of the segmented heart substructure.
Disclosure of Invention
The invention aims to solve the problem of low precision of heart substructure segmentation, and provides an automatic method for accurately segmenting the heart substructure.
The invention is realized by the following technical scheme: a heart segmentation method based on a multi-scale attention U-net network. Firstly, cutting and scaling preprocessing are carried out on an image so as to reduce training parameters and cover global information; then 30000 times of training is carried out on the AMU-Net network of the invention, and the trained parameters are stored; and finally, segmenting the test data set by using the trained weight to finally obtain a segmentation result graph.
(1) Data preprocessing: image preprocessing first re-encodes the label data of the CT images of 10 training volume data in the MM-WHS2017 dataset to make them suitable for multi-classification tasks. The data is then randomly cropped, the image is cropped to 256 x 16 size, and light data enhancement techniques are applied to the data and label. On a random basis, the data is rotated between-15 and +15 degrees and scaled in between by 0.9-1.1. This ensures slight robustness and variability of network training.
(2) A training stage: in the training stage, the AMU-Net network provided by the invention is subjected to parameter training. And (3) learning model parameters of the training set by using a Tensorflow deep learning framework, setting the batch processing size to be 4 and setting the training iteration number to be 30000 by adopting an Adam optimizer. Selecting 50% of data in the training set as a verification set, training the data by using a cross entropy loss function, and selecting a weight coefficient when the verification set is minimally lost as a final training weight in a training stage in an experiment to provide a weight parameter for a subsequent testing stage.
The main innovation of the AU-Net network provided by the invention is as follows:
a. introducing a multi-scale inclusion module: by introducing the inclusion modules of convolution kernels with different scales, as shown in fig. 1, feature information with different scales is extracted and fused, global information and local information are better combined, and more features are obtained.
b. An attention mechanism is introduced: by introducing an attention mechanism, non-information areas are filtered, areas containing useful information are emphasized, the resolution of characteristic images is enhanced, the expression capacity of detail characteristics is improved, and more detail characteristics can be obtained.
(3) And (3) a testing stage: randomly selecting 10% of training volume data in the MM-WHS2017 data set as a test set. Firstly, a CT image with the size of 256 multiplied by 16 is directly input into a testing stage, and the final weight parameters obtained in a training stage are used for carrying out heart substructure segmentation on the testing image, so that a segmented heart substructure result graph is finally obtained.
The invention provides an improved AMU-Net heart image segmentation algorithm aiming at the defects that the heart image segmentation precision of a U-shaped convolution network is not high and the boundary of each region is fuzzy. By introducing a multi-scale inclusion module, feature information of different scales is extracted and fused. And an attention mechanism is introduced, the reusability of the features is increased, the shallow features are fused with the corresponding high-level features, and finally a trainable end-to-end segmentation algorithm is formed. Compared with the classic U-Net segmentation algorithm, the algorithm has a finer structure, can effectively solve the problems of over-segmentation and under-segmentation of heart segmentation, and has higher accuracy of segmentation results.
Drawings
Fig. 1 is a schematic diagram of a multi-scale inclusion structure.
Fig. 2 is a flow chart of a network training method for cardiac segmentation.
Detailed Description
To verify the cardiac substructure segmentation performance of the present invention, we selected the MM-WHS2017 dataset for training and testing.
The method comprises the steps of preprocessing CT image data, using Spyder software, and performing image normalization processing by image rotation, translation transformation and contrast enhancement.
Step two, training the AMU-Net network in Spyder software, wherein the batch _ size is 4, the learning _ rate is 0.001, adopting an Adama optimizer, using L2 regularization to prevent overfitting, and setting the regularization coefficient to be 0.0005. 60000 epochs are trained, the training set and the verification set are divided by the ratio of 1:1, the two processes are discontinuously trained, network parameters are adjusted until the network converges, and the training is finished.
And step three, testing the AMU-Net network by adopting a test set of the MM-WHS2017 data set. To evaluate the segmentation results, 2 common evaluation criteria, similarity coefficient (Dice) and Jaccard index, were used, as in table 1.
TABLE 1 comparison with other people's work (Dice index)
Figure BDA0003223518220000031
Experimental results show that the algorithm structure of the invention is finer, the problems of over-segmentation and under-segmentation of heart segmentation can be effectively solved, the segmentation result precision is higher, and the completeness and the accuracy of the heart segmentation can be ensured.

Claims (1)

1. The invention discloses a heart segmentation method based on a multi-scale attention U-net network, which comprises the following steps of:
(1) data preprocessing: firstly, carrying out label data recoding on CT images of 10 training volume data in an MM-WHS2017 data set to enable the CT images to be suitable for a multi-classification task; then randomly cutting the data, cutting the image into 256 multiplied by 16 size, and applying optical data enhancement technology on the data and the label; on a random basis, the data is rotated between-15 and +15 degrees and scaled in the range of 0.9-1.1 therebetween; this ensures slight robustness and variability of network training;
(2) a training stage: in the training stage, parameter training is carried out on the AMU-Net network provided by the invention; learning model parameters of a training set by using a Tensorflow deep learning framework, setting the batch processing size to be 4 and the training iteration number to be 30000 by using an Adam optimizer; selecting 50% of data in a training set as a verification set, training the data by using a cross entropy loss function, and selecting a weight coefficient when the verification set is minimally lost as a final training weight in a training stage in an experiment to provide a weight parameter for a subsequent testing stage;
(3) and (3) a testing stage: randomly selecting 10% of training volume data in the MM-WHS2017 data set as a test set; firstly, a CT image with the size of 256 multiplied by 16 is directly input into a testing stage, and the final weight parameters obtained in a training stage are used for carrying out heart substructure segmentation on the testing image, so that a segmented heart substructure result graph is finally obtained.
CN202110964471.4A 2021-08-22 2021-08-22 Heart segmentation method based on multi-scale attention U-net network Withdrawn CN113744210A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202110964471.4A CN113744210A (en) 2021-08-22 2021-08-22 Heart segmentation method based on multi-scale attention U-net network

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202110964471.4A CN113744210A (en) 2021-08-22 2021-08-22 Heart segmentation method based on multi-scale attention U-net network

Publications (1)

Publication Number Publication Date
CN113744210A true CN113744210A (en) 2021-12-03

Family

ID=78732134

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202110964471.4A Withdrawn CN113744210A (en) 2021-08-22 2021-08-22 Heart segmentation method based on multi-scale attention U-net network

Country Status (1)

Country Link
CN (1) CN113744210A (en)

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN114066913A (en) * 2022-01-12 2022-02-18 广东工业大学 Heart image segmentation method and system

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN114066913A (en) * 2022-01-12 2022-02-18 广东工业大学 Heart image segmentation method and system
CN114066913B (en) * 2022-01-12 2022-04-22 广东工业大学 Heart image segmentation method and system

Similar Documents

Publication Publication Date Title
CN107154043B (en) Pulmonary nodule false positive sample inhibition method based on 3DCNN
Tang et al. A multi-stage framework with context information fusion structure for skin lesion segmentation
CN111882560B (en) Lung parenchyma CT image segmentation method based on weighted full convolution neural network
CN111612008A (en) Image segmentation method based on convolution network
CN115170582A (en) Liver image segmentation method based on multi-scale feature fusion and grid attention mechanism
CN113344933B (en) Glandular cell segmentation method based on multi-level feature fusion network
CN111784713A (en) Attention mechanism-introduced U-shaped heart segmentation method
Zhang et al. A novel denoising method for CT images based on U-net and multi-attention
CN113192076B (en) MRI brain tumor image segmentation method combining classification prediction and multi-scale feature extraction
Zhang et al. A novel denoising method for low-dose CT images based on transformer and CNN
CN114693935A (en) Medical image segmentation method based on automatic data augmentation
Wang et al. EAR-U-Net: EfficientNet and attention-based residual U-Net for automatic liver segmentation in CT
CN113744209A (en) Heart segmentation method based on multi-scale residual U-net network
CN114092450A (en) Real-time image segmentation method, system and device based on gastroscopy video
Priyadharshini et al. A Logical Design of Robust Methodology to Detect and Classify Melanoma Disease using Hybrid Deep Learning Principles
Mai et al. Few-shot transfer learning for hereditary retinal diseases recognition
CN115471512A (en) Medical image segmentation method based on self-supervision contrast learning
CN118430790A (en) Mammary tumor BI-RADS grading method based on multi-modal-diagram neural network
CN114066883A (en) Liver tumor segmentation method based on feature selection and residual fusion
CN113744210A (en) Heart segmentation method based on multi-scale attention U-net network
CN113822858A (en) Heart image segmentation method based on residual attention unit
CN116894943A (en) Double-constraint camouflage target detection method and system
CN114373080B (en) Hyperspectral classification method of lightweight hybrid convolution model based on global reasoning
CN111798455A (en) Thyroid nodule real-time segmentation method based on full convolution dense cavity network
Zhou et al. Subcellular structure segmentation from cryo-electron tomograms via machine learning

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
WW01 Invention patent application withdrawn after publication
WW01 Invention patent application withdrawn after publication

Application publication date: 20211203