CN114359194A - Multi-mode stroke infarct area image processing method based on improved U-Net network - Google Patents
Multi-mode stroke infarct area image processing method based on improved U-Net network Download PDFInfo
- Publication number
- CN114359194A CN114359194A CN202111606856.XA CN202111606856A CN114359194A CN 114359194 A CN114359194 A CN 114359194A CN 202111606856 A CN202111606856 A CN 202111606856A CN 114359194 A CN114359194 A CN 114359194A
- Authority
- CN
- China
- Prior art keywords
- attention
- net
- image
- network
- cerebral blood
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Granted
Links
- 206010061216 Infarction Diseases 0.000 title claims abstract description 41
- 230000007574 infarction Effects 0.000 title claims abstract description 41
- 238000003672 processing method Methods 0.000 title claims abstract description 16
- 230000011218 segmentation Effects 0.000 claims abstract description 35
- 238000002597 diffusion-weighted imaging Methods 0.000 claims abstract description 28
- 210000004556 brain Anatomy 0.000 claims abstract description 26
- 238000004364 calculation method Methods 0.000 claims abstract description 25
- 230000003727 cerebral blood flow Effects 0.000 claims abstract description 18
- 230000002490 cerebral effect Effects 0.000 claims abstract description 15
- 239000008280 blood Substances 0.000 claims abstract description 13
- 210000004369 blood Anatomy 0.000 claims abstract description 13
- 238000004458 analytical method Methods 0.000 claims abstract description 10
- 238000000034 method Methods 0.000 claims abstract description 9
- 238000012549 training Methods 0.000 claims description 25
- 230000006870 function Effects 0.000 claims description 19
- 238000012360 testing method Methods 0.000 claims description 19
- 230000007246 mechanism Effects 0.000 claims description 12
- 238000002372 labelling Methods 0.000 claims description 11
- 238000004422 calculation algorithm Methods 0.000 claims description 6
- 238000011176 pooling Methods 0.000 claims description 6
- 230000004913 activation Effects 0.000 claims description 5
- 230000006872 improvement Effects 0.000 claims description 5
- 238000011156 evaluation Methods 0.000 claims description 4
- 230000008569 process Effects 0.000 claims description 4
- ORILYTVJVMAKLC-UHFFFAOYSA-N Adamantane Natural products C1C(C2)CC3CC1CC2C3 ORILYTVJVMAKLC-UHFFFAOYSA-N 0.000 claims description 3
- 230000009467 reduction Effects 0.000 claims description 3
- 238000012795 verification Methods 0.000 claims 2
- 208000006011 Stroke Diseases 0.000 abstract description 18
- 238000010586 diagram Methods 0.000 abstract description 6
- 206010008190 Cerebrovascular accident Diseases 0.000 abstract description 2
- 238000002591 computed tomography Methods 0.000 description 14
- 230000010412 perfusion Effects 0.000 description 12
- 238000000113 differential scanning calorimetry Methods 0.000 description 10
- 208000032382 Ischaemic stroke Diseases 0.000 description 8
- 230000035945 sensitivity Effects 0.000 description 7
- 238000002474 experimental method Methods 0.000 description 6
- 230000003902 lesion Effects 0.000 description 6
- 230000000052 comparative effect Effects 0.000 description 4
- 238000001514 detection method Methods 0.000 description 4
- 238000005516 engineering process Methods 0.000 description 4
- 238000003384 imaging method Methods 0.000 description 4
- 238000013421 nuclear magnetic resonance imaging Methods 0.000 description 4
- 238000012545 processing Methods 0.000 description 4
- 238000004891 communication Methods 0.000 description 3
- 238000013135 deep learning Methods 0.000 description 3
- 238000003709 image segmentation Methods 0.000 description 3
- 238000002790 cross-validation Methods 0.000 description 2
- 238000011161 development Methods 0.000 description 2
- 230000018109 developmental process Effects 0.000 description 2
- 238000003745 diagnosis Methods 0.000 description 2
- 238000012821 model calculation Methods 0.000 description 2
- 238000002610 neuroimaging Methods 0.000 description 2
- 238000005457 optimization Methods 0.000 description 2
- 238000005070 sampling Methods 0.000 description 2
- 238000005481 NMR spectroscopy Methods 0.000 description 1
- 238000012896 Statistical algorithm Methods 0.000 description 1
- 238000013528 artificial neural network Methods 0.000 description 1
- 230000009286 beneficial effect Effects 0.000 description 1
- 238000003759 clinical diagnosis Methods 0.000 description 1
- 238000013170 computed tomography imaging Methods 0.000 description 1
- 238000010276 construction Methods 0.000 description 1
- 239000002872 contrast media Substances 0.000 description 1
- 230000007547 defect Effects 0.000 description 1
- 238000000586 desensitisation Methods 0.000 description 1
- 239000003814 drug Substances 0.000 description 1
- 230000036541 health Effects 0.000 description 1
- 230000001788 irregular Effects 0.000 description 1
- 238000010801 machine learning Methods 0.000 description 1
- 238000012986 modification Methods 0.000 description 1
- 230000004048 modification Effects 0.000 description 1
- 238000011084 recovery Methods 0.000 description 1
- 238000005728 strengthening Methods 0.000 description 1
- 230000009466 transformation Effects 0.000 description 1
- 238000010200 validation analysis Methods 0.000 description 1
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
- G06T7/0002—Inspection of images, e.g. flaw detection
- G06T7/0012—Biomedical image inspection
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/04—Architecture, e.g. interconnection topology
- G06N3/045—Combinations of networks
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/04—Architecture, e.g. interconnection topology
- G06N3/048—Activation functions
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/08—Learning methods
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
- G06T7/10—Segmentation; Edge detection
- G06T7/11—Region-based segmentation
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
- G06T7/30—Determination of transform parameters for the alignment of images, i.e. image registration
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/10—Image acquisition modality
- G06T2207/10072—Tomographic images
- G06T2207/10088—Magnetic resonance imaging [MRI]
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/20—Special algorithmic details
- G06T2207/20081—Training; Learning
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/20—Special algorithmic details
- G06T2207/20084—Artificial neural networks [ANN]
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/30—Subject of image; Context of image processing
- G06T2207/30004—Biomedical image processing
- G06T2207/30016—Brain
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/30—Subject of image; Context of image processing
- G06T2207/30004—Biomedical image processing
- G06T2207/30101—Blood vessel; Artery; Vein; Vascular
- G06T2207/30104—Vascular flow; Blood flow; Perfusion
Landscapes
- Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- Physics & Mathematics (AREA)
- General Physics & Mathematics (AREA)
- General Health & Medical Sciences (AREA)
- Health & Medical Sciences (AREA)
- Computing Systems (AREA)
- Mathematical Physics (AREA)
- Data Mining & Analysis (AREA)
- Evolutionary Computation (AREA)
- Biophysics (AREA)
- Molecular Biology (AREA)
- Biomedical Technology (AREA)
- General Engineering & Computer Science (AREA)
- Artificial Intelligence (AREA)
- Computational Linguistics (AREA)
- Software Systems (AREA)
- Life Sciences & Earth Sciences (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Medical Informatics (AREA)
- Nuclear Medicine, Radiotherapy & Molecular Imaging (AREA)
- Radiology & Medical Imaging (AREA)
- Quality & Reliability (AREA)
- Image Analysis (AREA)
- Image Processing (AREA)
Abstract
The invention discloses a multi-mode cerebral apoplexy infarct region image processing method based on an improved U-Net network, which comprises the steps of obtaining a brain medical image of a patient and inputting the brain medical image into an upper computer, wherein the brain medical image comprises a cerebral blood flow chart, a cerebral blood volume chart and a diffusion weighted imaging image, and matching 2-mode data of the cerebral blood flow chart and the cerebral blood volume chart of the same patient by using a 3D Slicer as a comprehensive analysis tool library with the diffusion weighted imaging image as a reference to obtain a registered NIFIT format file; using trained U-NetAttentionThe network divides the registered NIFIT format fileAnd cutting to obtain a two-dimensional segmentation result image, and displaying the output result image in the upper computer. According to the method, the attention degree of the region related to the final segmentation result of the infarct region in the bottom layer characteristic diagram is enhanced through calculation of the attention weight of the space and the channel, so that the prediction level of the model is improved.
Description
Technical Field
The invention relates to the technical field of medical image processing, in particular to an image processing method for multi-mode stroke infarction region segmentation based on an improved U-Net network.
Background
The stroke becomes the first threat to the health and life safety of middle-aged and elderly people in China. Perfusion images of Nuclear Magnetic Resonance Imaging (NMRI) and Computed Tomography (CT) are the criteria for a physician to determine the infarcted area of the brain of a patient. With the rapid development of medical image processing technology, digital Imaging and Communications DICOM (digital Imaging and Communications in medicine) is widely used in radiological diagnosis and treatment equipment, and the calculation of each diagnosis index of stroke increasingly depends on digital Imaging and Communications in DICOM medical image. At present, the clinical diagnosis of stroke still depends on manually labeled Nuclear Magnetic Resonance Images (NMRI) or Computed Tomography (CT) perfusion images, the method is time-consuming and easy to delay the state of an illness, and doctors with rich experience are needed, otherwise, the judgment of the state of an illness is easy to be inaccurate.
With the development of deep learning technology in recent years, the deep learning technology has been successful in the fields of target detection, image classification, image segmentation and the like. The application of deep learning techniques in medical image processing is also increasing, and higher accuracy is achieved in many problems than in conventional machine learning and statistical algorithms. Segmentation of an ischemic stroke infarct region is a typical medical image segmentation problem, and can be predicted by using a U-Net network model which is widely used in the field of medical image segmentation. However, the traditional prediction model only applies the brain image of the patient in a single mode, and does not fully mine rich information provided by data in different modes; in addition, due to the fact that the peduncle area has the conditions of extremely irregular edges and complex and variable form and position, the existing U-Net network does not deal with the problem in a targeted mode. Therefore, an image processing method based on multi-mode stroke infarct region segmentation of an improved U-Net network is needed.
Disclosure of Invention
The invention aims to solve the technical problem of providing an image processing method for multi-mode stroke infarct region segmentation based on an improved U-Net network, which is used for carrying out image processing on medical brain images of stroke patients.
In order to solve the technical problem, the invention provides an image processing method for multi-modal stroke infarct region segmentation based on an improved U-Net network, which comprises the following specific processes: acquiring a brain medical image of a patient and inputting the brain medical image into an upper computer, wherein the brain medical image comprises a cerebral blood flow chart, a cerebral blood volume chart and a diffusion weighted imaging image, and matching 2-mode data of the cerebral blood flow chart and the cerebral blood volume chart of the same patient by using a 3D Slicer in a comprehensive analysis tool library by taking the diffusion weighted imaging image as a reference to obtain an NIFIT format file after registration; using trained U-NetAttentionThe network segments the registered NIFIT format file to obtain a two-dimensional segmentation result image, and displays the output result image in an upper computer;
the U-NetAttentionThe network comprises a U-Net network as a baseline model, an attention mechanism module is introduced at a long connection position of the baseline model for improvement, and a target area is emphasized in a low-level feature map by calculating weights through the feature map of a codec.
The invention relates to an improvement of an image processing method for multi-modal stroke infarct area segmentation based on an improved U-Net network, which comprises the following steps:
the attention mechanism module includes a channel attention weight calculation and a spatial attention weight calculation:
(1) spatial attention weight calculation
The features of the codec are extracted by two 1 × 1 convolutions, respectively, and the attention distribution F and the attention score a are obtained as follows:
F(xl,xh)=σ1(wαxl+bα)+σ1(wβxh+bβ) (formula 1)
A=σ2(wγF(xl,xh)+bγ) (formula 2)
Wherein x islFor the low-level feature map from the encoder, xhFor high level feature maps from the decoder, xlAnd xhIs the input of the attention module, σ1As Relu activation function, σ2Activating a function for Sigmoid, wα、wβ、wγAnd bα、bβ、bγParameters and offsets representing the convolution, respectively;
(2) channel attention weight calculation
Compressing the space size of the whole feature map into 1 × 1 through global average pooling operation, generating the attention weight on the channel through convolution of the channel feature map subjected to global average pooling GAP operation, and multiplying the attention weight by the low-level features to finally realize the attention weight calculation on the channel:
where x (i, j) is an element in the feature map at position (i, j), and W and H represent the size of the two-dimensional space of the feature map.
The invention is further improved by an image processing method for multi-mode stroke infarct region segmentation based on an improved U-Net network, and comprises the following steps:
the trained U-NetAttentionThe network acquisition process comprises the following steps:
(1) the loss function L is defined as:
L=αLDSC+βLfocal_loss(formula 4)
Wherein, alpha and beta are the weights of the two loss functions respectively;
wherein, ygtFor the true annotation value of a pixel, y' is viaAnd (3) outputting final prediction of the Sigmoid activation function, wherein alpha and gamma are two adjusting coefficients, wherein alpha is used for balancing positive and negative samples, and gamma is used for adjusting the weight of the samples;
wherein X represents a real sample label region, Y represents a prediction output result region, | X | represents the area of the real sample label region, | Y | represents the area of the prediction output result region;
(2)、U-NetAttentionnetwork training
Acquiring medical brain images of a patient, including a cerebral blood flow chart, a cerebral blood volume chart and diffusion-weighted imaging image data, matching different modal data of the patient by using a 3D Slicer comprehensive analysis tool library and taking diffusion-weighted imaging images as a reference, and acquiring a registered NIFIT format file as a U-NetAttentionAn input data set for training and testing of the network; manually labeling the peduncle region image on the diffusion weighted imaging modality image to be used as a corresponding label, taking 75% of input data sets and the corresponding labels as training data sets, and taking the rest 25% of the input data sets and the labels as test data sets;
using an Adam algorithm as an optimization loss function L, adopting a four-fold cross validation mode, and using an averaged result as an index of a final evaluation algorithm; training initial learning rate is set to 1e-3The lowest learning rate is 1e-10And the dynamic learning rate reduction is realized in an exponential decay mode according to the number of training rounds, and 260 rounds of iteration are trained to obtain a trained pt model file;
(3)、U-NetAttentionnetwork testing
Using U-NetAttentionLoading a trained pt model file by a network, and taking a test data set as U-NetAttentionThe network input obtains a segmentation result, thereby verifying and obtaining the trained U-NetAttentionA network.
The invention relates to an improvement of an image processing method for multi-modal stroke infarct area segmentation based on an improved U-Net network, which comprises the following steps:
the U-Net network includes an encoder, a decoder, and a long connection.
It is emphasized that the image obtained by the present invention is only used for the observation of the user, and is not used as the basis for judging the cerebral apoplexy infarction.
The invention has the following beneficial effects:
the method comprises the steps of establishing a base line model for segmenting the infarct area based on a U-Net network coding and decoding structure aiming at multi-modal brain medical images, improving an original long connection attention module aiming at the defect that splicing operation at a long connection part of the base line model is simple, and strengthening the attention degree of a region related to a final infarct area segmentation result in a bottom layer characteristic diagram through calculation of space and channel attention weight so as to improve the prediction level of the model.
Drawings
The following describes embodiments of the present invention in further detail with reference to the accompanying drawings.
FIG. 1 is a schematic flow chart of an image processing method for multi-modal stroke infarct region segmentation based on an improved U-Net network according to the present invention;
FIG. 2 is a topology diagram of a baseline network U-Net network of the present invention;
FIG. 3 is a diagram of a attention mechanism module network architecture of the present invention;
FIG. 4 shows the manual labeling, U-Net, of comparative experiment 1AttentionAnd (4) segmenting the result image contrast map by the network and the baseline U-Net network.
Detailed Description
The invention will be further described with reference to specific examples, but the scope of the invention is not limited thereto:
s1, acquiring images and registering:
acquiring a CT perfusion multi-modal brain image and a lesion information image labeled under a Diffusion Weighted Imaging (DWI) mode of a patient, wherein the CT perfusion multi-modal brain image comprises images of 2 modes of cerebral blood flow volume (CBF) and Cerebral Blood Volume (CBV);
the data used in this example are from published datasets (458 slices) provided by the desensitization-treated Ischemic Stroke CT perfusion multimodal brain imaging (286 slices) and Ischemic Stroke Lesion Segmentation (Ischemic Stroke Lesion Segmentation)2018 contest provided by the present laboratory collaboration hospital, both in Neuroimaging Informatics Technology Initiative (NIFTI) format. Each patient acquired images within 8 hours of ischemic stroke onset, first the patient generated CT perfusion multimodality brain images in dynamic scans with 1-2 second intervals by using contrast agent. The CT perfusion scanning is obtained in the form of sparse regions (with axial distance of 5mm) covered with stroke lesions, the slice range is 2-22, a Diffusion Weighted Imaging (DWI) image is obtained through magnetic resonance scanning within a few hours after scanning, and manual labeling of the infarct region is carried out on the Diffusion Weighted Imaging (DWI) image to obtain labeling data. All images are manually screened twice by medical experts, and some data which do not meet the requirements, such as imaging errors, labeling errors and omissions, are eliminated.
Because of different data sources, different acquisition equipment protocols and the like, images need to be uniformly registered, namely, the images under different modes, different fields of view and different time are subjected to a series of spatial transformation, so that the two images have the maximum similarity, and the matching purpose is achieved; and matching the data of 2 modes including cerebral blood flow volume (CBF) under CT perfusion and Cerebral Blood Volume (CBV) of the same patient by using a 3D Slicer of a comprehensive analysis tool library of nuclear magnetic resonance imaging and computed tomography imaging with a Diffusion Weighted Imaging (DWI) image as a reference to obtain a registered NIFIT format file, so that all data of the same patient have the maximum similarity.
S2、U-NetAttentionConstruction of a network
Taking a U-Net network as a baseline model, improving the model at long connection positions based on an attention mechanism, calculating weights through feature maps of a codec to emphasize target areas in a low-level feature map, and forming a new network called U-NetAttentionA network.
S201: building U-Net network model
A U-Net network is built as a baseline network, and as shown in FIG. 2, the U-Net network comprises three parts, namely an encoder, a decoder and a long connection. The encoder is mainly responsible for extracting features, and a deep feature map is obtained through a series of convolution and down-sampling operations; the decoder is mainly responsible for restoring the image, restores the size of the image through a series of convolution and up-sampling operations and generates a prediction result; the long connection is mainly responsible for splicing the feature maps of the corresponding space sizes of the encoder and the decoder, so that a finer recovery result is obtained by utilizing the low-layer features of the encoder; the network structure of the U-Net network model is shown in FIG. 2.
S202: module for constructing attention mechanism
Introducing an attention mechanism module at a long connection position of the U-Net network model, calculating a weight through a feature map of a coder-decoder to emphasize a target area in a low-layer feature map and finally improve the feature representation of the decoder, wherein the network structure diagram of the attention mechanism module is shown in FIG. 3;
the calculation of the attention mechanism module can be divided into two parts, channel attention weight calculation and spatial attention weight calculation:
(1) spatial attention weight calculation
First, the features of the codec are extracted by two 1 × 1 convolutions, and the attention distribution F and the attention score a are obtained by the following formula:
F(xl,xh)=σ1(wαxl+bα)+σ1(wβxh+bβ) (formula 1)
A=σ2(wγF(xl,xh)+bγ) (formula 2)
Wherein x islFor the low-level feature map from the encoder, xhFor high level feature maps from the decoder, xlAnd xhIs the input of the attention module, σ1Activating a function for Relu to introduce non-linearity in attention calculation, increasing complexity of weight calculation, σ2Activating a function for Sigmoid for generating a weight map between (0, 1), wα、wβ、wγAnd bα、bβ、bγParameters and offsets representing the convolution, respectively;
(2) channel attention weight calculation
Compressing the space size of the whole feature map to 1 × 1 through global average pooling operation, wherein W and H in formula 3 represent the two-dimensional space size of the feature map, generating the attention weight on the channel through convolution of the channel feature map after Global Average Pooling (GAP) operation, and multiplying the attention weight by the low-level features to finally realize the attention weight calculation on the channel:
wherein x (i, j) is an element with position (i, j) in the feature diagram
S3:U-NetAttentionTraining and testing of models
S301: establishing a loss function
To increase U-NetAttentionThe identification capability of the model to the whole infarcted area is realized by introducing DSC (differential scanning calorimetry) Loss as another weight on the basis of Focal Loss, and using the Focal Loss and the Focal LossThe weighting of DSC Loss as a Loss function is defined as follows:
L=αLDSC+βLfocal_loss(formula 4)
Where α and β are used to adjust the weights of the two loss functions.
The calculation formula of Focal local is as follows:
wherein, ygtThe true labeled value of the pixel is obtained, y' is the final prediction output through a Sigmoid activation function, alpha and gamma are two adjusting coefficients, wherein alpha is used for balancing positive and negative samples, and gamma is used for adjusting the weight of the samples;
the DSC Loss is calculated as follows:
wherein X represents a real sample label region; y denotes a prediction output result region. | X | represents the area of the true sample label region and | Y | represents the area of the prediction output result region.
S302:U-NetAttentionNetwork training
Obtaining medical brain images of the patient according to step S1, including CT perfusion Cerebral Blood Flow (CBF), Cerebral Blood Volume (CBV)2 modal images and Diffusion Weighted Imaging (DWI) image data, inputting the images into an upper computer, then performing unified registration on the images, matching different modal data of the patient by using a comprehensive analysis tool library 3D Slicer with the Diffusion Weighted Imaging (DWI) image as a reference, and obtaining a registered NIFIT format file as U-NetAttentionAn input data set for training and testing of the network; meanwhile, manually labeling an peduncle region image on a Diffusion Weighted Imaging (DWI) image as a label corresponding to an input image, taking 75% of the input data set and the label as a training data set, and taking the rest 25% of the input data set and the label as a test data set;
the training data set and the test data set contain as far as possible image data of different patients.
Using an Adam algorithm as an optimization loss function L, and paying attention to the fact that in order to avoid the deviation of results caused by unreasonable segmentation of a training data set and a test data set, a four-fold cross validation mode is adopted in the experimental process, namely, the training data set is divided into four parts according to patients, three parts of the training data set are used as the training set and the other part of the training data set is used as a validation set each time, the experiment is repeated for four times, and the average result is used as the index of the final evaluation algorithm;
U-NetAttentionthe training of the network is realized based on a Tensorflow framework. Initial learning rate is set to 1e-3The lowest learning rate is 1e-10And a dynamic learning rate reduction is achieved in an exponentially decaying manner according to the number of training rounds. And (4) performing 260 rounds of iteration to obtain a trained pt model file, wherein weights and parameters of the neural network nodes are stored in the pt model file and are used for testing the test data set in the step S303.
S303:U-NetAttentionNetwork testing
Using U-Net in the test phaseAttentionAnd (3) loading the trained pt model file, namely loading the model weight obtained by training, and testing by using the multi-mode image of the test data set which does not participate in the training as U-NetAttentionThe network inputs to obtain a prediction result, and a two-dimensional segmentation result image with higher prediction accuracy than that of the U-Net baseline model is successfully obtained, so that the trained U-Net is verified and obtainedAttentionA network.
S04, online use
Acquiring a medical brain image of a patient, wherein the medical brain image comprises a cerebral blood flow volume (CBF) map, a Cerebral Blood Volume (CBV) map and Diffusion Weighted Imaging (DWI) image data, and inputting the medical brain image into an upper computer; matching different modal data of a patient by using a comprehensive analysis tool library 3D Slicer with DWI images as a reference to obtain a registered NIFIT format file; with the trained U-Net obtained in step S303AttentionSegmenting the registered NIFIT format file to obtain a two-dimensional segmentation result image, and displaying the two-dimensional segmentation result image in an upper computerAnd displaying and outputting the two-dimensional segmentation result image.
Comparative experiment 1:
acquiring medical brain images of the same patient, including a cerebral blood flow volume (CBF) map, a Cerebral Blood Volume (CBV) map and Diffusion Weighted Imaging (DWI) image data, and inputting the images into an upper computer; matching different modal data of a patient by using a comprehensive analysis tool library 3D Slicer with DWI images as a reference to obtain a registered NIFIT format file; then, the doctor inputs the registered NIFIT format files into U-Net respectively according to the actual infarct area marked by the DWI imageAttentionObtaining two-dimensional infarct region segmentation result images by a network and a baseline U-Net network, manually labeling and U-NetAttentionThe segmentation result images of the network and the baseline U-Net network infarct area are shown in FIG. 4, the first column is the image of the actual infarct area labeled by the doctor according to the DWI image, the second column is the segmentation result image of the baseline U-Net, and the third column is the U-Net in the inventionAttentionAnd (4) segmenting a result image of the model. As can be seen from FIG. 4, the division area of the base-line network U-Net is obviously larger than the actual infarction area, and is easy to shift and miss detection, and after the improvement of the long connection attention mechanism, the U-Net in the inventionAttentionThe conditions of missed detection and false detection of the infarct area by the model are reduced, the similarity of the predicted infarct area and the normal area is improved, and the similarity is closer to a real labeling result.
Comparative experiment 2:
data used in the comparison experiment are from desensitized Ischemic Stroke CT perfusion multi-modal brain images (286 slices) provided by a cooperative hospital and public data sets (458 slices) provided by an Ischemic Stroke Lesion Segmentation (Ischemic Stroke Segmentation)2018 competition, different modal data of a patient are matched by using a comprehensive analysis tool library 3D Slicer with DWI images as a reference, and registered NIFIT format files are obtained; and then inputting the registered NIFIT format file into a U-NetAttention network and a baseline U-Net network respectively so as to obtain a two-dimensional infarct area segmentation result image.
The DSC coefficient, Sensitivity (Sensitivity), and Hausdorff Distance (HD for short) are used as performance indexes for comparison, and the calculation method is described as follows:
1) DSC coefficient: setting a pixel set X on the segmented image of the infarct area obtained by model calculation, and a pixel set Y of the labeling image of the real infarct area, wherein 1 represents that the pixel belongs to the infarct area, and 0 represents that the pixel belongs to the normal area, and then calculating DSC coefficients of the two sets as follows:
wherein | X | + | Y | is the intersection of the sets X and Y, and | X | + | Y | is the addition of the point sets of the sets X and Y.
2) The Sensitivity (Sensitivity) represents how many pixels belonging to the infarct area are correctly predicted as the true value, and the calculation method is as follows:
wherein TP represents that the true value is a pixel belonging to the infarct area and the model is predicted as the infarct area, FP represents that the true value is a pixel belonging to the infarct area and the model is predicted as the normal area
3) The Housdov Distance (HD) is used for measuring the similarity between two point sets, and a pixel set X on an peduncle region segmentation image obtained by model calculation and a pixel set Y of a real peduncle region labeling image are calculated in the following mode:
dH(X,Y)=max{supx∈Xinfy∈Yd(x,y),supy∈Yinfx∈Xd (x, y) } (equation 9)
The DSC coefficient is the most important evaluation index, and when DSCs are close, Sensitivity (Sensitivity) and Hausdorff Distance (HD) indexes are used as supplementary comparisons, and the results of the comparative experiments are as follows:
the higher the DSC coefficient, the lower the HD value, and the lower the Sensitivity value, the higher the degree of area coincidence between the predicted infarction area and the actual infarction area, and the greater the similarity. From the results in the table, it can be known that the improved attention module enhances the attention of the model to the important region on the feature map through the calculation of the space and the channel weight, and improves the segmentation accuracy of the final infarct region.
Finally, it is also noted that the above-mentioned lists merely illustrate a few specific embodiments of the invention. It is obvious that the invention is not limited to the above embodiments, but that many variations are possible. All modifications which can be derived or suggested by a person skilled in the art from the disclosure of the present invention are to be considered within the scope of the invention.
Claims (4)
1. A multi-mode stroke infarct region image processing method based on an improved U-Net network is characterized by comprising the following steps:
acquiring a brain medical image of a patient and inputting the brain medical image into an upper computer, wherein the brain medical image comprises a cerebral blood flow chart, a cerebral blood volume chart and a diffusion weighted imaging image, and matching 2-mode data of the cerebral blood flow chart and the cerebral blood volume chart of the same patient by using a 3D Slicer in a comprehensive analysis tool library by taking the diffusion weighted imaging image as a reference to obtain an NIFIT format file after registration; using trained U-NetAttentionThe network segments the registered NIFIT format file to obtain a two-dimensional segmentation result image, and displays the output result image in an upper computer;
the U-NetAttentionThe network comprises a U-Net network as a baseline model, an attention mechanism module is introduced at a long connection position of the baseline model for improvement, and a target area is emphasized in a low-level feature map by calculating weights through the feature map of a codec.
2. The image processing method of multi-modal stroke infarct region segmentation based on the improved U-Net network according to claim 1, characterized in that:
the attention mechanism module includes a channel attention weight calculation and a spatial attention weight calculation:
(1) spatial attention weight calculation
The features of the codec are extracted by two 1 × 1 convolutions, respectively, and the attention distribution F and the attention score a are obtained as follows:
F(xl,xh)=σ1(wαxl+bα)+σ1(wβxh+bβ) (formula 1)
A=σ2(wγF(xl,xh)+bγ) (formula 2)
Wherein x islFor the low-level feature map from the encoder, xhFor high level feature maps from the decoder, xlAnd xhIs the input of the attention module, σ1As Relu activation function, σ2Activating a function for Sigmoid, wα、wβ、wγAnd bα、bβ、bγParameters and offsets representing the convolution, respectively;
(2) channel attention weight calculation
Compressing the space size of the whole feature map into 1 × 1 through global average pooling operation, generating the attention weight on the channel through convolution of the channel feature map subjected to global average pooling GAP operation, and multiplying the attention weight by the low-level features to finally realize the attention weight calculation on the channel:
where x (i, j) is an element in the feature map at position (i, j), and W and H represent the size of the two-dimensional space of the feature map.
3. The image processing method of multi-modal stroke infarct region segmentation based on the improved U-Net network according to claim 2, characterized in that:
the trained U-NetAttentionThe network acquisition process comprises the following steps:
(1) the loss function L is defined as:
L=αLDSC+βLfocal_loss(formula 4)
Wherein, alpha and beta are the weights of the two loss functions respectively;
wherein, ygtThe true labeled value of the pixel is obtained, y' is the final prediction output through a Sigmoid activation function, alpha and gamma are two adjusting coefficients, wherein alpha is used for balancing positive and negative samples, and gamma is used for adjusting the weight of the samples;
wherein X represents a real sample label region, Y represents a prediction output result region, | X | represents the area of the real sample label region, | Y | represents the area of the prediction output result region;
(2)、U-NetAttentionnetwork training
Acquiring medical brain images of a patient, including a cerebral blood flow chart, a cerebral blood volume chart and diffusion-weighted imaging image data, matching different modal data of the patient by using a 3D Slicer comprehensive analysis tool library and taking diffusion-weighted imaging images as a reference, and acquiring a registered NIFIT format file as a U-NetAttentionAn input data set for training and testing of the network; manually labeling the peduncle region image on the diffusion weighted imaging modality image to be used as a corresponding label, taking 75% of input data sets and the corresponding labels as training data sets, and taking the rest 25% of the input data sets and the labels as test data sets;
using Adam's algorithm as the bestChanging a loss function L, adopting a verification mode of four-fold cross verification, and taking an average result as an index of a final evaluation algorithm; training initial learning rate is set to le-3The lowest learning rate is le-10And the dynamic learning rate reduction is realized in an exponential decay mode according to the number of training rounds, and 260 rounds of iteration are trained to obtain a trained pt model file;
(3)、U-NetAttentionnetwork testing
Using U-NetAttentionLoading a trained pt model file by a network, and taking a test data set as U-NetAttentionThe network input obtains a segmentation result, thereby verifying and obtaining the trained U-NetAttentionA network.
4. The image processing method of multi-modal stroke infarct region segmentation based on the improved U-Net network according to claim 3, characterized in that:
the U-Net network includes an encoder, a decoder, and a long connection.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202111606856.XA CN114359194B (en) | 2021-12-27 | 2021-12-27 | Multimode cerebral apoplexy infarct region image processing method based on improved U-Net network |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202111606856.XA CN114359194B (en) | 2021-12-27 | 2021-12-27 | Multimode cerebral apoplexy infarct region image processing method based on improved U-Net network |
Publications (2)
Publication Number | Publication Date |
---|---|
CN114359194A true CN114359194A (en) | 2022-04-15 |
CN114359194B CN114359194B (en) | 2024-07-12 |
Family
ID=81100550
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN202111606856.XA Active CN114359194B (en) | 2021-12-27 | 2021-12-27 | Multimode cerebral apoplexy infarct region image processing method based on improved U-Net network |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN114359194B (en) |
Cited By (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN117422788A (en) * | 2023-12-19 | 2024-01-19 | 英瑞云医疗科技(烟台)有限公司 | Method for generating DWI image based on CT brain stem image |
Citations (11)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN109658401A (en) * | 2018-12-14 | 2019-04-19 | 上海商汤智能科技有限公司 | Image processing method and device, electronic equipment and storage medium |
CN110223285A (en) * | 2019-06-13 | 2019-09-10 | 中南大学湘雅二医院 | Imaging results prediction technique and system based on animal migration and neural network model |
CN110322528A (en) * | 2019-06-26 | 2019-10-11 | 浙江大学 | Nuclear magnetic resonance brain image reconstructing blood vessel method based on 3T, 7T |
US20200381096A1 (en) * | 2019-06-03 | 2020-12-03 | The Board Of Trustees Of The Leland Stanford Junior University | Methods of Predicting Disorder Progression for Control Arms Within an Experimental Trial |
CN112150449A (en) * | 2020-09-29 | 2020-12-29 | 太原理工大学 | Cerebral apoplexy focus segmentation method and system |
CN112418027A (en) * | 2020-11-11 | 2021-02-26 | 青岛科技大学 | Remote sensing image road extraction method for improving U-Net network |
CN112767374A (en) * | 2021-01-27 | 2021-05-07 | 天津理工大学 | Alzheimer disease focus region semantic segmentation algorithm based on MRI |
CN113095382A (en) * | 2021-03-30 | 2021-07-09 | 浙江大学 | Interpretable tuberculosis classification network identification method based on CT image |
CN113177943A (en) * | 2021-06-29 | 2021-07-27 | 中南大学 | Cerebral apoplexy CT image segmentation method |
WO2021184817A1 (en) * | 2020-03-16 | 2021-09-23 | 苏州科技大学 | Method for segmenting liver and focus thereof in medical image |
CN113627564A (en) * | 2021-08-23 | 2021-11-09 | 李永鑫 | Deep learning-based CT medical image processing model training method and diagnosis and treatment system |
-
2021
- 2021-12-27 CN CN202111606856.XA patent/CN114359194B/en active Active
Patent Citations (11)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN109658401A (en) * | 2018-12-14 | 2019-04-19 | 上海商汤智能科技有限公司 | Image processing method and device, electronic equipment and storage medium |
US20200381096A1 (en) * | 2019-06-03 | 2020-12-03 | The Board Of Trustees Of The Leland Stanford Junior University | Methods of Predicting Disorder Progression for Control Arms Within an Experimental Trial |
CN110223285A (en) * | 2019-06-13 | 2019-09-10 | 中南大学湘雅二医院 | Imaging results prediction technique and system based on animal migration and neural network model |
CN110322528A (en) * | 2019-06-26 | 2019-10-11 | 浙江大学 | Nuclear magnetic resonance brain image reconstructing blood vessel method based on 3T, 7T |
WO2021184817A1 (en) * | 2020-03-16 | 2021-09-23 | 苏州科技大学 | Method for segmenting liver and focus thereof in medical image |
CN112150449A (en) * | 2020-09-29 | 2020-12-29 | 太原理工大学 | Cerebral apoplexy focus segmentation method and system |
CN112418027A (en) * | 2020-11-11 | 2021-02-26 | 青岛科技大学 | Remote sensing image road extraction method for improving U-Net network |
CN112767374A (en) * | 2021-01-27 | 2021-05-07 | 天津理工大学 | Alzheimer disease focus region semantic segmentation algorithm based on MRI |
CN113095382A (en) * | 2021-03-30 | 2021-07-09 | 浙江大学 | Interpretable tuberculosis classification network identification method based on CT image |
CN113177943A (en) * | 2021-06-29 | 2021-07-27 | 中南大学 | Cerebral apoplexy CT image segmentation method |
CN113627564A (en) * | 2021-08-23 | 2021-11-09 | 李永鑫 | Deep learning-based CT medical image processing model training method and diagnosis and treatment system |
Non-Patent Citations (3)
Title |
---|
AN YANG等: "CT Images Recognition of Pulmonary Tuberculosis Based on Improved Faster RCNN and U-Net", 《2019 10TH INTERNATIONAL CONFERENCE ON INFORMATION TECHNOLOGY IN MEDICINE AND EDUCATION 》, 23 January 2020 (2020-01-23) * |
TAO SONG: "Generative Model-Based Ischemic Stroke Lesion Segmentation", 《IMAGE AND VIDEO PROCESSING》, 6 June 2019 (2019-06-06) * |
何洪林;谢峻;李毅;雷跳;钱俊;马桃林;: "基于裸眼3D视觉效果的脑血管CT图像快速三维重建方法研究", 医疗卫生装备, no. 09, 15 September 2015 (2015-09-15) * |
Cited By (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN117422788A (en) * | 2023-12-19 | 2024-01-19 | 英瑞云医疗科技(烟台)有限公司 | Method for generating DWI image based on CT brain stem image |
CN117422788B (en) * | 2023-12-19 | 2024-03-26 | 英瑞云医疗科技(烟台)有限公司 | Method for generating DWI image based on CT brain stem image |
Also Published As
Publication number | Publication date |
---|---|
CN114359194B (en) | 2024-07-12 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
Khan et al. | Deep neural architectures for medical image semantic segmentation | |
WO2016192612A1 (en) | Method for analysing medical treatment data based on deep learning, and intelligent analyser thereof | |
CN107016395B (en) | Identification system for sparsely expressed primary brain lymphomas and glioblastomas | |
CN112086197B (en) | Breast nodule detection method and system based on ultrasonic medicine | |
Xie et al. | Computer‐Aided System for the Detection of Multicategory Pulmonary Tuberculosis in Radiographs | |
EP3654343A1 (en) | Application of deep learning for medical imaging evaluation | |
CN111755118B (en) | Medical information processing method, device, electronic equipment and storage medium | |
CN114782307A (en) | Enhanced CT image colorectal cancer staging auxiliary diagnosis system based on deep learning | |
EP2996058A1 (en) | Method for automatically generating representations of imaging data and interactive visual imaging reports | |
Li et al. | Optical coherence tomography vulnerable plaque segmentation based on deep residual U-Net | |
CN112508884A (en) | Comprehensive detection device and method for cancerous region | |
CN116091490A (en) | Lung nodule detection method based on YOLOv4-CA-CBAM-K-means++ -SIOU | |
CN116433654A (en) | Improved U-Net network spine integral segmentation method | |
JP2024528381A (en) | Method and system for automatically tracking and interpreting medical image data | |
CN110570425A (en) | Lung nodule analysis method and device based on deep reinforcement learning algorithm | |
CN114359194B (en) | Multimode cerebral apoplexy infarct region image processing method based on improved U-Net network | |
Saeed et al. | 3D MRU-Net: A novel mobile residual U-Net deep learning model for spine segmentation using computed tomography images | |
CN113902738A (en) | Heart MRI segmentation method and system | |
Singha Deo et al. | Supremacy of attention based convolution neural network in classification of oral cancer using histopathological images | |
Zhang et al. | Femoral image segmentation based on two-stage convolutional network using 3D-DMFNet and 3D-ResUnet | |
CN117711576A (en) | Method and system for providing a template data structure for medical reports | |
Quy et al. | Multi-view digital mammography mass classification: A convolutional neural network model approach | |
CN116631584A (en) | General medical image report generation method and system, electronic equipment and readable storage medium | |
CN113362350B (en) | Method, device, terminal equipment and storage medium for segmenting cancer medical record image | |
WO2020099941A1 (en) | Application of deep learning for medical imaging evaluation |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
GR01 | Patent grant | ||
GR01 | Patent grant |