CN114359194A - Multi-mode stroke infarct area image processing method based on improved U-Net network - Google Patents

Multi-mode stroke infarct area image processing method based on improved U-Net network Download PDF

Info

Publication number
CN114359194A
CN114359194A CN202111606856.XA CN202111606856A CN114359194A CN 114359194 A CN114359194 A CN 114359194A CN 202111606856 A CN202111606856 A CN 202111606856A CN 114359194 A CN114359194 A CN 114359194A
Authority
CN
China
Prior art keywords
attention
net
image
network
cerebral blood
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN202111606856.XA
Other languages
Chinese (zh)
Other versions
CN114359194B (en
Inventor
金心宇
陈榕
赵晗
金昀程
陈智鸿
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Zhejiang University ZJU
Original Assignee
Zhejiang University ZJU
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Zhejiang University ZJU filed Critical Zhejiang University ZJU
Priority to CN202111606856.XA priority Critical patent/CN114359194B/en
Publication of CN114359194A publication Critical patent/CN114359194A/en
Application granted granted Critical
Publication of CN114359194B publication Critical patent/CN114359194B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/0002Inspection of images, e.g. flaw detection
    • G06T7/0012Biomedical image inspection
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/045Combinations of networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/048Activation functions
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/08Learning methods
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/10Segmentation; Edge detection
    • G06T7/11Region-based segmentation
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/30Determination of transform parameters for the alignment of images, i.e. image registration
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/10Image acquisition modality
    • G06T2207/10072Tomographic images
    • G06T2207/10088Magnetic resonance imaging [MRI]
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/20Special algorithmic details
    • G06T2207/20081Training; Learning
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/20Special algorithmic details
    • G06T2207/20084Artificial neural networks [ANN]
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/30Subject of image; Context of image processing
    • G06T2207/30004Biomedical image processing
    • G06T2207/30016Brain
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/30Subject of image; Context of image processing
    • G06T2207/30004Biomedical image processing
    • G06T2207/30101Blood vessel; Artery; Vein; Vascular
    • G06T2207/30104Vascular flow; Blood flow; Perfusion

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • General Health & Medical Sciences (AREA)
  • Health & Medical Sciences (AREA)
  • Computing Systems (AREA)
  • Mathematical Physics (AREA)
  • Data Mining & Analysis (AREA)
  • Evolutionary Computation (AREA)
  • Biophysics (AREA)
  • Molecular Biology (AREA)
  • Biomedical Technology (AREA)
  • General Engineering & Computer Science (AREA)
  • Artificial Intelligence (AREA)
  • Computational Linguistics (AREA)
  • Software Systems (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Medical Informatics (AREA)
  • Nuclear Medicine, Radiotherapy & Molecular Imaging (AREA)
  • Radiology & Medical Imaging (AREA)
  • Quality & Reliability (AREA)
  • Image Analysis (AREA)
  • Image Processing (AREA)

Abstract

The invention discloses a multi-mode cerebral apoplexy infarct region image processing method based on an improved U-Net network, which comprises the steps of obtaining a brain medical image of a patient and inputting the brain medical image into an upper computer, wherein the brain medical image comprises a cerebral blood flow chart, a cerebral blood volume chart and a diffusion weighted imaging image, and matching 2-mode data of the cerebral blood flow chart and the cerebral blood volume chart of the same patient by using a 3D Slicer as a comprehensive analysis tool library with the diffusion weighted imaging image as a reference to obtain a registered NIFIT format file; using trained U-NetAttentionThe network divides the registered NIFIT format fileAnd cutting to obtain a two-dimensional segmentation result image, and displaying the output result image in the upper computer. According to the method, the attention degree of the region related to the final segmentation result of the infarct region in the bottom layer characteristic diagram is enhanced through calculation of the attention weight of the space and the channel, so that the prediction level of the model is improved.

Description

Multi-mode stroke infarct area image processing method based on improved U-Net network
Technical Field
The invention relates to the technical field of medical image processing, in particular to an image processing method for multi-mode stroke infarction region segmentation based on an improved U-Net network.
Background
The stroke becomes the first threat to the health and life safety of middle-aged and elderly people in China. Perfusion images of Nuclear Magnetic Resonance Imaging (NMRI) and Computed Tomography (CT) are the criteria for a physician to determine the infarcted area of the brain of a patient. With the rapid development of medical image processing technology, digital Imaging and Communications DICOM (digital Imaging and Communications in medicine) is widely used in radiological diagnosis and treatment equipment, and the calculation of each diagnosis index of stroke increasingly depends on digital Imaging and Communications in DICOM medical image. At present, the clinical diagnosis of stroke still depends on manually labeled Nuclear Magnetic Resonance Images (NMRI) or Computed Tomography (CT) perfusion images, the method is time-consuming and easy to delay the state of an illness, and doctors with rich experience are needed, otherwise, the judgment of the state of an illness is easy to be inaccurate.
With the development of deep learning technology in recent years, the deep learning technology has been successful in the fields of target detection, image classification, image segmentation and the like. The application of deep learning techniques in medical image processing is also increasing, and higher accuracy is achieved in many problems than in conventional machine learning and statistical algorithms. Segmentation of an ischemic stroke infarct region is a typical medical image segmentation problem, and can be predicted by using a U-Net network model which is widely used in the field of medical image segmentation. However, the traditional prediction model only applies the brain image of the patient in a single mode, and does not fully mine rich information provided by data in different modes; in addition, due to the fact that the peduncle area has the conditions of extremely irregular edges and complex and variable form and position, the existing U-Net network does not deal with the problem in a targeted mode. Therefore, an image processing method based on multi-mode stroke infarct region segmentation of an improved U-Net network is needed.
Disclosure of Invention
The invention aims to solve the technical problem of providing an image processing method for multi-mode stroke infarct region segmentation based on an improved U-Net network, which is used for carrying out image processing on medical brain images of stroke patients.
In order to solve the technical problem, the invention provides an image processing method for multi-modal stroke infarct region segmentation based on an improved U-Net network, which comprises the following specific processes: acquiring a brain medical image of a patient and inputting the brain medical image into an upper computer, wherein the brain medical image comprises a cerebral blood flow chart, a cerebral blood volume chart and a diffusion weighted imaging image, and matching 2-mode data of the cerebral blood flow chart and the cerebral blood volume chart of the same patient by using a 3D Slicer in a comprehensive analysis tool library by taking the diffusion weighted imaging image as a reference to obtain an NIFIT format file after registration; using trained U-NetAttentionThe network segments the registered NIFIT format file to obtain a two-dimensional segmentation result image, and displays the output result image in an upper computer;
the U-NetAttentionThe network comprises a U-Net network as a baseline model, an attention mechanism module is introduced at a long connection position of the baseline model for improvement, and a target area is emphasized in a low-level feature map by calculating weights through the feature map of a codec.
The invention relates to an improvement of an image processing method for multi-modal stroke infarct area segmentation based on an improved U-Net network, which comprises the following steps:
the attention mechanism module includes a channel attention weight calculation and a spatial attention weight calculation:
(1) spatial attention weight calculation
The features of the codec are extracted by two 1 × 1 convolutions, respectively, and the attention distribution F and the attention score a are obtained as follows:
F(xl,xh)=σ1(wαxl+bα)+σ1(wβxh+bβ) (formula 1)
A=σ2(wγF(xl,xh)+bγ) (formula 2)
Wherein x islFor the low-level feature map from the encoder, xhFor high level feature maps from the decoder, xlAnd xhIs the input of the attention module, σ1As Relu activation function, σ2Activating a function for Sigmoid, wα、wβ、wγAnd bα、bβ、bγParameters and offsets representing the convolution, respectively;
(2) channel attention weight calculation
Compressing the space size of the whole feature map into 1 × 1 through global average pooling operation, generating the attention weight on the channel through convolution of the channel feature map subjected to global average pooling GAP operation, and multiplying the attention weight by the low-level features to finally realize the attention weight calculation on the channel:
Figure BDA0003434259000000021
where x (i, j) is an element in the feature map at position (i, j), and W and H represent the size of the two-dimensional space of the feature map.
The invention is further improved by an image processing method for multi-mode stroke infarct region segmentation based on an improved U-Net network, and comprises the following steps:
the trained U-NetAttentionThe network acquisition process comprises the following steps:
(1) the loss function L is defined as:
L=αLDSC+βLfocal_loss(formula 4)
Wherein, alpha and beta are the weights of the two loss functions respectively;
Figure BDA0003434259000000031
wherein, ygtFor the true annotation value of a pixel, y' is viaAnd (3) outputting final prediction of the Sigmoid activation function, wherein alpha and gamma are two adjusting coefficients, wherein alpha is used for balancing positive and negative samples, and gamma is used for adjusting the weight of the samples;
Figure BDA0003434259000000032
wherein X represents a real sample label region, Y represents a prediction output result region, | X | represents the area of the real sample label region, | Y | represents the area of the prediction output result region;
(2)、U-NetAttentionnetwork training
Acquiring medical brain images of a patient, including a cerebral blood flow chart, a cerebral blood volume chart and diffusion-weighted imaging image data, matching different modal data of the patient by using a 3D Slicer comprehensive analysis tool library and taking diffusion-weighted imaging images as a reference, and acquiring a registered NIFIT format file as a U-NetAttentionAn input data set for training and testing of the network; manually labeling the peduncle region image on the diffusion weighted imaging modality image to be used as a corresponding label, taking 75% of input data sets and the corresponding labels as training data sets, and taking the rest 25% of the input data sets and the labels as test data sets;
using an Adam algorithm as an optimization loss function L, adopting a four-fold cross validation mode, and using an averaged result as an index of a final evaluation algorithm; training initial learning rate is set to 1e-3The lowest learning rate is 1e-10And the dynamic learning rate reduction is realized in an exponential decay mode according to the number of training rounds, and 260 rounds of iteration are trained to obtain a trained pt model file;
(3)、U-NetAttentionnetwork testing
Using U-NetAttentionLoading a trained pt model file by a network, and taking a test data set as U-NetAttentionThe network input obtains a segmentation result, thereby verifying and obtaining the trained U-NetAttentionA network.
The invention relates to an improvement of an image processing method for multi-modal stroke infarct area segmentation based on an improved U-Net network, which comprises the following steps:
the U-Net network includes an encoder, a decoder, and a long connection.
It is emphasized that the image obtained by the present invention is only used for the observation of the user, and is not used as the basis for judging the cerebral apoplexy infarction.
The invention has the following beneficial effects:
the method comprises the steps of establishing a base line model for segmenting the infarct area based on a U-Net network coding and decoding structure aiming at multi-modal brain medical images, improving an original long connection attention module aiming at the defect that splicing operation at a long connection part of the base line model is simple, and strengthening the attention degree of a region related to a final infarct area segmentation result in a bottom layer characteristic diagram through calculation of space and channel attention weight so as to improve the prediction level of the model.
Drawings
The following describes embodiments of the present invention in further detail with reference to the accompanying drawings.
FIG. 1 is a schematic flow chart of an image processing method for multi-modal stroke infarct region segmentation based on an improved U-Net network according to the present invention;
FIG. 2 is a topology diagram of a baseline network U-Net network of the present invention;
FIG. 3 is a diagram of a attention mechanism module network architecture of the present invention;
FIG. 4 shows the manual labeling, U-Net, of comparative experiment 1AttentionAnd (4) segmenting the result image contrast map by the network and the baseline U-Net network.
Detailed Description
The invention will be further described with reference to specific examples, but the scope of the invention is not limited thereto:
embodiment 1, an image processing method for multi-modal stroke infarct region segmentation based on an improved U-Net network, as shown in fig. 1 to 3, includes obtaining a CT perfusion multi-modal brain medical image of a stroke patient and a lesion information image labeled in a Diffusion Weighted Imaging (DWI) format, and performing matching with a Diffusion Weighted Imaging (DWI) modal image as a reference, so that the images are matched with each otherAll data of the same patient have the greatest similarity; taking a U-Net network as a baseline model, improving the model at long connection positions based on an attention mechanism, calculating weights through feature maps of a codec to emphasize target areas in a low-level feature map, and forming a new network called U-NetAttentionA network; taking CT perfusion multi-modal brain images as input, taking marked focus information images as labels, and training the network to extract feature description of focus information, thereby improving the recognition accuracy rate of the model on the whole infarction region; the method comprises the following specific steps:
s1, acquiring images and registering:
acquiring a CT perfusion multi-modal brain image and a lesion information image labeled under a Diffusion Weighted Imaging (DWI) mode of a patient, wherein the CT perfusion multi-modal brain image comprises images of 2 modes of cerebral blood flow volume (CBF) and Cerebral Blood Volume (CBV);
the data used in this example are from published datasets (458 slices) provided by the desensitization-treated Ischemic Stroke CT perfusion multimodal brain imaging (286 slices) and Ischemic Stroke Lesion Segmentation (Ischemic Stroke Lesion Segmentation)2018 contest provided by the present laboratory collaboration hospital, both in Neuroimaging Informatics Technology Initiative (NIFTI) format. Each patient acquired images within 8 hours of ischemic stroke onset, first the patient generated CT perfusion multimodality brain images in dynamic scans with 1-2 second intervals by using contrast agent. The CT perfusion scanning is obtained in the form of sparse regions (with axial distance of 5mm) covered with stroke lesions, the slice range is 2-22, a Diffusion Weighted Imaging (DWI) image is obtained through magnetic resonance scanning within a few hours after scanning, and manual labeling of the infarct region is carried out on the Diffusion Weighted Imaging (DWI) image to obtain labeling data. All images are manually screened twice by medical experts, and some data which do not meet the requirements, such as imaging errors, labeling errors and omissions, are eliminated.
Because of different data sources, different acquisition equipment protocols and the like, images need to be uniformly registered, namely, the images under different modes, different fields of view and different time are subjected to a series of spatial transformation, so that the two images have the maximum similarity, and the matching purpose is achieved; and matching the data of 2 modes including cerebral blood flow volume (CBF) under CT perfusion and Cerebral Blood Volume (CBV) of the same patient by using a 3D Slicer of a comprehensive analysis tool library of nuclear magnetic resonance imaging and computed tomography imaging with a Diffusion Weighted Imaging (DWI) image as a reference to obtain a registered NIFIT format file, so that all data of the same patient have the maximum similarity.
S2、U-NetAttentionConstruction of a network
Taking a U-Net network as a baseline model, improving the model at long connection positions based on an attention mechanism, calculating weights through feature maps of a codec to emphasize target areas in a low-level feature map, and forming a new network called U-NetAttentionA network.
S201: building U-Net network model
A U-Net network is built as a baseline network, and as shown in FIG. 2, the U-Net network comprises three parts, namely an encoder, a decoder and a long connection. The encoder is mainly responsible for extracting features, and a deep feature map is obtained through a series of convolution and down-sampling operations; the decoder is mainly responsible for restoring the image, restores the size of the image through a series of convolution and up-sampling operations and generates a prediction result; the long connection is mainly responsible for splicing the feature maps of the corresponding space sizes of the encoder and the decoder, so that a finer recovery result is obtained by utilizing the low-layer features of the encoder; the network structure of the U-Net network model is shown in FIG. 2.
S202: module for constructing attention mechanism
Introducing an attention mechanism module at a long connection position of the U-Net network model, calculating a weight through a feature map of a coder-decoder to emphasize a target area in a low-layer feature map and finally improve the feature representation of the decoder, wherein the network structure diagram of the attention mechanism module is shown in FIG. 3;
the calculation of the attention mechanism module can be divided into two parts, channel attention weight calculation and spatial attention weight calculation:
(1) spatial attention weight calculation
First, the features of the codec are extracted by two 1 × 1 convolutions, and the attention distribution F and the attention score a are obtained by the following formula:
F(xl,xh)=σ1(wαxl+bα)+σ1(wβxh+bβ) (formula 1)
A=σ2(wγF(xl,xh)+bγ) (formula 2)
Wherein x islFor the low-level feature map from the encoder, xhFor high level feature maps from the decoder, xlAnd xhIs the input of the attention module, σ1Activating a function for Relu to introduce non-linearity in attention calculation, increasing complexity of weight calculation, σ2Activating a function for Sigmoid for generating a weight map between (0, 1), wα、wβ、wγAnd bα、bβ、bγParameters and offsets representing the convolution, respectively;
(2) channel attention weight calculation
Compressing the space size of the whole feature map to 1 × 1 through global average pooling operation, wherein W and H in formula 3 represent the two-dimensional space size of the feature map, generating the attention weight on the channel through convolution of the channel feature map after Global Average Pooling (GAP) operation, and multiplying the attention weight by the low-level features to finally realize the attention weight calculation on the channel:
Figure BDA0003434259000000061
wherein x (i, j) is an element with position (i, j) in the feature diagram
S3:U-NetAttentionTraining and testing of models
S301: establishing a loss function
To increase U-NetAttentionThe identification capability of the model to the whole infarcted area is realized by introducing DSC (differential scanning calorimetry) Loss as another weight on the basis of Focal Loss, and using the Focal Loss and the Focal LossThe weighting of DSC Loss as a Loss function is defined as follows:
L=αLDSC+βLfocal_loss(formula 4)
Where α and β are used to adjust the weights of the two loss functions.
The calculation formula of Focal local is as follows:
Figure BDA0003434259000000062
wherein, ygtThe true labeled value of the pixel is obtained, y' is the final prediction output through a Sigmoid activation function, alpha and gamma are two adjusting coefficients, wherein alpha is used for balancing positive and negative samples, and gamma is used for adjusting the weight of the samples;
the DSC Loss is calculated as follows:
Figure BDA0003434259000000063
wherein X represents a real sample label region; y denotes a prediction output result region. | X | represents the area of the true sample label region and | Y | represents the area of the prediction output result region.
S302:U-NetAttentionNetwork training
Obtaining medical brain images of the patient according to step S1, including CT perfusion Cerebral Blood Flow (CBF), Cerebral Blood Volume (CBV)2 modal images and Diffusion Weighted Imaging (DWI) image data, inputting the images into an upper computer, then performing unified registration on the images, matching different modal data of the patient by using a comprehensive analysis tool library 3D Slicer with the Diffusion Weighted Imaging (DWI) image as a reference, and obtaining a registered NIFIT format file as U-NetAttentionAn input data set for training and testing of the network; meanwhile, manually labeling an peduncle region image on a Diffusion Weighted Imaging (DWI) image as a label corresponding to an input image, taking 75% of the input data set and the label as a training data set, and taking the rest 25% of the input data set and the label as a test data set;
the training data set and the test data set contain as far as possible image data of different patients.
Using an Adam algorithm as an optimization loss function L, and paying attention to the fact that in order to avoid the deviation of results caused by unreasonable segmentation of a training data set and a test data set, a four-fold cross validation mode is adopted in the experimental process, namely, the training data set is divided into four parts according to patients, three parts of the training data set are used as the training set and the other part of the training data set is used as a validation set each time, the experiment is repeated for four times, and the average result is used as the index of the final evaluation algorithm;
U-NetAttentionthe training of the network is realized based on a Tensorflow framework. Initial learning rate is set to 1e-3The lowest learning rate is 1e-10And a dynamic learning rate reduction is achieved in an exponentially decaying manner according to the number of training rounds. And (4) performing 260 rounds of iteration to obtain a trained pt model file, wherein weights and parameters of the neural network nodes are stored in the pt model file and are used for testing the test data set in the step S303.
S303:U-NetAttentionNetwork testing
Using U-Net in the test phaseAttentionAnd (3) loading the trained pt model file, namely loading the model weight obtained by training, and testing by using the multi-mode image of the test data set which does not participate in the training as U-NetAttentionThe network inputs to obtain a prediction result, and a two-dimensional segmentation result image with higher prediction accuracy than that of the U-Net baseline model is successfully obtained, so that the trained U-Net is verified and obtainedAttentionA network.
S04, online use
Acquiring a medical brain image of a patient, wherein the medical brain image comprises a cerebral blood flow volume (CBF) map, a Cerebral Blood Volume (CBV) map and Diffusion Weighted Imaging (DWI) image data, and inputting the medical brain image into an upper computer; matching different modal data of a patient by using a comprehensive analysis tool library 3D Slicer with DWI images as a reference to obtain a registered NIFIT format file; with the trained U-Net obtained in step S303AttentionSegmenting the registered NIFIT format file to obtain a two-dimensional segmentation result image, and displaying the two-dimensional segmentation result image in an upper computerAnd displaying and outputting the two-dimensional segmentation result image.
Comparative experiment 1:
acquiring medical brain images of the same patient, including a cerebral blood flow volume (CBF) map, a Cerebral Blood Volume (CBV) map and Diffusion Weighted Imaging (DWI) image data, and inputting the images into an upper computer; matching different modal data of a patient by using a comprehensive analysis tool library 3D Slicer with DWI images as a reference to obtain a registered NIFIT format file; then, the doctor inputs the registered NIFIT format files into U-Net respectively according to the actual infarct area marked by the DWI imageAttentionObtaining two-dimensional infarct region segmentation result images by a network and a baseline U-Net network, manually labeling and U-NetAttentionThe segmentation result images of the network and the baseline U-Net network infarct area are shown in FIG. 4, the first column is the image of the actual infarct area labeled by the doctor according to the DWI image, the second column is the segmentation result image of the baseline U-Net, and the third column is the U-Net in the inventionAttentionAnd (4) segmenting a result image of the model. As can be seen from FIG. 4, the division area of the base-line network U-Net is obviously larger than the actual infarction area, and is easy to shift and miss detection, and after the improvement of the long connection attention mechanism, the U-Net in the inventionAttentionThe conditions of missed detection and false detection of the infarct area by the model are reduced, the similarity of the predicted infarct area and the normal area is improved, and the similarity is closer to a real labeling result.
Comparative experiment 2:
data used in the comparison experiment are from desensitized Ischemic Stroke CT perfusion multi-modal brain images (286 slices) provided by a cooperative hospital and public data sets (458 slices) provided by an Ischemic Stroke Lesion Segmentation (Ischemic Stroke Segmentation)2018 competition, different modal data of a patient are matched by using a comprehensive analysis tool library 3D Slicer with DWI images as a reference, and registered NIFIT format files are obtained; and then inputting the registered NIFIT format file into a U-NetAttention network and a baseline U-Net network respectively so as to obtain a two-dimensional infarct area segmentation result image.
The DSC coefficient, Sensitivity (Sensitivity), and Hausdorff Distance (HD for short) are used as performance indexes for comparison, and the calculation method is described as follows:
1) DSC coefficient: setting a pixel set X on the segmented image of the infarct area obtained by model calculation, and a pixel set Y of the labeling image of the real infarct area, wherein 1 represents that the pixel belongs to the infarct area, and 0 represents that the pixel belongs to the normal area, and then calculating DSC coefficients of the two sets as follows:
Figure BDA0003434259000000083
wherein | X | + | Y | is the intersection of the sets X and Y, and | X | + | Y | is the addition of the point sets of the sets X and Y.
2) The Sensitivity (Sensitivity) represents how many pixels belonging to the infarct area are correctly predicted as the true value, and the calculation method is as follows:
Figure BDA0003434259000000081
wherein TP represents that the true value is a pixel belonging to the infarct area and the model is predicted as the infarct area, FP represents that the true value is a pixel belonging to the infarct area and the model is predicted as the normal area
3) The Housdov Distance (HD) is used for measuring the similarity between two point sets, and a pixel set X on an peduncle region segmentation image obtained by model calculation and a pixel set Y of a real peduncle region labeling image are calculated in the following mode:
dH(X,Y)=max{supx∈Xinfy∈Yd(x,y),supy∈Yinfx∈Xd (x, y) } (equation 9)
The DSC coefficient is the most important evaluation index, and when DSCs are close, Sensitivity (Sensitivity) and Hausdorff Distance (HD) indexes are used as supplementary comparisons, and the results of the comparative experiments are as follows:
Figure BDA0003434259000000082
Figure BDA0003434259000000091
the higher the DSC coefficient, the lower the HD value, and the lower the Sensitivity value, the higher the degree of area coincidence between the predicted infarction area and the actual infarction area, and the greater the similarity. From the results in the table, it can be known that the improved attention module enhances the attention of the model to the important region on the feature map through the calculation of the space and the channel weight, and improves the segmentation accuracy of the final infarct region.
Finally, it is also noted that the above-mentioned lists merely illustrate a few specific embodiments of the invention. It is obvious that the invention is not limited to the above embodiments, but that many variations are possible. All modifications which can be derived or suggested by a person skilled in the art from the disclosure of the present invention are to be considered within the scope of the invention.

Claims (4)

1. A multi-mode stroke infarct region image processing method based on an improved U-Net network is characterized by comprising the following steps:
acquiring a brain medical image of a patient and inputting the brain medical image into an upper computer, wherein the brain medical image comprises a cerebral blood flow chart, a cerebral blood volume chart and a diffusion weighted imaging image, and matching 2-mode data of the cerebral blood flow chart and the cerebral blood volume chart of the same patient by using a 3D Slicer in a comprehensive analysis tool library by taking the diffusion weighted imaging image as a reference to obtain an NIFIT format file after registration; using trained U-NetAttentionThe network segments the registered NIFIT format file to obtain a two-dimensional segmentation result image, and displays the output result image in an upper computer;
the U-NetAttentionThe network comprises a U-Net network as a baseline model, an attention mechanism module is introduced at a long connection position of the baseline model for improvement, and a target area is emphasized in a low-level feature map by calculating weights through the feature map of a codec.
2. The image processing method of multi-modal stroke infarct region segmentation based on the improved U-Net network according to claim 1, characterized in that:
the attention mechanism module includes a channel attention weight calculation and a spatial attention weight calculation:
(1) spatial attention weight calculation
The features of the codec are extracted by two 1 × 1 convolutions, respectively, and the attention distribution F and the attention score a are obtained as follows:
F(xl,xh)=σ1(wαxl+bα)+σ1(wβxh+bβ) (formula 1)
A=σ2(wγF(xl,xh)+bγ) (formula 2)
Wherein x islFor the low-level feature map from the encoder, xhFor high level feature maps from the decoder, xlAnd xhIs the input of the attention module, σ1As Relu activation function, σ2Activating a function for Sigmoid, wα、wβ、wγAnd bα、bβ、bγParameters and offsets representing the convolution, respectively;
(2) channel attention weight calculation
Compressing the space size of the whole feature map into 1 × 1 through global average pooling operation, generating the attention weight on the channel through convolution of the channel feature map subjected to global average pooling GAP operation, and multiplying the attention weight by the low-level features to finally realize the attention weight calculation on the channel:
Figure FDA0003434258990000011
where x (i, j) is an element in the feature map at position (i, j), and W and H represent the size of the two-dimensional space of the feature map.
3. The image processing method of multi-modal stroke infarct region segmentation based on the improved U-Net network according to claim 2, characterized in that:
the trained U-NetAttentionThe network acquisition process comprises the following steps:
(1) the loss function L is defined as:
L=αLDSC+βLfocal_loss(formula 4)
Wherein, alpha and beta are the weights of the two loss functions respectively;
Figure FDA0003434258990000021
wherein, ygtThe true labeled value of the pixel is obtained, y' is the final prediction output through a Sigmoid activation function, alpha and gamma are two adjusting coefficients, wherein alpha is used for balancing positive and negative samples, and gamma is used for adjusting the weight of the samples;
Figure FDA0003434258990000022
wherein X represents a real sample label region, Y represents a prediction output result region, | X | represents the area of the real sample label region, | Y | represents the area of the prediction output result region;
(2)、U-NetAttentionnetwork training
Acquiring medical brain images of a patient, including a cerebral blood flow chart, a cerebral blood volume chart and diffusion-weighted imaging image data, matching different modal data of the patient by using a 3D Slicer comprehensive analysis tool library and taking diffusion-weighted imaging images as a reference, and acquiring a registered NIFIT format file as a U-NetAttentionAn input data set for training and testing of the network; manually labeling the peduncle region image on the diffusion weighted imaging modality image to be used as a corresponding label, taking 75% of input data sets and the corresponding labels as training data sets, and taking the rest 25% of the input data sets and the labels as test data sets;
using Adam's algorithm as the bestChanging a loss function L, adopting a verification mode of four-fold cross verification, and taking an average result as an index of a final evaluation algorithm; training initial learning rate is set to le-3The lowest learning rate is le-10And the dynamic learning rate reduction is realized in an exponential decay mode according to the number of training rounds, and 260 rounds of iteration are trained to obtain a trained pt model file;
(3)、U-NetAttentionnetwork testing
Using U-NetAttentionLoading a trained pt model file by a network, and taking a test data set as U-NetAttentionThe network input obtains a segmentation result, thereby verifying and obtaining the trained U-NetAttentionA network.
4. The image processing method of multi-modal stroke infarct region segmentation based on the improved U-Net network according to claim 3, characterized in that:
the U-Net network includes an encoder, a decoder, and a long connection.
CN202111606856.XA 2021-12-27 2021-12-27 Multimode cerebral apoplexy infarct region image processing method based on improved U-Net network Active CN114359194B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202111606856.XA CN114359194B (en) 2021-12-27 2021-12-27 Multimode cerebral apoplexy infarct region image processing method based on improved U-Net network

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202111606856.XA CN114359194B (en) 2021-12-27 2021-12-27 Multimode cerebral apoplexy infarct region image processing method based on improved U-Net network

Publications (2)

Publication Number Publication Date
CN114359194A true CN114359194A (en) 2022-04-15
CN114359194B CN114359194B (en) 2024-07-12

Family

ID=81100550

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202111606856.XA Active CN114359194B (en) 2021-12-27 2021-12-27 Multimode cerebral apoplexy infarct region image processing method based on improved U-Net network

Country Status (1)

Country Link
CN (1) CN114359194B (en)

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN117422788A (en) * 2023-12-19 2024-01-19 英瑞云医疗科技(烟台)有限公司 Method for generating DWI image based on CT brain stem image

Citations (11)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN109658401A (en) * 2018-12-14 2019-04-19 上海商汤智能科技有限公司 Image processing method and device, electronic equipment and storage medium
CN110223285A (en) * 2019-06-13 2019-09-10 中南大学湘雅二医院 Imaging results prediction technique and system based on animal migration and neural network model
CN110322528A (en) * 2019-06-26 2019-10-11 浙江大学 Nuclear magnetic resonance brain image reconstructing blood vessel method based on 3T, 7T
US20200381096A1 (en) * 2019-06-03 2020-12-03 The Board Of Trustees Of The Leland Stanford Junior University Methods of Predicting Disorder Progression for Control Arms Within an Experimental Trial
CN112150449A (en) * 2020-09-29 2020-12-29 太原理工大学 Cerebral apoplexy focus segmentation method and system
CN112418027A (en) * 2020-11-11 2021-02-26 青岛科技大学 Remote sensing image road extraction method for improving U-Net network
CN112767374A (en) * 2021-01-27 2021-05-07 天津理工大学 Alzheimer disease focus region semantic segmentation algorithm based on MRI
CN113095382A (en) * 2021-03-30 2021-07-09 浙江大学 Interpretable tuberculosis classification network identification method based on CT image
CN113177943A (en) * 2021-06-29 2021-07-27 中南大学 Cerebral apoplexy CT image segmentation method
WO2021184817A1 (en) * 2020-03-16 2021-09-23 苏州科技大学 Method for segmenting liver and focus thereof in medical image
CN113627564A (en) * 2021-08-23 2021-11-09 李永鑫 Deep learning-based CT medical image processing model training method and diagnosis and treatment system

Patent Citations (11)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN109658401A (en) * 2018-12-14 2019-04-19 上海商汤智能科技有限公司 Image processing method and device, electronic equipment and storage medium
US20200381096A1 (en) * 2019-06-03 2020-12-03 The Board Of Trustees Of The Leland Stanford Junior University Methods of Predicting Disorder Progression for Control Arms Within an Experimental Trial
CN110223285A (en) * 2019-06-13 2019-09-10 中南大学湘雅二医院 Imaging results prediction technique and system based on animal migration and neural network model
CN110322528A (en) * 2019-06-26 2019-10-11 浙江大学 Nuclear magnetic resonance brain image reconstructing blood vessel method based on 3T, 7T
WO2021184817A1 (en) * 2020-03-16 2021-09-23 苏州科技大学 Method for segmenting liver and focus thereof in medical image
CN112150449A (en) * 2020-09-29 2020-12-29 太原理工大学 Cerebral apoplexy focus segmentation method and system
CN112418027A (en) * 2020-11-11 2021-02-26 青岛科技大学 Remote sensing image road extraction method for improving U-Net network
CN112767374A (en) * 2021-01-27 2021-05-07 天津理工大学 Alzheimer disease focus region semantic segmentation algorithm based on MRI
CN113095382A (en) * 2021-03-30 2021-07-09 浙江大学 Interpretable tuberculosis classification network identification method based on CT image
CN113177943A (en) * 2021-06-29 2021-07-27 中南大学 Cerebral apoplexy CT image segmentation method
CN113627564A (en) * 2021-08-23 2021-11-09 李永鑫 Deep learning-based CT medical image processing model training method and diagnosis and treatment system

Non-Patent Citations (3)

* Cited by examiner, † Cited by third party
Title
AN YANG等: "CT Images Recognition of Pulmonary Tuberculosis Based on Improved Faster RCNN and U-Net", 《2019 10TH INTERNATIONAL CONFERENCE ON INFORMATION TECHNOLOGY IN MEDICINE AND EDUCATION 》, 23 January 2020 (2020-01-23) *
TAO SONG: "Generative Model-Based Ischemic Stroke Lesion Segmentation", 《IMAGE AND VIDEO PROCESSING》, 6 June 2019 (2019-06-06) *
何洪林;谢峻;李毅;雷跳;钱俊;马桃林;: "基于裸眼3D视觉效果的脑血管CT图像快速三维重建方法研究", 医疗卫生装备, no. 09, 15 September 2015 (2015-09-15) *

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN117422788A (en) * 2023-12-19 2024-01-19 英瑞云医疗科技(烟台)有限公司 Method for generating DWI image based on CT brain stem image
CN117422788B (en) * 2023-12-19 2024-03-26 英瑞云医疗科技(烟台)有限公司 Method for generating DWI image based on CT brain stem image

Also Published As

Publication number Publication date
CN114359194B (en) 2024-07-12

Similar Documents

Publication Publication Date Title
Khan et al. Deep neural architectures for medical image semantic segmentation
WO2016192612A1 (en) Method for analysing medical treatment data based on deep learning, and intelligent analyser thereof
CN107016395B (en) Identification system for sparsely expressed primary brain lymphomas and glioblastomas
CN112086197B (en) Breast nodule detection method and system based on ultrasonic medicine
Xie et al. Computer‐Aided System for the Detection of Multicategory Pulmonary Tuberculosis in Radiographs
EP3654343A1 (en) Application of deep learning for medical imaging evaluation
CN111755118B (en) Medical information processing method, device, electronic equipment and storage medium
CN114782307A (en) Enhanced CT image colorectal cancer staging auxiliary diagnosis system based on deep learning
EP2996058A1 (en) Method for automatically generating representations of imaging data and interactive visual imaging reports
Li et al. Optical coherence tomography vulnerable plaque segmentation based on deep residual U-Net
CN112508884A (en) Comprehensive detection device and method for cancerous region
CN116091490A (en) Lung nodule detection method based on YOLOv4-CA-CBAM-K-means++ -SIOU
CN116433654A (en) Improved U-Net network spine integral segmentation method
JP2024528381A (en) Method and system for automatically tracking and interpreting medical image data
CN110570425A (en) Lung nodule analysis method and device based on deep reinforcement learning algorithm
CN114359194B (en) Multimode cerebral apoplexy infarct region image processing method based on improved U-Net network
Saeed et al. 3D MRU-Net: A novel mobile residual U-Net deep learning model for spine segmentation using computed tomography images
CN113902738A (en) Heart MRI segmentation method and system
Singha Deo et al. Supremacy of attention based convolution neural network in classification of oral cancer using histopathological images
Zhang et al. Femoral image segmentation based on two-stage convolutional network using 3D-DMFNet and 3D-ResUnet
CN117711576A (en) Method and system for providing a template data structure for medical reports
Quy et al. Multi-view digital mammography mass classification: A convolutional neural network model approach
CN116631584A (en) General medical image report generation method and system, electronic equipment and readable storage medium
CN113362350B (en) Method, device, terminal equipment and storage medium for segmenting cancer medical record image
WO2020099941A1 (en) Application of deep learning for medical imaging evaluation

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant