CN117392137B - Intracranial aneurysm image detection method, system, equipment and medium - Google Patents

Intracranial aneurysm image detection method, system, equipment and medium Download PDF

Info

Publication number
CN117392137B
CN117392137B CN202311703651.2A CN202311703651A CN117392137B CN 117392137 B CN117392137 B CN 117392137B CN 202311703651 A CN202311703651 A CN 202311703651A CN 117392137 B CN117392137 B CN 117392137B
Authority
CN
China
Prior art keywords
module
image
loss
intracranial aneurysm
mlp
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN202311703651.2A
Other languages
Chinese (zh)
Other versions
CN117392137A (en
Inventor
刘庆一
王可心
王以军
白培瑞
韩超
杨光
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Shandong University of Science and Technology
Original Assignee
Shandong University of Science and Technology
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Shandong University of Science and Technology filed Critical Shandong University of Science and Technology
Priority to CN202311703651.2A priority Critical patent/CN117392137B/en
Publication of CN117392137A publication Critical patent/CN117392137A/en
Application granted granted Critical
Publication of CN117392137B publication Critical patent/CN117392137B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/0002Inspection of images, e.g. flaw detection
    • G06T7/0012Biomedical image inspection
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/0464Convolutional networks [CNN, ConvNet]
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/08Learning methods
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/10Segmentation; Edge detection
    • G06T7/12Edge-based segmentation
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/20Special algorithmic details
    • G06T2207/20081Training; Learning
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/20Special algorithmic details
    • G06T2207/20084Artificial neural networks [ANN]
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/30Subject of image; Context of image processing
    • G06T2207/30004Biomedical image processing
    • G06T2207/30096Tumor; Lesion
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/30Subject of image; Context of image processing
    • G06T2207/30004Biomedical image processing
    • G06T2207/30101Blood vessel; Artery; Vein; Vascular
    • YGENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
    • Y02TECHNOLOGIES OR APPLICATIONS FOR MITIGATION OR ADAPTATION AGAINST CLIMATE CHANGE
    • Y02TCLIMATE CHANGE MITIGATION TECHNOLOGIES RELATED TO TRANSPORTATION
    • Y02T10/00Road transport of goods or passengers
    • Y02T10/10Internal combustion engine [ICE] based vehicles
    • Y02T10/40Engine management systems

Landscapes

  • Engineering & Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • Theoretical Computer Science (AREA)
  • General Physics & Mathematics (AREA)
  • General Health & Medical Sciences (AREA)
  • Health & Medical Sciences (AREA)
  • Molecular Biology (AREA)
  • Computational Linguistics (AREA)
  • Software Systems (AREA)
  • Mathematical Physics (AREA)
  • General Engineering & Computer Science (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Artificial Intelligence (AREA)
  • Biomedical Technology (AREA)
  • Biophysics (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Data Mining & Analysis (AREA)
  • Evolutionary Computation (AREA)
  • Computing Systems (AREA)
  • Medical Informatics (AREA)
  • Nuclear Medicine, Radiotherapy & Molecular Imaging (AREA)
  • Radiology & Medical Imaging (AREA)
  • Quality & Reliability (AREA)
  • Magnetic Resonance Imaging Apparatus (AREA)

Abstract

The invention belongs to the technical field of medical image processing, and particularly discloses a method, a system, equipment and a medium for detecting intracranial aneurysm images. The invention improves the extraction of the deep context feature of the image, the edge of the shallow layer of the image, the outline and other feature information by introducing the context perception pyramid feature extraction module, the channel attention module and the multi-scale convolution module on the basis of the 3D U-Net model, provides more depth features with channel weight information for the decoder end, improves the global learning capacity of the model, is more suitable for the detection task of the small target of the intracranial aneurysm, and effectively improves the detection performance of the intracranial aneurysm. The lightweight 3D MAU-Net detection model based on the multi-layer perceptron provided by the invention has fewer parameters, higher training speed and lower calculation complexity on the premise of ensuring that the detection performance is not obviously affected.

Description

Intracranial aneurysm image detection method, system, equipment and medium
Technical Field
The invention belongs to the technical field of medical image processing, and particularly relates to a method, a system, equipment and a medium for detecting intracranial aneurysm images.
Background
Intracranial aneurysms have high mortality due to their special location of incidence, and early discovery of uncracked intracranial aneurysms is the best means to reduce mortality. At present, the 3D TOF-MRA (Three Dimensional Timeof Flight Magnetic Resonance Angiography, three-dimensional time-leaping method magnetic resonance vascular imaging) technology has great application potential in intracranial aneurysm examination and diagnosis, does not use contrast agent, and has the advantages of safety and noninvasive property. However, when the 3D TOF-MRA image is used for aneurysm detection, the problems of insensitivity to small intracranial aneurysms, long training time, high memory occupation and the like exist, meanwhile, the number of layers of the 3D MRA image is large, the manual film reading workload of doctors is greatly increased, and the sensitivity of the diagnosis of radiologists can be reduced due to high-intensity film reading work, so that a series of problems such as missed diagnosis and misdiagnosis are caused. The MIP (Maximum Intensity Projection ) image generated by the rotation projection of the 3D TOF-MRA image largely retains the density information of the original image, but the quality and reliability of the MIP image mainly depend on the quality and parameter selection of the original data, which requires professional skills and experience, and if the parameter selection is poor, the acquired related information amount is more limited.
In recent years, deep learning models such as Convolutional Neural Networks (CNNs), full Convolutional Networks (FCNs), and the like have exhibited excellent performance in medical image detection and segmentation. For the study of the deep learning algorithm of the intracranial aneurysm, the network model is constructed as a primary factor considered for improving the detection performance of the intracranial aneurysm, because in the study of the intracranial aneurysm on a 3D image, the algorithm is realized by adopting a basic structure of an encoder and a decoder, and the improvement is usually developed around the thought of adjusting the network structure, modifying the size of a convolution kernel, extracting and fusing multi-scale features, adding an attention mechanism, improving a loss function and the like, so that the network is helped to learn the depth features with finer granularity through the improvement, and the detection performance of the algorithm is improved. In addition, part of researchers also combine the traditional algorithm with the deep learning algorithm, and based on the original deep learning model, the traditional algorithm is embedded into the whole detection flow in a pre-processing or post-processing mode, so that the detection precision of the algorithm is improved. Still other researchers combine two or more lightweight deep learning models and propose a multi-stage based learning strategy to achieve fine segmentation and detection of intracranial aneurysms.
In conclusion, the appearance of computer-aided diagnosis provides an important reference for doctor diagnosis, and the existing intracranial aneurysm aided diagnosis technology is not mature enough, and has the problems of long detection time, insufficient accuracy and the like.
Researchers have made a lot of effort to improve the sensitivity and accuracy of intracranial aneurysm detection models. The detection model is designed based on a convolution network, so that the detection sensitivity of the detection model to the intracranial aneurysm is very high, an encoder and decoder structure similar to the Unet is proposed by Geng et al, and the detection model has higher detection rate to the intracranial aneurysm of the 3D TOF-MRA image, and is suitable for clinical screening of the intracranial aneurysm. Claux et al propose a two-stage Unet network detection model that concatenates two Unets, wherein the first Unet model is used to segment the cerebral vessels and generate a binary image, and then the segmented binary image is sent to the second Unet model for training, finally, the detection of intracranial aneurysms is realized, the sensitivity of the algorithm reaches 78%, and the recognition rate of positive samples reaches 62%. Although the deep learning method has higher detection accuracy, it also brings more calculation cost, and this defect is more remarkable in the case of 3D medical images.
Disclosure of Invention
The invention aims to provide an intracranial aneurysm image detection method, which utilizes a 3D MAU-Net detection model (3D Multi-scale attention U-Net, namely a Multi-scale attention detection model) to perform feature extraction on an original 3D TOF-MRA image, so that original features of the image are reserved to the greatest extent, the model is enabled to pay more attention to segmentation of small targets, the accuracy of intracranial aneurysm detection and segmentation is improved, and a linear network is used for replacing part of convolution modules of the original model, so that the complexity of the model is reduced and the training time of the model is shortened on the premise that the detection performance of the model is basically unchanged.
In order to achieve the above purpose, the invention adopts the following technical scheme:
an intracranial aneurysm image detection method comprises the following steps:
step 1, preprocessing original 3D TOF-MRA;
and 2, building a 3D MAU-Net detection model, sending the preprocessed image into the 3D MAU-Net detection model, and detecting the intracranial aneurysm of the image to obtain a segmentation result of the intracranial aneurysm.
In addition, on the basis of the intracranial aneurysm image detection method, the invention also provides an intracranial aneurysm image detection system which is suitable for the intracranial aneurysm image detection method, and the intracranial aneurysm image detection system adopts the following technical scheme:
an intracranial aneurysm image detection system comprising:
the preprocessing module is used for preprocessing the original 3D TOF-MRA;
and the intracranial aneurysm identification module is used for building a 3D MAU-Net detection model, sending the preprocessed image into the 3D MAU-Net detection model, and carrying out intracranial aneurysm detection on the image to obtain an intracranial aneurysm segmentation result.
In addition, on the basis of the intracranial aneurysm image detection method, the invention further provides computer equipment which comprises a memory and one or more processors.
The memory stores executable codes, and the processor is used for realizing the steps of the intracranial aneurysm image detection method when executing the executable codes.
Furthermore, on the basis of the intracranial aneurysm image detection method, the invention further provides a computer readable storage medium on which a program is stored.
The program, when executed by a processor, is configured to implement the steps of the intracranial aneurysm image detection method described above.
The invention has the following advantages:
the invention provides a 3D MAU-Net detection model, which is obtained by introducing a context perception pyramid feature extraction module, a channel attention module and a multi-scale convolution module on the basis of a 3D U-Net model, so that the extraction of feature information such as deep image context features, edges of shallow images and contours is deepened, more depth features with channel weight information are provided for a decoder end, the global learning capacity of the model is improved, the model is more suitable for intracranial aneurysm small target detection tasks, and the detection performance of intracranial aneurysms is effectively improved. Aiming at the common problems of the U-Net network, namely the characteristics of overlarge parameter quantity, overlong model training time and overhigh requirement on computer performance, and can not be well applied to the hospital nursing point for rapid image segmentation, the lightweight 3D MAU-Net detection model based on the multi-layer perceptron has fewer parameters, higher training speed and lower calculation complexity on the premise of ensuring that the detection performance is not obviously influenced.
Drawings
FIG. 1 is a flow chart of an automatic intracranial aneurysm detection model based on an attention mechanism and a multi-layer perceptron in an embodiment of the invention.
Fig. 2 is a schematic diagram of a network architecture of a 3D MAU-Net detection model constructed in an embodiment of the present invention.
Fig. 3 is a schematic diagram of a PCA module according to an embodiment of the present invention.
FIG. 4 is a schematic diagram of hole convolution in an embodiment of the present invention.
Fig. 5 is a schematic diagram of an MC module according to an embodiment of the invention.
FIG. 6 is a schematic diagram of a 3D Tokenized MLP in an embodiment of the invention.
Fig. 7 is a schematic diagram of a deep convolution in an embodiment of the present invention.
Detailed Description
The invention is described in further detail below with reference to the attached drawings and detailed description:
example 1
In this embodiment 1, an intracranial aneurysm image detection method is described, which proposes feature extraction of an original 3D TOF-MRA image, so as to preserve original features of the image to the greatest extent, and improve detection performance of the intracranial aneurysm. And the linear network is used for replacing part of convolution modules in the original model, so that the complexity of the model is reduced and the training time of the model is shortened on the premise of ensuring that the detection performance of the model is basically unchanged. Therefore, the invention provides an intracranial aneurysm automatic detection model based on an attention mechanism and a Multi-layer perceptron, firstly, a PCA module (Pyramid channel attention) and a MC module (Multi-scale convolution) are added in 3D U-Net to deeply extract features of different layers, then, a decoder is used for carrying out feature fusion among different layers, and finally, a final intracranial aneurysm detection result is obtained through deep supervision and guidance network training. Based on the model, the invention proposes to replace the original partial convolution module by a three-dimensional labeling multi-layer perceptron module (3D Tokenized MLP) so as to reduce the complexity of the model.
As shown in fig. 1, the intracranial aneurysm image detection method in the embodiment includes the following steps:
and step 1, preprocessing the original 3D TOF-MRA.
The preprocessing operations include resampling, data normalization, and random data enhancement.
Resampling is to normalize voxels of different sizes in the original image to the same size, setting the median pixel pitch to [0.7,0.43,0.43].
The data normalization is to scale data with different sizes into the same data interval through processing, and the calculation formula is as follows:
wherein the method comprises the steps ofIs the standard deviation of the two-dimensional image,as an average value of the values,a voxel value representing each point is represented,representing the normalized score.
Random data enhancement: when the algorithm performs training experiments, random data enhancement can perform random elastic deformation, random rotation and random scale transformation on the data in real time.
And 2, building a 3D MAU-Net detection model, sending the preprocessed image into the 3D MAU-Net detection model, and detecting the intracranial aneurysm of the image to obtain a segmentation result of the intracranial aneurysm.
The original TOF-MRA image is sent to the 3D MAU-Net detection model to obtain a segmentation result, so that the original characteristics of the image can be fully learned, and the detection performance of intracranial aneurysms is effectively improved.
The invention provides an intracranial aneurysm automatic detection model combining an attention mechanism with a 3D U-Net model in deep learning, and provides a pyramid channel attention module, a multi-scale convolution module and a 3D Tokenized MLP module.
Specifically, the 3D MAU-Net detection model is based on the traditional 3D U-Net model, a context awareness pyramid feature extraction module CPFE, a channel attention module CA and a multi-scale convolution module MC are introduced, extraction of feature information of deep context features, edges and contours of images is achieved, and in addition, the invention further provides that a three-dimensional labeling MLP (namely a 3D processed MLP) module is used for replacing a part of convolution module, so that the complexity of the 3D MAU-Net detection model is reduced fully on the premise that the detection performance is basically unchanged. The method can solve the problems that the existing manual interpretation method is easily influenced by subjective experience of doctors and wastes time and labor, and further improves the accuracy and efficiency of intracranial aneurysm detection.
After the step 1 is completed, the intracranial aneurysm image is pre-cut, and then is sent into an automatic intracranial aneurysm detection model based on an attention mechanism and a multi-layer perceptron, namely a 3D MAU-Net detection model. The detection model takes a 3D U-Net structure as a main network, is improved by introducing a PCA and MC functional module and a 3D enabled MLP module, so that the detection model is suitable for intracranial aneurysm detection, and the network architecture of the 3D MAU-Net detection model is shown as a figure 2 and comprises an encoder, a PCA module, an MC module, a decoder, a 3D enabled MLP module and a depth supervision module;
the processing flow of the preprocessed 3D TOF-MRA image in the 3D MAU-Net detection model is as follows:
firstly, sending the preprocessed 3D TOF-MRA image to an encoder, performing common convolution and residual connection to extract image features, and inputting the features extracted by the encoder to a 3D Tokenized MLP module to perform shifting MLP operation;
the encoder outputs five layers of features f1, f2, f3, f4 and f5 together with the 3D Tokenized MLP;
the advanced features f1 and f2 are subjected to feature learning through the MC module, extraction of image edge information is enhanced, and feature images of f1 'and f 2' are output; extracting the advanced features f3, f4 and f5 by using a PCA module, outputting f3 ', f4 ' and f5 ' feature images to obtain more complete bottom features, and then passing through a 3D enabled MLP module and a decoder module;
and f3 ' feature images and the upper layer S4 of the decoder are fused and input into the third layer decoder to obtain S3, f2 ' feature images and the upper layer S3 of the decoder are fused and input into the second layer decoder to obtain S2, f1 ' feature images and the upper layer S2 of the decoder are fused and input into the third layer decoder to obtain S1, the upper three layers of the decoder output S1, S2 and S3 are guided by a mixed loss function to train through a depth supervision module, and a final intracranial aneurysm detection result is obtained.
The encoder is formed by pooling two Conv3d convolution layers and a maximum value, the number of characteristic channels is doubled after each downsampling, and five layers of characteristics of outputs f1, f2, f3, f4 and f5 are activated through convolution and LeakyReLU.
Since the main improvement of 3D U-Net is based on feature fusion of the decoder, features extracted prior to entry into the decoder are critical for intracranial aneurysm detection. When the channel dimension performs feature learning, importance degrees of all channels are emphasized through different module structures, and feature extraction capacity of a deep learning network is improved at an encoder end, so that overall detection performance of a model is improved.
The PCA module structure is shown in fig. 3, the PCA module is composed of a context perception pyramid feature extraction module CPFE module and a channel attention module CA, the CPFE module firstly convolves a feature image from an encoder by utilizing cavities with convolution kernel sizes of 3 multiplied by 3 and different expansion rates (3, 5 and 7 respectively) to obtain multi-scale multi-receptive field features, then fuses the feature image obtained by the cavity convolutions with the feature image with the convolution kernel size of 1 multiplied by 1 to obtain an output feature layer P, then sends the fused feature image P into the CA module, gives different weights to different channels by utilizing the CA module, fully utilizes effective information, and finally inputs the output feature information into a decoder of the same layer. Wherein, definition of the hole convolution is as follows:
wherein the method comprises the steps ofIs an input signal which is provided with a signal,is an output signal which is provided with a signal,representing a length ofR corresponds to the samplingExpansion ratio of (2); in standard convolution, r=1.
Hole convolution as shown in FIG. 4, when the CA module assigns a larger weight to a channel that plays an important role in image detection, it is assumed thatFor PResetting, then transposing the reset P, and multiplying the transposed P and the reset P by matrixFinally, obtaining the channel attention map by using softmax
Wherein,representing the effect of the ith channel on the jth channel; thereafter, matrix multiplication is performed between X transpose and reset of P, and the resulting result is reset to
Finally, the result is combinedMultiplying and performing element-by-element summation operation on P to obtain final output
Wherein,representing the output characteristics of the channel,the importance weight representing the ith feature dimension,a bias term representing a jth neuron;representing the scale parameters, the weights are learned gradually starting from 0.
The formula shows that the final feature of each channel is a weighted sum of the features of all channels and the original features, which is beneficial to improving the discrimination between the features.
As shown in fig. 5, the MC module performs a convolution operation on a feature map obtained at a certain time by using 3 x 3 convolution, 5 x 5 convolution, and 7 x 7 convolution, while simultaneously convolving with 1 x 1 convolution, and obtaining a new feature map with different feature information, and then fusing the different feature maps to obtain a final output feature map. The multi-scale convolution does not change the size of the original feature map, but only enriches the features of the image through convolution operations of different convolutions, and extracts interesting feature information from the image from a global view angle, so that the performance of the model is improved. Because the multi-scale convolution fully utilizes convolution kernels with different sizes, not only can abundant context information be obtained in the feature extraction process, but also the edge contour information of the image can be better reserved.
As shown in FIG. 6, the 3D Tokenized MLP is a three-dimensional extension of the 2D Tokenized MLP block. In the 3D Tokenized MLP module, the feature map is processed as follows:
transmitting the feature map to a first Shifted multi-layer perceptron (Shifted MLP) to carry out width-wise shifting MLP operation, and carrying out depth convolution (DWConv) on the feature information after shifting; after a depth convolution, using a GELU activation function to output to the following Shifted MLP, respectively performing a shifting MLP operation on the height and a shifting MLP operation on the depth on the second shifting multi-layer perceptron Shifted MLP and the third shifting multi-layer perceptron Shifted MLP, performing the depth convolution again, and performing LN normalization on the feature map; here, the 3D token MLP uses a residual connection to linearly add the initial feature map token to the shifted MLP feature map and pass the output features to the next co-layer PCA module.
The mathematical computation process in the 3D Tokenized MLP module can be calculated as:
wherein X is shift Representing the output shifted in the X direction, Y shift Representing the output shifted in the Y direction, Z shift Representing the output shifted in the Z direction; token represents a marker location; t represents the original feature map token, H represents the height, W represents the width, D represents the depth, DWconv represents the depth convolution, LN represents the layer normalization, GELU is the activation function, X represents the output of the X-direction through the Tokenized MLP, Y represents the output of the Y-direction through the Tokenized MLP, and Z is the current output of the module.
When the depth convolution carries out convolution operation, the number of convolution kernels is consistent with the number of channels of the upper layer, and the number of output characteristic diagrams is ensured to be consistent with the number of input characteristic diagrams. As shown in fig. 7.
The decoder end is composed of functional modules, and the decoding stage and the encoding stage are mirror images.
In the decoding stage, the resolution is doubled once per upsampling. And the information of the encoding stage and the information of the decoding stage are combined through the jump connection, so that the image characteristic information quantity contained in the network is increased.
After the 3D MAU-Net detection model is built, the model needs to be trained, and the training process is as follows:
and sending the data set of the TOF-MRA image obtained after pretreatment into a 3D MAU-Net detection model for training.
The model complex training optimization algorithm adopts an ADAM algorithm with an initial learning rate lr of 0.001;
weight decay index of 10 -5 Setting a moving index average valueThe training loss is monitored every 30 iteration cycles as long asDoes not reduce by 5X 10 -4 Lr is reduced by a factor of 5; the network adopts a mixed loss function;
the number of the batch sizes is 30, the maximum iteration period of the network training is 300, each iteration period comprises 150 iterations, and when the iteration period reaches the maximum value or lr is less than 10 -8 The network stops training;
and (5) saving the model weight at the highest accuracy and lower loss under the verification set at present, and completing the training process.
In the training process of the 3D MAU-Net detection model, a Focal loss function and a Lovasz-Softmax loss function are adopted to jointly represent global loss and local loss, and the two losses are combined to be expressed as: loss=loss F +loss L
Wherein loss represents a loss function of the 3D MAU-Net detection model, loss F Table Focal loss, loss L Representing the lovassz-Softmax loss. Focal loss is used to deal with extreme foreground-background class imbalance problems, focusing loss on difficult-to-separate samples, starting from the difficulty of sample classification, defined as follows:
wherein y is i For input instance x i True category, p i To predict input instance x i Probability of belonging to real class, for extremely unbalanced class, addTo predict samples with low probability, and to switch positive samplesThe fluence becomes large, gamma is an adjustable factor, N represents the total number of samples, and alpha represents a weight parameter.
Lovasz-Softmax can promote the intersection ratio IoU score of the segmentation result, and IoU score is also called Jaccard index, which is defined as:
wherein the method comprises the steps ofRepresenting standard results, y represents network prediction results, so Jaccard loss is defined as:
optimizing Jaccard loss by Lovasz-Softmax is often used to fill in the gaps in Jaccard loss partitioning, recover small objects, and form a more reasonable global partitioning, defined as follows:
wherein p represents the number of pixels;representing the first of the network predictions𝑖Value of individual element->Represents the first of the standard results𝑖The values of the individual elements.
Furthermore, in order to verify the effectiveness of the method of the invention, the following experiments are also presented:
the data used in this experiment was provided by published data in a hospital and MICCAI2020 aneurysm detection and segmentation (ADAM) challenge, and the experimental data were 1043 TOF-MRA images, 500 negative samples and 543 positive samples.
S1, preprocessing a data set:
the label used in the experiment is a binarized image, and since the negative sample does not have the original label, a label image which is consistent with the original image in size and has a voxel value of 0 needs to be generated in the process of constructing the data set. All subject data were then divided into large, medium, small and non-intracranial aneurysms patients, with intracranial aneurysms greater than 7mm in diameter labeled large aneurysms, medium aneurysms between 3mm and 7mm in diameter labeled medium aneurysms, small aneurysms less than 3mm in diameter labeled small aneurysms, and subjects not having an intracranial aneurysm labeled normal human specimens. And split it into training and testing sets.
S2, sending the training set of the TOF-MRA image obtained after pretreatment into a lightweight 3D MAU-Net detection model based on MLP for training.
The network training optimization algorithm provided by the model adopts an ADAM algorithm with initial learning rate of lr=0.001, and weight attenuation index is 10 -5 Setting a moving index average valueThe training loss is monitored every 30 iteration cycles as long asDoes not reduce by 5X 10 -4 Lr is reduced by a factor of 5. The network employs a hybrid loss function. The number of the batch sizes is 30, the maximum iteration period of the network training is 300, each iteration period comprises 150 iterations, and when the iteration period reaches the maximum value or lr is less than 10 -8 The network stops training. And saving the model weight at the time of highest accuracy and lower loss under the verification set.
And S3, sending the test set into a lightweight 3D MAU-Net to obtain an intracranial aneurysm detection result. And carrying out visualization processing on the image to obtain a visualization result. The detection model outputs intracranial aneurysm binarization images, the original MIP images and the corresponding binarization images are simultaneously opened by amide software to splice the images, and the aneurysm images are marked by different colors, so that the observation is convenient.
Example 2
This embodiment 2 describes an intracranial aneurysm image detection system based on the same inventive concept as the intracranial aneurysm image detection method described in embodiment 1 above.
Specifically, an intracranial aneurysm image detection system comprising:
the preprocessing module is used for preprocessing the original 3D TOF-MRA;
and the intracranial aneurysm identification module is used for building a 3D MAU-Net detection model, sending the preprocessed image into the 3D MAU-Net detection model, and carrying out intracranial aneurysm detection on the image to obtain an intracranial aneurysm segmentation result.
It should be noted that, in the intracranial aneurysm image detection system, the implementation process of the functions and roles of each functional module is specifically detailed in the implementation process of the corresponding steps in the method in the above embodiment 1, and will not be described herein.
Example 3
Embodiment 3 describes a computer apparatus for implementing the intracranial aneurysm image detection method described in embodiment 1.
In particular, the computer device includes a memory and one or more processors. Executable code is stored in the memory for implementing the steps of the intracranial aneurysm image detection method described above when the executable code is executed by the processor.
In this embodiment, the computer device is any device or apparatus having data processing capability, which is not described herein.
Example 4
Embodiment 4 describes a computer-readable storage medium for implementing the intracranial aneurysm image detection method described in embodiment 1.
Specifically, the computer-readable storage medium in this embodiment 4 has stored thereon a program for implementing the steps of the intracranial aneurysm image detection method described above when executed by a processor.
The computer readable storage medium may be an internal storage unit of any device or apparatus having data processing capability, such as a hard disk or a memory, or may be an external storage device of any device having data processing capability, such as a plug-in hard disk, a Smart Media Card (SMC), an SD Card, a Flash memory Card (Flash Card), or the like, which are provided on the device.
The foregoing description is, of course, merely illustrative of preferred embodiments of the present invention, and it should be understood that the present invention is not limited to the above-described embodiments, but is intended to cover all modifications, equivalents and alternatives falling within the spirit and scope of the present invention as defined by the appended claims.

Claims (8)

1. An intracranial aneurysm image detection method is characterized by comprising the following steps:
step 1, preprocessing original 3D TOF-MRA;
step 2, building a 3D MAU-Net detection model, sending the preprocessed image into the 3D MAU-Net detection model, and detecting intracranial aneurysms on the image to obtain intracranial aneurysms segmentation results;
the built 3D MAU-Net detection model comprises an encoder, a PCA module, an MC module, a decoder, a 3D Tokenized MLP module and a depth supervision module;
the processing flow of the preprocessed 3D TOF-MRA image in the 3D MAU-Net detection model is as follows:
firstly, sending the preprocessed 3D TOF-MRA image to an encoder, performing common convolution and residual connection to extract image features, and inputting the features extracted by the encoder to a 3D Tokenized MLP module to perform shifting MLP operation;
the encoder outputs five layers of features f1, f2, f3, f4 and f5 together with the 3D Tokenized MLP;
the advanced features f1 and f2 are subjected to feature learning through the MC module, extraction of image edge information is enhanced, and feature images of f1 'and f 2' are output; extracting the advanced features f3, f4 and f5 by using a PCA module, outputting f3 ', f4 ' and f5 ' feature images to obtain more complete bottom features, and then passing through a 3D enabled MLP module and a decoder;
three layers of outputs S1, S2 and S3 on the decoder pass through a depth supervision module, and a mixed loss function guides a 3D MAU-Net detection model network to train so as to obtain a final intracranial aneurysm detection result;
the PCA module consists of a context awareness pyramid feature extraction module CPFE and a channel attention module CA; the CPFE module firstly convolves the feature map from the encoder by utilizing cavities with convolution kernel sizes of 3 multiplied by 3 and different expansion rates to obtain multi-scale multi-receptive field features, then fuses the feature map obtained by the cavity convolution with different expansion rates with the feature map with convolution kernel sizes of 1 multiplied by 1 to obtain an output feature layer P, then sends the fused feature map P into the CA module, gives different weights to different channels by utilizing the CA module, and finally inputs the output feature information into a decoder of the same layer.
2. The method for detecting an image of an intracranial aneurysm as claimed in claim 1,
the definition of the hole convolution is as follows:
wherein f [ i ]]Is an input signal g [ i ]]Is an output signal, h [ l ]]Representing a filter of length l, r corresponding to a sample f [ i ]]Expansion ratio of (2); suppose P ε R C×H×W R to P C×H Resetting, transposing the reset P, performing matrix multiplication on the transposed P and the reset P, and finally obtaining a channel attention map P epsilon R by using softmax C×C
Wherein x is ji Representing the effect of the ith channel on the jth channel; thereafter, matrix multiplication is performed between X transpose and reset of P, and the resulting result is reset to R C×H×W
Finally multiplying the result by betaAnd performs element-by-element summation operation on P to obtain final output A E R C×H×W
Wherein A is j Representing channel output characteristics, P i Importance weight, P, representing the ith feature dimension j A bias term representing a jth neuron; beta represents a scale parameter, and the weights are learned gradually from 0.
3. The method for detecting an image of an intracranial aneurysm as claimed in claim 1,
in a 3D processed MLP module, transmitting the feature map to a first Shifted MLP (multi-layer perceptron) to carry out width-wise shift MLP operation, and carrying out depth convolution DWConv on the feature information after shifting;
after a depth convolution, using a GELU activation function to output to the following Shifted MLP, respectively performing a shifting MLP operation on the height and a shifting MLP operation on the depth on the second shifting multi-layer perceptron Shifted MLP and the third shifting multi-layer perceptron Shifted MLP, performing the depth convolution again, and performing LN normalization on the feature map;
the 3D enabled MLP module uses a residual connection to linearly add the initial feature map token and the feature map after the shift MLP, and transmits the output features to the same-layer PCA module.
4. The method for detecting an image of an intracranial aneurysm as claimed in claim 1,
the training process of the 3D MAU-Net detection model is as follows:
sending the data set of the TOF-MRA image obtained after pretreatment into a 3D MAU-Net detection model for training;
wherein, the model complex training optimization algorithm adopts an ADAM algorithm with an initial learning rate lr of 0.001, and a weight attenuation index is 10 -5 Setting a moving index average valueMonitoring training loss once every 30 iteration cycles;
the network adopts a mixed loss function;
the number of the batch sizes is 30, the maximum iteration period of the network training is 300, each iteration period comprises 150 iterations, and when the iteration period reaches the maximum value or lr is less than 10 -8 The network stops training;
and saving the model weight to complete the training process.
5. The method for detecting an image of an intracranial aneurysm as claimed in claim 4,
in the training process of the 3D MAU-Net detection model, a Focal loss function and a Lovasz-Softmax loss function are adopted to jointly represent global loss and local loss, and the two losses are combined to be expressed as:
loss=loss F +loss L
wherein loss represents a loss function of the 3D MAU-Net detection model; loss of loss F Represents a Focal loss, loss L Representing the Lovasz-Softmax loss;
loss F the definition is as follows:
wherein y is i For input instance x i True category, p i To predict input instance x i The probability of belonging to the true class of the class, for the problem of extremely unbalanced class, adds (1-p i ) γ Predicting samples with small probability, so that the attention of positive samples is increased, gamma is an adjustable factor, N represents the total number of samples, and alpha represents a weight parameter;
Lovasz-Softmax can promote the intersection ratio IoU score of the segmentation result, and IoU score is also called Jaccard index, which is defined as:
wherein y is * Representing standard results, y represents network prediction results, so Jaccard loss is defined as:
loss J =Δ J (y * ,y)=1-J c (y * ,y);
optimizing Jaccard loss by Lovasz-Softmax is often used to fill in the gaps in Jaccard loss partitioning, recover small objects, and form a more reasonable global partitioning, defined as follows:
m i =max(1-y i ·y i * ,0);
wherein p represents the number of pixels; y is i Representing the value of the i-th element in the network prediction result, y i * Representing the value of the i-th element in the standard result.
6. An intracranial aneurysm image detection system for implementing the intracranial aneurysm image detection method as recited in any one of claims 1 to 5, the intracranial aneurysm image detection system comprising:
the preprocessing module is used for preprocessing the original 3D TOF-MRA;
and the intracranial aneurysm identification module is used for building a 3D MAU-Net detection model, sending the preprocessed image into the 3D MAU-Net detection model, and carrying out intracranial aneurysm detection on the image to obtain an intracranial aneurysm segmentation result.
7. A computer device comprising a memory and one or more processors, the memory having executable code stored therein, wherein the processor, when executing the executable code,
a step of implementing the intracranial aneurysm image detection method as recited in any one of claims 1 to 5.
8. A computer-readable storage medium having a program stored thereon, which when executed by a processor, implements the steps of the intracranial aneurysm image detection method as recited in any one of claims 1 to 5.
CN202311703651.2A 2023-12-13 2023-12-13 Intracranial aneurysm image detection method, system, equipment and medium Active CN117392137B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202311703651.2A CN117392137B (en) 2023-12-13 2023-12-13 Intracranial aneurysm image detection method, system, equipment and medium

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202311703651.2A CN117392137B (en) 2023-12-13 2023-12-13 Intracranial aneurysm image detection method, system, equipment and medium

Publications (2)

Publication Number Publication Date
CN117392137A CN117392137A (en) 2024-01-12
CN117392137B true CN117392137B (en) 2024-02-13

Family

ID=89468876

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202311703651.2A Active CN117392137B (en) 2023-12-13 2023-12-13 Intracranial aneurysm image detection method, system, equipment and medium

Country Status (1)

Country Link
CN (1) CN117392137B (en)

Citations (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2021104056A1 (en) * 2019-11-27 2021-06-03 中国科学院深圳先进技术研究院 Automatic tumor segmentation system and method, and electronic device
CN113436166A (en) * 2021-06-24 2021-09-24 深圳市铱硙医疗科技有限公司 Intracranial aneurysm detection method and system based on magnetic resonance angiography data
CN113706451A (en) * 2021-07-07 2021-11-26 杭州脉流科技有限公司 Method, device, system and computer-readable storage medium for intracranial aneurysm identification detection
CN115147404A (en) * 2022-08-06 2022-10-04 河南大学 Intracranial aneurysm segmentation method with dual-feature fusion MRA image
CN115829914A (en) * 2022-08-15 2023-03-21 首都医科大学 Method for automatically identifying and segmenting intracranial aneurysm in 3.0T high-resolution MRI T1 sequence
CN115953393A (en) * 2023-03-10 2023-04-11 济南市人民医院 Intracranial aneurysm detection system, equipment and storage medium based on multitask learning
CN116823833A (en) * 2023-08-30 2023-09-29 山东科技大学 Method, system and equipment for detecting intracranial aneurysm by using omnibearing MIP image

Family Cites Families (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US11580646B2 (en) * 2021-03-26 2023-02-14 Nanjing University Of Posts And Telecommunications Medical image segmentation method based on U-Net

Patent Citations (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2021104056A1 (en) * 2019-11-27 2021-06-03 中国科学院深圳先进技术研究院 Automatic tumor segmentation system and method, and electronic device
CN113436166A (en) * 2021-06-24 2021-09-24 深圳市铱硙医疗科技有限公司 Intracranial aneurysm detection method and system based on magnetic resonance angiography data
CN113706451A (en) * 2021-07-07 2021-11-26 杭州脉流科技有限公司 Method, device, system and computer-readable storage medium for intracranial aneurysm identification detection
CN115147404A (en) * 2022-08-06 2022-10-04 河南大学 Intracranial aneurysm segmentation method with dual-feature fusion MRA image
CN115829914A (en) * 2022-08-15 2023-03-21 首都医科大学 Method for automatically identifying and segmenting intracranial aneurysm in 3.0T high-resolution MRI T1 sequence
CN115953393A (en) * 2023-03-10 2023-04-11 济南市人民医院 Intracranial aneurysm detection system, equipment and storage medium based on multitask learning
CN116823833A (en) * 2023-08-30 2023-09-29 山东科技大学 Method, system and equipment for detecting intracranial aneurysm by using omnibearing MIP image

Non-Patent Citations (2)

* Cited by examiner, † Cited by third party
Title
深度迭代融合的脑部磁共振图像颅骨去除网络;姚发展;李智;王丽会;程欣宇;张健;;中国图象图形学报;20201016(第10期);全文 *
耿辰 ; 杨丽琴 ; 尹波 ; 李郁欣 ; 耿道颖 ; .基于深度学习的时间飞跃法磁共振血管成像对脑动脉瘤自动检测方法研究.上海医学.2020,(第05期),全文. *

Also Published As

Publication number Publication date
CN117392137A (en) 2024-01-12

Similar Documents

Publication Publication Date Title
CN113077471B (en) Medical image segmentation method based on U-shaped network
CN112017189B (en) Image segmentation method and device, computer equipment and storage medium
CN111612754B (en) MRI tumor optimization segmentation method and system based on multi-modal image fusion
Wang et al. FE-YOLOv5: Feature enhancement network based on YOLOv5 for small object detection
CN113642390B (en) Street view image semantic segmentation method based on local attention network
CN114445670B (en) Training method, device and equipment of image processing model and storage medium
CN114494296A (en) Brain glioma segmentation method and system based on fusion of Unet and Transformer
CN113496494B (en) Two-dimensional skeleton segmentation method and device based on DRR simulation data generation
CN116309648A (en) Medical image segmentation model construction method based on multi-attention fusion
CN116935044B (en) Endoscopic polyp segmentation method with multi-scale guidance and multi-level supervision
CN112488996B (en) Non-homogeneous three-dimensional esophageal cancer energy spectrum CT weak supervision automatic labeling method and system
CN117422715A (en) Global information-based breast ultrasonic tumor lesion area detection method
CN117934824A (en) Target region segmentation method and system for ultrasonic image and electronic equipment
CN113706544A (en) Medical image segmentation method based on complete attention convolution neural network
CN117557774A (en) Unmanned aerial vehicle image small target detection method based on improved YOLOv8
CN117252852A (en) Lightweight skin lesion segmentation method and system based on convolutional neural network and multi-layer perceptron
Qin et al. Dynamically aggregating MLPs and CNNs for skin lesion segmentation with geometry regularization
CN116823850A (en) Cardiac MRI segmentation method and system based on U-Net and transducer fusion improvement
CN118196628A (en) Enhanced semantic-location feature fusion network method, device and application for extracting backbone based on different pre-training features
Zheng et al. Smaformer: Synergistic multi-attention transformer for medical image segmentation
CN117392137B (en) Intracranial aneurysm image detection method, system, equipment and medium
Zhao et al. VCMix-Net: A hybrid network for medical image segmentation
CN115775252A (en) Magnetic resonance image cervical cancer tumor segmentation method based on global local cascade
Liu et al. Medical Image Recognition Based on Multiscale Cascade Segmentation Network MCSnet
CN117893934B (en) Improved UNet3+ network unmanned aerial vehicle image railway track line detection method and device

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant