CN117392137B - Intracranial aneurysm image detection method, system, equipment and medium - Google Patents

Intracranial aneurysm image detection method, system, equipment and medium Download PDF

Info

Publication number
CN117392137B
CN117392137B CN202311703651.2A CN202311703651A CN117392137B CN 117392137 B CN117392137 B CN 117392137B CN 202311703651 A CN202311703651 A CN 202311703651A CN 117392137 B CN117392137 B CN 117392137B
Authority
CN
China
Prior art keywords
module
loss
intracranial aneurysm
mlp
mau
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN202311703651.2A
Other languages
Chinese (zh)
Other versions
CN117392137A (en
Inventor
刘庆一
王可心
王以军
白培瑞
韩超
杨光
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Shandong University of Science and Technology
Original Assignee
Shandong University of Science and Technology
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Shandong University of Science and Technology filed Critical Shandong University of Science and Technology
Priority to CN202311703651.2A priority Critical patent/CN117392137B/en
Publication of CN117392137A publication Critical patent/CN117392137A/en
Application granted granted Critical
Publication of CN117392137B publication Critical patent/CN117392137B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/0002Inspection of images, e.g. flaw detection
    • G06T7/0012Biomedical image inspection
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/0464Convolutional networks [CNN, ConvNet]
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/08Learning methods
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/10Segmentation; Edge detection
    • G06T7/12Edge-based segmentation
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/20Special algorithmic details
    • G06T2207/20081Training; Learning
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/20Special algorithmic details
    • G06T2207/20084Artificial neural networks [ANN]
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/30Subject of image; Context of image processing
    • G06T2207/30004Biomedical image processing
    • G06T2207/30096Tumor; Lesion
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/30Subject of image; Context of image processing
    • G06T2207/30004Biomedical image processing
    • G06T2207/30101Blood vessel; Artery; Vein; Vascular
    • YGENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
    • Y02TECHNOLOGIES OR APPLICATIONS FOR MITIGATION OR ADAPTATION AGAINST CLIMATE CHANGE
    • Y02TCLIMATE CHANGE MITIGATION TECHNOLOGIES RELATED TO TRANSPORTATION
    • Y02T10/00Road transport of goods or passengers
    • Y02T10/10Internal combustion engine [ICE] based vehicles
    • Y02T10/40Engine management systems

Landscapes

  • Engineering & Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • Theoretical Computer Science (AREA)
  • General Physics & Mathematics (AREA)
  • General Health & Medical Sciences (AREA)
  • Health & Medical Sciences (AREA)
  • Molecular Biology (AREA)
  • Computational Linguistics (AREA)
  • Software Systems (AREA)
  • Mathematical Physics (AREA)
  • General Engineering & Computer Science (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Artificial Intelligence (AREA)
  • Biomedical Technology (AREA)
  • Biophysics (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Data Mining & Analysis (AREA)
  • Evolutionary Computation (AREA)
  • Computing Systems (AREA)
  • Medical Informatics (AREA)
  • Nuclear Medicine, Radiotherapy & Molecular Imaging (AREA)
  • Radiology & Medical Imaging (AREA)
  • Quality & Reliability (AREA)
  • Magnetic Resonance Imaging Apparatus (AREA)

Abstract

本发明属于医学图像处理技术领域,具体公开了一种颅内动脉瘤图像检测方法、系统、设备及介质。本发明在3D U‑Net模型的基础上通过引入了上下文感知金字塔特征提取模块、通道注意模块与多尺度卷积模块进行改进,加深了对图像深层上下文特征、图像浅层的边缘以及轮廓等特征信息的提取,为解码器端提供更多具有通道权重信息的深度特征,提高模型的全局学习能力,使其更适用于颅内动脉瘤小目标检测任务,有效地提高了颅内动脉瘤的检测性能。本发明所提基于多层感知机的轻量化3D MAU‑Net检测模型,能够在保证不显著影响检测性能得前提下,具有较少的参数、较高的训练速度以及较低的计算复杂度。

The invention belongs to the technical field of medical image processing, and specifically discloses an intracranial aneurysm image detection method, system, equipment and medium. This invention improves the 3D U-Net model by introducing a context-aware pyramid feature extraction module, a channel attention module and a multi-scale convolution module, and deepens the understanding of deep contextual features of the image, shallow edges and contours of the image, etc. The extraction of information provides the decoder with more deep features with channel weight information, improves the global learning ability of the model, makes it more suitable for intracranial aneurysm small target detection tasks, and effectively improves the detection of intracranial aneurysms. performance. The lightweight 3D MAU-Net detection model based on multi-layer perceptron proposed by the present invention can have fewer parameters, higher training speed and lower computational complexity without significantly affecting the detection performance.

Description

Intracranial aneurysm image detection method, system, equipment and medium
Technical Field
The invention belongs to the technical field of medical image processing, and particularly relates to a method, a system, equipment and a medium for detecting intracranial aneurysm images.
Background
Intracranial aneurysms have high mortality due to their special location of incidence, and early discovery of uncracked intracranial aneurysms is the best means to reduce mortality. At present, the 3D TOF-MRA (Three Dimensional Timeof Flight Magnetic Resonance Angiography, three-dimensional time-leaping method magnetic resonance vascular imaging) technology has great application potential in intracranial aneurysm examination and diagnosis, does not use contrast agent, and has the advantages of safety and noninvasive property. However, when the 3D TOF-MRA image is used for aneurysm detection, the problems of insensitivity to small intracranial aneurysms, long training time, high memory occupation and the like exist, meanwhile, the number of layers of the 3D MRA image is large, the manual film reading workload of doctors is greatly increased, and the sensitivity of the diagnosis of radiologists can be reduced due to high-intensity film reading work, so that a series of problems such as missed diagnosis and misdiagnosis are caused. The MIP (Maximum Intensity Projection ) image generated by the rotation projection of the 3D TOF-MRA image largely retains the density information of the original image, but the quality and reliability of the MIP image mainly depend on the quality and parameter selection of the original data, which requires professional skills and experience, and if the parameter selection is poor, the acquired related information amount is more limited.
In recent years, deep learning models such as Convolutional Neural Networks (CNNs), full Convolutional Networks (FCNs), and the like have exhibited excellent performance in medical image detection and segmentation. For the study of the deep learning algorithm of the intracranial aneurysm, the network model is constructed as a primary factor considered for improving the detection performance of the intracranial aneurysm, because in the study of the intracranial aneurysm on a 3D image, the algorithm is realized by adopting a basic structure of an encoder and a decoder, and the improvement is usually developed around the thought of adjusting the network structure, modifying the size of a convolution kernel, extracting and fusing multi-scale features, adding an attention mechanism, improving a loss function and the like, so that the network is helped to learn the depth features with finer granularity through the improvement, and the detection performance of the algorithm is improved. In addition, part of researchers also combine the traditional algorithm with the deep learning algorithm, and based on the original deep learning model, the traditional algorithm is embedded into the whole detection flow in a pre-processing or post-processing mode, so that the detection precision of the algorithm is improved. Still other researchers combine two or more lightweight deep learning models and propose a multi-stage based learning strategy to achieve fine segmentation and detection of intracranial aneurysms.
In conclusion, the appearance of computer-aided diagnosis provides an important reference for doctor diagnosis, and the existing intracranial aneurysm aided diagnosis technology is not mature enough, and has the problems of long detection time, insufficient accuracy and the like.
Researchers have made a lot of effort to improve the sensitivity and accuracy of intracranial aneurysm detection models. The detection model is designed based on a convolution network, so that the detection sensitivity of the detection model to the intracranial aneurysm is very high, an encoder and decoder structure similar to the Unet is proposed by Geng et al, and the detection model has higher detection rate to the intracranial aneurysm of the 3D TOF-MRA image, and is suitable for clinical screening of the intracranial aneurysm. Claux et al propose a two-stage Unet network detection model that concatenates two Unets, wherein the first Unet model is used to segment the cerebral vessels and generate a binary image, and then the segmented binary image is sent to the second Unet model for training, finally, the detection of intracranial aneurysms is realized, the sensitivity of the algorithm reaches 78%, and the recognition rate of positive samples reaches 62%. Although the deep learning method has higher detection accuracy, it also brings more calculation cost, and this defect is more remarkable in the case of 3D medical images.
Disclosure of Invention
The invention aims to provide an intracranial aneurysm image detection method, which utilizes a 3D MAU-Net detection model (3D Multi-scale attention U-Net, namely a Multi-scale attention detection model) to perform feature extraction on an original 3D TOF-MRA image, so that original features of the image are reserved to the greatest extent, the model is enabled to pay more attention to segmentation of small targets, the accuracy of intracranial aneurysm detection and segmentation is improved, and a linear network is used for replacing part of convolution modules of the original model, so that the complexity of the model is reduced and the training time of the model is shortened on the premise that the detection performance of the model is basically unchanged.
In order to achieve the above purpose, the invention adopts the following technical scheme:
an intracranial aneurysm image detection method comprises the following steps:
step 1, preprocessing original 3D TOF-MRA;
and 2, building a 3D MAU-Net detection model, sending the preprocessed image into the 3D MAU-Net detection model, and detecting the intracranial aneurysm of the image to obtain a segmentation result of the intracranial aneurysm.
In addition, on the basis of the intracranial aneurysm image detection method, the invention also provides an intracranial aneurysm image detection system which is suitable for the intracranial aneurysm image detection method, and the intracranial aneurysm image detection system adopts the following technical scheme:
an intracranial aneurysm image detection system comprising:
the preprocessing module is used for preprocessing the original 3D TOF-MRA;
and the intracranial aneurysm identification module is used for building a 3D MAU-Net detection model, sending the preprocessed image into the 3D MAU-Net detection model, and carrying out intracranial aneurysm detection on the image to obtain an intracranial aneurysm segmentation result.
In addition, on the basis of the intracranial aneurysm image detection method, the invention further provides computer equipment which comprises a memory and one or more processors.
The memory stores executable codes, and the processor is used for realizing the steps of the intracranial aneurysm image detection method when executing the executable codes.
Furthermore, on the basis of the intracranial aneurysm image detection method, the invention further provides a computer readable storage medium on which a program is stored.
The program, when executed by a processor, is configured to implement the steps of the intracranial aneurysm image detection method described above.
The invention has the following advantages:
the invention provides a 3D MAU-Net detection model, which is obtained by introducing a context perception pyramid feature extraction module, a channel attention module and a multi-scale convolution module on the basis of a 3D U-Net model, so that the extraction of feature information such as deep image context features, edges of shallow images and contours is deepened, more depth features with channel weight information are provided for a decoder end, the global learning capacity of the model is improved, the model is more suitable for intracranial aneurysm small target detection tasks, and the detection performance of intracranial aneurysms is effectively improved. Aiming at the common problems of the U-Net network, namely the characteristics of overlarge parameter quantity, overlong model training time and overhigh requirement on computer performance, and can not be well applied to the hospital nursing point for rapid image segmentation, the lightweight 3D MAU-Net detection model based on the multi-layer perceptron has fewer parameters, higher training speed and lower calculation complexity on the premise of ensuring that the detection performance is not obviously influenced.
Drawings
FIG. 1 is a flow chart of an automatic intracranial aneurysm detection model based on an attention mechanism and a multi-layer perceptron in an embodiment of the invention.
Fig. 2 is a schematic diagram of a network architecture of a 3D MAU-Net detection model constructed in an embodiment of the present invention.
Fig. 3 is a schematic diagram of a PCA module according to an embodiment of the present invention.
FIG. 4 is a schematic diagram of hole convolution in an embodiment of the present invention.
Fig. 5 is a schematic diagram of an MC module according to an embodiment of the invention.
FIG. 6 is a schematic diagram of a 3D Tokenized MLP in an embodiment of the invention.
Fig. 7 is a schematic diagram of a deep convolution in an embodiment of the present invention.
Detailed Description
The invention is described in further detail below with reference to the attached drawings and detailed description:
example 1
In this embodiment 1, an intracranial aneurysm image detection method is described, which proposes feature extraction of an original 3D TOF-MRA image, so as to preserve original features of the image to the greatest extent, and improve detection performance of the intracranial aneurysm. And the linear network is used for replacing part of convolution modules in the original model, so that the complexity of the model is reduced and the training time of the model is shortened on the premise of ensuring that the detection performance of the model is basically unchanged. Therefore, the invention provides an intracranial aneurysm automatic detection model based on an attention mechanism and a Multi-layer perceptron, firstly, a PCA module (Pyramid channel attention) and a MC module (Multi-scale convolution) are added in 3D U-Net to deeply extract features of different layers, then, a decoder is used for carrying out feature fusion among different layers, and finally, a final intracranial aneurysm detection result is obtained through deep supervision and guidance network training. Based on the model, the invention proposes to replace the original partial convolution module by a three-dimensional labeling multi-layer perceptron module (3D Tokenized MLP) so as to reduce the complexity of the model.
As shown in fig. 1, the intracranial aneurysm image detection method in the embodiment includes the following steps:
and step 1, preprocessing the original 3D TOF-MRA.
The preprocessing operations include resampling, data normalization, and random data enhancement.
Resampling is to normalize voxels of different sizes in the original image to the same size, setting the median pixel pitch to [0.7,0.43,0.43].
The data normalization is to scale data with different sizes into the same data interval through processing, and the calculation formula is as follows:
wherein the method comprises the steps ofIs the standard deviation of the two-dimensional image,as an average value of the values,a voxel value representing each point is represented,representing the normalized score.
Random data enhancement: when the algorithm performs training experiments, random data enhancement can perform random elastic deformation, random rotation and random scale transformation on the data in real time.
And 2, building a 3D MAU-Net detection model, sending the preprocessed image into the 3D MAU-Net detection model, and detecting the intracranial aneurysm of the image to obtain a segmentation result of the intracranial aneurysm.
The original TOF-MRA image is sent to the 3D MAU-Net detection model to obtain a segmentation result, so that the original characteristics of the image can be fully learned, and the detection performance of intracranial aneurysms is effectively improved.
The invention provides an intracranial aneurysm automatic detection model combining an attention mechanism with a 3D U-Net model in deep learning, and provides a pyramid channel attention module, a multi-scale convolution module and a 3D Tokenized MLP module.
Specifically, the 3D MAU-Net detection model is based on the traditional 3D U-Net model, a context awareness pyramid feature extraction module CPFE, a channel attention module CA and a multi-scale convolution module MC are introduced, extraction of feature information of deep context features, edges and contours of images is achieved, and in addition, the invention further provides that a three-dimensional labeling MLP (namely a 3D processed MLP) module is used for replacing a part of convolution module, so that the complexity of the 3D MAU-Net detection model is reduced fully on the premise that the detection performance is basically unchanged. The method can solve the problems that the existing manual interpretation method is easily influenced by subjective experience of doctors and wastes time and labor, and further improves the accuracy and efficiency of intracranial aneurysm detection.
After the step 1 is completed, the intracranial aneurysm image is pre-cut, and then is sent into an automatic intracranial aneurysm detection model based on an attention mechanism and a multi-layer perceptron, namely a 3D MAU-Net detection model. The detection model takes a 3D U-Net structure as a main network, is improved by introducing a PCA and MC functional module and a 3D enabled MLP module, so that the detection model is suitable for intracranial aneurysm detection, and the network architecture of the 3D MAU-Net detection model is shown as a figure 2 and comprises an encoder, a PCA module, an MC module, a decoder, a 3D enabled MLP module and a depth supervision module;
the processing flow of the preprocessed 3D TOF-MRA image in the 3D MAU-Net detection model is as follows:
firstly, sending the preprocessed 3D TOF-MRA image to an encoder, performing common convolution and residual connection to extract image features, and inputting the features extracted by the encoder to a 3D Tokenized MLP module to perform shifting MLP operation;
the encoder outputs five layers of features f1, f2, f3, f4 and f5 together with the 3D Tokenized MLP;
the advanced features f1 and f2 are subjected to feature learning through the MC module, extraction of image edge information is enhanced, and feature images of f1 'and f 2' are output; extracting the advanced features f3, f4 and f5 by using a PCA module, outputting f3 ', f4 ' and f5 ' feature images to obtain more complete bottom features, and then passing through a 3D enabled MLP module and a decoder module;
and f3 ' feature images and the upper layer S4 of the decoder are fused and input into the third layer decoder to obtain S3, f2 ' feature images and the upper layer S3 of the decoder are fused and input into the second layer decoder to obtain S2, f1 ' feature images and the upper layer S2 of the decoder are fused and input into the third layer decoder to obtain S1, the upper three layers of the decoder output S1, S2 and S3 are guided by a mixed loss function to train through a depth supervision module, and a final intracranial aneurysm detection result is obtained.
The encoder is formed by pooling two Conv3d convolution layers and a maximum value, the number of characteristic channels is doubled after each downsampling, and five layers of characteristics of outputs f1, f2, f3, f4 and f5 are activated through convolution and LeakyReLU.
Since the main improvement of 3D U-Net is based on feature fusion of the decoder, features extracted prior to entry into the decoder are critical for intracranial aneurysm detection. When the channel dimension performs feature learning, importance degrees of all channels are emphasized through different module structures, and feature extraction capacity of a deep learning network is improved at an encoder end, so that overall detection performance of a model is improved.
The PCA module structure is shown in fig. 3, the PCA module is composed of a context perception pyramid feature extraction module CPFE module and a channel attention module CA, the CPFE module firstly convolves a feature image from an encoder by utilizing cavities with convolution kernel sizes of 3 multiplied by 3 and different expansion rates (3, 5 and 7 respectively) to obtain multi-scale multi-receptive field features, then fuses the feature image obtained by the cavity convolutions with the feature image with the convolution kernel size of 1 multiplied by 1 to obtain an output feature layer P, then sends the fused feature image P into the CA module, gives different weights to different channels by utilizing the CA module, fully utilizes effective information, and finally inputs the output feature information into a decoder of the same layer. Wherein, definition of the hole convolution is as follows:
wherein the method comprises the steps ofIs an input signal which is provided with a signal,is an output signal which is provided with a signal,representing a length ofR corresponds to the samplingExpansion ratio of (2); in standard convolution, r=1.
Hole convolution as shown in FIG. 4, when the CA module assigns a larger weight to a channel that plays an important role in image detection, it is assumed thatFor PResetting, then transposing the reset P, and multiplying the transposed P and the reset P by matrixFinally, obtaining the channel attention map by using softmax
Wherein,representing the effect of the ith channel on the jth channel; thereafter, matrix multiplication is performed between X transpose and reset of P, and the resulting result is reset to
Finally, the result is combinedMultiplying and performing element-by-element summation operation on P to obtain final output
Wherein,representing the output characteristics of the channel,the importance weight representing the ith feature dimension,a bias term representing a jth neuron;representing the scale parameters, the weights are learned gradually starting from 0.
The formula shows that the final feature of each channel is a weighted sum of the features of all channels and the original features, which is beneficial to improving the discrimination between the features.
As shown in fig. 5, the MC module performs a convolution operation on a feature map obtained at a certain time by using 3 x 3 convolution, 5 x 5 convolution, and 7 x 7 convolution, while simultaneously convolving with 1 x 1 convolution, and obtaining a new feature map with different feature information, and then fusing the different feature maps to obtain a final output feature map. The multi-scale convolution does not change the size of the original feature map, but only enriches the features of the image through convolution operations of different convolutions, and extracts interesting feature information from the image from a global view angle, so that the performance of the model is improved. Because the multi-scale convolution fully utilizes convolution kernels with different sizes, not only can abundant context information be obtained in the feature extraction process, but also the edge contour information of the image can be better reserved.
As shown in FIG. 6, the 3D Tokenized MLP is a three-dimensional extension of the 2D Tokenized MLP block. In the 3D Tokenized MLP module, the feature map is processed as follows:
transmitting the feature map to a first Shifted multi-layer perceptron (Shifted MLP) to carry out width-wise shifting MLP operation, and carrying out depth convolution (DWConv) on the feature information after shifting; after a depth convolution, using a GELU activation function to output to the following Shifted MLP, respectively performing a shifting MLP operation on the height and a shifting MLP operation on the depth on the second shifting multi-layer perceptron Shifted MLP and the third shifting multi-layer perceptron Shifted MLP, performing the depth convolution again, and performing LN normalization on the feature map; here, the 3D token MLP uses a residual connection to linearly add the initial feature map token to the shifted MLP feature map and pass the output features to the next co-layer PCA module.
The mathematical computation process in the 3D Tokenized MLP module can be calculated as:
wherein X is shift Representing the output shifted in the X direction, Y shift Representing the output shifted in the Y direction, Z shift Representing the output shifted in the Z direction; token represents a marker location; t represents the original feature map token, H represents the height, W represents the width, D represents the depth, DWconv represents the depth convolution, LN represents the layer normalization, GELU is the activation function, X represents the output of the X-direction through the Tokenized MLP, Y represents the output of the Y-direction through the Tokenized MLP, and Z is the current output of the module.
When the depth convolution carries out convolution operation, the number of convolution kernels is consistent with the number of channels of the upper layer, and the number of output characteristic diagrams is ensured to be consistent with the number of input characteristic diagrams. As shown in fig. 7.
The decoder end is composed of functional modules, and the decoding stage and the encoding stage are mirror images.
In the decoding stage, the resolution is doubled once per upsampling. And the information of the encoding stage and the information of the decoding stage are combined through the jump connection, so that the image characteristic information quantity contained in the network is increased.
After the 3D MAU-Net detection model is built, the model needs to be trained, and the training process is as follows:
and sending the data set of the TOF-MRA image obtained after pretreatment into a 3D MAU-Net detection model for training.
The model complex training optimization algorithm adopts an ADAM algorithm with an initial learning rate lr of 0.001;
weight decay index of 10 -5 Setting a moving index average valueThe training loss is monitored every 30 iteration cycles as long asDoes not reduce by 5X 10 -4 Lr is reduced by a factor of 5; the network adopts a mixed loss function;
the number of the batch sizes is 30, the maximum iteration period of the network training is 300, each iteration period comprises 150 iterations, and when the iteration period reaches the maximum value or lr is less than 10 -8 The network stops training;
and (5) saving the model weight at the highest accuracy and lower loss under the verification set at present, and completing the training process.
In the training process of the 3D MAU-Net detection model, a Focal loss function and a Lovasz-Softmax loss function are adopted to jointly represent global loss and local loss, and the two losses are combined to be expressed as: loss=loss F +loss L
Wherein loss represents a loss function of the 3D MAU-Net detection model, loss F Table Focal loss, loss L Representing the lovassz-Softmax loss. Focal loss is used to deal with extreme foreground-background class imbalance problems, focusing loss on difficult-to-separate samples, starting from the difficulty of sample classification, defined as follows:
wherein y is i For input instance x i True category, p i To predict input instance x i Probability of belonging to real class, for extremely unbalanced class, addTo predict samples with low probability, and to switch positive samplesThe fluence becomes large, gamma is an adjustable factor, N represents the total number of samples, and alpha represents a weight parameter.
Lovasz-Softmax can promote the intersection ratio IoU score of the segmentation result, and IoU score is also called Jaccard index, which is defined as:
wherein the method comprises the steps ofRepresenting standard results, y represents network prediction results, so Jaccard loss is defined as:
optimizing Jaccard loss by Lovasz-Softmax is often used to fill in the gaps in Jaccard loss partitioning, recover small objects, and form a more reasonable global partitioning, defined as follows:
wherein p represents the number of pixels;representing the first of the network predictions𝑖Value of individual element->Represents the first of the standard results𝑖The values of the individual elements.
Furthermore, in order to verify the effectiveness of the method of the invention, the following experiments are also presented:
the data used in this experiment was provided by published data in a hospital and MICCAI2020 aneurysm detection and segmentation (ADAM) challenge, and the experimental data were 1043 TOF-MRA images, 500 negative samples and 543 positive samples.
S1, preprocessing a data set:
the label used in the experiment is a binarized image, and since the negative sample does not have the original label, a label image which is consistent with the original image in size and has a voxel value of 0 needs to be generated in the process of constructing the data set. All subject data were then divided into large, medium, small and non-intracranial aneurysms patients, with intracranial aneurysms greater than 7mm in diameter labeled large aneurysms, medium aneurysms between 3mm and 7mm in diameter labeled medium aneurysms, small aneurysms less than 3mm in diameter labeled small aneurysms, and subjects not having an intracranial aneurysm labeled normal human specimens. And split it into training and testing sets.
S2, sending the training set of the TOF-MRA image obtained after pretreatment into a lightweight 3D MAU-Net detection model based on MLP for training.
The network training optimization algorithm provided by the model adopts an ADAM algorithm with initial learning rate of lr=0.001, and weight attenuation index is 10 -5 Setting a moving index average valueThe training loss is monitored every 30 iteration cycles as long asDoes not reduce by 5X 10 -4 Lr is reduced by a factor of 5. The network employs a hybrid loss function. The number of the batch sizes is 30, the maximum iteration period of the network training is 300, each iteration period comprises 150 iterations, and when the iteration period reaches the maximum value or lr is less than 10 -8 The network stops training. And saving the model weight at the time of highest accuracy and lower loss under the verification set.
And S3, sending the test set into a lightweight 3D MAU-Net to obtain an intracranial aneurysm detection result. And carrying out visualization processing on the image to obtain a visualization result. The detection model outputs intracranial aneurysm binarization images, the original MIP images and the corresponding binarization images are simultaneously opened by amide software to splice the images, and the aneurysm images are marked by different colors, so that the observation is convenient.
Example 2
This embodiment 2 describes an intracranial aneurysm image detection system based on the same inventive concept as the intracranial aneurysm image detection method described in embodiment 1 above.
Specifically, an intracranial aneurysm image detection system comprising:
the preprocessing module is used for preprocessing the original 3D TOF-MRA;
and the intracranial aneurysm identification module is used for building a 3D MAU-Net detection model, sending the preprocessed image into the 3D MAU-Net detection model, and carrying out intracranial aneurysm detection on the image to obtain an intracranial aneurysm segmentation result.
It should be noted that, in the intracranial aneurysm image detection system, the implementation process of the functions and roles of each functional module is specifically detailed in the implementation process of the corresponding steps in the method in the above embodiment 1, and will not be described herein.
Example 3
Embodiment 3 describes a computer apparatus for implementing the intracranial aneurysm image detection method described in embodiment 1.
In particular, the computer device includes a memory and one or more processors. Executable code is stored in the memory for implementing the steps of the intracranial aneurysm image detection method described above when the executable code is executed by the processor.
In this embodiment, the computer device is any device or apparatus having data processing capability, which is not described herein.
Example 4
Embodiment 4 describes a computer-readable storage medium for implementing the intracranial aneurysm image detection method described in embodiment 1.
Specifically, the computer-readable storage medium in this embodiment 4 has stored thereon a program for implementing the steps of the intracranial aneurysm image detection method described above when executed by a processor.
The computer readable storage medium may be an internal storage unit of any device or apparatus having data processing capability, such as a hard disk or a memory, or may be an external storage device of any device having data processing capability, such as a plug-in hard disk, a Smart Media Card (SMC), an SD Card, a Flash memory Card (Flash Card), or the like, which are provided on the device.
The foregoing description is, of course, merely illustrative of preferred embodiments of the present invention, and it should be understood that the present invention is not limited to the above-described embodiments, but is intended to cover all modifications, equivalents and alternatives falling within the spirit and scope of the present invention as defined by the appended claims.

Claims (8)

1.一种颅内动脉瘤图像检测方法,其特征在于,包括如下步骤:1. An intracranial aneurysm image detection method, characterized by comprising the following steps: 步骤1.对原始的3D TOF-MRA进行预处理;Step 1. Preprocess the original 3D TOF-MRA; 步骤2.搭建3D MAU-Net检测模型,将预处理后的图像送入3D MAU-Net检测模型中,对图像进行颅内动脉瘤检测,得到颅内动脉瘤分割结果;Step 2. Build a 3D MAU-Net detection model, send the preprocessed images into the 3D MAU-Net detection model, detect intracranial aneurysms on the images, and obtain the intracranial aneurysm segmentation results; 搭建的3D MAU-Net检测模型包括编码器、PCA模块、MC模块、解码器、3D Tokenized MLP模块以及深度监督模块;The built 3D MAU-Net detection model includes the encoder, PCA module, MC module, decoder, 3D Tokenized MLP module and deep supervision module; 预处理后的3D TOF-MRA图像在3D MAU-Net检测模型中的处理流程如下:The processing flow of preprocessed 3D TOF-MRA images in the 3D MAU-Net detection model is as follows: 首先预处理后的3D TOF-MRA图像送入到编码器,进行普通卷积与残差连接提取图像特征,经过编码器提取的特征输入到3D Tokenized MLP模块,进行移位MLP操作;First, the preprocessed 3D TOF-MRA image is sent to the encoder, and ordinary convolution and residual connection are performed to extract image features. The features extracted by the encoder are input to the 3D Tokenized MLP module to perform the shift MLP operation; 编码器与3D Tokenized MLP共输出f1、f2、f3、f4、f5五层特征;The encoder and 3D Tokenized MLP output a total of five layers of features: f1, f2, f3, f4, and f5; 将高级特征f1、f2通过MC模块进行特征学习,加强对图像边缘信息的提取,并输出f1`、f2`特征图像;用PCA模块将高级特征f3、f4、f5进行提取,输出f3`、f4`、f5`特征图像以获取更完整的底层特征,再通过3D Tokenized MLP模块与解码器;Use the MC module to perform feature learning on high-level features f1 and f2, strengthen the extraction of image edge information, and output f1`, f2` feature images; use the PCA module to extract high-level features f3, f4, f5, and output f3`, f4 `, f5` feature images to obtain more complete underlying features, and then pass through the 3D Tokenized MLP module and decoder; 将解码器上三层输出S1、S2、S3通过深度监督模块,由混合损失函数指导3D MAU-Net检测模型网络训练,得到最终的颅内动脉瘤检测结果;The outputs S1, S2, and S3 of the upper three layers of the decoder are passed through the deep supervision module, and the hybrid loss function guides the 3D MAU-Net detection model network training to obtain the final intracranial aneurysm detection result; 所述PCA模块由上下文感知金字塔特征提取模块CPFE和通道注意模块CA组成;其中,CPFE模块首先将来自编码器的特征图,利用卷积核大小为3×3×3、膨胀率不同的空洞卷积,获取多尺度多感受野特征,然后将不同膨胀率空洞卷积得到的特征图与经过卷积核大小为1×1×1的特征图相融合得到输出特征层P,之后将经过融合的特征图P送入CA模块中,利用CA模块对不同通道给予不同权重,最后将输出的特征信息输入到同层的解码器中。The PCA module is composed of the context-aware pyramid feature extraction module CPFE and the channel attention module CA; among them, the CPFE module first uses the feature map from the encoder to use dilated convolutions with a convolution kernel size of 3×3×3 and different expansion rates. product to obtain multi-scale multi-receptive field features, and then fuse the feature maps obtained by atrous convolution with different expansion rates with the feature map with a convolution kernel size of 1×1×1 to obtain the output feature layer P, and then the fused The feature map P is sent to the CA module, which uses the CA module to give different weights to different channels, and finally inputs the output feature information into the decoder of the same layer. 2.根据权利要求1所述的颅内动脉瘤图像检测方法,其特征在于,2. The intracranial aneurysm image detection method according to claim 1, characterized in that, 空洞卷积的定义如下:The definition of atrous convolution is as follows: 其中f[i]是输入信号,g[i]是输出信号,h[l]表示长度为l的滤波器,r对应采样f[i]的膨胀率;假设P∈RC×H×W,对P的RC×H进行重置,然后对重置后的P进行转置,将转置后的P和重置后的P执行矩阵乘法,最后利用softmax得到通道注意映射P∈RC×Cwhere f[i] is the input signal, g[i] is the output signal, h[l] represents a filter of length l, and r corresponds to the expansion rate of sampling f[i]; assuming P∈R C×H×W , Reset R C×H of P, then transpose the reset P, perform matrix multiplication on the transposed P and the reset P, and finally use softmax to obtain the channel attention mapping P∈R C× C : 其中xji表示第i个通道对第j个通道的影响;之后,在X转置和P的重置之间执行矩阵乘法,并将得到的结果重置为RC×H×W where x ji represents the influence of the i-th channel on the j-th channel; after that, matrix multiplication is performed between the transposition of X and the reset of P, and the result is reset to R C×H×W ; 最后将结果与β相乘并对P执行逐元素求和运算,得到最终输出A∈RC×H×WFinally, the result is multiplied by β and an element-wise summation operation is performed on P to obtain the final output A∈R C×H×W : 其中,Aj表示通道输出特征,Pi表示第i个特征维度的重要性权重,Pj表示第j个神经元的偏置项;β表示比例参数,逐渐从0开始学习权重。Among them, A j represents the channel output feature, Pi represents the importance weight of the i-th feature dimension, P j represents the bias term of the j-th neuron; β represents the proportion parameter, and gradually starts learning the weight from 0. 3.根据权利要求1所述的颅内动脉瘤图像检测方法,其特征在于,3. The intracranial aneurysm image detection method according to claim 1, characterized in that, 在3D Tokenized MLP模块中,将特征图传递到第一个移位多层感知机Shifted MLP进行宽度上的移位MLP操作,移位后将特征信息进行深度卷积DWConv;In the 3D Tokenized MLP module, the feature map is passed to the first shifted multi-layer perceptron Shifted MLP to perform a width shift MLP operation. After the shift, the feature information is depth-convolved DWConv; 经过一个深度卷积之后,使用GELU激活函数输出到下面的Shifted MLP,在第二个移位多层感知机Shifted MLP和第三个移位多层感知机Shifted MLP分别进行高度上的移位MLP和深度上的移位MLP操作,再次进行深度卷积,将特征图进行LN归一化;After a depth convolution, the GELU activation function is used to output to the following Shifted MLP. The second shifted multi-layer perceptron Shifted MLP and the third shifted multi-layer perceptron Shifted MLP perform height-shifted MLP respectively. And the shift MLP operation in depth, perform depth convolution again, and perform LN normalization on the feature map; 其中,3D Tokenized MLP模块运用了一个残差连接,将最开始的特征图token与经过移位MLP后的特征图进行线性相加,并将输出特征传递到同层PCA模块。Among them, the 3D Tokenized MLP module uses a residual connection to linearly add the initial feature map token and the feature map after the shifted MLP, and transfer the output features to the PCA module of the same layer. 4.根据权利要求1所述的颅内动脉瘤图像检测方法,其特征在于,4. The intracranial aneurysm image detection method according to claim 1, characterized in that, 对3D MAU-Net检测模型进行训练的过程如下:The process of training the 3D MAU-Net detection model is as follows: 将预处理后得到的TOF-MRA图像的数据集送入3D MAU-Net检测模型进行训练;Send the preprocessed TOF-MRA image data set to the 3D MAU-Net detection model for training; 其中,模型络训练优化算法采用初始学习率lr为0.001的ADAM算法,权重衰减指数为10-5,设置移动指数平均值每30个迭代周期监测一次训练损失;Among them, the model network training optimization algorithm uses the ADAM algorithm with an initial learning rate lr of 0.001, a weight attenuation index of 10 -5 , and a moving index average value. Monitor training loss every 30 iterations; 网络采用混合损失函数;The network uses a hybrid loss function; 批大小数量为30,网络训练最大迭代周期为300,每个迭代周期包含150次迭代,当迭代周期达到最大值或lr<10-8时网络停止训练;The batch size is 30, and the maximum iteration cycle of network training is 300. Each iteration cycle contains 150 iterations. When the iteration cycle reaches the maximum value or lr <10 -8 , the network stops training; 保存模型权重,完成训练过程。Save the model weights and complete the training process. 5.根据权利要求4所述的颅内动脉瘤图像检测方法,其特征在于,5. The intracranial aneurysm image detection method according to claim 4, characterized in that, 在3D MAU-Net检测模型训练过程中,采用Focal loss损失函数和Lovasz-Softmaxloss损失函数来共同表示全局损失与局部损失,将两种损失结合表示为:During the training process of the 3D MAU-Net detection model, the Focal loss function and the Lovasz-Softmaxloss loss function are used to jointly represent the global loss and local loss. The combination of the two losses is expressed as: loss=lossF+lossLloss=loss F +loss L ; 其中,loss代表3D MAU-Net检测模型的损失函数;lossF代表Focal loss,lossL代表Lovasz-Softmax loss损失;Among them, loss represents the loss function of the 3D MAU-Net detection model; loss F represents Focal loss, and loss L represents Lovasz-Softmax loss; lossF定义如下:loss F is defined as follows: 其中yi为输入实例xi的真实类别,pi为预测输入实例xi属于类别真实类别的概率,对于类别极度不平衡的问题,添加了(1-pi)γ来预测概率小的样本,使正样本的关注度变大,γ为可调节因子,N表示样本总数,α表示权重参数;where y i is the true category of the input instance x i , and p i is the probability that the input instance x i belongs to the true category of the category. For the problem of extremely imbalanced categories, (1-p i ) γ is added to predict samples with small probability. , to increase the attention of positive samples, γ is an adjustable factor, N represents the total number of samples, and α represents the weight parameter; Lovasz-Softmax能够提升分割结果的交并比IoU分数,IoU分数也称Jaccard指数,Jaccard指数定义为: Lovasz-Softmax can improve the IoU score of the segmentation result. The IoU score is also called the Jaccard index. The Jaccard index is defined as: 其中y*代表标准结果,y代表网络预测结果,因此Jaccard loss定义为:where y * represents the standard result and y represents the network prediction result, so Jaccard loss is defined as: lossJ=ΔJ(y*,y)=1-Jc(y*,y);loss JJ (y * , y) = 1-J c (y * , y); Lovasz-Softmax对Jaccard loss进行优化常用于填补Jaccard loss分割中的空白,恢复小对象,并形成更合理的全局分割,其定义如下:Lovasz-Softmax optimization of Jaccard loss is often used to fill the gaps in Jaccard loss segmentation, restore small objects, and form a more reasonable global segmentation. It is defined as follows: mi=max(1-yi·yi *,0);m i =max(1-y i ·y i * ,0); 其中,p表示像素数量;yi代表网络预测结果中第i个元素的值,yi *代表标准结果中第i个元素的值。Among them, p represents the number of pixels; y i represents the value of the i-th element in the network prediction result, and y i * represents the value of the i-th element in the standard result. 6.一种用于实现如权利要求1至5任一项所述的颅内动脉瘤图像检测方法的颅内动脉瘤图像检测系统,其特征在于,所述颅内动脉瘤图像检测系统包括:6. An intracranial aneurysm image detection system for implementing the intracranial aneurysm image detection method according to any one of claims 1 to 5, characterized in that the intracranial aneurysm image detection system includes: 预处理模块,用于对原始的3D TOF-MRA进行预处理;Preprocessing module, used to preprocess original 3D TOF-MRA; 以及颅内动脉瘤识别模块,用于搭建3D MAU-Net检测模型,将预处理后的图像送入3DMAU-Net检测模型中,对图像进行颅内动脉瘤检测,得到颅内动脉瘤分割结果。And the intracranial aneurysm identification module is used to build a 3D MAU-Net detection model, send the preprocessed images into the 3DMAU-Net detection model, detect intracranial aneurysms on the images, and obtain the intracranial aneurysm segmentation results. 7.一种计算机设备,包括存储器和一个或多个处理器,所述存储器中存储有可执行代码,其特征在于,所述处理器执行所述可执行代码时,7. A computer device, comprising a memory and one or more processors, with executable code stored in the memory, characterized in that when the processor executes the executable code, 实现如权利要求1至5任一项所述的颅内动脉瘤图像检测方法的步骤。The steps of implementing the intracranial aneurysm image detection method according to any one of claims 1 to 5. 8.一种计算机可读存储介质,其上存储有程序,其特征在于,该程序被处理器执行时,实现如权利要求1至5任一项所述的颅内动脉瘤图像检测方法的步骤。8. A computer-readable storage medium with a program stored thereon, characterized in that when the program is executed by a processor, the steps of the intracranial aneurysm image detection method according to any one of claims 1 to 5 are implemented. .
CN202311703651.2A 2023-12-13 2023-12-13 Intracranial aneurysm image detection method, system, equipment and medium Active CN117392137B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202311703651.2A CN117392137B (en) 2023-12-13 2023-12-13 Intracranial aneurysm image detection method, system, equipment and medium

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202311703651.2A CN117392137B (en) 2023-12-13 2023-12-13 Intracranial aneurysm image detection method, system, equipment and medium

Publications (2)

Publication Number Publication Date
CN117392137A CN117392137A (en) 2024-01-12
CN117392137B true CN117392137B (en) 2024-02-13

Family

ID=89468876

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202311703651.2A Active CN117392137B (en) 2023-12-13 2023-12-13 Intracranial aneurysm image detection method, system, equipment and medium

Country Status (1)

Country Link
CN (1) CN117392137B (en)

Families Citing this family (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN119991681B (en) * 2025-04-17 2025-06-20 山东科技大学 Intracranial aneurysm image detection method, system, device and storage medium

Citations (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2021104056A1 (en) * 2019-11-27 2021-06-03 中国科学院深圳先进技术研究院 Automatic tumor segmentation system and method, and electronic device
CN113436166A (en) * 2021-06-24 2021-09-24 深圳市铱硙医疗科技有限公司 Intracranial aneurysm detection method and system based on magnetic resonance angiography data
CN113706451A (en) * 2021-07-07 2021-11-26 杭州脉流科技有限公司 Method, device, system and computer-readable storage medium for intracranial aneurysm identification detection
CN115147404A (en) * 2022-08-06 2022-10-04 河南大学 A dual-feature fusion method for intracranial aneurysm segmentation in MRA images
CN115829914A (en) * 2022-08-15 2023-03-21 首都医科大学 Method for automatically identifying and segmenting intracranial aneurysm in 3.0T high-resolution MRI T1 sequence
CN115953393A (en) * 2023-03-10 2023-04-11 济南市人民医院 Intracranial aneurysm detection system, equipment and storage medium based on multitask learning
CN116823833A (en) * 2023-08-30 2023-09-29 山东科技大学 Comprehensive MIP image intracranial aneurysm detection method, system and equipment

Family Cites Families (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US11580646B2 (en) * 2021-03-26 2023-02-14 Nanjing University Of Posts And Telecommunications Medical image segmentation method based on U-Net

Patent Citations (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2021104056A1 (en) * 2019-11-27 2021-06-03 中国科学院深圳先进技术研究院 Automatic tumor segmentation system and method, and electronic device
CN113436166A (en) * 2021-06-24 2021-09-24 深圳市铱硙医疗科技有限公司 Intracranial aneurysm detection method and system based on magnetic resonance angiography data
CN113706451A (en) * 2021-07-07 2021-11-26 杭州脉流科技有限公司 Method, device, system and computer-readable storage medium for intracranial aneurysm identification detection
CN115147404A (en) * 2022-08-06 2022-10-04 河南大学 A dual-feature fusion method for intracranial aneurysm segmentation in MRA images
CN115829914A (en) * 2022-08-15 2023-03-21 首都医科大学 Method for automatically identifying and segmenting intracranial aneurysm in 3.0T high-resolution MRI T1 sequence
CN115953393A (en) * 2023-03-10 2023-04-11 济南市人民医院 Intracranial aneurysm detection system, equipment and storage medium based on multitask learning
CN116823833A (en) * 2023-08-30 2023-09-29 山东科技大学 Comprehensive MIP image intracranial aneurysm detection method, system and equipment

Non-Patent Citations (2)

* Cited by examiner, † Cited by third party
Title
深度迭代融合的脑部磁共振图像颅骨去除网络;姚发展;李智;王丽会;程欣宇;张健;;中国图象图形学报;20201016(第10期);全文 *
耿辰 ; 杨丽琴 ; 尹波 ; 李郁欣 ; 耿道颖 ; .基于深度学习的时间飞跃法磁共振血管成像对脑动脉瘤自动检测方法研究.上海医学.2020,(第05期),全文. *

Also Published As

Publication number Publication date
CN117392137A (en) 2024-01-12

Similar Documents

Publication Publication Date Title
Wang et al. FE-YOLOv5: Feature enhancement network based on YOLOv5 for small object detection
Chen et al. Attention-based context aggregation network for monocular depth estimation
CN116580241B (en) Image processing method and system based on dual-branch multi-scale semantic segmentation network
CN118134952A (en) A medical image segmentation method based on feature interaction
CN113205524A (en) Blood vessel image segmentation method, device and equipment based on U-Net
CN113538458A (en) U-Net image segmentation method based on FTL loss function and attention
CN112634211A (en) MRI (magnetic resonance imaging) image segmentation method, device and equipment based on multiple neural networks
CN112052877A (en) Image fine-grained classification method based on cascade enhanced network
CN114359293A (en) A 3D MRI brain tumor segmentation method based on deep learning
CN116823833B (en) Comprehensive MIP image intracranial aneurysm detection method, system and equipment
CN117934824A (en) Target region segmentation method and system for ultrasonic image and electronic equipment
CN116091929A (en) Semantic Segmentation Method of Remote Sensing Image Combined with Unet and Transformer
CN117392137B (en) Intracranial aneurysm image detection method, system, equipment and medium
CN118898773A (en) Medical image segmentation method and device based on multi-scale feature fusion
CN116503726A (en) Multi-scale light smoke image segmentation method and device
Gao et al. Robust lane line segmentation based on group feature enhancement
Qin et al. Dynamically aggregating MLPs and CNNs for skin lesion segmentation with geometry regularization
CN116229217A (en) Infrared target detection method applied to complex environment
CN119151963B (en) CT image segmentation method and system based on improved Swin-Unet
CN119495004A (en) Water extraction method from remote sensing images based on multi-scale feature fusion of compressed pyramid gating units
CN111047571B (en) Image salient target detection method with self-adaptive selection training process
CN118072017A (en) Lung CT image lung nodule full-automatic segmentation method and system based on deep learning
CN116258756B (en) A self-supervised monocular depth estimation method and system
CN116524237A (en) Multi-mode image classification system and method based on interaction between modes and in modes
Farsi et al. Improving Deep Learning-based Saliency Detection Using Channel Attention Module

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant