CN114581459A - Improved 3D U-Net model-based segmentation method for image region of interest of preschool child lung - Google Patents

Improved 3D U-Net model-based segmentation method for image region of interest of preschool child lung Download PDF

Info

Publication number
CN114581459A
CN114581459A CN202210117033.9A CN202210117033A CN114581459A CN 114581459 A CN114581459 A CN 114581459A CN 202210117033 A CN202210117033 A CN 202210117033A CN 114581459 A CN114581459 A CN 114581459A
Authority
CN
China
Prior art keywords
image
model
net
interest
region
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN202210117033.9A
Other languages
Chinese (zh)
Inventor
俞刚
李哲明
黄坚
沈忱
李竞
杨丽
柴象飞
左盼莉
钱宝鑫
余卓
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Zhejiang University ZJU
Original Assignee
Zhejiang University ZJU
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Zhejiang University ZJU filed Critical Zhejiang University ZJU
Priority to CN202210117033.9A priority Critical patent/CN114581459A/en
Publication of CN114581459A publication Critical patent/CN114581459A/en
Pending legal-status Critical Current

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/10Segmentation; Edge detection
    • G06T7/11Region-based segmentation
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/045Combinations of networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/08Learning methods
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T5/00Image enhancement or restoration
    • G06T5/20Image enhancement or restoration using local operators
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/10Image acquisition modality
    • G06T2207/10072Tomographic images
    • G06T2207/10081Computed x-ray tomography [CT]
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/20Special algorithmic details
    • G06T2207/20024Filtering details
    • G06T2207/20032Median filtering
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/20Special algorithmic details
    • G06T2207/20081Training; Learning
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/20Special algorithmic details
    • G06T2207/20084Artificial neural networks [ANN]
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/20Special algorithmic details
    • G06T2207/20092Interactive image processing based on input by user
    • G06T2207/20104Interactive definition of region of interest [ROI]
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/30Subject of image; Context of image processing
    • G06T2207/30004Biomedical image processing
    • G06T2207/30061Lung

Landscapes

  • Engineering & Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • Theoretical Computer Science (AREA)
  • General Physics & Mathematics (AREA)
  • General Health & Medical Sciences (AREA)
  • Molecular Biology (AREA)
  • Biophysics (AREA)
  • Computational Linguistics (AREA)
  • Data Mining & Analysis (AREA)
  • Evolutionary Computation (AREA)
  • Artificial Intelligence (AREA)
  • Biomedical Technology (AREA)
  • Computing Systems (AREA)
  • General Engineering & Computer Science (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Mathematical Physics (AREA)
  • Software Systems (AREA)
  • Health & Medical Sciences (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Apparatus For Radiation Diagnosis (AREA)

Abstract

The invention provides a method for segmenting a region of interest of a preschool child lung image based on an improved 3D U-Net model, which comprises the following steps: (1) collecting CT image data of preschool children patients for preprocessing; (2) dividing the preprocessed image into a training set, a verification set and a test set; (3) constructing a segmentation model, wherein the segmentation model adopts an improved 3D U-Net network model, a channelized Transformer module is designed between an encoder and a decoder of the 3D U-Net network model, and a UCTransNet framework is constructed to replace a jump connection in U-Net so as to better fuse the characteristics of the encoder; (4) sending the preprocessed training set into a constructed segmentation model for training; (5) and inputting the pre-school child lung image to be segmented into the trained segmentation model to obtain the region of interest of the lung image. The method can solve the problem that the image quality is inferior to that of adults caused by excessive motion when preschool children take CT images, so that the interested area of the lung is automatically and accurately segmented from the CT images.

Description

Improved 3D U-Net model-based segmentation method for image region of interest of preschool child lung
Technical Field
The invention belongs to the field of medical artificial intelligence, and particularly relates to a method for segmenting a region of interest of a preschool child lung image based on an improved 3D U-Net model.
Background
Computer vision technology is commonly used in the field of image rapid intelligent processing, such as image classification, target detection and target retrieval, computer vision simulates a human vision mechanism, and has the advantages of high detection speed and low cost.
In recent years, with the application of deep learning in the field of computer vision, particularly the field of medical imaging, breakthrough progress is made, and a data-driven based deep learning technology enables a computer to realize full-automatic segmentation of a target region by combining the technologies of imaging and medical image processing with the analysis and calculation of the computer, so that the segmentation of organ tissues or focuses in medical images can be used for quantitative analysis of medical information and diagnosis assistance of doctors; but also three-dimensional reconstruction for computer surgical guidance.
For example, chinese patent publication No. CN110097550A discloses a medical image segmentation method based on deep learning, which includes: acquiring a historical Magnetic Resonance Imaging (MRI) modality image; dividing the historical Magnetic Resonance Imaging (MRI) modality image into a training set and a test set; in the down-sampling process, inputting two adjacent characteristic layers with different resolutions in any one of the historical MRI modal images in a training set into a neural network model for multi-level characteristic re-extraction and aggregation, and determining a segmented MRI modal image; two adjacent feature layers with different resolutions comprise a low-resolution feature layer and a high-resolution feature layer; two adjacent characteristic layers with different resolutions sequentially pass through a residual convolution unit, a resolution fusion unit and an aggregation unit to determine the segmented MRI modal image.
Chinese patent publication No. CN112950582A discloses a 3D lung lesion segmentation method based on deep learning, which is to pre-process a lung nodule dicom image by acquiring the dicom image; three-dimensionally stacking the preprocessed dicom images to obtain 3D image blocks, and cutting the 3D image blocks; performing feature extraction on the cut 3D image through a pre-trained spherical segmentation model to obtain a regression subgraph; and calculating the product of the centrality and the probability of the regression subgraph to obtain a plurality of central point coordinates, and obtaining coordinates of the regression point through the central point coordinates to obtain a segmentation result.
A doctor automatically and accurately segments the interested lung region from the CT image, obtains a three-dimensional lung model from the image, designs a treatment scheme and is very important for further lesion detection. However, for children, the image quality is not as good as that of adults due to excessive motion during CT taking, and the existing segmentation method has low segmentation precision on the region of interest of the lung image of children.
Disclosure of Invention
The invention provides a method for segmenting a lung image region of interest of a preschool child based on an improved 3D U-Net model, which can relieve the problem that the image quality is not good for adults caused by excessive motion when the preschool child takes a CT image, so that the lung region of interest can be automatically and accurately segmented from the CT image.
A method for segmenting a region of interest of a preschool child lung image based on an improved 3D U-Net model comprises the following steps:
(1) collecting CT image data of preschool child patients, and preprocessing the image data;
(2) dividing the preprocessed image data into a training set, a verification set and a test set;
(3) constructing a segmentation model, wherein the segmentation model adopts an improved 3D U-Net network model, a channelized Transformer module is designed between an encoder and a decoder of the 3D U-Net network model, and a UCTransNet framework is constructed to replace a jump connection in U-Net so as to better fuse the characteristics of the encoder;
the channelized Transformer module consists of a multi-scale channel fusion submodule and a multi-scale channel attention submodule;
(4) sending the preprocessed training set into the constructed segmentation model for training, and adjusting the hyper-parameters of the segmentation model by using the preprocessed verification set to obtain the trained segmentation model;
(5) and inputting the pre-school child lung image to be segmented into the trained segmentation model to obtain the region of interest of the lung image.
In the step (1), the preprocessing includes cutting the image, cutting off irrelevant areas, resampling the image, normalizing the image in scale, and filtering out noise by using a median filter.
In the step (2), the preprocessed image data is divided into a training set, a verification set and a test set according to the ratio of 7:2: 1.
In the step (3), 5 encoders E are arranged in the 3D U-Net network model1~E5And 4 decoders D1~D4
The multi-scale channel fusion submodule includes 2 steps, first multi-scale feature embedding, and then using a multi-layer perceptron.
The structure of the multi-scale channel fusion submodule is as follows:
first 4 encoders E of 3D U-Net network model1~E4The output is T1~T4First, features are reshaped into unfolded 2D patch sequences of different sizes using layer normalization, with the patches in placeThe same region mapped to the encoder features at 4 scales; three elements were introduced in the self-attention of the Transformer: query, Key, and Value; in the process, the original channel size is kept, then the outputs of 4 layers are combined to be used as Key and Value, and the Key and the Value are unified into a feature space through space transformation; mapping Query, Key and Value to a uniform scale of encoder features through layer standardization, and then inputting the features into a multi-layer sensor MLP with a residual error structure to obtain an output O1~O4
The structure of the multi-scale channel attention submodule is as follows:
outputting O to the ith stage of TransformeriAnd ith decoder feature map DiAs an input for channel cross attention; spatial compression is performed by the global average pooling layer, producing the vector g (x) and its kth channel, using this operation to embed global spatial information, and then generating an attention Mask:
Mi=L1*σ(Oi)+L2*σ(Di)
wherein L1 and L2 are the weights of 2 linear layers and the ReLU operator; channel attention maps are constructed using a single linear layer and Sigmoid functions, with the resulting vectors used to recalibrate or fire OiTo
Figure BDA0003496805720000041
Wherein σ (M) is activatedi) Representing the importance of each channel; finally, the Mask is processed
Figure BDA0003496805720000042
With up-sampling feature D of the i-th decoderiAre connected together.
And (4) training the segmentation model by adopting a supervision training method.
Compared with the prior art, the invention has the following beneficial effects:
according to the invention, a channelized Transformer module CTrans is designed between a 3D U-Net encoder (E1-E5) and a decoder (D1-D4) to construct a UCTransNet frame to replace skip connection (skip connection) in U-Net so as to better fuse the characteristics of the encoder and reduce semantic difference, thereby realizing accurate automatic segmentation of medical images.
Drawings
FIG. 1 is a flowchart of a method for segmenting a region of interest of a preschool child lung image based on an improved 3D U-Net model according to the present invention;
FIG. 2 is a general block diagram of a segmentation model according to the present invention;
FIG. 3 is a schematic diagram of a network structure of a multi-scale channel fusion submodule in a segmentation model;
FIG. 4 is a network structure diagram of a multi-scale channel attention submodule in a segmentation model.
Detailed Description
The invention will be described in further detail below with reference to the drawings and examples, which are intended to facilitate the understanding of the invention without limiting it in any way.
As shown in fig. 1, a method for segmenting a region of interest of a lung image of a preschool child based on an improved 3D U-Net model includes the following steps:
1. image pre-processing
Collecting preschool children CT image data, cutting the image, cutting off irrelevant areas, resampling the image, carrying out scale normalization on the image, and filtering noise by adopting a median filter.
2. Data packet
70% of the data set was used as the training set, 20% of the data set was used as the validation set, and 10% of the data set was used as the test set.
3. Model construction
As shown in fig. 2, a segmentation model is constructed, and the segmentation model adopts a modified 3D U-Net network model. In the figure, sections E1-E5 and D1-D4 constitute an encoder and a decoder of 3D U-Net, respectively. The invention is realized by a 3D U-Net encoder (E)1~E5) And a decoder (D)1~D4) A channelized Transformer (CTrans) module is designed to construct a UCTransNet framework to replace a skip connection in U-Net for better performanceAnd the characteristics of the encoder are fused, and the semantic gap is reduced.
CTrans consists of two modules: one multi-scale channel fusion submodule is used for carrying out multi-scale channel fusion with a Transformer, and the other multi-channel attention submodule is used for better solving the characteristic of semantic inconsistency between the channelized Transformer and a U-Net decoder so as to eliminate ambiguity. Therefore, the connection composed of multi-scale channel fusion and multi-channel attention provided by the invention can replace the original skip connections, solve the semantic blank and realize accurate automatic segmentation of the medical image.
In order to solve the problem of skip connection, the invention provides a novel channel cross fusion transform-multi-scale channel fusion, and the features of a multi-scale encoder are fused by using the advantage of long-dependence modeling of the transform. The multi-scale channel fusion sub-module comprises 2 steps, firstly multi-scale feature embedding, and finally multi-layer perceptron (MLP) is used.
The network structure of the multi-scale channel fusion submodule is shown in FIG. 3, given the output T of 4 decoders1~T4First, we use layer normalization to map features reshape into an expanded 2D patch sequence of different sizes so that these patches can be mapped to the same region of the encoder feature at 4 scales. A very critical contribution in the Transformer is self-attention, which is to construct an attention model by using the relation of input samples themselves. Three very important elements are introduced into self-attention: query, Key, and Value. In this process, the original channel size is maintained, and then the outputs of the 4 layers are merged as Key and Value. In self-association, both Key and Value are a transformation of the input sequence itself, and both itself are used as Key and Value. Key and Value can be unified into a feature space through certain spatial transformation. Query, Key and Value are mapped to a uniform scale of encoder features through layer normalization, and then the features are input into a multi-layer perceptron (MLP) with residual structure to encode channel relationships and dependencies, using multi-scale features to extract features from each U-Net encoder level.
Network of multi-channel attention sub-modulesThe structure is shown in FIG. 4, mathematically, the i-th stage transform is output as OiAnd ith-level decoder feature map DiAs an input for channel cross attention. Spatial compression is performed by the Global Average Pooling (GAP) layer, producing the vector g (x) and its kth channel, embedding global spatial information using this operation, and then generating an attention Mask:
Mi=L1*σ(Oi)+L2*σ(Di)
where L1 and L2 are the weights of the 2 linear layers and the ReLU operator. Experience with ECA-Net has shown that avoiding dimensionality reduction is important to learning channel attention, using a single linear layer and Sigmoid functions to construct a channel attention map. The resulting vector is used to recalibrate or excite OiTo
Figure BDA0003496805720000071
In which sigma (M) is activatedi) Indicating the importance of each channel. Finally, the Mask is processed
Figure BDA0003496805720000072
With up-sampling feature D of the i-th decoderiAre connected together.
The invention provides a new visual angle to improve the performance of semantic segmentation, namely, the semantic and resolution difference between low-level and high-level features is made up through more effective feature fusion and multi-scale channel cross attention so as to capture more complex channel dependence. UCTransNet reconsiders the method of the Transformer self-attention mechanism from a channel perspective.
4. Model training and segmentation testing
When the segmentation model is trained, the training set is sent into an improved 3D U-Net segmentation network; the verification set adjusts the hyper-parameters of the model, an optimizer is used for updating the parameters, the network is optimized, the learning rate is automatically adjusted, and a trained segmentation network is obtained; the test set is used to estimate the generalization ability of the model after the learning process is completed.
And training the model by adopting a supervised training method, in order to obtain a corresponding image label, delineating a lung field region in the CT image by an experienced sonographer, and verifying another expert label to ensure the accuracy of the CT image.
5. Evaluation phase
And (3) evaluating the lung field segmentation condition in each patient image: the evaluation of the segmentation task is essentially a pixel-level classification problem. The Accuracy of segmentation (ACC) is divided by the number of correctly classified pixels (TP) by the total number of pixels. When the model outputs a classification in a region that does not contain lung fields, False Positives (FP) are counted. Finally, the segmentation performance of the intelligent lung field segmentation model on the test set is evaluated by adopting a Dice coefficient, Jaccard accuracy, Average Surface Distance (ASD) and Hough Distance (HD).
The embodiments described above are intended to illustrate the technical solutions and advantages of the present invention, and it should be understood that the above-mentioned embodiments are only specific embodiments of the present invention, and are not intended to limit the present invention, and any modifications, additions and equivalents made within the scope of the principles of the present invention should be included in the scope of the present invention.

Claims (8)

1. A method for segmenting a region of interest of a preschool child lung image based on an improved 3D U-Net model is characterized by comprising the following steps:
(1) collecting CT image data of preschool child patients, and preprocessing the image data;
(2) dividing the preprocessed image data into a training set, a verification set and a test set;
(3) constructing a segmentation model, wherein the segmentation model adopts an improved 3D U-Net network model, a channelized Transformer module is designed between an encoder and a decoder of the 3D U-Net network model, and a UCTransNet framework is constructed to replace a jump connection in U-Net so as to better fuse the characteristics of the encoder;
the channelized Transformer module consists of a multi-scale channel fusion submodule and a multi-scale channel attention submodule;
(4) sending the preprocessed training set into the constructed segmentation model for training, and adjusting the hyper-parameters of the segmentation model by using the preprocessed verification set to obtain the trained segmentation model;
(5) and inputting the pre-school child lung image to be segmented into the trained segmentation model to obtain the region of interest of the lung image.
2. The method for segmenting the region of interest of the lung image of the preschool child based on the improved 3D U-Net model as claimed in claim 1, wherein in the step (1), the preprocessing comprises clipping the image, clipping the irrelevant region, resampling the image, normalizing the image in a scale, and filtering out the noise by using a median filter.
3. The method for segmenting the lung image region of interest of the preschool child based on the improved 3D U-Net model is characterized in that in the step (2), the preprocessed image data is divided into a training set, a verification set and a test set according to the ratio of 7:2: 1.
4. The method for segmenting ROI (region of interest) of preschool children based on improved 3D U-Net model of claim 1, wherein in step (3), 5 encoders E are provided in the 3D U-Net network model1~E5And 4 decoders D1~D4
5. The method for segmenting the region of interest of the pulmonary image of the preschool child based on the improved 3D U-Net model is characterized in that in the step (3), the multi-scale channel fusion submodule comprises 2 steps of firstly embedding the multi-scale features and then using a multi-layer perceptron.
6. The method for segmenting the region of interest of the pulmonary image of the preschool child based on the improved 3D U-Net model according to claim 5, wherein the multi-scale channel fusion submodule has the following structure:
first 4 encoders E of 3D U-Net network model1~E4The output is T1~T4First, features are reshaped into unwrapped 2D patch sequences of different sizes using layer normalization, such that these patches map to the same region of the encoder feature at 4 scales; three elements were introduced in self-attention of the Transformer: query, Key, and Value; in the process, the original channel size is kept, then the outputs of 4 layers are combined to be used as Key and Value, and the Key and the Value are unified into a feature space through space transformation; mapping Query, Key and Value to a uniform scale of encoder features through layer standardization, and then inputting the features into a multi-layer sensor MLP with a residual error structure to obtain an output O1~O4
7. The method for segmenting the region of interest of the pulmonary image of the preschool child based on the improved 3D U-Net model according to claim 6, wherein in the step (3), the structure of the multi-scale channel attention submodule is as follows:
outputting O to the ith stage of TransformeriAnd ith-level decoder feature map DiAs an input for channel cross attention; spatial compression is performed by the global mean pooling layer, producing the vector g (x) and its kth channel, using this operation to embed global spatial information, and then generating an attention Mask:
Mi=L1*σ(Oi)+L2*σ(Di)
wherein L1 and L2 are the weights of 2 linear layers and the ReLU operator; channel attention maps are constructed using a single linear layer and Sigmoid functions, with the resulting vectors used to recalibrate or fire OiTo
Figure FDA0003496805710000031
Wherein σ (M) is activatedi) Representing the importance of each channel; finally, the Mask is processed
Figure FDA0003496805710000032
With up-sampling feature D of the i-th decoderiAre connected together.
8. The method for segmenting the lung image region of interest of the preschool child based on the improved 3D U-Net model according to claim 1, wherein in the step (4), a supervised training method is used for training the segmentation model.
CN202210117033.9A 2022-02-08 2022-02-08 Improved 3D U-Net model-based segmentation method for image region of interest of preschool child lung Pending CN114581459A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202210117033.9A CN114581459A (en) 2022-02-08 2022-02-08 Improved 3D U-Net model-based segmentation method for image region of interest of preschool child lung

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202210117033.9A CN114581459A (en) 2022-02-08 2022-02-08 Improved 3D U-Net model-based segmentation method for image region of interest of preschool child lung

Publications (1)

Publication Number Publication Date
CN114581459A true CN114581459A (en) 2022-06-03

Family

ID=81768980

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202210117033.9A Pending CN114581459A (en) 2022-02-08 2022-02-08 Improved 3D U-Net model-based segmentation method for image region of interest of preschool child lung

Country Status (1)

Country Link
CN (1) CN114581459A (en)

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN116645336A (en) * 2023-05-10 2023-08-25 烟台大学 MRI brain image gland pituitary segmentation method
CN118379303A (en) * 2024-06-21 2024-07-23 华侨大学 U-Net pathological cell segmentation method and device based on multiscale transducer multi-attention

Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN109598727A (en) * 2018-11-28 2019-04-09 北京工业大学 A kind of CT image pulmonary parenchyma three-dimensional semantic segmentation method based on deep neural network
US20210110258A1 (en) * 2019-10-14 2021-04-15 Samsung Electronics Co., Ltd. Method and apparatus with model training and/or sequence recognition
CN112991269A (en) * 2021-02-07 2021-06-18 复旦大学 Identification and classification method for lung CT image
CN113469094A (en) * 2021-07-13 2021-10-01 上海中科辰新卫星技术有限公司 Multi-mode remote sensing data depth fusion-based earth surface coverage classification method
CN113888744A (en) * 2021-10-14 2022-01-04 浙江大学 Image semantic segmentation method based on Transformer visual upsampling module

Patent Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN109598727A (en) * 2018-11-28 2019-04-09 北京工业大学 A kind of CT image pulmonary parenchyma three-dimensional semantic segmentation method based on deep neural network
US20210110258A1 (en) * 2019-10-14 2021-04-15 Samsung Electronics Co., Ltd. Method and apparatus with model training and/or sequence recognition
CN112991269A (en) * 2021-02-07 2021-06-18 复旦大学 Identification and classification method for lung CT image
CN113469094A (en) * 2021-07-13 2021-10-01 上海中科辰新卫星技术有限公司 Multi-mode remote sensing data depth fusion-based earth surface coverage classification method
CN113888744A (en) * 2021-10-14 2022-01-04 浙江大学 Image semantic segmentation method based on Transformer visual upsampling module

Non-Patent Citations (5)

* Cited by examiner, † Cited by third party
Title
COAST_S: "如何理解 Transformer 中的 Query、Key 与 Valu", 《HTTPS://BLOG.CSDN.NET/WEIXIN_43996337/ARTICLE/DETAILS/121187666》 *
D.BLANC 等: "Artificial intelligence solution to classify pulmonary nodules on CT", 《DIAGNOSTIC AND INTERVENTIONAL IMAGING》 *
HAONAN WANG 等: "UCTransNet: Rethinking the Skip Connections in U-Net from a Channel-wise Perspective with Transformer", 《ARXIV:2109.04335V3》 *
JINGJING WANG 等: "DA-ResUNet: a novel method for brain tumor segmentation based on U-Net with residual block and CBAM", 《EVENT: FOURTH INTERNATIONAL CONFERENCE ON IMAGE, VIDEO PROCESSING, AND ARTIFICIAL INTELLIGENCE (IVPAI 2021)》 *
OLIVIER PETIT 等: "U-Net Transformer: Self and Cross Attention for Medical Image Segmentation", 《ARXIV:2103.06104V2》 *

Cited By (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN116645336A (en) * 2023-05-10 2023-08-25 烟台大学 MRI brain image gland pituitary segmentation method
CN116645336B (en) * 2023-05-10 2024-05-07 烟台大学 MRI brain image gland pituitary segmentation method
CN118379303A (en) * 2024-06-21 2024-07-23 华侨大学 U-Net pathological cell segmentation method and device based on multiscale transducer multi-attention

Similar Documents

Publication Publication Date Title
CN113506334B (en) Multi-mode medical image fusion method and system based on deep learning
CN113450294A (en) Multi-modal medical image registration and fusion method and device and electronic equipment
Sreelakshmy et al. [Retracted] An Automated Deep Learning Model for the Cerebellum Segmentation from Fetal Brain Images
CN113177943B (en) Cerebral apoplexy CT image segmentation method
CN112529909A (en) Tumor image brain region segmentation method and system based on image completion
CN115496771A (en) Brain tumor segmentation method based on brain three-dimensional MRI image design
CN114119516B (en) Virus focus segmentation method based on migration learning and cascade self-adaptive cavity convolution
CN114494296A (en) Brain glioma segmentation method and system based on fusion of Unet and Transformer
CN114581459A (en) Improved 3D U-Net model-based segmentation method for image region of interest of preschool child lung
Liu et al. A fully automatic segmentation algorithm for CT lung images based on random forest
CN113096137B (en) Adaptive segmentation method and system for OCT (optical coherence tomography) retinal image field
CN112884788B (en) Cup optic disk segmentation method and imaging method based on rich context network
CN115147600A (en) GBM multi-mode MR image segmentation method based on classifier weight converter
CN114882048A (en) Image segmentation method and system based on wavelet scattering learning network
CN113456031A (en) Training device and prediction device of brain state prediction model and electronic equipment
CN116309615A (en) Multi-mode MRI brain tumor image segmentation method
CN117274599A (en) Brain magnetic resonance segmentation method and system based on combined double-task self-encoder
CN118037791A (en) Construction method and application of multi-mode three-dimensional medical image segmentation registration model
CN116485853A (en) Medical image registration method and device based on deep learning neural network
CN116030063A (en) Classification diagnosis system, method, electronic device and medium for MRI image
CN115410032A (en) OCTA image classification structure training method based on self-supervision learning
CN115294023A (en) Liver tumor automatic segmentation method and device
CN115409812A (en) CT image automatic classification method based on fusion time attention mechanism
Mani Deep learning models for semantic multi-modal medical image segmentation
CN112967295A (en) Image processing method and system based on residual error network and attention mechanism

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
RJ01 Rejection of invention patent application after publication

Application publication date: 20220603

RJ01 Rejection of invention patent application after publication