CN114581459A - Improved 3D U-Net model-based segmentation method for image region of interest of preschool child lung - Google Patents
Improved 3D U-Net model-based segmentation method for image region of interest of preschool child lung Download PDFInfo
- Publication number
- CN114581459A CN114581459A CN202210117033.9A CN202210117033A CN114581459A CN 114581459 A CN114581459 A CN 114581459A CN 202210117033 A CN202210117033 A CN 202210117033A CN 114581459 A CN114581459 A CN 114581459A
- Authority
- CN
- China
- Prior art keywords
- image
- model
- net
- interest
- region
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Pending
Links
- 230000011218 segmentation Effects 0.000 title claims abstract description 44
- 238000000034 method Methods 0.000 title claims abstract description 30
- 210000004072 lung Anatomy 0.000 title claims abstract description 28
- 238000012549 training Methods 0.000 claims abstract description 22
- 238000012360 testing method Methods 0.000 claims abstract description 10
- 238000012795 verification Methods 0.000 claims abstract description 8
- 238000007781 pre-processing Methods 0.000 claims abstract description 6
- 230000004927 fusion Effects 0.000 claims description 16
- 230000008569 process Effects 0.000 claims description 6
- 239000013598 vector Substances 0.000 claims description 6
- 238000010606 normalization Methods 0.000 claims description 5
- 238000005070 sampling Methods 0.000 claims description 4
- 230000009466 transformation Effects 0.000 claims description 4
- 238000012952 Resampling Methods 0.000 claims description 3
- 230000006835 compression Effects 0.000 claims description 3
- 238000007906 compression Methods 0.000 claims description 3
- 238000001914 filtration Methods 0.000 claims description 3
- 230000006870 function Effects 0.000 claims description 3
- 238000011176 pooling Methods 0.000 claims description 3
- 238000013507 mapping Methods 0.000 claims description 2
- 230000002685 pulmonary effect Effects 0.000 claims 3
- 238000002595 magnetic resonance imaging Methods 0.000 description 5
- 238000013135 deep learning Methods 0.000 description 4
- 230000008901 benefit Effects 0.000 description 3
- 238000001514 detection method Methods 0.000 description 3
- 238000010586 diagram Methods 0.000 description 3
- 238000005516 engineering process Methods 0.000 description 3
- 230000002776 aggregation Effects 0.000 description 2
- 238000004220 aggregation Methods 0.000 description 2
- 238000011156 evaluation Methods 0.000 description 2
- 238000000605 extraction Methods 0.000 description 2
- 230000003902 lesion Effects 0.000 description 2
- 230000007246 mechanism Effects 0.000 description 2
- 238000012545 processing Methods 0.000 description 2
- 206010056342 Pulmonary mass Diseases 0.000 description 1
- 238000007792 addition Methods 0.000 description 1
- 238000004458 analytical method Methods 0.000 description 1
- 238000013473 artificial intelligence Methods 0.000 description 1
- 230000009286 beneficial effect Effects 0.000 description 1
- 238000004364 calculation method Methods 0.000 description 1
- 238000010276 construction Methods 0.000 description 1
- 238000003745 diagnosis Methods 0.000 description 1
- 238000002059 diagnostic imaging Methods 0.000 description 1
- 238000003709 image segmentation Methods 0.000 description 1
- 238000003384 imaging method Methods 0.000 description 1
- 238000012986 modification Methods 0.000 description 1
- 230000004048 modification Effects 0.000 description 1
- 238000003062 neural network model Methods 0.000 description 1
- 210000000056 organ Anatomy 0.000 description 1
- 238000004445 quantitative analysis Methods 0.000 description 1
- 230000009467 reduction Effects 0.000 description 1
- 238000010200 validation analysis Methods 0.000 description 1
- 230000000007 visual effect Effects 0.000 description 1
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
- G06T7/10—Segmentation; Edge detection
- G06T7/11—Region-based segmentation
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/04—Architecture, e.g. interconnection topology
- G06N3/045—Combinations of networks
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/08—Learning methods
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T5/00—Image enhancement or restoration
- G06T5/20—Image enhancement or restoration using local operators
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/10—Image acquisition modality
- G06T2207/10072—Tomographic images
- G06T2207/10081—Computed x-ray tomography [CT]
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/20—Special algorithmic details
- G06T2207/20024—Filtering details
- G06T2207/20032—Median filtering
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/20—Special algorithmic details
- G06T2207/20081—Training; Learning
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/20—Special algorithmic details
- G06T2207/20084—Artificial neural networks [ANN]
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/20—Special algorithmic details
- G06T2207/20092—Interactive image processing based on input by user
- G06T2207/20104—Interactive definition of region of interest [ROI]
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/30—Subject of image; Context of image processing
- G06T2207/30004—Biomedical image processing
- G06T2207/30061—Lung
Landscapes
- Engineering & Computer Science (AREA)
- Physics & Mathematics (AREA)
- Theoretical Computer Science (AREA)
- General Physics & Mathematics (AREA)
- General Health & Medical Sciences (AREA)
- Molecular Biology (AREA)
- Biophysics (AREA)
- Computational Linguistics (AREA)
- Data Mining & Analysis (AREA)
- Evolutionary Computation (AREA)
- Artificial Intelligence (AREA)
- Biomedical Technology (AREA)
- Computing Systems (AREA)
- General Engineering & Computer Science (AREA)
- Life Sciences & Earth Sciences (AREA)
- Mathematical Physics (AREA)
- Software Systems (AREA)
- Health & Medical Sciences (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Apparatus For Radiation Diagnosis (AREA)
Abstract
The invention provides a method for segmenting a region of interest of a preschool child lung image based on an improved 3D U-Net model, which comprises the following steps: (1) collecting CT image data of preschool children patients for preprocessing; (2) dividing the preprocessed image into a training set, a verification set and a test set; (3) constructing a segmentation model, wherein the segmentation model adopts an improved 3D U-Net network model, a channelized Transformer module is designed between an encoder and a decoder of the 3D U-Net network model, and a UCTransNet framework is constructed to replace a jump connection in U-Net so as to better fuse the characteristics of the encoder; (4) sending the preprocessed training set into a constructed segmentation model for training; (5) and inputting the pre-school child lung image to be segmented into the trained segmentation model to obtain the region of interest of the lung image. The method can solve the problem that the image quality is inferior to that of adults caused by excessive motion when preschool children take CT images, so that the interested area of the lung is automatically and accurately segmented from the CT images.
Description
Technical Field
The invention belongs to the field of medical artificial intelligence, and particularly relates to a method for segmenting a region of interest of a preschool child lung image based on an improved 3D U-Net model.
Background
Computer vision technology is commonly used in the field of image rapid intelligent processing, such as image classification, target detection and target retrieval, computer vision simulates a human vision mechanism, and has the advantages of high detection speed and low cost.
In recent years, with the application of deep learning in the field of computer vision, particularly the field of medical imaging, breakthrough progress is made, and a data-driven based deep learning technology enables a computer to realize full-automatic segmentation of a target region by combining the technologies of imaging and medical image processing with the analysis and calculation of the computer, so that the segmentation of organ tissues or focuses in medical images can be used for quantitative analysis of medical information and diagnosis assistance of doctors; but also three-dimensional reconstruction for computer surgical guidance.
For example, chinese patent publication No. CN110097550A discloses a medical image segmentation method based on deep learning, which includes: acquiring a historical Magnetic Resonance Imaging (MRI) modality image; dividing the historical Magnetic Resonance Imaging (MRI) modality image into a training set and a test set; in the down-sampling process, inputting two adjacent characteristic layers with different resolutions in any one of the historical MRI modal images in a training set into a neural network model for multi-level characteristic re-extraction and aggregation, and determining a segmented MRI modal image; two adjacent feature layers with different resolutions comprise a low-resolution feature layer and a high-resolution feature layer; two adjacent characteristic layers with different resolutions sequentially pass through a residual convolution unit, a resolution fusion unit and an aggregation unit to determine the segmented MRI modal image.
Chinese patent publication No. CN112950582A discloses a 3D lung lesion segmentation method based on deep learning, which is to pre-process a lung nodule dicom image by acquiring the dicom image; three-dimensionally stacking the preprocessed dicom images to obtain 3D image blocks, and cutting the 3D image blocks; performing feature extraction on the cut 3D image through a pre-trained spherical segmentation model to obtain a regression subgraph; and calculating the product of the centrality and the probability of the regression subgraph to obtain a plurality of central point coordinates, and obtaining coordinates of the regression point through the central point coordinates to obtain a segmentation result.
A doctor automatically and accurately segments the interested lung region from the CT image, obtains a three-dimensional lung model from the image, designs a treatment scheme and is very important for further lesion detection. However, for children, the image quality is not as good as that of adults due to excessive motion during CT taking, and the existing segmentation method has low segmentation precision on the region of interest of the lung image of children.
Disclosure of Invention
The invention provides a method for segmenting a lung image region of interest of a preschool child based on an improved 3D U-Net model, which can relieve the problem that the image quality is not good for adults caused by excessive motion when the preschool child takes a CT image, so that the lung region of interest can be automatically and accurately segmented from the CT image.
A method for segmenting a region of interest of a preschool child lung image based on an improved 3D U-Net model comprises the following steps:
(1) collecting CT image data of preschool child patients, and preprocessing the image data;
(2) dividing the preprocessed image data into a training set, a verification set and a test set;
(3) constructing a segmentation model, wherein the segmentation model adopts an improved 3D U-Net network model, a channelized Transformer module is designed between an encoder and a decoder of the 3D U-Net network model, and a UCTransNet framework is constructed to replace a jump connection in U-Net so as to better fuse the characteristics of the encoder;
the channelized Transformer module consists of a multi-scale channel fusion submodule and a multi-scale channel attention submodule;
(4) sending the preprocessed training set into the constructed segmentation model for training, and adjusting the hyper-parameters of the segmentation model by using the preprocessed verification set to obtain the trained segmentation model;
(5) and inputting the pre-school child lung image to be segmented into the trained segmentation model to obtain the region of interest of the lung image.
In the step (1), the preprocessing includes cutting the image, cutting off irrelevant areas, resampling the image, normalizing the image in scale, and filtering out noise by using a median filter.
In the step (2), the preprocessed image data is divided into a training set, a verification set and a test set according to the ratio of 7:2: 1.
In the step (3), 5 encoders E are arranged in the 3D U-Net network model1~E5And 4 decoders D1~D4。
The multi-scale channel fusion submodule includes 2 steps, first multi-scale feature embedding, and then using a multi-layer perceptron.
The structure of the multi-scale channel fusion submodule is as follows:
first 4 encoders E of 3D U-Net network model1~E4The output is T1~T4First, features are reshaped into unfolded 2D patch sequences of different sizes using layer normalization, with the patches in placeThe same region mapped to the encoder features at 4 scales; three elements were introduced in the self-attention of the Transformer: query, Key, and Value; in the process, the original channel size is kept, then the outputs of 4 layers are combined to be used as Key and Value, and the Key and the Value are unified into a feature space through space transformation; mapping Query, Key and Value to a uniform scale of encoder features through layer standardization, and then inputting the features into a multi-layer sensor MLP with a residual error structure to obtain an output O1~O4。
The structure of the multi-scale channel attention submodule is as follows:
outputting O to the ith stage of TransformeriAnd ith decoder feature map DiAs an input for channel cross attention; spatial compression is performed by the global average pooling layer, producing the vector g (x) and its kth channel, using this operation to embed global spatial information, and then generating an attention Mask:
Mi=L1*σ(Oi)+L2*σ(Di)
wherein L1 and L2 are the weights of 2 linear layers and the ReLU operator; channel attention maps are constructed using a single linear layer and Sigmoid functions, with the resulting vectors used to recalibrate or fire OiToWherein σ (M) is activatedi) Representing the importance of each channel; finally, the Mask is processedWith up-sampling feature D of the i-th decoderiAre connected together.
And (4) training the segmentation model by adopting a supervision training method.
Compared with the prior art, the invention has the following beneficial effects:
according to the invention, a channelized Transformer module CTrans is designed between a 3D U-Net encoder (E1-E5) and a decoder (D1-D4) to construct a UCTransNet frame to replace skip connection (skip connection) in U-Net so as to better fuse the characteristics of the encoder and reduce semantic difference, thereby realizing accurate automatic segmentation of medical images.
Drawings
FIG. 1 is a flowchart of a method for segmenting a region of interest of a preschool child lung image based on an improved 3D U-Net model according to the present invention;
FIG. 2 is a general block diagram of a segmentation model according to the present invention;
FIG. 3 is a schematic diagram of a network structure of a multi-scale channel fusion submodule in a segmentation model;
FIG. 4 is a network structure diagram of a multi-scale channel attention submodule in a segmentation model.
Detailed Description
The invention will be described in further detail below with reference to the drawings and examples, which are intended to facilitate the understanding of the invention without limiting it in any way.
As shown in fig. 1, a method for segmenting a region of interest of a lung image of a preschool child based on an improved 3D U-Net model includes the following steps:
1. image pre-processing
Collecting preschool children CT image data, cutting the image, cutting off irrelevant areas, resampling the image, carrying out scale normalization on the image, and filtering noise by adopting a median filter.
2. Data packet
70% of the data set was used as the training set, 20% of the data set was used as the validation set, and 10% of the data set was used as the test set.
3. Model construction
As shown in fig. 2, a segmentation model is constructed, and the segmentation model adopts a modified 3D U-Net network model. In the figure, sections E1-E5 and D1-D4 constitute an encoder and a decoder of 3D U-Net, respectively. The invention is realized by a 3D U-Net encoder (E)1~E5) And a decoder (D)1~D4) A channelized Transformer (CTrans) module is designed to construct a UCTransNet framework to replace a skip connection in U-Net for better performanceAnd the characteristics of the encoder are fused, and the semantic gap is reduced.
CTrans consists of two modules: one multi-scale channel fusion submodule is used for carrying out multi-scale channel fusion with a Transformer, and the other multi-channel attention submodule is used for better solving the characteristic of semantic inconsistency between the channelized Transformer and a U-Net decoder so as to eliminate ambiguity. Therefore, the connection composed of multi-scale channel fusion and multi-channel attention provided by the invention can replace the original skip connections, solve the semantic blank and realize accurate automatic segmentation of the medical image.
In order to solve the problem of skip connection, the invention provides a novel channel cross fusion transform-multi-scale channel fusion, and the features of a multi-scale encoder are fused by using the advantage of long-dependence modeling of the transform. The multi-scale channel fusion sub-module comprises 2 steps, firstly multi-scale feature embedding, and finally multi-layer perceptron (MLP) is used.
The network structure of the multi-scale channel fusion submodule is shown in FIG. 3, given the output T of 4 decoders1~T4First, we use layer normalization to map features reshape into an expanded 2D patch sequence of different sizes so that these patches can be mapped to the same region of the encoder feature at 4 scales. A very critical contribution in the Transformer is self-attention, which is to construct an attention model by using the relation of input samples themselves. Three very important elements are introduced into self-attention: query, Key, and Value. In this process, the original channel size is maintained, and then the outputs of the 4 layers are merged as Key and Value. In self-association, both Key and Value are a transformation of the input sequence itself, and both itself are used as Key and Value. Key and Value can be unified into a feature space through certain spatial transformation. Query, Key and Value are mapped to a uniform scale of encoder features through layer normalization, and then the features are input into a multi-layer perceptron (MLP) with residual structure to encode channel relationships and dependencies, using multi-scale features to extract features from each U-Net encoder level.
Network of multi-channel attention sub-modulesThe structure is shown in FIG. 4, mathematically, the i-th stage transform is output as OiAnd ith-level decoder feature map DiAs an input for channel cross attention. Spatial compression is performed by the Global Average Pooling (GAP) layer, producing the vector g (x) and its kth channel, embedding global spatial information using this operation, and then generating an attention Mask:
Mi=L1*σ(Oi)+L2*σ(Di)
where L1 and L2 are the weights of the 2 linear layers and the ReLU operator. Experience with ECA-Net has shown that avoiding dimensionality reduction is important to learning channel attention, using a single linear layer and Sigmoid functions to construct a channel attention map. The resulting vector is used to recalibrate or excite OiToIn which sigma (M) is activatedi) Indicating the importance of each channel. Finally, the Mask is processedWith up-sampling feature D of the i-th decoderiAre connected together.
The invention provides a new visual angle to improve the performance of semantic segmentation, namely, the semantic and resolution difference between low-level and high-level features is made up through more effective feature fusion and multi-scale channel cross attention so as to capture more complex channel dependence. UCTransNet reconsiders the method of the Transformer self-attention mechanism from a channel perspective.
4. Model training and segmentation testing
When the segmentation model is trained, the training set is sent into an improved 3D U-Net segmentation network; the verification set adjusts the hyper-parameters of the model, an optimizer is used for updating the parameters, the network is optimized, the learning rate is automatically adjusted, and a trained segmentation network is obtained; the test set is used to estimate the generalization ability of the model after the learning process is completed.
And training the model by adopting a supervised training method, in order to obtain a corresponding image label, delineating a lung field region in the CT image by an experienced sonographer, and verifying another expert label to ensure the accuracy of the CT image.
5. Evaluation phase
And (3) evaluating the lung field segmentation condition in each patient image: the evaluation of the segmentation task is essentially a pixel-level classification problem. The Accuracy of segmentation (ACC) is divided by the number of correctly classified pixels (TP) by the total number of pixels. When the model outputs a classification in a region that does not contain lung fields, False Positives (FP) are counted. Finally, the segmentation performance of the intelligent lung field segmentation model on the test set is evaluated by adopting a Dice coefficient, Jaccard accuracy, Average Surface Distance (ASD) and Hough Distance (HD).
The embodiments described above are intended to illustrate the technical solutions and advantages of the present invention, and it should be understood that the above-mentioned embodiments are only specific embodiments of the present invention, and are not intended to limit the present invention, and any modifications, additions and equivalents made within the scope of the principles of the present invention should be included in the scope of the present invention.
Claims (8)
1. A method for segmenting a region of interest of a preschool child lung image based on an improved 3D U-Net model is characterized by comprising the following steps:
(1) collecting CT image data of preschool child patients, and preprocessing the image data;
(2) dividing the preprocessed image data into a training set, a verification set and a test set;
(3) constructing a segmentation model, wherein the segmentation model adopts an improved 3D U-Net network model, a channelized Transformer module is designed between an encoder and a decoder of the 3D U-Net network model, and a UCTransNet framework is constructed to replace a jump connection in U-Net so as to better fuse the characteristics of the encoder;
the channelized Transformer module consists of a multi-scale channel fusion submodule and a multi-scale channel attention submodule;
(4) sending the preprocessed training set into the constructed segmentation model for training, and adjusting the hyper-parameters of the segmentation model by using the preprocessed verification set to obtain the trained segmentation model;
(5) and inputting the pre-school child lung image to be segmented into the trained segmentation model to obtain the region of interest of the lung image.
2. The method for segmenting the region of interest of the lung image of the preschool child based on the improved 3D U-Net model as claimed in claim 1, wherein in the step (1), the preprocessing comprises clipping the image, clipping the irrelevant region, resampling the image, normalizing the image in a scale, and filtering out the noise by using a median filter.
3. The method for segmenting the lung image region of interest of the preschool child based on the improved 3D U-Net model is characterized in that in the step (2), the preprocessed image data is divided into a training set, a verification set and a test set according to the ratio of 7:2: 1.
4. The method for segmenting ROI (region of interest) of preschool children based on improved 3D U-Net model of claim 1, wherein in step (3), 5 encoders E are provided in the 3D U-Net network model1~E5And 4 decoders D1~D4。
5. The method for segmenting the region of interest of the pulmonary image of the preschool child based on the improved 3D U-Net model is characterized in that in the step (3), the multi-scale channel fusion submodule comprises 2 steps of firstly embedding the multi-scale features and then using a multi-layer perceptron.
6. The method for segmenting the region of interest of the pulmonary image of the preschool child based on the improved 3D U-Net model according to claim 5, wherein the multi-scale channel fusion submodule has the following structure:
first 4 encoders E of 3D U-Net network model1~E4The output is T1~T4First, features are reshaped into unwrapped 2D patch sequences of different sizes using layer normalization, such that these patches map to the same region of the encoder feature at 4 scales; three elements were introduced in self-attention of the Transformer: query, Key, and Value; in the process, the original channel size is kept, then the outputs of 4 layers are combined to be used as Key and Value, and the Key and the Value are unified into a feature space through space transformation; mapping Query, Key and Value to a uniform scale of encoder features through layer standardization, and then inputting the features into a multi-layer sensor MLP with a residual error structure to obtain an output O1~O4。
7. The method for segmenting the region of interest of the pulmonary image of the preschool child based on the improved 3D U-Net model according to claim 6, wherein in the step (3), the structure of the multi-scale channel attention submodule is as follows:
outputting O to the ith stage of TransformeriAnd ith-level decoder feature map DiAs an input for channel cross attention; spatial compression is performed by the global mean pooling layer, producing the vector g (x) and its kth channel, using this operation to embed global spatial information, and then generating an attention Mask:
Mi=L1*σ(Oi)+L2*σ(Di)
wherein L1 and L2 are the weights of 2 linear layers and the ReLU operator; channel attention maps are constructed using a single linear layer and Sigmoid functions, with the resulting vectors used to recalibrate or fire OiToWherein σ (M) is activatedi) Representing the importance of each channel; finally, the Mask is processedWith up-sampling feature D of the i-th decoderiAre connected together.
8. The method for segmenting the lung image region of interest of the preschool child based on the improved 3D U-Net model according to claim 1, wherein in the step (4), a supervised training method is used for training the segmentation model.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202210117033.9A CN114581459A (en) | 2022-02-08 | 2022-02-08 | Improved 3D U-Net model-based segmentation method for image region of interest of preschool child lung |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202210117033.9A CN114581459A (en) | 2022-02-08 | 2022-02-08 | Improved 3D U-Net model-based segmentation method for image region of interest of preschool child lung |
Publications (1)
Publication Number | Publication Date |
---|---|
CN114581459A true CN114581459A (en) | 2022-06-03 |
Family
ID=81768980
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN202210117033.9A Pending CN114581459A (en) | 2022-02-08 | 2022-02-08 | Improved 3D U-Net model-based segmentation method for image region of interest of preschool child lung |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN114581459A (en) |
Cited By (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN116645336A (en) * | 2023-05-10 | 2023-08-25 | 烟台大学 | MRI brain image gland pituitary segmentation method |
CN118379303A (en) * | 2024-06-21 | 2024-07-23 | 华侨大学 | U-Net pathological cell segmentation method and device based on multiscale transducer multi-attention |
Citations (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN109598727A (en) * | 2018-11-28 | 2019-04-09 | 北京工业大学 | A kind of CT image pulmonary parenchyma three-dimensional semantic segmentation method based on deep neural network |
US20210110258A1 (en) * | 2019-10-14 | 2021-04-15 | Samsung Electronics Co., Ltd. | Method and apparatus with model training and/or sequence recognition |
CN112991269A (en) * | 2021-02-07 | 2021-06-18 | 复旦大学 | Identification and classification method for lung CT image |
CN113469094A (en) * | 2021-07-13 | 2021-10-01 | 上海中科辰新卫星技术有限公司 | Multi-mode remote sensing data depth fusion-based earth surface coverage classification method |
CN113888744A (en) * | 2021-10-14 | 2022-01-04 | 浙江大学 | Image semantic segmentation method based on Transformer visual upsampling module |
-
2022
- 2022-02-08 CN CN202210117033.9A patent/CN114581459A/en active Pending
Patent Citations (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN109598727A (en) * | 2018-11-28 | 2019-04-09 | 北京工业大学 | A kind of CT image pulmonary parenchyma three-dimensional semantic segmentation method based on deep neural network |
US20210110258A1 (en) * | 2019-10-14 | 2021-04-15 | Samsung Electronics Co., Ltd. | Method and apparatus with model training and/or sequence recognition |
CN112991269A (en) * | 2021-02-07 | 2021-06-18 | 复旦大学 | Identification and classification method for lung CT image |
CN113469094A (en) * | 2021-07-13 | 2021-10-01 | 上海中科辰新卫星技术有限公司 | Multi-mode remote sensing data depth fusion-based earth surface coverage classification method |
CN113888744A (en) * | 2021-10-14 | 2022-01-04 | 浙江大学 | Image semantic segmentation method based on Transformer visual upsampling module |
Non-Patent Citations (5)
Title |
---|
COAST_S: "如何理解 Transformer 中的 Query、Key 与 Valu", 《HTTPS://BLOG.CSDN.NET/WEIXIN_43996337/ARTICLE/DETAILS/121187666》 * |
D.BLANC 等: "Artificial intelligence solution to classify pulmonary nodules on CT", 《DIAGNOSTIC AND INTERVENTIONAL IMAGING》 * |
HAONAN WANG 等: "UCTransNet: Rethinking the Skip Connections in U-Net from a Channel-wise Perspective with Transformer", 《ARXIV:2109.04335V3》 * |
JINGJING WANG 等: "DA-ResUNet: a novel method for brain tumor segmentation based on U-Net with residual block and CBAM", 《EVENT: FOURTH INTERNATIONAL CONFERENCE ON IMAGE, VIDEO PROCESSING, AND ARTIFICIAL INTELLIGENCE (IVPAI 2021)》 * |
OLIVIER PETIT 等: "U-Net Transformer: Self and Cross Attention for Medical Image Segmentation", 《ARXIV:2103.06104V2》 * |
Cited By (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN116645336A (en) * | 2023-05-10 | 2023-08-25 | 烟台大学 | MRI brain image gland pituitary segmentation method |
CN116645336B (en) * | 2023-05-10 | 2024-05-07 | 烟台大学 | MRI brain image gland pituitary segmentation method |
CN118379303A (en) * | 2024-06-21 | 2024-07-23 | 华侨大学 | U-Net pathological cell segmentation method and device based on multiscale transducer multi-attention |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN113506334B (en) | Multi-mode medical image fusion method and system based on deep learning | |
CN113450294A (en) | Multi-modal medical image registration and fusion method and device and electronic equipment | |
Sreelakshmy et al. | [Retracted] An Automated Deep Learning Model for the Cerebellum Segmentation from Fetal Brain Images | |
CN113177943B (en) | Cerebral apoplexy CT image segmentation method | |
CN112529909A (en) | Tumor image brain region segmentation method and system based on image completion | |
CN115496771A (en) | Brain tumor segmentation method based on brain three-dimensional MRI image design | |
CN114119516B (en) | Virus focus segmentation method based on migration learning and cascade self-adaptive cavity convolution | |
CN114494296A (en) | Brain glioma segmentation method and system based on fusion of Unet and Transformer | |
CN114581459A (en) | Improved 3D U-Net model-based segmentation method for image region of interest of preschool child lung | |
Liu et al. | A fully automatic segmentation algorithm for CT lung images based on random forest | |
CN113096137B (en) | Adaptive segmentation method and system for OCT (optical coherence tomography) retinal image field | |
CN112884788B (en) | Cup optic disk segmentation method and imaging method based on rich context network | |
CN115147600A (en) | GBM multi-mode MR image segmentation method based on classifier weight converter | |
CN114882048A (en) | Image segmentation method and system based on wavelet scattering learning network | |
CN113456031A (en) | Training device and prediction device of brain state prediction model and electronic equipment | |
CN116309615A (en) | Multi-mode MRI brain tumor image segmentation method | |
CN117274599A (en) | Brain magnetic resonance segmentation method and system based on combined double-task self-encoder | |
CN118037791A (en) | Construction method and application of multi-mode three-dimensional medical image segmentation registration model | |
CN116485853A (en) | Medical image registration method and device based on deep learning neural network | |
CN116030063A (en) | Classification diagnosis system, method, electronic device and medium for MRI image | |
CN115410032A (en) | OCTA image classification structure training method based on self-supervision learning | |
CN115294023A (en) | Liver tumor automatic segmentation method and device | |
CN115409812A (en) | CT image automatic classification method based on fusion time attention mechanism | |
Mani | Deep learning models for semantic multi-modal medical image segmentation | |
CN112967295A (en) | Image processing method and system based on residual error network and attention mechanism |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
RJ01 | Rejection of invention patent application after publication |
Application publication date: 20220603 |
|
RJ01 | Rejection of invention patent application after publication |