CN111612790A - Medical image segmentation method based on T-shaped attention structure - Google Patents

Medical image segmentation method based on T-shaped attention structure Download PDF

Info

Publication number
CN111612790A
CN111612790A CN202010355011.7A CN202010355011A CN111612790A CN 111612790 A CN111612790 A CN 111612790A CN 202010355011 A CN202010355011 A CN 202010355011A CN 111612790 A CN111612790 A CN 111612790A
Authority
CN
China
Prior art keywords
training
attention
network structure
layer
medical image
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN202010355011.7A
Other languages
Chinese (zh)
Other versions
CN111612790B (en
Inventor
颜成钢
杨祥宇
杨建�
张二四
孙垚棋
张继勇
张勇东
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Hangzhou Dianzi University
Original Assignee
Hangzhou Dianzi University
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Hangzhou Dianzi University filed Critical Hangzhou Dianzi University
Priority to CN202010355011.7A priority Critical patent/CN111612790B/en
Publication of CN111612790A publication Critical patent/CN111612790A/en
Application granted granted Critical
Publication of CN111612790B publication Critical patent/CN111612790B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/10Segmentation; Edge detection
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/045Combinations of networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/08Learning methods
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/10Image acquisition modality
    • G06T2207/10004Still image; Photographic image
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/10Image acquisition modality
    • G06T2207/10024Color image
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/20Special algorithmic details
    • G06T2207/20021Dividing image into blocks, subimages or windows
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/20Special algorithmic details
    • G06T2207/20081Training; Learning
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/20Special algorithmic details
    • G06T2207/20084Artificial neural networks [ANN]
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/30Subject of image; Context of image processing
    • G06T2207/30004Biomedical image processing
    • G06T2207/30041Eye; Retina; Ophthalmic
    • YGENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
    • Y02TECHNOLOGIES OR APPLICATIONS FOR MITIGATION OR ADAPTATION AGAINST CLIMATE CHANGE
    • Y02TCLIMATE CHANGE MITIGATION TECHNOLOGIES RELATED TO TRANSPORTATION
    • Y02T10/00Road transport of goods or passengers
    • Y02T10/10Internal combustion engine [ICE] based vehicles
    • Y02T10/40Engine management systems

Abstract

The invention provides a medical image segmentation method based on a T-shaped attention structure. The invention provides an improved U-Net model for medical image segmentation, wherein a gate control attention mechanism with a T attention module is introduced into a decoding network structure of a U-Net network, and a ResNet residual module is added into the encoding network structure and the decoding network structure; through the gated attention mechanism with the T attention module, the global feature information of the shallow input image can be extracted, and the problem of insufficient image features caused by direct jumping is solved. And the T attention modules share model parameters, so that the time and space complexity is reduced. And a residual error module is added for each convolution layer, so that the problems of gradient loss and explosion in the training process are solved, and the optimization is easier during network training.

Description

Medical image segmentation method based on T-shaped attention structure
Technical Field
The invention belongs to the technical field of deep learning, particularly relates to a novel U-Net medical image segmentation method based on a T-shaped attention structure, and aims at medical image segmentation.
Background
At present, due to the rapid development of deep learning, great influence is generated on many fields. As a difficult visual task, medical image segmentation encounters the diversity and individual characteristics of imaging data sets that are important reasons for the persistent interest: the data set is extremely different when considering the queue size, image dimensions, image size, voxel intensity range and intensity interpretation. Class labels in images may be highly unbalanced, labels may be ambiguous, and the quality of expert annotations varies greatly between different data sets. Furthermore, some data sets are very inhomogeneous in terms of image geometry, or slice misplacement may occur.
In the current medical image segmentation field, the application is the most extensive, the best effect is U-Net, and the adopted encoder (down sampling) -decoder (up sampling) structure and jump connection are a very classical design method. However, the skip structure of the U-Net network has a disadvantage that the result of up-sampling at each step is connected to the decoding layer without processing, so that the shallow layer characteristics may not be sufficiently transformed.
In recent years, the semantic segmentation framework based on the full convolution network has made remarkable progress, but some problems still remain to be solved, and the medical image may have high correlation among pixels of the whole image, so that the required receptive field is larger than that of a general image, and the global receptive field can generate the most abundant information. However, the network generated by FCN limits the local receptive field and small range of context information, and the hole convolution in the proposed aspp can only collect information of some surrounding pixels, and cannot generate tight context information, as does pyramid pooling.
Some attention-based methods proposed by this expert scholarly require the generation of large attention maps to measure the relationship of each pair of pixels. The effect is improved, but the complexity of time and space is high, the calculated amount and the occupied memory are large, and the training difficulty is increased.
Disclosure of Invention
The invention provides a medical image segmentation method based on a T-shaped attention structure, aiming at the problems that the possible transformation of shallow features is insufficient and the calculated amount of an attention mechanism and the occupied memory are reduced, and the gradient disappears and explodes in training caused by the direct fusion of the traditional U-Net intermediate jump structure.
The invention comprises the following steps:
firstly, preprocessing medical image data, and dividing the data to obtain a training set and a test set;
step two, improving the U-Net network to obtain a prediction model:
a gate control attention mechanism with a T attention module is introduced into a decoding network structure of the U-Net network, and a ResNet residual module is added into the encoding network structure and the decoding network structure;
inputting the training set data into a prediction model for training;
and inputting the training set into a prediction model, and adopting a random initialization and random gradient descent optimization method. Setting an initial learning rate, a segmentation layer learning rate, momentum and a weight attenuation coefficient. Training according to a set training strategy to obtain a trained network model;
inputting the test set data set into the trained prediction model in the third step to obtain segmentation data;
further, preprocessing the acquired medical image data, and dividing the data to obtain a training set, a verification set and a test set, wherein the specific operations are as follows;
unifying the size of the medical image into n x n square shape, taking eighty percent of samples in the medical image data set as a training set, and taking the rest twenty percent of samples as a testing set. In the training and testing stage, a patch-based method based on image blocking is adopted, and each picture of the training set is divided into a plurality of small blocks, and each small block is called a patch. The training set was divided into patch 90% as training data and 10% as validation data.
Preferably, the prediction model in step two includes: the encoding network structure comprises 4 convolutional layers and 3 maximum pooling layers, and the decoding network structure comprises 3 upsampling layers and 3 convolutional layers. Wherein each convolutional layer contains three convolutional block operations, replacing the last two convolutional blocks of each convolutional layer in the coding network structure and the decoding network structure with residual convolutional blocks. In order to combine the prediction of the last layer and the previous shallower layer of the coding network structure, jump structures are added to connect the last layer and the previous shallower layer of the coding network structure, and a gating attention structure is added to each layer of jump structures to perform shallow input feature extraction.
The gated attention structure firstly fuses an input coded image and a convolved decoded image, performs dimensionality reduction operation by using a 1 x 1 convolution layer, inputs the dimensionality reduced image into two series T-shaped attention modules, and fuses an output obtained by the T-shaped attention modules and the input coded image to obtain a feature image containing global information.
Furthermore, the T-shaped attention modules collect relevant information from the horizontal direction and the vertical direction, the representativeness of a pixel level is enhanced, for a given local feature H, 1 × 1 convolution operation is firstly carried out to obtain Q, K and V, each pixel point in the Q and all pixels in the same row and column corresponding to the point in the K are subjected to fusion operation to obtain A, the result obtained by the fusion operation of the A and the V is further fused with the H to obtain the relevance of each pixel point and the pixel points in the same row and column, the two T-shaped attention modules are connected in series to obtain global information, and the two T-shaped attention modules share parameter information.
And further, inputting training set data into the prediction model for training, and adopting a random initialization and random gradient descent optimization method, wherein the random gradient descent optimization method comprises the steps of randomly extracting a group of samples from the input training set for training, updating once according to the gradient after training, then re-extracting a group of samples for training, and updating once again. And setting the initial learning rate, the segmentation layer learning rate, the momentum, the weight attenuation coefficient and the iteration times in sequence. And inputting the verification set after the training is finished to obtain a verification result.
The invention has the following beneficial effects:
the invention provides an improved U-Net model for medical image segmentation, which can extract global feature information of a shallow input image through a gating attention mechanism with a T attention module and solves the problem of insufficient image features caused by directly performing a jump structure. And the T attention modules share model parameters, so that the time and space complexity is reduced. And a residual error module is added for each convolution layer, so that the problems of gradient loss and explosion in the training process are solved, and the optimization is easier during network training.
Drawings
FIG. 1 is a block diagram of an embodiment of the method of the present invention;
FIG. 2 is a diagram of a gated attention structure in an embodiment of the method of the present invention;
FIG. 3 is a T-shaped attention module diagram according to an embodiment of the method of the present invention.
Detailed Description
For the purpose of better explaining the present invention and to facilitate understanding, the present invention will be described in detail by way of specific embodiments with reference to the accompanying drawings. The invention is intended to cover alternatives, modifications, equivalents and alternatives which may be included within the spirit and scope of the invention as defined by the appended claims. In the following detailed description of the present invention, certain specific details are set forth in order to provide a better understanding of the present invention. It will be apparent to one skilled in the art that the present invention may be practiced without these specific details.
The following detailed description of embodiments of the invention refers to the accompanying drawings.
Example one
As shown in fig. 1, a method for segmenting a medical image based on a T-shaped attention structure includes the following steps:
firstly, preprocessing medical image data, and dividing the data to obtain a training set and a test set;
unifying the size of the medical image into n x n square shape, taking eighty percent of samples in the medical image data set as a training set, and taking the rest twenty percent of samples as a testing set. Due to the data scarcity of the data set, an image-based patch-based method is adopted in the training and testing stage, and each picture of the training set is divided into a plurality of small blocks, and each small block is called a patch. The training set was divided into patch 90% as training data and 10% as validation data.
In this example, the medical data set used a DRIVE data set consisting of 40 color retinal images, 32 of which were used for training and the remaining 8 were used for testing. Each original image is of size 565 x 584 pixels, and in order to be a square shape with the medical image size being uniform to n x n, the image is cut to contain only data from column 9 to column 574, and each image is of size 565 x 565 pixels. Each picture of the training set is divided into a plurality of patches, each patch being called a patch. Randomly partitioning the training set image into 320000 patches, wherein 288000 patches are used for training and the remaining 32000 are used for verification;
step two, improving the U-Net network to obtain a prediction model:
the U-Net network model comprises an encoding network structure and a decoding network structure, wherein the encoding structure is used for extracting the characteristics of input characteristics, and the decoding structure predicts a segmentation model of an input image on the basis of extracting the characteristics, and the segmentation model has the same size as the input image. A gated attention mechanism with a T attention module is introduced into a decoding network structure of the U-Net network and is used for transferring image characteristics of each layer of the encoding network to the decoding network. In order to solve the problems of gradient loss and gradient explosion in training, a ResNet residual module is added into an encoding network structure and a decoding network structure;
the U-Net network model comprises: the encoding network structure comprises 4 convolutional layers and 3 maximum pooling layers, and the decoding network structure comprises 3 upsampling layers and 3 convolutional layers. Wherein each convolutional layer contains three convolutional block operations, replacing the last two convolutional blocks of each convolutional layer in the coding network structure and the decoding network structure with residual convolutional blocks. In order to combine the prediction of the last layer and the previous shallower layer of the coding network structure, jump structures are added to connect the last layer and the previous shallower layer of the coding network structure, and a gating attention structure is added to each layer of jump structures to perform shallow input feature extraction.
The gated attention structure firstly fuses an input coded image and a convolved decoded image, performs dimensionality reduction operation by using a 1 x 1 convolution layer, inputs the dimensionality reduced image into two series T-shaped attention modules, and fuses an output obtained by the T-shaped attention modules and the input coded image to obtain a feature image containing global information.
In order to reduce the computational complexity and obtain sufficient global information, the T-shaped attention modules collect relevant information from the horizontal direction and the vertical direction and enhance the representativeness of pixel levels, for a given local feature H, 1 × 1 convolution operation is firstly carried out to obtain Q, K and V, each pixel point in Q and all pixels in the same row and column corresponding to the point in K are subjected to fusion operation to obtain A, the result obtained by the fusion operation of A and V is fused with H to obtain the relevance of each pixel point and the same pixel point in the same row and column, two T-shaped attention modules are connected in series to obtain the global information, and the two T-shaped attention modules share parameter information.
Inputting the training set data into a prediction model for training;
inputting a training set into a prediction model, and adopting a random initialization and random gradient descent optimization method, wherein a group of samples are randomly extracted from the input training set for training, the training is carried out once according to the gradient after the training, then a group of samples are extracted again for training, the updating is carried out once again, under the condition of large sample quantity, a model with a loss value within an acceptable range can be obtained without training all the samples, and the samples are randomly disturbed in each iteration process. Setting an initial learning rate, a segmentation layer learning rate, momentum and a weight attenuation coefficient. Training according to a set training strategy to obtain a trained network model;
the batch size for stochastic gradient descent optimization was 32, epoch was 150, initial learning rate was 0.001, final segmentation layer learning rate was 0.01, momentum was 0.9, and weight decay was 0.0005.
And inputting the verification set after the training is finished to obtain a verification result.
Inputting the test set data set into the trained prediction model in the third step to obtain segmentation data;
establishing evaluation indexes including Accuracy (AC), Sensitivity (SE), Specificity (SP), F1 score, Dice Coefficient (DC) and Jacard index (JA), wherein the evaluation indexes are calculated in the following mode:
Figure BDA0002473151620000061
Figure BDA0002473151620000062
wherein
Figure BDA0002473151620000063
Wherein TP is true positive, TN is true negative, FP is false positive, and FN is false negative.

Claims (5)

1. A medical image segmentation method based on a T-shaped attention structure is characterized by comprising the following steps:
firstly, preprocessing medical image data, and dividing the data to obtain a training set and a test set;
step two, improving the U-Net network to obtain a prediction model:
a gate control attention mechanism with a T attention module is introduced into a decoding network structure of the U-Net network, and a ResNet residual module is added into the encoding network structure and the decoding network structure;
inputting the training set data into a prediction model for training;
inputting the training set into a prediction model, and adopting a random initialization and random gradient descent optimization method; setting an initial learning rate, a segmentation layer learning rate, momentum and a weight attenuation coefficient; training according to a set training strategy to obtain a trained network model;
and step four, inputting the test set data set into the well-trained prediction model in the step three to obtain the segmentation data.
2. The medical image segmentation method based on the T-shaped attention structure according to claim 1, characterized in that, further, the steps of preprocessing the acquired medical image data and dividing the data to obtain a training set, a verification set and a test set are as follows;
unifying the sizes of the medical images into a square shape of n x n, taking eighty percent of samples in the medical image data set as a training set, and taking the rest twenty percent of samples as a testing set; in the training and testing stage, a patch-based method based on image blocking is adopted, each picture of a training set is divided into a plurality of small blocks, and each small block is called a patch; the training set was divided into patch 90% as training data and 10% as validation data.
3. A T-shaped attention structure-based medical image segmentation method as claimed in claim 2, wherein preferably, the prediction model in step two comprises: the network structure comprises an encoding network structure and a decoding network structure, wherein the encoding network structure comprises 4 convolutional layers and 3 maximum pooling layers, and the decoding network structure comprises 3 upsampling layers and 3 convolutional layers; wherein each convolutional layer comprises three convolutional block operations, the last two convolutional blocks of each convolutional layer in the coding network structure and the decoding network structure are replaced by residual convolutional blocks; in order to combine the prediction of the last layer and the previous shallower layer of the coding network structure, a jump structure is added to connect the last layer and the previous shallower layer, and a gating attention structure is added to each layer of jump structure to extract the input features of the shallower layer;
the gated attention structure firstly fuses an input coded image and a convolved decoded image, performs dimensionality reduction operation by using a 1 x 1 convolution layer, inputs the dimensionality reduced image into two series T-shaped attention modules, and fuses an output obtained by the T-shaped attention modules and the input coded image to obtain a characteristic image containing global information;
furthermore, in order to reduce the computational complexity and obtain sufficient global information, the T-shaped attention module collects relevant information from the horizontal and vertical directions, enhances the representativeness of the pixel level, firstly performs 1 × 1 convolution operation on a given local feature H to obtain Q, K and V, performs fusion operation on each pixel point in Q and all pixels in the same row and column corresponding to the point in K to obtain a, fuses the result obtained by performing the fusion operation on a and V and H to obtain the relevance between each pixel point and the pixel point in the same row and column, and connects two T-shaped attention modules in series to obtain the global information, and the two T-shaped attention modules share parameter information.
4. The method for segmenting the medical image based on the T-shaped attention structure according to claim 3, characterized in that, further, in the third step, the training set data is input into the prediction model for training, and the random initialization and random gradient descent optimization method is adopted, wherein, firstly, a group of samples are randomly extracted from the input training set for training, after the training, the group of samples are updated according to the gradient, then, a group of samples are extracted again for training, and then, the samples are updated again; setting an initial learning rate, a segmentation layer learning rate, momentum, a weight attenuation coefficient and iteration times in sequence; and inputting the verification set after the training is finished to obtain a verification result.
5. The method of claim 4, wherein the batch size of the stochastic gradient descent optimization is 32, the epoch is 150, the initial learning rate is 0.001, the last segmentation layer learning rate is 0.01, the momentum is 0.9, and the weight attenuation is 0.0005.
CN202010355011.7A 2020-04-29 2020-04-29 Medical image segmentation method based on T-shaped attention structure Active CN111612790B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202010355011.7A CN111612790B (en) 2020-04-29 2020-04-29 Medical image segmentation method based on T-shaped attention structure

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202010355011.7A CN111612790B (en) 2020-04-29 2020-04-29 Medical image segmentation method based on T-shaped attention structure

Publications (2)

Publication Number Publication Date
CN111612790A true CN111612790A (en) 2020-09-01
CN111612790B CN111612790B (en) 2023-10-17

Family

ID=72205614

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202010355011.7A Active CN111612790B (en) 2020-04-29 2020-04-29 Medical image segmentation method based on T-shaped attention structure

Country Status (1)

Country Link
CN (1) CN111612790B (en)

Cited By (13)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN112183507A (en) * 2020-11-30 2021-01-05 北京沃东天骏信息技术有限公司 Image segmentation method, device, equipment and storage medium
CN112381172A (en) * 2020-11-28 2021-02-19 桂林电子科技大学 InSAR interference image phase unwrapping method based on U-net
CN112651979A (en) * 2021-01-11 2021-04-13 华南农业大学 Lung X-ray image segmentation method, system, computer equipment and storage medium
CN112669285A (en) * 2020-12-29 2021-04-16 中山大学 Fundus image blood vessel segmentation method based on shared decoder and residual error tower type structure
CN113139972A (en) * 2021-03-22 2021-07-20 杭州电子科技大学 Cerebral apoplexy MRI image focus region segmentation method based on artificial intelligence
CN113160240A (en) * 2021-03-09 2021-07-23 温州医科大学附属眼视光医院 Cyclic hopping deep learning network
CN113159056A (en) * 2021-05-21 2021-07-23 中国科学院深圳先进技术研究院 Image segmentation method, device, equipment and storage medium
CN113240691A (en) * 2021-06-10 2021-08-10 南京邮电大学 Medical image segmentation method based on U-shaped network
CN113326851A (en) * 2021-05-21 2021-08-31 中国科学院深圳先进技术研究院 Image feature extraction method and device, electronic equipment and storage medium
CN113837193A (en) * 2021-09-23 2021-12-24 中南大学 Zinc flotation froth image segmentation algorithm based on improved U-Net network
CN114757938A (en) * 2022-05-16 2022-07-15 国网四川省电力公司电力科学研究院 Transformer oil leakage identification method and system
CN115131612A (en) * 2022-07-02 2022-09-30 哈尔滨理工大学 Retina OCT image classification method based on recursive residual error network
CN115410070A (en) * 2022-11-01 2022-11-29 宏景科技股份有限公司 Crop disease condition training and evaluating method based on improved Unet network structure

Citations (13)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20120189161A1 (en) * 2011-01-25 2012-07-26 Electronics And Telecommunications Research Institute Visual attention apparatus and control method based on mind awareness and display apparatus using the visual attention apparatus
CN103345763A (en) * 2013-06-25 2013-10-09 西安理工大学 Method for calculating motion attention based on multiscale variable-block
CN107291945A (en) * 2017-07-12 2017-10-24 上海交通大学 The high-precision image of clothing search method and system of view-based access control model attention model
WO2018068532A1 (en) * 2016-10-11 2018-04-19 京东方科技集团股份有限公司 Image encoding and decoding devices, image processing system, image encoding and decoding methods, and training method
CN108537793A (en) * 2018-04-17 2018-09-14 电子科技大学 A kind of pulmonary nodule detection method based on improved u-net networks
CN109191472A (en) * 2018-08-28 2019-01-11 杭州电子科技大学 Based on the thymocyte image partition method for improving U-Net network
CN109509178A (en) * 2018-10-24 2019-03-22 苏州大学 A kind of OCT image choroid dividing method based on improved U-net network
CN110189334A (en) * 2019-05-28 2019-08-30 南京邮电大学 The medical image cutting method of the full convolutional neural networks of residual error type based on attention mechanism
CN110298844A (en) * 2019-06-17 2019-10-01 艾瑞迈迪科技石家庄有限公司 X-ray contrastographic picture blood vessel segmentation and recognition methods and device
US10430946B1 (en) * 2019-03-14 2019-10-01 Inception Institute of Artificial Intelligence, Ltd. Medical image segmentation and severity grading using neural network architectures with semi-supervised learning techniques
CN110675406A (en) * 2019-09-16 2020-01-10 南京信息工程大学 CT image kidney segmentation algorithm based on residual double-attention depth network
EP3611699A1 (en) * 2018-08-14 2020-02-19 Siemens Healthcare GmbH Image segmentation using deep learning techniques
CN110889852A (en) * 2018-09-07 2020-03-17 天津大学 Liver segmentation method based on residual error-attention deep neural network

Patent Citations (13)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20120189161A1 (en) * 2011-01-25 2012-07-26 Electronics And Telecommunications Research Institute Visual attention apparatus and control method based on mind awareness and display apparatus using the visual attention apparatus
CN103345763A (en) * 2013-06-25 2013-10-09 西安理工大学 Method for calculating motion attention based on multiscale variable-block
WO2018068532A1 (en) * 2016-10-11 2018-04-19 京东方科技集团股份有限公司 Image encoding and decoding devices, image processing system, image encoding and decoding methods, and training method
CN107291945A (en) * 2017-07-12 2017-10-24 上海交通大学 The high-precision image of clothing search method and system of view-based access control model attention model
CN108537793A (en) * 2018-04-17 2018-09-14 电子科技大学 A kind of pulmonary nodule detection method based on improved u-net networks
EP3611699A1 (en) * 2018-08-14 2020-02-19 Siemens Healthcare GmbH Image segmentation using deep learning techniques
CN109191472A (en) * 2018-08-28 2019-01-11 杭州电子科技大学 Based on the thymocyte image partition method for improving U-Net network
CN110889852A (en) * 2018-09-07 2020-03-17 天津大学 Liver segmentation method based on residual error-attention deep neural network
CN109509178A (en) * 2018-10-24 2019-03-22 苏州大学 A kind of OCT image choroid dividing method based on improved U-net network
US10430946B1 (en) * 2019-03-14 2019-10-01 Inception Institute of Artificial Intelligence, Ltd. Medical image segmentation and severity grading using neural network architectures with semi-supervised learning techniques
CN110189334A (en) * 2019-05-28 2019-08-30 南京邮电大学 The medical image cutting method of the full convolutional neural networks of residual error type based on attention mechanism
CN110298844A (en) * 2019-06-17 2019-10-01 艾瑞迈迪科技石家庄有限公司 X-ray contrastographic picture blood vessel segmentation and recognition methods and device
CN110675406A (en) * 2019-09-16 2020-01-10 南京信息工程大学 CT image kidney segmentation algorithm based on residual double-attention depth network

Non-Patent Citations (4)

* Cited by examiner, † Cited by third party
Title
DONG LI ETAL: "Unified Spatio-Temporal Attention Networks for Action Recognition in Videos", IEEE TRANSACTIONS ON MULTIMEDIA *
严广宇;刘正熙;: "基于混合注意力的实时语义分割算法", 现代计算机, no. 10 *
梅旭璋;江红;孙军;: "基于密集注意力网络的视网膜血管图像分割", 计算机工程, no. 03 *
边小勇等: "基于尺度注意力网络的遥感图像场景分类", 计算机应用 *

Cited By (23)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN112381172A (en) * 2020-11-28 2021-02-19 桂林电子科技大学 InSAR interference image phase unwrapping method based on U-net
CN112381172B (en) * 2020-11-28 2022-09-16 桂林电子科技大学 InSAR interference image phase unwrapping method based on U-net
CN112183507B (en) * 2020-11-30 2021-03-19 北京沃东天骏信息技术有限公司 Image segmentation method, device, equipment and storage medium
CN112183507A (en) * 2020-11-30 2021-01-05 北京沃东天骏信息技术有限公司 Image segmentation method, device, equipment and storage medium
CN112669285B (en) * 2020-12-29 2022-03-08 中山大学 Fundus image blood vessel segmentation method based on shared decoder and residual error tower type structure
CN112669285A (en) * 2020-12-29 2021-04-16 中山大学 Fundus image blood vessel segmentation method based on shared decoder and residual error tower type structure
CN112651979A (en) * 2021-01-11 2021-04-13 华南农业大学 Lung X-ray image segmentation method, system, computer equipment and storage medium
CN112651979B (en) * 2021-01-11 2023-10-10 华南农业大学 Lung X-ray image segmentation method, system, computer equipment and storage medium
CN113160240A (en) * 2021-03-09 2021-07-23 温州医科大学附属眼视光医院 Cyclic hopping deep learning network
CN113139972A (en) * 2021-03-22 2021-07-20 杭州电子科技大学 Cerebral apoplexy MRI image focus region segmentation method based on artificial intelligence
WO2022242131A1 (en) * 2021-05-21 2022-11-24 中国科学院深圳先进技术研究院 Image segmentation method and apparatus, device, and storage medium
CN113326851A (en) * 2021-05-21 2021-08-31 中国科学院深圳先进技术研究院 Image feature extraction method and device, electronic equipment and storage medium
CN113159056A (en) * 2021-05-21 2021-07-23 中国科学院深圳先进技术研究院 Image segmentation method, device, equipment and storage medium
CN113326851B (en) * 2021-05-21 2023-10-27 中国科学院深圳先进技术研究院 Image feature extraction method and device, electronic equipment and storage medium
CN113159056B (en) * 2021-05-21 2023-11-21 中国科学院深圳先进技术研究院 Image segmentation method, device, equipment and storage medium
CN113240691A (en) * 2021-06-10 2021-08-10 南京邮电大学 Medical image segmentation method based on U-shaped network
WO2022257408A1 (en) * 2021-06-10 2022-12-15 南京邮电大学 Medical image segmentation method based on u-shaped network
CN113837193A (en) * 2021-09-23 2021-12-24 中南大学 Zinc flotation froth image segmentation algorithm based on improved U-Net network
CN113837193B (en) * 2021-09-23 2023-09-01 中南大学 Zinc flotation froth image segmentation method based on improved U-Net network
CN114757938A (en) * 2022-05-16 2022-07-15 国网四川省电力公司电力科学研究院 Transformer oil leakage identification method and system
CN114757938B (en) * 2022-05-16 2023-09-15 国网四川省电力公司电力科学研究院 Transformer oil leakage identification method and system
CN115131612A (en) * 2022-07-02 2022-09-30 哈尔滨理工大学 Retina OCT image classification method based on recursive residual error network
CN115410070A (en) * 2022-11-01 2022-11-29 宏景科技股份有限公司 Crop disease condition training and evaluating method based on improved Unet network structure

Also Published As

Publication number Publication date
CN111612790B (en) 2023-10-17

Similar Documents

Publication Publication Date Title
CN111612790B (en) Medical image segmentation method based on T-shaped attention structure
CN113159051B (en) Remote sensing image lightweight semantic segmentation method based on edge decoupling
US20220004744A1 (en) Human posture detection method and apparatus, device and storage medium
CN111524144B (en) Intelligent lung nodule diagnosis method based on GAN and Unet network
CN107742061A (en) A kind of prediction of protein-protein interaction mthods, systems and devices
CN113468996B (en) Camouflage object detection method based on edge refinement
EP4016454A1 (en) Three-dimensional edge detection method and apparatus, storage medium and computer device
CN107247952B (en) Deep supervision-based visual saliency detection method for cyclic convolution neural network
CN111046793B (en) Tomato disease identification method based on deep convolutional neural network
CN112348830B (en) Multi-organ segmentation method based on improved 3D U-Net
CN112686902A (en) Two-stage calculation method for brain glioma identification and segmentation in nuclear magnetic resonance image
CN112288749A (en) Skull image segmentation method based on depth iterative fusion depth learning model
CN115393293A (en) Electron microscope red blood cell segmentation and positioning method based on UNet network and watershed algorithm
CN115761297A (en) Method for automatically identifying landslide by attention neural network based on edge guidance
CN114511710A (en) Image target detection method based on convolutional neural network
CN114092815A (en) Remote sensing intelligent extraction method for large-range photovoltaic power generation facility
CN116012653A (en) Method and system for classifying hyperspectral images of attention residual unit neural network
CN111310820A (en) Foundation meteorological cloud chart classification method based on cross validation depth CNN feature integration
CN117422691A (en) Concrete crack detection method and system combining linear guide and grid optimization
CN112837276A (en) Brain glioma segmentation method based on cascaded deep neural network model
CN116778318A (en) Convolutional neural network remote sensing image road extraction model and method
CN114140472B (en) Cross-level information fusion medical image segmentation method
CN113962332A (en) Salient target identification method based on self-optimization fusion feedback
CN114529794A (en) Infrared and visible light image fusion method, system and medium
CN114581789A (en) Hyperspectral image classification method and system

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant