CN114897914A - Semi-supervised CT image segmentation method based on confrontation training - Google Patents

Semi-supervised CT image segmentation method based on confrontation training Download PDF

Info

Publication number
CN114897914A
CN114897914A CN202210259206.0A CN202210259206A CN114897914A CN 114897914 A CN114897914 A CN 114897914A CN 202210259206 A CN202210259206 A CN 202210259206A CN 114897914 A CN114897914 A CN 114897914A
Authority
CN
China
Prior art keywords
image
generator
discriminator
loss
mask
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN202210259206.0A
Other languages
Chinese (zh)
Other versions
CN114897914B (en
Inventor
孙仕亮
丁超越
赵静
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
East China Normal University
Original Assignee
East China Normal University
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by East China Normal University filed Critical East China Normal University
Priority to CN202210259206.0A priority Critical patent/CN114897914B/en
Publication of CN114897914A publication Critical patent/CN114897914A/en
Application granted granted Critical
Publication of CN114897914B publication Critical patent/CN114897914B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/10Segmentation; Edge detection
    • G06T7/11Region-based segmentation
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/10Image acquisition modality
    • G06T2207/10072Tomographic images
    • G06T2207/10081Computed x-ray tomography [CT]
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/20Special algorithmic details
    • G06T2207/20081Training; Learning
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/30Subject of image; Context of image processing
    • G06T2207/30004Biomedical image processing
    • G06T2207/30061Lung
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/30Subject of image; Context of image processing
    • G06T2207/30004Biomedical image processing
    • G06T2207/30096Tumor; Lesion
    • YGENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
    • Y02TECHNOLOGIES OR APPLICATIONS FOR MITIGATION OR ADAPTATION AGAINST CLIMATE CHANGE
    • Y02TCLIMATE CHANGE MITIGATION TECHNOLOGIES RELATED TO TRANSPORTATION
    • Y02T10/00Road transport of goods or passengers
    • Y02T10/10Internal combustion engine [ICE] based vehicles
    • Y02T10/40Engine management systems

Abstract

The invention relates to the technical field of image processing, in particular to a semi-supervised CT image segmentation method based on confrontation training, which comprises the following steps: firstly, acquiring a three-dimensional CT image of a lung, and establishing a diseased data set labeled at a voxel level, a diseased data set not labeled and a healthy data set; and then, sequentially inputting the CT images in the three data sets to a generator and a divider to respectively obtain a synthesized healthy image and a division mask. Then splicing the mask area of the synthesized health image and the reverse mask area of the input image to obtain a recovered health image; setting the value of the mask region of the synthesized healthy image to 0 results in an anti-mask image. And finally, the two discriminators respectively supervise the recovered health image and the anti-mask image in an antagonistic training mode to improve the segmentation and generator effects of the segmenter. The invention realizes that a segmentation model with accurate performance is obtained by using a small amount of voxel level marking samples, and the specially designed segmenter effectively improves the feature representation capability.

Description

Semi-supervised CT image segmentation method based on confrontation training
Technical Field
The invention relates to the technical field of computers, in particular to a semi-supervised CT image segmentation method based on confrontation training.
Background
The background art involves: lung CT image segmentation and attention mechanism.
1) Lung CT image segmentation
Lung CT has greater accuracy in diagnosing lung disease and has been used in many applications for medical images of the lungs using deep learning techniques. However, much of the previous work has been directed to the task of classifying images of the lungs, which do not reflect the location and size of the lesion as well as the task of segmentation. Segmentation of the infected region using CT images may help radiologists to better quantify the lesion area. Quantitative analysis of the segmentation mask of the lesion area can lead to a series of meaningful diagnosis results related to the lung.
At present, deep learning has been primarily applied to the task of CT segmentation of the lungs, but this method requires a large amount of training data. Clinically, the three-dimensional CT image data annotation of the lung is very difficult, and it takes more than 3 hours for a radiologist to annotate a lung CT volume. Therefore, scarcity of voxel-level labeled lung CT images is an important issue. Image synthesis and data enhancement can alleviate the problem of voxel level label loss. Active learning and self-learning may provide pseudo-labels for unlabeled data to optimize the segmentation model. Generation of countermeasure networks (GANs) and Class Activation Maps (CAMs) deal with the absence of pixel-level or voxel-level labels by training the segmentation model using data of weak labels, such as slice-level labels or volume-level labels with CT. However, training a model using pseudo-labels may introduce noise, affecting the performance of the model, and the model is trained only on data with weak labels, often failing to obtain a satisfactory segmentation mask.
2) Attention mechanism
The basic idea of attention mechanism in computer vision is to let the system learn to pay attention and be able to ignore irrelevant information and concentrate on critical information. In recent years, attention models have been widely used in various fields such as image processing, speech recognition, and natural language processing. As deep learning progresses to this day, it becomes increasingly important to build a neural network with attention mechanisms. On the one hand, this neural network can learn the attention mechanism autonomously, and on the other hand, the attention mechanism can help to understand the world seen by the neural network. In recent years, much research work on the combination of deep learning and visual attention mechanisms has focused on the use of masks to form the attention mechanism. The principle of masking is to identify key features in the image data by another layer of new weights. Through learning and training, the deep neural network can learn the regions needing attention in each new image.
Disclosure of Invention
The invention aims to solve the problem of label shortage of voxel-level CT image data and provides a semi-supervised CT image segmentation method based on countermeasure training.
The specific technical scheme for realizing the purpose of the invention is as follows:
a semi-supervised CT image segmentation method based on antagonistic training comprises the following steps:
step 1) obtaining a lung three-dimensional CT image, establishing a diseased data set V marked at a voxel level, a diseased data set D not marked and a healthy data set H, and setting the marked diseased image as I v Belongs to V, and the unmarked diseased image is I d E D, healthy image I h ∈H;
Step 2) for any input image in the data set, carrying out image preprocessing operations including image cutting, resampling and standardization on the input image to obtain an image I i Inputting the data into a generator to obtain a reconstruction result I of the generator g The generator is aimed at i Reconstructing a healthy image I g (ii) a Will I i Input to a segmenter to obtain a segmentation mask
Figure BDA0003550095730000021
For a voxel-level labeled diseased dataset V, the segmentation mask is supervised using the true label Y in the dataset V
Figure BDA0003550095730000022
Step 3) masking the segmentation mask
Figure BDA0003550095730000023
Negation is carried out to obtain a reverse mask
Figure BDA0003550095730000024
Step 4) taking generator result I g Is a mask prediction region of
Figure BDA0003550095730000025
A portion of the number 1 of the first to the second,
Figure BDA0003550095730000026
and an input image I i Is the inverse mask of
Figure BDA0003550095730000027
A portion of the number 1 of the first to the second,
Figure BDA0003550095730000028
adding the two images according to elements to obtain a composite image I p
Step 5) taking generator result I g Is predicted in the reverse mask, i.e.
Figure BDA0003550095730000029
And an input image I i Is predicted in the reverse mask, i.e.
Figure BDA00035500957300000210
Respectively obtain images I gh And image I ih
Step 6) constructing a semi-supervised CT image segmentation model, wherein the model comprises a generator G, a segmenter S and a discriminator D 1 And a discriminator D 2 (ii) a The input of the model is simultaneously input into a generator G and a divider S, the divider S obtains a division mask of the CT image, a voxel with the value of 1 represents a lesion area, and a voxel with the value of 0 represents a healthy area; the generator G is used for inputting a CT image, generating a CT image after the CT image is recovered to be healthy, and then combining the generated image with a segmentation mask and the image input by a model to generate a recovered healthy image and an anti-mask image so as to approximate a real image to cheat the discriminator D 1 And a discriminator D 2 (ii) a Discriminator D 1 And D 2 The function of (1) is to judge whether the input image is from a real image; by competing against each other, the generator G and the discriminator D 1 And D 2 Optimizing the weight in an iterative mode to improve the performance;
the semi-supervised CT image segmentation model comprises four losses, namely supervision loss, reconstruction loss, discrimination loss 1 and discrimination loss 2; the surveillance loss is effective when the input CT image belongs to a voxel-level labeled diseased data set; the reconstruction loss is the mean square error MSE loss between the model input and the generator G result to supervise the generator G to produce a more realistic healthy CT image; judging whether the loss 1 is a countermeasure loss, and judging whether the input image is a real image or a synthesized image so as to supervise the quality of the recovered healthy image; judging whether the loss 2 is a countermeasure loss, and judging whether the input image is from a real model input image or an image generated by a generator G so as to supervise the quality of the reverse mask image; continuously optimizing the loss using a model optimization method based on antagonistic training until the loss converges;
and 7) after the model training is finished, giving a three-dimensional CT image to be segmented, and inputting the three-dimensional CT image to the segmenter S to obtain the voxel-level three-dimensional segmentation mask of the image.
The step 1) specifically comprises the following steps:
step a 1: acquiring a three-dimensional CT image sample by professional equipment to obtain an original data set;
step a 2: dividing an original data set into a diseased data set and a healthy data set, randomly sampling one fifth of data in the diseased data, and labeling the data by an artificial expert to be used as a diseased data set V labeled in a voxel level; the rest image data are used as an unmarked diseased data set D, and all the healthy image data form a healthy data set H;
step a 3: sequentially inputting images in the diseased data set V labeled at the voxel level, the diseased data set D not labeled and the healthy data set H to a divider and a generator, and recording an input CT image of the model as I i
The image preprocessing operation in step 2) specifically includes:
(1) image cropping
Image cropping cuts a three-dimensional medical image to a non-zero area of the image, namely, a minimum three-dimensional boundary box is searched in the image, the value outside the boundary box area is 0, and the boundary box is used for cropping the image;
(2) resampling
Resampling can solve the problem that the actual space represented by different voxel images is not consistent in size in a three-dimensional CT data set. In order to unify the resolution sizes of all the CT images, the resolutions of different CT images are scaled by resampling so as to unify the resolutions to 0.5mm multiplied by 0.5 mm;
(3) standardization
In order to have the same distributed gray-scale values for each image, the minimum and maximum values of the gray-scale values of the CT image are set to 300 and 3000; the value of the gradation value less than 300 is increased to 300 and the value of the gradation value greater than 3000 is decreased to 3000; the voxel values of the CT image are then normalized to obtain values between [0,1], and these values are then scaled to the [0,255] interval.
The generator G and the discriminator D in the step 6) 1 And D 2 The design of (2) follows the structure of CycleGAN, and in addition, the final discrimination result, namely 1 or 0, is obtained by using the fully connected layer as a classification network at the last layer, wherein 1 represents that the discriminator considers the input image to be a real image, and 0 represents that the discriminator considers the input image to be a generated or synthesized image.
The segmenter S in the step 6) designs a feature enhancement module for enhancing the feature representation of the encoder based on the 3DU-Net, wherein the feature enhancement module comprises channel attention and space attention. (ii) a To balance memory usage and segmentation accuracy, four 2-fold downsampling is used in the segmenter S; the divider S adopts a pyramid pooling structure of the dense cavity space, can combine features with different scales by using expansion rates with different sizes, and well realizes reuse of the features; the dilation convolution (scaled convolution) in segmenter S uses dilation rates of 3, 6, and 12; in order to improve the quality of the learned features, the features with the same scale are fused in a quick connection mode; in addition, features of the same scale are fused in a quick connection mode, and therefore the quality of the learned features is improved. The number of channels is set to 16 in consideration of the memory usage of the 3D segmentation model.
Step 6) the model optimization method based on the countermeasure training adopts a gradient descent algorithm based on an Adam optimizer to iteratively optimize the segmenter S, the generator G and the discriminator D 1 And a discriminator D 2 The method specifically comprises the following steps:
(1) initialization discriminator D 1 And D 2 The weight parameter of (2); the iteration iter is set to 1, and the maximum iteration iter is set max
(2) Optimizing segmenter S and generator G: freezing discriminator D 1 And D 2 Parameters of the defreezing divider S and the generator G, calculating loss and optimizing the model, iter plus one;
(3) optimization discriminator D 1 : freezing segmenter S, generator G and discriminator D 2 Parameter of (2), thawing discriminator D 1 Calculating the loss and optimizing the model, iter plus one;
(4) optimization discriminator D 2 : freezing segmenter S, generator G and discriminator D 1 Parameter of (2), thawing discriminator D 2 Calculating the loss and optimizing the model, iter plus one;
repeating (1) - (4), wherein iter is larger than iter max Or until the loss converges.
The design of the generator and the discriminator of the invention follows the structure of cycleGAN, and in addition, the final discrimination result is obtained by connecting the classifier by using a full connection layer at the last layer. A spectral normalization operation is added to the discriminator. Regarding the memory usage of 3D images, the number of fundamental channels is reduced from 64 to 16.
Compared with the prior art, the invention has the beneficial effects that:
1. the invention provides a semi-supervised CT image segmentation method based on confrontation training, which is used for segmenting a focus region in a lung three-dimensional CT image. Compared with a two-dimensional divider, the three-dimensional divider can combine information between image layers, so that the change continuity between interlayer image masks can be ensured. The present invention can be trained using only a small amount of voxel-level annotated data and the remaining unlabeled data.
2. The invention optimizes the generator, the segmenter and the discriminator in the model in a mode of countertraining, and enables the segmenter to learn the characteristic information of healthy lungs and focuses from the annotated data of the voxel level so as to gradually improve the segmentation performance of the model, thereby greatly reducing the number of voxel level labels required by training the model.
3. In order to better deal with the problem of low contrast between the infected lung area and normal tissue, the present invention designs a feature enhancement module to deal with ambiguous boundaries. In particular, channel attention is used to implicitly enhance contrast between features, thereby highlighting boundary information for lesion regions, and spatial attention generation spatial attention is used to highlight important regions. The feature enhancement module effectively enhances the feature representation of the lesion area by fusing these features.
Drawings
FIG. 1 is a flow chart of the present invention;
FIG. 2 is a diagram of an antagonistic semi-supervised model of the present invention;
FIG. 3 is a schematic diagram of a segmenter according to the present invention;
FIG. 4 is a schematic view of a feature enhancement module of the present invention that combines channel and spatial attention.
Detailed Description
The present invention will be described in further detail with reference to the following detailed description and accompanying drawings. The procedures, conditions, experimental methods and the like for carrying out the present invention are general knowledge and common general knowledge in the art except for the contents specifically mentioned below, and the present invention is not particularly limited.
The whole process of the present invention includes steps 1) -7), please refer to fig. 1 and 2. Fig. 1 is a flow chart of a method for segmenting a semi-supervised CT image based on antagonistic training, and fig. 2 is a schematic diagram of an antagonistic semi-supervised model.
The process of the invention:
step 1) obtaining lungsThree-dimensional CT image of the patient, establishing a diseased data set V marked at voxel level, a diseased data set D not marked, a healthy data set S, and setting the marked diseased image as I v Belongs to V, and the unmarked diseased image is I d E.g. D, health image I h ∈S。
Step 2) inputting all input data into a generator to obtain a healthy reconstructed version I g Input to a segmenter to obtain a segmentation mask
Figure BDA0003550095730000051
Diseased data set I for voxel level labeling v Supervision of segmentation masks using a true label Y in the data set V
Figure BDA0003550095730000052
Step 3) masking the segmentation mask
Figure BDA0003550095730000053
Get the inverse to
Figure BDA0003550095730000054
Step 4) taking generator result I g Is predicted (i.e. by masking
Figure BDA0003550095730000055
A portion of the number 1 of the first to the second,
Figure BDA0003550095730000056
) And an input image I i Is predicted (i.e. is predicted by the inverse mask of (c))
Figure BDA0003550095730000057
A portion that is a number 1 of the segment,
Figure BDA0003550095730000058
) Adding the two images according to elements to obtain a composite image I p
Specifically, an unmarked disease data is set as I d Belongs to E D, and the ill data marked on a voxel level isI v Belongs to V, and one piece of health image data is I h E.g. H. The lesion region predicted by the segmenter is replaced with the corresponding region reconstructed by the generator while the uninfected region is preserved, called pseudo-healthy image I p . The pseudo-healthy image is calculated by the following formula:
1)
Figure BDA0003550095730000059
where φ is a function that generates a pseudo-health image;
Figure BDA00035500957300000510
is the probabilistic segmentation mask predicted by segmenter S; image I generated by the generator g =G(I d ;θ G );θ S And theta G Learnable parameters for segmenter S and generator G, respectively.
Step 5) taking generator result I g Is predicted in the reverse mask
Figure BDA00035500957300000511
And an input image I i Is predicted in the reverse mask
Figure BDA00035500957300000512
Respectively obtain I gh And I ih
Because there is no supervisory signal to constrain I g May be used, which may lead to problems with the quality of the generated image being uncontrollable. Although they are not used to form a composite image, this affects the performance of the generator, which is a bottleneck in improving the final performance of the segmenter. The invention utilizes the reconstruction result of the generator to obtain I from the health area predicted by the divider g Generated healthy area I gh And I i Health area I of ih
2)
Figure BDA0003550095730000061
Figure BDA0003550095730000062
Then I gh And I ih Is inputted to a discriminator D 2 ,I gh Referred to as composite healthy area image sum I ih Referred to as the true healthy area image.
Step 6) constructing a three-dimensional confrontation semi-supervised model, and iteratively optimizing a generator, a divider and a discriminator D 1 And a discriminator D 2 Until the total loss converges. The confrontation training can lead the model to be trained in a semi-supervised learning mode, namely, the network can learn information which is beneficial to the segmentation accuracy from the unlabeled data.
To fool the arbiter D completely 1 And D 2 The segmenter needs to segment all the infected regions, and the generator needs to generate pseudo-healthy images in both the segmented predicted diseased regions and healthy regions. In contrast, discriminator D 1 For distinguishing and real health images I d D, discriminator D 2 To distinguish the composite healthy area image I gh And a real healthy area image I ih . The model was trained in an antagonistic manner with the following maximum minimum, antagonistic losses were as follows:
Figure BDA0003550095730000063
wherein theta is G And theta S Learnable parameters of the generator G and the divider S respectively;
Figure BDA0003550095730000064
and
Figure BDA0003550095730000065
are respectively a discriminator D 1 And a discriminator D 1 A learnable parameter of (c);
Figure BDA0003550095730000066
yang (Yang)
Figure BDA0003550095730000067
Is a loss function.
Marking the composite image as 0, the real health image objective function
Figure BDA0003550095730000068
By
5)
Figure BDA0003550095730000069
6) Wherein the content of the first and second substances,
Figure BDA00035500957300000610
is D 1 The predicted result of (2); i is p Calculated by formula (1);
Figure BDA00035500957300000611
representation discriminator D 1 Learnable parameters of (c);
Figure BDA00035500957300000612
and
Figure BDA00035500957300000613
representing a mathematical expectation.
Label the synthetic healthy region as 0 and the real healthy region as 1, wherein the objective function
Figure BDA00035500957300000614
The calculation is as follows:
7)
Figure BDA00035500957300000615
wherein the content of the first and second substances,
Figure BDA00035500957300000616
is discriminator D 2 The predicted result of (2); image I gh And I sh Calculated from equations 2 and 3, respectively;
Figure BDA00035500957300000617
represents D 2 May be used to learn the parameters.
The segmenter of the present invention is described with reference to FIG. 3. Fig. 3 is a schematic diagram of a divider.
The invention proposes a feature enhancement module that uses channel and spatial attention to adapt the optimization features to optimize the segmenter, fig. 4 being a schematic diagram of the feature enhancement module that combines channel and spatial attention. In particular, the channel attention module learns global parameters to highlight useful boundary information. The spatial attention module calculates an attention weight map of the lung region. The channel attention, spatial attention, may improve the representation of features of the region of interest, enabling salient features and suppression of unnecessary features. Given an intermediate feature mapping
Figure BDA0003550095730000071
As input, where C is the number of channels, D is the feature dimension, H is the image height, and W is the image width. Feature enhancement module infers 1D channel attention in turn
Figure BDA0003550095730000072
And 2D spatial attention mapping
Figure BDA0003550095730000073
The whole attention process can be summarized as follows:
attention of the channel: to enhance the contrast of features, spatial information on the feature map is first aggregated using an average pooling and max pooling operation, generating two different spatial context descriptors, an average pooled feature and a max pooled feature. The two descriptors are forwarded to a multi-layer perceptron network sharing parameters, and finally the result is subjected to a sigmoid function to generate a channel attention M c (F) In that respect The shared network is composed of a multi-layered perceptron (MLP) with one hidden layer. The channel attention mainly comprises an average pooling operation, a maximum pooling operation, a multi-layer perceptron of a shared parameter and a sigmoid function. An intermediate optimization feature for channel attention generation is
Figure BDA0003550095730000074
The calculation process is as follows:
8)
Figure BDA0003550095730000075
wherein F is the input characteristic of the channel attention; sigma represents a sigmoid activation function; AvgPool is the average pooling function; MaxPool is the maximum pooling function.
Spatial attention is as follows: the purpose of spatial attention is to discard unimportant features and highlight features of interest that are useful for segmenting COVID19 infections. Spatial attention generation spatial attention M using spatial relationships between features s (F'), which is complementary to the channel attention. To compute spatial attention, the average pooling and maximum pooling operations are first taken along the channel axis and concatenated to generate a valid feature descriptor. Applying pooling along the channel axis can effectively highlight the feature region of interest. On concatenated feature descriptors, convolutional layers are used to generate spatial attention, which encodes information where emphasis or suppression is needed. The final optimization feature of spatial attention generation is F ",
Figure BDA0003550095730000076
wherein σ represents a sigmoid activation function; conv is a convolution operation of size 7x7x 7;
Figure BDA0003550095730000077
representing multiplication by element. F "is obtained by element-by-element multiplication between F' and the attention map.
Due to the synthesized image I p With images I generated by a prediction mask sum generator g In connection with, therefore, the
Figure BDA0003550095730000078
Can be fed back to the segmentor S and the generator G in order to optimize the parameters of both modules. In addition, also addA basic segmentation loss is obtained
Figure BDA00035500957300000713
The difference between the output Y of S and the true mask GT for measuring a small number of voxel level label samples:
Figure BDA00035500957300000710
wherein CEL is the cross-entropy loss;
Figure BDA00035500957300000711
is voxel level label data I v The prediction mask of (3); y is v Is voxel level label data I v The real tag of (1). Prediction mask
Figure BDA00035500957300000712
Wherein S (-) is a segmenter θ S Learnable parameters representing a segmenter.
In order to improve the stability of the resistance training in the step 6), the invention further sets auxiliary constraint. First, a health image I h Input to a segmentor and generator. Health image data I h Is predicted by the prediction mask S (I) h ;θ S ) And health image data I h True label Y of h The cross entropy loss between is added to
Figure BDA0003550095730000081
Figure BDA0003550095730000082
Secondly, to further improve the performance of the generator, the method uses reconstruction losses
Figure BDA0003550095730000083
To constrain the output of the generator:
Figure BDA0003550095730000084
where MSE (-) is a mean square error function; g (I) h ;θ G ) Is represented by I h The reconstruction result obtained after being input to the generator G; theta G Are learnable parameters of the generator G. Further, it also relates to
Figure BDA0003550095730000085
Introducing additional loss function
Figure BDA0003550095730000086
Figure BDA0003550095730000087
Wherein, G (I) d ;θ G ) Is represented by d The reconstruction result obtained after being input to the generator G;
Figure BDA0003550095730000088
is discriminator D 1 May be used to learn the parameters.
So as to pass through I d To generate an image I g Input D 1 To further improve the generating effect. When the input of the segmenter S is a healthy image, the forward propagation process of synthesis and discrimination is not required, and calculation is not required
Figure BDA00035500957300000818
In addition, the original disease image I d Is input to a discriminator D 1 To keep the discriminator D during the training process 1 Sensitivity and distinguishability to lesion signals, the dropout rate is set to be fixed at 0.5. Will constrain the loss function
Figure BDA00035500957300000810
Is added to
Figure BDA00035500957300000819
Discriminator D 1 The goal of (1) is to be able to distinguish between images of a patient and a healthy person:
Figure BDA00035500957300000812
to summarize the above, the penalty function is extended by adding four new penalties as auxiliary constraints
Figure BDA00035500957300000813
And
Figure BDA00035500957300000814
the final objective function is defined as follows:
Figure BDA00035500957300000815
wherein λ is S Is used for balancing
Figure BDA00035500957300000816
And
Figure BDA00035500957300000817
is determined.
And 7) realizing the CT image segmentation based on the trained segmenter.
The protection of the present invention is not limited to the above embodiments. Variations and advantages that may occur to those skilled in the art are intended to be included within the invention without departing from the spirit and scope of the inventive concept, and the scope of the invention is to be determined by the appended claims.

Claims (6)

1. A semi-supervised CT image segmentation method based on antagonistic training is characterized by comprising the following steps:
step 1) obtaining a lung three-dimensional CT image, establishing a diseased data set V marked at a voxel level, a diseased data set D not marked and a healthy data set H, and setting a marked diseased image as I v E.g. y, the unmarked diseased image is I d E.g. D, health image I h ∈H;
Step 2) forAny one input image in the data set is subjected to image preprocessing operations including image cropping, resampling and standardization to obtain an image I i Inputting the data into a generator to obtain a reconstruction result I of the generator g The generator is aimed at i Reconstructing a healthy image I g (ii) a Will I i Input to a segmenter to obtain a segmentation mask
Figure FDA0003550095720000011
For a voxel-level labeled diseased dataset V, the segmentation mask is supervised using the true label Y in the dataset V
Figure FDA0003550095720000012
Step 3) masking the segmentation mask
Figure FDA0003550095720000013
Negation is carried out to obtain a reverse mask
Figure FDA0003550095720000014
Step 4) taking generator result I g Is a mask prediction region of
Figure FDA0003550095720000015
A portion that is a number 1 of the segment,
Figure FDA0003550095720000016
and an input image I i Is the inverse mask of
Figure FDA0003550095720000017
A portion of the number 1 of the first to the second,
Figure FDA0003550095720000018
adding the two images according to elements to obtain a composite image I p
Step 5) taking generator result I g Is predicted in the reverse mask, i.e.
Figure FDA0003550095720000019
And an input image I i Is predicted in the reverse mask, i.e.
Figure FDA00035500957200000110
Respectively obtain images I gh And image I ih
Step 6) constructing a semi-supervised CT image segmentation model, wherein the model comprises a generator G, a segmenter S and a discriminator D 1 And a discriminator D 2 (ii) a The input of the model is simultaneously input into a generator G and a divider S, the divider S obtains a division mask of the CT image, a voxel with the value of 1 represents a lesion area, and a voxel with the value of 0 represents a healthy area; the generator G is used for inputting a CT image, generating a CT image after the CT image is recovered to be healthy, and then combining the generated image with a segmentation mask and the image input by a model to generate a recovered healthy image and an anti-mask image so as to approximate a real image to cheat the discriminator D 1 And a discriminator D 2 (ii) a Discriminator D 1 And D 2 The function of (1) is to judge whether the input image is from a real image; by competing against each other, the generator G and the discriminator D 1 And D 2 Optimizing the weight in an iterative mode to improve the performance;
the semi-supervised CT image segmentation model comprises four losses, namely supervision loss, reconstruction loss, discrimination loss 1 and discrimination loss 2; the surveillance loss is effective when the input CT image belongs to a voxel-level labeled diseased data set; the reconstruction loss is the mean square error MSE loss between the model input and the generator G result to supervise the generator G to produce a more realistic healthy CT image; judging whether the loss 1 is a countermeasure loss, and judging whether the input image is a real image or a synthesized image so as to monitor the quality of the recovered healthy image; judging whether the loss 2 is a countermeasure loss, and judging whether the input image is from a real model input image or an image generated by a generator G so as to supervise the quality of the reverse mask image; continuously optimizing the loss using a model optimization method based on antagonistic training until the loss converges;
and 7) after the model training is finished, giving a three-dimensional CT image to be segmented, and inputting the three-dimensional CT image to the segmenter S to obtain the voxel-level three-dimensional segmentation mask of the image.
2. The method of claim 1, wherein the step 1) comprises:
step a 1: acquiring a three-dimensional CT image sample by professional equipment to obtain an original data set;
step a 2: dividing an original data set into a diseased data set and a healthy data set, randomly sampling one fifth of data in the diseased data, and labeling the data by an artificial expert to be used as a diseased data set V labeled in a voxel level; the remaining image data serve as the unlabeled diseased data set D, and all healthy image data constitute the healthy data set H.
3. The method of claim 1, wherein the image preprocessing operation in step 2) specifically comprises:
(1) image cropping
Image cropping cuts a three-dimensional medical image to a non-zero area of the image, namely, a minimum three-dimensional boundary box is searched in the image, the value outside the boundary box area is 0, and the boundary box is used for cropping the image;
(2) resampling
Scaling the different CT image resolutions using resampling to unify the resolutions to 0.5mm by 0.5 mm;
(3) standardization
Setting the minimum value and the maximum value of the CT image gray value to be 300 and 3000; the value of the gradation value less than 300 is increased to 300 and the value of the gradation value greater than 3000 is decreased to 3000; the voxel values of the CT image are then normalized to obtain values between [0,1], and these values are then scaled to the [0,255] interval.
4. The method of claim 1, wherein the generator G and the judging in step 6) are performed by a semi-supervised CT image segmentation method based on the interactive trainingPin D 1 And D 2 The design of (2) follows the structure of CycleGAN, and in addition, the final discrimination result, namely 1 or 0, is obtained by using the fully connected layer as a classification network at the last layer, wherein 1 represents that the discriminator considers the input image to be a real image, and 0 represents that the discriminator considers the input image to be a generated or synthesized image.
5. The method as claimed in claim 1, wherein the segmenter S in step 6) is based on 3DU-Net, and a feature enhancement module is designed in the segmenter S for enhancing the feature representation of the encoder, and the feature enhancement module includes channel attention and spatial attention. (ii) a To balance memory usage and segmentation accuracy, four 2-fold downsampling is used in the segmenter S; the divider S adopts a pyramid pooling structure of the dense cavity space, can combine features with different scales by using expansion rates with different sizes, and well realizes reuse of the features; the dilation convolution (scaled convolution) in segmenter S uses dilation rates of 3, 6, and 12; the features with the same scale are fused in a quick connection mode; the number of channels is set to 16.
6. The confrontation training-based semi-supervised CT image segmentation method of claim 1, wherein the confrontation training-based model optimization method of step 6) adopts a gradient descent algorithm based on Adam optimizer to iteratively optimize the segmenter S, the generator G and the discriminator D 1 And a discriminator D 2 The method specifically comprises the following steps:
(1) initialization discriminator D 1 And D 2 The weight parameter of (2); the iteration iter is set to 1, and the maximum iteration iter is set max
(2) Optimizing segmenter S and generator G: freezing discriminator D 1 And D 2 Parameters of the defreezing divider S and the generator G, calculating loss and optimizing the model, iter plus one;
(3) optimization discriminator D 1 : freezing segmenter S, generator G and discriminator D 2 Radix Ginseng (radix Ginseng)Number, unfreezing discriminator D 1 Calculating the loss and optimizing the model, iter plus one;
(4) optimization discriminator D 2 : freezing segmenter S, generator G and discriminator D 1 Parameter of (2), thawing discriminator D 2 Calculating the loss and optimizing the model, iter plus one;
repeating (1) - (4), wherein iter is larger than iter max Or until the loss converges.
CN202210259206.0A 2022-03-16 2022-03-16 Semi-supervised CT image segmentation method based on countermeasure training Active CN114897914B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202210259206.0A CN114897914B (en) 2022-03-16 2022-03-16 Semi-supervised CT image segmentation method based on countermeasure training

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202210259206.0A CN114897914B (en) 2022-03-16 2022-03-16 Semi-supervised CT image segmentation method based on countermeasure training

Publications (2)

Publication Number Publication Date
CN114897914A true CN114897914A (en) 2022-08-12
CN114897914B CN114897914B (en) 2023-07-07

Family

ID=82715871

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202210259206.0A Active CN114897914B (en) 2022-03-16 2022-03-16 Semi-supervised CT image segmentation method based on countermeasure training

Country Status (1)

Country Link
CN (1) CN114897914B (en)

Cited By (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN115187819A (en) * 2022-08-23 2022-10-14 北京医准智能科技有限公司 Training method and device for image classification model, electronic equipment and storage medium
CN115861252A (en) * 2022-12-14 2023-03-28 深圳技术大学 Semi-supervised medical image organ segmentation method based on counterstudy strategy
CN116739890A (en) * 2023-06-26 2023-09-12 强联智创(北京)科技有限公司 Method and equipment for training generation model for generating healthy blood vessel image
CN117078703A (en) * 2023-08-30 2023-11-17 深圳扬奇医芯智能科技有限公司 CT image segmentation method and system based on MRI guidance
CN117095395A (en) * 2023-10-19 2023-11-21 北京智源人工智能研究院 Model training method and device for heart ultrasonic image segmentation and segmentation method
CN117726642A (en) * 2024-02-07 2024-03-19 中国科学院宁波材料技术与工程研究所 High reflection focus segmentation method and device for optical coherence tomography image

Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN109949317A (en) * 2019-03-06 2019-06-28 东南大学 Based on the semi-supervised image instance dividing method for gradually fighting study
CN112419327A (en) * 2020-12-10 2021-02-26 复旦大学附属肿瘤医院 Image segmentation method, system and device based on generation countermeasure network
US20210089845A1 (en) * 2019-09-20 2021-03-25 Samsung Electronics Co., Ltd. Teaching gan (generative adversarial networks) to generate per-pixel annotation
CN112734764A (en) * 2021-03-31 2021-04-30 电子科技大学 Unsupervised medical image segmentation method based on countermeasure network
WO2022041307A1 (en) * 2020-08-31 2022-03-03 温州医科大学 Method and system for constructing semi-supervised image segmentation framework

Patent Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN109949317A (en) * 2019-03-06 2019-06-28 东南大学 Based on the semi-supervised image instance dividing method for gradually fighting study
US20210089845A1 (en) * 2019-09-20 2021-03-25 Samsung Electronics Co., Ltd. Teaching gan (generative adversarial networks) to generate per-pixel annotation
WO2022041307A1 (en) * 2020-08-31 2022-03-03 温州医科大学 Method and system for constructing semi-supervised image segmentation framework
CN112419327A (en) * 2020-12-10 2021-02-26 复旦大学附属肿瘤医院 Image segmentation method, system and device based on generation countermeasure network
CN112734764A (en) * 2021-03-31 2021-04-30 电子科技大学 Unsupervised medical image segmentation method based on countermeasure network

Non-Patent Citations (2)

* Cited by examiner, † Cited by third party
Title
DAIQING LI等: "Semantic Segmentation with Generative Models: Semi-Supervised Learning and Strong Out-of-Domain Generalization", 《HTTPS://ARXIV.ORG/ABS/2104.05833》, pages 1 - 12 *
张粲 等: "基于对抗学习网络的半监督医学图像分割方法", 《工业控制计算机》, no. 9, pages 57 - 59 *

Cited By (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN115187819A (en) * 2022-08-23 2022-10-14 北京医准智能科技有限公司 Training method and device for image classification model, electronic equipment and storage medium
CN115861252A (en) * 2022-12-14 2023-03-28 深圳技术大学 Semi-supervised medical image organ segmentation method based on counterstudy strategy
CN115861252B (en) * 2022-12-14 2023-09-22 深圳技术大学 Semi-supervised medical image organ segmentation method based on countermeasure learning strategy
CN116739890A (en) * 2023-06-26 2023-09-12 强联智创(北京)科技有限公司 Method and equipment for training generation model for generating healthy blood vessel image
CN117078703A (en) * 2023-08-30 2023-11-17 深圳扬奇医芯智能科技有限公司 CT image segmentation method and system based on MRI guidance
CN117095395A (en) * 2023-10-19 2023-11-21 北京智源人工智能研究院 Model training method and device for heart ultrasonic image segmentation and segmentation method
CN117095395B (en) * 2023-10-19 2024-02-09 北京智源人工智能研究院 Model training method and device for heart ultrasonic image segmentation and segmentation method
CN117726642A (en) * 2024-02-07 2024-03-19 中国科学院宁波材料技术与工程研究所 High reflection focus segmentation method and device for optical coherence tomography image

Also Published As

Publication number Publication date
CN114897914B (en) 2023-07-07

Similar Documents

Publication Publication Date Title
CN111476292B (en) Small sample element learning training method for medical image classification processing artificial intelligence
CN114897914B (en) Semi-supervised CT image segmentation method based on countermeasure training
WO2019200747A1 (en) Method and device for segmenting proximal femur, computer apparatus, and storage medium
WO2020133636A1 (en) Method and system for intelligent envelope detection and warning in prostate surgery
Tataru et al. Deep Learning for abnormality detection in Chest X-Ray images
Yao et al. Pneumonia detection using an improved algorithm based on faster r-cnn
JP7312510B1 (en) Whole-slide pathological image classification system and construction method considering tumor microenvironment
CN112926696A (en) Interpretable local migration mutual learning method based on attention diagram
CN116129141A (en) Medical data processing method, apparatus, device, medium and computer program product
Wu et al. Cascaded fully convolutional DenseNet for automatic kidney segmentation in ultrasound images
CN117036288A (en) Tumor subtype diagnosis method for full-slice pathological image
Duvieusart et al. Multimodal cardiomegaly classification with image-derived digital biomarkers
CN111753736A (en) Human body posture recognition method, device, equipment and medium based on packet convolution
CN116452793A (en) Multi-view and multi-level-based green coding and decoding significant target detection method
CN115409812A (en) CT image automatic classification method based on fusion time attention mechanism
CN114419061A (en) Method and system for segmenting pulmonary artery and vein blood vessels
Zhang et al. Nucleus image segmentation method based on GAN network and FCN model
Hao et al. Bone age estimation with x-ray images based on efficientnet pre-training model
Bhandari et al. Chest abnormality detection from x-ray using deep learning
Shen et al. HarDNet and Dual-Code Attention Mechanism based model for medical images segmentation
CN113379735B (en) Labeling method and system for CT colon image content
CN117393100B (en) Diagnostic report generation method, model training method, system, equipment and medium
Zhang et al. MSAA-Net: a multi-scale attention-aware U-Net is used to segment the liver
CN116740041B (en) CTA scanning image analysis system and method based on machine vision
Baskar DeepNet model empowered cuckoo search algorithm for the effective identification of lung cancer nodules

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant