CN114140368B - Multi-mode medical image synthesis method based on generation type countermeasure network - Google Patents
Multi-mode medical image synthesis method based on generation type countermeasure network Download PDFInfo
- Publication number
- CN114140368B CN114140368B CN202111465819.1A CN202111465819A CN114140368B CN 114140368 B CN114140368 B CN 114140368B CN 202111465819 A CN202111465819 A CN 202111465819A CN 114140368 B CN114140368 B CN 114140368B
- Authority
- CN
- China
- Prior art keywords
- network
- modal
- trained
- self
- image
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Active
Links
- 238000001308 synthesis method Methods 0.000 title abstract description 9
- 238000006243 chemical reaction Methods 0.000 claims abstract description 104
- 238000012549 training Methods 0.000 claims abstract description 78
- 238000000034 method Methods 0.000 claims abstract description 41
- 230000015572 biosynthetic process Effects 0.000 claims abstract description 25
- 238000003786 synthesis reaction Methods 0.000 claims abstract description 25
- 239000002131 composite material Substances 0.000 claims abstract description 15
- 238000012360 testing method Methods 0.000 claims abstract description 12
- 238000010586 diagram Methods 0.000 claims description 20
- 238000010276 construction Methods 0.000 claims description 13
- 230000008569 process Effects 0.000 claims description 11
- 230000001502 supplementing effect Effects 0.000 claims description 10
- 230000008485 antagonism Effects 0.000 claims description 7
- 238000004364 calculation method Methods 0.000 claims description 4
- 239000000203 mixture Substances 0.000 claims description 3
- 238000012544 monitoring process Methods 0.000 claims description 3
- 230000002194 synthesizing effect Effects 0.000 abstract description 12
- 230000009286 beneficial effect Effects 0.000 abstract description 3
- 230000006870 function Effects 0.000 description 10
- 238000000605 extraction Methods 0.000 description 7
- 239000011159 matrix material Substances 0.000 description 6
- 238000003384 imaging method Methods 0.000 description 5
- 230000009466 transformation Effects 0.000 description 5
- 230000004913 activation Effects 0.000 description 4
- 238000013135 deep learning Methods 0.000 description 4
- 208000003174 Brain Neoplasms Diseases 0.000 description 3
- 230000006872 improvement Effects 0.000 description 3
- 230000011218 segmentation Effects 0.000 description 3
- 239000013589 supplement Substances 0.000 description 3
- 206010028980 Neoplasm Diseases 0.000 description 2
- 210000001175 cerebrospinal fluid Anatomy 0.000 description 2
- 230000000295 complement effect Effects 0.000 description 2
- 230000000694 effects Effects 0.000 description 2
- 230000004927 fusion Effects 0.000 description 2
- 230000000153 supplemental effect Effects 0.000 description 2
- 238000005481 NMR spectroscopy Methods 0.000 description 1
- 238000013528 artificial neural network Methods 0.000 description 1
- 210000004556 brain Anatomy 0.000 description 1
- 239000002872 contrast media Substances 0.000 description 1
- 238000012217 deletion Methods 0.000 description 1
- 230000037430 deletion Effects 0.000 description 1
- 238000013461 design Methods 0.000 description 1
- 238000003745 diagnosis Methods 0.000 description 1
- 239000003814 drug Substances 0.000 description 1
- 230000008921 facial expression Effects 0.000 description 1
- 230000014509 gene expression Effects 0.000 description 1
- 238000002075 inversion recovery Methods 0.000 description 1
- 239000007788 liquid Substances 0.000 description 1
- 238000013508 migration Methods 0.000 description 1
- 230000005012 migration Effects 0.000 description 1
- 238000012545 processing Methods 0.000 description 1
- 230000000750 progressive effect Effects 0.000 description 1
- 238000011160 research Methods 0.000 description 1
- 230000009897 systematic effect Effects 0.000 description 1
- 238000010200 validation analysis Methods 0.000 description 1
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T5/00—Image enhancement or restoration
- G06T5/50—Image enhancement or restoration by the use of more than one image, e.g. averaging, subtraction
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F18/00—Pattern recognition
- G06F18/20—Analysing
- G06F18/24—Classification techniques
- G06F18/241—Classification techniques relating to the classification model, e.g. parametric or non-parametric approaches
- G06F18/2415—Classification techniques relating to the classification model, e.g. parametric or non-parametric approaches based on parametric or probabilistic models, e.g. based on likelihood ratio or false acceptance rate versus a false rejection rate
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F18/00—Pattern recognition
- G06F18/20—Analysing
- G06F18/25—Fusion techniques
- G06F18/253—Fusion techniques of extracted features
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/04—Architecture, e.g. interconnection topology
- G06N3/045—Combinations of networks
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/04—Architecture, e.g. interconnection topology
- G06N3/048—Activation functions
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/08—Learning methods
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T3/00—Geometric image transformation in the plane of the image
- G06T3/40—Scaling the whole image or part thereof
- G06T3/4038—Scaling the whole image or part thereof for image mosaicing, i.e. plane images composed of plane sub-images
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
- G06T7/0002—Inspection of images, e.g. flaw detection
- G06T7/0012—Biomedical image inspection
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2200/00—Indexing scheme for image data processing or generation, in general
- G06T2200/32—Indexing scheme for image data processing or generation, in general involving image mosaicing
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/10—Image acquisition modality
- G06T2207/10072—Tomographic images
- G06T2207/10088—Magnetic resonance imaging [MRI]
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/20—Special algorithmic details
- G06T2207/20212—Image combination
- G06T2207/20221—Image fusion; Image merging
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/30—Subject of image; Context of image processing
- G06T2207/30004—Biomedical image processing
- G06T2207/30016—Brain
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/30—Subject of image; Context of image processing
- G06T2207/30004—Biomedical image processing
- G06T2207/30096—Tumor; Lesion
Abstract
The invention relates to a multi-mode medical image synthesis method based on a generation type countermeasure network. The method comprises the following steps: constructing a modal attention generation type countermeasure network; the modal attention-generating countermeasure network includes a self-representation network and an image conversion network implemented by the generating countermeasure network; training and testing the modal attention generation type countermeasure network by adopting a multi-modal medical image data set to generate a trained modal attention generation type countermeasure network; and inputting the existing multi-modal medical image into the trained modal attention generation type countermeasure network, and outputting a composite image of the target modality. The method is used for synthesizing the missing medical mode images, the quality of the synthesis of the multi-mode medical images can be improved, and the complete multi-mode medical images are beneficial to doctors to make more accurate decisions.
Description
Technical Field
The invention relates to the technical field of image synthesis, in particular to a multi-mode medical image synthesis method based on a generation type countermeasure network.
Background
With the diversification of data acquisition means, more and more fields are beginning to use multi-modal data. For example, in the field of precise medicine, brain nuclear magnetic resonance can obtain image data of four modes, namely T1 weighted imaging (T1), gd contrast agent imaging (T1 Gd), T2 weighted imaging (T2) and liquid attenuation inversion recovery sequence imaging (T2-FLAIR), and a doctor can make more precise diagnosis by combining medical images of multiple modes. The actual situation is limited by the factors of limited medical conditions, insufficient scanning time and cost limitation, and systematic errors, even partial modality deletion and the like may occur in the multi-modality data acquisition process, so that partial modality images are finally unavailable. These unavailable modality images may cause deviations in the physician decision.
How to synthesize unusable modalities with existing, complete modality image data has become an important topic of research in computer vision and image processing. Complementing the missing modalities in multi-modality image data can be seen as a special class of image conversion problems. Image transformation refers to the transformation of an image from one image domain to another, similar tasks include style migration, semantic segmentation, deblurring, super resolution, etc. Different image fields refer to various imaging differences between lighting conditions, facial expressions, sensors, etc. Researchers have invested significant effort in developing efficient image conversion algorithms and have made significant progress in methods based on generating a countermeasure network, such as Pix2Pix. However, these methods are applicable to single modality inputs only, and do not consider the valid information contained in other available modalities in a multi-modality scenario in terms of model inputs and network design. CollaGAN (Collaborative GENERATIVE ADVERSARIAL Networks) is one of a few image conversion models designed for multiple modes, dongwook Lee et al propose multiple-cycle consistency loss for the multiple mode scene on the basis of CycleGAN aiming at the multiple mode scene, so that the function of generating target mode images from any multiple mode input is realized. CollaGAN has the disadvantage that no modeling is directed to modality complementarity, which makes it impossible to make full use of the information of the input model, and thus the quality of the composite image is low when applied to multi-modality medical image synthesis.
Disclosure of Invention
The invention aims to provide a multi-modal medical image synthesis method based on a generation type countermeasure network, so as to improve the quality of multi-modal medical image synthesis.
In order to achieve the above object, the present invention provides the following solutions:
a method of multimodal medical image synthesis based on a generated countermeasure network, comprising:
Constructing a modal attention generation type countermeasure network; the modal attention-generating countermeasure network includes a self-representation network and an image conversion network implemented by the generating countermeasure network;
Training and testing the modal attention generation type countermeasure network by adopting a multi-modal medical image data set to generate a trained modal attention generation type countermeasure network;
and inputting the existing multi-modal medical image into the trained modal attention generation type countermeasure network, and outputting a composite image of the target modality.
Optionally, the constructing the modal attention generation type countermeasure network specifically includes:
Constructing a self-representing network; the self-presenting network includes an encoder and a decoder;
Constructing an image conversion network; the image conversion network comprises a generator and a discriminator; the generator includes a multi-encoder, a modal attention module, and a decoder.
Optionally, the method for implanting in the modal attention module specifically includes:
acquiring a plurality of feature maps corresponding to medical images of other modes except the target mode;
Calculating the cross-modal channel weight according to the feature maps;
supplementing the information of the other modes into each input mode according to the cross-mode channel weight so as to obtain a new feature diagram after the supplementary information of a plurality of modes;
and splicing the new feature graphs after the plurality of modal supplementary information to generate a spliced feature graph.
Optionally, the training and testing the modal attention generation type countermeasure network by using the multi-modal medical image dataset to generate a trained modal attention generation type countermeasure network specifically includes:
Inputting the multi-modal medical image dataset into an encoder of the self-representing network, training the self-representing network using a random gradient descent algorithm, generating a pre-trained self-representing network;
inputting medical images of a target mode into the pre-trained self-expression network, inputting medical images of other modes except the target mode and masks corresponding to the target mode into a multi-encoder of the image conversion network, training the image conversion network by using a random gradient descent algorithm, and monitoring the training of the image conversion network by the self-expression network in the training process to generate a trained image conversion network; the pre-trained self-representing network and the trained image conversion network together comprise the trained modal attention-generating countermeasure network.
Optionally, the inputting the existing multi-modal medical image into the trained modal attention generation type countermeasure network outputs a composite image of the target modality, specifically including:
and inputting the existing multi-modal medical image and the mask corresponding to the target modality into a multi-encoder of the trained image conversion network, and outputting a composite image of the target modality by a generator of the trained image conversion network.
A multi-modal medical image composition system based on a generative countermeasure network, comprising:
the modal attention generation type antagonism network construction module is used for constructing a modal attention generation type antagonism network; the modal attention-generating countermeasure network includes a self-representation network and an image conversion network implemented by the generating countermeasure network;
the system comprises a modal attention generation type countermeasure network training module, a modal attention generation type countermeasure network training module and a modal attention generation type countermeasure network training module, wherein the modal attention generation type countermeasure network training module is used for training and testing the modal attention generation type countermeasure network by adopting a multi-modal medical image data set to generate a trained modal attention generation type countermeasure network;
and the multi-modal image synthesis module is used for inputting the existing multi-modal medical images into the trained modal attention generation type countermeasure network and outputting the synthesized images of the target modalities.
Optionally, the modal attention generation type countering network construction module specifically includes:
A self-expression network construction unit for constructing a self-expression network; the self-presenting network includes an encoder and a decoder;
an image conversion network construction unit for constructing an image conversion network; the image conversion network comprises a generator and a discriminator; the generator includes a multi-encoder, a modal attention module, and a decoder.
Optionally, the modal attention module specifically includes:
The characteristic map acquisition unit is used for acquiring a plurality of characteristic maps corresponding to medical images of other modes except the target mode;
A channel weight calculation unit, configured to calculate a cross-modal channel weight according to the plurality of feature graphs;
the modal information complementation unit is used for supplementing the information of the other modalities into each input modality according to the cross-modal channel weight so as to obtain a new feature map after the plurality of modal supplementary information;
And the characteristic diagram splicing unit is used for splicing the new characteristic diagrams after the plurality of modal supplementary information to generate spliced characteristic diagrams.
Optionally, the modal attention generation type countermeasure network training module specifically includes:
A self-expression network training unit, configured to input the multi-modal medical image dataset into an encoder of the self-expression network, train the self-expression network using a random gradient descent algorithm, and generate a pre-trained self-expression network;
an image conversion network training unit, configured to input a medical image of a target modality into the pre-trained self-expression network, input medical images of other modalities except the target modality and masks corresponding to the target modality into a multi-encoder of the image conversion network, train the image conversion network using a random gradient descent algorithm, and monitor training of the image conversion network by the self-expression network during training, so as to generate a trained image conversion network; the pre-trained self-representing network and the trained image conversion network together comprise the trained modal attention-generating countermeasure network.
Optionally, the multi-mode image synthesis module specifically includes:
and the multi-mode image synthesis unit is used for inputting the masks corresponding to the existing multi-mode medical images and the target modes into a multi-encoder of the trained image conversion network, and outputting the synthesized image of the target modes by a generator of the trained image conversion network.
According to the specific embodiment provided by the invention, the invention discloses the following technical effects:
The invention provides a multi-mode medical image synthesis method based on a generated type countermeasure network, which comprises the following steps: constructing a modal attention generation type countermeasure network; the modal attention-generating countermeasure network includes a self-representation network and an image conversion network implemented by the generating countermeasure network; training and testing the modal attention generation type countermeasure network by adopting a multi-modal medical image data set to generate a trained modal attention generation type countermeasure network; and inputting the existing multi-modal medical image into the trained modal attention generation type countermeasure network, and outputting a composite image of the target modality. The method is used for synthesizing the missing medical mode images, the quality of the synthesis of the multi-mode medical images can be improved, and the complete multi-mode medical images are beneficial to doctors to make more accurate decisions.
Drawings
In order to more clearly illustrate the embodiments of the present invention or the technical solutions of the prior art, the drawings that are needed in the embodiments will be briefly described below, it being obvious that the drawings in the following description are only some embodiments of the present invention, and that other drawings may be obtained according to these drawings without inventive effort for a person skilled in the art.
FIG. 1 is a flow chart of a method for synthesizing a multimodal medical image based on a generated challenge network in accordance with the present invention;
FIG. 2 is a schematic diagram of a modal attention-generating type countermeasure network according to the present invention;
FIG. 3 is a schematic diagram of a training process of a modal attention generating type countermeasure network provided by the present invention;
FIG. 4 is a schematic diagram of one embodiment of a modal attention-generating countermeasure network provided by the present invention;
FIG. 5 is a schematic diagram illustrating the complementation of model information in an embodiment of a modal attention generation type countermeasure network according to the present invention;
Fig. 6 is a block diagram of a multi-modal medical image synthesis system based on a generated countermeasure network according to the present invention.
Detailed Description
The following description of the embodiments of the present invention will be made clearly and completely with reference to the accompanying drawings, in which it is apparent that the embodiments described are only some embodiments of the present invention, but not all embodiments. All other embodiments, which can be made by those skilled in the art based on the embodiments of the invention without making any inventive effort, are intended to be within the scope of the invention.
The invention aims to provide a multi-modal medical image synthesis method based on a generation type countermeasure network, so as to improve the quality of multi-modal medical image synthesis.
In order that the above-recited objects, features and advantages of the present invention will become more readily apparent, a more particular description of the invention will be rendered by reference to the appended drawings and appended detailed description.
Fig. 1 is a flowchart of a method for synthesizing a multi-modal medical image based on a generated countermeasure network according to the present invention. As shown in fig. 1, the method for synthesizing a multi-modal medical image based on a generated type countermeasure network of the present invention includes:
step 101: a modal attention generating type countermeasure network is constructed.
A model of a Modal Attention GENERATIVE ADVERSARIAL Networks (MAGAN) proposed by the present invention is constructed, and the model MAGAN includes a self-expression network and an image conversion network (also referred to as a conversion network in the present invention) implemented by the generated type of network, where the image conversion network includes a Modal Attention module proposed by the present invention.
Fig. 2 is a schematic diagram of a modal attention generation type countermeasure network according to the present invention. Referring to fig. 2, the modal attention generation type countermeasure network MAGAN includes a conversion network and a self-expression network, and the conversion network incorporates the modal attention module proposed by the present invention to better utilize complementary information between modalities. The structure of the self-presenting network and the switching network is shown in fig. 2. Wherein the self-expression network is composed of an encoder and a decoder; the conversion network consists of a generator and a discriminator, wherein the generator consists of n-1 encoders, a modal attention module and a decoder, and the encoders and the decoder are connected through the modal attention module provided by the invention.
Specifically, the step 101 constructs a modal attention generation type countermeasure network, including:
Step 1.1: constructing a self-representing network; the self-presenting network includes an encoder and a decoder.
The self-expression network is based on a self-encoder implementation and consists of an encoder and a decoder.
Step 1.2: constructing an image conversion network; the image conversion network comprises a generator and a discriminator; the generator includes a multi-encoder, a modal attention module, and a decoder.
A switching network as shown in fig. 2 is constructed. The transformation network is based on Pix2Pix (image transformation classical model based on deep learning, known to the person skilled in the art) implementation, and is consistent with other generation-antagonism networks, the transformation network being composed of a generator G and a arbiter D. In order to adapt to the input of a plurality of modes, the single encoder structure of the generator G is changed into a multi-encoder structure shown on the left side of fig. 2, the number of encoders is n-1, and n is the total mode number of a specific application field. Furthermore, in order to fully mine the correlation among the modes, the invention adds the mode attention module in the characteristic fusion stage of the generator, and the mode attention module can automatically fuse the information from a plurality of input modes according to the information required by the target mode, thereby improving the quality of the synthesized mode image. The modal attention module is connected with a multi-encoder branch and a decoder, and the multi-encoder, the modal attention module and the decoder together form a generator of the conversion network.
The model MAGAN in the present invention is composed of a self-presenting network and a conversion network, and the improvement to the model is focused on the implementation of the conversion network. The conversion network is realized based on Pix2Pix, and the improvement is divided into two aspects, namely, in order to use a plurality of inputs, the input of the network is changed into a multi-encoder structure from an original single encoder; secondly, a mode attention module originally created by the method is added, and feature fusion is carried out on multi-branch extraction features aiming at a target mode. The principles and implementation of the modal attention module will be described in detail below.
Each channel in the feature map can be seen as a pattern extracted from the input by the deep neural network. The information contained in each input modality is different for any channel. Taking the medical image as an example, if the T-th channel is a texture feature for a tumor, it is apparent that T1Gd contains the most abundant and valuable information; t2 is the most important modality if the channel is a feature with respect to cerebrospinal fluid. One natural idea is to add tumor-related information in T1Gd to the T2 modality and cerebrospinal fluid-related information in T2 to the T1Gd modality. This is in effect the complementarity between the utilized modalities.
It is worth noting that it is necessary to know which modality contains the most rich information, or more attention to which modality, for a given channel or mode before supplementing other modalities. In the following description, the problem is formed as medical image data of n modalities altogether, with the objective of providing n-1 images belonging to n-1 modalities respectively, and synthesizing an image of the nth modality (i.e., the target modality).
The mode attention module is realized by adopting an algorithm, and the algorithm implanted in the mode attention module specifically comprises the following steps:
Step 1.2.1: and acquiring a plurality of feature maps corresponding to medical images of other modes except the target mode.
Referring to fig. 2, without loss of generality, medical images from the n modalities described above are denoted by x 1,x2,...,xn, respectively, where medical images from n-1 input modalities (i.e., other modalities than the target modality) are denoted by x 1,x2,...,xn-1, and medical images from n modalities, i.e., the target modality, are denoted by x n.
As shown in fig. 2, the masks corresponding to x 1,x2,...,xn-1 and the target mode are respectively input to n-1 encoders of the conversion network, so as to obtain feature images f 1,f2,...,fn-1, that is, a plurality of feature images f 1,f2,...,fn-1 corresponding to medical images of modes other than the target mode.
Step 1.2.2: and calculating the cross-modal channel weight according to the feature maps.
For a set of samples { x 1,x1,...,xn } from n modalities, equation (1) is used to calculate the cross-modality channel weights (denoted by a):
A=S2(σ((δ([f1,f2,...,fn-1]W1)W2) (1)
here, f k=Ek(xk) represents a feature map corresponding to the medical image x k of the kth modality other than the target modality, k E (1, 2,..n-1), E k (·) represents the output of the kth encoding branch E k of the generator G. The notation [. Cndot. ] indicates that tensors are stitched, e.g., [ f 1,f2,...,fn-1 ] indicates that the n-1 tensors f 1,f2,...,fn-1 are flattened and then stitched. W 1 and W 2 are two fully connected layers. Sigma and delta are the ReLU and sigmoid activation functions, respectively (two activation functions in the deep learning field, well known to those skilled in the art). S 2 is a function of softmax operations for each row of the weight matrix a. From the result, element a i,j is the attention of the ith channel to the jth modality, in other words, the weight of the jth modality when supplementing the ith channel with information.
Step 1.2.3: and supplementing the information of the other modes into each input mode according to the cross-mode channel weight so as to obtain a new feature map after the supplementary information of a plurality of modes.
Next, supplementing information of other modes into each input mode by using a cross-mode channel weight matrix a according to a formula (2):
Where a j represents the multiplication between the scalar and the feature map or the channel multiplication between the vector and the feature map, and a is the j-th column of the channel weight matrix a. The information of the current mode can be reserved by setting balance parameters gamma while the information is supplemented, and the value of gamma defaults to 0.5. Representing the new feature map after the kth modality supplemental information.
Step 1.2.4: and splicing the new feature graphs after the plurality of modal supplementary information to generate a spliced feature graph.
New feature map obtained by supplementing multiple modes according to formula (3)Splicing to generate a spliced characteristic diagram f all:
Aiming at the problem that Pix2Pix is only suitable for single-mode input, the invention modifies the network structure and adds the supervision of the self-expression network; aiming at the problem that CollaGAN does not fully utilize the mode complementarity, a mode attention module is provided; these two improvements improve the quality of multi-modality medical image synthesis.
Step 102: training and testing the modal attention generation type countermeasure network by adopting the multi-modal medical image data set, and generating the trained modal attention generation type countermeasure network.
Fig. 3 is a schematic diagram of a training process of a modal attention generation type countermeasure network provided by the present invention. Referring to fig. 3, the image synthesis network is trained and tested using a multi-modality medical image dataset, and the training of the network supervision switching network is self-represented during the training process. The multi-modal medical image dataset used in the training of the present invention is medical image dataset BraTS2020, which is a public dataset provided by an international brain tumor segmentation challenge (Brain Tumor Segmentation, abbreviated as BraTS), the training set contains 369 sets of samples, and the test set contains 125 sets of samples. Each set of samples contained four modalities of brain tumor nmr scans of T1, T1Gd, T2-FLAIR.
Training the self-expression network by using target modal data in the training set to obtain a pre-trained self-expression network SR; training the conversion network by using all data of the training set, and extracting the characteristics of the SR supervision generator of the self-expression network in the process to obtain the conversion network after training.
The user can synthesize the image of any target mode by using the trained conversion network, namely, inputting medical images respectively belonging to n-1 modes and masks corresponding to the target modes to the conversion network, and outputting the synthesized image of the target mode by the conversion network.
As mentioned previously, the present invention introduces a self-representing network (implemented with a self-encoder) to drive the training of the switching network. The self-encoder is an unsupervised feature extraction model with strong expression capacity in the field of deep learning, and the difficulty of the multi-branch feature extraction structure of the conversion network in extracting effective features is considered to be larger than that of the original single branch, so that the pre-trained self-expression network SR is used for driving the feature extraction of the generator. Specifically, the following losses are calculated during the training phase:
In equation (4) we assume that the nth modality is the target modality, E SR (·) represents the output of the self-representative network encoder.
In addition, the total loss of the model conversion network generator is:
Wherein the method comprises the steps of And L GAN are loss functions already defined in Pix2Pix, which are common knowledge in the art and are not described in detail herein. Lambda is the hyper-parameter of balance L SR, defaulting to 1. The calculation formula for self-representing network loss is/>I.e. the L1 norm between the generated image and the original. Equations (4) and (5) are modal attention generating type against the constraints of the training phase of the network, with the training goal being to minimize loss from phase to phase.
Specifically, the step 102 trains and tests the modal attention generation type countermeasure network by using the multi-modal medical image dataset, and generates a trained modal attention generation type countermeasure network, which specifically includes:
Step 2.1: inputting the multi-modal medical image dataset into an encoder of the self-representing network, training the self-representing network using a random gradient descent algorithm, generating a pre-trained self-representing network.
All images are taken from the training set and input into the encoder of the self-representative network, which is trained using SGD (stationary GRADIENT DESCENT, random gradient descent), resulting in a trained self-representative network SR, hereinafter referred to as pre-trained SR.
Specifically, referring to fig. 3, in the embodiment of the present invention, a training set in BraTS2020 is used to train the self-expression network SR, where the parameter of SR is W SR, and a specific training procedure of SR is as follows:
Step 1): inputting iteration times T, randomly initializing a parameter W SR, and initializing the current iteration times iter;
Step 2): traversing all images of the training set, and repeating the step 3 and the step 4 until the iter reaches T;
step 3): inputting the current images I to SR by the formula Calculating loss;
Step 4): w SR is updated using an SGD (random gradient descent) algorithm, updating iter=iter+1.
Step 2.2: inputting medical images of a target mode into the pre-trained self-expression network, inputting medical images of other modes except the target mode and masks corresponding to the target mode into a multi-encoder of the image conversion network, training the image conversion network by using a random gradient descent algorithm, and monitoring the training of the image conversion network by the self-expression network in the training process to generate a trained image conversion network; the pre-trained self-representing network and the trained image conversion network together comprise the trained modal attention-generating countermeasure network.
Referring to fig. 3, after the training of the self-presenting network SR is completed, the switching network is trained next. The above-mentioned requirement is that a mask corresponding to the target modality is added to the network input so that the model can determine which modality of image to synthesize. The mask adopts one-hot coding, taking BraTS datasets as an example, and defining that T1, T1Gd, T2 and T2-FLAIR are respectively 1 st to 4 th modes, then the mask corresponding to the T1 mode is tensor of dimension 4 x h x w (h and w are respectively width and height of an image), the mask can be regarded as being formed by splicing matrices of 4 layers h x w, then the value of the matrix of the 1 st layer is set to be 1, the rest values are set to be 0, namely the mask corresponding to the T1 mode, and then the mask corresponding to the T1Gd, the T2 and the T2-FLAIR is obtained by analogy. In training and using a network conversion network, a mask corresponding to a target modality needs to be stitched behind each input modality. The training process of the switching network is as follows:
Step 1): inputting iteration times T, parameters W SR of a pre-trained SR, randomly initializing a generator parameter W G and a discriminator parameter W D, and initializing a current iteration times item;
step 2): repeating the step3 and the step 4 until the iteration times reach T;
Step 3): because MAGAN provided by the invention can receive the input of any mode, each mode can be used as a target mode in sequence, other modes are used as input modes to synthesize images, the loss of a generator is calculated by using a formula L G, and the loss of a discriminator is calculated by using a formula L GAN;
Step 4): w G and W D are alternately updated using an SGD (random gradient descent) algorithm, with iter=iter+1 updated.
More specifically, referring to fig. 2 and 3, the detailed training steps of the image conversion network are as follows:
Step (1): the medical images from the n-1 input modalities are denoted by x 1,x2,...,xn-1 and the medical images from the n modalities, i.e. the target modality, are denoted by x n. As shown in fig. 2, the masks corresponding to x 1,x2,...,xn-1 and the target modes are respectively input into n-1 encoders to respectively obtain feature maps f 1,f2,...,fn-1;
Step (2): inputting x n into a pre-trained encoder of the self-expression network to obtain a feature map E SR(xn), and calculating the difference between E SR(xn) and f 1,f2,...,fn-1 by adopting a formula (4) to obtain a part L SR of the generator loss, thereby completing supervision of feature extraction by the self-expression network;
Step (3): inputting the f 1,f2,...,fn-1 into a modal attention module to perform modal information complementation to obtain a spliced characteristic diagram f all;
Step (4): inputting f all into a decoder of a conversion network to obtain a synthetic image of a target mode The function of the generator is to generate/>, as close as possible to the real x n
Step (5): will beInput to a arbiter, which essentially is a classifier, with the goal of distinguishing as much as possible between x n and/>
Step (6): repeating the steps (1) to (4), and alternately optimizing the generator and the discriminator by using an SGD algorithm until the appointed iteration times T are reached, so as to finally obtain the trained image conversion network.
Step 103: and inputting the existing multi-modal medical image into the trained modal attention generation type countermeasure network, and outputting a composite image of the target modality.
In the using stage of the model, the user can synthesize an image of any target mode by using the conversion network after training is completed, and the method is similar to the model training: the medical images belonging to n-1 modes and the masks corresponding to the target modes are input into n-1 encoders of a conversion network generator G, and the G outputs the synthesized images of the target modes. For example, in fig. 2, x 1,x2,...,xn-1 is input to n-1 encoders of the generator, respectively, and the output of the generator is a composite image of the nth mode (i.e., the target mode)
Thus, the step 103 inputs the existing multi-modal medical image into the trained modal attention generation type countermeasure network, and outputs a composite image of the target modality, specifically including:
and inputting the existing multi-modal medical image and the mask corresponding to the target modality into a multi-encoder of the trained image conversion network, and outputting a composite image of the target modality by a generator of the trained image conversion network.
The invention provides an image synthesis framework MAGAN in a multi-mode scene: the user inputs images of a plurality of modes and a target mode corresponding mask, and the model MAGAN automatically synthesizes medical images corresponding to the target mode; the method introduces a modal attention module, and effectively improves the quality of the synthesized image by modeling the correlation among the modalities; the present invention proposes a method for extracting features related to a target modality using a self-representative network constraint model, which allows MAGAN (modality attention generating type countermeasure network) to extract only features related to the target modality from the input. In summary, the multi-modal medical image synthesis model MAGAN based on the generated countermeasure network can receive any modal image as input and synthesize a target modal image, is mainly used for synthesizing a missing modal image in the multi-modal medical image, can improve the quality of synthesizing the missing medical modal image, and is beneficial to a doctor to make more accurate decisions.
An embodiment of a method for synthesizing a multi-modal medical image based on a generated countermeasure network according to the present invention is given below, and fig. 4 is a schematic diagram of an embodiment of a modal attention generated countermeasure network according to the present invention, where n=4 is assumed. Referring to fig. 4, a specific embodiment of the method comprises the steps of:
Step one: the network model shown in fig. 4 is constructed, and the network model comprises a conversion network and a self-expression network, wherein the conversion network is added with the modal attention module provided by the invention so as to better utilize the complementary information among modalities. The structure of the self-presenting network and the switching network is shown in fig. 4: the self-expression network consists of an encoder and a decoder; the conversion network consists of a generator and a arbiter, wherein the generator consists of n-1 encoders and one decoder, the encoders and the decoders being connected by the modal attention module presented herein.
Step two: training the self-expression network by using target modal data in the training set to obtain a pre-trained self-expression network; training the conversion network by using all data of the training set, and extracting the characteristics of the SR supervision generator of the self-expression network in the process to obtain the conversion network for completing training.
As shown in fig. 4, for a set of samples from BraTS2020, the medical images of the 4 modes are denoted by x 1,x2,x3 and x 4 respectively, in fig. 4 x 1,x2,x3 is input to 3 encoders of the generator (we assume n=4, and hence n-1=3 here) respectively, and the output of the generator is a composite image of the 4 th mode
In this particular embodiment, the training process of the switching network is as follows:
1. Respectively inputting masks corresponding to the x 1,x2,x3 and the target modes into 3 encoders to respectively obtain a feature map f 1,f2,f3;
2. Inputting the true x 4 into a pre-trained decoder of the self-expression network to obtain a feature map E SR(x4), and calculating the difference between E SR(x4) and f 1,f2,f3 by adopting a formula (4) to obtain a part L SR of the generator loss, thereby completing supervision of feature extraction by the self-expression network;
3. Inputting f 1,f2,f3 into a modal attention module, and carrying out modal information complementation to obtain f all; the method comprises the following steps:
Using equation (1) to calculate the cross-modal channel weight (denoted as a):
A=S2(σ((δ([f1,f2,f3]W1)W2)
Information of other modalities is next supplemented into each input modality using a. Fig. 5 is a schematic diagram illustrating complementation of model information in an embodiment of a modal attention generation type countermeasure network according to the present invention. In fig. 5, the 3 rd channel of the 1 st modality is supplemented with information of other modalities, the 2 nd modality supplements the 3 rd channel of the 1 st modality with weight a 3,2, and the 3 rd modality supplements the 3 rd channel of the 1 st modality with weight a 3,3. In this way, a new feature f k comp after the kth modality supplemental information can be obtained:
Here, f k=Ek(xk),Ek (·) represents the output of generator G kth encoding branch E k, [ ·, ·, · ] represents the concatenation after flattening the 3 tensors. W 1 and W 2 are two fully connected layers. Sigma and delta are the ReLU and sigmoid activation functions, respectively (two activation functions in the deep learning field, well known to those skilled in the art). S 2 is a function of softmax operation on each row of the weight matrix a, representing multiplication between scalar and feature map or channel multiplication between vector and feature map, a j is the j-th column of the weight matrix a. The information of the current mode can be reserved by setting balance parameters gamma while the information is supplemented, and the value of gamma defaults to 0.5.
Finally, since the decoder is of a single-branch structure, the mode attention module needs to splice the feature images and send the feature images into the decoder, and the output of the mode attention module is as follows:
fall=[f1 comp,f2 comp,f3 comp]
4. Inputting f all into a decoder of a conversion network to obtain a synthetic image of a target mode The function of the generator is to generate/>, as close as possible to x 4
5. Will beInput to a arbiter, which essentially is a classifier, with the goal of distinguishing as much as possible between x 4 and/>
6. Repeating the steps 1, 4, and alternately optimizing the generator and the discriminator by using the SGD algorithm until the appointed iteration times are reached, and finally obtaining the trained conversion network.
Step three: the user can synthesize the image of any target mode by using the conversion network obtained in the last step, specifically, the image belonging to n-1 modes and the mask corresponding to the target mode are input to the conversion network, and the conversion network outputs the synthesized image of the target mode.
For example, x 1,x2,x3 is input to 3 encoders of a generator, and the output of the generator is a composite image of the 4 th mode
In summary, according to the multi-mode medical image synthesis method based on the generation type countermeasure network, through a single model, any mode image can be used as input, and a target mode image (generally a mode with a missing or poor quality) can be synthesized. The method can solve the problem of missing of a certain mode image in the multi-mode medical image scene.
The experimental results of the embodiment of the present invention on BraTS's 2020 validation set are shown in table 1 and table 2, table 1 compares the scores of MAGAN and Pix2Pix proposed by the present invention on the structural similarity (structural similarity index, SSIM), and table 2 compares the scores of MAGAN and Pix2Pix proposed by the present invention on the feature similarity (feature similarity index, FSIM) (SSIM and FSIM are both methods for measuring image similarity, which are well known to those skilled in the art).
Table 1 SSIM score comparison
Table 2 FSIM score comparison
As can be seen from the comparison of the data in tables 1 and 2, the structural similarity and the characteristic similarity of the invention MAGAN are greatly improved in the synthetic scenes of various modes.
Based on the method for synthesizing the multi-modal medical image based on the generated type countermeasure network, the invention also provides a system for synthesizing the multi-modal medical image based on the generated type countermeasure network. Fig. 6 is a block diagram of a multi-modal medical image synthesis system based on a generated countermeasure network according to the present invention, and referring to fig. 6, the system includes:
a modal attention generation type antagonism network construction module 601 for constructing a modal attention generation type antagonism network; the modal attention-generating countermeasure network includes a self-representation network and an image conversion network implemented by the generating countermeasure network;
The modal attention generation type countermeasure network training module 602 is configured to train and test the modal attention generation type countermeasure network by using a multi-modal medical image dataset, and generate a trained modal attention generation type countermeasure network;
The multi-mode image synthesis module 603 is configured to input an existing multi-mode medical image into the trained modal attention generation type countermeasure network, and output a synthesized image of the target modality.
The modal attention generation type countering network construction module 601 specifically includes:
A self-expression network construction unit for constructing a self-expression network; the self-presenting network includes an encoder and a decoder;
an image conversion network construction unit for constructing an image conversion network; the image conversion network comprises a generator and a discriminator; the generator includes a multi-encoder, a modal attention module, and a decoder.
The modal attention module specifically includes:
The characteristic map acquisition unit is used for acquiring a plurality of characteristic maps corresponding to medical images of other modes except the target mode;
A channel weight calculation unit, configured to calculate a cross-modal channel weight according to the plurality of feature graphs;
the modal information complementation unit is used for supplementing the information of the other modalities into each input modality according to the cross-modal channel weight so as to obtain a new feature map after the plurality of modal supplementary information;
And the characteristic diagram splicing unit is used for splicing the new characteristic diagrams after the plurality of modal supplementary information to generate spliced characteristic diagrams.
The modal attention generation type countermeasure network training module 602 specifically includes:
A self-expression network training unit, configured to input the multi-modal medical image dataset into an encoder of the self-expression network, train the self-expression network using a random gradient descent algorithm, and generate a pre-trained self-expression network;
an image conversion network training unit, configured to input a medical image of a target modality into the pre-trained self-expression network, input medical images of other modalities except the target modality and masks corresponding to the target modality into a multi-encoder of the image conversion network, train the image conversion network using a random gradient descent algorithm, and monitor training of the image conversion network by the self-expression network during training, so as to generate a trained image conversion network; the pre-trained self-representing network and the trained image conversion network together comprise the trained modal attention-generating countermeasure network.
The multi-modal image composition module 603 specifically includes:
and the multi-mode image synthesis unit is used for inputting the masks corresponding to the existing multi-mode medical images and the target modes into a multi-encoder of the trained image conversion network, and outputting the synthesized image of the target modes by a generator of the trained image conversion network.
The invention provides a multi-modal medical image synthesis method and a system based on a generation type countermeasure network, which provide a multi-modal medical image synthesis model based on the generation type countermeasure network, and the quality of synthesized modal images can be improved by improving the structure of the existing image synthesis model so as to adapt to multi-modal input and fully utilize information of multiple modes; the invention provides a mode attention module, which supplements information required by generating a target mode for each input mode according to the complementarity among modes, solves the problem that the existing model CollaGAN does not model the correlation of the input modes, and effectively improves the quality of synthesized mode images; the invention also introduces the self-expression network to guide the training of the generator, and filters out irrelevant information in the characteristic extraction stage, thereby accelerating the convergence rate of the model and further improving the quality of the synthesized modal image.
In summary, the method and the system of the invention can adapt to the image synthesis network input by multiple modes, add the mode attention module, introduce the self-expression network to guide the training of the generator, and finally achieve the purpose of synthesizing the high-quality medical mode image with missing.
In the present specification, each embodiment is described in a progressive manner, and each embodiment is mainly described in a different point from other embodiments, and identical and similar parts between the embodiments are all enough to refer to each other. For the system disclosed in the embodiment, since it corresponds to the method disclosed in the embodiment, the description is relatively simple, and the relevant points refer to the description of the method section.
The principles and embodiments of the present invention have been described herein with reference to specific examples, the description of which is intended only to assist in understanding the methods of the present invention and the core ideas thereof; also, it is within the scope of the present invention to be modified by those of ordinary skill in the art in light of the present teachings. In view of the foregoing, this description should not be construed as limiting the invention.
Claims (6)
1. A method of multi-modal medical image synthesis based on a generated countermeasure network, comprising:
Constructing a modal attention generation type countermeasure network; the modal attention-generating countermeasure network includes a self-representation network and an image conversion network implemented by the generating countermeasure network;
the construction modality attention generation type countermeasure network specifically includes:
Constructing a self-representing network; the self-presenting network includes an encoder and a decoder;
Constructing an image conversion network; the image conversion network comprises a generator and a discriminator; the generator includes a multi-encoder, a modal attention module, and a decoder;
the method for implanting the modal attention module specifically comprises the following steps:
acquiring a plurality of feature maps corresponding to medical images of other modes except the target mode;
Calculating the cross-modal channel weight according to the feature maps;
supplementing the information of the other modes into each input mode according to the cross-mode channel weight so as to obtain a new feature diagram after the supplementary information of a plurality of modes;
Splicing the new feature graphs after the plurality of modal supplementary information to generate a spliced feature graph;
Training and testing the modal attention generation type countermeasure network by adopting a multi-modal medical image data set to generate a trained modal attention generation type countermeasure network;
and inputting the existing multi-modal medical image into the trained modal attention generation type countermeasure network, and outputting a composite image of the target modality.
2. The method according to claim 1, wherein the training and testing the modal attention-generating countermeasure network with the multimodal medical image dataset to generate a trained modal attention-generating countermeasure network, specifically comprises:
Inputting the multi-modal medical image dataset into an encoder of the self-representing network, training the self-representing network using a random gradient descent algorithm, generating a pre-trained self-representing network;
inputting medical images of a target mode into the pre-trained self-expression network, inputting medical images of other modes except the target mode and masks corresponding to the target mode into a multi-encoder of the image conversion network, training the image conversion network by using a random gradient descent algorithm, and monitoring the training of the image conversion network by the self-expression network in the training process to generate a trained image conversion network; the pre-trained self-representing network and the trained image conversion network together comprise the trained modal attention-generating countermeasure network.
3. The method according to claim 2, wherein said inputting an existing multi-modal medical image into the trained modal attention generating countermeasure network outputs a composite image of a target modality, comprising in particular:
and inputting the existing multi-modal medical image and the mask corresponding to the target modality into a multi-encoder of the trained image conversion network, and outputting a composite image of the target modality by a generator of the trained image conversion network.
4. A multi-modal medical image composition system based on a generative countermeasure network, comprising:
the modal attention generation type antagonism network construction module is used for constructing a modal attention generation type antagonism network; the modal attention-generating countermeasure network includes a self-representation network and an image conversion network implemented by the generating countermeasure network;
The modal attention generation type antagonism network construction module specifically comprises:
A self-expression network construction unit for constructing a self-expression network; the self-presenting network includes an encoder and a decoder;
An image conversion network construction unit for constructing an image conversion network; the image conversion network comprises a generator and a discriminator; the generator includes a multi-encoder, a modal attention module, and a decoder;
the modal attention module specifically includes:
The characteristic map acquisition unit is used for acquiring a plurality of characteristic maps corresponding to medical images of other modes except the target mode;
A channel weight calculation unit, configured to calculate a cross-modal channel weight according to the plurality of feature graphs;
the modal information complementation unit is used for supplementing the information of the other modalities into each input modality according to the cross-modal channel weight so as to obtain a new feature map after the plurality of modal supplementary information;
The feature map splicing unit is used for splicing the new feature maps after the plurality of modal supplementary information to generate a spliced feature map;
the system comprises a modal attention generation type countermeasure network training module, a modal attention generation type countermeasure network training module and a modal attention generation type countermeasure network training module, wherein the modal attention generation type countermeasure network training module is used for training and testing the modal attention generation type countermeasure network by adopting a multi-modal medical image data set to generate a trained modal attention generation type countermeasure network;
and the multi-modal image synthesis module is used for inputting the existing multi-modal medical images into the trained modal attention generation type countermeasure network and outputting the synthesized images of the target modalities.
5. The system of claim 4, wherein the modal attention generation type countermeasure network training module specifically comprises:
A self-expression network training unit, configured to input the multi-modal medical image dataset into an encoder of the self-expression network, train the self-expression network using a random gradient descent algorithm, and generate a pre-trained self-expression network;
an image conversion network training unit, configured to input a medical image of a target modality into the pre-trained self-expression network, input medical images of other modalities except the target modality and masks corresponding to the target modality into a multi-encoder of the image conversion network, train the image conversion network using a random gradient descent algorithm, and monitor training of the image conversion network by the self-expression network during training, so as to generate a trained image conversion network; the pre-trained self-representing network and the trained image conversion network together comprise the trained modal attention-generating countermeasure network.
6. The system of claim 5, wherein the multi-modality image synthesis module specifically comprises:
and the multi-mode image synthesis unit is used for inputting the masks corresponding to the existing multi-mode medical images and the target modes into a multi-encoder of the trained image conversion network, and outputting the synthesized image of the target modes by a generator of the trained image conversion network.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202111465819.1A CN114140368B (en) | 2021-12-03 | 2021-12-03 | Multi-mode medical image synthesis method based on generation type countermeasure network |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202111465819.1A CN114140368B (en) | 2021-12-03 | 2021-12-03 | Multi-mode medical image synthesis method based on generation type countermeasure network |
Publications (2)
Publication Number | Publication Date |
---|---|
CN114140368A CN114140368A (en) | 2022-03-04 |
CN114140368B true CN114140368B (en) | 2024-04-23 |
Family
ID=80387588
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN202111465819.1A Active CN114140368B (en) | 2021-12-03 | 2021-12-03 | Multi-mode medical image synthesis method based on generation type countermeasure network |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN114140368B (en) |
Citations (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN110706302A (en) * | 2019-10-11 | 2020-01-17 | 中山市易嘀科技有限公司 | System and method for text synthesis image |
WO2021017372A1 (en) * | 2019-08-01 | 2021-02-04 | 中国科学院深圳先进技术研究院 | Medical image segmentation method and system based on generative adversarial network, and electronic equipment |
US10990848B1 (en) * | 2019-12-27 | 2021-04-27 | Sap Se | Self-paced adversarial training for multimodal and 3D model few-shot learning |
CN112734764A (en) * | 2021-03-31 | 2021-04-30 | 电子科技大学 | Unsupervised medical image segmentation method based on countermeasure network |
CN113705400A (en) * | 2021-08-18 | 2021-11-26 | 中山大学 | Single-mode face living body detection method based on multi-mode face training |
-
2021
- 2021-12-03 CN CN202111465819.1A patent/CN114140368B/en active Active
Patent Citations (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
WO2021017372A1 (en) * | 2019-08-01 | 2021-02-04 | 中国科学院深圳先进技术研究院 | Medical image segmentation method and system based on generative adversarial network, and electronic equipment |
CN110706302A (en) * | 2019-10-11 | 2020-01-17 | 中山市易嘀科技有限公司 | System and method for text synthesis image |
US10990848B1 (en) * | 2019-12-27 | 2021-04-27 | Sap Se | Self-paced adversarial training for multimodal and 3D model few-shot learning |
CN112734764A (en) * | 2021-03-31 | 2021-04-30 | 电子科技大学 | Unsupervised medical image segmentation method based on countermeasure network |
CN113705400A (en) * | 2021-08-18 | 2021-11-26 | 中山大学 | Single-mode face living body detection method based on multi-mode face training |
Non-Patent Citations (2)
Title |
---|
梁建青 ; 胡清华.基于半监督距离学习和多模态信息的图像聚类.第十三届中国Rough集与软计算学术会议、第七届中国Web智能学术研讨会、第 七届中国粒计算学术研讨会联合学术会议.2013,全文. * |
注意力机制引导的多模态心脏图像分割;杨琬琪;周子奇;郭心娜;南京师大学报(自然科学版);20191231;第42卷(第003期);全文 * |
Also Published As
Publication number | Publication date |
---|---|
CN114140368A (en) | 2022-03-04 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN110097550B (en) | Medical image segmentation method and system based on deep learning | |
CN111476805B (en) | Cross-source unsupervised domain adaptive segmentation model based on multiple constraints | |
CN109635774B (en) | Face synthesis method based on generation of confrontation network | |
CN110544239B (en) | Multi-modal MRI conversion method, system and medium for generating countermeasure network based on conditions | |
CN113343705B (en) | Text semantic based detail preservation image generation method and system | |
CN113314205B (en) | Efficient medical image labeling and learning system | |
CN110544264A (en) | Temporal bone key anatomical structure small target segmentation method based on 3D deep supervision mechanism | |
CN109711426A (en) | A kind of pathological picture sorter and method based on GAN and transfer learning | |
CN111754532B (en) | Image segmentation model searching method, device, computer equipment and storage medium | |
CN108765512B (en) | Confrontation image generation method based on multi-level features | |
CN112651360B (en) | Skeleton action recognition method under small sample | |
CN113724206B (en) | Fundus image blood vessel segmentation method and system based on self-supervision learning | |
CN112215339B (en) | Medical data expansion method based on generation countermeasure network | |
CN110516724A (en) | Visualize the high-performance multilayer dictionary learning characteristic image processing method of operation scene | |
CN112686898A (en) | Automatic radiotherapy target area segmentation method based on self-supervision learning | |
JP2022530868A (en) | Target object attribute prediction method based on machine learning, related equipment and computer programs | |
CN115409937A (en) | Facial video expression migration model construction method based on integrated nerve radiation field and expression migration method and system | |
Cheng et al. | DDU-Net: A dual dense U-structure network for medical image segmentation | |
Cai et al. | Dstunet: Unet with efficient dense swin transformer pathway for medical image segmentation | |
CN115661165A (en) | Glioma fusion segmentation system and method based on attention enhancement coding and decoding network | |
CN111862261A (en) | FLAIR modal magnetic resonance image generation method and system | |
CN117421591A (en) | Multi-modal characterization learning method based on text-guided image block screening | |
CN112489048B (en) | Automatic optic nerve segmentation method based on depth network | |
CN109934796A (en) | A kind of automatic delineation method of organ based on Deep integrating study | |
CN114140368B (en) | Multi-mode medical image synthesis method based on generation type countermeasure network |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
GR01 | Patent grant | ||
GR01 | Patent grant |