WO2021022752A1 - Multimodal three-dimensional medical image fusion method and system, and electronic device - Google Patents

Multimodal three-dimensional medical image fusion method and system, and electronic device Download PDF

Info

Publication number
WO2021022752A1
WO2021022752A1 PCT/CN2019/125430 CN2019125430W WO2021022752A1 WO 2021022752 A1 WO2021022752 A1 WO 2021022752A1 CN 2019125430 W CN2019125430 W CN 2019125430W WO 2021022752 A1 WO2021022752 A1 WO 2021022752A1
Authority
WO
WIPO (PCT)
Prior art keywords
image
generator
mri
pet
classifier
Prior art date
Application number
PCT/CN2019/125430
Other languages
French (fr)
Chinese (zh)
Inventor
王书强
王鸿飞
陈卓
余雯
Original Assignee
深圳先进技术研究院
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by 深圳先进技术研究院 filed Critical 深圳先进技术研究院
Publication of WO2021022752A1 publication Critical patent/WO2021022752A1/en

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/21Design or setup of recognition systems or techniques; Extraction of features in feature space; Blind source separation
    • G06F18/214Generating training patterns; Bootstrap methods, e.g. bagging or boosting
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/24Classification techniques
    • G06F18/241Classification techniques relating to the classification model, e.g. parametric or non-parametric approaches
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/25Fusion techniques
    • G06F18/253Fusion techniques of extracted features
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/045Combinations of networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/08Learning methods
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T5/00Image enhancement or restoration
    • G06T5/50Image enhancement or restoration using two or more images, e.g. averaging or subtraction
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/0002Inspection of images, e.g. flaw detection
    • G06T7/0012Biomedical image inspection
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V30/00Character recognition; Recognising digital ink; Document-oriented image-based pattern recognition
    • G06V30/10Character recognition
    • G06V30/19Recognition using electronic means
    • G06V30/192Recognition using electronic means using simultaneous comparisons or correlations of the image signals with a plurality of references
    • G06V30/194References adjustable by an adaptive method, e.g. learning
    • GPHYSICS
    • G16INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR SPECIFIC APPLICATION FIELDS
    • G16HHEALTHCARE INFORMATICS, i.e. INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR THE HANDLING OR PROCESSING OF MEDICAL OR HEALTHCARE DATA
    • G16H30/00ICT specially adapted for the handling or processing of medical images
    • G16H30/20ICT specially adapted for the handling or processing of medical images for handling medical images, e.g. DICOM, HL7 or PACS
    • GPHYSICS
    • G16INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR SPECIFIC APPLICATION FIELDS
    • G16HHEALTHCARE INFORMATICS, i.e. INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR THE HANDLING OR PROCESSING OF MEDICAL OR HEALTHCARE DATA
    • G16H50/00ICT specially adapted for medical diagnosis, medical simulation or medical data mining; ICT specially adapted for detecting, monitoring or modelling epidemics or pandemics
    • G16H50/20ICT specially adapted for medical diagnosis, medical simulation or medical data mining; ICT specially adapted for detecting, monitoring or modelling epidemics or pandemics for computer-aided diagnosis, e.g. based on medical expert systems
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/10Image acquisition modality
    • G06T2207/10072Tomographic images
    • G06T2207/10088Magnetic resonance imaging [MRI]
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/10Image acquisition modality
    • G06T2207/10072Tomographic images
    • G06T2207/10104Positron emission tomography [PET]
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/20Special algorithmic details
    • G06T2207/20081Training; Learning
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/20Special algorithmic details
    • G06T2207/20084Artificial neural networks [ANN]
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/20Special algorithmic details
    • G06T2207/20212Image combination
    • G06T2207/20221Image fusion; Image merging
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/30Subject of image; Context of image processing
    • G06T2207/30004Biomedical image processing
    • G06T2207/30016Brain

Definitions

  • This application belongs to the technical field of medical image processing, and particularly relates to a multi-modal three-dimensional medical image fusion method, system and electronic equipment.
  • PET Positron Emission Computed Tomography
  • CT Computer Resonance Imaging
  • PET Positron Emission Computed Tomography
  • the PET system can detect gamma rays emitted indirectly from the radioactive tracer.
  • the tracer is injected into the human body through biologically active molecules, and then computer analysis technology is used to construct a three-dimensional PET image of the tracer concentration in the human body.
  • the PET collection process poses an inevitable potential threat to human health. Investigations show that a PET scan of the brain can increase the risk of cancer by 0.04%.
  • condition-GAN conditional-based generative confrontation network
  • cGAN conditional-based generative confrontation network
  • Berkeley AI Lab scholar Phillip et al. based on cGAN, used different image styles as constraints, and realized the style transfer from image to image by optimizing the generator network through adversarial training strategies.
  • the image style transfer network based on cGAN is restricted by the task requirements.
  • the input image and output image must be paired and have the same content.
  • Zhu et al. combined the two cGANs to design a cycle-based generation confrontation network CycleGAN, and increased the penalty of the loss function, which improved the effect of image style transfer and was no longer restricted by content consistency.
  • Pan et al. designed an MRI-PET synthesis model based on 3D-CycleGAN, which achieved accurate synthesis from MRI to PET. And using synthetic PET and MRI for feature fusion for AD (Alzheimer, Alzheimer's disease) and MCI (mild cognition impairment, mild cognitive impairment) diagnosis.
  • AD Alzheimer's disease
  • MCI mild cognitive impairment
  • the existing generative confrontation model usually consists of a classification network and a discriminant network.
  • the discriminant network takes into account the tasks of sample authenticity and pattern classification. This will cause conflicts between sample generation and the convergence point of pattern classification.
  • traditional generative models cannot be used to deal with multitasking problems.
  • the existing cross-modal image synthesis method based on the conditional generation confrontation network uses the image of a given modal as the conditional constraint information without considering the label information of the sample.
  • the existing cross-modal image synthesis and classification diagnosis research is to design independent models for the two tasks and train them separately, without considering the correlation between the two in the optimization process.
  • the present application provides a multi-modal three-dimensional medical image fusion method, system and electronic equipment, which aim to solve one of the above-mentioned technical problems in the prior art at least to a certain extent.
  • a multi-modal three-dimensional medical image fusion method includes the following steps:
  • Step a construct a multi-task generative confrontation network, the multi-task generative confrontation network including a generator, a discriminator and a classifier;
  • Step b Training the multi-task generation confrontation network according to the subject's MRI image, PET image and diagnostic tag information, so that the multi-task generation confrontation network automatically learns the associated features between the MRI image and the PET image;
  • Step c Input the MRI image of the person to be inspected into the trained multi-task generating confrontation network, the generator synthesizes the corresponding PET image according to the MRI image, and inputs the MRI image of the person to be inspected and the synthesized PET image into the classifier, The classifier merges the MRI image of the person to be detected and the synthesized PET image and outputs the disease classification prediction label of the person to be detected.
  • the training of the multi-task generation confrontation network specifically includes:
  • Step b1 Construct the confrontation loss function of the multi-task generation confrontation network; the confrontation loss in the training process is represented by an improved maximum-minimum cost function:
  • (C, G, D) respectively represent the classifier, discriminator and generator
  • (x, y, z) respectively represent the MRI image, PET image and diagnostic label information
  • ⁇ ⁇ (0, 1) is a Constant, used to control the proportion of classifier and generator loss in the training process
  • E (x,y,z) ⁇ p(x,y,z) [log D(x,y,z)] represents the discriminator
  • the samples from the real data distribution are judged as real samples; Indicates that the discriminator has identified a pair of pseudo samples in the output data space of the classifier; Indicates that the discriminator recognizes the pseudo sample label pair from the generator, where G(x,z) represents the data distribution generated by the generator;
  • the convergence point of the sample distribution p c (x,y,z) of the generator is limited to p(x,y), so that the global optimum of the model satisfies the sample distribution generated by the generator G and the classifier C is the same as the real data distribution;
  • Step b3 Introduce the generator supervision loss, and use the gradient mutual information between the target image and the generated image as a similarity measure:
  • I (A, B) and G (A, B) respectively represent the gradient information and gradient difference between the generated image and the target image.
  • the technical solution adopted in the embodiment of the application further includes: in the step b, the training the multi-task generation confrontation network according to the subject's MRI image, PET image, and diagnostic tag information also includes: the generation The MRI image is used as a conditional constraint, and the random noise input of the same dimension as the target image is mapped to the PET image.
  • the discriminator determines whether the input sample distribution (x, y, z) comes from the real data distribution or the pseudo data distribution, so
  • the classifier takes the joint distribution of MRI and PET images as input and predicts its label type; in the data-driven mode, with the gradual optimization of the generator, the discriminator updates the network parameters to identify the pseudo data distribution generated by the generator ; With the optimization of the discriminator, the classifier is optimized to make the predicted disease classification and prediction label tend to be real data without being judged as fake data by the discriminator, and then reversely affect the training of the generator; through iterative adversarial training,
  • the generator learns the potential correlation features between the MRI image and the PET image, thereby synthesizing the corresponding PET image from the input MRI image, and makes the classifier extract the key feature information from the input MRI image and the PET image. Predict the corresponding disease classification prediction label.
  • the technical solution adopted in the embodiment of the application further includes: the generator adopts a U-Net network structure, which includes an encoder and a decoder with a symmetrical network structure; in the step c, the generator synthesizes corresponding MRI images
  • the PET image specifically includes: outputting the feature map of the MRI image through the feature extraction operation of the encoder multi-layer convolution; the decoder performs the multi-layer deconvolution operation on the feature map output by the encoder, and generates the feature map The feature map of the same size as the corresponding position of the encoder undergoes multiple splicing operations, and finally the target reconstructed image is output, which is the synthesized PET image.
  • the technical solution adopted in the embodiment of the present application further includes: in the step c, the classifier fused the MRI image of the subject to be detected and the synthesized PET image and then output the disease classification prediction label of the subject to be detected specifically includes:
  • the feature extraction network extracts the feature value of the MRI image, performs convolution operation on the synthesized PET image, and extracts the feature value of the PET image; splicing the feature value of the MRI image and the PET image to form the spliced feature value.
  • the fully connected layer performs fusion and high-dimensional abstraction on the spliced feature values; the fused feature information is operated by the Softmax function to obtain the corresponding disease classification prediction label.
  • a multi-modal three-dimensional medical image fusion system including:
  • Model building module used to construct a multi-task generative confrontation network, which includes a generator, a discriminator, and a classifier;
  • Model training module used to train the multi-task generation confrontation network according to the subject's MRI images, PET images and diagnostic tag information, so that the multi-task generation confrontation network can automatically learn the association between MRI images and PET images feature;
  • Model application module used to input the MRI image of the person to be inspected into a trained multi-task generation confrontation network.
  • the generator synthesizes the corresponding PET image according to the MRI image, and inputs the MRI image of the person to be inspected and the synthesized PET image A classifier, which merges the MRI image of the person to be detected and the synthesized PET image and outputs the disease classification prediction label of the person to be detected.
  • the model training module includes:
  • Loss function construction unit used to construct the confrontation loss function of the multi-task generation confrontation network; the confrontation loss in the training process is represented by an improved maximum-minimum cost function:
  • (C, G, D) respectively represent the classifier, discriminator and generator, (x, y, z) respectively represent the MRI image, PET image and diagnostic label information;
  • ⁇ ⁇ (0, 1) is a Constant, used to control the proportion of classifier and generator loss in the training process,
  • E (x,y,z) ⁇ p(x,y,z) [log D(x,y,z)] represents the discriminator
  • the samples from the real data distribution are judged as real samples; Indicates that the discriminator has identified a pair of pseudo samples in the output data space of the classifier; It means that the discriminator recognizes the pseudo-sample label pair from the generator, where x represents the MRI modal image of the sample, z represents the sample label, and G(x,z) represents the PET modal image synthesized by the conditional generation network.
  • Generator optimization unit used to introduce generator supervision loss, using the gradient mutual information between the target image and the generated image as a similarity measure:
  • I (A, B) and G (A, B) respectively represent the gradient information and gradient difference between the generated image and the target image.
  • the technical solution adopted in the embodiment of the application further includes: the model training module trains the multi-task generation confrontation network specifically: the generator uses the MRI image as a conditional constraint, and maps the random noise input of the same dimension as the target image into For PET images, the discriminator determines whether the input sample distribution (x, y, z) comes from a real data distribution or a pseudo data distribution, and the classifier takes the joint distribution of MRI images and PET images as input, and predicts its label type ; In the data-driven mode, with the gradual optimization of the generator, the discriminator updates the network parameters to identify the pseudo data distribution generated by the generator; with the optimization of the discriminator, the classifier is optimized to encourage the classification of the disease to predict the label trend Based on real data without being judged as pseudo data by the discriminator, it then acts in the reverse direction on the training of the generator; through iterative confrontation training, the generator learns the potential correlation characteristics between MRI images and PET images, thereby The input MRI images are synthesized to obtain corresponding PET images, and the classifier is
  • the technical solution adopted by the embodiment of the application further includes: the generator adopts a U-Net network structure, which includes an encoder and a decoder with a symmetric network structure; the generator synthesizing the corresponding PET image according to the MRI image specifically includes: The feature extraction operation of the encoder multi-layer convolution to output the feature map of the MRI image; the decoder performs the multi-layer deconvolution operation on the feature map output by the encoder, and the generated feature map is the same size as the corresponding position of the encoder Perform multiple stitching operations on the feature map of, and finally output the target reconstructed image, which is the synthesized PET image.
  • the technical solution adopted in the embodiment of the present application further includes: the classifier integrates the MRI image of the subject to be detected and the synthesized PET image and then outputs the disease classification prediction label of the subject to be detected. Specifically, it includes: extracting the MRI image by using a feature extraction network. Eigenvalues, and perform convolution operations on the synthesized PET images to extract the eigenvalues of the PET images; splicing the MRI images and the eigenvalues of the PET images to form the spliced characteristic values, and the fully connected layer pairs the spliced Feature values are fused and high-dimensional abstracted; the fused feature information is operated by the Softmax function to obtain the corresponding disease classification prediction label.
  • the classifier integrates the MRI image of the subject to be detected and the synthesized PET image and then outputs the disease classification prediction label of the subject to be detected. Specifically, it includes: extracting the MRI image by using a feature extraction network. Eigenvalues, and perform convolution operations on the synthesized PET
  • an electronic device including:
  • At least one processor At least one processor
  • a memory communicatively connected with the at least one processor; wherein,
  • the memory stores instructions executable by the one processor, and the instructions are executed by the at least one processor, so that the at least one processor can execute the following of the above-mentioned multi-modal three-dimensional medical image fusion method operating:
  • Step a construct a multi-task generative confrontation network, the multi-task generative confrontation network including a generator, a discriminator and a classifier;
  • Step b Training the multi-task generation confrontation network according to the subject's MRI image, PET image and diagnostic tag information, so that the multi-task generation confrontation network automatically learns the associated features between the MRI image and the PET image;
  • Step c Input the MRI image of the person to be inspected into the trained multi-task generating confrontation network, the generator synthesizes the corresponding PET image according to the MRI image, and inputs the MRI image of the person to be inspected and the synthesized PET image into the classifier, The classifier merges the MRI image of the person to be detected and the synthesized PET image and outputs the disease classification prediction label of the person to be detected.
  • the beneficial effects produced by the embodiments of the present application are: the multi-modal three-dimensional medical image fusion method, system and electronic equipment of the embodiments of the present application propose a multi-task generation confrontation model, which is based on the subject's lesion location Synthesize the pattern image in the PET image from the MRI image, and fuse the real MRI image and the synthesized PET image to obtain more key features for classification and diagnosis, and classify the disease types according to the key features.
  • this application has at least the following beneficial effects:
  • the trained generative model learns the associated features of MRI and PET imaging, and can synthesize the corresponding PET from the MRI of the examinee.
  • the classification model fuses the characteristic information of MRI and synthetic PET to classify and diagnose diseases, avoiding the high cost of PET collection At the same time as the risk of radiation exposure, functional imaging features are effectively integrated, which can achieve higher classification accuracy.
  • This application considers the joint distribution of the three attributes of MRI, PET, and diagnostic tags.
  • the model can extract richer correlation feature information between multimodal imaging and classification diagnosis, and improve image generation errors and classification diagnosis performance. Through the cumulative training of a large number of cases, the accuracy and robustness of the prediction model are gradually improved.
  • the multi-task generative confrontation network proposed by the present invention can also be used in other collaborative optimization application scenarios.
  • FIG. 1 is a flowchart of a multi-modal three-dimensional medical image fusion method according to an embodiment of the present application
  • Figure 2 is the overall framework diagram of the multi-task generation confrontation network
  • Figure 3 is a schematic diagram of the network structure of the generator
  • Figure 4 is a schematic diagram of the network structure of the classifier
  • Figure 5 is an application flow chart of a multi-task generating confrontation network
  • FIG. 6 is a schematic structural diagram of a multi-modal three-dimensional medical image fusion system according to an embodiment of the present application.
  • FIG. 7 is a schematic diagram of the hardware device structure of the multi-modal three-dimensional medical image fusion method provided by an embodiment of the present application.
  • the multi-modal three-dimensional medical image fusion method in the embodiments of the present application proposes a multi-task generation confrontation model (Multi- Task GAN, MT-GAN), predict the pattern image in PET imaging based on the MRI image of the subject's lesion, and realize the confrontation between the cross-modal image synthesis network and the multi-modal fusion classification network in the data-driven mode Co-training, the optimized system successfully learned the potential correlation features between MRI imaging, PET imaging and disease diagnosis, and solved the conflict problem between the generation network and the discrimination network convergence point in the traditional generative confrontation model.
  • Multi- Task GAN MT-GAN
  • This application integrates multi-source medical imaging features without the need for PET collection by the examinee, which can more accurately assist doctors in clinical diagnosis.
  • the following examples are based on MRI and PET imaging of Alzheimer’s disease as an example, but the scope of application of this application is not limited to diseases of Alzheimer’s disease and MRI-PET imaging. It can also be widely used in CT-PET, MRI-CT and other modal images of other diseases.
  • FIG. 1 is a flowchart of a multi-modal three-dimensional medical image fusion method according to an embodiment of the present application.
  • the multi-modal three-dimensional medical image fusion method of the embodiment of the present application includes the following steps:
  • Step 100 Collect MRI images and PET images of the subject, and preprocess the collected MRI images and PET images to obtain a data set for training the model;
  • step 100 the acquisition of MRI images and PET images is specifically: selecting Alzheimer's disease (AD) to be tested, mild cognitive impairment (MCI) to be tested, and normal elderly (Normal) as the recipients.
  • the examinee collects the MRI image and PET image of his brain as the original data set, and in the clinical observation and diagnosis of each subject, a professional physician will give the diagnosis information, and use the diagnosis information as the Diagnostic label information.
  • the preprocessing of the original data set is specifically: using FSL, SPM and other technologies to perform redundant tissue removal and image correction processing on the collected brain MRI and PET, and use FSL brain image processing tools to perform linear registration operations on MRI and PET images. Make the anatomical points of MRI and PET images in the diagnostic sense reach the same spatial position.
  • Step 200 Construct a multi-task generating confrontation network
  • the multi-task generation confrontation network framework is shown in FIG. 2.
  • the multi-task generation confrontation network needs to consider the three attributes of MRI image, PET image and diagnostic label information of each subject to be detected. It includes classifier C, generator G, and discriminator D.
  • Generator G is used to pass real MRI images Synthesize the corresponding PET images;
  • the discriminator D is used to determine whether the data pattern distribution comes from the real data or the pseudo sample distribution;
  • the classifier C is used to fuse the MRI image and the synthesized PET image to output the disease classification prediction label of the subject .
  • the network structure of the generator is shown in FIG. 3.
  • the generator adopts the U-Net network structure, and the U-Net model is designed based on a jump-connected full convolutional network.
  • the main idea is to design an encoder and decoder with a symmetric network structure to have the same number and size of feature maps , And combine the corresponding feature maps of the encoder and the decoder through skip connection, which can retain the feature information in the down-sampling process to the maximum extent, thereby improving the efficiency of feature expression.
  • MRI images and PET images come from the same sample and share a large amount of primary feature information between them. Therefore, the U-net model is very suitable for complex feature mapping between the two modal images.
  • the generator synthesizes the corresponding PET image with real MRI image samples as follows:
  • the decoder reconstructs the feature map output by the encoder; first, the 1024 feature maps output by the encoder are deconvolved to generate 512 2 ⁇ 2 ⁇ 2 feature maps, which correspond to the encoder Feature maps of the same size at the same location are stitched together. After 6 layers of deconvolution operation and splicing operation in turn, the final output is the target reconstructed image of 128 ⁇ 128 ⁇ 128 size, which is the synthesized PET image.
  • the classifier in the embodiment of the present application selects a relatively simple multi-modal fusion classification network, and its structure is shown in FIG. 4.
  • the processing flow of the multi-modal image data by the classifier is:
  • the feature extraction network is used to extract the feature information of the MRI image; first, the two convolutional layers of the 2 ⁇ 2 ⁇ 2 size convolution kernel are used to extract the primary features of the image to generate 32 feature maps, and then a layer of window size is 2 The ⁇ 2 ⁇ 2 pooling layer reduces the dimensionality of the feature map. Subsequently, a 3 ⁇ 3 ⁇ 3 size convolution kernel is used to extract advanced features.
  • the third and fourth convolution layers use 64 convolution kernels respectively, and the extracted features are pooled and reduced in dimension, and then 128 convolution kernels are used for Higher-dimensional feature extraction.
  • the integrated fusion feature information is operated by the Softmax function to obtain the corresponding label prediction type (that is, the probability of the predicted image data corresponding to the disease level).
  • Step 300 Train the multi-task generation confrontation network according to the subject's MRI image, PET image and diagnostic tag information;
  • step 300 the training process of the multi-task generating confrontation network includes the following steps:
  • Step 301 Construct the model's anti-loss function
  • the adversarial loss of the training process can be represented by an improved minimax cost function:
  • ⁇ (0,1) is a constant, which is used to control the proportion of classifier and generator loss in the training process, that is, the relative importance in the confrontation training task.
  • E (x,y,z) ⁇ p(x,y,z) [log D(x,y,z)] means that the discriminator judges the samples from the real data distribution as real samples; Indicates that the discriminator has identified a pair of pseudo samples in the output data space of the classifier; It means that the discriminator recognizes the pseudo-sample label pair from the generator, where x represents the MRI modal image of the sample, z represents the sample label, and G(x,z) represents the PET modal image synthesized by the conditional generation network.
  • the multi-task generating confrontation network's confrontation loss function is constructed.
  • the equilibrium of the adversarial game shows that when one of the generator G and the classifier C reaches the optimum, the other also approaches the optimum.
  • Step 303 Introduce generator supervision loss
  • I (A, B) and G (A, B) respectively represent the gradient information and gradient difference between the generated image and the target image.
  • Step 304 Divide the data set of 900 subjects into a training set and a test set, train the multi-task generation confrontation network through the training set, and test the performance of the multi-task generation confrontation network through the test set;
  • step 304 there are 700 sample data in the training set, and 200 sample data in the test set.
  • the specific training process of the model is as follows: In the data-driven mode, as the generator G is gradually optimized, the discriminator D needs to update the network parameters to identify the pseudo data distribution generated by the generator G; as the discriminator D optimizes the incentive classification
  • the device C is optimized so that the predicted disease classification prediction label tends to be real data without being judged as fake data by the discriminator D, and then acts on the training of the generator G in reverse.
  • the generator G and the classifier C are optimized in the confrontation training, and in the process of the three confrontation games, the classifier and the generator are better than the independent training. Performance.
  • Step 400 Input the MRI image of the person to be detected into the trained multi-task generating confrontation network, and the multi-task generating confrontation network outputs the disease classification prediction label of the person to be detected;
  • step 400 after the confrontation training, the generator G learns the potential correlation features between the MRI image and the PET image, and can more accurately synthesize the corresponding PET image from the input MRI image.
  • the parameters of the classifier are also optimized, and key feature information can be extracted from the input MRI images and PET images, and the corresponding disease classification prediction labels can be predicted based on the features.
  • the application process of the multi-task generating confrontation network specifically includes the following steps:
  • Step 401 Collect MRI images of the person to be tested
  • Step 402 Input the MRI image into the trained generator for synthesis, and the generator synthesizes the corresponding PET image according to the MRI image;
  • Step 403 Input the MRI image and the synthesized PET image into the trained classifier, and the classifier outputs the disease classification prediction label of the person to be detected.
  • FIG. 6 is a schematic structural diagram of a multi-modal three-dimensional medical image fusion system according to an embodiment of the present application.
  • the multi-modal three-dimensional medical image fusion system of the embodiment of the present application includes a data acquisition module, a model construction module, a model training module, and a model application module.
  • Data acquisition module used to acquire MRI images and PET images of subjects, and preprocess the acquired MRI images and PET images to obtain a data set for training the model; among them, the acquisition of MRI images and PET images is specifically: Select 300 people each for Alzheimer's disease (AD) to be tested, mild cognitive impairment (MCI) to be tested, and normal elderly (Normal) as subjects, and collect MRI and PET images of their brains As the original data set, and in the clinical observation and diagnosis of each subject, professional physicians will give diagnostic information, and use the diagnostic information as the diagnostic label information of each subject.
  • AD Alzheimer's disease
  • MCI mild cognitive impairment
  • Normal normal elderly
  • the preprocessing of MRI and PET images is specifically: using FSL, SPM and other technologies to perform redundant tissue removal and image correction processing on the collected brain MRI and PET, and use FSL brain image processing tools to linearly register MRI and PET images Operation to make the anatomical points of MRI and PET images in the diagnostic sense reach the same spatial position.
  • Model building module used to build a multi-task generative confrontation network; among them, the multi-task generative confrontation network includes a classifier C, a generator G, and a discriminator D.
  • the generator G is used to synthesize corresponding PET images from real MRI images;
  • the device D is used to determine whether the data pattern distribution comes from real data or a pseudo sample distribution;
  • the classifier C is used to fuse the MRI image and the synthesized PET image to output the disease classification prediction label of the subject to be detected.
  • the generator adopts the U-Net network structure.
  • the U-Net model is designed based on a jump-connected full convolutional network.
  • the main idea is to design an encoder and a decoder with a symmetric network structure to have the same
  • the number and size of feature maps, and the corresponding feature maps of the encoder and decoder are combined through skip connection, which can retain the feature information in the downsampling process to the maximum extent, thereby improving the efficiency of feature expression.
  • MRI images and PET images come from the same sample and share a large amount of primary feature information between them. Therefore, the U-net model is very suitable for complex feature mapping between the two modal images.
  • the generator synthesizes the corresponding PET image with real MRI image samples as follows:
  • the decoder reconstructs the feature map output by the encoder; first, the 1024 feature maps output by the encoder are deconvolved to generate 512 2 ⁇ 2 ⁇ 2 feature maps, which correspond to the encoder Feature maps of the same size at the same location are stitched together. After 6 layers of deconvolution operation and splicing operation in turn, the final output is the target reconstructed image of 128 ⁇ 128 ⁇ 128 size, which is the synthesized PET image.
  • the classifier in this embodiment of the application selects a relatively simple multi-modal fusion classification network, and the classifier processes the multi-modal image data as follows:
  • Extract the feature information of the MRI image first use the two convolutional layers of the 2 ⁇ 2 ⁇ 2 size convolution kernel to extract the primary features of the image to generate 32 feature maps, and then use a layer of window size of 2 ⁇ 2 ⁇ 2
  • the pooling layer reduces the dimensionality of the feature map.
  • a 3 ⁇ 3 ⁇ 3 size convolution kernel is used to extract advanced features.
  • the third and fourth convolution layers use 64 convolution kernels respectively, and the extracted features are pooled and reduced in dimension, and then 128 convolution kernels are used for Higher-dimensional feature extraction.
  • the integrated fusion feature information is operated by the Softmax function to obtain the corresponding label prediction type (that is, the probability of the predicted image data corresponding to the disease level).
  • Model training module used to train the multi-task generation confrontation network based on the subject’s MRI images, PET images and diagnostic tag information; the model training module includes:
  • Loss function construction unit used to construct the model's adversarial loss function; in the actual application process of the model, the MRI data of each person to be tested needs to be collected, so the prediction process of the generator and the classifier are respectively the following conditional distributions:
  • the adversarial loss of the training process can be represented by an improved minimax cost function:
  • ⁇ (0,1) is a constant, which is used to control the proportion of classifier and generator loss in the training process, that is, the relative importance in the confrontation training task.
  • E (x,y,z) ⁇ p(x,y,z) [log D(x,y,z)] means that the discriminator judges the samples from the real data distribution as real samples; Indicates that the discriminator has identified a pair of pseudo samples in the output data space of the classifier; As a result, the multi-task generating confrontation network's confrontation loss function is constructed.
  • the equilibrium of the adversarial game shows that when one of the generator G and the classifier C reaches the optimum, the other also approaches the optimum.
  • Generator optimization unit used to introduce generator supervision loss; in generator training, in addition to the need to generate samples to make the discriminator difficult to identify from the loss function design, it is also necessary to ensure that the generated samples are as similar as possible to the target image.
  • This application uses the gradient mutual information between the target image and the generated image as a similarity measure:
  • I(A,B) and G(A,B) respectively represent the gradient information and gradient difference between the generated image and the target image.
  • Model training unit used to divide the data set of 900 subjects into a training set and a test set, train the multi-task generating confrontation network through the training set, and test the performance of the multi-task generating confrontation network through the test set; ,
  • the sample data in the training set is 700, and the sample data in the test set is 200.
  • the specific training process of the model is as follows: In the data-driven mode, as the generator G is gradually optimized, the discriminator D needs to update the network parameters to identify the pseudo data distribution generated by the generator G; as the discriminator D optimizes the incentive classification
  • the device C is optimized so that the predicted disease classification prediction label tends to be real data without being judged as fake data by the discriminator D, and then acts on the training of the generator G in reverse.
  • the generator G and the classifier C are optimized in the confrontation training, and in the process of the three confrontation games, the classifier and the generator are better than the independent training. Performance.
  • Model application module used to input the MRI image of the person to be detected into the trained multi-task generation confrontation network, and the multi-task generation confrontation network outputs the disease classification prediction label of the person to be detected; after the confrontation training, the generator G learns the MRI image
  • the potential associated features with PET images can be more accurately synthesized from the input MRI images to obtain the corresponding PET images.
  • the parameters of the classifier are also optimized, and key feature information can be extracted from the input MRI images and PET images, and the corresponding disease classification prediction labels can be predicted based on the features.
  • the application process of the multi-task generation confrontation network is specifically: collecting the MRI image of the person to be detected, inputting the MRI image into the trained generator for synthesis, and the generator synthesizing the corresponding PET image according to the MRI image; combining the MRI image with The synthesized PET image is input to the trained classifier, and the classifier outputs the disease classification prediction label of the person to be detected.
  • FIG. 7 is a schematic diagram of the hardware device structure of the multi-modal three-dimensional medical image fusion method provided by an embodiment of the present application.
  • the device includes one or more processors and memory. Taking a processor as an example, the device may also include: an input system and an output system.
  • the processor, the memory, the input system, and the output system may be connected by a bus or in other ways.
  • the connection by a bus is taken as an example.
  • the memory can be used to store non-transitory software programs, non-transitory computer executable programs, and modules.
  • the processor executes various functional applications and data processing of the electronic device by running non-transitory software programs, instructions, and modules stored in the memory, that is, realizing the processing methods of the foregoing method embodiments.
  • the memory may include a program storage area and a data storage area, where the program storage area can store an operating system and an application program required by at least one function; the data storage area can store data and the like.
  • the memory may include a high-speed random access memory, and may also include a non-transitory memory, such as at least one magnetic disk storage device, a flash memory device, or other non-transitory solid state storage devices.
  • the storage may optionally include storage remotely arranged with respect to the processor, and these remote storages may be connected to the processing system through a network. Examples of the aforementioned networks include, but are not limited to, the Internet, corporate intranets, local area networks, mobile communication networks, and combinations thereof.
  • the input system can receive input digital or character information, and generate signal input.
  • the output system may include display devices such as a display screen.
  • the one or more modules are stored in the memory, and when executed by the one or more processors, the following operations of any of the foregoing method embodiments are performed:
  • Step a construct a multi-task generative confrontation network, the multi-task generative confrontation network including a generator, a discriminator and a classifier;
  • Step b Training the multi-task generating confrontation network according to the subject's MRI image, PET image and diagnostic tag information, so that the multi-task generating confrontation network automatically learns the associated features between the MRI image and the PET image;
  • Step c Input the MRI image of the person to be inspected into the trained multi-task generating confrontation network, the generator synthesizes the corresponding PET image according to the MRI image, and inputs the MRI image of the person to be inspected and the synthesized PET image into the classifier, The classifier merges the MRI image of the person to be detected and the synthesized PET image and outputs the disease classification prediction label of the person to be detected.
  • the embodiments of the present application provide a non-transitory (non-volatile) computer storage medium, the computer storage medium stores computer executable instructions, and the computer executable instructions can perform the following operations:
  • Step a construct a multi-task generative confrontation network, the multi-task generative confrontation network including a generator, a discriminator and a classifier;
  • Step b Training the multi-task generation confrontation network according to the subject's MRI image, PET image and diagnostic tag information, so that the multi-task generation confrontation network automatically learns the associated features between the MRI image and the PET image;
  • Step c Input the MRI image of the person to be inspected into the trained multi-task generating confrontation network, the generator synthesizes the corresponding PET image according to the MRI image, and inputs the MRI image of the person to be inspected and the synthesized PET image into the classifier, The classifier merges the MRI image of the person to be detected and the synthesized PET image and outputs the disease classification prediction label of the person to be detected.
  • the embodiment of the present application provides a computer program product, the computer program product includes a computer program stored on a non-transitory computer-readable storage medium, the computer program includes program instructions, when the program instructions are executed by a computer To make the computer do the following:
  • Step a construct a multi-task generative confrontation network, the multi-task generative confrontation network including a generator, a discriminator and a classifier;
  • Step b Training the multi-task generation confrontation network according to the subject's MRI image, PET image and diagnostic tag information, so that the multi-task generation confrontation network automatically learns the associated features between the MRI image and the PET image;
  • Step c Input the MRI image of the person to be inspected into the trained multi-task generating confrontation network, the generator synthesizes the corresponding PET image according to the MRI image, and inputs the MRI image of the person to be inspected and the synthesized PET image into the classifier, The classifier merges the MRI image of the person to be detected and the synthesized PET image and outputs the disease classification prediction label of the person to be detected.
  • the multi-modal three-dimensional medical image fusion method, system and electronic equipment of the embodiments of the present application propose a multi-task generation confrontation model, which synthesizes the pattern image in the PET image according to the MRI image of the subject’s lesion, and merges it After real MRI images and synthetic PET images, more key features for classification and diagnosis are obtained, and disease types are classified according to the key features.
  • this application has at least the following beneficial effects:
  • the trained generative model learns the associated features of MRI and PET imaging, and can synthesize the corresponding PET from the MRI of the examinee.
  • the classification model fuses the characteristic information of MRI and synthetic PET to classify and diagnose diseases, avoiding the high cost of PET collection At the same time as the risk of radiation exposure, functional imaging features are effectively integrated, which can achieve higher classification accuracy.
  • This application considers the joint distribution of the three attributes of MRI, PET, and diagnostic tags.
  • the model can extract richer associated feature information between multimodal imaging and classification diagnosis, and improve image generation errors and classification diagnosis performance. Through the cumulative training of a large number of cases, the accuracy and robustness of the prediction model are gradually improved.
  • the multi-task generative confrontation network proposed by the present invention can also be used in other collaborative optimization application scenarios.

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • Data Mining & Analysis (AREA)
  • Health & Medical Sciences (AREA)
  • General Physics & Mathematics (AREA)
  • Evolutionary Computation (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • General Engineering & Computer Science (AREA)
  • Artificial Intelligence (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • General Health & Medical Sciences (AREA)
  • Medical Informatics (AREA)
  • Biomedical Technology (AREA)
  • Evolutionary Biology (AREA)
  • Bioinformatics & Computational Biology (AREA)
  • Bioinformatics & Cheminformatics (AREA)
  • Public Health (AREA)
  • Computational Linguistics (AREA)
  • Biophysics (AREA)
  • Mathematical Physics (AREA)
  • Software Systems (AREA)
  • Databases & Information Systems (AREA)
  • Computing Systems (AREA)
  • Nuclear Medicine, Radiotherapy & Molecular Imaging (AREA)
  • Radiology & Medical Imaging (AREA)
  • Epidemiology (AREA)
  • Molecular Biology (AREA)
  • Primary Health Care (AREA)
  • Multimedia (AREA)
  • Quality & Reliability (AREA)
  • Pathology (AREA)
  • Magnetic Resonance Imaging Apparatus (AREA)
  • Nuclear Medicine (AREA)
  • Measuring And Recording Apparatus For Diagnosis (AREA)

Abstract

Disclosed are a multimodal three-dimensional medical image fusion method and system, and an electronic device. The method comprises: constructing a multi-task generative adversarial network, wherein the multi-task generative adversarial network comprises a generator, a discriminator and a classifier; training the multi-task generative adversarial network according to an MRI image, a PET image and diagnosis label information of a subject; and inputting an MRI image of a person to be tested into the trained multi-task generative adversarial network, the generator synthesizing a corresponding PET image according to the MRI image and inputting the MRI image of said person and the synthesized PET image into the classifier, and outputting a disease classification prediction label of said person after the MRI image of said person and the synthesized PET image are fused. By means of the method, the problem in a traditional generative adversarial network of a conflict between loss function convergence points, which possibly occurs when the performance of a generator and the performance of a classifier are taken into consideration, is solved, such that the generator and the classifier can be optimal at the same time.

Description

一种多模态三维医学影像融合方法、系统及电子设备Multimodal three-dimensional medical image fusion method, system and electronic equipment 技术领域Technical field
本申请属于医学影像处理技术领域,特别涉及一种多模态三维医学影像融合方法、系统及电子设备。This application belongs to the technical field of medical image processing, and particularly relates to a multi-modal three-dimensional medical image fusion method, system and electronic equipment.
背景技术Background technique
与MRI(Magnetic Resonance Imaging,磁共振成像)、CT(Computed Tomography,电子计算机断层扫描)等影像不同,PET(Positron Emission Computed Tomography,正电子发射型计算机断层显像)是一种对人体代谢过程进行体内观察的功能性成像技术,已逐渐广泛用于临床诊断和早期干预。具体而言,PET系统可以检测从放射性示踪剂间接发射的伽马射线,首先将示踪剂通过生物活性分子注入人体,然后利用计算机分析技术构建人体内示踪剂浓度的三维PET影像。PET的采集过程对人体健康造成不可避免的潜在威胁,调查显示,一次脑部PET扫描可以使患癌风险增加0.04%。虽然这个数字很小,但是在治疗过程中的反复扫描可以使患癌风险成倍增加。由于结构性影像MRI和功能性影像PET具有互补性,因此融合两种模态的影像数据可以获取更多用于分类诊断的关键特征,多模态融合的方法是当前提升辅助诊断模型性能的有效方法之一。但由于PET数据采集昂贵,样本量不够充足,很难获取足够的数据量训练模型。Different from MRI (Magnetic Resonance Imaging), CT (Computed Tomography) and other images, PET (Positron Emission Computed Tomography, Positron Emission Computed Tomography) is a kind of The functional imaging technology of in vivo observation has gradually been widely used in clinical diagnosis and early intervention. Specifically, the PET system can detect gamma rays emitted indirectly from the radioactive tracer. First, the tracer is injected into the human body through biologically active molecules, and then computer analysis technology is used to construct a three-dimensional PET image of the tracer concentration in the human body. The PET collection process poses an inevitable potential threat to human health. Investigations show that a PET scan of the brain can increase the risk of cancer by 0.04%. Although this number is small, repeated scans during treatment can double the risk of cancer. Because structural imaging MRI and functional imaging PET are complementary, fusing the image data of the two modalities can obtain more key features for classification and diagnosis. The multimodal fusion method is currently effective in improving the performance of auxiliary diagnosis models. One of the methods. However, due to the expensive PET data collection and insufficient sample size, it is difficult to obtain sufficient data to train the model.
生成式对抗网络最早由LanGoodfellow等人于2014年提出,继而引领了GAN改进和应用研究的热潮。2016年,Salimans等人对GAN训练和应用过程中出现的问题进行了理论分析和解释,并给出了经验性的解决方案(Improved-GAN)。Odena等人对GAN网络进行改进并应用于半监督学习中,利用无标签的数据和对抗训练策略提升分类器性能。当GAN用于图像生成时, 通常面临生成图像语义难以控制和无法保证图像多样性的问题。为此,Mehdi等人设计了基于条件控制的生成式对抗网络(conditional-GAN,cGAN),可以将数据标签或多模态属性作为条件变量,用于指导图像的生成。伯克利AI实验室学者Phillip等人在cGAN的基础上,将图像的不同风格作为约束条件,通过对抗训练策略优化生成器网络实现了由图像到图像的风格迁移。但是,基于cGAN的图像风格迁移网络受到任务要求的限制,输入图像和输出图像必须是成对的且内容一致。Zhu等人将两个cGAN组合设计了循环式生成对抗网络CycleGAN,并增加了损失函数的惩罚相,提升了图像风格迁移的效果且不再受内容一致性的限制。The generative confrontation network was first proposed by LanGoodfellow and others in 2014, which led to the upsurge of GAN improvement and application research. In 2016, Salimans and others conducted a theoretical analysis and explanation of the problems that occurred during GAN training and application, and gave an empirical solution (Improved-GAN). Odena et al. improved the GAN network and applied it to semi-supervised learning, using unlabeled data and adversarial training strategies to improve the performance of the classifier. When GAN is used for image generation, it usually faces the problem that the semantics of the generated image is difficult to control and the image diversity cannot be guaranteed. To this end, Mehdi et al. designed a conditional-based generative confrontation network (conditional-GAN, cGAN), which can use data labels or multimodal attributes as conditional variables to guide image generation. Berkeley AI Lab scholar Phillip et al., based on cGAN, used different image styles as constraints, and realized the style transfer from image to image by optimizing the generator network through adversarial training strategies. However, the image style transfer network based on cGAN is restricted by the task requirements. The input image and output image must be paired and have the same content. Zhu et al. combined the two cGANs to design a cycle-based generation confrontation network CycleGAN, and increased the penalty of the loss function, which improved the effect of image style transfer and was no longer restricted by content consistency.
Wang等人借鉴cGAN的基本框架并采用U-Net结构作为生成器网络,实现了由低计量造影剂采集的PET生成高质量PET影像。并提出了一种递进式的生成网络架构,逐级提高合成影像的质量,这项研究对于降低PET成像成本和对人体的潜在辐射威胁具有重要意义。Nie等人融合cGAN网络和FCN实现了脑部MRI到CT影像的合成。针对腹腔部位医学影像空间复杂度高、合成误差大等问题,Hiasa等人设计了基于CycleGAN的跨模态影像合成方法,实现了由MRI到CT的迁移,降低了合成影像与真实影像之间的误差。最近,Pan等人设计了基于3D-CycleGAN的MRI-PET合成模型,实现了由MRI到PET的精准合成。并利用合成的PET和MRI进行特征融合用于AD(Alzheimer,阿尔兹海默症)和MCI(mild cognition impairment,轻度认知功能损害)诊断。Wang et al. borrowed the basic framework of cGAN and adopted the U-Net structure as the generator network to realize the generation of high-quality PET images from PET collected by low-meter contrast agents. And proposed a progressive generation network architecture to gradually improve the quality of the synthesized image. This research is of great significance for reducing the cost of PET imaging and the potential radiation threat to the human body. Nie et al. fused cGAN network and FCN to realize the synthesis of brain MRI to CT image. Aiming at the problems of high spatial complexity and large synthesis errors in abdominal medical imaging, Hiasa et al. designed a cross-modal image synthesis method based on CycleGAN, which realized the migration from MRI to CT, and reduced the difference between the synthesized image and the real image. error. Recently, Pan et al. designed an MRI-PET synthesis model based on 3D-CycleGAN, which achieved accurate synthesis from MRI to PET. And using synthetic PET and MRI for feature fusion for AD (Alzheimer, Alzheimer's disease) and MCI (mild cognition impairment, mild cognitive impairment) diagnosis.
然而,现有的生成对抗模型通常由分类网络和判别网络组成,在处理分类问题时判别网络同时兼顾样本真伪判别和模式分类的任务。这样会造成样本生成和模式分类收敛点的冲突问题,换言之,传统的生成模型无法用于处理多任务问题。同时,现有的基于条件生成对抗网络的跨模态影像合成方法是以给定 模态的影像作为条件约束信息,而没有考虑样本的标签信息。另外,现有的跨模态影像合成和分类诊断研究是对两个任务设计独立的模型并分别训练,没有考虑二者在优化过程中的关联性。However, the existing generative confrontation model usually consists of a classification network and a discriminant network. When dealing with classification problems, the discriminant network takes into account the tasks of sample authenticity and pattern classification. This will cause conflicts between sample generation and the convergence point of pattern classification. In other words, traditional generative models cannot be used to deal with multitasking problems. At the same time, the existing cross-modal image synthesis method based on the conditional generation confrontation network uses the image of a given modal as the conditional constraint information without considering the label information of the sample. In addition, the existing cross-modal image synthesis and classification diagnosis research is to design independent models for the two tasks and train them separately, without considering the correlation between the two in the optimization process.
发明内容Summary of the invention
本申请提供了一种多模态三维医学影像融合方法、系统及电子设备,旨在至少在一定程度上解决现有技术中的上述技术问题之一。The present application provides a multi-modal three-dimensional medical image fusion method, system and electronic equipment, which aim to solve one of the above-mentioned technical problems in the prior art at least to a certain extent.
为了解决上述问题,本申请提供了如下技术方案:In order to solve the above-mentioned problems, this application provides the following technical solutions:
一种多模态三维医学影像融合方法,包括以下步骤:A multi-modal three-dimensional medical image fusion method includes the following steps:
步骤a:构建多任务生成对抗网络,所述多任务生成对抗网络包括生成器、判别器和分类器;Step a: construct a multi-task generative confrontation network, the multi-task generative confrontation network including a generator, a discriminator and a classifier;
步骤b:根据受试者的MRI影像、PET影像和诊断标签信息对所述多任务生成对抗网络进行训练,使所述多任务生成对抗网络自动学习MRI影像和PET影像之间的关联特征;Step b: Training the multi-task generation confrontation network according to the subject's MRI image, PET image and diagnostic tag information, so that the multi-task generation confrontation network automatically learns the associated features between the MRI image and the PET image;
步骤c:将待检测者的MRI影像输入训练好的多任务生成对抗网络,所述生成器根据MRI影像合成对应的PET影像,并将待检测者的MRI影像与合成的PET影像输入分类器,所述分类器将待检测者的MRI影像与合成的PET影像进入融合后输出待检测者的疾病分类预测标签。Step c: Input the MRI image of the person to be inspected into the trained multi-task generating confrontation network, the generator synthesizes the corresponding PET image according to the MRI image, and inputs the MRI image of the person to be inspected and the synthesized PET image into the classifier, The classifier merges the MRI image of the person to be detected and the synthesized PET image and outputs the disease classification prediction label of the person to be detected.
本申请实施例采取的技术方案还包括:在所述步骤b中,所述对多任务生成对抗网络进行训练具体包括:The technical solution adopted in the embodiment of the present application further includes: in the step b, the training of the multi-task generation confrontation network specifically includes:
步骤b1:构建所述多任务生成对抗网络的对抗损失函数;训练过程的对抗损失用改进的极大极小代价函数进行表示:Step b1: Construct the confrontation loss function of the multi-task generation confrontation network; the confrontation loss in the training process is represented by an improved maximum-minimum cost function:
Figure PCTCN2019125430-appb-000001
Figure PCTCN2019125430-appb-000001
上述公式中,(C,G,D)分别表示分类器、判别器和生成器,(x,y,z)分别表示MRI影像、PET影像和诊断标签信息;α∈(0,1)是一个常量,用于控制分类器和生成器损失在训练过程中所占比重,E (x,y,z)~p(x,y,z)[log D(x,y,z)]表示判别器将来自于真实数据分布中的样本判定为真实样本;
Figure PCTCN2019125430-appb-000002
Figure PCTCN2019125430-appb-000003
表示判别器识别出有分类器输出数据空间中的伪样本对;
Figure PCTCN2019125430-appb-000004
表示判别器将自生成器的伪样本标签对识别出来,其中G(x,z)表示生成器产生的数据分布;
In the above formula, (C, G, D) respectively represent the classifier, discriminator and generator, (x, y, z) respectively represent the MRI image, PET image and diagnostic label information; α ∈ (0, 1) is a Constant, used to control the proportion of classifier and generator loss in the training process, E (x,y,z)~p(x,y,z) [log D(x,y,z)] represents the discriminator The samples from the real data distribution are judged as real samples;
Figure PCTCN2019125430-appb-000002
Figure PCTCN2019125430-appb-000003
Indicates that the discriminator has identified a pair of pseudo samples in the output data space of the classifier;
Figure PCTCN2019125430-appb-000004
Indicates that the discriminator recognizes the pseudo sample label pair from the generator, where G(x,z) represents the data distribution generated by the generator;
步骤b2:对分类器引入监督学习下的交叉熵损失K c=E (x,y,z)~p(x,y,z)[-log p c(x,y,z)],将分类器的样本分布p c(x,y,z)的收敛点限定在p(x,y)附近,使模型的全局最优满足生成器G和分类器C产生的样本分布与真实数据分布相同; Step b2: Introduce the cross-entropy loss K c =E (x,y,z)~p(x,y,z) [-log p c (x,y,z)] under supervised learning to the classifier, and classify The convergence point of the sample distribution p c (x,y,z) of the generator is limited to p(x,y), so that the global optimum of the model satisfies the sample distribution generated by the generator G and the classifier C is the same as the real data distribution;
步骤b3:引入生成器监督损失,利用目标图像与生成图像之间的梯度互信息作为相似性度量:Step b3: Introduce the generator supervision loss, and use the gradient mutual information between the target image and the generated image as a similarity measure:
K g=NI(A,B)=G(A,B)·I(A,B) K g =NI(A,B)=G(A,B)·I(A,B)
Figure PCTCN2019125430-appb-000005
Figure PCTCN2019125430-appb-000005
Figure PCTCN2019125430-appb-000006
Figure PCTCN2019125430-appb-000006
上述公式中,I(A,B)和G(A,B)分别表示生成图像与目标图像之间的梯度信息和梯度差值。In the above formula, I (A, B) and G (A, B) respectively represent the gradient information and gradient difference between the generated image and the target image.
本申请实施例采取的技术方案还包括:在所述步骤b中,所述根据受试者的MRI影像、PET影像和诊断标签信息对所述多任务生成对抗网络进行训练还包括:所述生成器以MRI影像作为条件约束,将与目标图像同样维度的随机噪声输入映射为PET影像,所述判别器判定输入的样本分布(x,y,z)来自于真实数据分布还是伪数据分布,所述分类器以MRI影像和PET影像的联合分布作为输入,并预测其标签类型;在数据驱动模式下,随着生成器的逐渐优化,判别器更新网络参数以识别出生成器产生的伪数据分布;随着判别器的优化激励分类器优化使其预测的疾病分类预测标签趋向于真实数据而不会被判别器判定为伪数据,继而反向作用于生成器的训练;通过迭代对抗训练,使得所述生成器学习到MRI影像与PET影像之间的潜在关联特征,从而由输入的MRI影像合成得到相应的PET影像,并使得所述分类器从输入的MRI影像和PET影像提取关键特征信息并预测对应的疾病分类预测标签。The technical solution adopted in the embodiment of the application further includes: in the step b, the training the multi-task generation confrontation network according to the subject's MRI image, PET image, and diagnostic tag information also includes: the generation The MRI image is used as a conditional constraint, and the random noise input of the same dimension as the target image is mapped to the PET image. The discriminator determines whether the input sample distribution (x, y, z) comes from the real data distribution or the pseudo data distribution, so The classifier takes the joint distribution of MRI and PET images as input and predicts its label type; in the data-driven mode, with the gradual optimization of the generator, the discriminator updates the network parameters to identify the pseudo data distribution generated by the generator ; With the optimization of the discriminator, the classifier is optimized to make the predicted disease classification and prediction label tend to be real data without being judged as fake data by the discriminator, and then reversely affect the training of the generator; through iterative adversarial training, The generator learns the potential correlation features between the MRI image and the PET image, thereby synthesizing the corresponding PET image from the input MRI image, and makes the classifier extract the key feature information from the input MRI image and the PET image. Predict the corresponding disease classification prediction label.
本申请实施例采取的技术方案还包括:所述生成器采用U-Net网络结构,其包括网络结构对称的编码器和解码器;在所述步骤c中,所述生成器根据MRI影像合成对应的PET影像具体包括:通过编码器多层卷积的特征提取运算,输出MRI影像的特征图;所述解码器对编码器输出的特征图进行多层反卷积运算,并将产生的特征图与编码器对应位置相同大小的特征图进行多次拼接操作,最终输出目标重构图像,即为合成的PET影像。The technical solution adopted in the embodiment of the application further includes: the generator adopts a U-Net network structure, which includes an encoder and a decoder with a symmetrical network structure; in the step c, the generator synthesizes corresponding MRI images The PET image specifically includes: outputting the feature map of the MRI image through the feature extraction operation of the encoder multi-layer convolution; the decoder performs the multi-layer deconvolution operation on the feature map output by the encoder, and generates the feature map The feature map of the same size as the corresponding position of the encoder undergoes multiple splicing operations, and finally the target reconstructed image is output, which is the synthesized PET image.
本申请实施例采取的技术方案还包括:在所述步骤c中,所述分类器将待检测者的MRI影像与合成的PET影像进入融合后输出待检测者的疾病分类预测标签具体包括:采用特征提取网络提取MRI影像的特征值,并对合成的PET影像进行卷积运算,提取PET影像的特征值;将所述MRI影像和PET影像的特征值进行拼接,组成拼接后的特征值,由全连接层对拼接后的特征值进行融合 和高维抽象;将融合后的特征信息经过Softmax函数运算得到对应的疾病分类预测标签。The technical solution adopted in the embodiment of the present application further includes: in the step c, the classifier fused the MRI image of the subject to be detected and the synthesized PET image and then output the disease classification prediction label of the subject to be detected specifically includes: The feature extraction network extracts the feature value of the MRI image, performs convolution operation on the synthesized PET image, and extracts the feature value of the PET image; splicing the feature value of the MRI image and the PET image to form the spliced feature value. The fully connected layer performs fusion and high-dimensional abstraction on the spliced feature values; the fused feature information is operated by the Softmax function to obtain the corresponding disease classification prediction label.
本申请实施例采取的另一技术方案为:一种多模态三维医学影像融合系统,包括:Another technical solution adopted by the embodiment of the present application is: a multi-modal three-dimensional medical image fusion system, including:
模型构建模块:用于构建多任务生成对抗网络,所述多任务生成对抗网络包括生成器、判别器和分类器;Model building module: used to construct a multi-task generative confrontation network, which includes a generator, a discriminator, and a classifier;
模型训练模块:用于根据受试者的MRI影像、PET影像和诊断标签信息对所述多任务生成对抗网络进行训练,使所述多任务生成对抗网络自动学习MRI影像和PET影像之间的关联特征;Model training module: used to train the multi-task generation confrontation network according to the subject's MRI images, PET images and diagnostic tag information, so that the multi-task generation confrontation network can automatically learn the association between MRI images and PET images feature;
模型应用模块:用于将待检测者的MRI影像输入训练好的多任务生成对抗网络,所述生成器根据MRI影像合成对应的PET影像,并将待检测者的MRI影像与合成的PET影像输入分类器,所述分类器将待检测者的MRI影像与合成的PET影像进入融合后输出待检测者的疾病分类预测标签。Model application module: used to input the MRI image of the person to be inspected into a trained multi-task generation confrontation network. The generator synthesizes the corresponding PET image according to the MRI image, and inputs the MRI image of the person to be inspected and the synthesized PET image A classifier, which merges the MRI image of the person to be detected and the synthesized PET image and outputs the disease classification prediction label of the person to be detected.
本申请实施例采取的技术方案还包括:所述模型训练模块包括:The technical solution adopted in the embodiment of the application further includes: the model training module includes:
损失函数构建单元:用于构建所述多任务生成对抗网络的对抗损失函数;训练过程的对抗损失用改进的极大极小代价函数进行表示:Loss function construction unit: used to construct the confrontation loss function of the multi-task generation confrontation network; the confrontation loss in the training process is represented by an improved maximum-minimum cost function:
Figure PCTCN2019125430-appb-000007
Figure PCTCN2019125430-appb-000007
上述公式中,(C,G,D)分别表示分类器、判别器和生成器,(x,y,z)分别表示MRI影像、PET影像和诊断标签信息;α∈(0,1)是一个常量,用于控制分类器和生成器损失在训练过程中所占比重,E (x,y,z)~p(x,y,z)[log D(x,y,z)]表示判别器 将来自于真实数据分布中的样本判定为真实样本;
Figure PCTCN2019125430-appb-000008
Figure PCTCN2019125430-appb-000009
表示判别器识别出有分类器输出数据空间中的伪样本对;
Figure PCTCN2019125430-appb-000010
表示判别器将自生成器的伪样本标签对识别出来,其中x表示受试样本的MRI模态影像,z表示样本标签,G(x,z)表示条件生成网络合成的PET模态影像。
In the above formula, (C, G, D) respectively represent the classifier, discriminator and generator, (x, y, z) respectively represent the MRI image, PET image and diagnostic label information; α ∈ (0, 1) is a Constant, used to control the proportion of classifier and generator loss in the training process, E (x,y,z)~p(x,y,z) [log D(x,y,z)] represents the discriminator The samples from the real data distribution are judged as real samples;
Figure PCTCN2019125430-appb-000008
Figure PCTCN2019125430-appb-000009
Indicates that the discriminator has identified a pair of pseudo samples in the output data space of the classifier;
Figure PCTCN2019125430-appb-000010
It means that the discriminator recognizes the pseudo-sample label pair from the generator, where x represents the MRI modal image of the sample, z represents the sample label, and G(x,z) represents the PET modal image synthesized by the conditional generation network.
分类器优化单元:用于对分类器引入监督学习下的交叉熵损失K c=E (x,y,z)~p(x,y,z)[-log p c(x,y,z)],将分类器的样本分布p c(x,y,z)的收敛点限定在p(x,y)附近,使模型的全局最优满足生成器G和分类器C产生的样本分布与真实数据分布相同; Classifier optimization unit: used to introduce the cross entropy loss under supervised learning to the classifier K c =E (x,y,z)~p(x,y,z) [-log p c (x,y,z) ], the convergent point of the sample distribution p c (x,y,z) of the classifier is limited to p(x,y), so that the global optimum of the model satisfies the sample distribution generated by generator G and classifier C and the real The data distribution is the same;
生成器优化单元:用于引入生成器监督损失,利用目标图像与生成图像之间的梯度互信息作为相似性度量:Generator optimization unit: used to introduce generator supervision loss, using the gradient mutual information between the target image and the generated image as a similarity measure:
K g=NI(A,B)=G(A,B)·I(A,B) K g =NI(A,B)=G(A,B)·I(A,B)
Figure PCTCN2019125430-appb-000011
Figure PCTCN2019125430-appb-000011
Figure PCTCN2019125430-appb-000012
Figure PCTCN2019125430-appb-000012
上述公式中,I(A,B)和G(A,B)分别表示生成图像与目标图像之间的梯度信息和梯度差值。In the above formula, I (A, B) and G (A, B) respectively represent the gradient information and gradient difference between the generated image and the target image.
本申请实施例采取的技术方案还包括:所述模型训练模块对多任务生成对抗网络进行训练具体为:所述生成器以MRI影像作为条件约束,将与目标图像同样维度的随机噪声输入映射为PET影像,所述判别器判定输入的样本分布(x,y,z)来自于真实数据分布还是伪数据分布,所述分类器以MRI影像和PET影像的联合分布作为输入,并预测其标签类型;在数据驱动模式下,随着生成器的逐渐优化,判别器更新网络参数以识别出生成器产生的伪数据分布;随着 判别器的优化激励分类器优化使其预测的疾病分类预测标签趋向于真实数据而不会被判别器判定为伪数据,继而反向作用于生成器的训练;通过迭代对抗训练,使得所述生成器学习到MRI影像与PET影像之间的潜在关联特征,从而由输入的MRI影像合成得到相应的PET影像,并使得所述分类器从输入的MRI影像和PET影像提取关键特征信息并预测对应的疾病分类预测标签。The technical solution adopted in the embodiment of the application further includes: the model training module trains the multi-task generation confrontation network specifically: the generator uses the MRI image as a conditional constraint, and maps the random noise input of the same dimension as the target image into For PET images, the discriminator determines whether the input sample distribution (x, y, z) comes from a real data distribution or a pseudo data distribution, and the classifier takes the joint distribution of MRI images and PET images as input, and predicts its label type ; In the data-driven mode, with the gradual optimization of the generator, the discriminator updates the network parameters to identify the pseudo data distribution generated by the generator; with the optimization of the discriminator, the classifier is optimized to encourage the classification of the disease to predict the label trend Based on real data without being judged as pseudo data by the discriminator, it then acts in the reverse direction on the training of the generator; through iterative confrontation training, the generator learns the potential correlation characteristics between MRI images and PET images, thereby The input MRI images are synthesized to obtain corresponding PET images, and the classifier is made to extract key feature information from the input MRI images and PET images and predict the corresponding disease classification prediction labels.
本申请实施例采取的技术方案还包括:所述生成器采用U-Net网络结构,其包括网络结构对称的编码器和解码器;所述生成器根据MRI影像合成对应的PET影像具体包括:通过编码器多层卷积的特征提取运算,输出MRI影像的特征图;所述解码器对编码器输出的特征图进行多层反卷积运算,并将产生的特征图与编码器对应位置相同大小的特征图进行多次拼接操作,最终输出目标重构图像,即为合成的PET影像。The technical solution adopted by the embodiment of the application further includes: the generator adopts a U-Net network structure, which includes an encoder and a decoder with a symmetric network structure; the generator synthesizing the corresponding PET image according to the MRI image specifically includes: The feature extraction operation of the encoder multi-layer convolution to output the feature map of the MRI image; the decoder performs the multi-layer deconvolution operation on the feature map output by the encoder, and the generated feature map is the same size as the corresponding position of the encoder Perform multiple stitching operations on the feature map of, and finally output the target reconstructed image, which is the synthesized PET image.
本申请实施例采取的技术方案还包括:所述分类器将待检测者的MRI影像与合成的PET影像进入融合后输出待检测者的疾病分类预测标签具体包括:采用特征提取网络提取MRI影像的特征值,并对合成的PET影像进行卷积运算,提取PET影像的特征值;将所述MRI影像和PET影像的特征值进行拼接,组成拼接后的特征值,由全连接层对拼接后的特征值进行融合和高维抽象;将融合后的特征信息经过Softmax函数运算得到对应的疾病分类预测标签。The technical solution adopted in the embodiment of the present application further includes: the classifier integrates the MRI image of the subject to be detected and the synthesized PET image and then outputs the disease classification prediction label of the subject to be detected. Specifically, it includes: extracting the MRI image by using a feature extraction network. Eigenvalues, and perform convolution operations on the synthesized PET images to extract the eigenvalues of the PET images; splicing the MRI images and the eigenvalues of the PET images to form the spliced characteristic values, and the fully connected layer pairs the spliced Feature values are fused and high-dimensional abstracted; the fused feature information is operated by the Softmax function to obtain the corresponding disease classification prediction label.
本申请实施例采取的又一技术方案为:一种电子设备,包括:Another technical solution adopted by the embodiments of the present application is: an electronic device, including:
至少一个处理器;以及At least one processor; and
与所述至少一个处理器通信连接的存储器;其中,A memory communicatively connected with the at least one processor; wherein,
所述存储器存储有可被所述一个处理器执行的指令,所述指令被所述至少一个处理器执行,以使所述至少一个处理器能够执行上述的多模态三维医学影像融合方法的以下操作:The memory stores instructions executable by the one processor, and the instructions are executed by the at least one processor, so that the at least one processor can execute the following of the above-mentioned multi-modal three-dimensional medical image fusion method operating:
步骤a:构建多任务生成对抗网络,所述多任务生成对抗网络包括生成器、判别器和分类器;Step a: construct a multi-task generative confrontation network, the multi-task generative confrontation network including a generator, a discriminator and a classifier;
步骤b:根据受试者的MRI影像、PET影像和诊断标签信息对所述多任务生成对抗网络进行训练,使所述多任务生成对抗网络自动学习MRI影像和PET影像之间的关联特征;Step b: Training the multi-task generation confrontation network according to the subject's MRI image, PET image and diagnostic tag information, so that the multi-task generation confrontation network automatically learns the associated features between the MRI image and the PET image;
步骤c:将待检测者的MRI影像输入训练好的多任务生成对抗网络,所述生成器根据MRI影像合成对应的PET影像,并将待检测者的MRI影像与合成的PET影像输入分类器,所述分类器将待检测者的MRI影像与合成的PET影像进入融合后输出待检测者的疾病分类预测标签。Step c: Input the MRI image of the person to be inspected into the trained multi-task generating confrontation network, the generator synthesizes the corresponding PET image according to the MRI image, and inputs the MRI image of the person to be inspected and the synthesized PET image into the classifier, The classifier merges the MRI image of the person to be detected and the synthesized PET image and outputs the disease classification prediction label of the person to be detected.
相对于现有技术,本申请实施例产生的有益效果在于:本申请实施例的多模态三维医学影像融合方法、系统及电子设备提出了一种多任务生成对抗模型,根据受试者病灶部位的MRI影像合成得到其在PET影像中的模式图像,并融合真实的MRI影像与合成的PET影像后获取更多用于分类诊断的关键特征,根据关键特征对疾病类型进行分类。相对于现有技术,本申请至少具有以下有益效果:Compared with the prior art, the beneficial effects produced by the embodiments of the present application are: the multi-modal three-dimensional medical image fusion method, system and electronic equipment of the embodiments of the present application propose a multi-task generation confrontation model, which is based on the subject's lesion location Synthesize the pattern image in the PET image from the MRI image, and fuse the real MRI image and the synthesized PET image to obtain more key features for classification and diagnosis, and classify the disease types according to the key features. Compared with the prior art, this application has at least the following beneficial effects:
1、通过设置单独的判别器,其唯一作用是识别数据分布的真伪,解决了传统生成对抗网络在兼顾生成器和分类器性能时可能出现的损失函数收敛点的冲突问题,可以使生成器和分类器同时达到最优。1. By setting up a separate discriminator, its sole function is to identify the authenticity of the data distribution, which solves the conflict problem of the convergence point of the loss function that may occur when the traditional generative confrontation network takes into account the performance of the generator and the classifier, and can make the generator At the same time as the classifier achieve the optimal.
2、可实现跨模态影像合成模型和多模态融合分类模型的一步式协同训练,可以实现更优的训练效果。训练好的生成模型学习到MRI与PET成像的关联特征,可由待检测者的MRI合成其相应的PET,分类模型融合MRI与合成PET的特征信息进行疾病类型的分类诊断,避免了PET采集高昂成本和辐射暴露风险的同时,有效融合了功能性成像特征,可以实现更高的分类精度。2. It can realize the one-step collaborative training of cross-modal image synthesis model and multi-modal fusion classification model, which can achieve better training effects. The trained generative model learns the associated features of MRI and PET imaging, and can synthesize the corresponding PET from the MRI of the examinee. The classification model fuses the characteristic information of MRI and synthetic PET to classify and diagnose diseases, avoiding the high cost of PET collection At the same time as the risk of radiation exposure, functional imaging features are effectively integrated, which can achieve higher classification accuracy.
3、本申请考虑了MRI、PET和诊断标签三种属性的联合分布,模型可以提取到多模态成像和分类诊断之间更丰富的关联特征信息,提升图像生成误差和分类诊断性能。通过大量病例的累积训练,逐步提高预测模型的准确率和鲁棒性。3. This application considers the joint distribution of the three attributes of MRI, PET, and diagnostic tags. The model can extract richer correlation feature information between multimodal imaging and classification diagnosis, and improve image generation errors and classification diagnosis performance. Through the cumulative training of a large number of cases, the accuracy and robustness of the prediction model are gradually improved.
4、本发明所提出的多任务生成对抗网络也可用于其他协同优化的应用场景。4. The multi-task generative confrontation network proposed by the present invention can also be used in other collaborative optimization application scenarios.
附图说明Description of the drawings
图1是本申请实施例的多模态三维医学影像融合方法的流程图;FIG. 1 is a flowchart of a multi-modal three-dimensional medical image fusion method according to an embodiment of the present application;
图2是多任务生成对抗网络总体框架图;Figure 2 is the overall framework diagram of the multi-task generation confrontation network;
图3是生成器的网络结构示意图;Figure 3 is a schematic diagram of the network structure of the generator;
图4是分类器的网络结构示意图;Figure 4 is a schematic diagram of the network structure of the classifier;
图5是多任务生成对抗网络的应用流程图;Figure 5 is an application flow chart of a multi-task generating confrontation network;
图6是本申请实施例的多模态三维医学影像融合系统的结构示意图;6 is a schematic structural diagram of a multi-modal three-dimensional medical image fusion system according to an embodiment of the present application;
图7是本申请实施例提供的多模态三维医学影像融合方法的硬件设备结构示意图。FIG. 7 is a schematic diagram of the hardware device structure of the multi-modal three-dimensional medical image fusion method provided by an embodiment of the present application.
具体实施方式detailed description
为了使本申请的目的、技术方案及优点更加清楚明白,以下结合附图及实施例,对本申请进行进一步详细说明。应当理解,此处所描述的具体实施例仅用以解释本申请,并不用于限定本申请。In order to make the purpose, technical solutions, and advantages of this application clearer, the following further describes this application in detail with reference to the accompanying drawings and embodiments. It should be understood that the specific embodiments described herein are only used to explain the application, and not used to limit the application.
针对多模态影像在临床诊断中的互补性和PET采集过程中的高成本及辐射暴露风险,本申请实施例的多模态三维医学影像融合方法提出了一种多任务生成对抗模型(Multi-Task GAN,MT-GAN),根据受试者病灶部位的MRI影 像预测出其在PET成像中的模式图像,在数据驱动模式下,实现跨模态影像合成网络和多模态融合分类网络的对抗协同训练,优化后的系统成功地学习到了MRI成像、PET成像和疾病诊断之间的潜在关联特征,解决了传统生成对抗模型中生成网络和判别网络收敛点的冲突问题。本申请在无需待检测者进行PET采集的情况下融合了多源医学影像特征,可以更加精准地辅助医生进行临床诊断。为了清楚说明本申请的具体实施方案,以下实施例基于MRI和PET影像的阿尔茨海默症为例进行阐释,但本申请的应用范围并不仅限于病种阿尔茨海默症和MRI-PET影像,也可以广泛应用于其他疾病的CT-PET、MRI-CT等其他模态影像。In view of the complementarity of multi-modal images in clinical diagnosis and the high cost and radiation exposure risks in the PET acquisition process, the multi-modal three-dimensional medical image fusion method in the embodiments of the present application proposes a multi-task generation confrontation model (Multi- Task GAN, MT-GAN), predict the pattern image in PET imaging based on the MRI image of the subject's lesion, and realize the confrontation between the cross-modal image synthesis network and the multi-modal fusion classification network in the data-driven mode Co-training, the optimized system successfully learned the potential correlation features between MRI imaging, PET imaging and disease diagnosis, and solved the conflict problem between the generation network and the discrimination network convergence point in the traditional generative confrontation model. This application integrates multi-source medical imaging features without the need for PET collection by the examinee, which can more accurately assist doctors in clinical diagnosis. In order to clearly illustrate the specific implementation of this application, the following examples are based on MRI and PET imaging of Alzheimer’s disease as an example, but the scope of application of this application is not limited to diseases of Alzheimer’s disease and MRI-PET imaging. It can also be widely used in CT-PET, MRI-CT and other modal images of other diseases.
请参阅图1,是本申请实施例的多模态三维医学影像融合方法的流程图。本申请实施例的多模态三维医学影像融合方法包括以下步骤:Please refer to FIG. 1, which is a flowchart of a multi-modal three-dimensional medical image fusion method according to an embodiment of the present application. The multi-modal three-dimensional medical image fusion method of the embodiment of the present application includes the following steps:
步骤100:采集受试者的MRI影像和PET影像,并对采集的MRI影像和PET影像进行预处理,得到用于训练模型的数据集;Step 100: Collect MRI images and PET images of the subject, and preprocess the collected MRI images and PET images to obtain a data set for training the model;
步骤100中,采集MRI影像和PET影像具体为:分别选择阿尔茨海默症(AD)待检测者、轻度认知障碍(MCI)待检测者和正常老年人(Normal)各300位作为受试者,采集其脑部的MRI影像和PET影像作为原始数据集,并在每位受试者的临床观察和诊断中,由专业医师给出诊断信息,将诊断信息作为每位受试者的诊断标签信息。In step 100, the acquisition of MRI images and PET images is specifically: selecting Alzheimer's disease (AD) to be tested, mild cognitive impairment (MCI) to be tested, and normal elderly (Normal) as the recipients. The examinee collects the MRI image and PET image of his brain as the original data set, and in the clinical observation and diagnosis of each subject, a professional physician will give the diagnosis information, and use the diagnosis information as the Diagnostic label information.
原始数据集预处理具体为:采用FSL、SPM等技术对采集的脑部MRI和PET进行冗余组织剔除和图像校正处理,并利用FSL脑图像处理工具对MRI和PET影像进行线性配准操作,使MRI和PET影像在诊断意义上的解剖点达到空间位置的一致。The preprocessing of the original data set is specifically: using FSL, SPM and other technologies to perform redundant tissue removal and image correction processing on the collected brain MRI and PET, and use FSL brain image processing tools to perform linear registration operations on MRI and PET images. Make the anatomical points of MRI and PET images in the diagnostic sense reach the same spatial position.
步骤200:构建多任务生成对抗网络;Step 200: Construct a multi-task generating confrontation network;
步骤200中,多任务生成对抗网络框架如图2所示。多任务生成对抗网络需要考虑每个待检测者的MRI影像、PET影像和诊断标签信息三种属性,其包括分类器C、生成器G和判别器D,生成器G用于通过真实的MRI影像合成对应的PET影像;判别器D用于判定数据模式分布是来自于真实数据还是伪样本分布;分类器C用于将MRI影像和合成的PET影像进行融合后输出待检测者的疾病分类预测标签。将MRI影像、PET影像和诊断标签信息分别标记为(x,y,z),则在多任务生成对抗网络中共包含真实数据分布p true(x,y,z)、生成器的样本分布p G(x,y g,z)和分类器的样本分布p c(x,y,z l)三种数据分布。生成器以MRI影像作为条件约束,将与目标图像同样维度的随机噪声输入映射为PET影像,实现p G(x,y g,z)=p true(x,y,z)的条件特征映射;分类器则以MRI影像和PET影像的联合分布作为输入,预测其标签类型,实现p c(x,y,z l)=p true(x,y,z)的条件特征映射;判别器则根据输入的样本分布(x,y,z)判定其是否来自于真实数据分布,实质是一个二分类问题。 In step 200, the multi-task generation confrontation network framework is shown in FIG. 2. The multi-task generation confrontation network needs to consider the three attributes of MRI image, PET image and diagnostic label information of each subject to be detected. It includes classifier C, generator G, and discriminator D. Generator G is used to pass real MRI images Synthesize the corresponding PET images; the discriminator D is used to determine whether the data pattern distribution comes from the real data or the pseudo sample distribution; the classifier C is used to fuse the MRI image and the synthesized PET image to output the disease classification prediction label of the subject . Mark the MRI image, PET image, and diagnostic label information as (x, y, z), then the real data distribution p true (x, y, z) and the generator sample distribution p G are included in the multi-task generation confrontation network (x,y g ,z) and the sample distribution of the classifier p c (x,y,z l ) three data distributions. The generator takes the MRI image as a conditional constraint, and maps the random noise input of the same dimension as the target image to the PET image to realize the conditional feature mapping of p G (x, y g , z) = p true (x, y, z); The classifier takes the joint distribution of MRI and PET images as input, predicts its label type, and realizes the conditional feature mapping of p c (x, y, z l ) = p true (x, y, z); the discriminator is based on The input sample distribution (x, y, z) determines whether it comes from the real data distribution, which is essentially a binary classification problem.
本申请实施例中,生成器的网络结构如图3所示。生成器采用U-Net网络结构,U-Net模型基于跳跃式连接的全卷积网络而设计,其主要思路是设计网络结构对称的编码器和解码器,使其具有相同数量和大小的特征图,并通过跳跃式连接将编码器和解码器的对应特征图相结合,可以最大限度保留降采样过程中的特征信息,从而提高特征表达的效率。MRI影像和PET影像来自于同一样本,它们之间共享大量的初级特征信息,因此U-net模型很适合用于两种模态影像之间的复杂特征映射。In the embodiment of this application, the network structure of the generator is shown in FIG. 3. The generator adopts the U-Net network structure, and the U-Net model is designed based on a jump-connected full convolutional network. The main idea is to design an encoder and decoder with a symmetric network structure to have the same number and size of feature maps , And combine the corresponding feature maps of the encoder and the decoder through skip connection, which can retain the feature information in the down-sampling process to the maximum extent, thereby improving the efficiency of feature expression. MRI images and PET images come from the same sample and share a large amount of primary feature information between them. Therefore, the U-net model is very suitable for complex feature mapping between the two modal images.
基于上述U-Net网络结构,生成器通过真实的MRI影像样本合成对应的PET影像的方式具体为:Based on the above-mentioned U-Net network structure, the generator synthesizes the corresponding PET image with real MRI image samples as follows:
(1)通过编码器提取MRI影像的特征信息;以样本128×128×128大小的MRI影像作为输入,经过64个2×2×2大小卷积核的特征提取运算,在三个维度上滑动步长为2,输出64个大小为64×64×64的特征图。然后利用128个2×2×2大小的卷积核对其进行卷积运算,产生128个32×32×32大小的特征图。以此类推,依次经过编码器6层卷积的特征提取运算,输出1024个特征图。(1) Extract the feature information of the MRI image through the encoder; take the sample 128×128×128 size MRI image as input, after 64 2×2×2 size convolution kernels feature extraction operations, slide in three dimensions The step size is 2, and 64 feature maps with a size of 64×64×64 are output. Then 128 2×2×2 size convolution kernels are used to perform convolution operation on it, and 128 32×32×32 size feature maps are generated. By analogy, after the encoder's 6-layer convolution feature extraction operation, 1024 feature maps are output.
(2)解码器对编码器输出的特征图进行重构;首先,对编码器输出的1024个特征图进行反卷积运算,产生512个2×2×2的特征图,并与编码器对应位置相同大小的特征图进行拼接。依次经过6层反卷积运算和拼接操作,最终输出128×128×128大小的目标重构图像,即为合成的PET影像。(2) The decoder reconstructs the feature map output by the encoder; first, the 1024 feature maps output by the encoder are deconvolved to generate 512 2×2×2 feature maps, which correspond to the encoder Feature maps of the same size at the same location are stitched together. After 6 layers of deconvolution operation and splicing operation in turn, the final output is the target reconstructed image of 128×128×128 size, which is the synthesized PET image.
为了降低模型的运算复杂度、提高网络的协同训练效率,本申请实施例中的分类器选择了一种相对简单的多模态融合分类网络,其结构如图4所示。分类器对多模态影像数据的处理流程为:In order to reduce the computational complexity of the model and improve the efficiency of collaborative training of the network, the classifier in the embodiment of the present application selects a relatively simple multi-modal fusion classification network, and its structure is shown in FIG. 4. The processing flow of the multi-modal image data by the classifier is:
(1)采用特征提取网络提取MRI影像的特征信息;首先利用2×2×2大小卷积核的两个卷积层提取影像的初级特征产生32个特征图,再利用一层窗口大小为2×2×2的池化层对特征图进行降维。随后采用3×3×3大小的卷积核提取高级特征,第三和第四个卷积层分别采用64个卷积核,提取到的特征池化降维,之后采用128个卷积核进行更高维特征提取。(1) The feature extraction network is used to extract the feature information of the MRI image; first, the two convolutional layers of the 2×2×2 size convolution kernel are used to extract the primary features of the image to generate 32 feature maps, and then a layer of window size is 2 The ×2×2 pooling layer reduces the dimensionality of the feature map. Subsequently, a 3×3×3 size convolution kernel is used to extract advanced features. The third and fourth convolution layers use 64 convolution kernels respectively, and the extracted features are pooled and reduced in dimension, and then 128 convolution kernels are used for Higher-dimensional feature extraction.
(2)采用同样结构的特征提取网络对PET影像进行卷积运算,产生128个特征值。(2) Using the same structure of feature extraction network to perform convolution operation on PET images to generate 128 feature values.
(3)将从MRI影像和PET影像提取到的特征值进行拼接,组成256个特征值,由包含54个节点的全连接层对两种模态的特征信息进行融合和高维抽象。(3) The feature values extracted from the MRI image and the PET image are spliced together to form 256 feature values, and the feature information of the two modalities is fused and high-dimensional abstracted by a fully connected layer containing 54 nodes.
(4)将整合后的融合特征信息经过Softmax函数运算得到对应的标签预测类型(即预测影像数据对应疾病等级的概率)。(4) The integrated fusion feature information is operated by the Softmax function to obtain the corresponding label prediction type (that is, the probability of the predicted image data corresponding to the disease level).
步骤300:根据受试者的MRI影像、PET影像和诊断标签信息对多任务生成对抗网络进行训练;Step 300: Train the multi-task generation confrontation network according to the subject's MRI image, PET image and diagnostic tag information;
步骤300中,多任务生成对抗网络的训练过程包括以下步骤:In step 300, the training process of the multi-task generating confrontation network includes the following steps:
步骤301:构建模型的对抗损失函数;Step 301: Construct the model's anti-loss function;
在模型的实际应用过程中,需对每个待检测者的MRI数据进行采集,因此生成器和分类器的预测过程分别为如下条件分布:In the actual application process of the model, it is necessary to collect the MRI data of each person to be tested, so the prediction process of the generator and the classifier are respectively the following conditional distributions:
p g(x,y g)=p(y|x)p(x)  (1) p g (x,y g )=p(y|x)p(x) (1)
p c(x,y,z)=p[z|(x,y)]p(x,y)  (2) p c (x,y,z)=p[z|(x,y)]p(x,y) (2)
训练过程的对抗损失可以用改进的极大极小代价函数进行表示:The adversarial loss of the training process can be represented by an improved minimax cost function:
Figure PCTCN2019125430-appb-000013
Figure PCTCN2019125430-appb-000013
公式(3)中,α∈(0,1)是一个常量,用于控制分类器和生成器损失在训练过程中所占比重,即在对抗训练任务中的相对重要性。E (x,y,z)~p(x,y,z)[log D(x,y,z)]表示判别器将来自于真实数据分布中的样本判定为真实样本;
Figure PCTCN2019125430-appb-000014
表示判别器识别出有分类器输出数据空间中的伪样本对;
Figure PCTCN2019125430-appb-000015
表示判别器将自生成器的伪样本标签对识别出来,其中x表示受试样本的MRI模态影像,z表示样本标签,G(x,z)表示条件生成网络合成的PET模态影像。由此构建了多任务生成对抗网络的对抗损失函数。
In formula (3), α∈(0,1) is a constant, which is used to control the proportion of classifier and generator loss in the training process, that is, the relative importance in the confrontation training task. E (x,y,z)~p(x,y,z) [log D(x,y,z)] means that the discriminator judges the samples from the real data distribution as real samples;
Figure PCTCN2019125430-appb-000014
Indicates that the discriminator has identified a pair of pseudo samples in the output data space of the classifier;
Figure PCTCN2019125430-appb-000015
It means that the discriminator recognizes the pseudo-sample label pair from the generator, where x represents the MRI modal image of the sample, z represents the sample label, and G(x,z) represents the PET modal image synthesized by the conditional generation network. As a result, the multi-task generating confrontation network's confrontation loss function is constructed.
步骤302:引入分类器监督损失;由一般对抗生成网络的优化原理可知, 模型当且仅当p(x,y,z)=(1-α)p g(x,y g,z)+αp c(x,y,z)时达到纳什均衡。对抗博弈的均衡表明,当生成器G和分类器C中的其中一个达到最优时,另一个也趋近于最优。但事实上,模型的全局最优应当满足生成器G和分类器C产生的样本分布与真实数据分布相同,即p(x,y,z)=p g(x,y,z)=p c(x,y,z)。但上述损失函数的解是p(x,y,z)=(1-α)p g(x,y,z)+αp c(x,y,z)的子集,无法保证p(x,y,z)=p g(x,y,z)=p c(x,y,z)。因此,本申请通过在训练中对分类器引入监督学习下的交叉熵损失K c=E (x,y,z)~p(x,y,z)[-logp c(x,y,z)],从而将p c(x,y,z)的收敛点限定在p(x,y)附近,继而保证损失函数的解是全局最优解。 Step 302: Introduce the classifier supervision loss; from the optimization principle of the general confrontation generation network, it can be known that the model is only if p(x,y,z)=(1-α)p g (x,y g ,z)+αp Nash equilibrium is reached at c (x, y, z). The equilibrium of the adversarial game shows that when one of the generator G and the classifier C reaches the optimum, the other also approaches the optimum. But in fact, the global optimum of the model should satisfy that the sample distributions generated by generator G and classifier C are the same as the real data distribution, that is, p(x,y,z)=p g (x,y,z)=p c (x,y,z). But the solution of the above loss function is a subset of p(x,y,z)=(1-α)p g (x,y,z)+αp c (x,y,z), which cannot guarantee p(x, y, z) = p g (x, y, z) = p c (x, y, z). Therefore, this application introduces the cross-entropy loss K c =E (x,y,z)~p(x,y,z) [-logp c (x,y,z) under supervised learning to the classifier in training ], thus limiting the convergence point of p c (x, y, z) to near p (x, y), and then ensuring that the solution of the loss function is the global optimal solution.
步骤303:引入生成器监督损失;Step 303: Introduce generator supervision loss;
在生成器训练中,从损失函数设计上除了需要生成样本让判别器难以识别,还要保证生成样本与目标图像尽可能相似。本申请利用目标图像与生成图像之间的梯度互信息作为相似性度量:In generator training, in addition to the need to generate samples to make the discriminator difficult to identify from the loss function design, it is also necessary to ensure that the generated samples are as similar as possible to the target image. This application uses the gradient mutual information between the target image and the generated image as a similarity measure:
K g=NI(A,B)=G(A,B)·I(A,B)  (4) K g =NI(A,B)=G(A,B)·I(A,B) (4)
Figure PCTCN2019125430-appb-000016
Figure PCTCN2019125430-appb-000016
Figure PCTCN2019125430-appb-000017
Figure PCTCN2019125430-appb-000017
上述公式中,I(A,B)和G(A,B)分别表示生成图像与目标图像之间的梯度信息和梯度差值。In the above formula, I (A, B) and G (A, B) respectively represent the gradient information and gradient difference between the generated image and the target image.
综上所述,本申请所提出的多任务生成对抗网络的目标函数为:In summary, the objective function of the multi-task generative confrontation network proposed in this application is:
Figure PCTCN2019125430-appb-000018
Figure PCTCN2019125430-appb-000018
步骤304:将900个受试者的数据集划分训练集和测试集,通过训练集对多任务生成对抗网络进行训练,并通过测试集对多任务生成对抗网络的性能进 行测试;Step 304: Divide the data set of 900 subjects into a training set and a test set, train the multi-task generation confrontation network through the training set, and test the performance of the multi-task generation confrontation network through the test set;
步骤304中,训练集中的样本数据为700个,测试集中样本数据为200个。模型的训练过程具体为:在数据驱动模式下,随着生成器G的逐渐优化,判别器D需更新网络参数以识别出生成器G产生的伪数据分布;随着判别器D的优化激励分类器C优化使其预测的疾病分类预测标签趋向于真实数据而不会被判别器D判定为伪数据,继而反向作用于生成器G的训练。通过如此对多任务生成对抗网络进行迭代训练,使得生成器G和分类器C在对抗训练中达到最优,且在三者的对抗博弈过程中,使得分类器和生成器取得比单独训练更好的性能。In step 304, there are 700 sample data in the training set, and 200 sample data in the test set. The specific training process of the model is as follows: In the data-driven mode, as the generator G is gradually optimized, the discriminator D needs to update the network parameters to identify the pseudo data distribution generated by the generator G; as the discriminator D optimizes the incentive classification The device C is optimized so that the predicted disease classification prediction label tends to be real data without being judged as fake data by the discriminator D, and then acts on the training of the generator G in reverse. Through the iterative training of the multi-task generation confrontation network in this way, the generator G and the classifier C are optimized in the confrontation training, and in the process of the three confrontation games, the classifier and the generator are better than the independent training. Performance.
步骤400:将待检测者的MRI影像输入训练好的多任务生成对抗网络,多任务生成对抗网络输出待检测者的疾病分类预测标签;Step 400: Input the MRI image of the person to be detected into the trained multi-task generating confrontation network, and the multi-task generating confrontation network outputs the disease classification prediction label of the person to be detected;
步骤400中,通过对抗训练后,生成器G学习到了MRI影像与PET影像之间的潜在关联特征,可以更加准确地由输入的MRI影像合成得到相应的PET影像。分类器的参数也实现最优化,可以从输入的MRI影像和PET影像提取关键特征信息并基于该特征预测对应的疾病分类预测标签。In step 400, after the confrontation training, the generator G learns the potential correlation features between the MRI image and the PET image, and can more accurately synthesize the corresponding PET image from the input MRI image. The parameters of the classifier are also optimized, and key feature information can be extracted from the input MRI images and PET images, and the corresponding disease classification prediction labels can be predicted based on the features.
具体请一并参阅图5,多任务生成对抗网络的应用流程具体包括以下步骤:Please refer to Figure 5 for details. The application process of the multi-task generating confrontation network specifically includes the following steps:
步骤401:采集待检测者的MRI影像;Step 401: Collect MRI images of the person to be tested;
步骤402:将MRI影像输入训练好的生成器中进行合成,生成器根据MRI影像合成对应的PET影像;Step 402: Input the MRI image into the trained generator for synthesis, and the generator synthesizes the corresponding PET image according to the MRI image;
步骤403:将MRI影像与合成的PET影像输入到训练好的分类器中,分类器输出待检测者的疾病分类预测标签。Step 403: Input the MRI image and the synthesized PET image into the trained classifier, and the classifier outputs the disease classification prediction label of the person to be detected.
请参阅图6,是本申请实施例的多模态三维医学影像融合系统的结构示意图。本申请实施例的多模态三维医学影像融合系统包括数据采集模块、模型构 建模块、模型训练模块和模型应用模块。Please refer to FIG. 6, which is a schematic structural diagram of a multi-modal three-dimensional medical image fusion system according to an embodiment of the present application. The multi-modal three-dimensional medical image fusion system of the embodiment of the present application includes a data acquisition module, a model construction module, a model training module, and a model application module.
数据采集模块:用于采集受试者的MRI影像和PET影像,并对采集的MRI影像和PET影像进行预处理,得到用于训练模型的数据集;其中,采集MRI影像和PET影像具体为:分别选择阿尔茨海默症(AD)待检测者、轻度认知障碍(MCI)待检测者和正常老年人(Normal)各300位作为受试者,采集其脑部的MRI影像和PET影像作为原始数据集,并在每位受试者的临床观察和诊断中,由专业医师给出诊断信息,将诊断信息作为每位受试者的诊断标签信息。Data acquisition module: used to acquire MRI images and PET images of subjects, and preprocess the acquired MRI images and PET images to obtain a data set for training the model; among them, the acquisition of MRI images and PET images is specifically: Select 300 people each for Alzheimer's disease (AD) to be tested, mild cognitive impairment (MCI) to be tested, and normal elderly (Normal) as subjects, and collect MRI and PET images of their brains As the original data set, and in the clinical observation and diagnosis of each subject, professional physicians will give diagnostic information, and use the diagnostic information as the diagnostic label information of each subject.
MRI影像和PET影像预处理具体为:采用FSL、SPM等技术对采集的脑部MRI和PET进行冗余组织剔除和图像校正处理,并利用FSL脑图像处理工具对MRI和PET影像进行线性配准操作,使MRI和PET影像在诊断意义上的解剖点达到空间位置的一致。The preprocessing of MRI and PET images is specifically: using FSL, SPM and other technologies to perform redundant tissue removal and image correction processing on the collected brain MRI and PET, and use FSL brain image processing tools to linearly register MRI and PET images Operation to make the anatomical points of MRI and PET images in the diagnostic sense reach the same spatial position.
模型构建模块:用于构建多任务生成对抗网络;其中,多任务生成对抗网络包括分类器C、生成器G和判别器D,生成器G用于通过真实的MRI影像合成对应的PET影像;判别器D用于判定数据模式分布是来自于真实数据还是伪样本分布;分类器C用于将MRI影像和合成的PET影像进行融合后输出待检测者的疾病分类预测标签。将MRI影像、PET影像和诊断标签信息分别标记为(x,y,z),则在多任务生成对抗网络中共包含真实数据分布p true(x,y,z)、生成器的样本分布p G(x,y g,z)和分类器的样本分布p c(x,y,z l)三种数据分布。生成器以MRI影像作为条件约束,将与目标图像同样维度的随机噪声输入映射为PET影像,实现p G(x,y g,z)=p true(x,y,z)的条件特征映射;分类器则以MRI影像和PET影像的联合分布作为输入,预测其标签类型,实现p c(x,y,z l)=p true(x,y,z) 的条件特征映射;判别器则根据输入的样本分布(x,y,z)判定其是否来自于真实数据分布,实质是一个二分类问题。 Model building module: used to build a multi-task generative confrontation network; among them, the multi-task generative confrontation network includes a classifier C, a generator G, and a discriminator D. The generator G is used to synthesize corresponding PET images from real MRI images; The device D is used to determine whether the data pattern distribution comes from real data or a pseudo sample distribution; the classifier C is used to fuse the MRI image and the synthesized PET image to output the disease classification prediction label of the subject to be detected. Mark the MRI image, PET image, and diagnostic label information as (x, y, z), then the real data distribution p true (x, y, z) and the generator sample distribution p G are included in the multi-task generation confrontation network (x,y g ,z) and the sample distribution of the classifier p c (x,y,z l ) three data distributions. The generator takes the MRI image as a conditional constraint, and maps the random noise input of the same dimension as the target image to the PET image to realize the conditional feature mapping of p G (x, y g , z) = p true (x, y, z); The classifier takes the joint distribution of MRI images and PET images as input, predicts its label type, and realizes the conditional feature mapping of p c (x,y,z l )=p true (x,y,z); the discriminator is based on The input sample distribution (x, y, z) determines whether it comes from the real data distribution, which is essentially a binary classification problem.
本申请实施例中,生成器采用U-Net网络结构,U-Net模型基于跳跃式连接的全卷积网络而设计,其主要思路是设计网络结构对称的编码器和解码器,使其具有相同数量和大小的特征图,并通过跳跃式连接将编码器和解码器的对应特征图相结合,可以最大限度保留降采样过程中的特征信息,从而提高特征表达的效率。MRI影像和PET影像来自于同一样本,它们之间共享大量的初级特征信息,因此U-net模型很适合用于两种模态影像之间的复杂特征映射。In the embodiment of this application, the generator adopts the U-Net network structure. The U-Net model is designed based on a jump-connected full convolutional network. The main idea is to design an encoder and a decoder with a symmetric network structure to have the same The number and size of feature maps, and the corresponding feature maps of the encoder and decoder are combined through skip connection, which can retain the feature information in the downsampling process to the maximum extent, thereby improving the efficiency of feature expression. MRI images and PET images come from the same sample and share a large amount of primary feature information between them. Therefore, the U-net model is very suitable for complex feature mapping between the two modal images.
基于上述U-Net网络结构,生成器通过真实的MRI影像样本合成对应的PET影像的方式具体为:Based on the above-mentioned U-Net network structure, the generator synthesizes the corresponding PET image with real MRI image samples as follows:
(1)通过编码器提取MRI影像的特征信息;以样本128×128×128大小的MRI影像作为输入,经过64个2×2×2大小卷积核的特征提取运算,在三个维度上滑动步长为2,输出64个大小为64×64×64的特征图。然后利用128个2×2×2大小的卷积核对其进行卷积运算,产生128个32×32×32大小的特征图。以此类推,依次经过编码器6层卷积的特征提取运算,输出1024个特征图。(1) Extract the feature information of the MRI image through the encoder; take the sample 128×128×128 size MRI image as input, after 64 2×2×2 size convolution kernels feature extraction operations, slide in three dimensions The step size is 2, and 64 feature maps with a size of 64×64×64 are output. Then 128 2×2×2 size convolution kernels are used to perform convolution operation on it, and 128 32×32×32 size feature maps are generated. By analogy, after the encoder's 6-layer convolution feature extraction operation, 1024 feature maps are output.
(2)解码器对编码器输出的特征图进行重构;首先,对编码器输出的1024个特征图进行反卷积运算,产生512个2×2×2的特征图,并与编码器对应位置相同大小的特征图进行拼接。依次经过6层反卷积运算和拼接操作,最终输出128×128×128大小的目标重构图像,即为合成的PET影像。(2) The decoder reconstructs the feature map output by the encoder; first, the 1024 feature maps output by the encoder are deconvolved to generate 512 2×2×2 feature maps, which correspond to the encoder Feature maps of the same size at the same location are stitched together. After 6 layers of deconvolution operation and splicing operation in turn, the final output is the target reconstructed image of 128×128×128 size, which is the synthesized PET image.
为了降低模型的运算复杂度、提高网络的协同训练效率,本申请实施例中的分类器选择了一种相对简单的多模态融合分类网络,分类器对多模态影像数据的处理流程为:In order to reduce the computational complexity of the model and improve the efficiency of collaborative training of the network, the classifier in this embodiment of the application selects a relatively simple multi-modal fusion classification network, and the classifier processes the multi-modal image data as follows:
(1)提取MRI影像的特征信息;首先利用2×2×2大小卷积核的两个卷积层提取影像的初级特征产生32个特征图,再利用一层窗口大小为2×2×2的池化层对特征图进行降维。随后采用3×3×3大小的卷积核提取高级特征,第三和第四个卷积层分别采用64个卷积核,提取到的特征池化降维,之后采用128个卷积核进行更高维特征提取。(1) Extract the feature information of the MRI image; first use the two convolutional layers of the 2×2×2 size convolution kernel to extract the primary features of the image to generate 32 feature maps, and then use a layer of window size of 2×2×2 The pooling layer reduces the dimensionality of the feature map. Subsequently, a 3×3×3 size convolution kernel is used to extract advanced features. The third and fourth convolution layers use 64 convolution kernels respectively, and the extracted features are pooled and reduced in dimension, and then 128 convolution kernels are used for Higher-dimensional feature extraction.
(2)采用同样结构的特征提取网络对PET影像进行卷积运算,产生128个特征值。(2) Using the same structure of feature extraction network to perform convolution operation on PET images to generate 128 feature values.
(3)将从MRI影像和PET影像提取到的特征值进行拼接,组成256个特征值,由包含54个节点的全连接层对两种模态的特征信息进行融合和高维抽象。(3) The feature values extracted from the MRI image and the PET image are spliced together to form 256 feature values, and the feature information of the two modalities is fused and high-dimensional abstracted by a fully connected layer containing 54 nodes.
(4)将整合后的融合特征信息经过Softmax函数运算得到对应的标签预测类型(即预测影像数据对应疾病等级的概率)。(4) The integrated fusion feature information is operated by the Softmax function to obtain the corresponding label prediction type (that is, the probability of the predicted image data corresponding to the disease level).
模型训练模块:用于根据受试者的MRI影像、PET影像和诊断标签信息对多任务生成对抗网络进行训练;模型训练模块包括:Model training module: used to train the multi-task generation confrontation network based on the subject’s MRI images, PET images and diagnostic tag information; the model training module includes:
损失函数构建单元:用于构建模型的对抗损失函数;在模型的实际应用过程中,需对每个待检测者的MRI数据进行采集,因此生成器和分类器的预测过程分别为如下条件分布:Loss function construction unit: used to construct the model's adversarial loss function; in the actual application process of the model, the MRI data of each person to be tested needs to be collected, so the prediction process of the generator and the classifier are respectively the following conditional distributions:
p g(x,y g)=p(y|x)p(x)  (1) p g (x,y g )=p(y|x)p(x) (1)
p c(x,y,z)=p[z|(x,y)]p(x,y)  (2) p c (x,y,z)=p[z|(x,y)]p(x,y) (2)
训练过程的对抗损失可以用改进的极大极小代价函数进行表示:The adversarial loss of the training process can be represented by an improved minimax cost function:
Figure PCTCN2019125430-appb-000019
Figure PCTCN2019125430-appb-000019
公式(3)中,α∈(0,1)是一个常量,用于控制分类器和生成器损失在训 练过程中所占比重,即在对抗训练任务中的相对重要性。E (x,y,z)~p(x,y,z)[log D(x,y,z)]表示判别器将来自于真实数据分布中的样本判定为真实样本;
Figure PCTCN2019125430-appb-000020
表示判别器识别出有分类器输出数据空间中的伪样本对;
Figure PCTCN2019125430-appb-000021
由此构建了多任务生成对抗网络的对抗损失函数。
In formula (3), α∈(0,1) is a constant, which is used to control the proportion of classifier and generator loss in the training process, that is, the relative importance in the confrontation training task. E (x,y,z)~p(x,y,z) [log D(x,y,z)] means that the discriminator judges the samples from the real data distribution as real samples;
Figure PCTCN2019125430-appb-000020
Indicates that the discriminator has identified a pair of pseudo samples in the output data space of the classifier;
Figure PCTCN2019125430-appb-000021
As a result, the multi-task generating confrontation network's confrontation loss function is constructed.
分类器优化单元:用于引入分类器监督损失;由一般对抗生成网络的优化原理可知,模型当且仅当p(x,y,z)=(1-α)p g(x,y g,z)+αp c(x,y,z)时达到纳什均衡。对抗博弈的均衡表明,当生成器G和分类器C中的其中一个达到最优时,另一个也趋近于最优。但事实上,模型的全局最优应当满足生成器G和分类器C产生的样本分布与真实数据分布相同,即p(x,y,z)=p g(x,y,z)=p c(x,y,z)。但上述损失函数的解是p(x,y,z)=(1-α)p g(x,y,z)+αp c(x,y,z)的子集,无法保证p(x,y,z)=p g(x,y,z)=p c(x,y,z)。因此,本申请通过在训练中对分类器引入监督学习下的交叉熵损失K c=E (x,y,z)~p(x,y,z)[-log p c(x,y,z)],从而将p c(x,y,z)的收敛点限定在p(x,y)附近,继而保证损失函数的解是全局最优解。 Classifier optimization unit: used to introduce classifier supervision loss; it can be known from the optimization principle of general confrontation generation network that the model is and only if p(x,y,z)=(1-α)p g (x,y g , z)+αp c (x, y, z) reaches the Nash equilibrium. The equilibrium of the adversarial game shows that when one of the generator G and the classifier C reaches the optimum, the other also approaches the optimum. But in fact, the global optimum of the model should satisfy that the sample distributions generated by generator G and classifier C are the same as the real data distribution, that is, p(x,y,z)=p g (x,y,z)=p c (x,y,z). But the solution of the above loss function is a subset of p(x,y,z)=(1-α)p g (x,y,z)+αp c (x,y,z), which cannot guarantee p(x, y, z) = p g (x, y, z) = p c (x, y, z). Therefore, this application introduces the cross-entropy loss K c =E (x,y,z)~p(x,y,z) [-log p c (x,y,z ) under supervised learning to the classifier in training )], thus limiting the convergence point of p c (x,y,z) near p(x,y), and then ensuring that the solution of the loss function is the global optimal solution.
生成器优化单元:用于引入生成器监督损失;在生成器训练中,从损失函数设计上除了需要生成样本让判别器难以识别,还要保证生成样本与目标图像尽可能相似。本申请利用目标图像与生成图像之间的梯度互信息作为相似性度量:Generator optimization unit: used to introduce generator supervision loss; in generator training, in addition to the need to generate samples to make the discriminator difficult to identify from the loss function design, it is also necessary to ensure that the generated samples are as similar as possible to the target image. This application uses the gradient mutual information between the target image and the generated image as a similarity measure:
K g=NI(A,B)=G(A,B)·I(A,B)  (4) K g =NI(A,B)=G(A,B)·I(A,B) (4)
Figure PCTCN2019125430-appb-000022
Figure PCTCN2019125430-appb-000022
Figure PCTCN2019125430-appb-000023
Figure PCTCN2019125430-appb-000023
上述公式中,I(A,B)和G(A,B)分别表示生成图像与目标图像之间的梯度信 息和梯度差值。In the above formula, I(A,B) and G(A,B) respectively represent the gradient information and gradient difference between the generated image and the target image.
综上所述,本申请所提出的多任务生成对抗网络的目标函数为:In summary, the objective function of the multi-task generative confrontation network proposed in this application is:
Figure PCTCN2019125430-appb-000024
Figure PCTCN2019125430-appb-000024
模型训练单元:用于将900个受试者的数据集划分训练集和测试集,通过训练集对多任务生成对抗网络进行训练,并通过测试集对多任务生成对抗网络的性能进行测试;其中,训练集中的样本数据为700个,测试集中样本数据为200个。模型的训练过程具体为:在数据驱动模式下,随着生成器G的逐渐优化,判别器D需更新网络参数以识别出生成器G产生的伪数据分布;随着判别器D的优化激励分类器C优化使其预测的疾病分类预测标签趋向于真实数据而不会被判别器D判定为伪数据,继而反向作用于生成器G的训练。通过如此对多任务生成对抗网络进行迭代训练,使得生成器G和分类器C在对抗训练中达到最优,且在三者的对抗博弈过程中,使得分类器和生成器取得比单独训练更好的性能。Model training unit: used to divide the data set of 900 subjects into a training set and a test set, train the multi-task generating confrontation network through the training set, and test the performance of the multi-task generating confrontation network through the test set; , The sample data in the training set is 700, and the sample data in the test set is 200. The specific training process of the model is as follows: In the data-driven mode, as the generator G is gradually optimized, the discriminator D needs to update the network parameters to identify the pseudo data distribution generated by the generator G; as the discriminator D optimizes the incentive classification The device C is optimized so that the predicted disease classification prediction label tends to be real data without being judged as fake data by the discriminator D, and then acts on the training of the generator G in reverse. Through the iterative training of the multi-task generation confrontation network in this way, the generator G and the classifier C are optimized in the confrontation training, and in the process of the three confrontation games, the classifier and the generator are better than the independent training. Performance.
模型应用模块:用于将待检测者的MRI影像输入训练好的多任务生成对抗网络,多任务生成对抗网络输出待检测者的疾病分类预测标签;通过对抗训练后,生成器G学习到了MRI影像与PET影像之间的潜在关联特征,可以更加准确地由输入的MRI影像合成得到相应的PET影像。分类器的参数也实现最优化,可以从输入的MRI影像和PET影像提取关键特征信息并基于该特征预测对应的疾病分类预测标签。Model application module: used to input the MRI image of the person to be detected into the trained multi-task generation confrontation network, and the multi-task generation confrontation network outputs the disease classification prediction label of the person to be detected; after the confrontation training, the generator G learns the MRI image The potential associated features with PET images can be more accurately synthesized from the input MRI images to obtain the corresponding PET images. The parameters of the classifier are also optimized, and key feature information can be extracted from the input MRI images and PET images, and the corresponding disease classification prediction labels can be predicted based on the features.
具体的,多任务生成对抗网络的应用过程具体为:采集待检测者的MRI影像,将MRI影像输入训练好的生成器中进行合成,生成器根据MRI影像合 成对应的PET影像;将MRI影像与合成的PET影像输入到训练好的分类器中,分类器输出待检测者的疾病分类预测标签。Specifically, the application process of the multi-task generation confrontation network is specifically: collecting the MRI image of the person to be detected, inputting the MRI image into the trained generator for synthesis, and the generator synthesizing the corresponding PET image according to the MRI image; combining the MRI image with The synthesized PET image is input to the trained classifier, and the classifier outputs the disease classification prediction label of the person to be detected.
图7是本申请实施例提供的多模态三维医学影像融合方法的硬件设备结构示意图。如图7所示,该设备包括一个或多个处理器以及存储器。以一个处理器为例,该设备还可以包括:输入系统和输出系统。FIG. 7 is a schematic diagram of the hardware device structure of the multi-modal three-dimensional medical image fusion method provided by an embodiment of the present application. As shown in Figure 7, the device includes one or more processors and memory. Taking a processor as an example, the device may also include: an input system and an output system.
处理器、存储器、输入系统和输出系统可以通过总线或者其他方式连接,图7中以通过总线连接为例。The processor, the memory, the input system, and the output system may be connected by a bus or in other ways. In FIG. 7, the connection by a bus is taken as an example.
存储器作为一种非暂态计算机可读存储介质,可用于存储非暂态软件程序、非暂态计算机可执行程序以及模块。处理器通过运行存储在存储器中的非暂态软件程序、指令以及模块,从而执行电子设备的各种功能应用以及数据处理,即实现上述方法实施例的处理方法。As a non-transitory computer-readable storage medium, the memory can be used to store non-transitory software programs, non-transitory computer executable programs, and modules. The processor executes various functional applications and data processing of the electronic device by running non-transitory software programs, instructions, and modules stored in the memory, that is, realizing the processing methods of the foregoing method embodiments.
存储器可以包括存储程序区和存储数据区,其中,存储程序区可存储操作系统、至少一个功能所需要的应用程序;存储数据区可存储数据等。此外,存储器可以包括高速随机存取存储器,还可以包括非暂态存储器,例如至少一个磁盘存储器件、闪存器件、或其他非暂态固态存储器件。在一些实施例中,存储器可选包括相对于处理器远程设置的存储器,这些远程存储器可以通过网络连接至处理系统。上述网络的实例包括但不限于互联网、企业内部网、局域网、移动通信网及其组合。The memory may include a program storage area and a data storage area, where the program storage area can store an operating system and an application program required by at least one function; the data storage area can store data and the like. In addition, the memory may include a high-speed random access memory, and may also include a non-transitory memory, such as at least one magnetic disk storage device, a flash memory device, or other non-transitory solid state storage devices. In some embodiments, the storage may optionally include storage remotely arranged with respect to the processor, and these remote storages may be connected to the processing system through a network. Examples of the aforementioned networks include, but are not limited to, the Internet, corporate intranets, local area networks, mobile communication networks, and combinations thereof.
输入系统可接收输入的数字或字符信息,以及产生信号输入。输出系统可包括显示屏等显示设备。The input system can receive input digital or character information, and generate signal input. The output system may include display devices such as a display screen.
所述一个或者多个模块存储在所述存储器中,当被所述一个或者多个处理器执行时,执行上述任一方法实施例的以下操作:The one or more modules are stored in the memory, and when executed by the one or more processors, the following operations of any of the foregoing method embodiments are performed:
步骤a:构建多任务生成对抗网络,所述多任务生成对抗网络包括生成器、判别器和分类器;Step a: construct a multi-task generative confrontation network, the multi-task generative confrontation network including a generator, a discriminator and a classifier;
步骤b:根据受试者的MRI影像、PET影像和诊断标签信息对所述多任务 生成对抗网络进行训练,使所述多任务生成对抗网络自动学习MRI影像和PET影像之间的关联特征;Step b: Training the multi-task generating confrontation network according to the subject's MRI image, PET image and diagnostic tag information, so that the multi-task generating confrontation network automatically learns the associated features between the MRI image and the PET image;
步骤c:将待检测者的MRI影像输入训练好的多任务生成对抗网络,所述生成器根据MRI影像合成对应的PET影像,并将待检测者的MRI影像与合成的PET影像输入分类器,所述分类器将待检测者的MRI影像与合成的PET影像进入融合后输出待检测者的疾病分类预测标签。Step c: Input the MRI image of the person to be inspected into the trained multi-task generating confrontation network, the generator synthesizes the corresponding PET image according to the MRI image, and inputs the MRI image of the person to be inspected and the synthesized PET image into the classifier, The classifier merges the MRI image of the person to be detected and the synthesized PET image and outputs the disease classification prediction label of the person to be detected.
上述产品可执行本申请实施例所提供的方法,具备执行方法相应的功能模块和有益效果。未在本实施例中详尽描述的技术细节,可参见本申请实施例提供的方法。The above-mentioned products can execute the methods provided in the embodiments of the present application, and have functional modules and beneficial effects corresponding to the execution methods. For technical details not described in detail in this embodiment, please refer to the method provided in the embodiment of this application.
本申请实施例提供了一种非暂态(非易失性)计算机存储介质,所述计算机存储介质存储有计算机可执行指令,该计算机可执行指令可执行以下操作:The embodiments of the present application provide a non-transitory (non-volatile) computer storage medium, the computer storage medium stores computer executable instructions, and the computer executable instructions can perform the following operations:
步骤a:构建多任务生成对抗网络,所述多任务生成对抗网络包括生成器、判别器和分类器;Step a: construct a multi-task generative confrontation network, the multi-task generative confrontation network including a generator, a discriminator and a classifier;
步骤b:根据受试者的MRI影像、PET影像和诊断标签信息对所述多任务生成对抗网络进行训练,使所述多任务生成对抗网络自动学习MRI影像和PET影像之间的关联特征;Step b: Training the multi-task generation confrontation network according to the subject's MRI image, PET image and diagnostic tag information, so that the multi-task generation confrontation network automatically learns the associated features between the MRI image and the PET image;
步骤c:将待检测者的MRI影像输入训练好的多任务生成对抗网络,所述生成器根据MRI影像合成对应的PET影像,并将待检测者的MRI影像与合成的PET影像输入分类器,所述分类器将待检测者的MRI影像与合成的PET影像进入融合后输出待检测者的疾病分类预测标签。Step c: Input the MRI image of the person to be inspected into the trained multi-task generating confrontation network, the generator synthesizes the corresponding PET image according to the MRI image, and inputs the MRI image of the person to be inspected and the synthesized PET image into the classifier, The classifier merges the MRI image of the person to be detected and the synthesized PET image and outputs the disease classification prediction label of the person to be detected.
本申请实施例提供了一种计算机程序产品,所述计算机程序产品包括存储在非暂态计算机可读存储介质上的计算机程序,所述计算机程序包括程序指令,当所述程序指令被计算机执行时,使所述计算机执行以下操作:The embodiment of the present application provides a computer program product, the computer program product includes a computer program stored on a non-transitory computer-readable storage medium, the computer program includes program instructions, when the program instructions are executed by a computer To make the computer do the following:
步骤a:构建多任务生成对抗网络,所述多任务生成对抗网络包括生成器、判别器和分类器;Step a: construct a multi-task generative confrontation network, the multi-task generative confrontation network including a generator, a discriminator and a classifier;
步骤b:根据受试者的MRI影像、PET影像和诊断标签信息对所述多任务生成对抗网络进行训练,使所述多任务生成对抗网络自动学习MRI影像和PET影像之间的关联特征;Step b: Training the multi-task generation confrontation network according to the subject's MRI image, PET image and diagnostic tag information, so that the multi-task generation confrontation network automatically learns the associated features between the MRI image and the PET image;
步骤c:将待检测者的MRI影像输入训练好的多任务生成对抗网络,所述生成器根据MRI影像合成对应的PET影像,并将待检测者的MRI影像与合成的PET影像输入分类器,所述分类器将待检测者的MRI影像与合成的PET影像进入融合后输出待检测者的疾病分类预测标签。Step c: Input the MRI image of the person to be inspected into the trained multi-task generating confrontation network, the generator synthesizes the corresponding PET image according to the MRI image, and inputs the MRI image of the person to be inspected and the synthesized PET image into the classifier, The classifier merges the MRI image of the person to be detected and the synthesized PET image and outputs the disease classification prediction label of the person to be detected.
本申请实施例的多模态三维医学影像融合方法、系统及电子设备提出了一种多任务生成对抗模型,根据受试者病灶部位的MRI影像合成得到其在PET影像中的模式图像,并融合真实的MRI影像与合成的PET影像后获取更多用于分类诊断的关键特征,根据关键特征对疾病类型进行分类。相对于现有技术,本申请至少具有以下有益效果:The multi-modal three-dimensional medical image fusion method, system and electronic equipment of the embodiments of the present application propose a multi-task generation confrontation model, which synthesizes the pattern image in the PET image according to the MRI image of the subject’s lesion, and merges it After real MRI images and synthetic PET images, more key features for classification and diagnosis are obtained, and disease types are classified according to the key features. Compared with the prior art, this application has at least the following beneficial effects:
1、通过设置单独的判别器,其唯一作用是识别数据分布的真伪,解决了传统生成对抗网络在兼顾生成器和分类器性能时可能出现的损失函数收敛点的冲突问题,可以使生成器和分类器同时达到最优。1. By setting up a separate discriminator, its sole function is to identify the authenticity of the data distribution, which solves the conflict problem of the convergence point of the loss function that may occur when the traditional generative confrontation network takes into account the performance of the generator and the classifier, and can make the generator At the same time as the classifier achieve the optimal.
2、可实现跨模态影像合成模型和多模态融合分类模型的一步式协同训练,可以实现更优的训练效果。训练好的生成模型学习到MRI与PET成像的关联特征,可由待检测者的MRI合成其相应的PET,分类模型融合MRI与合成PET的特征信息进行疾病类型的分类诊断,避免了PET采集高昂成本和辐射暴露风险的同时,有效融合了功能性成像特征,可以实现更高的分类精度。2. It can realize the one-step collaborative training of cross-modal image synthesis model and multi-modal fusion classification model, which can achieve better training effects. The trained generative model learns the associated features of MRI and PET imaging, and can synthesize the corresponding PET from the MRI of the examinee. The classification model fuses the characteristic information of MRI and synthetic PET to classify and diagnose diseases, avoiding the high cost of PET collection At the same time as the risk of radiation exposure, functional imaging features are effectively integrated, which can achieve higher classification accuracy.
3、本申请考虑了MRI、PET和诊断标签三种属性的联合分布,模型可以 提取到多模态成像和分类诊断之间更丰富的关联特征信息,提升图像生成误差和分类诊断性能。通过大量病例的累积训练,逐步提高预测模型的准确率和鲁棒性。3. This application considers the joint distribution of the three attributes of MRI, PET, and diagnostic tags. The model can extract richer associated feature information between multimodal imaging and classification diagnosis, and improve image generation errors and classification diagnosis performance. Through the cumulative training of a large number of cases, the accuracy and robustness of the prediction model are gradually improved.
4、本发明所提出的多任务生成对抗网络也可用于其他协同优化的应用场景。4. The multi-task generative confrontation network proposed by the present invention can also be used in other collaborative optimization application scenarios.
以上所述仅是本发明的优选实施方式,应当指出,对于本技术领域的普通技术人员来说,在不脱离本发明原理的前提下,还可以做出若干改进和润饰,这些改进和润饰也应视为本发明的保护范围。The above are only the preferred embodiments of the present invention. It should be pointed out that for those of ordinary skill in the art, without departing from the principle of the present invention, several improvements and modifications can be made, and these improvements and modifications are also It should be regarded as the protection scope of the present invention.

Claims (11)

  1. 一种多模态三维医学影像融合方法,其特征在于,包括以下步骤:A multi-modal three-dimensional medical image fusion method is characterized in that it comprises the following steps:
    步骤a:构建多任务生成对抗网络,所述多任务生成对抗网络包括生成器、判别器和分类器;Step a: construct a multi-task generative confrontation network, the multi-task generative confrontation network including a generator, a discriminator and a classifier;
    步骤b:根据受试者的MRI影像、PET影像和诊断标签信息对所述多任务生成对抗网络进行训练,使所述多任务生成对抗网络自动学习MRI影像和PET影像之间的关联特征;Step b: Training the multi-task generation confrontation network according to the subject's MRI image, PET image and diagnostic tag information, so that the multi-task generation confrontation network automatically learns the associated features between the MRI image and the PET image;
    步骤c:将待检测者的MRI影像输入训练好的多任务生成对抗网络,所述生成器根据MRI影像合成对应的PET影像,并将待检测者的MRI影像与合成的PET影像输入分类器,所述分类器将待检测者的MRI影像与合成的PET影像进入融合后输出待检测者的疾病分类预测标签。Step c: Input the MRI image of the person to be inspected into the trained multi-task generating confrontation network, the generator synthesizes the corresponding PET image according to the MRI image, and inputs the MRI image of the person to be inspected and the synthesized PET image into the classifier, The classifier merges the MRI image of the person to be detected and the synthesized PET image and outputs the disease classification prediction label of the person to be detected.
  2. 根据权利要求1所述的多模态三维医学影像融合方法,其特征在于,在所述步骤b中,所述对多任务生成对抗网络进行训练具体包括:The multi-modal three-dimensional medical image fusion method according to claim 1, wherein in the step b, the training of the multi-task generation confrontation network specifically includes:
    步骤b1:构建所述多任务生成对抗网络的对抗损失函数;训练过程的对抗损失用改进的极大极小代价函数进行表示:Step b1: Construct the confrontation loss function of the multi-task generation confrontation network; the confrontation loss in the training process is represented by an improved maximum-minimum cost function:
    Figure PCTCN2019125430-appb-100001
    Figure PCTCN2019125430-appb-100001
    上述公式中,(C,G,D)分别表示分类器、判别器和生成器,(x,y,z)分别表示MRI影像、PET影像和诊断标签信息;α∈(0,1)是一个常量,用于控制分类器和生成器损失在训练过程中所占比重,E (x,y,z)~p(x,y,z)[logD(x,y,z)]表示判别器将来自于真实数据分布中的样本判定为真实样本;
    Figure PCTCN2019125430-appb-100002
    表示判别器识别出有分类器输出数据空间中的伪样本对;
    Figure PCTCN2019125430-appb-100003
    表示判别器将自生成器的伪样本标签对识别出来,其中x表示受试样本的MRI模态影像,z表示样本标签,G(x,z)表示条件生成网络合成的PET模态影像;
    In the above formula, (C, G, D) respectively represent the classifier, discriminator and generator, (x, y, z) respectively represent the MRI image, PET image and diagnostic label information; α ∈ (0, 1) is a Constant, used to control the proportion of classifier and generator loss in the training process, E (x,y,z)~p(x,y,z) [logD(x,y,z)] means that the discriminator will The samples from the real data distribution are judged as real samples;
    Figure PCTCN2019125430-appb-100002
    Indicates that the discriminator has identified a pair of pseudo samples in the output data space of the classifier;
    Figure PCTCN2019125430-appb-100003
    Indicates that the discriminator recognizes the pseudo sample label pair from the generator, where x represents the MRI modal image of the specimen, z represents the sample label, and G(x,z) represents the PET modal image synthesized by the conditional generation network;
    步骤b2:对分类器引入监督学习下的交叉熵损失K c=E (x,y,z)~p(x,y,z)[-log p c(x,y,z)],将分类器的样本分布p c(x,y,z)的收敛点限定在p(x,y)附近,使模型的全局最优满足生成器G和分类器C产生的样本分布与真实数据分布相同; Step b2: Introduce the cross-entropy loss K c =E (x,y,z)~p(x,y,z) [-log p c (x,y,z)] under supervised learning to the classifier, and classify The convergence point of the sample distribution p c (x,y,z) of the generator is limited to p(x,y), so that the global optimum of the model satisfies the sample distribution generated by the generator G and the classifier C is the same as the real data distribution;
    步骤b3:引入生成器监督损失,利用目标图像与生成图像之间的梯度互信息作为相似性度量:Step b3: Introduce the generator supervision loss, and use the gradient mutual information between the target image and the generated image as a similarity measure:
    K g=NI(A,B)=G(A,B)·I(A,B) K g =NI(A,B)=G(A,B)·I(A,B)
    Figure PCTCN2019125430-appb-100004
    Figure PCTCN2019125430-appb-100004
    Figure PCTCN2019125430-appb-100005
    Figure PCTCN2019125430-appb-100005
    上述公式中,I(A,B)和G(A,B)分别表示生成图像与目标图像之间的梯度信息和梯度差值。In the above formula, I (A, B) and G (A, B) respectively represent the gradient information and gradient difference between the generated image and the target image.
  3. 根据权利要求2所述的多模态三维医学影像融合方法,其特征在于,在所述步骤b中,所述根据受试者的MRI影像、PET影像和诊断标签信息对所述多任务生成对抗网络进行训练还包括:所述生成器以MRI影像作为条件约束,将与目标图像同样维度的随机噪声输入映射为PET影像,所述判别器判定输入的样本分布(x,y,z)来自于真实数据分布还是伪数据分布,所述分类器以MRI影像和PET影像的联合分布作为输入,并预测其标签类型;在数据驱动模式下,随着生成器的逐渐优化,判别器更新网络参数以识别出生成器产生的伪数据分布; 随着判别器的优化激励分类器优化使其预测的疾病分类预测标签趋向于真实数据而不会被判别器判定为伪数据,继而反向作用于生成器的训练;通过迭代对抗训练,使得所述生成器学习到MRI影像与PET影像之间的潜在关联特征,从而由输入的MRI影像合成得到相应的PET影像,并使得所述分类器从输入的MRI影像和PET影像提取关键特征信息并预测对应的疾病分类预测标签。The multi-modal three-dimensional medical image fusion method according to claim 2, wherein in the step b, the MRI image, PET image, and diagnostic tag information of the subject are used to generate a confrontation against the multi-task The network training also includes: the generator uses the MRI image as a conditional constraint to map the random noise input of the same dimension as the target image to the PET image, and the discriminator determines that the input sample distribution (x, y, z) comes from Real data distribution or pseudo data distribution. The classifier takes the joint distribution of MRI and PET images as input and predicts its label type; in the data-driven mode, with the gradual optimization of the generator, the discriminator updates the network parameters to Recognize the pseudo data distribution generated by the generator; with the optimization of the discriminator, the classifier is optimized to make the predicted disease classification and prediction label tend to be real data without being judged as pseudo data by the discriminator, and then the generator is reversed Training; through iterative adversarial training, the generator learns the potential correlation features between the MRI image and the PET image, thereby synthesizing the corresponding PET image from the input MRI image, and making the classifier from the input MRI Images and PET images extract key feature information and predict corresponding disease classification prediction labels.
  4. 根据权利要求1至3任一项所述的多模态三维医学影像融合方法,其特征在于,所述生成器采用U-Net网络结构,其包括网络结构对称的编码器和解码器;在所述步骤c中,所述生成器根据MRI影像合成对应的PET影像具体包括:通过编码器多层卷积的特征提取运算,输出MRI影像的特征图;所述解码器对编码器输出的特征图进行多层反卷积运算,并将产生的特征图与编码器对应位置相同大小的特征图进行多次拼接操作,最终输出目标重构图像,即为合成的PET影像。The multi-modal three-dimensional medical image fusion method according to any one of claims 1 to 3, wherein the generator adopts a U-Net network structure, which includes an encoder and a decoder with a symmetric network structure; In the step c, the generator synthesizing the corresponding PET image according to the MRI image specifically includes: outputting the feature map of the MRI image through the feature extraction operation of the encoder multi-layer convolution; the decoder outputting the feature map to the encoder Perform a multi-layer deconvolution operation, and perform multiple stitching operations on the generated feature map and the feature map of the same size at the corresponding position of the encoder, and finally output the target reconstructed image, which is the synthesized PET image.
  5. 根据权利要求4所述的多模态三维医学影像融合方法,其特征在于,在所述步骤c中,所述分类器将待检测者的MRI影像与合成的PET影像进入融合后输出待检测者的疾病分类预测标签具体包括:采用特征提取网络提取MRI影像的特征值,并对合成的PET影像进行卷积运算,提取PET影像的特征值;将所述MRI影像和PET影像的特征值进行拼接,组成拼接后的特征值,由全连接层对拼接后的特征值进行融合和高维抽象;将融合后的特征信息经过Softmax函数运算得到对应的疾病分类预测标签。The multi-modal three-dimensional medical image fusion method according to claim 4, wherein in the step c, the classifier merges the MRI image of the subject to be detected and the synthesized PET image and then outputs the subject to be detected The disease classification prediction label specifically includes: extracting the feature value of the MRI image using a feature extraction network, and performing a convolution operation on the synthesized PET image to extract the feature value of the PET image; and stitching the feature value of the MRI image and the PET image , Compose the spliced feature values, and the fully connected layer will fusion and high-dimensional abstraction of the spliced feature values; the fused feature information is operated by the Softmax function to obtain the corresponding disease classification prediction label.
  6. 一种多模态三维医学影像融合系统,其特征在于,包括:A multi-modal three-dimensional medical image fusion system is characterized by comprising:
    模型构建模块:用于构建多任务生成对抗网络,所述多任务生成对抗网络包括生成器、判别器和分类器;Model building module: used to construct a multi-task generative confrontation network, which includes a generator, a discriminator, and a classifier;
    模型训练模块:用于根据受试者的MRI影像、PET影像和诊断标签信息对所述多任务生成对抗网络进行训练,使所述多任务生成对抗网络自动学习MRI 影像和PET影像之间的关联特征;Model training module: used to train the multi-task generation confrontation network according to the subject’s MRI images, PET images and diagnostic tag information, so that the multi-task generation confrontation network can automatically learn the association between MRI images and PET images feature;
    模型应用模块:用于将待检测者的MRI影像输入训练好的多任务生成对抗网络,所述生成器根据MRI影像合成对应的PET影像,并将待检测者的MRI影像与合成的PET影像输入分类器,所述分类器将待检测者的MRI影像与合成的PET影像进入融合后输出待检测者的疾病分类预测标签。Model application module: used to input the MRI image of the person to be inspected into a trained multi-task generation confrontation network. The generator synthesizes the corresponding PET image according to the MRI image, and inputs the MRI image of the person to be inspected and the synthesized PET image A classifier, which merges the MRI image of the person to be detected and the synthesized PET image and outputs the disease classification prediction label of the person to be detected.
  7. 根据权利要求6所述的多模态三维医学影像融合系统,其特征在于,所述模型训练模块包括:The multi-modal three-dimensional medical image fusion system according to claim 6, wherein the model training module comprises:
    损失函数构建单元:用于构建所述多任务生成对抗网络的对抗损失函数;训练过程的对抗损失用改进的极大极小代价函数进行表示:Loss function construction unit: used to construct the confrontation loss function of the multi-task generation confrontation network; the confrontation loss in the training process is represented by an improved maximum-minimum cost function:
    Figure PCTCN2019125430-appb-100006
    Figure PCTCN2019125430-appb-100006
    上述公式中,(C,G,D)分别表示分类器、判别器和生成器,(x,y,z)分别表示MRI影像、PET影像和诊断标签信息;α∈(0,1)是一个常量,用于控制分类器和生成器损失在训练过程中所占比重,E (x,y,z)~p(x,y,z)[log D(x,y,z)]表示判别器将来自于真实数据分布中的样本判定为真实样本;
    Figure PCTCN2019125430-appb-100007
    表示判别器识别出有分类器输出数据空间中的伪样本对;
    Figure PCTCN2019125430-appb-100008
    表示判别器将自生成器的伪样本标签对识别出来,其中x表示受试样本的MRI模态影像,z表示样本标签,G(x,z)表示条件生成网络合成的PET模态影像;
    In the above formula, (C, G, D) respectively represent the classifier, discriminator and generator, (x, y, z) respectively represent the MRI image, PET image and diagnostic label information; α ∈ (0, 1) is a Constant, used to control the proportion of classifier and generator loss in the training process, E (x,y,z)~p(x,y,z) [log D(x,y,z)] represents the discriminator The samples from the real data distribution are judged as real samples;
    Figure PCTCN2019125430-appb-100007
    Indicates that the discriminator has identified a pair of pseudo samples in the output data space of the classifier;
    Figure PCTCN2019125430-appb-100008
    Indicates that the discriminator recognizes the pseudo sample label pair from the generator, where x represents the MRI modal image of the specimen, z represents the sample label, and G(x,z) represents the PET modal image synthesized by the conditional generation network;
    分类器优化单元:用于对分类器引入监督学习下的交叉熵损失K c=E (x,y,z)~p(x,y,z)[-log p c(x,y,z)],将分类器的样本分布p c(x,y,z)的收敛点限定在p(x,y)附近,使模型的全局最优满足生成器G和分类器C产生的样本分布与真 实数据分布相同; Classifier optimization unit: used to introduce the cross entropy loss under supervised learning to the classifier K c =E (x,y,z)~p(x,y,z) [-log p c (x,y,z) ], the convergent point of the sample distribution p c (x,y,z) of the classifier is limited to p(x,y), so that the global optimum of the model satisfies the sample distribution generated by generator G and classifier C and the real The data distribution is the same;
    生成器优化单元:用于引入生成器监督损失,利用目标图像与生成图像之间的梯度互信息作为相似性度量:Generator optimization unit: used to introduce generator supervision loss, using the gradient mutual information between the target image and the generated image as a similarity measure:
    K g=NI(A,B)=G(A,B)·I(A,B) K g =NI(A,B)=G(A,B)·I(A,B)
    Figure PCTCN2019125430-appb-100009
    Figure PCTCN2019125430-appb-100009
    Figure PCTCN2019125430-appb-100010
    Figure PCTCN2019125430-appb-100010
    上述公式中,I(A,B)和G(A,B)分别表示生成图像与目标图像之间的梯度信息和梯度差值。In the above formula, I (A, B) and G (A, B) respectively represent the gradient information and gradient difference between the generated image and the target image.
  8. 根据权利要求7所述的多模态三维医学影像融合系统,其特征在于,所述模型训练模块对多任务生成对抗网络进行训练具体为:所述生成器以MRI影像作为条件约束,将与目标图像同样维度的随机噪声输入映射为PET影像,所述判别器判定输入的样本分布(x,y,z)来自于真实数据分布还是伪数据分布,所述分类器以MRI影像和PET影像的联合分布作为输入,并预测其标签类型;在数据驱动模式下,随着生成器的逐渐优化,判别器更新网络参数以识别出生成器产生的伪数据分布;随着判别器的优化激励分类器优化使其预测的疾病分类预测标签趋向于真实数据而不会被判别器判定为伪数据,继而反向作用于生成器的训练;通过迭代对抗训练,使得所述生成器学习到MRI影像与PET影像之间的潜在关联特征,从而由输入的MRI影像合成得到相应的PET影像,并使得所述分类器从输入的MRI影像和PET影像提取关键特征信息并预测对应的疾病分类预测标签。The multi-modal three-dimensional medical image fusion system according to claim 7, wherein the training of the multi-task generation confrontation network by the model training module is specifically: the generator takes the MRI image as a conditional constraint, and compares it with the target The random noise input of the same dimension of the image is mapped to the PET image. The discriminator determines whether the input sample distribution (x, y, z) comes from the real data distribution or the pseudo data distribution. The classifier uses a combination of MRI and PET images. The distribution is used as input and its label type is predicted; in the data-driven mode, with the gradual optimization of the generator, the discriminator updates the network parameters to identify the pseudo data distribution generated by the generator; with the optimization of the discriminator, the classifier is optimized to stimulate Make its predicted disease classification and prediction labels tend to be real data without being judged as fake data by the discriminator, and then reversely affect the training of the generator; through iterative confrontation training, the generator can learn MRI images and PET images Therefore, the corresponding PET image is synthesized from the input MRI image, and the classifier extracts key feature information from the input MRI image and PET image and predicts the corresponding disease classification prediction label.
  9. 根据权利要求6至8任一项所述的多模态三维医学影像融合系统,其特征在于,所述生成器采用U-Net网络结构,其包括网络结构对称的编码器和解码器; 所述生成器根据MRI影像合成对应的PET影像具体包括:通过编码器多层卷积的特征提取运算,输出MRI影像的特征图;所述解码器对编码器输出的特征图进行多层反卷积运算,并将产生的特征图与编码器对应位置相同大小的特征图进行多次拼接操作,最终输出目标重构图像,即为合成的PET影像。The multi-modal three-dimensional medical image fusion system according to any one of claims 6 to 8, wherein the generator adopts a U-Net network structure, which includes an encoder and a decoder with a symmetric network structure; The generator synthesizing the corresponding PET image according to the MRI image specifically includes: outputting the feature map of the MRI image through the feature extraction operation of the encoder multi-layer convolution; the decoder performs the multi-layer deconvolution operation on the feature map output by the encoder , And perform multiple stitching operations on the generated feature map and the feature map of the same size at the corresponding position of the encoder, and finally output the target reconstructed image, which is the synthesized PET image.
  10. 根据权利要求9所述的多模态三维医学影像融合系统,其特征在于,所述分类器将待检测者的MRI影像与合成的PET影像进入融合后输出待检测者的疾病分类预测标签具体包括:采用特征提取网络提取MRI影像的特征值,并对合成的PET影像进行卷积运算,提取PET影像的特征值;将所述MRI影像和PET影像的特征值进行拼接,组成拼接后的特征值,由全连接层对拼接后的特征值进行融合和高维抽象;将融合后的特征信息经过Softmax函数运算得到对应的疾病分类预测标签。The multi-modal three-dimensional medical image fusion system according to claim 9, wherein the classifier merges the MRI image of the subject to be detected and the synthesized PET image and then outputs the disease classification prediction label of the subject to be detected. : Use the feature extraction network to extract the feature values of the MRI image, and perform convolution operations on the synthesized PET image to extract the feature value of the PET image; stitch the feature values of the MRI image and the PET image to form the stitched feature value , The fully connected layer fusion and high-dimensional abstraction of the spliced feature values; the fused feature information is operated by the Softmax function to obtain the corresponding disease classification prediction label.
  11. 一种电子设备,包括:An electronic device including:
    至少一个处理器;以及At least one processor; and
    与所述至少一个处理器通信连接的存储器;其中,A memory communicatively connected with the at least one processor; wherein,
    所述存储器存储有可被所述一个处理器执行的指令,所述指令被所述至少一个处理器执行,以使所述至少一个处理器能够执行上述1至5任一项所述的多模态三维医学影像融合方法的以下操作:The memory stores instructions that can be executed by the one processor, and the instructions are executed by the at least one processor, so that the at least one processor can execute the multi-mode operation described in any one of 1 to 5 above. The following operations of the three-dimensional medical image fusion method:
    步骤a:构建多任务生成对抗网络,所述多任务生成对抗网络包括生成器、判别器和分类器;Step a: construct a multi-task generative confrontation network, the multi-task generative confrontation network including a generator, a discriminator and a classifier;
    步骤b:根据受试者的MRI影像、PET影像和诊断标签信息对所述多任务生成对抗网络进行训练,使所述多任务生成对抗网络自动学习MRI影像和PET影像之间的关联特征;Step b: Training the multi-task generation confrontation network according to the subject's MRI image, PET image and diagnostic tag information, so that the multi-task generation confrontation network automatically learns the associated features between the MRI image and the PET image;
    步骤c:将待检测者的MRI影像输入训练好的多任务生成对抗网络,所述 生成器根据MRI影像合成对应的PET影像,并将待检测者的MRI影像与合成的PET影像输入分类器,所述分类器将待检测者的MRI影像与合成的PET影像进入融合后输出待检测者的疾病分类预测标签。Step c: Input the MRI image of the person to be inspected into the trained multi-task generating confrontation network, the generator synthesizes the corresponding PET image according to the MRI image, and inputs the MRI image of the person to be inspected and the synthesized PET image into the classifier, The classifier merges the MRI image of the person to be detected and the synthesized PET image and outputs the disease classification prediction label of the person to be detected.
PCT/CN2019/125430 2019-08-07 2019-12-14 Multimodal three-dimensional medical image fusion method and system, and electronic device WO2021022752A1 (en)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
CN201910727072.9 2019-08-07
CN201910727072.9A CN110580695B (en) 2019-08-07 2019-08-07 Multi-mode three-dimensional medical image fusion method and system and electronic equipment

Publications (1)

Publication Number Publication Date
WO2021022752A1 true WO2021022752A1 (en) 2021-02-11

Family

ID=68810603

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/CN2019/125430 WO2021022752A1 (en) 2019-08-07 2019-12-14 Multimodal three-dimensional medical image fusion method and system, and electronic device

Country Status (2)

Country Link
CN (1) CN110580695B (en)
WO (1) WO2021022752A1 (en)

Cited By (46)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN112966112A (en) * 2021-03-25 2021-06-15 支付宝(杭州)信息技术有限公司 Text classification model training and text classification method and device based on counterstudy
CN113052243A (en) * 2021-03-30 2021-06-29 浙江工业大学 Target detection method based on CycleGAN and condition distribution self-adaption
CN113052930A (en) * 2021-03-12 2021-06-29 北京医准智能科技有限公司 Chest DR dual-energy digital subtraction image generation method
CN113095038A (en) * 2021-05-08 2021-07-09 杭州王道控股有限公司 Font generation method and device for generating countermeasure network based on multitask discriminator
CN113223068A (en) * 2021-05-31 2021-08-06 西安电子科技大学 Multi-modal image registration method and system based on depth global features
CN113326731A (en) * 2021-04-22 2021-08-31 南京大学 Cross-domain pedestrian re-identification algorithm based on momentum network guidance
CN113470777A (en) * 2021-06-04 2021-10-01 江苏大学 Tumor auxiliary diagnosis report generation method and device, electronic equipment and storage medium
CN113488183A (en) * 2021-06-30 2021-10-08 南京云上数融技术有限公司 Multi-modal characteristic fusion cognitive system, equipment and storage medium for fever diseases
CN113496482A (en) * 2021-05-21 2021-10-12 郑州大学 Poison driving test paper image segmentation model, positioning segmentation method and portable device
CN113538533A (en) * 2021-06-22 2021-10-22 南方医科大学 Spine registration method, spine registration device, spine registration equipment and computer storage medium
CN113989474A (en) * 2021-12-09 2022-01-28 北京环境特性研究所 Infrared small target intelligent identification method based on progressive hierarchical fusion network
CN113989697A (en) * 2021-09-24 2022-01-28 天津大学 Short video frequency classification method and device based on multi-mode self-supervision deep countermeasure network
CN114022729A (en) * 2021-10-27 2022-02-08 华中科技大学 Heterogeneous image matching positioning method and system based on twin network and supervised training
CN114036356A (en) * 2021-10-13 2022-02-11 中国科学院信息工程研究所 Unbalanced traffic classification method and system based on confrontation generation network traffic enhancement
CN114038055A (en) * 2021-10-27 2022-02-11 电子科技大学长三角研究院(衢州) Image generation method based on contrast learning and generation countermeasure network
CN114332577A (en) * 2021-12-31 2022-04-12 福州大学 Colorectal cancer image classification method and system combining deep learning and image omics
CN114332287A (en) * 2022-03-11 2022-04-12 之江实验室 Method, device, equipment and medium for reconstructing PET (positron emission tomography) image based on transformer feature sharing
CN114373532A (en) * 2021-12-31 2022-04-19 华南理工大学 Multi-modal medical image translation method for generating countermeasure network based on target perception
CN114387481A (en) * 2021-12-30 2022-04-22 天翼物联科技有限公司 Medical image cross-modal synthesis system and method based on multi-source countermeasure strategy
CN114779661A (en) * 2022-04-22 2022-07-22 北京科技大学 Chemical synthesis robot system based on multi-classification generation confrontation imitation learning algorithm
CN114821206A (en) * 2022-06-30 2022-07-29 山东建筑大学 Multi-modal image fusion classification method and system based on confrontation complementary features
CN114821157A (en) * 2022-04-01 2022-07-29 山东大学 Multi-modal image classification method based on hybrid model network
CN114882252A (en) * 2022-05-20 2022-08-09 中国人民解放军国防科技大学 Semi-supervised remote sensing image change detection method and device and computer equipment
CN114926702A (en) * 2022-04-16 2022-08-19 西北工业大学深圳研究院 Small sample image classification method based on depth attention measurement
CN114972878A (en) * 2022-06-14 2022-08-30 云南大学 Method for detecting migratable confrontation sample independent of attack
CN114974518A (en) * 2022-04-15 2022-08-30 浙江大学 Multi-mode data fusion lung nodule image recognition method and device
CN115187467A (en) * 2022-05-31 2022-10-14 北京昭衍新药研究中心股份有限公司 Enhanced virtual image data generation method based on generation countermeasure network
CN115272261A (en) * 2022-08-05 2022-11-01 广州大学 Multi-modal medical image fusion method based on deep learning
CN115578370A (en) * 2022-10-28 2023-01-06 深圳市铱硙医疗科技有限公司 Metabolism region anomaly detection method and device based on brain images
CN115588487A (en) * 2022-11-07 2023-01-10 重庆邮电大学 Medical image data set making method based on federal learning and generation countermeasure network
US20230031910A1 (en) * 2020-12-09 2023-02-02 Shenzhen Institutes Of Advanced Technology Apriori guidance network for multitask medical image synthesis
CN115859175A (en) * 2023-02-16 2023-03-28 合肥综合性国家科学中心人工智能研究院(安徽省人工智能实验室) Hydraulic shock absorber equipment abnormity detection method based on cross-mode generative learning
CN115984622A (en) * 2023-01-10 2023-04-18 深圳大学 Classification method based on multi-mode and multi-example learning, prediction method and related device
CN116524248A (en) * 2023-04-17 2023-08-01 首都医科大学附属北京友谊医院 Medical data processing device, method and classification model training device
CN116883995A (en) * 2023-07-07 2023-10-13 广东食品药品职业学院 Identification system of breast cancer molecular subtype
CN117174240A (en) * 2023-10-26 2023-12-05 中国科学技术大学 Medical image report generation method based on large model field migration
CN117201693A (en) * 2023-11-01 2023-12-08 长春汽车工业高等专科学校 Internet of things image compression method, device, terminal equipment and medium
CN117315376A (en) * 2023-11-28 2023-12-29 聊城莱柯智能机器人有限公司 Machine learning-based mechanical part industrial quality inspection method
CN117315425A (en) * 2023-10-12 2023-12-29 无锡市第五人民医院 Fusion method and system of multi-mode magnetic resonance images
CN117408330A (en) * 2023-12-14 2024-01-16 合肥高维数据技术有限公司 Federal knowledge distillation method and device for non-independent co-distributed data
CN117540489A (en) * 2023-11-13 2024-02-09 重庆大学 Airfoil pneumatic data calculation method and system based on multitask learning
CN117765292A (en) * 2023-12-26 2024-03-26 哈尔滨理工大学 Incomplete multi-view remote sensing data clustering method based on graph roll manifold regularization pseudo-label guidance
CN117936105A (en) * 2024-03-25 2024-04-26 杭州安鸿科技股份有限公司 Multimode melanoma immunotherapy prediction method based on deep learning network
CN118015551A (en) * 2024-04-09 2024-05-10 山东世融信息科技有限公司 Floating island type monitoring system applied to field ecological wetland
CN118279158A (en) * 2024-06-03 2024-07-02 之江实验室 Quality improvement method and device for magnetic resonance brain image and computer equipment
CN118506132A (en) * 2024-07-17 2024-08-16 成都师范学院 Method and device for multitasking weather proximity prediction against strong convection of network

Families Citing this family (37)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN111090764B (en) * 2019-12-20 2023-06-23 中南大学 Image classification method and device based on multitask learning and graph convolution neural network
CN111353499B (en) * 2020-02-24 2022-08-19 上海交通大学 Multi-modal medical image segmentation method, system, storage medium and electronic device
WO2021179189A1 (en) * 2020-03-11 2021-09-16 深圳先进技术研究院 Visualization method and device for evaluating brain addiction traits, and medium
CN111383217B (en) * 2020-03-11 2023-08-29 深圳先进技术研究院 Visual method, device and medium for brain addiction character evaluation
CN111640510A (en) * 2020-04-09 2020-09-08 之江实验室 Disease prognosis prediction system based on deep semi-supervised multitask learning survival analysis
JP7457292B2 (en) * 2020-06-01 2024-03-28 日本電気株式会社 Brain image analysis device, control method, and program
CN111739635A (en) * 2020-06-10 2020-10-02 四川大学华西医院 Diagnosis auxiliary model for acute ischemic stroke and image processing method
CN111815030B (en) * 2020-06-11 2024-02-06 浙江工商大学 Multi-target feature prediction method based on small amount of questionnaire survey data
CN111783796A (en) * 2020-06-19 2020-10-16 哈尔滨工业大学 PET/CT image recognition system based on depth feature fusion
CN112052874B (en) * 2020-07-31 2023-11-17 山东大学 Physiological data classification method and system based on generation countermeasure network
CN112085687B (en) * 2020-09-10 2023-12-01 浙江大学 Method for converting T1 to STIR image based on detail enhancement
CN112669408A (en) * 2020-11-23 2021-04-16 上海眼控科技股份有限公司 Multi-mode live-action map image generation method and device, computer equipment and medium
CN112614198A (en) * 2020-11-23 2021-04-06 上海眼控科技股份有限公司 Multi-modal edge entity image conversion method and device, computer equipment and medium
CN113516615B (en) * 2020-11-24 2024-03-01 阿里巴巴集团控股有限公司 Sample generation method, system, equipment and storage medium
CN112353381B (en) * 2020-11-24 2022-06-28 杭州冉曼智能科技有限公司 Alzheimer's disease comprehensive diagnosis system based on multi-modal brain image
CN112465058A (en) * 2020-12-07 2021-03-09 中国计量大学 Multi-modal medical image classification method under improved GoogLeNet neural network
CN112508775B (en) * 2020-12-10 2024-08-20 深圳先进技术研究院 MRI-PET image mode conversion method and system based on cyclic generation countermeasure network
WO2022120731A1 (en) * 2020-12-10 2022-06-16 深圳先进技术研究院 Mri-pet image modality conversion method and system based on cyclic generative adversarial network
CN112508175B (en) * 2020-12-10 2024-09-17 深圳先进技术研究院 Multi-task learning generation type countermeasure network generation method and system for low-dose PET reconstruction
CN112700859A (en) * 2020-12-15 2021-04-23 贵州小宝健康科技有限公司 Medical diagnosis assisting method and system based on medical images
WO2022126480A1 (en) * 2020-12-17 2022-06-23 深圳先进技术研究院 High-energy image synthesis method and device based on wasserstein generative adversarial network model
CN112633378B (en) * 2020-12-24 2022-06-28 电子科技大学 Intelligent detection method and system for multi-modal image fetal corpus callosum
CN112699809B (en) * 2020-12-31 2023-08-01 深圳数联天下智能科技有限公司 Vaccinia category identification method, device, computer equipment and storage medium
CN112801297B (en) * 2021-01-20 2021-11-16 哈尔滨工业大学 Machine learning model adversity sample generation method based on conditional variation self-encoder
CN113192605B (en) * 2021-04-13 2024-06-28 复旦大学附属中山医院 Medical image classification method, medical image retrieval method and device
CN113205566A (en) * 2021-04-23 2021-08-03 复旦大学 Abdomen three-dimensional medical image conversion generation method based on deep learning
CN113205567A (en) * 2021-04-25 2021-08-03 复旦大学 Method for synthesizing CT image by MRI image based on deep learning
CN113705662B (en) * 2021-08-26 2024-08-02 中国银联股份有限公司 Collaborative training method, device and computer readable storage medium
CN113724880A (en) * 2021-11-03 2021-11-30 深圳先进技术研究院 Abnormal brain connection prediction system, method and device and readable storage medium
CN114758783B (en) * 2022-03-31 2024-07-16 大连理工大学 Elasticity imaging method, device, computer equipment and storage medium for deep learning
CN114821014B (en) * 2022-05-17 2024-06-21 湖南大学 Multi-mode and countermeasure learning-based multi-task target detection and identification method and device
CN115588504A (en) * 2022-10-28 2023-01-10 大连大学附属中山医院 Monitoring management system based on molecular image imaging technology
WO2024113170A1 (en) * 2022-11-29 2024-06-06 中国科学院深圳先进技术研究院 Cycle generative adversarial network-based medical image cross-modal synthesis method and apparatus
CN116309591B (en) * 2023-05-19 2023-08-25 杭州健培科技有限公司 Medical image 3D key point detection method, model training method and device
CN116433795B (en) * 2023-06-14 2023-08-29 之江实验室 Multi-mode image generation method and device based on countermeasure generation network
CN117115045B (en) * 2023-10-24 2024-01-09 吉林大学 Method for improving medical image data quality based on Internet generation type artificial intelligence
CN117911844A (en) * 2024-03-20 2024-04-19 中国科学院自动化研究所 Multi-mode medical image labeling method and device

Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN109035356A (en) * 2018-07-05 2018-12-18 四川大学 A kind of system and method based on PET pattern imaging
WO2019051227A1 (en) * 2017-09-08 2019-03-14 The General Hospital Corporation System and method for utilizing general-purpose graphics processing units (gpgpu) architecture for medical image processing
CN109523584A (en) * 2018-10-26 2019-03-26 上海联影医疗科技有限公司 Image processing method, device, multi-mode imaging system, storage medium and equipment

Family Cites Families (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20170337682A1 (en) * 2016-05-18 2017-11-23 Siemens Healthcare Gmbh Method and System for Image Registration Using an Intelligent Artificial Agent
JP7295022B2 (en) * 2017-10-03 2023-06-20 株式会社根本杏林堂 Blood vessel extraction device and blood vessel extraction method
CN108198179A (en) * 2018-01-03 2018-06-22 华南理工大学 A kind of CT medical image pulmonary nodule detection methods for generating confrontation network improvement
US10482600B2 (en) * 2018-01-16 2019-11-19 Siemens Healthcare Gmbh Cross-domain image analysis and cross-domain image synthesis using deep image-to-image networks and adversarial networks
CN109961491B (en) * 2019-04-12 2023-05-26 上海联影医疗科技股份有限公司 Multi-mode image truncation compensation method, device, computer equipment and medium

Patent Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2019051227A1 (en) * 2017-09-08 2019-03-14 The General Hospital Corporation System and method for utilizing general-purpose graphics processing units (gpgpu) architecture for medical image processing
CN109035356A (en) * 2018-07-05 2018-12-18 四川大学 A kind of system and method based on PET pattern imaging
CN109523584A (en) * 2018-10-26 2019-03-26 上海联影医疗科技有限公司 Image processing method, device, multi-mode imaging system, storage medium and equipment

Cited By (68)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20230031910A1 (en) * 2020-12-09 2023-02-02 Shenzhen Institutes Of Advanced Technology Apriori guidance network for multitask medical image synthesis
US11915401B2 (en) * 2020-12-09 2024-02-27 Shenzhen Institutes Of Advanced Technology Apriori guidance network for multitask medical image synthesis
CN113052930A (en) * 2021-03-12 2021-06-29 北京医准智能科技有限公司 Chest DR dual-energy digital subtraction image generation method
CN112966112A (en) * 2021-03-25 2021-06-15 支付宝(杭州)信息技术有限公司 Text classification model training and text classification method and device based on counterstudy
CN112966112B (en) * 2021-03-25 2023-08-08 支付宝(杭州)信息技术有限公司 Text classification model training and text classification method and device based on countermeasure learning
CN113052243A (en) * 2021-03-30 2021-06-29 浙江工业大学 Target detection method based on CycleGAN and condition distribution self-adaption
CN113326731A (en) * 2021-04-22 2021-08-31 南京大学 Cross-domain pedestrian re-identification algorithm based on momentum network guidance
CN113326731B (en) * 2021-04-22 2024-04-19 南京大学 Cross-domain pedestrian re-identification method based on momentum network guidance
CN113095038B (en) * 2021-05-08 2024-04-16 杭州王道控股有限公司 Font generation method and device for generating countermeasure network based on multi-task discriminator
CN113095038A (en) * 2021-05-08 2021-07-09 杭州王道控股有限公司 Font generation method and device for generating countermeasure network based on multitask discriminator
CN113496482B (en) * 2021-05-21 2022-10-04 郑州大学 Toxic driving test paper image segmentation model, positioning segmentation method and portable device
CN113496482A (en) * 2021-05-21 2021-10-12 郑州大学 Poison driving test paper image segmentation model, positioning segmentation method and portable device
CN113223068A (en) * 2021-05-31 2021-08-06 西安电子科技大学 Multi-modal image registration method and system based on depth global features
CN113223068B (en) * 2021-05-31 2024-02-02 西安电子科技大学 Multi-mode image registration method and system based on depth global features
CN113470777B (en) * 2021-06-04 2024-04-09 江苏大学 Tumor auxiliary diagnosis report generation method and device, electronic equipment and storage medium
CN113470777A (en) * 2021-06-04 2021-10-01 江苏大学 Tumor auxiliary diagnosis report generation method and device, electronic equipment and storage medium
CN113538533A (en) * 2021-06-22 2021-10-22 南方医科大学 Spine registration method, spine registration device, spine registration equipment and computer storage medium
CN113488183B (en) * 2021-06-30 2023-10-31 吾征智能技术(北京)有限公司 Heating disease multi-mode feature fusion cognitive system, equipment and storage medium
CN113488183A (en) * 2021-06-30 2021-10-08 南京云上数融技术有限公司 Multi-modal characteristic fusion cognitive system, equipment and storage medium for fever diseases
CN113989697A (en) * 2021-09-24 2022-01-28 天津大学 Short video frequency classification method and device based on multi-mode self-supervision deep countermeasure network
CN113989697B (en) * 2021-09-24 2024-06-07 天津大学 Short video classification method and device based on multi-mode self-supervision deep countermeasure network
CN114036356A (en) * 2021-10-13 2022-02-11 中国科学院信息工程研究所 Unbalanced traffic classification method and system based on confrontation generation network traffic enhancement
CN114022729A (en) * 2021-10-27 2022-02-08 华中科技大学 Heterogeneous image matching positioning method and system based on twin network and supervised training
CN114038055A (en) * 2021-10-27 2022-02-11 电子科技大学长三角研究院(衢州) Image generation method based on contrast learning and generation countermeasure network
CN113989474A (en) * 2021-12-09 2022-01-28 北京环境特性研究所 Infrared small target intelligent identification method based on progressive hierarchical fusion network
CN114387481A (en) * 2021-12-30 2022-04-22 天翼物联科技有限公司 Medical image cross-modal synthesis system and method based on multi-source countermeasure strategy
CN114387481B (en) * 2021-12-30 2024-03-29 天翼物联科技有限公司 Medical image cross-mode synthesis system and method based on multisource countermeasure strategy
CN114332577A (en) * 2021-12-31 2022-04-12 福州大学 Colorectal cancer image classification method and system combining deep learning and image omics
CN114373532A (en) * 2021-12-31 2022-04-19 华南理工大学 Multi-modal medical image translation method for generating countermeasure network based on target perception
CN114332287B (en) * 2022-03-11 2022-07-15 之江实验室 Method, device, equipment and medium for reconstructing PET (positron emission tomography) image based on transformer feature sharing
CN114332287A (en) * 2022-03-11 2022-04-12 之江实验室 Method, device, equipment and medium for reconstructing PET (positron emission tomography) image based on transformer feature sharing
CN114821157A (en) * 2022-04-01 2022-07-29 山东大学 Multi-modal image classification method based on hybrid model network
CN114974518A (en) * 2022-04-15 2022-08-30 浙江大学 Multi-mode data fusion lung nodule image recognition method and device
CN114926702B (en) * 2022-04-16 2024-03-19 西北工业大学深圳研究院 Small sample image classification method based on depth attention measurement
CN114926702A (en) * 2022-04-16 2022-08-19 西北工业大学深圳研究院 Small sample image classification method based on depth attention measurement
CN114779661A (en) * 2022-04-22 2022-07-22 北京科技大学 Chemical synthesis robot system based on multi-classification generation confrontation imitation learning algorithm
CN114882252A (en) * 2022-05-20 2022-08-09 中国人民解放军国防科技大学 Semi-supervised remote sensing image change detection method and device and computer equipment
CN115187467A (en) * 2022-05-31 2022-10-14 北京昭衍新药研究中心股份有限公司 Enhanced virtual image data generation method based on generation countermeasure network
CN114972878B (en) * 2022-06-14 2024-05-14 云南大学 Attack-independent movable challenge sample detection method
CN114972878A (en) * 2022-06-14 2022-08-30 云南大学 Method for detecting migratable confrontation sample independent of attack
CN114821206B (en) * 2022-06-30 2022-09-13 山东建筑大学 Multi-modal image fusion classification method and system based on confrontation complementary features
CN114821206A (en) * 2022-06-30 2022-07-29 山东建筑大学 Multi-modal image fusion classification method and system based on confrontation complementary features
CN115272261A (en) * 2022-08-05 2022-11-01 广州大学 Multi-modal medical image fusion method based on deep learning
CN115578370B (en) * 2022-10-28 2023-05-09 深圳市铱硙医疗科技有限公司 Brain image-based metabolic region abnormality detection method and device
CN115578370A (en) * 2022-10-28 2023-01-06 深圳市铱硙医疗科技有限公司 Metabolism region anomaly detection method and device based on brain images
CN115588487A (en) * 2022-11-07 2023-01-10 重庆邮电大学 Medical image data set making method based on federal learning and generation countermeasure network
CN115984622A (en) * 2023-01-10 2023-04-18 深圳大学 Classification method based on multi-mode and multi-example learning, prediction method and related device
CN115984622B (en) * 2023-01-10 2023-12-29 深圳大学 Multi-mode and multi-example learning classification method, prediction method and related device
CN115859175A (en) * 2023-02-16 2023-03-28 合肥综合性国家科学中心人工智能研究院(安徽省人工智能实验室) Hydraulic shock absorber equipment abnormity detection method based on cross-mode generative learning
CN116524248A (en) * 2023-04-17 2023-08-01 首都医科大学附属北京友谊医院 Medical data processing device, method and classification model training device
CN116524248B (en) * 2023-04-17 2024-02-13 首都医科大学附属北京友谊医院 Medical data processing device, method and classification model training device
CN116883995A (en) * 2023-07-07 2023-10-13 广东食品药品职业学院 Identification system of breast cancer molecular subtype
CN117315425A (en) * 2023-10-12 2023-12-29 无锡市第五人民医院 Fusion method and system of multi-mode magnetic resonance images
CN117315425B (en) * 2023-10-12 2024-03-26 无锡市第五人民医院 Fusion method and system of multi-mode magnetic resonance images
CN117174240B (en) * 2023-10-26 2024-02-09 中国科学技术大学 Medical image report generation method based on large model field migration
CN117174240A (en) * 2023-10-26 2023-12-05 中国科学技术大学 Medical image report generation method based on large model field migration
CN117201693A (en) * 2023-11-01 2023-12-08 长春汽车工业高等专科学校 Internet of things image compression method, device, terminal equipment and medium
CN117201693B (en) * 2023-11-01 2024-01-16 长春汽车工业高等专科学校 Internet of things image compression method, device, terminal equipment and medium
CN117540489A (en) * 2023-11-13 2024-02-09 重庆大学 Airfoil pneumatic data calculation method and system based on multitask learning
CN117315376B (en) * 2023-11-28 2024-02-13 聊城莱柯智能机器人有限公司 Machine learning-based mechanical part industrial quality inspection method
CN117315376A (en) * 2023-11-28 2023-12-29 聊城莱柯智能机器人有限公司 Machine learning-based mechanical part industrial quality inspection method
CN117408330B (en) * 2023-12-14 2024-03-15 合肥高维数据技术有限公司 Federal knowledge distillation method and device for non-independent co-distributed data
CN117408330A (en) * 2023-12-14 2024-01-16 合肥高维数据技术有限公司 Federal knowledge distillation method and device for non-independent co-distributed data
CN117765292A (en) * 2023-12-26 2024-03-26 哈尔滨理工大学 Incomplete multi-view remote sensing data clustering method based on graph roll manifold regularization pseudo-label guidance
CN117936105A (en) * 2024-03-25 2024-04-26 杭州安鸿科技股份有限公司 Multimode melanoma immunotherapy prediction method based on deep learning network
CN118015551A (en) * 2024-04-09 2024-05-10 山东世融信息科技有限公司 Floating island type monitoring system applied to field ecological wetland
CN118279158A (en) * 2024-06-03 2024-07-02 之江实验室 Quality improvement method and device for magnetic resonance brain image and computer equipment
CN118506132A (en) * 2024-07-17 2024-08-16 成都师范学院 Method and device for multitasking weather proximity prediction against strong convection of network

Also Published As

Publication number Publication date
CN110580695B (en) 2022-06-21
CN110580695A (en) 2019-12-17

Similar Documents

Publication Publication Date Title
WO2021022752A1 (en) Multimodal three-dimensional medical image fusion method and system, and electronic device
Li et al. Deep learning attention mechanism in medical image analysis: Basics and beyonds
Parvaiz et al. Vision Transformers in medical computer vision—A contemplative retrospection
WO2023077603A1 (en) Prediction system, method and apparatus for abnormal brain connectivity, and readable storage medium
Tagaris et al. Machine learning for neurodegenerative disorder diagnosis—survey of practices and launch of benchmark dataset
US20230342918A1 (en) Image-driven brain atlas construction method, apparatus, device and storage medium
Zuo et al. Multimodal representations learning and adversarial hypergraph fusion for early Alzheimer’s disease prediction
Du et al. Cardiac-DeepIED: Automatic pixel-level deep segmentation for cardiac bi-ventricle using improved end-to-end encoder-decoder network
Liu et al. Assessing clinical progression from subjective cognitive decline to mild cognitive impairment with incomplete multi-modal neuroimages
CN112674720A (en) Alzheimer disease pre-diagnosis method based on 3D convolutional neural network
Zhan et al. LR-cGAN: Latent representation based conditional generative adversarial network for multi-modality MRI synthesis
Chen et al. Generative adversarial U-Net for domain-free medical image augmentation
Peña-Solórzano et al. Findings from machine learning in clinical medical imaging applications–Lessons for translation to the forensic setting
Aslan On the use of deep learning methods on medical images
Jin et al. A hybrid deep learning method for early and late mild cognitive impairment diagnosis with incomplete multimodal data
Asif et al. StoneNet: an efficient lightweight model based on depthwise separable convolutions for kidney stone detection from CT images
Li et al. Ensemble of convolutional neural networks and multilayer perceptron for the diagnosis of mild cognitive impairment and Alzheimer's disease
Kong et al. Data enhancement based on M2-Unet for liver segmentation in Computed Tomography
CN118230038A (en) Eye socket lymphoproliferative disease classification and identification method and system based on image analysis
Hwang et al. Real-world prediction of preclinical Alzheimer’s disease with a deep generative model
Li et al. A review of deep learning-based information fusion techniques for multimodal medical image classification
Zhang et al. An end-to-end multimodal 3D CNN framework with multi-level features for the prediction of mild cognitive impairment
Mahapatra Multimodal generalized zero shot learning for gleason grading using self-supervised learning
Pallawi et al. Study of Alzheimer’s disease brain impairment and methods for its early diagnosis: a comprehensive survey
Al-qaness et al. Chest X-ray Images for Lung Disease Detection Using Deep Learning Techniques: A Comprehensive Survey

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 19940462

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: DE

122 Ep: pct application non-entry in european phase

Ref document number: 19940462

Country of ref document: EP

Kind code of ref document: A1

32PN Ep: public notification in the ep bulletin as address of the adressee cannot be established

Free format text: NOTING OF LOSS OF RIGHTS PURSUANT TO RULE 112(1) EPC (EPO FORM 1205A DATED 16/02/2023)