WO2021184195A1 - 医学图像重建方法、医学图像重建网络训练方法和装置 - Google Patents
医学图像重建方法、医学图像重建网络训练方法和装置 Download PDFInfo
- Publication number
- WO2021184195A1 WO2021184195A1 PCT/CN2020/079678 CN2020079678W WO2021184195A1 WO 2021184195 A1 WO2021184195 A1 WO 2021184195A1 CN 2020079678 W CN2020079678 W CN 2020079678W WO 2021184195 A1 WO2021184195 A1 WO 2021184195A1
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- image
- network
- vector
- image reconstruction
- real
- Prior art date
Links
- 238000012549 training Methods 0.000 title claims abstract description 90
- 238000000034 method Methods 0.000 title claims abstract description 82
- 239000013598 vector Substances 0.000 claims abstract description 187
- 238000000605 extraction Methods 0.000 claims abstract description 42
- 230000006870 function Effects 0.000 claims description 77
- 238000005259 measurement Methods 0.000 claims description 25
- 238000012886 linear function Methods 0.000 claims description 18
- 238000005457 optimization Methods 0.000 claims description 18
- 238000011478 gradient descent method Methods 0.000 claims description 10
- 239000011159 matrix material Substances 0.000 claims description 4
- 230000008447 perception Effects 0.000 claims description 4
- 238000002599 functional magnetic resonance imaging Methods 0.000 description 20
- 238000004590 computer program Methods 0.000 description 19
- 230000008569 process Effects 0.000 description 13
- 238000010586 diagram Methods 0.000 description 7
- 210000004556 brain Anatomy 0.000 description 5
- 238000013473 artificial intelligence Methods 0.000 description 4
- 238000013135 deep learning Methods 0.000 description 4
- 238000005516 engineering process Methods 0.000 description 4
- 238000010191 image analysis Methods 0.000 description 4
- 230000001537 neural effect Effects 0.000 description 4
- 238000012545 processing Methods 0.000 description 4
- 230000004044 response Effects 0.000 description 4
- 230000008878 coupling Effects 0.000 description 3
- 238000010168 coupling process Methods 0.000 description 3
- 238000005859 coupling reaction Methods 0.000 description 3
- 238000011160 research Methods 0.000 description 3
- 238000004458 analytical method Methods 0.000 description 2
- QVGXLLKOCUKJST-UHFFFAOYSA-N atomic oxygen Chemical compound [O] QVGXLLKOCUKJST-UHFFFAOYSA-N 0.000 description 2
- 239000008280 blood Substances 0.000 description 2
- 210000004369 blood Anatomy 0.000 description 2
- 210000003710 cerebral cortex Anatomy 0.000 description 2
- 238000004891 communication Methods 0.000 description 2
- 230000002708 enhancing effect Effects 0.000 description 2
- 230000000004 hemodynamic effect Effects 0.000 description 2
- 238000002595 magnetic resonance imaging Methods 0.000 description 2
- 239000002184 metal Substances 0.000 description 2
- 238000002610 neuroimaging Methods 0.000 description 2
- 229910052760 oxygen Inorganic materials 0.000 description 2
- 239000001301 oxygen Substances 0.000 description 2
- 210000002442 prefrontal cortex Anatomy 0.000 description 2
- 210000004515 ventral tegmental area Anatomy 0.000 description 2
- 238000012952 Resampling Methods 0.000 description 1
- 230000009471 action Effects 0.000 description 1
- 230000009286 beneficial effect Effects 0.000 description 1
- 230000015572 biosynthetic process Effects 0.000 description 1
- 238000013461 design Methods 0.000 description 1
- 238000001514 detection method Methods 0.000 description 1
- 238000002059 diagnostic imaging Methods 0.000 description 1
- 230000000694 effects Effects 0.000 description 1
- 230000004927 fusion Effects 0.000 description 1
- 230000003993 interaction Effects 0.000 description 1
- 230000007246 mechanism Effects 0.000 description 1
- 238000012986 modification Methods 0.000 description 1
- 230000004048 modification Effects 0.000 description 1
- 230000002093 peripheral effect Effects 0.000 description 1
- 230000000087 stabilizing effect Effects 0.000 description 1
- 238000003786 synthesis reaction Methods 0.000 description 1
- 230000002194 synthesizing effect Effects 0.000 description 1
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T11/00—2D [Two Dimensional] image generation
- G06T11/003—Reconstruction from projections, e.g. tomography
- G06T11/005—Specific pre-processing for tomographic reconstruction, e.g. calibration, source positioning, rebinning, scatter correction, retrospective gating
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V10/00—Arrangements for image or video recognition or understanding
- G06V10/70—Arrangements for image or video recognition or understanding using pattern recognition or machine learning
- G06V10/82—Arrangements for image or video recognition or understanding using pattern recognition or machine learning using neural networks
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T11/00—2D [Two Dimensional] image generation
- G06T11/003—Reconstruction from projections, e.g. tomography
- G06T11/008—Specific post-processing after tomographic reconstruction, e.g. voxelisation, metal artifact correction
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T9/00—Image coding
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V10/00—Arrangements for image or video recognition or understanding
- G06V10/70—Arrangements for image or video recognition or understanding using pattern recognition or machine learning
- G06V10/74—Image or video pattern matching; Proximity measures in feature spaces
- G06V10/761—Proximity, similarity or dissimilarity measures
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V10/00—Arrangements for image or video recognition or understanding
- G06V10/70—Arrangements for image or video recognition or understanding using pattern recognition or machine learning
- G06V10/77—Processing image or video features in feature spaces; using data integration or data reduction, e.g. principal component analysis [PCA] or independent component analysis [ICA] or self-organising maps [SOM]; Blind source separation
- G06V10/774—Generating sets of training patterns; Bootstrap methods, e.g. bagging or boosting
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V10/00—Arrangements for image or video recognition or understanding
- G06V10/70—Arrangements for image or video recognition or understanding using pattern recognition or machine learning
- G06V10/77—Processing image or video features in feature spaces; using data integration or data reduction, e.g. principal component analysis [PCA] or independent component analysis [ICA] or self-organising maps [SOM]; Blind source separation
- G06V10/776—Validation; Performance evaluation
-
- G—PHYSICS
- G01—MEASURING; TESTING
- G01R—MEASURING ELECTRIC VARIABLES; MEASURING MAGNETIC VARIABLES
- G01R33/00—Arrangements or instruments for measuring magnetic variables
- G01R33/20—Arrangements or instruments for measuring magnetic variables involving magnetic resonance
- G01R33/44—Arrangements or instruments for measuring magnetic variables involving magnetic resonance using nuclear magnetic resonance [NMR]
- G01R33/48—NMR imaging systems
- G01R33/4806—Functional imaging of brain activation
-
- G—PHYSICS
- G01—MEASURING; TESTING
- G01R—MEASURING ELECTRIC VARIABLES; MEASURING MAGNETIC VARIABLES
- G01R33/00—Arrangements or instruments for measuring magnetic variables
- G01R33/20—Arrangements or instruments for measuring magnetic variables involving magnetic resonance
- G01R33/44—Arrangements or instruments for measuring magnetic variables involving magnetic resonance using nuclear magnetic resonance [NMR]
- G01R33/48—NMR imaging systems
- G01R33/54—Signal processing systems, e.g. using pulse sequences ; Generation or control of pulse sequences; Operator console
- G01R33/56—Image enhancement or correction, e.g. subtraction or averaging techniques, e.g. improvement of signal-to-noise ratio and resolution
- G01R33/5608—Data processing and visualization specially adapted for MR, e.g. for feature analysis and pattern recognition on the basis of measured MR data, segmentation of measured MR data, edge contour detection on the basis of measured MR data, for enhancing measured MR data in terms of signal-to-noise ratio by means of noise filtering or apodization, for enhancing measured MR data in terms of resolution by means for deblurring, windowing, zero filling, or generation of gray-scaled images, colour-coded images or images displaying vectors instead of pixels
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2210/00—Indexing scheme for image generation or computer graphics
- G06T2210/41—Medical
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V2201/00—Indexing scheme relating to image or video recognition or understanding
- G06V2201/03—Recognition of patterns in medical or anatomical images
Definitions
- the embodiments of the present application belong to the field of image processing technology, and in particular relate to a medical image reconstruction method, a medical image reconstruction network training method and device.
- Functional magnetic resonance imaging is an emerging neuroimaging method. Its principle is to use magnetic resonance imaging to measure the changes in hemodynamics caused by neuronal activity. As a non-interventional technology, it can accurately locate specific cortical areas of the brain and capture changes in blood oxygen that can reflect neuronal activity.
- fMRI functional magnetic resonance imaging
- the number of images that can be acquired is often limited in certain application scenarios, which greatly limits The application of artificial intelligence methods that rely on large amounts of data such as deep learning in the field of medical image analysis.
- a promising solution is to use existing artificial intelligence methods to use limited real image samples to learn to reconstruct corresponding medical images from Gaussian hidden layer vectors, so as to achieve the purpose of enhancing the sample size and supporting subsequent image analysis tasks.
- the generative adversarial network is the current best-performing generative model, which has gradually become a research hotspot of deep learning, and has begun to be applied to the field of medical imaging.
- the traditional generative adversarial network can generate new images with diversity by learning the distribution of real data, but it also has the problem of difficulty in network training and difficulty in achieving optimal convergence.
- One aspect of the embodiments of the present application provides a medical image reconstruction network training method, which includes:
- an image reconstruction network performing image reconstruction based on the feature code vector to obtain a first image, and performing image reconstruction based on the first hidden layer vector of the real image sample to obtain a second image;
- the performing feature encoding extraction on a real image sample to obtain a feature encoding vector of the real image sample includes:
- the feature extraction of the real image sample based on the image coding network to obtain the feature coding vector of the real image sample includes:
- the extracted features are processed through a linear function to obtain the feature encoding vector of the real image sample.
- the linear function is a piecewise linear function.
- the piecewise linear function is a ReLU function.
- the method further includes:
- the image coding network is optimized based on the vector discrimination result.
- the optimizing the image coding network based on the vector discrimination result includes:
- the performing confrontation training on the image coding network based on the vector discrimination result includes:
- the first loss function is:
- L C is the first loss function
- z e is the feature coding vector
- z r is the first hidden layer vector
- C represents the image coding network
- E is the mathematical expectation.
- the optimizing the image reconstruction network according to the image discrimination result includes:
- the performing confrontation training on the image reconstruction network according to the image discrimination result includes:
- the structural similarity measurement loss function and the perception measurement loss function, the second loss function of the image reconstruction network is determined, and the network parameters of the image reconstruction network are updated by the gradient descent method. Rebuild the network for training;
- the second loss function is:
- L G is the second loss function
- z e is the feature coding vector
- z r is the first hidden layer vector
- C represents the image coding network
- D is the image discrimination network
- G is In the image reconstruction network
- E is the mathematical expectation
- L SSIM is the structural similarity measurement loss function
- L perceptual is the perceptual measurement loss function
- X real represents the real image
- ⁇ 1 and ⁇ 2 are the weight coefficients
- ⁇ is the Gram matrix
- L D is the loss function of the image discrimination network.
- the performing image reconstruction based on the feature coding vector to obtain the first image through the image generation network, and performing image reconstruction based on the first hidden layer vector of the real image sample to obtain the second image includes:
- the feature encoding vector and the first hidden layer vector are input to the image reconstruction network to obtain the first image and the second image; wherein, the convolutional layer of the image generation network is upsampled by neighbors Three-dimensional separable convolutional layer.
- a second aspect of the embodiments of the present application provides a medical image reconstruction method, which includes:
- image reconstruction is performed on the image to be reconstructed based on the second hidden layer vector.
- the third aspect of the embodiments of the present application provides a medical image reconstruction network training device, which includes:
- the feature encoding extraction module is used to perform feature encoding extraction on real image samples to obtain feature encoding vectors of the real image samples;
- the first image reconstruction module is configured to perform image reconstruction based on the feature code vector through an image reconstruction network to obtain a first image, and perform image reconstruction based on the first hidden layer vector of the real image sample to obtain a second image;
- the first optimization module is configured to perform image discrimination on the real image sample, the first image and the second image through an image discrimination network, and optimize the image generation network according to the result of the image discrimination.
- a fourth aspect of the embodiments of the present application provides a medical image reconstruction device, which includes:
- Hidden layer vector acquisition module for acquiring the second hidden layer vector of the image to be reconstructed
- the second image reconstruction module is configured to perform image reconstruction on the reconstructed image based on the second hidden layer vector through the trained image reconstruction network.
- a fifth aspect of the embodiments of the present application provides a terminal device, including a memory, a processor, and computer-readable instructions stored in the memory and running on the processor, and the processor executes the computer-readable instructions. The following steps are implemented when ordering:
- an image reconstruction network performing image reconstruction based on the feature code vector to obtain a first image, and performing image reconstruction based on the first hidden layer vector of the real image sample to obtain a second image;
- the performing feature encoding extraction on a real image sample to obtain a feature encoding vector of the real image sample includes:
- the processor further implements the following steps when executing the computer-readable instructions:
- the image coding network is optimized based on the vector discrimination result.
- a sixth aspect of the embodiments of the present application provides a terminal device.
- the terminal device includes a memory, a processor, and computer-readable instructions that are stored in the memory and run on the processor, and the processor executes all When the computer-readable instructions are described, the following steps are implemented:
- image reconstruction is performed on the image to be reconstructed based on the second hidden layer vector.
- a seventh aspect of the embodiments of the present application provides a computer storage medium, the computer-readable storage medium stores computer-readable instructions, wherein the computer-readable instructions are executed by a processor to implement the following steps:
- an image reconstruction network performing image reconstruction based on the feature code vector to obtain a first image, and performing image reconstruction based on the first hidden layer vector of the real image sample to obtain a second image;
- the eighth aspect of the embodiments of the present application provides a computer storage medium, the computer-readable storage medium stores computer-readable instructions, wherein the computer-readable instructions are executed by a processor to implement the following steps:
- image reconstruction is performed on the image to be reconstructed based on the second hidden layer vector.
- the ninth aspect of the embodiments of the present application provides a computer program product, which when the computer program product runs on a terminal device, causes the terminal device to execute any one of the above-mentioned methods in the first aspect, or causes the terminal device to execute the above-mentioned second aspect Any one of the methods.
- feature encoding and extraction are performed on real image samples to obtain the feature encoding vector of the real image sample
- the first image is obtained by image reconstruction based on the feature encoding vector through the image reconstruction network, based on the hidden layer of the real image sample
- the vector is used for image reconstruction to obtain the second image.
- the real image sample, the first image, and the second image are distinguished through the image distinguishing network, and the image reconstruction network is optimized according to the image distinguishing result, and the optimized
- the image reconstruction network is used for image reconstruction, and introduces prior knowledge guidance from real images for the generation of the confrontation network, so as to stabilize the training of the image reconstruction network, and is easy to achieve optimal convergence, thereby solving the problem of difficulty in training the generation of the confrontation network.
- Fig. 1 is a schematic diagram of an application scenario provided by an embodiment of the present application
- FIG. 2 is a schematic flowchart of a medical image reconstruction network training method provided by an embodiment of the present application
- FIG. 3 is a schematic flowchart of a medical image reconstruction network training method provided by an embodiment of the present application.
- FIG. 4 is a schematic flowchart of a medical image reconstruction network training method provided by an embodiment of the present application.
- FIG. 5 is a schematic flowchart of a medical image reconstruction method provided by an embodiment of the present application.
- Fig. 6 is a schematic diagram of a medical image reconstruction process provided by an embodiment of the present application.
- FIG. 7 is a schematic structural diagram of a medical image reconstruction network training device provided by an embodiment of the present application.
- FIG. 8 is a schematic structural diagram of a medical image reconstruction device provided by an embodiment of the present application.
- FIG. 9 is a schematic structural diagram of a terminal device provided by an embodiment of the present application.
- the term “if” can be construed as “when” or “once” or “in response to determination” or “in response to detecting “.
- the phrase “if determined” or “if detected [described condition or event]” can be interpreted as meaning “once determined” or “in response to determination” or “once detected [described condition or event]” depending on the context ]” or “in response to detection of [condition or event described]”.
- Functional magnetic resonance imaging fMRI is an emerging neuroimaging method, its principle is to use magnetic resonance imaging to measure the changes in hemodynamics caused by neuronal activity. As a non-interventional technology, it can accurately locate specific cortical areas of the brain and capture changes in blood oxygen that can reflect neuronal activity.
- fMRI image acquisition and long scanning time and some special patients cannot perform it (such as those with metal objects in the body cannot accept scanning)
- the number of images that can be acquired is often limited in certain application scenarios, which greatly limits The application of artificial intelligence methods that rely on large amounts of data such as deep learning in the field of medical image analysis.
- a promising solution is to use existing artificial intelligence methods to use limited real image samples to learn to reconstruct corresponding medical images from Gaussian hidden layer vectors, so as to achieve the purpose of enhancing the sample size and supporting subsequent image analysis tasks.
- the generative adversarial network is currently the best-performing generative model. It was first proposed by Lan Goodfellow et al. in 2014. It can capture the potential distribution of real data through the generator, so as to achieve the purpose of generating real data distribution samples from the hidden layer space. . Since then, generative adversarial networks have gradually become a research hotspot in deep learning and have begun to be applied in various fields.
- the generative adversarial network can generate new images with diversity by learning the real data distribution, but its biggest problem is that it is difficult to train the network and it is not easy to achieve optimal convergence.
- the purpose of generating the confrontation network is to make the data distribution fitted by the generator close to the real data distribution.
- the inventor of the present application found in the research that the generation network introduced without any prior knowledge does not know the real data distribution at all, and can only be judged based on The true and false feedback of the device is tested again and again.
- this problem does not exist. It can first extract the encoding feature vector of the real image, and at the same time perform variational inference through resampling, and decode the hidden vector according to the variational result. generate.
- the embodiment of the present application introduces the encoding feature vector of the variational autoencoder as a feature prior knowledge about the real image into the training of the generative confrontation network, giving the generative network a comparison.
- a clear optimization direction to solve the problem of difficult, time-consuming, and easy to crash training.
- this application further introduces a separate encoding discriminator, so that the optimization process of the variational autoencoder is also incorporated into the "generation-adversarial" system to solve the difference between the objective function of the variational reasoning and the generative adversarial network. There is an optimization conflict between.
- the embodiments of the present application can be applied to the exemplary scenario shown in FIG. 1.
- the terminal 10 and the server 20 constitute application scenarios of the above-mentioned medical image reconstruction network training method and medical image reconstruction method.
- the terminal 10 is used to obtain a real image sample of the subject and send the real image sample to the server 20;
- the server 20 is used to perform feature encoding extraction on the real image sample to obtain the feature encoding vector of the real image sample, and pass Image reconstruction network, image reconstruction based on feature coding vector to obtain the first image, image reconstruction based on the hidden layer vector of the real image sample to obtain the second image, and image discrimination network for the real image sample, the first image and the second image Discrimination, and optimize the image reconstruction network based on the results of the image discrimination, use the optimized image generation network for image reconstruction work, and introduce prior knowledge guidance from the real image for the generation of the confrontation network, thereby stabilizing the image reconstruction network Training is easy to achieve optimal convergence, thereby solving the problem of difficulty in training the generation of adversarial networks.
- FIG. 2 is a schematic flowchart of a medical image reconstruction network training method provided by an embodiment of the present application. Referring to FIG. 2, the medical image reconstruction network training method is detailed as follows:
- step 101 feature encoding extraction is performed on a real image sample to obtain a feature encoding vector of the real image sample.
- step 101 feature extraction may be performed on the above-mentioned real image sample through an image coding network to obtain the feature coding vector of the above-mentioned real image sample.
- the feature extraction of the real image sample through the image coding network to obtain the feature encoding vector of the real image sample may specifically include:
- step 1011 hierarchical feature extraction is performed on the real image sample through the multiple three-dimensional convolutional layers of the image coding network.
- step 1012 the extracted features are processed through a linear function to obtain the feature encoding vector of the real image sample.
- real image samples can be generated into 3D images in a time series, and the 3D images can be sequentially input into the image coding network, and the 3D images can be layered using multiple 3D convolutional layers of the image coding network Feature extraction, and synthesize the linear and non-linear features of the three-dimensional image through a linear function to obtain the feature encoding representation vector of the real image sample.
- the above-mentioned linear function is a piecewise linear function. Specifically, the linear and non-linear features of the three-dimensional image are processed by the piecewise linear function to obtain the feature encoding representation vector of the real image sample.
- the piecewise linear function may be a ReLU function. Specifically, the linear feature and the non-linear feature of the three-dimensional image are processed through the ReLU function to obtain the feature encoding representation vector of the real image sample.
- step 102 through an image reconstruction network, image reconstruction is performed based on the feature code vector to obtain a first image, and image reconstruction is performed based on the first hidden layer vector of the real image sample to obtain a second image.
- the above-mentioned feature encoding vector and the above-mentioned first hidden layer vector may be input to the above-mentioned image reconstruction network to obtain the above-mentioned first image and the above-mentioned second image; wherein, the convolution of the image generation network in this embodiment of the application
- the layer is a three-dimensional separable convolutional layer that is upsampled by neighbors.
- the feature code vector extracted from the real image sample and the first hidden layer vector sampled from the Gaussian distribution of the real image sample may be used as the input of the image reconstruction network, and the feature code vector and the first hidden layer can be used as input to the image reconstruction network.
- the first image and the second image are obtained by stepwise reconstruction in the vector.
- a three-dimensional separable convolutional layer with neighbor upsampling is used to replace the deconvolutional layer in the traditional image reconstruction network, which can reduce the number of learnable parameters, and can improve the quality of the generated fMRI image, so that the reconstruction The image has fewer artifacts, clearer structure, etc.
- step 103 image discrimination is performed on the real image sample, the first image, and the second image through an image discrimination network, and the image reconstruction network is optimized according to the result of the image discrimination.
- the real image samples, the first image, and the second image can all be used as the input of the image discrimination network, and the image reconstruction network can be optimized according to the discrimination results of the image discrimination network, and the "generation-adversarial" training can be constructed, and The optimized and trained image reconstruction network is used for image reconstruction.
- Step 103 is executed again after the image, and the execution is repeated in turn.
- the above medical image reconstruction network training method is to perform feature encoding extraction on real image samples to obtain feature encoding vectors of the above real image samples.
- image reconstruction is performed based on the feature encoding vectors to obtain the first image, based on the above real image samples
- the second image is obtained by image reconstruction with the hidden layer vector of, and at the same time, the real image sample, the first image and the second image are distinguished through the image discrimination network, and the image reconstruction network is optimized according to the result of the image discrimination.
- the optimized image reconstruction network is used for image reconstruction work, and introduces prior knowledge guidance from real images for the generation of the confrontation network, so as to stabilize the training of the image reconstruction network, and is easy to achieve optimal convergence, thereby solving the difficulty of generating the confrontation network training The problem.
- FIG. 4 is a schematic flowchart of a medical image reconstruction network training method provided by an embodiment of this application.
- the medical image reconstruction network training method may further include:
- step 104 vector discrimination is performed on the feature coding vector and the first hidden layer vector through the coding feature discrimination network.
- step 105 the image coding network is optimized based on the vector discrimination result.
- the feature coding vector and the first hidden layer vector of the real image sample can be used to optimize the image coding network through steps 104 and 105, and then the optimized image coding network can be used as The image coding network in step 101 can be used to perform step 101 again; this loop is repeated to optimize the image coding network.
- the image coding network may be subjected to confrontation training based on the vector discrimination result, so as to optimize the above-mentioned image coding network.
- a coding feature discrimination network with the same structure as the image discrimination network can be constructed, and the feature coding vector obtained from real image samples and the first hidden layer vector sampled from the Gaussian distribution are used as the input of the coding feature discrimination network.
- the coding feature discrimination network and the image coding network also constitute a "generation-antagonism" training relationship, instead of variational inference, to solve the problem of training conflicts between variational inference and generative confrontation objective function.
- the above-mentioned adversarial training of the image coding network based on the vector discrimination result may specifically include: calculating the voxel-by-voxel difference between the second image and the real image sample, and performing gradient descent Method to update the network parameters of the image coding network until the voxel-by-voxel difference is less than or equal to the preset threshold to realize the training of the image coding network, wherein the voxel-by-voxel difference is the first of the image coding network A loss function.
- the coding feature discrimination network is introduced to replace the original variational reasoning process.
- the training process of the above image coding network first calculate the voxel-by-voxel difference between the reconstructed fMRI image and the real fMRI image, and then update the network parameters of the above-mentioned image coding network through the gradient descent method, so that the voxel-by-voxel difference is less than or equal to the first A preset threshold is enough; secondly, the Wasserstein distance is selected as the measurement tool of the real image distribution and the reconstructed image distribution in the first loss function, and the gradient penalty item is introduced to crop the discriminator network gradient to further stabilize the network training.
- the first loss function may be:
- L C is the first loss function
- z e is the feature coding vector
- z r is the first hidden layer vector
- C represents the image coding network
- E is the mathematical expectation.
- the optimization of the image reconstruction network according to the image discrimination result in step 103 may specifically be: performing confrontation training on the image reconstruction network according to the image discrimination result.
- performing confrontation training on the image reconstruction network according to the image discrimination result may include: determining the second image reconstruction network according to the image discrimination result, structural similarity measurement loss function, and perception measurement loss function Loss function, and update the network parameters of the image reconstruction network through a gradient descent method, and train the image reconstruction network.
- the image reconstruction network is subjected to confrontation training according to the image discrimination result. Specifically, if the discrimination result of the image discrimination network is closer to the real image, then only the image reconstruction network needs to be first performed on the network parameters by using the gradient descent method. The preset amplitude is updated or not updated; if the judgment result of the image discrimination network is closer to the reconstructed image, the image reconstruction network needs to update the network parameters with the second preset amplitude, and the second preset amplitude is greater than the first preset amplitude .
- the structural similarity measurement loss and the perceptual measurement loss are also introduced to ensure that the characteristics of the reconstructed image are more in line with the real image.
- the second loss function may be:
- L G is the second loss function
- z e is the feature coding vector
- z r is the first hidden layer vector
- C represents the image coding network
- D is the image discrimination network
- G is In the image reconstruction network
- E is the mathematical expectation
- L SSIM is the structural similarity measurement loss function
- L perceptual is the perceptual measurement loss function
- X real represents the real image
- ⁇ 1 and ⁇ 2 are the weight coefficients
- ⁇ is the Gram matrix
- L D is the loss function of the image discrimination network.
- the image overlap ratio (SOR) technical indicator can be used to evaluate the closeness of the reconstructed image reconstructed by the image reconstruction network to the real image.
- the trained image reconstruction network can reconstruct high-quality medical image samples from the Gaussian hidden layer vector, which can enhance the image sample size and facilitate subsequent analysis.
- FIG. 5 is a schematic flowchart of a medical image reconstruction method provided by an embodiment of the present application. Referring to FIG. 5, the medical image reconstruction method is described in detail as follows:
- step 201 the second hidden layer vector of the image to be reconstructed is obtained.
- step 202 image reconstruction is performed on the image to be reconstructed based on the second hidden layer vector through the trained image reconstruction network.
- the medical image reconstruction method described above is to perform feature encoding extraction on real image samples to obtain the feature encoding vector of the real image sample.
- image reconstruction is performed based on the feature encoding vector to obtain the first image.
- the layer vector is used for image reconstruction to obtain the second image.
- the real image sample, the first image, and the second image are distinguished through the image discriminating network, and the image reconstruction network is trained and optimized according to the image discriminating result, and passed
- the image reconstruction network after training optimization is based on the second hidden layer vector to reconstruct the image to be reconstructed, and introduces the prior knowledge guidance from the real image for the generation of the confrontation network, so as to stabilize the training of the image reconstruction network and easily achieve optimal convergence , So as to solve the problem of the difficulty of generating the confrontation network training, and the reconstructed image is closer to the real image.
- the process of medical image reconstruction may include the following steps:
- step 301 feature extraction is performed on the real image sample based on the image coding network to obtain the feature coding vector of the real image sample.
- step 302 through the image reconstruction network, the first image is obtained by image reconstruction based on the feature code vector, and the second image is obtained by image reconstruction based on the first hidden layer vector of the real image sample.
- step 303 the real image sample, the first image, and the second image are discriminated by the image discriminating network, and the image reconstruction network is trained and optimized according to the image discriminating result.
- the image reconstruction network after training and optimization is used as the image reconstruction network in step 302 to perform the next image reconstruction.
- step 304 the feature encoding vector in step 301 and the first hidden layer vector of the real image sample are subjected to vector discrimination through the encoding feature discrimination network.
- step 305 based on the vector discrimination result, the image coding network is optimized, and the optimized image coding network is used as the image coding network in step 301 to perform feature extraction on the next real image sample.
- step 306 after the image reconstruction network training and optimization are completed through the real image samples, the second hidden layer vector of the image to be reconstructed is acquired.
- step 307 image reconstruction is performed on the image to be reconstructed based on the second hidden layer vector through the trained image reconstruction network.
- the real fMRI image x real of the rat brain area is developed into a three-dimensional image in a time series, and input into the image coding network in turn, and the three-dimensional image is layered using multiple three-dimensional convolutional layers of the image coding network Feature extraction, and synthesize linear and non-linear features through the ReLU function, and output the feature encoding vector z e of the real fMRI image.
- the feature encoding vector z e obtained by extracting the real fMRI image and the hidden layer vector z r sampled from the Gaussian distribution are both used as the input of the image reconstruction network, which are reconstructed step by step from z e and z r respectively fMRI images x rec and x rand .
- the convolution of the image reconstruction network is a three-dimensional separable convolutional layer with nearest neighbor upsampling.
- the use of a three-dimensional separable convolution operation with nearest neighbor upsampling instead of the traditional deconvolution layer can reduce the number of learnable parameters and Improve the quality of the reconstructed fMRI image, so that the reconstructed image has fewer artifacts and the brain structure is clearer.
- the real fMRI image x real , image x rec and image x rand are all used as the input of the image discrimination network, and the image reconstructor is optimized according to the discrimination result of the image discrimination network to construct "generation-adversarial" training .
- a coding feature discrimination network with the same structure as the image discrimination network is constructed, and the feature representation vector z e encoded from the real fMRI image x real and the hidden layer vector z r sampled from the Gaussian distribution are used as its input to make the encoding
- the feature discrimination network and the image coding network also constitute a "generation-antagonism" training relationship to replace the variational reasoning and solve the training conflict between the variational reasoning and the generation against the objective function.
- the fourth step is to select the optimal loss function to train and optimize the network.
- this embodiment cleverly introduces the coding feature discrimination network to replace the traditional variational reasoning process, and only needs to minimize the voxel difference between the reconstructed fMRI image and the real fMRI image; moreover, this application
- the Wasserstein distance is selected as the measurement tool of the real image distribution and the reconstructed image distribution in the loss function, and the gradient penalty item is introduced to crop the discriminator network gradient to further stabilize the image coding network training.
- this application also introduces structural similarity measurement loss and perceptual measurement loss to ensure that the reconstructed image is in the rat ventral tegmental area (VTA), prefrontal cortex (PFC), etc.
- VTA ventral tegmental area
- PFC prefrontal cortex
- the characteristics of the key area correspond to the real image.
- the loss function formula of each network is as follows:
- the loss function of the image coding network is:
- the loss function of the image discrimination network is:
- the loss function of the image reconstruction network is:
- L SSIM is the structural similarity measurement loss function
- L perceptual represents the perceptual measurement loss function
- this program intends to evaluate the closeness of the reconstructed image to the real image through the image overlap ratio (SOR) technical indicator.
- the trained image reconstruction network reconstructs high-quality medical image samples from the Gaussian hidden layer vector of the image to be reconstructed, which enhances the image sample size and facilitates subsequent analysis.
- the embodiment of this application proposes a medical image reconstruction network training method that integrates a variational autoencoder and a generative confrontation network. Compared with the traditional generative confrontation network, this application introduces a method from the real world through the fusion variational autoencoder. The image is guided by prior knowledge, so as to solve the difficult problem of generating confrontation network training.
- a separate coding discrimination network is added between the variational autoencoder and the generative confrontation network. Its goal is to replace the function of variational inference, so that the encoding feature vector of the variational encoder can be trained as a confrontation.
- the method approximates the original Gaussian hidden layer vector, so as to resolve the conflict between the variational reasoning and the objective function of the generated confrontation network.
- FIG. 7 shows the structural block diagram of the medical image reconstruction network training device provided by the embodiment of the present application. part.
- the medical image reconstruction network training device in the embodiment of the present application may include a feature code extraction module 401, a first image reconstruction module 402 and an optimization module 403.
- the feature encoding extraction module 401 is configured to perform feature encoding extraction on real image samples to obtain feature encoding vectors of the real image samples;
- the first image reconstruction module 402 is configured to perform image reconstruction based on the feature code vector through an image reconstruction network to obtain a first image, and perform image reconstruction based on the first hidden layer vector of the real image sample to obtain a second image;
- the first optimization module 403 is configured to perform image discrimination on the real image sample, the first image and the second image through an image discrimination network, and optimize the image reconstruction network according to the result of the image discrimination.
- the feature coding extraction module 401 may be used to: perform feature extraction on the real image sample based on an image coding network to obtain a feature coding vector of the real image sample.
- the feature encoding extraction module 401 may be specifically used for:
- the extracted features are processed through a linear function to obtain the feature encoding vector of the real image sample.
- the linear function is a piecewise linear function.
- the piecewise linear function is a ReLU function.
- the medical image reconstruction network training device may further include a second optimization module; the second optimization module is used to:
- the image coding network is optimized based on the vector discrimination result.
- the optimizing the image coding network based on the vector discrimination result includes:
- the performing confrontation training on the image coding network based on the vector discrimination result includes:
- the voxel-by-voxel difference is the first loss function of the image coding network, and the first loss function is:
- L C is the first loss function
- z e is the feature coding vector
- z r is the first hidden layer vector
- C represents the image coding network
- E is the mathematical expectation.
- the first optimization module 403 may be used for:
- the performing confrontation training on the image reconstruction network according to the image discrimination result may include:
- the structural similarity measurement loss function and the perception measurement loss function, the second loss function of the image reconstruction network is determined, and the network parameters of the image reconstruction network are updated by the gradient descent method. Rebuild the network for training;
- the second loss function is:
- L G is the second loss function
- z e is the feature coding vector
- z r is the first hidden layer vector
- C represents the image coding network
- D is the image discrimination network
- G is In the image reconstruction network
- E is the mathematical expectation
- L SSIM is the structural similarity measurement loss function
- L perceptual is the perceptual measurement loss function
- X real represents the real image
- ⁇ 1 and ⁇ 2 are the weight coefficients
- ⁇ is the Gram matrix
- L D is the loss function of the image discrimination network.
- the first image reconstruction module 402 may be specifically used for:
- the feature encoding vector and the first hidden layer vector are input to the image reconstruction network to obtain the first image and the second image; wherein, the convolutional layer of the image generation network is upsampled by neighbors Three-dimensional separable convolutional layer.
- FIG. 8 shows a structural block diagram of the medical image reconstruction device provided by an embodiment of the present application. For ease of description, only the parts related to the embodiment of the present application are shown.
- the medical image reconstruction device in the embodiment of the present application may include a hidden layer vector acquisition module 501 and a second image reconstruction module 502.
- the hidden layer vector obtaining module 501 is used to obtain the second hidden layer vector of the image to be reconstructed;
- the second image reconstruction module 502 is configured to perform image reconstruction on the image to be reconstructed based on the second hidden layer vector through the trained image reconstruction network.
- the terminal device 600 may include: at least one processor 610, a memory 620, and stored in the memory 620 and available on the at least one processor 610.
- a running computer program when the processor 610 executes the computer program, the steps in any of the foregoing method embodiments are implemented, such as steps 101 to 103 in the embodiment shown in FIG. 2, or the embodiment shown in FIG. 5 Step 201 to step 202 in.
- the processor 610 executes the computer program
- the functions of the modules/units in the foregoing device embodiments are implemented, for example, the functions of the modules 401 to 403 shown in FIG. 7 or the functions of the modules 501 to 502 shown in FIG. 8.
- the computer program may be divided into one or more modules/units, and one or more modules/units are stored in the memory 620 and executed by the processor 610 to complete the application.
- the one or more modules/units may be a series of computer program segments capable of completing specific functions, and the program segments are used to describe the execution process of the computer program in the terminal device 600.
- FIG. 9 is only an example of a terminal device, and does not constitute a limitation on the terminal device. It may include more or less components than those shown in the figure, or a combination of certain components, or different components, such as Input and output equipment, network access equipment, bus, etc.
- the processor 610 may be a central processing unit (Central Processing Unit, CPU), other general-purpose processors, digital signal processors (Digital Signal Processor, DSP), application specific integrated circuits (Application Specific Integrated Circuit, ASIC), ready-made Field-Programmable Gate Array (FPGA) or other programmable logic devices, discrete gates or transistor logic devices, discrete hardware components, etc.
- the general-purpose processor may be a microprocessor or the processor may also be any conventional processor or the like.
- the memory 620 may be an internal storage unit of the terminal device, or an external storage device of the terminal device, such as a plug-in hard disk, a smart media card (SMC), a secure digital (SD) card, and a flash memory card. (Flash Card) and so on.
- the memory 620 is used to store the computer program and other programs and data required by the terminal device.
- the memory 620 may also be used to temporarily store data that has been output or will be output.
- the bus can be an Industry Standard Architecture (ISA) bus, Peripheral Component (PCI) bus, or Extended Industry Standard Architecture (EISA) bus, etc.
- ISA Industry Standard Architecture
- PCI Peripheral Component
- EISA Extended Industry Standard Architecture
- the bus can be divided into address bus, data bus, control bus and so on.
- the buses in the drawings of this application are not limited to only one bus or one type of bus.
- the embodiments of the present application also provide a computer-readable storage medium, the computer-readable storage medium stores a computer program, and when the computer program is executed by a processor, the steps in each embodiment of the above-mentioned medical image reconstruction network training method are implemented , Or implement the steps in each embodiment of the above-mentioned medical image reconstruction method.
- the embodiments of the present application provide a computer program product.
- the computer program product runs on a mobile terminal, the mobile terminal executes the steps in each embodiment of the above-mentioned medical image reconstruction network training method, or realizes the above-mentioned medical image reconstruction method. Steps in each embodiment.
- the integrated unit is implemented in the form of a software functional unit and sold or used as an independent product, it can be stored in a computer readable storage medium.
- the computer program can be stored in a computer-readable storage medium. When executed by the processor, the steps of the foregoing method embodiments can be implemented.
- the computer program includes computer program code, and the computer program code may be in the form of source code, object code, executable file, or some intermediate forms.
- the computer-readable medium may at least include: any entity or device capable of carrying the computer program code to the photographing device/terminal device, recording medium, computer memory, read-only memory (ROM, Read-Only Memory), and random access memory (RAM, Random Access Memory), electric carrier signal, telecommunications signal and software distribution medium.
- ROM read-only memory
- RAM random access memory
- electric carrier signal telecommunications signal and software distribution medium.
- U disk mobile hard disk, floppy disk or CD-ROM, etc.
- computer-readable media cannot be electrical carrier signals and telecommunication signals.
- the disclosed apparatus/network equipment and method may be implemented in other ways.
- the device/network device embodiments described above are only illustrative.
- the division of the modules or units is only a logical function division, and there may be other divisions in actual implementation, such as multiple units.
- components can be combined or integrated into another system, or some features can be omitted or not implemented.
- the displayed or discussed mutual coupling or direct coupling or communication connection may be indirect coupling or communication connection through some interfaces, devices or units, and may be in electrical, mechanical or other forms.
- the units described as separate components may or may not be physically separated, and the components displayed as units may or may not be physical units, that is, they may be located in one place, or they may be distributed on multiple network units. Some or all of the units may be selected according to actual needs to achieve the objectives of the solutions of the embodiments.
Landscapes
- Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- Physics & Mathematics (AREA)
- General Physics & Mathematics (AREA)
- Evolutionary Computation (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Multimedia (AREA)
- Health & Medical Sciences (AREA)
- Artificial Intelligence (AREA)
- Computing Systems (AREA)
- Databases & Information Systems (AREA)
- General Health & Medical Sciences (AREA)
- Medical Informatics (AREA)
- Software Systems (AREA)
- Image Analysis (AREA)
Abstract
Description
Claims (20)
- 一种医学图像重建网络训练方法,其中,所述医学图像重建网络训练方法包括:对真实图像样本进行特征编码提取,得到所述真实图像样本的特征编码向量;通过图像重建网络,基于所述特征编码向量进行图像重建得到第一图像,基于所述真实图像样本的第一隐层向量进行图像重建得到第二图像;通过图像判别网络对所述真实图像样本、所述第一图像和所述第二图像进行图像判别,并根据图像判别结果对所述图像重建网络进行优化。
- 如权利要求1所述的医学图像重建网络训练方法,其中,所述对真实图像样本进行特征编码提取,得到所述真实图像样本的特征编码向量,包括:基于图像编码网络对所述真实图像样本进行特征提取,得到所述真实图像样本的特征编码向量。
- 如权利要求2所述的医学图像重建网络训练方法,其中,所述基于图像编码网络对所述真实图像样本进行特征提取,得到所述真实图像样本的特征编码向量,包括:通过所述图像编码网络的多个三维卷积层对所述真实图像样本进行分层特征提取;通过线性函数对提取到的特征进行处理,得到所述真实图像样本的特征编码向量。
- 如权利要求3所述的医学图像重建网络训练方法,其中,所述线性函数为分段线性函数。
- 如权利要求4所述的医学图像重建网络训练方法,其中,所述分段线性函数为ReLU函数。
- 如权利要求2所述的医学图像重建网络训练方法,其中,所述方法还包括:通过编码特征判别网络对所述特征编码向量和所述第一隐层向量进行向量判别;基于向量判别结果对所述图像编码网络进行优化。
- 如权利要求6所述的医学图像重建网络训练方法,其中,所述基于向量判别结果对所述图像编码网络进行优化,包括:基于所述向量判别结果对所述图像编码网络进行对抗训练。
- 如权利要求1所述的医学图像重建网络训练方法,其中,所述根据图像判别结果对所述图像重建网络进行优化,包括:根据所述图像判别结果对所述图像重建网络进行对抗训练。
- 如权利要求9所述的医学图像重建网络训练方法,其中,所述根据所述图像判别结果对所述图像重建网络进行对抗训练,包括:根据所述图像判别结果、结构相似性度量损失函数和感知度量损失函数,确定所述图像重建网络的第二损失函数,并通过梯度下降法更新所述图像重建网络的网络参数,对所述图像重建网络进行训练;其中,所述第二损失函数为:L G为所述第二损失函数,z e为所述特征编码向量,z r为所述第一隐层向量,C表征所述图像编码网络,D为所述图像判别网络,G为所述图像重建网络,E为数学期望,L SSIM为结构相似性度量损失函数,L perceptual代表感知度量损失函数,X real表征所述真实图像,λ 1和λ 2为权重系数,Φ为Gram矩阵,L D为图像判别网络的损失函数。
- 如权利要求1所述的医学图像重建网络训练方法,其中,所述通过图像生成网络,基于所述特征编码向量进行图像重建得到第一图像,基于所述真实图像样本的第一隐层向量进行图像重建得到第二图像,包括:将所述特征编码向量和所述第一隐层向量输入所述图像重建网络,得到所述第一图像和所述第二图像;其中,所述图像生成网络的卷积层为近邻上采样的三维可分离卷积层。
- 一种医学图像重建方法,其中,所述医学图像重建方法包括:获取待重建图像的第二隐层向量;通过训练后的图像重建网络,基于所述第二隐层向量对所述待重建图像进行图像重建。
- 一种医学图像重建网络训练装置,其中,所述医学图像重建网络训练装置包括:特征编码提取模块,用于对真实图像样本进行特征编码提取,得到所述真实图像样本的特征编码向量;第一图像重建模块,用于通过图像重建网络,基于所述特征编码向量进行图像重建得到第一图像,基于所述真实图像样本的第一隐层向量进行图像重建得到第二图像;第一优化模块,用于通过图像判别网络对所述真实图像样本、所述第一图像和所述第二图像进行图像判别,并根据图像判别结果对所述图像生成网络进行优化。
- 一种医学图像重建装置,其中,所述医学图像重建装置包括:隐层向量获取模块,用于获取待重建图像的第二隐层向量;第二图像重建模块,用于通过训练后的图像重建网络,基于所述第二隐层向量对所述 带重建图像进行图像重建。
- 一种终端设备,其中,所述终端设备包括存储器、处理器以及存储在所述存储器中并可在所述处理器上运行的计算机可读指令,所述处理器执行所述计算机可读指令时实现如下步骤:对真实图像样本进行特征编码提取,得到所述真实图像样本的特征编码向量;通过图像重建网络,基于所述特征编码向量进行图像重建得到第一图像,基于所述真实图像样本的第一隐层向量进行图像重建得到第二图像;通过图像判别网络对所述真实图像样本、所述第一图像和所述第二图像进行图像判别,并根据图像判别结果对所述图像生成网络进行优化。
- 如权利要求15所述的终端设备,其中,所述对真实图像样本进行特征编码提取,得到所述真实图像样本的特征编码向量,包括:基于图像编码网络对所述真实图像样本进行特征提取,得到所述真实图像样本的特征编码向量。
- 如权利要求16所述的终端设备,其中,所述处理器执行所述计算机可读指令时还实现如下步骤:通过编码特征判别网络对所述特征编码向量和所述第一隐层向量进行向量判别;基于向量判别结果对所述图像编码网络进行优化。
- 一种终端设备,其中,所述终端设备包括存储器、处理器以及存储在所述存储器中并可在所述处理器上运行的计算机可读指令,所述处理器执行所述计算机可读指令时实现如下步骤:获取待重建图像的第二隐层向量;通过训练后的图像重建网络,基于所述第二隐层向量对所述待重建图像进行图像重建。
- 一种计算机可读存储介质,所述计算机可读存储介质存储有计算机可读指令,其中,所述计算机可读指令被处理器执行时实现如下步骤:对真实图像样本进行特征编码提取,得到所述真实图像样本的特征编码向量;通过图像重建网络,基于所述特征编码向量进行图像重建得到第一图像,基于所述真实图像样本的第一隐层向量进行图像重建得到第二图像;通过图像判别网络对所述真实图像样本、所述第一图像和所述第二图像进行图像判别,并根据图像判别结果对所述图像生成网络进行优化。
- 一种计算机可读存储介质,所述计算机可读存储介质存储有计算机可读指令,其中,所述计算机可读指令被处理器执行时实现如下步骤:获取待重建图像的第二隐层向量;通过训练后的图像重建网络,基于所述第二隐层向量对所述待重建图像进行图像重建。
Priority Applications (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US17/791,099 US20230032472A1 (en) | 2020-03-17 | 2020-03-17 | Method and apparatus for reconstructing medical image |
PCT/CN2020/079678 WO2021184195A1 (zh) | 2020-03-17 | 2020-03-17 | 医学图像重建方法、医学图像重建网络训练方法和装置 |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
PCT/CN2020/079678 WO2021184195A1 (zh) | 2020-03-17 | 2020-03-17 | 医学图像重建方法、医学图像重建网络训练方法和装置 |
Publications (1)
Publication Number | Publication Date |
---|---|
WO2021184195A1 true WO2021184195A1 (zh) | 2021-09-23 |
Family
ID=77772180
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
PCT/CN2020/079678 WO2021184195A1 (zh) | 2020-03-17 | 2020-03-17 | 医学图像重建方法、医学图像重建网络训练方法和装置 |
Country Status (2)
Country | Link |
---|---|
US (1) | US20230032472A1 (zh) |
WO (1) | WO2021184195A1 (zh) |
Cited By (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN115393534A (zh) * | 2022-10-31 | 2022-11-25 | 深圳市宝润科技有限公司 | 基于深度学习的锥束三维dr重建的方法及系统 |
CN117993500A (zh) * | 2024-04-07 | 2024-05-07 | 江西为易科技有限公司 | 基于人工智能的医学教学数据管理方法及系统 |
Families Citing this family (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN116385330B (zh) * | 2023-06-06 | 2023-09-15 | 之江实验室 | 一种利用图知识引导的多模态医学影像生成方法和装置 |
Citations (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN108537743A (zh) * | 2018-03-13 | 2018-09-14 | 杭州电子科技大学 | 一种基于生成对抗网络的面部图像增强方法 |
CN109559358A (zh) * | 2018-10-22 | 2019-04-02 | 天津大学 | 一种基于卷积自编码的图像样本升采样方法 |
CN109685863A (zh) * | 2018-12-11 | 2019-04-26 | 帝工(杭州)科技产业有限公司 | 一种重建医学乳房图像的方法 |
CN110148194A (zh) * | 2019-05-07 | 2019-08-20 | 北京航空航天大学 | 图像重建方法和装置 |
WO2019169594A1 (en) * | 2018-03-08 | 2019-09-12 | Intel Corporation | Methods and apparatus to generate three-dimensional (3d) model for 3d scene reconstruction |
CN110490807A (zh) * | 2019-08-27 | 2019-11-22 | 中国人民公安大学 | 图像重建方法、装置及存储介质 |
Family Cites Families (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US10607319B2 (en) * | 2017-04-06 | 2020-03-31 | Pixar | Denoising monte carlo renderings using progressive neural networks |
US11257259B2 (en) * | 2017-08-15 | 2022-02-22 | Siemens Healthcare Gmbh | Topogram prediction from surface data in medical imaging |
US10679351B2 (en) * | 2017-08-18 | 2020-06-09 | Samsung Electronics Co., Ltd. | System and method for semantic segmentation of images |
US11508169B2 (en) * | 2020-01-08 | 2022-11-22 | Palo Alto Research Center Incorporated | System and method for synthetic image generation with localized editing |
-
2020
- 2020-03-17 US US17/791,099 patent/US20230032472A1/en active Pending
- 2020-03-17 WO PCT/CN2020/079678 patent/WO2021184195A1/zh active Application Filing
Patent Citations (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
WO2019169594A1 (en) * | 2018-03-08 | 2019-09-12 | Intel Corporation | Methods and apparatus to generate three-dimensional (3d) model for 3d scene reconstruction |
CN108537743A (zh) * | 2018-03-13 | 2018-09-14 | 杭州电子科技大学 | 一种基于生成对抗网络的面部图像增强方法 |
CN109559358A (zh) * | 2018-10-22 | 2019-04-02 | 天津大学 | 一种基于卷积自编码的图像样本升采样方法 |
CN109685863A (zh) * | 2018-12-11 | 2019-04-26 | 帝工(杭州)科技产业有限公司 | 一种重建医学乳房图像的方法 |
CN110148194A (zh) * | 2019-05-07 | 2019-08-20 | 北京航空航天大学 | 图像重建方法和装置 |
CN110490807A (zh) * | 2019-08-27 | 2019-11-22 | 中国人民公安大学 | 图像重建方法、装置及存储介质 |
Cited By (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN115393534A (zh) * | 2022-10-31 | 2022-11-25 | 深圳市宝润科技有限公司 | 基于深度学习的锥束三维dr重建的方法及系统 |
CN117993500A (zh) * | 2024-04-07 | 2024-05-07 | 江西为易科技有限公司 | 基于人工智能的医学教学数据管理方法及系统 |
Also Published As
Publication number | Publication date |
---|---|
US20230032472A1 (en) | 2023-02-02 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN111462264B (zh) | 医学图像重建方法、医学图像重建网络训练方法和装置 | |
Biffi et al. | Explainable anatomical shape analysis through deep hierarchical generative models | |
WO2021184195A1 (zh) | 医学图像重建方法、医学图像重建网络训练方法和装置 | |
US20230342918A1 (en) | Image-driven brain atlas construction method, apparatus, device and storage medium | |
WO2021186592A1 (ja) | 診断支援装置及びモデル生成装置 | |
CN112435341B (zh) | 三维重建网络的训练方法及装置、三维重建方法及装置 | |
CN112470190A (zh) | 用于改进低剂量体积对比增强mri的系统和方法 | |
Zhan et al. | LR-cGAN: Latent representation based conditional generative adversarial network for multi-modality MRI synthesis | |
CN112949654A (zh) | 图像检测方法及相关装置、设备 | |
CN118247284B (zh) | 图像处理模型的训练方法、图像处理方法 | |
CN114155232A (zh) | 颅内出血区域检测方法、装置、计算机设备及存储介质 | |
CN115272295A (zh) | 基于时域-空域联合状态的动态脑功能网络分析方法及系统 | |
Yerukalareddy et al. | Brain tumor classification based on mr images using GAN as a pre-trained model | |
CN117274599A (zh) | 一种基于组合双任务自编码器的脑磁共振分割方法及系统 | |
CN113724185B (zh) | 用于图像分类的模型处理方法、装置及存储介质 | |
Tiago et al. | A domain translation framework with an adversarial denoising diffusion model to generate synthetic datasets of echocardiography images | |
CN113850796A (zh) | 基于ct数据的肺部疾病识别方法及装置、介质和电子设备 | |
CN118037615A (zh) | 一种肿瘤分割引导的磁共振图像翻译方法、系统、设备及介质 | |
CN111383217B (zh) | 大脑成瘾性状评估的可视化方法、装置及介质 | |
CN117333371A (zh) | 基于生成对抗网络的自适应域医学图像跨模态转换方法 | |
CN114463320B (zh) | 一种磁共振成像脑胶质瘤idh基因预测方法及系统 | |
CN115691770A (zh) | 基于条件分数的跨模态医学影像补全方法、装置及设备 | |
KR102400568B1 (ko) | 인코더를 이용한 이미지의 특이 영역 분석 방법 및 장치 | |
Shihabudeen et al. | Autoencoder Network based CT and MRI Medical Image Fusion | |
CN113223104B (zh) | 一种基于因果关系的心脏mr图像插补方法及系统 |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
121 | Ep: the epo has been informed by wipo that ep was designated in this application |
Ref document number: 20926334 Country of ref document: EP Kind code of ref document: A1 |
|
NENP | Non-entry into the national phase |
Ref country code: DE |
|
122 | Ep: pct application non-entry in european phase |
Ref document number: 20926334 Country of ref document: EP Kind code of ref document: A1 |
|
122 | Ep: pct application non-entry in european phase |
Ref document number: 20926334 Country of ref document: EP Kind code of ref document: A1 |
|
122 | Ep: pct application non-entry in european phase |
Ref document number: 20926334 Country of ref document: EP Kind code of ref document: A1 |
|
32PN | Ep: public notification in the ep bulletin as address of the adressee cannot be established |
Free format text: NOTING OF LOSS OF RIGHTS PURSUANT TO RULE 112(1) EPC (EPO FORM 1205A DATED 04-07-2023) |