CN112102385A - Multi-modal liver magnetic resonance image registration system based on deep learning - Google Patents
Multi-modal liver magnetic resonance image registration system based on deep learning Download PDFInfo
- Publication number
- CN112102385A CN112102385A CN202010845127.9A CN202010845127A CN112102385A CN 112102385 A CN112102385 A CN 112102385A CN 202010845127 A CN202010845127 A CN 202010845127A CN 112102385 A CN112102385 A CN 112102385A
- Authority
- CN
- China
- Prior art keywords
- image
- registration
- network
- training
- gan
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Granted
Links
- 210000004185 liver Anatomy 0.000 title claims abstract description 17
- 238000013135 deep learning Methods 0.000 title claims abstract description 9
- 238000000034 method Methods 0.000 claims abstract description 69
- 238000012549 training Methods 0.000 claims abstract description 37
- 230000006870 function Effects 0.000 claims abstract description 17
- 238000007781 pre-processing Methods 0.000 claims abstract description 17
- 230000008569 process Effects 0.000 claims abstract description 16
- 238000012216 screening Methods 0.000 claims abstract description 14
- 238000013507 mapping Methods 0.000 claims abstract description 5
- 230000009466 transformation Effects 0.000 claims description 22
- PXFBZOLANLWPMH-UHFFFAOYSA-N 16-Epiaffinine Natural products C1C(C2=CC=CC=C2N2)=C2C(=O)CC2C(=CC)CN(C)C1C2CO PXFBZOLANLWPMH-UHFFFAOYSA-N 0.000 claims description 20
- 230000004913 activation Effects 0.000 claims description 13
- 239000011159 matrix material Substances 0.000 claims description 11
- 238000004088 simulation Methods 0.000 claims description 10
- 230000000694 effects Effects 0.000 claims description 6
- 210000000056 organ Anatomy 0.000 claims description 6
- 238000012545 processing Methods 0.000 claims description 5
- 238000012952 Resampling Methods 0.000 claims description 3
- 238000002939 conjugate gradient method Methods 0.000 claims description 2
- 238000007796 conventional method Methods 0.000 abstract description 2
- 230000003902 lesion Effects 0.000 description 11
- 238000005070 sampling Methods 0.000 description 10
- 241000257303 Hymenoptera Species 0.000 description 9
- 238000002597 diffusion-weighted imaging Methods 0.000 description 9
- 206010028980 Neoplasm Diseases 0.000 description 6
- 238000010586 diagram Methods 0.000 description 6
- 238000004458 analytical method Methods 0.000 description 5
- 238000013535 dynamic contrast enhanced MRI Methods 0.000 description 3
- QVGXLLKOCUKJST-UHFFFAOYSA-N atomic oxygen Chemical compound [O] QVGXLLKOCUKJST-UHFFFAOYSA-N 0.000 description 2
- 239000008280 blood Substances 0.000 description 2
- 210000004369 blood Anatomy 0.000 description 2
- 201000011510 cancer Diseases 0.000 description 2
- 238000003745 diagnosis Methods 0.000 description 2
- 238000000605 extraction Methods 0.000 description 2
- 238000002599 functional magnetic resonance imaging Methods 0.000 description 2
- 238000002595 magnetic resonance imaging Methods 0.000 description 2
- 238000010606 normalization Methods 0.000 description 2
- 229910052760 oxygen Inorganic materials 0.000 description 2
- 239000001301 oxygen Substances 0.000 description 2
- 238000004445 quantitative analysis Methods 0.000 description 2
- INGWEZCOABYORO-UHFFFAOYSA-N 2-(furan-2-yl)-7-methyl-1h-1,8-naphthyridin-4-one Chemical compound N=1C2=NC(C)=CC=C2C(O)=CC=1C1=CC=CO1 INGWEZCOABYORO-UHFFFAOYSA-N 0.000 description 1
- 238000012935 Averaging Methods 0.000 description 1
- 210000003484 anatomy Anatomy 0.000 description 1
- 230000036770 blood supply Effects 0.000 description 1
- 238000004364 calculation method Methods 0.000 description 1
- 230000008859 change Effects 0.000 description 1
- 238000006243 chemical reaction Methods 0.000 description 1
- 239000002872 contrast media Substances 0.000 description 1
- 108010002255 deoxyhemoglobin Proteins 0.000 description 1
- 230000001419 dependent effect Effects 0.000 description 1
- 238000006073 displacement reaction Methods 0.000 description 1
- 230000002708 enhancing effect Effects 0.000 description 1
- 238000011156 evaluation Methods 0.000 description 1
- PCHJSUWPFVWCPO-UHFFFAOYSA-N gold Chemical compound [Au] PCHJSUWPFVWCPO-UHFFFAOYSA-N 0.000 description 1
- 230000000004 hemodynamic effect Effects 0.000 description 1
- 238000002372 labelling Methods 0.000 description 1
- 201000007270 liver cancer Diseases 0.000 description 1
- 210000005229 liver cell Anatomy 0.000 description 1
- 208000019423 liver disease Diseases 0.000 description 1
- 208000014018 liver neoplasm Diseases 0.000 description 1
- 238000012986 modification Methods 0.000 description 1
- 230000004048 modification Effects 0.000 description 1
- 230000000877 morphologic effect Effects 0.000 description 1
- 238000005457 optimization Methods 0.000 description 1
- 230000008520 organization Effects 0.000 description 1
- 230000036285 pathological change Effects 0.000 description 1
- 231100000915 pathological change Toxicity 0.000 description 1
- 230000035699 permeability Effects 0.000 description 1
- 230000001766 physiological effect Effects 0.000 description 1
- 238000011524 similarity measure Methods 0.000 description 1
- 238000013519 translation Methods 0.000 description 1
- 238000010200 validation analysis Methods 0.000 description 1
- 230000000007 visual effect Effects 0.000 description 1
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
- G06T7/30—Determination of transform parameters for the alignment of images, i.e. image registration
- G06T7/38—Registration of image sequences
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T3/00—Geometric image transformations in the plane of the image
- G06T3/14—Transformations for image registration, e.g. adjusting or mapping for alignment of images
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T3/00—Geometric image transformations in the plane of the image
- G06T3/14—Transformations for image registration, e.g. adjusting or mapping for alignment of images
- G06T3/147—Transformations for image registration, e.g. adjusting or mapping for alignment of images using affine transformations
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/10—Image acquisition modality
- G06T2207/10072—Tomographic images
- G06T2207/10088—Magnetic resonance imaging [MRI]
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/20—Special algorithmic details
- G06T2207/20081—Training; Learning
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/20—Special algorithmic details
- G06T2207/20084—Artificial neural networks [ANN]
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/30—Subject of image; Context of image processing
- G06T2207/30004—Biomedical image processing
- G06T2207/30056—Liver; Hepatic
Landscapes
- Engineering & Computer Science (AREA)
- Physics & Mathematics (AREA)
- General Physics & Mathematics (AREA)
- Theoretical Computer Science (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Image Analysis (AREA)
- Image Processing (AREA)
Abstract
The invention belongs to the technical field of magnetic resonance image registration, and particularly relates to a multi-modal liver magnetic resonance image registration system based on deep learning. The system of the invention comprises: the image preprocessing module is used for preprocessing a source image and a target image and carrying out pre-registration by using a traditional method; the training sample screening module is used for screening out a sample for GAN network training from the sample pre-registered by the conventional method; the GAN network module is used for registering the multi-mode liver magnetic resonance image; the GAN network includes a discriminator network and a generator network; according to the invention, the GAN network is pre-trained to learn the mapping relation between image domains with different contrast ratios, so that the function of converting one contrast image into another contrast image is realized; and then, in the registration process, the trained GAN network is utilized to assist the traditional registration method, so that errors caused by modal differences in registration are eliminated, the registration time is shortened, and the registration accuracy is improved.
Description
Technical Field
The invention belongs to the technical field of magnetic resonance image registration, and particularly relates to a liver magnetic resonance image registration method system with different contrasts, namely different modalities.
Background
Liver cancer is a common tumor in China, is the sixth most common cancer in the world, is also the cancer with the second highest Asian fatality rate, and has important significance in early and accurate diagnosis of liver lesions. In the process of diagnosing liver lesions, comprehensive and comprehensive judgment is generally performed according to the characteristics and advantages of different images by combining image information of multiple modalities. For example, the magnetic resonance Dynamic enhanced MRI (Dynamic Contrast enhanced MRI, DCE-MRI) can carry out quantitative analysis on the permeability of tumor vessels, not only can provide the morphological characteristics of liver lesions, but also can reflect the blood supply state of the lesions; diffusion Weighted Imaging (DWI) can carry out quantitative analysis on the characteristics and the tissue structure of liver cells, provides more information than conventional MRI, and has great significance for determining pathological changes and physiological activities of lesion parts; in recent years, blood oxygen level dependent functional magnetic resonance imaging (BOLD-fMRI) with deoxyhemoglobin as an endogenous contrast agent can reflect changes of hemodynamics and blood oxygen content, and can be used for diagnosis and severity evaluation of liver diseases. The purpose of image registration is to align images of different modalities spatially or temporally to more accurately locate the lesion.
The traditional multi-modal registration method mostly adopts a linear transformation model and a nonlinear transformation model. The linear transformation model comprises rigid registration, similarity registration and affine registration through adjustment of a linear change matrix and a translation vector, but the method cannot take local differences between a source image and a target image into account. In nonlinear transformation, a deformation field-based spline function, such as a thin-plate spline method and a B-spline method, is commonly used, and an optimal smooth displacement field is obtained by solving a mapping relation between corresponding control points of a target image and a reference image, however, as the number of the control points increases, a coefficient matrix becomes larger and larger, and calculation accuracy and stability are reduced.
For multi-modality magnetic resonance image registration, the registration difficulty mainly lies in contrast difference between different modality images, and difference in anatomical structure of corresponding organs is caused due to different time points for acquiring different modality images, which both causes great difficulty in extracting feature space. Therefore, the multi-modal registration directly using the conventional registration method usually cannot achieve higher accuracy, and the speed is still to be improved.
Disclosure of Invention
The invention aims to provide a multi-modal liver magnetic resonance image registration system which can reduce the difficulty of feature extraction between multi-modal images and realize more accurate and faster registration.
According to the multi-modal liver magnetic resonance image registration system provided by the invention, the pre-trained GAN network is utilized to convert the source modal image into the target modal image, so that the difficulty of extracting the feature space is reduced, and the auxiliary traditional registration method is enabled to carry out more accurate and faster registration.
The multi-mode liver magnetic resonance image registration system provided by the invention is based on deep learning and specifically comprises the following modules.
The image preprocessing module is used for preprocessing a source image and a target image and carrying out pre-registration by using a traditional method, namely carrying out rigid registration, affine registration and nonlinear registration in sequence;
in the image preprocessing module, the preprocessing modes mainly include three types:
(1) because some images have the problem of incomplete images of the first layer and the last layer in the acquisition process, 3-10 layers of the beginning and the end of each 3D image sample are removed;
(2) re-sampling all 2D images fed into the network to a resolution of 128 x 128 to 1024 x 1024;
(3) and normalizing the 2D image according to the value range of the pixel points of the image pair and the output range of the activation function of the network structure, and normalizing the pixel value to be in the range of [0,1] or [ -1,0 ].
And the training sample screening module is used for screening the sample pre-registered by the conventional method, screening out a sample with good hierarchical correspondence and clear organ tissues as a training sample of the GAN network.
The GAN network module is trained and used for registering the multi-mode liver magnetic resonance image; the GAN network is built by adopting but not limited to a pix2pix architecture; the constructed GAN network mainly comprises a discriminator network and a generator network; wherein:
the generator network architecture, as shown in fig. 4, adopts a structure similar to U-Net, and the generator network includes: 5-12 convolutional layers, 5-15 deconvolution layers; each convolutional layer selects a suitable activation function, including but not limited to Relu function, tanh function, etc.
The identification network structure adopts a patch-GAN structure as shown in FIG. 5, that is, a feature matrix with a small size is obtained after a piece of image is subjected to multilayer convolution, probability values of each feature value in the matrix between 0 and 1 are obtained after sigmoid activation function processing, and the average value of the probability values is the identification result output by the identification network; the authentication network includes: convolutional layers, 5-10 strided convolutional layers, and final convolutional layers.
The system of the invention has the following working procedures:
firstly, preprocessing a source image and a target image by an image preprocessing module, and performing pre-registration by using a traditional method, namely sequentially performing rigid registration, affine registration and nonlinear registration;
secondly, a training sample screening module screens the sample pre-registered by the traditional method, and screens out a sample with good hierarchical correspondence and clear organ tissues as a training sample of the GAN network; the screening can adopt manual screening;
thirdly, training the GAN network by using the screened training samples;
in the training process, the training results are compared, and the image preprocessing method, the network training method and the data enhancement method are adjusted to achieve the best training effect.
The image preprocessing adjusting method comprises three modes:
(1) because some images have the problem of incomplete images of the first layer and the last layer in the acquisition process, 3-10 layers of the beginning and the end of each 3D image sample are removed;
(2) resampling all 2D images fed into the network to a resolution of 128 x 128 to 1024 x 1024;
(3) and normalizing the 2D image according to the value range of the pixel points of the image pair and the output range of the activation function of the network structure, and normalizing the pixel value to be in the range of [0,1] or [ -1,0 ].
The method for enhancing the adjustment data specifically uses, but is not limited to, methods of turning over, rotating, affine transformation, noise increasing, cutting, color disturbance, scaling and the like, and specifically enhances the following modes:
turning over: the folding mode comprises three modes of horizontal folding, vertical folding and diagonal folding, and one of the three folding modes is randomly selected for data enhancement;
rotating: selecting an angle threshold value theta (the range of theta is 120 degrees to 180 degrees), and randomly rotating at an angle from-theta degrees to theta degrees;
affine transformation: respectively randomly taking four points in a circular area with four end points of the image as the circle center and n pixels in diameter as transformation vertexes of affine transformation to carry out affine transformation;
scaling transformation: the image scaling is performed by taking values randomly within a certain percentage range (typically 10% -15%) of the image size.
The network training method is adjusted by using, but not limited to, a conjugate gradient method, an alternating direction multiplier method, a random gradient method and other training methods, setting a proper batch size and a proper learning rate, and performing network convergence after training for a certain number of rounds.
(IV) performing auxiliary registration of the GAN network, wherein the specific process is as follows:
(1) inputting a three-dimensional source modal image slice to be registered into a trained GAN network in the form of a two-dimensional image, outputting a corresponding two-dimensional target modal simulation image, and integrating the two-dimensional simulation image into a three-dimensional simulation image;
(2) performing traditional registration on the target modal simulation image obtained in the last step and a target modal image to be registered, namely sequentially performing rigid registration, affine registration and nonlinear registration to obtain a corresponding pixel mapping relation, namely a deformation field;
(3) and fusing the deformation field obtained in the last step with the source image to be registered to complete the GAN network auxiliary registration.
Compared with the prior art, the system has the following advantages:
(1) a novel auxiliary registration system based on deep learning is provided. The traditional multi-modal registration method uses mutual information as a loss function, and whether the mutual information can realize accurate registration depends on whether a good initialization process exists or not, namely whether a sufficient number of correct pairing relations can be known at the beginning of optimization. According to the invention, the target modal image is generated through the GAN network, and a good initialization process is provided for the traditional registration method through modal conversion, so that the registration precision is improved;
(2) the registration time of the conventional registration method is shortened. Because a GAN network is used for converting a source mode into a target mode and inputting the target mode into the traditional registration method, image features are easier to extract, so that the registration time is shortened, and the average registration time is shortened by 14 seconds through statistics of the registration time of a number of samples;
(3) the GAN network has excellent learning effect on images of various modes, is not limited to DWI, DCE, fMRI, even CT and other images, so that the GAN network with different functions can be obtained by only feeding the images of the corresponding modes for training so as to assist in the registration among different modes;
(4) network training data is readily available. The GAN network has low requirements on training data, and only needs that the input source image has good level correspondence with the target image and clear organization outline. By changing the structure of the GAN network (e.g., Cycle-GAN), training can be accomplished even with unpaired datasets.
Drawings
Fig. 1 is a diagram of a GAN training process.
Fig. 2 is a GAN assisted registration flow diagram.
Fig. 3 shows the result of the conventional pre-registration.
Fig. 4 is a diagram of a generator network structure in the present invention.
Fig. 5 is a diagram showing the structure of a discriminator network in the present invention.
Fig. 6 is a diagram illustrating the effect of generating a simulated DCE for a GAN network.
Fig. 7 shows the generation of the deformation field after ANTs registration.
Fig. 8 is a comparison of GAN assisted registration and ANTs registration results. Wherein (a), (b) and (c) are three cases.
Detailed description of the invention
In the following, a batch of samples of 100 patients with lesion labeling is taken as an example for analysis, wherein the samples comprise DCE-MRI images and DWI images, and the invention is further explained by combining the drawings. Several variations and modifications of the following process are within the scope of the invention without departing from the spirit of the invention. The specific working flow of the system of the invention is as follows.
Firstly, preprocessing raw DWI and DCE data of a training sample, and performing pre-registration by using a traditional method, namely performing rigid registration, affine registration and nonlinear registration in sequence.
And (II) screening the samples subjected to the traditional pre-registration (manual screening), wherein the screened samples have good hierarchical correspondence, for example, as shown in fig. 3, the samples with clear organ tissues are used as training samples of the GAN network. The samples were finally divided into two groups, 92% of the samples for network training and 8% for final outcome validation.
Thirdly, building a GAN network based on a pix2pix architecture, wherein a generator network architecture of the GAN network is shown in fig. 4, and a structure similar to U-Net is adopted, and the structure comprises two stages of down sampling and up sampling; in a down-sampling stage, image feature extraction is carried out through the convolution layer; in the up-sampling stage, the network uses 'jump connection', namely, the dropout output of the previous layer is connected with the output of the symmetrical down-sampling convolution layer of the previous layer to be used as the input of the anti-convolution layer, so that the information extracted by down-sampling can be continuously memorized again when up-sampling is carried out, and the generated image can keep some information of the original image as much as possible. The network down-sampling stage comprises 3 step convolutional layers with the step length of 2 and convolutional kernels of 64, 128 and 256 respectively, the features of the input image with the step length of 256,256 and 1 are extracted to be (32, 32 and 256), and the features are continuously extracted to be (1, 1 and 512) through 5 convolutional layers with the step length of 2 and the convolutional kernels of 512; in the up-sampling stage, the feature matrix is up-sampled to (16, 256) by 4 deconvolution layers with the step size of 2 and the convolution kernel of 512, and then restored to (256, 1) by 4 step deconvolution layers with the step size of 2 and the convolution kernel of 256,128,64,1, that is, an image with the same size as the original image is output. Except the last convolution layer, each convolution layer contains real rule regularization and ReLU activation, and the final activation is carried out at the tail end of the generator by adopting a tanh activation function. In the upsampling phase, the dropout ratio for each deconvolution layer is set to 0.5.
The authentication network structure is shown in fig. 5. In order to make more accurate judgment on the local part of an image, a patch-GAN structure is adopted in the pix2pix identification network, namely, a feature matrix with smaller size is obtained after multi-layer convolution of one image, probability value between 0 and 1 corresponding to each feature value in the matrix is obtained after sigmoid activation function processing, and the average value of the probability value is the identification result output by the identification network. The specific implementation process is as follows: the original DWI image and the original DCE or the simulated DCE image are input into a discriminator network in pairs, firstly, 4 layers of step-by-step convolutional layers with the step length of 2 and convolutional kernels of 64, 128, 256 and 512 are processed to obtain a characteristic diagram of (16, 16, 512), and then, the convolutional layers with the step length of 1 and the convolutional kernels of 1 are processed to obtain a characteristic matrix of (16, 16, 1). And then, carrying out sigmoid activation function processing on each characteristic value in the characteristic matrix to obtain a corresponding probability value, and averaging the probability values to obtain the final output of the identification network.
Through multiple adjustments, the final pretreatment modes adopted by the invention include the following three types:
(1) because some images have the problem of incomplete images of the head layer and the tail layer in the acquisition process, the initial 3 layers and the tail 3 layers of each 3D image sample are removed;
(2) resampling all 2D images fed into the network to 256 x 256;
(3) the pixel point range of each picture is [0,255], and because the generator finally adopts the tanh activation function with the output of (-1, 1), each 2D image is subjected to normalization processing, and the pixel value is normalized to be in the range of [ -1,1 ].
Adopting on-line data enhancement, respectively adopting four methods of turning, rotating, affine transformation and zooming, and the specific parameters are as follows:
turning over: the turning modes comprise three types, namely horizontal turning, vertical turning and diagonal turning, and one of the three turning modes is randomly selected for data enhancement;
rotating: -180 to 180 degrees random angular rotation;
perspective transformation: taking four end points of the image as circle centers respectively, randomly taking four points in a circular area with the diameter of 22 pixels, namely within 5 percent of the total pixels, and taking the four points as transformation vertexes of affine transformation to perform affine transformation;
scaling transformation: randomly taking values in the range of 10% -15% of the image size, and zooming the image.
The network was trained in a Tesla V100 GPU with a batch (batch size) set to 80, a learning rate of 2e-4, and a training round number of 20000, with the best results being obtained, as shown in fig. 6.
And (V) cutting 7 three-dimensional DWI images to be registered into slices, inputting the slices into a GAN network in the form of two-dimensional images, outputting corresponding two-dimensional simulation DCE mode images, integrating the two-dimensional images into a three-dimensional simulation DCE image, performing three-dimensional pixel normalization, and eliminating pixel jump among levels.
And (VI) registering the simulated DCE image and the target DCE image by using a traditional method, namely sequentially carrying out rigid registration, affine registration and nonlinear registration to obtain a corresponding pixel mapping relation, namely a deformation field, as shown in FIG. 7.
And (seventhly), fusing the deformation field obtained in the last step with the DWI image to be registered, and finishing the GAN network auxiliary registration process.
(eighth), only 7 samples are registered by using the traditional method, and are visually compared with the registration result assisted by using the GAN network, and 3 samples with obvious effect in 7 cases are taken for display, as shown in fig. 8. In the result graph, a Mask layer is a lesion position marked on DCE influence by a doctor, GAN + ANTs and ANTs respectively represent the result of using GAN auxiliary registration and using only a traditional method for registration, and a yellow frame region is a region where a lesion is located. Observing the registration results of the three samples, compared with the registration method only using ANTs, the registration method of GAN + ANTs can register the focus on the DWI into the same layer, and the focus position marked on the DCE image by a doctor are overlapped by more than 70%; however, no lesion can be observed in the result of ANTs registration, and the levels cannot be accurately corresponded.
(nine) analyzing results: tumor positions of DWI images registered by the two methods are respectively marked, and compared with the tumor positions marked by doctors, similarity measures are respectively calculated, and the Dice index is adopted for similarity analysis in the embodiment. And calculating the time consumed by the two registration methods, and evaluating from the speed perspective. The results of the specific analyses of 7 samples are shown in Table 1.
As can be seen from table 1, in the 5 samples (6 lesions) with non-zero Dice values, the GAN network-based auxiliary registration method all optimizes the registration result to a certain extent, wherein the Dice value increase range is between 0.04 and 0.45; and for three samples (with the Dice value of 0) which cannot be registered with the focuses only by using ANTs, the registered focus area can be coincided with a small part of the gold standard. In the registration time, 3 samples in 5 samples are improved in registration time, the mean registration time based on the GAN network from the overall analysis is 742.88 seconds, and the mean registration time of the traditional registration method is 756.16 seconds, which is slightly improved.
Therefore, from the visual image and Dice index analysis, the images generated by the GAN network and the DCE are jointly used as ANTs registration input, the registration effect can be achieved, the accuracy of most cases can be improved, and the registration time is obviously shortened.
TABLE 1 comparison of two method dice indexes with registration time in six samples
Claims (4)
1. A multi-modal liver magnetic resonance image registration system based on deep learning is characterized by specifically comprising the following modules:
the image preprocessing module is used for preprocessing a source image and a target image and carrying out pre-registration by using a traditional method, namely carrying out rigid registration, affine registration and nonlinear registration in sequence;
the training sample screening module is used for screening the sample pre-registered by the traditional method, screening out a sample with good hierarchical correspondence and clear organ tissues as a training sample of the GAN network;
the GAN network module is trained and used for registering the multi-mode liver magnetic resonance image; the GAN network is built by adopting a pix2pix architecture; the constructed GAN network comprises a discriminator network and a generator network; wherein:
the generator network adopts a structure similar to U-Net, and comprises: 5-12 convolutional layers, 5-15 deconvolution layers; selecting a proper activation function for each convolution layer;
the identification network adopts a patch-GAN structure, namely, an image is convoluted by multiple layers to obtain a feature matrix with smaller size, probability values of each feature value in the matrix between 0 and 1 are obtained after sigmoid activation function processing, and the average value of the probability values is the judgment result output by the identification network; the authentication network includes: convolutional layers, 5-10 strided convolutional layers, and final convolutional layers;
the system has the following working flows:
firstly, preprocessing a source image and a target image by an image preprocessing module, and performing pre-registration by using a traditional method, namely sequentially performing rigid registration, affine registration and nonlinear registration;
secondly, a training sample screening module screens the sample pre-registered by the traditional method, and screens out a sample with good hierarchical correspondence and clear organ tissues as a training sample of the GAN network;
thirdly, training the GAN network by using the screened training samples;
in the training process, the training results are compared, and an image preprocessing method, a network training method and a data enhancement method are adjusted to achieve the best training effect;
(IV) performing auxiliary registration of the GAN network, wherein the specific process is as follows:
(1) inputting a three-dimensional source modal image slice to be registered into a trained GAN network in the form of a two-dimensional image, outputting a corresponding two-dimensional target modal simulation image, and integrating the two-dimensional simulation image into a three-dimensional simulation image;
(2) performing traditional registration on the target modal simulation image obtained in the last step and a target modal image to be registered, namely sequentially performing rigid registration, affine registration and nonlinear registration to obtain a corresponding pixel mapping relation, namely a deformation field;
(3) and fusing the deformation field obtained in the last step with the source image to be registered to complete the GAN network auxiliary registration.
2. The deep learning-based multi-modal liver magnetic resonance image registration system according to claim 1, wherein the image preprocessing method is adjusted in three ways:
(1) for the condition that the first and last layers of images are incomplete in the acquisition process of some images, removing 3-10 layers of the beginning and the end of each 3D image sample;
(2) resampling to a resolution of 128 x 128 to 1024 x 1024 for all 2D images fed into the network;
(3) and normalizing the 2D image according to the value range of the pixel points of the image pair and the output range of the activation function of the network result, and normalizing the pixel value to be in the range of [0,1] or [ -1,0 ].
3. The deep learning based multi-modal liver magnetic resonance image registration system according to claim 1, wherein the adjustment data enhancement method specifically comprises: turning over, rotating, affine transformation, noise increasing, cutting, color disturbance and scaling;
turning over: the method comprises three modes of horizontal turning, vertical turning and diagonal turning, and one of the three turning modes is randomly selected for data enhancement;
rotating: selecting an angle threshold value theta, and rotating at a random angle from-theta degrees to theta degrees; θ ranges between 120 ° and 180 °;
affine transformation: respectively randomly taking four points in a circular area with four end points of the image as the circle center and n pixels in diameter as transformation vertexes of affine transformation to carry out affine transformation;
scaling transformation: randomly taking a value within a certain percentage range of the image size, and zooming the image; the percentage ranges from 10% to 15%.
4. The deep learning-based multi-modal liver magnetic resonance image registration system according to claim 1, wherein the network training method is adjusted, specifically, a conjugate gradient method, an alternating direction multiplier method, a random gradient method, etc. are selected as the network training methods, an appropriate batch size and learning rate are set, and the network converges after a certain number of training rounds.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202010845127.9A CN112102385B (en) | 2020-08-20 | 2020-08-20 | Multi-modal liver magnetic resonance image registration system based on deep learning |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202010845127.9A CN112102385B (en) | 2020-08-20 | 2020-08-20 | Multi-modal liver magnetic resonance image registration system based on deep learning |
Publications (2)
Publication Number | Publication Date |
---|---|
CN112102385A true CN112102385A (en) | 2020-12-18 |
CN112102385B CN112102385B (en) | 2023-02-10 |
Family
ID=73753352
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN202010845127.9A Active CN112102385B (en) | 2020-08-20 | 2020-08-20 | Multi-modal liver magnetic resonance image registration system based on deep learning |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN112102385B (en) |
Cited By (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN112598718A (en) * | 2020-12-31 | 2021-04-02 | 北京深睿博联科技有限责任公司 | Unsupervised multi-view multi-mode intelligent glasses image registration method and device |
CN112907639A (en) * | 2021-01-20 | 2021-06-04 | 云南电网有限责任公司电力科学研究院 | X-ray image registration method for power equipment |
CN113487657A (en) * | 2021-07-29 | 2021-10-08 | 广州柏视医疗科技有限公司 | Deep learning-based mode conversion method |
WO2023283795A1 (en) * | 2021-07-12 | 2023-01-19 | Shanghaitech University | Method for high-resolution image reconstruction |
CN116402865A (en) * | 2023-06-06 | 2023-07-07 | 之江实验室 | Multi-mode image registration method, device and medium using diffusion model |
EP4394796A1 (en) * | 2022-12-29 | 2024-07-03 | Tempus AI, Inc. | Multimodal cardio disease state predictions combining electrocardiogram, echocardiogram, clinical and demographical information relating to a patient |
Citations (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20180075581A1 (en) * | 2016-09-15 | 2018-03-15 | Twitter, Inc. | Super resolution using a generative adversarial network |
CN109754447A (en) * | 2018-12-28 | 2019-05-14 | 上海联影智能医疗科技有限公司 | Image generating method, device, equipment and storage medium |
CN110021037A (en) * | 2019-04-17 | 2019-07-16 | 南昌航空大学 | A kind of image non-rigid registration method and system based on generation confrontation network |
CN110298871A (en) * | 2019-06-10 | 2019-10-01 | 东软医疗系统股份有限公司 | Method for registering images and device |
CN111047629A (en) * | 2019-11-04 | 2020-04-21 | 中国科学院深圳先进技术研究院 | Multi-modal image registration method and device, electronic equipment and storage medium |
CN111210465A (en) * | 2019-12-31 | 2020-05-29 | 上海联影智能医疗科技有限公司 | Image registration method and device, computer equipment and readable storage medium |
-
2020
- 2020-08-20 CN CN202010845127.9A patent/CN112102385B/en active Active
Patent Citations (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20180075581A1 (en) * | 2016-09-15 | 2018-03-15 | Twitter, Inc. | Super resolution using a generative adversarial network |
CN109754447A (en) * | 2018-12-28 | 2019-05-14 | 上海联影智能医疗科技有限公司 | Image generating method, device, equipment and storage medium |
CN110021037A (en) * | 2019-04-17 | 2019-07-16 | 南昌航空大学 | A kind of image non-rigid registration method and system based on generation confrontation network |
CN110298871A (en) * | 2019-06-10 | 2019-10-01 | 东软医疗系统股份有限公司 | Method for registering images and device |
CN111047629A (en) * | 2019-11-04 | 2020-04-21 | 中国科学院深圳先进技术研究院 | Multi-modal image registration method and device, electronic equipment and storage medium |
CN111210465A (en) * | 2019-12-31 | 2020-05-29 | 上海联影智能医疗科技有限公司 | Image registration method and device, computer equipment and readable storage medium |
Non-Patent Citations (1)
Title |
---|
陈爱国: "《教育神经学-视野中的体育教育创新》", 30 December 2016 * |
Cited By (10)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN112598718A (en) * | 2020-12-31 | 2021-04-02 | 北京深睿博联科技有限责任公司 | Unsupervised multi-view multi-mode intelligent glasses image registration method and device |
CN112907639A (en) * | 2021-01-20 | 2021-06-04 | 云南电网有限责任公司电力科学研究院 | X-ray image registration method for power equipment |
CN112907639B (en) * | 2021-01-20 | 2024-04-26 | 云南电网有限责任公司电力科学研究院 | Power equipment X-ray image registration method |
WO2023283795A1 (en) * | 2021-07-12 | 2023-01-19 | Shanghaitech University | Method for high-resolution image reconstruction |
CN113487657A (en) * | 2021-07-29 | 2021-10-08 | 广州柏视医疗科技有限公司 | Deep learning-based mode conversion method |
CN113487657B (en) * | 2021-07-29 | 2022-02-01 | 广州柏视医疗科技有限公司 | Deep learning-based mode conversion method |
WO2023005186A1 (en) * | 2021-07-29 | 2023-02-02 | 广州柏视医疗科技有限公司 | Modal transformation method based on deep learning |
EP4394796A1 (en) * | 2022-12-29 | 2024-07-03 | Tempus AI, Inc. | Multimodal cardio disease state predictions combining electrocardiogram, echocardiogram, clinical and demographical information relating to a patient |
CN116402865A (en) * | 2023-06-06 | 2023-07-07 | 之江实验室 | Multi-mode image registration method, device and medium using diffusion model |
CN116402865B (en) * | 2023-06-06 | 2023-09-15 | 之江实验室 | Multi-mode image registration method, device and medium using diffusion model |
Also Published As
Publication number | Publication date |
---|---|
CN112102385B (en) | 2023-02-10 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN112102385B (en) | Multi-modal liver magnetic resonance image registration system based on deep learning | |
CN109166133B (en) | Soft tissue organ image segmentation method based on key point detection and deep learning | |
Mahapatra et al. | Image super-resolution using progressive generative adversarial networks for medical image analysis | |
CN110930416B (en) | MRI image prostate segmentation method based on U-shaped network | |
CN110555835B (en) | Brain slice image region division method and device | |
CN113826143A (en) | Feature point detection | |
CN113239755B (en) | Medical hyperspectral image classification method based on space-spectrum fusion deep learning | |
Mahapatra et al. | Progressive generative adversarial networks for medical image super resolution | |
CN108764342B (en) | Semantic segmentation method for optic discs and optic cups in fundus image | |
He et al. | Few-shot learning for deformable medical image registration with perception-correspondence decoupling and reverse teaching | |
CN111488912B (en) | Laryngeal disease diagnosis system based on deep learning neural network | |
CN112150564B (en) | Medical image fusion algorithm based on deep convolution neural network | |
CN115830016B (en) | Medical image registration model training method and equipment | |
CN112488971A (en) | Medical image fusion method for generating countermeasure network based on spatial attention mechanism and depth convolution | |
CN112329844A (en) | Image object classification method and related device, equipment and storage medium | |
CN108665474B (en) | B-COSFIRE-based retinal vessel segmentation method for fundus image | |
Li et al. | A dense connection encoding–decoding convolutional neural network structure for semantic segmentation of thymoma | |
CN115393239A (en) | Multi-mode fundus image registration and fusion method and system | |
Wu et al. | Automatic symmetry detection from brain MRI based on a 2-channel convolutional neural network | |
Feng et al. | Retinal mosaicking with vascular bifurcations detected on vessel mask by a convolutional network | |
Zhao et al. | Attention residual convolution neural network based on U-net (AttentionResU-Net) for retina vessel segmentation | |
CN117422788B (en) | Method for generating DWI image based on CT brain stem image | |
CN113362360B (en) | Ultrasonic carotid plaque segmentation method based on fluid velocity field | |
Li et al. | Deep attention super-resolution of brain magnetic resonance images acquired under clinical protocols | |
Liu et al. | Weakly-supervised localization and classification of biomarkers in OCT images with integrated reconstruction and attention |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
GR01 | Patent grant | ||
GR01 | Patent grant |