CN117994143A - Multi-mode MR image synthesis method, system, storage medium and equipment - Google Patents
Multi-mode MR image synthesis method, system, storage medium and equipment Download PDFInfo
- Publication number
- CN117994143A CN117994143A CN202410037947.3A CN202410037947A CN117994143A CN 117994143 A CN117994143 A CN 117994143A CN 202410037947 A CN202410037947 A CN 202410037947A CN 117994143 A CN117994143 A CN 117994143A
- Authority
- CN
- China
- Prior art keywords
- mode
- image
- modality
- encoder
- images
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Granted
Links
- 238000001308 synthesis method Methods 0.000 title claims abstract description 18
- 238000003860 storage Methods 0.000 title claims abstract description 10
- 238000012549 training Methods 0.000 claims abstract description 25
- 230000015572 biosynthetic process Effects 0.000 claims abstract description 19
- 238000003786 synthesis reaction Methods 0.000 claims abstract description 19
- 238000012545 processing Methods 0.000 claims abstract description 13
- 230000004927 fusion Effects 0.000 claims description 18
- 238000000034 method Methods 0.000 claims description 18
- 238000010586 diagram Methods 0.000 claims description 8
- 230000000295 complement effect Effects 0.000 claims description 7
- 238000004590 computer program Methods 0.000 claims description 6
- 238000011176 pooling Methods 0.000 claims description 6
- 238000005070 sampling Methods 0.000 claims description 6
- 230000006870 function Effects 0.000 claims description 5
- 230000003993 interaction Effects 0.000 claims description 5
- 230000008569 process Effects 0.000 claims description 4
- 238000011160 research Methods 0.000 description 7
- 230000000873 masking effect Effects 0.000 description 5
- 230000000007 visual effect Effects 0.000 description 5
- JXSJBGJIGXNWCI-UHFFFAOYSA-N diethyl 2-[(dimethoxyphosphorothioyl)thio]succinate Chemical compound CCOC(=O)CC(SP(=S)(OC)OC)C(=O)OCC JXSJBGJIGXNWCI-UHFFFAOYSA-N 0.000 description 4
- 238000005481 NMR spectroscopy Methods 0.000 description 3
- 238000003759 clinical diagnosis Methods 0.000 description 3
- 238000013461 design Methods 0.000 description 3
- 238000000605 extraction Methods 0.000 description 3
- 238000003384 imaging method Methods 0.000 description 3
- 238000002372 labelling Methods 0.000 description 3
- 238000002595 magnetic resonance imaging Methods 0.000 description 3
- 208000037004 Myoclonic-astatic epilepsy Diseases 0.000 description 2
- 238000003745 diagnosis Methods 0.000 description 2
- 238000000874 microwave-assisted extraction Methods 0.000 description 2
- 238000012986 modification Methods 0.000 description 2
- 230000004048 modification Effects 0.000 description 2
- 230000002194 synthesizing effect Effects 0.000 description 2
- 210000001519 tissue Anatomy 0.000 description 2
- 230000009471 action Effects 0.000 description 1
- 230000002776 aggregation Effects 0.000 description 1
- 238000004220 aggregation Methods 0.000 description 1
- 238000004458 analytical method Methods 0.000 description 1
- 210000003484 anatomy Anatomy 0.000 description 1
- 230000009286 beneficial effect Effects 0.000 description 1
- 230000008901 benefit Effects 0.000 description 1
- 238000004891 communication Methods 0.000 description 1
- 239000002131 composite material Substances 0.000 description 1
- 230000007547 defect Effects 0.000 description 1
- 238000012217 deletion Methods 0.000 description 1
- 230000037430 deletion Effects 0.000 description 1
- 238000002059 diagnostic imaging Methods 0.000 description 1
- 230000000694 effects Effects 0.000 description 1
- 230000002708 enhancing effect Effects 0.000 description 1
- 238000010191 image analysis Methods 0.000 description 1
- 239000007943 implant Substances 0.000 description 1
- 230000006872 improvement Effects 0.000 description 1
- 238000012423 maintenance Methods 0.000 description 1
- 238000004519 manufacturing process Methods 0.000 description 1
- 238000013507 mapping Methods 0.000 description 1
- 239000000463 material Substances 0.000 description 1
- 239000002184 metal Substances 0.000 description 1
- 239000000203 mixture Substances 0.000 description 1
- 238000007781 pre-processing Methods 0.000 description 1
- 210000004872 soft tissue Anatomy 0.000 description 1
- 238000000638 solvent extraction Methods 0.000 description 1
- 238000012360 testing method Methods 0.000 description 1
- 230000009466 transformation Effects 0.000 description 1
- 238000009966 trimming Methods 0.000 description 1
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T5/00—Image enhancement or restoration
- G06T5/50—Image enhancement or restoration using two or more images, e.g. averaging or subtraction
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
- G06T7/0002—Inspection of images, e.g. flaw detection
- G06T7/0012—Biomedical image inspection
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/10—Image acquisition modality
- G06T2207/10072—Tomographic images
- G06T2207/10088—Magnetic resonance imaging [MRI]
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/20—Special algorithmic details
- G06T2207/20212—Image combination
- G06T2207/20221—Image fusion; Image merging
Landscapes
- Engineering & Computer Science (AREA)
- Physics & Mathematics (AREA)
- General Physics & Mathematics (AREA)
- Theoretical Computer Science (AREA)
- Health & Medical Sciences (AREA)
- General Health & Medical Sciences (AREA)
- Medical Informatics (AREA)
- Nuclear Medicine, Radiotherapy & Molecular Imaging (AREA)
- Radiology & Medical Imaging (AREA)
- Quality & Reliability (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Magnetic Resonance Imaging Apparatus (AREA)
Abstract
The invention relates to a multi-mode MR image synthesis method, a system, a storage medium and equipment, wherein MR images of different modes are obtained and divided into mutually non-overlapping blocks, images of at least two modes are randomly aligned and shielded, an alignment mark and a corresponding position in each mode image are extracted based on an encoder, an image of a target mode is obtained based on a decoder, characteristics in the image of the target mode are utilized to be subjected to consistency processing, and a synthesized image of a real mode is obtained through training; during training, an edge map corresponding to the mode image is obtained by using a generator and is spliced with the original mode image to form the characteristic representation of the set mode. By means of limited pairing data synthesis missing modalities, the time of data pairing can be saved and the cost of the MR device reduced.
Description
Technical Field
The invention relates to the technical field of image synthesis, in particular to a multi-mode MR image synthesis method, a multi-mode MR image synthesis system, a storage medium and multi-mode MR image synthesis equipment.
Background
The statements in this section merely provide background information related to the present disclosure and may not necessarily constitute prior art.
MR images, i.e. magnetic resonance images, are a non-invasive, non-radiative imaging technique that uses signals generated by nuclei under the action of strong magnetic fields and radio frequency pulses to generate high resolution images of human tissue. Its high contrast, multi-planar imaging and superior performance on soft tissue make it an indispensable tool in clinical and scientific research. While this technique has some limitations, the manufacturing and maintenance costs of the equipment are high; acquiring magnetic resonance images generally requires a long scan, and for some patients, particularly patients that cannot remain stationary, may lead to discomfort leading to reduced imaging quality; some patients carrying metal implants or instruments present a certain safety risk in acquiring magnetic resonance images.
Disclosure of Invention
In order to solve the technical problems in the background art, the invention provides a multi-mode MR image synthesis method, a system, a storage medium and equipment, wherein under the condition of limited pairing data, a multi-mode medical image synthesis frame (MCF-Net) with patch complementary pre-training (PC-MAE) is introduced, and the limitation of the traditional method on noise, motion artifact and cost is overcome by fusing information of different modes of a magnetic resonance image, so that the cost of acquiring the image is reduced, and the quality of the image is improved.
In order to achieve the above purpose, the present invention adopts the following technical scheme:
A first aspect of the invention provides a method of multi-modality MR image synthesis comprising the steps of:
MR images of different modes are obtained and divided into mutually non-overlapping blocks, at least two of the images of the modes are randomly aligned and shielded, an alignment mark and a corresponding position in each mode image are extracted based on an encoder, an image of a target mode is obtained based on a decoder, characteristics in the image of the target mode are subjected to consistency processing, and a synthesized image of a real mode is obtained through training;
During training, an edge map corresponding to the mode image is obtained by using a generator and is spliced with the original mode image to form the characteristic representation of the set mode.
Further, the alignment mark and the corresponding position in each modal image are extracted based on the encoder, specifically: and acquiring the alignment mark and the corresponding position in the mode image by utilizing the encoder corresponding to each mode, and acquiring high-dimensional data coding representation of different mode images by using the shared encoder.
Further, the image of the target mode is obtained based on the decoder, specifically: and performing cross-modal interaction by a shared decoder corresponding to the shared encoder, and obtaining images of corresponding target modalities based on the specific decoder.
Furthermore, the characteristic representation of the mode is set, downsampling is realized through branches of the encoder corresponding to each mode, the decoder corresponding to each mode improves the characteristic resolution, and the image of the target mode is obtained through fusion.
Further, obtaining an image of the target mode through fusion comprises the following steps:
Determining a characteristic representation acquired by an encoder corresponding to each mode, and an up-sampling characteristic representation acquired by a decoder corresponding to each mode;
in the space attention guiding stage, splicing the characteristic representation acquired by the encoder and the up-sampling characteristic representation acquired by the decoder, and obtaining a space attention map and scaling through point-by-point convolution operation;
a spatial multiplication is performed between the spatial attention map and the input feature to obtain a filtered feature.
Further, obtaining an image of the target mode through fusion further comprises:
in the channel attention guiding stage, the filtered characteristics are subjected to pooling treatment to obtain a channel context descriptor;
And recombining the channel context descriptors to obtain a fused characteristic diagram and taking the fused characteristic diagram as input of the next up-sampling stage.
Further, the consistency processing specifically includes: and balancing a loss function through weight parameters to realize consistency processing by measuring the difference between the synthesized image and the true value based on the characteristic consistency loss by using the fused target mode image and the corresponding true value of the missing mode.
A second aspect of the invention provides a multi-modality MR image synthesis system comprising:
A patch complementary pre-training module configured to: MR images of different modes are obtained and divided into mutually non-overlapping blocks, at least two of the images of the modes are randomly aligned and shielded, an alignment mark and a corresponding position in each mode image are extracted based on an encoder, an image of a target mode is obtained based on a decoder, characteristics in the image of the target mode are subjected to consistency processing, and a synthesized image of a real mode is obtained through training;
an edge enhancement trim module configured to: during training, an edge map corresponding to the mode image is obtained by using a generator and is spliced with the original mode image to form the characteristic representation of the set mode.
A third aspect of the present invention provides a computer readable storage medium having stored thereon a computer program which when executed by a processor performs the steps of the above-described multi-modality MR image synthesis method.
A fourth aspect of the invention provides a computer device comprising a memory, a processor and a computer program stored on the memory and executable on the processor, the processor implementing the steps in the above-described multi-modality MR image synthesis method when the program is executed.
Compared with the prior art, the above technical scheme has the following beneficial effects:
Since MR images are mostly used in medical image processing, due to the high cost of data acquisition and labeling, only a small amount of paired data is often available, and at the same time, the missing modality may cause incomplete medical image data, which may affect subsequent clinical diagnosis and scientific research. Therefore, each block contains complete anatomical structure information by dividing a limited modal image into mutually non-overlapping blocks, simulating differences and deletions in an actual medical image by randomly aligning and masking, and reconstructing two complete modal images (images of target modalities) which are not masked originally by using a corresponding encoder and decoder, thereby determining weights; and when the method is operated, a generator is utilized to obtain edge images corresponding to the two input mode images and splice the edge images with the original mode images, the characteristics are further extracted, and the synthesized third mode image is obtained by utilizing the weight obtained in the prior art through characteristic fusion, so that the time for data pairing can be saved, and the cost of the MR equipment can be reduced.
Drawings
The accompanying drawings, which are included to provide a further understanding of the invention and are incorporated in and constitute a part of this specification, illustrate embodiments of the invention and together with the description serve to explain the invention.
FIG. 1 is a schematic diagram of an overall architecture of a multi-modality MR image synthesis process provided by one or more embodiments of the present invention;
FIG. 2 is a schematic diagram of a multi-modal collaborative fusion (MCF-Net) network architecture according to one or more embodiments of the present invention;
FIG. 3 is a schematic diagram of an architecture of an attention directing fusion module (ADF) provided by one or more embodiments of the invention;
FIG. 4 is a visual result of a T1 and T2 synthetic FLAIR task provided by one or more embodiments of the present invention;
FIG. 5 is a visual result of a T2 and FLAIR synthesis T1 task provided by one or more embodiments of the present invention;
FIG. 6 is a visual result of a T1 and FLAIR synthesis T2 task provided by one or more embodiments of the present invention.
Detailed Description
The invention will be further described with reference to the drawings and examples.
It should be noted that the following detailed description is exemplary and is intended to provide further explanation of the invention. Unless defined otherwise, all technical and scientific terms used herein have the same meaning as commonly understood by one of ordinary skill in the art to which this invention belongs.
The magnetic resonance imaging (Magnetic Resonance Imaging, abbreviated as MRI or MR) uses the principle of nuclear magnetic resonance (nuclearmagneticresonance, abbreviated as NMR) to detect the emitted electromagnetic waves by externally applied gradient magnetic fields according to the different attenuation of the released energy in different structural environments inside the material, so as to obtain the positions and types of the nuclei constituting the object, and can be drawn into an image of the internal structure of the object, which is commonly used for detecting human tissues by medical institutions.
The following embodiments provide a multi-mode MR image synthesis method, system, storage medium and apparatus, introducing a multi-mode medical image synthesis framework (MCF-Net) with patch complementary pre-training (PC-MAE) under the condition of limited pairing data, and overcoming the limitation of the traditional method on noise, motion artifact and cost by fusing the information of different modes of the magnetic resonance image, thereby reducing the cost of acquiring the image and improving the quality of the image.
Embodiment one:
as shown in fig. 1-6, the multi-modality MR image synthesis method includes the steps of:
Step 1: data acquisition and preprocessing.
1.1 Data acquisition: collecting MR images of different modalities (e.g., the published BRATS2020 data) from a medical imaging database, from which three modality images are selected, such as T1 weighted, T2 weighted, and FLAIR images;
1.2 image block partitioning: dividing the image of each modality into mutually non-overlapping blocks, typically of fixed size, to ensure that each block contains complete anatomical information;
1.3 random pair Ji Zhebi: randomly aligned masking of two modal images after division into a plurality of blocks, namely masking blocks of a first modal image, and not masking blocks at the same position of a second input modal image (so-called masking some blocks, namely adding great noise to small blocks of the images to cause the small blocks to lose original information so as to simulate differences and defects in actual medical images)
Step 2: feature extraction and decoding.
2.1 Encoder feature extraction: arranging two image blocks with non-occluded modes into a row in a one-dimensional vector form respectively, inputting the image blocks into a visual transducer encoder respectively, and extracting an alignment mark and corresponding position information in each mode image; then splicing and inputting the characteristics output by the two specific encoders into the shared encoder and decoder;
2.2 decoder decodes: based on the characteristics and the position information obtained in the steps, the information is decoded by two specific decoders, so that the reconstruction of the two-mode image is realized (the image is segmented and shielded at first, the non-shielded image blocks are input, and the information of the non-shielded image blocks is extracted, so that the two complete original non-shielded mode images are reconstructed and recovered).
The method comprises the steps of pre-training, carrying out shielding reconstruction on two modal images, training a model, extracting the characteristics of the model, and obtaining pre-trained weights for subsequent synthesis tasks.
Step 3: edge enhancement and weight reading.
3.1: And reading in the weight of the pre-training, and continuing to train the downlink task on the basis. And respectively generating edge maps corresponding to the two input modes by utilizing a Sobel generator, respectively splicing the two edge maps with the two input mode images, and inputting the two edge maps into an encoder of the visual transducer together.
Step 4: target modality synthesis and consistency processing.
4.1 Fusion of features: further extracting features through an MCF-Net model, and fusing the feature information obtained from the front side together through an ADF module to finally obtain a third target mode image;
4.2 consistency handling: the synthesized image and the real image are aligned in characteristics through the pre-trained encoder, so that the difference between the two modes is reduced, and the synthesized image is more lifelike.
The synthesis process of this embodiment may be divided into two parts, patch complementary pre-training and edge enhancement fine tuning.
Patch complement pre-training (PatchComplementaryPre-training). The input image is firstly divided into a series of non-overlapping blocks, and then the two input modes are randomly aligned and shielded, so that the patch at the same position in the two modes is ensured to have information complementarity. The unmasked patches are projected into the D-dimensional embeddings, respectively, while the masked patches are masked.
In order to enable the encoders to learn the modal differences by mapping the input to the feature space, the aligned visible markers and their respective positions and modal embeddings are provided to the respective encoders. Feature fusion and cross-modal interaction of the visible patch is then facilitated by sharing the encoder.
In general, MAE encoders benefit from learning a generic encoder that captures high-dimensional data encoded representations of different modality data. Because of the differences between the two modalities, specialized decoders are required to decode the advanced potential information of the respective modalities. The purpose of the additional shared decoder layer is ultimately to focus the encoder more on feature extraction while ignoring details of modal interactions. Since MAEs employ an asymmetric auto-encoder design in which the mask token does not pass through the shared encoder, the mask token is passed in addition to the visible token through the shared decoder. Without such a design, the entire decoder branch would be partitioned and the mask flags for the different modalities would not participate in feature fusion. After sharing the decoder, a special decoder is designed for the different modalities to achieve a better reconstruction.
Loss is measured using Mean Square Error (MSE), respectively denoted asAnd/>The input of the shared decoder includes the complete set of coded visible features and mask marks for both modalities, and the shared decoder performs cross-modality interactions on these potential representations. Then, at a separate decoder stage, the decoder maps back to the image, with the final penalty of: /(I)The detailed architecture is shown in fig. 1.
Edge-enhanced trimming (Edge-ENHANCEDFINE-Tuning). And generating an edge map corresponding to the input mode image by using a Sobel generator, connecting the generated edge map with the original mode image, and forming a characteristic representation of a specific mode by utilizing a pre-training stage in a multi-mode complementation mode. These feature representations are input into two independent encoder branches of an MCF-Net (multi-modal collaborative fusion network), each branch containing multiple downsampling stages. At each stage, swinTransformer layers and patch merging modules are adopted, the resolution of the feature map is reduced through continuous operation, and the feature dimension is doubled, so that a layered feature representation is formed.
Notably, the large-scale branching includes an additional downsampling stage as compared to the small-scale branching, further enhancing the feature capture capability of the network. In order to maintain important information of a specific modality in the final composition, a decoder is employed, including a patch expansion operation and SwinTransformer layers, to improve feature resolution. By introducing a jump connection, a tight communication between the encoder and the decoder is achieved, helping to preserve more modality-specific information. The design enables the MCF-Net to capture the association between each mode more flexibly and effectively in the process of feature fusion, and provides more powerful performance for multi-mode synthesis tasks, and the detailed architecture is shown in fig. 1-2.
The following are two important modules that introduce MCF-Net (multi-modal collaborative fusion network) inside:
1) Attention directed fusion module (Attention-DrivenFusionModule): in order to better fuse the features of a specific mode in the multi-modes extracted by the dual-branch encoder, an attention-directed Aggregation (ADF) module is proposed to effectively integrate the multi-mode features, and propagate the fused features into the decoder by skipping the connection. Inspired by the existing attention-based feature selection module, the proposed AD F module consists of two key components, spatial attention-directed fusion, channel attention-directed fusion, as shown in fig. 3, representing the specific modal features of the dual-branch encoder as F a∈RC×(H×W) and F b∈RC ×(H×W), and the upsampled features from the decoder as F s∈RC×(H×W). In the spatial attention guidance stage, F a and F b are respectively connected with F s, and then connected features are input into a point-by-point convolution operation to generate two spatial attention force diagrams M a∈RH×W and M b∈RH×W.
Spatial attention is striven scaled to [0,1] using a Sigmoid function. To suppress noise features while emphasizing information features, spatial multiplication is performed between the spatial attention map and the input features, resulting in filtered features F 'a and F' b. In the channel attention guidance phase, global average pooling and global maximum pooling are applied to the filtered input features F 'a and F' b, respectively, to obtain channel context descriptorsAnd/>And respectively carrying out addition operation on the maximum pooling and average pooling results to obtain: /(I)They were then combined into F p=[Pa,Pb]∈R2C×1, and then softmax manipulation was performed along the channel: /(I)Wherein/>And/>Each representing the ith element of the corresponding channel attention map. Here, there is/> And/>The i-th element in P a and P b are denoted, respectively. Information features are obtained by fusing the dual-branch features, which can be processed by: wherein F represents the fused feature map. Finally, connections F and F s are used as inputs for the next upsampling stage, the detailed architecture is shown in FIG. 3.
2) And a feature consistency module: in order to obtain a rich and generic representation of the modal characteristics learned from large-scale data, a pre-training encoder is chosen as the core of the characteristic consistency module. Specifically, the image to be synthesizedAnd the corresponding true values y of the missing modes are fed into the pre-trained PC-MAE encoder, where x, E and G represent the input image, the pre-trained encoder and the proposed MCF-Net, respectively. Let F j (y) and/>The j-th transformation layer of the feature consistency module is derived from y and/>, respectivelyThe output of the multi-level features is extracted, and the perceived difference between the synthesized image and the true value is measured by using the feature consistency loss, namely the method can be defined by the following formula:
Wherein, Indicating the desire, l indicates the number of transform layers of the feature consistency module. Feature consistency loss is used in the framework, enabling more emphasis on content and style similarity between images. Further, the conventional pixel-by-pixel difference between the composite image and the true value can be measured by the following formula:
then, the overall objective function of the fine-tuning framework is formed by linear combination:
Where λ is the weight parameter used to balance the two loss functions.
The core goal of this embodiment is to generate a missing modality with limited paired data, which is often available only in small amounts due to the high cost of data acquisition and labeling in medical image processing. The present embodiment can thus utilize limited pairing data to generate missing modalities in order to better utilize such data in medical image analysis and diagnosis.
1. The data pairing time can be saved: conventional supervised learning methods typically require a large amount of paired multi-modal data to train. However, in practical applications, it is often expensive or impractical to obtain sufficient pairing data. The method provided by the embodiment can effectively utilize limited pairing data.
2. Can support medical diagnosis and research: the missing modalities may lead to incomplete medical image data, which can affect clinical diagnosis and scientific research. By synthesizing the missing modalities, doctors and researchers can obtain more comprehensive image information, helping to make more accurate diagnoses and conduct more intensive studies.
3. Can save cost: obtaining complete multimodal medical image data generally requires more cost and time, including additional scan time and equipment. By synthesizing the missing modalities, the required information can be obtained without increasing cost and time, thereby improving the efficiency of medical care.
4. Can enhance the case data: for medical research and education, it is very valuable to have more case data. The synthetic missing modality can be used to augment existing medical image datasets, enabling researchers to conduct more extensive research and analysis.
Furthermore, in some cases, a physician may need multimodal information to formulate a treatment plan for a patient. The resultant missing modalities may provide the physician with complete information that helps to better plan the treatment plan.
By testing the above procedure with three different compositing tasks as shown in fig. 4-6, in each of the figures from left to right, columns 1-2 represent the input of the model, column 3 represents the actual result to be composited, column 4 represents the result of the model compositing, and columns 5-8 represent the results composited by other comparison methods, and by comparison, the embodiment can be found to have good effects on all three tasks and also to highlight its effectiveness in the multi-modal MR image compositing.
Embodiment two:
A multi-modality MR image synthesis system comprising:
A patch complementary pre-training module configured to: MR images of different modes are obtained and divided into mutually non-overlapping blocks, at least two of the images of the modes are randomly aligned and shielded, an alignment mark and a corresponding position in each mode image are extracted based on an encoder, an image of a target mode is obtained based on a decoder, characteristics in the image of the target mode are subjected to consistency processing, and a synthesized image of a real mode is obtained through training;
an edge enhancement trim module configured to: during training, an edge map corresponding to the mode image is obtained by using a generator and is spliced with the original mode image to form the characteristic representation of the set mode.
MR images are mostly used in medical image processing, because of the high cost of data acquisition and labeling, only a small amount of paired data is often available, and at the same time, the missing modality may cause incomplete medical image data, which may affect subsequent clinical diagnosis and scientific research, and the missing modality is synthesized by limited paired data, so that the time of data pairing can be saved and the cost of the MR device can be reduced.
Embodiment III:
The present embodiment provides a computer-readable storage medium having stored thereon a computer program which, when executed by a processor, implements the steps in the multi-modality MR image synthesis method as described in the above embodiment.
Embodiment four:
the present embodiment provides a computer device, including a memory, a processor and a computer program stored on the memory and executable on the processor, where the processor implements the steps in the multi-modality MR image synthesis method according to the above embodiment when executing the program.
The steps involved in the second to fourth embodiments correspond to the first embodiment, and the detailed description of the second embodiment refers to the related description section of the first embodiment. The term "computer-readable storage medium" should be taken to include a single medium or multiple media including one or more sets of instructions; it should also be understood to include any medium capable of storing, encoding or carrying a set of instructions for execution by a processor and that cause the processor to perform any one of the methods of the present invention.
The above description is only of the preferred embodiments of the present invention and is not intended to limit the present invention, but various modifications and variations can be made to the present invention by those skilled in the art. Any modification, equivalent replacement, improvement, etc. made within the spirit and principle of the present invention should be included in the protection scope of the present invention.
Claims (10)
1. The multi-mode MR image synthesis method is characterized by comprising the following steps:
MR images of different modes are obtained and divided into mutually non-overlapping blocks, at least two of the images of the modes are randomly aligned and shielded, an alignment mark and a corresponding position in each mode image are extracted based on an encoder, an image of a target mode is obtained based on a decoder, characteristics in the image of the target mode are subjected to consistency processing, and a synthesized image of a real mode is obtained through training;
During training, an edge map corresponding to the mode image is obtained by using a generator and is spliced with the original mode image to form the characteristic representation of the set mode.
2. The multi-modality MR image synthesis method of claim 1, wherein the alignment markers and corresponding positions in each modality image are extracted based on an encoder, in particular: and acquiring the alignment mark and the corresponding position in the mode image by utilizing the encoder corresponding to each mode, and acquiring high-dimensional data coding representation of different mode images by using the shared encoder.
3. The multi-modality MR image synthesis method of claim 1, wherein the decoder-based acquisition of images of the target modality is: and performing cross-modal interaction by a shared decoder corresponding to the shared encoder, and obtaining images of corresponding target modalities based on the specific decoder.
4. The multi-mode MR image synthesis method of claim 1, wherein a characteristic representation of a mode is set, downsampling is performed through branches of an encoder corresponding to each mode, a decoder corresponding to each mode improves a characteristic resolution, and an image of a target mode is obtained through fusion.
5. The method of multi-modality MR image synthesis as claimed in claim 4, wherein obtaining an image of the target modality via fusion includes:
Determining a characteristic representation acquired by an encoder corresponding to each mode, and an up-sampling characteristic representation acquired by a decoder corresponding to each mode;
in the space attention guiding stage, splicing the characteristic representation acquired by the encoder and the up-sampling characteristic representation acquired by the decoder, and obtaining a space attention map and scaling through point-by-point convolution operation;
a spatial multiplication is performed between the spatial attention map and the input feature to obtain a filtered feature.
6. The method of multi-modality MR image synthesis as claimed in claim 4, wherein the image of the target modality is obtained via fusion, further comprising:
in the channel attention guiding stage, the filtered characteristics are subjected to pooling treatment to obtain a channel context descriptor;
And recombining the channel context descriptors to obtain a fused characteristic diagram and taking the fused characteristic diagram as input of the next up-sampling stage.
7. The multi-modality MR image synthesis method of claim 1, wherein the consistency process is specifically: and balancing a loss function through weight parameters to realize consistency processing by measuring the difference between the synthesized image and the true value based on the characteristic consistency loss by using the fused target mode image and the corresponding true value of the missing mode.
8. A multi-modality MR image synthesis system, comprising:
A patch complementary pre-training module configured to: MR images of different modes are obtained and divided into mutually non-overlapping blocks, at least two of the images of the modes are randomly aligned and shielded, an alignment mark and a corresponding position in each mode image are extracted based on an encoder, an image of a target mode is obtained based on a decoder, characteristics in the image of the target mode are subjected to consistency processing, and a synthesized image of a real mode is obtained through training;
an edge enhancement trim module configured to: during training, an edge map corresponding to the mode image is obtained by using a generator and is spliced with the original mode image to form the characteristic representation of the set mode.
9. A computer readable storage medium, on which a computer program is stored, which program, when being executed by a processor, carries out the steps of the multi-modality MR image synthesis method as claimed in any one of claims 1-7.
10. Computer device, characterized in that it comprises a memory, a processor and a computer program stored on the memory and executable on the processor, which processor, when executing the program, realizes the steps in the multi-modality MR image synthesis method according to any of claims 1-7.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202410037947.3A CN117994143B (en) | 2024-01-09 | 2024-01-09 | Multi-mode MR image synthesis method, system, storage medium and equipment |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202410037947.3A CN117994143B (en) | 2024-01-09 | 2024-01-09 | Multi-mode MR image synthesis method, system, storage medium and equipment |
Publications (2)
Publication Number | Publication Date |
---|---|
CN117994143A true CN117994143A (en) | 2024-05-07 |
CN117994143B CN117994143B (en) | 2024-08-20 |
Family
ID=90886560
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN202410037947.3A Active CN117994143B (en) | 2024-01-09 | 2024-01-09 | Multi-mode MR image synthesis method, system, storage medium and equipment |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN117994143B (en) |
Citations (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN114972231A (en) * | 2022-05-17 | 2022-08-30 | 华中科技大学 | Multi-modal MR image segmentation method based on prior-posterior probability encoder |
CN115861464A (en) * | 2022-12-01 | 2023-03-28 | 南方医科大学 | Pseudo CT (computed tomography) synthesis method based on multimode MRI (magnetic resonance imaging) synchronous generation |
CN115965638A (en) * | 2022-12-25 | 2023-04-14 | 衡阳师范学院 | Twin self-distillation method and system for automatically segmenting modal-deficient brain tumor image |
-
2024
- 2024-01-09 CN CN202410037947.3A patent/CN117994143B/en active Active
Patent Citations (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN114972231A (en) * | 2022-05-17 | 2022-08-30 | 华中科技大学 | Multi-modal MR image segmentation method based on prior-posterior probability encoder |
CN115861464A (en) * | 2022-12-01 | 2023-03-28 | 南方医科大学 | Pseudo CT (computed tomography) synthesis method based on multimode MRI (magnetic resonance imaging) synchronous generation |
CN115965638A (en) * | 2022-12-25 | 2023-04-14 | 衡阳师范学院 | Twin self-distillation method and system for automatically segmenting modal-deficient brain tumor image |
Non-Patent Citations (2)
Title |
---|
SALMAN UH DAR ET AL.: "Image synthesis in multi-contrast MRI with conditional generative adversarial networks", 《IEEE TRANS. MED. IMAGING》, vol. 38, no. 10, 26 February 2019 (2019-02-26), pages 2375, XP011748194, DOI: 10.1109/TMI.2019.2901750 * |
YONGHAO LI ET AL.: "Multi-Scale Transformer Network With Edge-Aware Pre-Training for Cross-Modality MR Image Synthesis", 《IEEE TRANSACTIONS ON MEDICAL IMAGING》, vol. 42, no. 11, 20 June 2023 (2023-06-20), pages 3395 - 3407 * |
Also Published As
Publication number | Publication date |
---|---|
CN117994143B (en) | 2024-08-20 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
Wang et al. | DiCyc: GAN-based deformation invariant cross-domain information fusion for medical image synthesis | |
Yurt et al. | mustGAN: multi-stream generative adversarial networks for MR image synthesis | |
Chartsias et al. | Multimodal MR synthesis via modality-invariant latent representation | |
Zhan et al. | Multi-modal MRI image synthesis via GAN with multi-scale gate mergence | |
Yoon et al. | Sadm: Sequence-aware diffusion model for longitudinal medical image generation | |
CN116309648A (en) | Medical image segmentation model construction method based on multi-attention fusion | |
CN114565816B (en) | Multi-mode medical image fusion method based on global information fusion | |
Ristea et al. | CyTran: A cycle-consistent transformer with multi-level consistency for non-contrast to contrast CT translation | |
Hu et al. | Cross‐model transformer method for medical image synthesis | |
Tavse et al. | A systematic literature review on applications of GAN-synthesized images for brain MRI | |
Wu et al. | Image synthesis in contrast MRI based on super resolution reconstruction with multi-refinement cycle-consistent generative adversarial networks | |
CN118037615A (en) | Tumor segmentation guided magnetic resonance image translation method, system, equipment and medium | |
CN116152235A (en) | Cross-modal synthesis method for medical image from CT (computed tomography) to PET (positron emission tomography) of lung cancer | |
Wang et al. | IGNFusion: an unsupervised information gate network for multimodal medical image fusion | |
CN113554728B (en) | Semi-supervised-based multi-sequence 3T to 7T magnetic resonance image generation method | |
Yang et al. | MGDUN: An interpretable network for multi-contrast MRI image super-resolution reconstruction | |
Wang et al. | MSE-Fusion: Weakly supervised medical image fusion with modal synthesis and enhancement | |
CN117475268A (en) | Multimode medical image fusion method based on SGDD GAN | |
CN117611602A (en) | Fiber bundle segmentation method based on U-NET network double-encoder structure | |
CN117994143B (en) | Multi-mode MR image synthesis method, system, storage medium and equipment | |
CN114298979B (en) | Method for generating hepatonuclear magnetic image sequence guided by description of focal lesion symptom | |
Li et al. | CPFTransformer: transformer fusion context pyramid medical image segmentation network | |
CN115965785A (en) | Image segmentation method, device, equipment, program product and medium | |
CN112529949A (en) | Method and system for generating DWI image based on T2 image | |
Chen et al. | MS-UNet: Swin Transformer U-Net with Multi-scale Nested Decoder for Medical Image Segmentation with Small Training Data |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
GR01 | Patent grant | ||
GR01 | Patent grant |