CN116645336A - MRI brain image gland pituitary segmentation method - Google Patents
MRI brain image gland pituitary segmentation method Download PDFInfo
- Publication number
- CN116645336A CN116645336A CN202310522974.5A CN202310522974A CN116645336A CN 116645336 A CN116645336 A CN 116645336A CN 202310522974 A CN202310522974 A CN 202310522974A CN 116645336 A CN116645336 A CN 116645336A
- Authority
- CN
- China
- Prior art keywords
- pituitary
- dimensional
- image
- mask
- cross
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Granted
Links
- 230000011218 segmentation Effects 0.000 title claims abstract description 59
- 230000001817 pituitary effect Effects 0.000 title claims abstract description 53
- 238000000034 method Methods 0.000 title claims abstract description 38
- 210000004556 brain Anatomy 0.000 title claims abstract description 20
- 210000004907 gland Anatomy 0.000 title description 2
- 206010062767 Hypophysitis Diseases 0.000 claims abstract description 55
- 210000003635 pituitary gland Anatomy 0.000 claims abstract description 55
- 238000007781 pre-processing Methods 0.000 claims abstract description 5
- 238000004891 communication Methods 0.000 claims description 9
- 238000003709 image segmentation Methods 0.000 claims description 8
- 238000004364 calculation method Methods 0.000 claims description 7
- 238000013527 convolutional neural network Methods 0.000 description 10
- 238000012549 training Methods 0.000 description 7
- 238000013135 deep learning Methods 0.000 description 3
- 201000010099 disease Diseases 0.000 description 3
- 208000037265 diseases, disorders, signs and symptoms Diseases 0.000 description 3
- 230000006870 function Effects 0.000 description 3
- ORILYTVJVMAKLC-UHFFFAOYSA-N Adamantane Natural products C1C(C2)CC3CC1CC2C3 ORILYTVJVMAKLC-UHFFFAOYSA-N 0.000 description 2
- 238000013136 deep learning model Methods 0.000 description 2
- 238000000605 extraction Methods 0.000 description 2
- 230000004927 fusion Effects 0.000 description 2
- 230000009467 reduction Effects 0.000 description 2
- 238000000638 solvent extraction Methods 0.000 description 2
- 238000005481 NMR spectroscopy Methods 0.000 description 1
- 230000006978 adaptation Effects 0.000 description 1
- 230000003044 adaptive effect Effects 0.000 description 1
- 238000013459 approach Methods 0.000 description 1
- 230000009286 beneficial effect Effects 0.000 description 1
- 238000002790 cross-validation Methods 0.000 description 1
- 238000010586 diagram Methods 0.000 description 1
- 230000000694 effects Effects 0.000 description 1
- 238000005516 engineering process Methods 0.000 description 1
- 238000001914 filtration Methods 0.000 description 1
- 238000003384 imaging method Methods 0.000 description 1
- 230000003993 interaction Effects 0.000 description 1
- 230000004807 localization Effects 0.000 description 1
- 230000014759 maintenance of location Effects 0.000 description 1
- 230000000877 morphologic effect Effects 0.000 description 1
- 238000010606 normalization Methods 0.000 description 1
- 238000005457 optimization Methods 0.000 description 1
- 210000000056 organ Anatomy 0.000 description 1
- 230000008569 process Effects 0.000 description 1
- 238000012545 processing Methods 0.000 description 1
- 238000011158 quantitative evaluation Methods 0.000 description 1
- 238000013519 translation Methods 0.000 description 1
- 238000010200 validation analysis Methods 0.000 description 1
- 230000000007 visual effect Effects 0.000 description 1
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
- G06T7/0002—Inspection of images, e.g. flaw detection
- G06T7/0012—Biomedical image inspection
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/04—Architecture, e.g. interconnection topology
- G06N3/0464—Convolutional networks [CNN, ConvNet]
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/08—Learning methods
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T17/00—Three dimensional [3D] modelling, e.g. data description of 3D objects
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
- G06T7/10—Segmentation; Edge detection
- G06T7/11—Region-based segmentation
-
- G—PHYSICS
- G16—INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR SPECIFIC APPLICATION FIELDS
- G16H—HEALTHCARE INFORMATICS, i.e. INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR THE HANDLING OR PROCESSING OF MEDICAL OR HEALTHCARE DATA
- G16H30/00—ICT specially adapted for the handling or processing of medical images
- G16H30/20—ICT specially adapted for the handling or processing of medical images for handling medical images, e.g. DICOM, HL7 or PACS
-
- G—PHYSICS
- G16—INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR SPECIFIC APPLICATION FIELDS
- G16H—HEALTHCARE INFORMATICS, i.e. INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR THE HANDLING OR PROCESSING OF MEDICAL OR HEALTHCARE DATA
- G16H30/00—ICT specially adapted for the handling or processing of medical images
- G16H30/40—ICT specially adapted for the handling or processing of medical images for processing medical images, e.g. editing
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/10—Image acquisition modality
- G06T2207/10072—Tomographic images
- G06T2207/10088—Magnetic resonance imaging [MRI]
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/20—Special algorithmic details
- G06T2207/20021—Dividing image into blocks, subimages or windows
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/20—Special algorithmic details
- G06T2207/20081—Training; Learning
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/20—Special algorithmic details
- G06T2207/20084—Artificial neural networks [ANN]
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/30—Subject of image; Context of image processing
- G06T2207/30004—Biomedical image processing
- G06T2207/30016—Brain
Landscapes
- Engineering & Computer Science (AREA)
- Physics & Mathematics (AREA)
- Theoretical Computer Science (AREA)
- Health & Medical Sciences (AREA)
- General Health & Medical Sciences (AREA)
- General Physics & Mathematics (AREA)
- Medical Informatics (AREA)
- Nuclear Medicine, Radiotherapy & Molecular Imaging (AREA)
- Radiology & Medical Imaging (AREA)
- Software Systems (AREA)
- Life Sciences & Earth Sciences (AREA)
- Data Mining & Analysis (AREA)
- Public Health (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Primary Health Care (AREA)
- Artificial Intelligence (AREA)
- Biomedical Technology (AREA)
- Biophysics (AREA)
- Computational Linguistics (AREA)
- Epidemiology (AREA)
- Evolutionary Computation (AREA)
- Molecular Biology (AREA)
- Computing Systems (AREA)
- General Engineering & Computer Science (AREA)
- Mathematical Physics (AREA)
- Computer Graphics (AREA)
- Quality & Reliability (AREA)
- Geometry (AREA)
- Magnetic Resonance Imaging Apparatus (AREA)
- Image Analysis (AREA)
Abstract
A method of MRI brain image adenohypophysis segmentation, comprising; preprocessing an MRI brain image acquired from a hospital, converting original Dicom format MRI data acquired from the hospital into a three-dimensional image in an NIFTI format by using ITK-SNAP software, and converting the image in the DICOM format into the three-dimensional image in the NIFTI format; cutting and dividing the three-dimensional image in the NIFTI format into sub-blocks; inputting the sub-blocks into VT-Unet to obtain a segmentation result, finally calculating a 3D boundary box of the pituitary gland, and applying the coordinates of the boundary box to the NIFTI image to obtain a pituitary region, thereby completing pituitary positioning; using the output of the pituitary gland positioning of the previous stage, generating a 2D slice according to the cross section direction as the input of the stage; finally, outputting a two-dimensional label of the pituitary gland, and then stacking all the two-dimensional slices in the cross section direction to reconstruct a three-dimensional image, thereby obtaining the final three-dimensional pituitary gland label. The method has the characteristics of high speed, high accuracy, good robustness, high efficiency and good generalization.
Description
Technical Field
The invention belongs to the technical field of clinical pituitary gland segmentation, and particularly relates to a method for segmenting pituitary gland by using an MRI brain image.
Background
Nuclear Magnetic Resonance (MRI) images are currently accepted as the best imaging method for displaying the morphological characteristics of the pituitary gland due to the fact that the MRI images are non-radiative and have high tissue contrast.
The existing pituitary gland segmentation method is mainly manual segmentation, and no method special for full-automatic segmentation of the pituitary gland exists at present. However, artificial-based pituitary segmentation suffers from the problems of being time consuming, subjective and inaccurate. In recent years, deep learning technology has been developed rapidly, and Convolutional Neural Networks (CNNs) have achieved great success in the field of medical image segmentation. The symmetrical codec structure used by U-Net and its variants to improve detail retention is the dominant architecture for medical image segmentation and is widely used for organ and tissue segmentation. ViT (Vision Transformer) is a new trend in the field of computer vision at present, and modeling remote dependency relationships with its unique global attention provides a new idea for computer vision and is gradually used for medical image segmentation. However, neither CNN nor transducer is applied to the pituitary segmentation field. Although CNN has strong feature extraction capability, the CNN has a great promotion space for the reservation of global information; transformer is leading in the field of natural language image processing, but has drawbacks in terms of both data amount requirements and generalization ability of models for both computer vision and freshness.
Disclosure of Invention
In order to overcome the technical problems, the invention aims to provide the MRI brain image adenohypophysis segmentation method which has the characteristics of high speed, high accuracy, good robustness, high efficiency and good generalization.
In order to achieve the above purpose, the technical scheme adopted by the invention is as follows:
an MRI brain image adenohypophysis segmentation method comprises the following steps;
step 1: positioning of the pituitary gland:
preprocessing an MRI brain image acquired from a hospital, converting original Dicom format MRI data acquired from the hospital into a three-dimensional image in an NIFTI format by using ITK-SNAP software, and converting the image in the DICOM format into the three-dimensional image in the NIFTI format;
clipping a three-dimensional image in NIFTI format to a size of 80 x 11, and dividing the image into 49 sub-blocks with a size of 32 x 11; then inputting the 49 sub-blocks into VT-Unet to obtain a segmentation result of the 49 sub-blocks, combining 49 outputs into a mask with the size of 80 multiplied by 11 according to a segmentation method, finally calculating a 3D boundary frame of the pituitary gland mask, expanding the coronal position and the sagittal position of the boundary frame to 32 and expanding two layers back and forth on the cross section to ensure that the boundary frame completely contains the pituitary gland region, and applying the coordinates of the boundary frame to an NIFTI image with the size of 80 multiplied by 11 to obtain the pituitary gland region, thereby completing the positioning of the pituitary gland;
step 2: pituitary segmentation:
using the output of the pituitary positioning of the previous stage, 2D slices (32 x 32) are generated from the cross-sectional direction as input to this stage;
finally, outputting a two-dimensional label of the pituitary gland, and then stacking all the two-dimensional slices in the cross section direction to reconstruct a three-dimensional image, thereby obtaining the final three-dimensional pituitary gland label.
In the step 1, the MRI brain image is obtained by using a Siemens Skyra3.0T superconducting magnetic resonance scanner, and the pituitary scan uses the following sequences and parameters, and the total scan time is about 16 minutes; sagittal T1WI sequence (tr=641 ms, te=9.9 ms), sagittal T2WI sequence (tr=3800 ms, te=75 ms).
In the step 1, a sliding window with a size of 32×32×11 is firstly defined, the sliding window is initially fixed at the upper right corner, then the sliding window slides leftwards or downwards, and each time, the sliding unit is too small, so that adjacent image blocks contain a large amount of repeated information, and training resources are wasted; too large a sliding unit may result in too little training data, considering comprehensively that 8 units per sliding are selected, the sub-blocks are divided by the sliding window, and finally 49 sub-blocks are obtained, and each sub-block has a size of 32×32×11. Marking the position of each sub-block before division;
then inputting the 49 subblocks into the VT-Unet to obtain a segmentation result of the 49 subblocks, wherein the segmentation result of each subblock is a mask with the size of 32 multiplied by 11;
mask i =VTUnet(patch i ),i∈(1,...,49) (1)
wherein VTUnet represents VT-Unet, and is a three-dimensional volume transducer model suitable for medical image segmentation i Representing the ith sub-block, mask i Representing a segmentation result corresponding to the ith sub-block;
the 49 outputs are combined into the mask with the size of 80 multiplied by 11 according to the dividing method, and as the mask with the size of 80 multiplied by 11 obtained by the combination has a plurality of discrete points, the subsequent calculation amount can be increased, only a main part is reserved by removing the discrete points by using the method of taking the maximum communication area, and the reserved main part is the pituitary mask;
ademask=f(mask) (2)
wherein f represents the maximum communication function, mask represents 80×80×11 size mask, and adepack represents pituitary mask obtained by removing the maximum communication area;
finally, calculating a 3D boundary box of the pituitary gland mask, expanding the coronal position and the sagittal position of the boundary box to 32 sizes and expanding the boundary box back and forth on the cross section to ensure that the boundary box completely contains the pituitary gland region, and applying the coordinates of the boundary box to an NIFTI image of 80 multiplied by 11 to obtain the pituitary gland region, thereby completing the positioning of the pituitary gland;
in the step 2, using the output of the pituitary gland positioning of the previous stage, a 2D slice (32 multiplied by 32) is generated as the input of the stage according to the cross section direction;
x i =X[a,b,i],i∈(1,...,N) (3)
wherein X represents the pituitary gland region obtained by locating the pituitary gland, and X i The ith slice of X in the cross-sectional direction, a represents the coronal direction of X, b represents the delocalization direction of X, i represents the cross-sectional direction of X, and N represents the number of cross-sectional layers of X. Wherein a and b are both 32;
finally, outputting a two-dimensional label of the pituitary gland, and then stacking all two-dimensional slices in the cross section direction to reconstruct a three-dimensional image, so that the final three-dimensional label of the pituitary gland is obtained;
X[a,b,i]=x i ,i∈(1,...,N) (4)
wherein X represents a three-dimensional label reconstructed in the cross-sectional direction, X i The ith two-dimensional label, a represents the coronal bit direction of X, b represents the bit losing direction of X, i represents the cross section direction of X, and the number of N two-dimensional labels X.
The invention has the beneficial effects of.
The invention adopts a segmentation method based on two stages of deep learning, and has the advantages of high segmentation speed and high segmentation result accuracy (DICE= 0.9013). The data set uses brain MRI images of different diseases and different sequences, so that the robustness and generalization are good. Quantitative and qualitative assessment of pituitary segmentation are shown in table 1 and fig. 4. In table 1, the segmentation index of the method of the patent is higher than that of the traditional CNN model, viT (visual transducer) model and HyBird (hybrid CNN and transducer) model, and higher segmentation accuracy is shown. Fig. 4 shows that the method of this patent has a better segmentation effect than other methods.
Description of the drawings:
FIG. 1 is a schematic representation of an image segmentation sub-block in pituitary positioning according to the present invention.
FIG. 2 is a schematic diagram of an adenohypophysis segmentation convolutional neural network (PIT-Former) according to the present invention.
FIG. 3 is an expanded view of the pituitary positioning frame produced in the pituitary positioning stage of the present invention.
FIG. 4 is a qualitative assessment of pituitary segmentation according to the present invention. Pituitary segmentation examples of different methods for randomly selecting three patients, a1-a2, b1-b2, c1-c2 represent T1WI and T2WI, respectively, of three patients, PIT-force being the method presented in this section. The white line encircled portion represents the label.
Detailed Description
The following description of the embodiments of the present invention will be made clearly and completely with reference to the accompanying drawings, in which it is apparent that the embodiments described are only some embodiments of the present invention, but not all embodiments. All other embodiments, which can be made by those skilled in the art based on the embodiments of the invention without making any inventive effort, are intended to be within the scope of the invention.
An MRI brain image adenohypophysis segmentation method comprises the following steps;
step 1: positioning of the pituitary gland:
preprocessing an MRI brain image acquired from a hospital, converting original Dicom format MRI data acquired from the hospital into a three-dimensional image in an NIFTI format by using ITK-SNAP software, and converting the image in the DICOM format into the three-dimensional image in the NIFTI format;
clipping a three-dimensional image in NIFTI format to a size of 80 x 11, and dividing the image into 49 sub-blocks with a size of 32 x 11; then inputting the 49 sub-blocks into VT-Unet to obtain a segmentation result of the 49 sub-blocks, combining 49 outputs into a mask with the size of 80 multiplied by 11 according to a segmentation method, finally calculating a 3D boundary frame of the pituitary gland mask, expanding the coronal position and the sagittal position of the boundary frame to 32 and expanding two layers back and forth on the cross section to ensure that the boundary frame completely contains the pituitary gland region, and applying the coordinates of the boundary frame to an NIFTI image with the size of 80 multiplied by 11 to obtain the pituitary gland region, thereby completing the positioning of the pituitary gland;
specific expansion method as shown in fig. 3, the inner frame represents before expansion, and the outer frame represents after expansion.
Step 2: pituitary segmentation:
the purpose of pituitary segmentation is to accurately segment the pituitary gland on specific sections determined at the pituitary positioning stage, pituitary segmentation using PIT-force (FIG. 2). PIT-Former is a two-dimensional network fusing CNN and a transducer;
using the output of the pituitary positioning of the previous stage, 2D slices (32 x 32) are generated from the cross-sectional direction as input to this stage;
finally, outputting a two-dimensional label of the pituitary gland, and then stacking all the two-dimensional slices in the cross section direction to reconstruct a three-dimensional image, thereby obtaining the final three-dimensional pituitary gland label.
In the step 1, the MRI brain image is obtained by using a Siemens Skyra3.0T superconducting magnetic resonance scanner, and the pituitary scan uses the following sequences and parameters, and the total scan time is about 16 minutes; sagittal T1WI sequence (tr=641 ms, te=9.9 ms), sagittal T2WI sequence (tr=3800 ms, te=75 ms).
In the step 1, a sliding window with a size of 32×32×11 is firstly defined, the sliding window is initially fixed at the upper right corner (white), then the sliding window slides leftwards or downwards, and each time, the sliding unit is too small, so that adjacent image blocks contain a large amount of repeated information, and training resources are wasted; too large a sliding unit may result in too little training data, considering comprehensively that 8 units per sliding are selected, the sub-blocks are divided by the sliding window, and finally 49 sub-blocks are obtained, and each sub-block has a size of 32×32×11. Marking the position of each sub-block before division;
then inputting the 49 subblocks into the VT-Unet to obtain a segmentation result of the 49 subblocks, wherein the segmentation result of each subblock is a mask with the size of 32 multiplied by 11;
mask i =VTUnet(patch i ),i∈(1,...,49) (1)
wherein VTUnet represents VT-Unet, and is a three-dimensional volume transducer model suitable for medical image segmentation i Representing the ith sub-block, mask i Representing a segmentation result corresponding to the ith sub-block;
the 49 outputs are combined into the mask with the size of 80 multiplied by 11 according to the dividing method, and as the mask with the size of 80 multiplied by 11 obtained by the combination has a plurality of discrete points, the subsequent calculation amount can be increased, only a main part is reserved by removing the discrete points by using the method of taking the maximum communication area, and the reserved main part is the pituitary mask;
ademask=f(mask) (2)
wherein f represents the maximum communication function, mask represents 80×80×11 size mask, and adepack represents pituitary mask obtained by removing the maximum communication area;
finally, calculating a 3D boundary box of the pituitary gland mask, expanding the coronal position and the sagittal position of the boundary box to 32 sizes and expanding the boundary box back and forth on the cross section to ensure that the boundary box completely contains the pituitary gland region, and applying the coordinates of the boundary box to an NIFTI image of 80 multiplied by 11 to obtain the pituitary gland region, thereby completing the positioning of the pituitary gland;
in the step 2, using the output of the pituitary gland positioning of the previous stage, a 2D slice (32 multiplied by 32) is generated as the input of the stage according to the cross section direction;
x i =X[a,b,i],i∈(1,...,N) (3)
wherein X represents the pituitary gland region obtained by locating the pituitary gland, and X i The ith slice of X in the cross-sectional direction, a represents the coronal direction of X, b represents the delocalization direction of X, i represents the cross-sectional direction of X, and N represents the number of cross-sectional layers of X. Wherein a and b are both 32;
finally, outputting a two-dimensional label of the pituitary gland, and then stacking all two-dimensional slices in the cross section direction to reconstruct a three-dimensional image, so that the final three-dimensional label of the pituitary gland is obtained;
X[a,b,i]=x i ,i∈(1,...,N) (4)
wherein X represents a three-dimensional label reconstructed in the cross-sectional direction, X i The ith two-dimensional label, a represents the coronal bit direction of X, b represents the bit losing direction of X, i represents the cross section direction of X, and the number of N two-dimensional labels X.
Examples:
one MRI brain image acquired was cropped to a size of 80 x 11. The image is divided into 49 subblocks with the size of 32 x 11 and is input to the VT-Unet. The 49 outputs were then combined into a mask of size 80 x 11 according to the partitioning method, and finally the mask was taken to the maximum connected area to obtain the final pituitary mask and the 3D bounding box of the pituitary mask was calculated. The coronal and sagittal planes of the bounding box are expanded to a 32 x 32 size and two layers anteriorly and posteriorly in cross-section as the final pituitary positioning output. The output is used to generate 2D slices (32 x 32) from the axial direction as input to the pituitary segmentation stage. Finally, stacking all two-dimensional slices output by the pituitary segmentation in a third dimension to reconstruct a three-dimensional image, and obtaining a final three-dimensional pituitary gland label;
a novel two-stage pituitary segmentation method based on deep learning. The two-stage pituitary segmentation method generally includes two steps, pituitary localization and pituitary segmentation.
As shown in fig. 1: in order to obtain good adaptation of the proposed adenohypophysis segmentation method, the data collected includes three-dimensional brain images (TIWI and T2 WI) of two sequences of three different diseases (dwarf group, precocious group and normal group). The data of a plurality of different diseases are used for participating in training, so that the robustness of the model is improved. Since the proportion of the pituitary gland is too small and the calculation efficiency is improved in order to reduce the calculation amount, in the preprocessing process of the data, the original data is cut into the size of 80 x 11, and compared with the original image, the calculation amount of the model is reduced, and the fitting speed of a network is improved. The preprocessed brain MRI image is divided into 49 sub-blocks with the size of 32 x 11, the sub-blocks are input into an adenohypophysis positioning network, and the output is 49 sub-masks with the size of 32 x 11. The 49 sub-masks are combined into a mask of 80 x 11 size according to the partitioning method, and then the mask is taken to be the largest connected region to obtain the final pituitary mask and the 3D bounding box of the pituitary mask is calculated. The coronal and sagittal planes of the bounding box are expanded to a 32 x 32 size and two layers anteriorly and posteriorly in cross-section as the final output of pituitary positioning. We then input the output 2-dimensional slice into the pituitary segmentation model for pituitary segmentation.
The pituitary segmentation network (PIT-Former) is a two-dimensional network of U-Net architecture fused with CNN and Transfomer. The network includes 4-layer downsampling and 4-layer upsampling. We use ECA blocks to capture local cross-channel interactions and avoid dimension reduction methods to extract features, aimed at guaranteeing efficiency and effectiveness. The characteristics of the encoder are better fused by using a Channel cross fusion transducer (Channel-wise Cross fusion Transformer, CCT block), so that the semantic gap is reduced to improve the segmentation performance. Channel cross-attention module (channel-wise Cross Attention model, CCA block) avoidance dimension reduction approach is used to guide channel and information filtering of the transducer features and decoder features are used to eliminate ambiguity to facilitate feature extraction. The input to the network is a two-dimensional slice of 32 x 32 size, the feature map is doubled after each downsampling, and the feature map becomes 4*4 size after the fourth downsampling. The last feature map is then obtained by four upsampling and CCA blocks. Finally, a Softmax operation is carried out to obtain a final output result. The output result is a two-dimensional mask with the size of 32 x 32.
Normalization is to normalize the data to between [0,1] by centering on the minimum and scaling on the polar difference (maximum-minimum). The training data is subjected to data enhancement by using rotation, translation and overturn so as to reduce the occurrence of the overfitting phenomenon and improve the robustness of the model. In addition, optimization was performed using ADAM. ADAM is an adaptive low-order moment estimation optimizer that utilizes Nesterov momentum. The deep learning model was trained in end-to-end mode using training and validation data, with a batch size of 16, a learning rate of 1e-4, and implemented on a GPU using python 3.7.0 and GeForce RTX 2080ti 11 gb.
In addition, the DICE Coefficients (DCs) at the pixel level are applied to the final feature map for loss function calculation. DC is a statistic for measuring the degree of spatial overlap between two samples.
Ranging from 0 (indicating no spatial overlap) to 1 (indicating complete spatial overlap).
Where a represents a manual label and B represents a label automatically segmented using a deep learning model.
And finally reconstructing the segmentation result into a three-dimensional image to obtain a final pituitary segmentation result.
TABLE 1
Table 1 is a quantitative evaluation table of pituitary segmentation according to the present invention. In order to verify the performance of the PIT-Former network, 7 image segmentation methods were compared on the basis of 5 indexes such as Dice, HD, HD, MD and ASSD. Model performance was tested using 10 fold cross validation.
Claims (4)
1. An MRI brain image adenohypophysis segmentation method is characterized by comprising the following steps of;
step 1: positioning of the pituitary gland:
preprocessing an MRI brain image acquired from a hospital, converting original Dicom format MRI data acquired from the hospital into a three-dimensional image in an NIFTI format by using ITK-SNAP software, and converting the image in the DICOM format into the three-dimensional image in the NIFTI format;
clipping a three-dimensional image in NIFTI format to a size of 80 x 11, and dividing the image into 49 sub-blocks with a size of 32 x 11; then inputting the 49 sub-blocks into VT-Unet to obtain a segmentation result of the 49 sub-blocks, combining 49 outputs into a mask with the size of 80 multiplied by 11 according to a segmentation method, finally calculating a 3D boundary frame of the pituitary gland mask, expanding the coronal position and the sagittal position of the boundary frame to 32 and expanding two layers back and forth on the cross section to ensure that the boundary frame completely contains the pituitary gland region, and applying the coordinates of the boundary frame to an NIFTI image with the size of 80 multiplied by 11 to obtain the pituitary gland region, thereby completing the positioning of the pituitary gland;
step 2: pituitary segmentation:
using the output of the pituitary positioning of the previous stage, 2D slices (32 x 32) are generated from the cross-sectional direction as input to this stage;
finally, outputting a two-dimensional label of the pituitary gland, and then stacking all the two-dimensional slices in the cross section direction to reconstruct a three-dimensional image, thereby obtaining the final three-dimensional pituitary gland label.
2. The method of claim 1, wherein in step 1, the MRI brain image is obtained by using a siemens skyra3.0t superconducting magnetic resonance scanner, and the pituitary scan uses the following sequences and parameters for a total scan time of about 16 minutes; sagittal T1WI sequence (tr=641 ms, te=9.9 ms), sagittal T2WI sequence (tr=3800 ms, te=75 ms).
3. The method of claim 1, wherein in step 1, a sliding window of size 32×32×11 is defined first, the sliding window is initially fixed at the upper right corner, then it slides left or down, each time by 8 units, the sliding window is used to divide the sub-blocks, and finally 49 sub-blocks are obtained, and each sub-block has size 32×32×11. Marking the position of each sub-block before division;
then inputting the 49 subblocks into the VT-Unet to obtain a segmentation result of the 49 subblocks, wherein the segmentation result of each subblock is a mask with the size of 32 multiplied by 11;
mask i =VTUnet(patch i ),i∈(1,...,49) (1)
wherein VTUnet represents VT-Unet, and is a three-dimensional volume transducer model suitable for medical image segmentation i Representing the ith sub-block, mask i Representing a segmentation result corresponding to the ith sub-block;
the 49 outputs are combined into the mask with the size of 80 multiplied by 11 according to the dividing method, and as the mask with the size of 80 multiplied by 11 obtained by the combination has a plurality of discrete points, the subsequent calculation amount can be increased, only a main part is reserved by removing the discrete points by using the method of taking the maximum communication area, and the reserved main part is the pituitary mask;
ademask=f(mask) (2)
wherein f represents the maximum communication function, mask represents 80×80×11 size mask, and adepack represents pituitary mask obtained by removing the maximum communication area;
finally, calculating a 3D boundary box of the pituitary mask, expanding the coronal position and the sagittal position of the boundary box to 32 sizes and expanding the boundary box back and forth in a cross section to ensure that the boundary box completely contains the pituitary region, and applying the coordinates of the boundary box to an NIFTI image of 80 multiplied by 11 to obtain the pituitary region, thereby completing the pituitary positioning.
4. The MRI brain image pituitary segmentation method according to claim 1, wherein in the step 1, the output of the pituitary positioning of the previous stage is used in the step 2, and a 2D slice (32×32) is generated as an input of the stage according to the cross-sectional direction;
x i =X[a,b,i],i∈(1,...,N) (3)
wherein X represents the pituitary gland region obtained by locating the pituitary gland, and X i The ith slice of X in the cross-sectional direction, a represents the coronal direction of X, b represents the delocalization direction of X, i represents the cross-sectional direction of X, and N represents the number of cross-sectional layers of X. Wherein a and b are both 32;
finally, outputting a two-dimensional label of the pituitary gland, and then stacking all two-dimensional slices in the cross section direction to reconstruct a three-dimensional image, so that the final three-dimensional label of the pituitary gland is obtained;
X[a,b,i]=x i ,i∈(1,...,N) (4)
wherein X represents a three-dimensional label reconstructed in the cross-sectional direction, X i The ith two-dimensional label, a represents the coronal bit direction of X, b represents the bit losing direction of X, i represents the cross section direction of X, and the number of N two-dimensional labels X.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202310522974.5A CN116645336B (en) | 2023-05-10 | 2023-05-10 | MRI brain image gland pituitary segmentation method |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202310522974.5A CN116645336B (en) | 2023-05-10 | 2023-05-10 | MRI brain image gland pituitary segmentation method |
Publications (2)
Publication Number | Publication Date |
---|---|
CN116645336A true CN116645336A (en) | 2023-08-25 |
CN116645336B CN116645336B (en) | 2024-05-07 |
Family
ID=87622202
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN202310522974.5A Active CN116645336B (en) | 2023-05-10 | 2023-05-10 | MRI brain image gland pituitary segmentation method |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN116645336B (en) |
Citations (10)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
WO2017210690A1 (en) * | 2016-06-03 | 2017-12-07 | Lu Le | Spatial aggregation of holistically-nested convolutional neural networks for automated organ localization and segmentation in 3d medical scans |
CN107464250A (en) * | 2017-07-03 | 2017-12-12 | 深圳市第二人民医院 | Tumor of breast automatic division method based on three-dimensional MRI image |
CN114066843A (en) * | 2021-11-12 | 2022-02-18 | 烟台大学 | CT brain image segmentation and hematoma measurement method |
CN114581459A (en) * | 2022-02-08 | 2022-06-03 | 浙江大学 | Improved 3D U-Net model-based segmentation method for image region of interest of preschool child lung |
CN114972378A (en) * | 2022-05-24 | 2022-08-30 | 南昌航空大学 | Brain tumor MRI image segmentation method based on mask attention mechanism |
CN115331071A (en) * | 2022-07-19 | 2022-11-11 | 长沙理工大学 | Tuberculous meningoencephalitis prediction method and system based on multi-scale feature map |
US20220370033A1 (en) * | 2021-05-05 | 2022-11-24 | Board Of Trustees Of Southern Illinois University | Three-dimensional modeling and assessment of cardiac tissue |
EP4141790A1 (en) * | 2021-08-30 | 2023-03-01 | Siemens Healthcare GmbH | Method, device and system for automated segmentation of prostate in medical images for tumor detection |
WO2023060944A1 (en) * | 2021-10-11 | 2023-04-20 | 浙江大学 | Liver ct image segmentation system and algorithm based on hybrid supervised learning |
CN116071383A (en) * | 2023-02-23 | 2023-05-05 | 重庆邮电大学 | Hippocampus subzone segmentation method and system based on ultra-high field magnetic resonance image reconstruction |
-
2023
- 2023-05-10 CN CN202310522974.5A patent/CN116645336B/en active Active
Patent Citations (10)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
WO2017210690A1 (en) * | 2016-06-03 | 2017-12-07 | Lu Le | Spatial aggregation of holistically-nested convolutional neural networks for automated organ localization and segmentation in 3d medical scans |
CN107464250A (en) * | 2017-07-03 | 2017-12-12 | 深圳市第二人民医院 | Tumor of breast automatic division method based on three-dimensional MRI image |
US20220370033A1 (en) * | 2021-05-05 | 2022-11-24 | Board Of Trustees Of Southern Illinois University | Three-dimensional modeling and assessment of cardiac tissue |
EP4141790A1 (en) * | 2021-08-30 | 2023-03-01 | Siemens Healthcare GmbH | Method, device and system for automated segmentation of prostate in medical images for tumor detection |
WO2023060944A1 (en) * | 2021-10-11 | 2023-04-20 | 浙江大学 | Liver ct image segmentation system and algorithm based on hybrid supervised learning |
CN114066843A (en) * | 2021-11-12 | 2022-02-18 | 烟台大学 | CT brain image segmentation and hematoma measurement method |
CN114581459A (en) * | 2022-02-08 | 2022-06-03 | 浙江大学 | Improved 3D U-Net model-based segmentation method for image region of interest of preschool child lung |
CN114972378A (en) * | 2022-05-24 | 2022-08-30 | 南昌航空大学 | Brain tumor MRI image segmentation method based on mask attention mechanism |
CN115331071A (en) * | 2022-07-19 | 2022-11-11 | 长沙理工大学 | Tuberculous meningoencephalitis prediction method and system based on multi-scale feature map |
CN116071383A (en) * | 2023-02-23 | 2023-05-05 | 重庆邮电大学 | Hippocampus subzone segmentation method and system based on ultra-high field magnetic resonance image reconstruction |
Non-Patent Citations (2)
Title |
---|
HIMASHI PEIRIS 等: "A Robust Volumetric Transformer for Accurate 3D Tumor Segmentation", 《MICCAI 2022》, pages 162 - 171 * |
戚永军 等: "基于深度混合卷积模型的肺结节检测方法", 《计算机应用》, vol. 40, no. 10, 10 October 2020 (2020-10-10), pages 2904 - 2909 * |
Also Published As
Publication number | Publication date |
---|---|
CN116645336B (en) | 2024-05-07 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN113012172B (en) | AS-UNet-based medical image segmentation method and system | |
CN111563902B (en) | Lung lobe segmentation method and system based on three-dimensional convolutional neural network | |
JP2023550844A (en) | Liver CT automatic segmentation method based on deep shape learning | |
CN110060235A (en) | A kind of thyroid nodule ultrasonic image division method based on deep learning | |
CN112258488A (en) | Medical image focus segmentation method | |
CN111369574B (en) | Thoracic organ segmentation method and device | |
CN114066843B (en) | CT brain image segmentation and hematoma measurement method | |
CN111080657A (en) | CT image organ segmentation method based on convolutional neural network multi-dimensional fusion | |
CN116188452A (en) | Medical image interlayer interpolation and three-dimensional reconstruction method | |
CN113763406B (en) | Infant brain MRI (magnetic resonance imaging) segmentation method based on semi-supervised learning | |
CN113160229A (en) | Pancreas segmentation method and device based on hierarchical supervision cascade pyramid network | |
CN117710671A (en) | Medical image segmentation method based on segmentation large model fine adjustment | |
CN117115132A (en) | Oral cavity CBCT image tooth and soft tissue segmentation model method based on improved U-Net model | |
CN116152500A (en) | Full-automatic tooth CBCT image segmentation method based on deep learning | |
CN116596949A (en) | Medical image segmentation method based on conditional diffusion model | |
CN117237641A (en) | Polyp segmentation method and system based on dual-branch feature fusion network | |
CN116452618A (en) | Three-input spine CT image segmentation method | |
CN115496732A (en) | Semi-supervised heart semantic segmentation algorithm | |
CN116645336B (en) | MRI brain image gland pituitary segmentation method | |
CN115272389A (en) | Aortic dissection method with intimal valve attention module | |
CN111667488B (en) | Medical image segmentation method based on multi-angle U-Net | |
CN117876370B (en) | CT image kidney tumor segmentation system based on three-dimensional axial transducer model | |
CN117611601B (en) | Text-assisted semi-supervised 3D medical image segmentation method | |
CN115018864B (en) | Three-stage liver tumor image segmentation method based on self-adaptive preprocessing | |
CN116362995A (en) | Tooth image restoration method and system based on standard prior |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
GR01 | Patent grant | ||
GR01 | Patent grant | ||
TR01 | Transfer of patent right | ||
TR01 | Transfer of patent right |
Effective date of registration: 20240625 Address after: 264003 Yingrui Industrial Park, building 5, No. 66, Feilong Road, Laishan District, Yantai City, Shandong Province Patentee after: Shandong Zhongjia Yingrui Medical Technology Co.,Ltd. Country or region after: China Address before: No.30 Qingquan Road, Laishan District, Yantai City, Shandong Province Patentee before: Yantai University Country or region before: China |