CN116645336B - MRI brain image gland pituitary segmentation method - Google Patents

MRI brain image gland pituitary segmentation method Download PDF

Info

Publication number
CN116645336B
CN116645336B CN202310522974.5A CN202310522974A CN116645336B CN 116645336 B CN116645336 B CN 116645336B CN 202310522974 A CN202310522974 A CN 202310522974A CN 116645336 B CN116645336 B CN 116645336B
Authority
CN
China
Prior art keywords
pituitary gland
dimensional
image
mask
pituitary
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN202310522974.5A
Other languages
Chinese (zh)
Other versions
CN116645336A (en
Inventor
郑强
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Yantai University
Original Assignee
Yantai University
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Yantai University filed Critical Yantai University
Priority to CN202310522974.5A priority Critical patent/CN116645336B/en
Publication of CN116645336A publication Critical patent/CN116645336A/en
Application granted granted Critical
Publication of CN116645336B publication Critical patent/CN116645336B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/0002Inspection of images, e.g. flaw detection
    • G06T7/0012Biomedical image inspection
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/0464Convolutional networks [CNN, ConvNet]
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/08Learning methods
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T17/00Three dimensional [3D] modelling, e.g. data description of 3D objects
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/10Segmentation; Edge detection
    • G06T7/11Region-based segmentation
    • GPHYSICS
    • G16INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR SPECIFIC APPLICATION FIELDS
    • G16HHEALTHCARE INFORMATICS, i.e. INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR THE HANDLING OR PROCESSING OF MEDICAL OR HEALTHCARE DATA
    • G16H30/00ICT specially adapted for the handling or processing of medical images
    • G16H30/20ICT specially adapted for the handling or processing of medical images for handling medical images, e.g. DICOM, HL7 or PACS
    • GPHYSICS
    • G16INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR SPECIFIC APPLICATION FIELDS
    • G16HHEALTHCARE INFORMATICS, i.e. INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR THE HANDLING OR PROCESSING OF MEDICAL OR HEALTHCARE DATA
    • G16H30/00ICT specially adapted for the handling or processing of medical images
    • G16H30/40ICT specially adapted for the handling or processing of medical images for processing medical images, e.g. editing
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/10Image acquisition modality
    • G06T2207/10072Tomographic images
    • G06T2207/10088Magnetic resonance imaging [MRI]
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/20Special algorithmic details
    • G06T2207/20021Dividing image into blocks, subimages or windows
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/20Special algorithmic details
    • G06T2207/20081Training; Learning
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/20Special algorithmic details
    • G06T2207/20084Artificial neural networks [ANN]
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/30Subject of image; Context of image processing
    • G06T2207/30004Biomedical image processing
    • G06T2207/30016Brain

Landscapes

  • Engineering & Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • Theoretical Computer Science (AREA)
  • Health & Medical Sciences (AREA)
  • General Health & Medical Sciences (AREA)
  • General Physics & Mathematics (AREA)
  • Medical Informatics (AREA)
  • Nuclear Medicine, Radiotherapy & Molecular Imaging (AREA)
  • Radiology & Medical Imaging (AREA)
  • Software Systems (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Data Mining & Analysis (AREA)
  • Public Health (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Primary Health Care (AREA)
  • Artificial Intelligence (AREA)
  • Biomedical Technology (AREA)
  • Biophysics (AREA)
  • Computational Linguistics (AREA)
  • Epidemiology (AREA)
  • Evolutionary Computation (AREA)
  • Molecular Biology (AREA)
  • Computing Systems (AREA)
  • General Engineering & Computer Science (AREA)
  • Mathematical Physics (AREA)
  • Computer Graphics (AREA)
  • Quality & Reliability (AREA)
  • Geometry (AREA)
  • Image Analysis (AREA)
  • Magnetic Resonance Imaging Apparatus (AREA)

Abstract

A method of MRI brain image adenohypophysis segmentation, comprising; preprocessing an MRI brain image acquired from a hospital, converting original Dicom format MRI data acquired from the hospital into a three-dimensional image in an NIFTI format by using ITK-SNAP software, and converting the image in the DICOM format into the three-dimensional image in the NIFTI format; cutting and dividing the three-dimensional image in the NIFTI format into sub-blocks; inputting the subblocks into VT-Unet to obtain a segmentation result, finally calculating a 3D boundary box of the pituitary gland, and applying the coordinates of the boundary box to the NIFTI image to obtain a pituitary gland region, thereby completing pituitary gland positioning; using the output of the pituitary gland positioning of the previous stage, generating a 2D slice according to the cross section direction as the input of the stage; finally, outputting a two-dimensional label of the pituitary gland, and then stacking all the two-dimensional slices in the cross section direction to reconstruct a three-dimensional image, thereby obtaining the final three-dimensional pituitary gland label. The method has the characteristics of high speed, high accuracy, good robustness, high efficiency and good generalization.

Description

MRI brain image gland pituitary segmentation method
Technical Field
The invention belongs to the technical field of clinical pituitary gland segmentation, and particularly relates to a method for segmenting pituitary gland by using an MRI brain image.
Background
Nuclear Magnetic Resonance (MRI) images are currently accepted as the best imaging method for displaying the morphological characteristics of the pituitary gland due to the fact that the MRI images are non-radiative and have high tissue contrast.
The existing pituitary gland segmentation method is mainly manual segmentation, and no method special for full-automatic segmentation of the pituitary gland exists at present. However, artificial-based pituitary segmentation suffers from the problems of being time consuming, subjective and inaccurate. In recent years, deep learning technology has been developed rapidly, and Convolutional Neural Networks (CNNs) have achieved great success in the field of medical image segmentation. The symmetrical codec structure used by U-Net and its variants to improve detail retention is the dominant architecture for medical image segmentation and is widely used for organ and tissue segmentation. ViT (Vision Transformer) is a new trend in the field of computer vision at present, which models remote dependency relationships with their unique global attention, provides a new idea for computer vision and is gradually used for medical image segmentation. However, neither CNN nor transducer is applied to the pituitary segmentation field. Although CNN has strong feature extraction capability, the CNN has a great promotion space for the reservation of global information; transformer is leading in the field of natural language image processing, but has drawbacks in terms of both data amount requirements and generalization ability of models for both computer vision and freshness.
Disclosure of Invention
In order to overcome the technical problems, the invention aims to provide the MRI brain image adenohypophysis segmentation method which has the characteristics of high speed, high accuracy, good robustness, high efficiency and good generalization.
In order to achieve the above purpose, the technical scheme adopted by the invention is as follows:
an MRI brain image adenohypophysis segmentation method comprises the following steps;
Step 1: positioning of the pituitary gland:
Preprocessing an MRI brain image acquired from a hospital, converting original Dicom format MRI data acquired from the hospital into a three-dimensional image in an NIFTI format by using ITK-SNAP software, and converting the image in the DICOM format into the three-dimensional image in the NIFTI format;
clipping a three-dimensional image in NIFTI format to a size of 80 x 11, and dividing the image into 49 sub-blocks with a size of 32 x 11; then inputting the 49 subblocks into VT-Unet to obtain a segmentation result of the 49 subblocks, combining 49 outputs into a mask with the size of 80 multiplied by 11 according to a segmentation method, finally calculating a 3D boundary frame of the pituitary gland mask, expanding the coronal position and the sagittal position of the boundary frame to 32 sizes and expanding two layers back and forth on the cross section to ensure that the boundary frame completely contains the pituitary gland region, and applying the coordinates of the boundary frame to an NIFTI image with the size of 80 multiplied by 11 to obtain the pituitary gland region, thereby completing the positioning of the pituitary gland;
Step 2: pituitary segmentation:
using the output of the pituitary positioning of the previous stage, 2D slices (32 x 32) are generated from the cross-sectional direction as input to this stage;
finally, outputting a two-dimensional label of the pituitary gland, and then stacking all the two-dimensional slices in the cross section direction to reconstruct a three-dimensional image, thereby obtaining the final three-dimensional pituitary gland label.
In the step 1, MRI brain images are obtained by using SIEMENS SKYRA 3.0.0T superconducting magnetic resonance scanner, and pituitary scanning is performed using the following sequences and parameters, and the total scanning time is about 16 minutes; sagittal T1WI sequence (tr=641 ms, te=9.9 ms), sagittal T2WI sequence (tr=3800 ms, te=75 ms).
In the step 1, a sliding window with a size of 32×32×11 is firstly defined, the sliding window is initially fixed at the upper right corner, then the sliding window slides leftwards or downwards, and each time, the sliding unit is too small, so that adjacent image blocks contain a large amount of repeated information, and training resources are wasted; too large a sliding unit may result in too little training data, considering comprehensively that 8 units per sliding are selected, the sub-blocks are divided by the sliding window, and finally 49 sub-blocks are obtained, and each sub-block has a size of 32×32×11. Marking the position of each sub-block before division;
then inputting the 49 subblocks into VT-Unet to obtain a segmentation result of the 49 subblocks, wherein the segmentation result of each subblock is a mask with the size of 32 multiplied by 11;
maski=VTUnet(patchi),i∈(1,...,49) (1)
Wherein VTUnet represents VT-Unet, which is a three-dimensional volume transducer model suitable for medical image segmentation, patch i represents the ith sub-block, and mask i represents the segmentation result corresponding to the ith sub-block;
The 49 outputs are combined into the mask with the size of 80 multiplied by 11 according to the dividing method, and as the mask with the size of 80 multiplied by 11 obtained by the combination has a plurality of discrete points, the subsequent calculation amount can be increased, only a main part is reserved by removing the discrete points by using the method of taking the maximum communication area, and the reserved main part is the pituitary mask;
ademask=f(mask) (2)
wherein f represents the maximum communication function, mask represents 80×80×11 size mask, ademask represents pituitary mask obtained by removing the maximum communication region;
Finally, calculating a 3D boundary box of the pituitary gland mask, expanding the coronal position and the sagittal position of the boundary box to 32 sizes and expanding the boundary box back and forth on the cross section to ensure that the boundary box completely contains the pituitary gland region, and applying the coordinates of the boundary box to an NIFTI image of 80 multiplied by 11 to obtain the pituitary gland region, thereby completing the positioning of the pituitary gland;
in the step 2, using the output of the pituitary gland positioning of the previous stage, a 2D slice (32 multiplied by 32) is generated as the input of the stage according to the cross section direction;
xi=X[a,b,i],i∈(1,...,N) (3)
wherein X represents an adenohypophysis region obtained by locating an adenohypophysis, X i represents an ith slice of X in a cross section direction, a represents a coronal position direction of X, b represents an aphrontal position direction of X, i represents a cross section direction of X, and N represents a cross section layer number of X. Wherein a and b are both 32;
Finally, outputting a two-dimensional label of the pituitary gland, and then stacking all two-dimensional slices in the cross section direction to reconstruct a three-dimensional image, so that the final three-dimensional label of the pituitary gland is obtained;
X[a,b,i]=xi,i∈(1,...,N) (4)
wherein X represents a three-dimensional label reconstructed in the cross section direction, X i is an ith two-dimensional label, a represents the coronal bit direction of X, b represents the losing bit direction of X, i represents the cross section direction of X, and the number of N two-dimensional labels X.
The invention has the beneficial effects of.
The invention adopts a segmentation method based on two stages of deep learning, and has the advantages of high segmentation speed and high segmentation result accuracy (DICE= 0.9013). The data set uses brain MRI images of different diseases and different sequences, so that the robustness and generalization are good. Quantitative and qualitative assessment of pituitary segmentation are shown in table 1 and fig. 4. In table 1, the segmentation index of the method of this patent is higher than that of the traditional CNN model, viT (visual transducer) model and HyBird (hybrid CNN and transducer) model, showing higher segmentation accuracy. Fig. 4 shows that the method of this patent has a better segmentation effect than other methods.
Description of the drawings:
FIG. 1 is a schematic representation of an image segmentation sub-block in pituitary positioning according to the present invention.
FIG. 2 is a schematic diagram of an adenohypophysis segmentation convolutional neural network (PIT-Former) according to the present invention.
FIG. 3 is an expanded view of the pituitary positioning frame produced in the pituitary positioning stage of the present invention.
FIG. 4 is a qualitative assessment of pituitary segmentation according to the present invention. Pituitary segmentation examples of different methods for randomly selecting three patients, a1-a2, b1-b2, c1-c2 represent T1WI and T2WI, respectively, of three patients, PIT-force being the method presented in this section. The white line encircled portion represents the label.
Detailed Description
The following description of the embodiments of the present invention will be made clearly and completely with reference to the accompanying drawings, in which it is apparent that the embodiments described are only some embodiments of the present invention, but not all embodiments. All other embodiments, which can be made by those skilled in the art based on the embodiments of the invention without making any inventive effort, are intended to be within the scope of the invention.
An MRI brain image adenohypophysis segmentation method comprises the following steps;
Step 1: positioning of the pituitary gland:
Preprocessing an MRI brain image acquired from a hospital, converting original Dicom format MRI data acquired from the hospital into a three-dimensional image in an NIFTI format by using ITK-SNAP software, and converting the image in the DICOM format into the three-dimensional image in the NIFTI format;
clipping a three-dimensional image in NIFTI format to a size of 80 x 11, and dividing the image into 49 sub-blocks with a size of 32 x 11; then inputting the 49 subblocks into VT-Unet to obtain a segmentation result of the 49 subblocks, combining 49 outputs into a mask with the size of 80 multiplied by 11 according to a segmentation method, finally calculating a 3D boundary frame of the pituitary gland mask, expanding the coronal position and the sagittal position of the boundary frame to 32 sizes and expanding two layers back and forth on the cross section to ensure that the boundary frame completely contains the pituitary gland region, and applying the coordinates of the boundary frame to an NIFTI image with the size of 80 multiplied by 11 to obtain the pituitary gland region, thereby completing the positioning of the pituitary gland;
Specific expansion method as shown in fig. 3, the inner frame represents before expansion, and the outer frame represents after expansion.
Step 2: pituitary segmentation:
The purpose of pituitary segmentation is to accurately segment the pituitary gland on specific sections determined at the pituitary positioning stage, pituitary segmentation using PIT-force (FIG. 2). PIT-Former is a two-dimensional network fusing CNN and a transducer;
using the output of the pituitary positioning of the previous stage, 2D slices (32 x 32) are generated from the cross-sectional direction as input to this stage;
finally, outputting a two-dimensional label of the pituitary gland, and then stacking all the two-dimensional slices in the cross section direction to reconstruct a three-dimensional image, thereby obtaining the final three-dimensional pituitary gland label.
In the step 1, MRI brain images are obtained by using SIEMENS SKYRA 3.0.0T superconducting magnetic resonance scanner, and pituitary scanning is performed using the following sequences and parameters, and the total scanning time is about 16 minutes; sagittal T1WI sequence (tr=641 ms, te=9.9 ms), sagittal T2WI sequence (tr=3800 ms, te=75 ms).
In the step 1, a sliding window with a size of 32×32×11 is firstly defined, the sliding window is initially fixed at the upper right corner (white), then the sliding window slides leftwards or downwards, and each time, the sliding unit is too small, so that adjacent image blocks contain a large amount of repeated information, and training resources are wasted; too large a sliding unit may result in too little training data, considering comprehensively that 8 units per sliding are selected, the sub-blocks are divided by the sliding window, and finally 49 sub-blocks are obtained, and each sub-block has a size of 32×32×11. Marking the position of each sub-block before division;
then inputting the 49 subblocks into VT-Unet to obtain a segmentation result of the 49 subblocks, wherein the segmentation result of each subblock is a mask with the size of 32 multiplied by 11;
maski=VTUnet(patchi),i∈(1,...,49) (1)
Wherein VTUnet represents VT-Unet, which is a three-dimensional volume transducer model suitable for medical image segmentation, patch i represents the ith sub-block, and mask i represents the segmentation result corresponding to the ith sub-block;
The 49 outputs are combined into the mask with the size of 80 multiplied by 11 according to the dividing method, and as the mask with the size of 80 multiplied by 11 obtained by the combination has a plurality of discrete points, the subsequent calculation amount can be increased, only a main part is reserved by removing the discrete points by using the method of taking the maximum communication area, and the reserved main part is the pituitary mask;
ademask=f(mask) (2)
wherein f represents the maximum communication function, mask represents 80×80×11 size mask, ademask represents pituitary mask obtained by removing the maximum communication region;
Finally, calculating a 3D boundary box of the pituitary gland mask, expanding the coronal position and the sagittal position of the boundary box to 32 sizes and expanding the boundary box back and forth on the cross section to ensure that the boundary box completely contains the pituitary gland region, and applying the coordinates of the boundary box to an NIFTI image of 80 multiplied by 11 to obtain the pituitary gland region, thereby completing the positioning of the pituitary gland;
in the step 2, using the output of the pituitary gland positioning of the previous stage, a 2D slice (32 multiplied by 32) is generated as the input of the stage according to the cross section direction;
xi=X[a,b,i],i∈(1,...,N) (3)
wherein X represents an adenohypophysis region obtained by locating an adenohypophysis, X i represents an ith slice of X in a cross section direction, a represents a coronal position direction of X, b represents an aphrontal position direction of X, i represents a cross section direction of X, and N represents a cross section layer number of X. Wherein a and b are both 32;
Finally, outputting a two-dimensional label of the pituitary gland, and then stacking all two-dimensional slices in the cross section direction to reconstruct a three-dimensional image, so that the final three-dimensional label of the pituitary gland is obtained;
X[a,b,i]=xi,i∈(1,...,N) (4)
wherein X represents a three-dimensional label reconstructed in the cross section direction, X i is an ith two-dimensional label, a represents the coronal bit direction of X, b represents the losing bit direction of X, i represents the cross section direction of X, and the number of N two-dimensional labels X.
Examples:
One MRI brain image acquired was cropped to a size of 80 x 11. The image is divided into 49 sub-blocks of 32 x 11 size and input to VT-Unet. The 49 outputs were then combined into a mask of size 80 x 11 according to the partitioning method, and finally the mask was taken to the maximum connected area to obtain the final pituitary mask and the 3D bounding box of the pituitary mask was calculated. The coronal and sagittal planes of the bounding box are expanded to a 32 x 32 size and two layers anteriorly and posteriorly in cross-section as the final pituitary positioning output. The output is used to generate 2D slices (32 x 32) from the axial direction as input to the pituitary segmentation stage. Finally, stacking all two-dimensional slices output by the pituitary segmentation in a third dimension to reconstruct a three-dimensional image, and obtaining a final three-dimensional pituitary gland label;
A novel two-stage pituitary segmentation method based on deep learning. The two-stage pituitary segmentation method generally includes two steps, pituitary localization and pituitary segmentation.
As shown in fig. 1: in order to obtain good adaptation of the proposed adenohypophysis segmentation method, the data collected includes two sequential (TIWI and T2 WI) three-dimensional brain images of three different diseases (dwarf, precocious and normal). The data of a plurality of different diseases are used for participating in training, so that the robustness of the model is improved. Since the proportion of the pituitary gland is too small and the calculation efficiency is improved in order to reduce the calculation amount, in the preprocessing process of the data, the original data is cut into the size of 80 x 11, and compared with the original image, the calculation amount of the model is reduced, and the fitting speed of a network is improved. The preprocessed brain MRI image is divided into 49 sub-blocks with the size of 32 x 11, the sub-blocks are input into an adenohypophysis positioning network, and the output is 49 sub-masks with the size of 32 x 11. The 49 sub-masks are combined into a mask of 80 x 11 size according to the partitioning method, and then the mask is taken to be the largest connected region to obtain the final pituitary mask and the 3D bounding box of the pituitary mask is calculated. The coronal and sagittal planes of the bounding box are expanded to a 32 x 32 size and two layers anteriorly and posteriorly in cross-section as the final output of pituitary positioning. We then input the output 2-dimensional slice into the pituitary segmentation model for pituitary segmentation.
The pituitary segmentation network (PIT-Former) is a two-dimensional network of U-Net architecture fused with CNN and Transfomer. The network includes 4-layer downsampling and 4-layer upsampling. We use ECA blocks to capture local cross-channel interactions and avoid dimension reduction methods to extract features, aimed at guaranteeing efficiency and effectiveness. The characteristics of the encoder are better fused by using a Channel cross fusion transducer (Channel-wise Cross fusion Transformer, CCT block), so that the semantic gap is reduced to improve the segmentation performance. Channel cross-attention module (channel-wise Cross Attention model, CCA block) avoidance dimension reduction approach is used to guide channel and information filtering of the transducer features and decoder features are used to eliminate ambiguity to facilitate feature extraction. The input to the network is a two-dimensional slice of 32 x 32 size, the feature map is doubled after each downsampling, and the feature map becomes 4*4 size after the fourth downsampling. The last feature map is then obtained by four upsampling and CCA blocks. Finally, a Softmax operation is carried out to obtain a final output result. The output result is a two-dimensional mask with the size of 32 x 32.
Normalization is to normalize the data to between [0,1] by centering on the minimum and scaling on the polar difference (maximum-minimum). The training data is subjected to data enhancement by using rotation, translation and overturn so as to reduce the occurrence of the overfitting phenomenon and improve the robustness of the model. In addition, optimization was performed using ADAM. ADAM is an adaptive low-order moment estimation optimizer that utilizes Nesterov momentum. The deep learning model was trained in end-to-end mode using training and validation data, batchsize size 16, learning rate 1e-4, and implemented on a GPU using python 3.7.0 and GeForce RTX 2080ti 11 gb.
In addition, the DICE Coefficients (DCs) at the pixel level are applied to the final feature map for loss function calculation. DC is a statistic for measuring the degree of spatial overlap between two samples.
Ranging from 0 (indicating no spatial overlap) to 1 (indicating complete spatial overlap).
Where a represents a manual label and B represents a label automatically segmented using a deep learning model.
And finally reconstructing the segmentation result into a three-dimensional image to obtain a final pituitary segmentation result.
TABLE 1
Table 1 is a quantitative evaluation table of pituitary segmentation according to the present invention. In order to verify the performance of the PIT-Former network, 7 image segmentation methods were compared on the basis of Dice, HD, HD indexes such as MD and ASSD. Model performance was tested using 10 fold cross validation.

Claims (2)

1. An MRI brain image adenohypophysis segmentation method is characterized by comprising the following steps of;
Step 1: positioning of the pituitary gland:
Preprocessing an MRI brain image acquired from a hospital, converting original Dicom format MRI data acquired from the hospital into a three-dimensional image in an NIFTI format by using ITK-SNAP software, and converting the image in the DICOM format into the three-dimensional image in the NIFTI format;
clipping a three-dimensional image in NIFTI format to a size of 80 x 11, and dividing the image into 49 sub-blocks with a size of 32 x 11; then inputting the 49 subblocks into VT-Unet to obtain a segmentation result of the 49 subblocks, combining 49 outputs into a mask with the size of 80 multiplied by 11 according to a segmentation method, finally calculating a 3D boundary frame of the pituitary gland mask, expanding the coronal position and the sagittal position of the boundary frame to 32 sizes and expanding two layers back and forth on the cross section to ensure that the boundary frame completely contains the pituitary gland region, and applying the coordinates of the boundary frame to an NIFTI image with the size of 80 multiplied by 11 to obtain the pituitary gland region, thereby completing the positioning of the pituitary gland;
Step 2: pituitary segmentation:
using the output of the pituitary gland positioning of the previous stage, generating a 2D slice according to the cross section direction as the input of the stage;
Finally, outputting a two-dimensional label of the pituitary gland, and then stacking all two-dimensional slices in the cross section direction to reconstruct a three-dimensional image, so that the final three-dimensional label of the pituitary gland is obtained;
In the step 1, a sliding window with a size of 32×32×11 is first defined, the sliding window is initially fixed at the upper right corner, then it slides leftwards or downwards, each time by 8 units, the sliding window is used for dividing sub-blocks, and finally 49 sub-blocks are obtained, and the size of each sub-block is 32×32×11; marking the position of each sub-block before division;
then inputting the 49 subblocks into VT-Unet to obtain a segmentation result of the 49 subblocks, wherein the segmentation result of each subblock is a mask with the size of 32 multiplied by 11;
maski=VTUnet(patchi),i∈(1,...,49) (1)
Wherein VTUnet represents VT-Unet, which is a three-dimensional volume transducer model suitable for medical image segmentation, patch i represents the ith sub-block, and mask i represents the segmentation result corresponding to the ith sub-block;
The 49 outputs are combined into the mask with the size of 80 multiplied by 11 according to the dividing method, and as the mask with the size of 80 multiplied by 11 obtained by the combination has a plurality of discrete points, the subsequent calculation amount can be increased, only a main part is reserved by removing the discrete points by using the method of taking the maximum communication area, and the reserved main part is the pituitary mask;
ademask=f(mask) (2)
wherein f represents the maximum communication function, mask represents 80×80×11 size mask, ademask represents pituitary mask obtained by removing the maximum communication region;
Finally, calculating a 3D boundary box of the pituitary gland mask, expanding the coronal position and the sagittal position of the boundary box to 32 sizes and expanding the boundary box back and forth on the cross section to ensure that the boundary box completely contains the pituitary gland region, and applying the coordinates of the boundary box to an NIFTI image of 80 multiplied by 11 to obtain the pituitary gland region, thereby completing the positioning of the pituitary gland;
in the step 2, using the output of the pituitary gland positioning in the previous stage, generating a 2D slice according to the cross section direction as the input of the stage;
xi=X[a,b,i],i∈(1,...,N) (3)
Wherein X represents an adenohypophysis region obtained by locating an adenohypophysis, X i represents an ith slice of X in the cross section direction, a represents the coronal direction of X, b represents the sagittal direction of X, i represents the cross section direction of X, and N represents the number of layers of X in the cross section; wherein a and b are both 32;
Finally, outputting a two-dimensional label of the pituitary gland, and then stacking all two-dimensional slices in the cross section direction to reconstruct a three-dimensional image, so that the final three-dimensional label of the pituitary gland is obtained;
Y[a,b,i]=yi,i∈(1,...,N) (4)
wherein Y represents a three-dimensional label reconstructed in the cross-sectional direction, Y i represents an ith two-dimensional label, a represents the coronal direction of X, b represents the sagittal direction of X, i represents the cross-sectional direction of X, and N represents the number of two-dimensional labels Y.
2. The method according to claim 1, wherein in step 1, the MRI brain image is obtained by using SIEMENS SKYRA 3.0.0T superconducting magnetic resonance scanner, and the pituitary scan uses the following sequences and parameters for a total scan time of 16 minutes; sagittal T1WI sequence, tr=641 ms, te=9.9 ms, sagittal T2WI sequence, tr=3800 ms, te=75 ms.
CN202310522974.5A 2023-05-10 2023-05-10 MRI brain image gland pituitary segmentation method Active CN116645336B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202310522974.5A CN116645336B (en) 2023-05-10 2023-05-10 MRI brain image gland pituitary segmentation method

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202310522974.5A CN116645336B (en) 2023-05-10 2023-05-10 MRI brain image gland pituitary segmentation method

Publications (2)

Publication Number Publication Date
CN116645336A CN116645336A (en) 2023-08-25
CN116645336B true CN116645336B (en) 2024-05-07

Family

ID=87622202

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202310522974.5A Active CN116645336B (en) 2023-05-10 2023-05-10 MRI brain image gland pituitary segmentation method

Country Status (1)

Country Link
CN (1) CN116645336B (en)

Citations (9)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2017210690A1 (en) * 2016-06-03 2017-12-07 Lu Le Spatial aggregation of holistically-nested convolutional neural networks for automated organ localization and segmentation in 3d medical scans
CN107464250A (en) * 2017-07-03 2017-12-12 深圳市第二人民医院 Tumor of breast automatic division method based on three-dimensional MRI image
CN114066843A (en) * 2021-11-12 2022-02-18 烟台大学 CT brain image segmentation and hematoma measurement method
CN114581459A (en) * 2022-02-08 2022-06-03 浙江大学 Improved 3D U-Net model-based segmentation method for image region of interest of preschool child lung
CN114972378A (en) * 2022-05-24 2022-08-30 南昌航空大学 Brain tumor MRI image segmentation method based on mask attention mechanism
CN115331071A (en) * 2022-07-19 2022-11-11 长沙理工大学 Tuberculous meningoencephalitis prediction method and system based on multi-scale feature map
EP4141790A1 (en) * 2021-08-30 2023-03-01 Siemens Healthcare GmbH Method, device and system for automated segmentation of prostate in medical images for tumor detection
WO2023060944A1 (en) * 2021-10-11 2023-04-20 浙江大学 Liver ct image segmentation system and algorithm based on hybrid supervised learning
CN116071383A (en) * 2023-02-23 2023-05-05 重庆邮电大学 Hippocampus subzone segmentation method and system based on ultra-high field magnetic resonance image reconstruction

Family Cites Families (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20220370033A1 (en) * 2021-05-05 2022-11-24 Board Of Trustees Of Southern Illinois University Three-dimensional modeling and assessment of cardiac tissue

Patent Citations (9)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2017210690A1 (en) * 2016-06-03 2017-12-07 Lu Le Spatial aggregation of holistically-nested convolutional neural networks for automated organ localization and segmentation in 3d medical scans
CN107464250A (en) * 2017-07-03 2017-12-12 深圳市第二人民医院 Tumor of breast automatic division method based on three-dimensional MRI image
EP4141790A1 (en) * 2021-08-30 2023-03-01 Siemens Healthcare GmbH Method, device and system for automated segmentation of prostate in medical images for tumor detection
WO2023060944A1 (en) * 2021-10-11 2023-04-20 浙江大学 Liver ct image segmentation system and algorithm based on hybrid supervised learning
CN114066843A (en) * 2021-11-12 2022-02-18 烟台大学 CT brain image segmentation and hematoma measurement method
CN114581459A (en) * 2022-02-08 2022-06-03 浙江大学 Improved 3D U-Net model-based segmentation method for image region of interest of preschool child lung
CN114972378A (en) * 2022-05-24 2022-08-30 南昌航空大学 Brain tumor MRI image segmentation method based on mask attention mechanism
CN115331071A (en) * 2022-07-19 2022-11-11 长沙理工大学 Tuberculous meningoencephalitis prediction method and system based on multi-scale feature map
CN116071383A (en) * 2023-02-23 2023-05-05 重庆邮电大学 Hippocampus subzone segmentation method and system based on ultra-high field magnetic resonance image reconstruction

Non-Patent Citations (2)

* Cited by examiner, † Cited by third party
Title
A Robust Volumetric Transformer for Accurate 3D Tumor Segmentation;Himashi Peiris 等;《MICCAI 2022》;参见第162-171页 *
基于深度混合卷积模型的肺结节检测方法;戚永军 等;《计算机应用》;20201010;第40卷(第10期);第2904-2909页 *

Also Published As

Publication number Publication date
CN116645336A (en) 2023-08-25

Similar Documents

Publication Publication Date Title
CN113012172B (en) AS-UNet-based medical image segmentation method and system
CN111563902B (en) Lung lobe segmentation method and system based on three-dimensional convolutional neural network
CN110097131B (en) Semi-supervised medical image segmentation method based on countermeasure cooperative training
JP2023550844A (en) Liver CT automatic segmentation method based on deep shape learning
CN112258488A (en) Medical image focus segmentation method
CN111369574B (en) Thoracic organ segmentation method and device
CN114066843B (en) CT brain image segmentation and hematoma measurement method
CN113763406B (en) Infant brain MRI (magnetic resonance imaging) segmentation method based on semi-supervised learning
CN112734755A (en) Lung lobe segmentation method based on 3D full convolution neural network and multitask learning
CN116309648A (en) Medical image segmentation model construction method based on multi-attention fusion
CN115578406B (en) CBCT jaw bone region segmentation method and system based on context fusion mechanism
CN111091575A (en) Medical image segmentation method based on reinforcement learning method
CN116596949A (en) Medical image segmentation method based on conditional diffusion model
CN113160229A (en) Pancreas segmentation method and device based on hierarchical supervision cascade pyramid network
CN116468732A (en) Lung CT image segmentation method and imaging method based on deep learning
CN116152500A (en) Full-automatic tooth CBCT image segmentation method based on deep learning
CN116452618A (en) Three-input spine CT image segmentation method
WO2024104035A1 (en) Long short-term memory self-attention model-based three-dimensional medical image segmentation method and system
CN111667488B (en) Medical image segmentation method based on multi-angle U-Net
CN116645336B (en) MRI brain image gland pituitary segmentation method
CN117115132A (en) Oral cavity CBCT image tooth and soft tissue segmentation model method based on improved U-Net model
CN116362995A (en) Tooth image restoration method and system based on standard prior
CN115018864B (en) Three-stage liver tumor image segmentation method based on self-adaptive preprocessing
CN115496732A (en) Semi-supervised heart semantic segmentation algorithm
CN116109605B (en) Medical image tumor segmentation system, training set construction method and model training method

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant