CN114519770A - Lumbosacral plexus nerve segmentation and three-dimensional visualization method - Google Patents
Lumbosacral plexus nerve segmentation and three-dimensional visualization method Download PDFInfo
- Publication number
- CN114519770A CN114519770A CN202210016806.4A CN202210016806A CN114519770A CN 114519770 A CN114519770 A CN 114519770A CN 202210016806 A CN202210016806 A CN 202210016806A CN 114519770 A CN114519770 A CN 114519770A
- Authority
- CN
- China
- Prior art keywords
- image
- feature map
- convolution
- layer
- dimensional
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Pending
Links
- 230000011218 segmentation Effects 0.000 title claims abstract description 38
- 210000002988 lumbosacral plexus Anatomy 0.000 title claims abstract description 27
- 238000007794 visualization technique Methods 0.000 title claims abstract description 12
- 238000002595 magnetic resonance imaging Methods 0.000 claims abstract description 17
- 238000012545 processing Methods 0.000 claims abstract description 16
- 238000010606 normalization Methods 0.000 claims abstract description 10
- 238000013528 artificial neural network Methods 0.000 claims abstract description 8
- 239000003814 drug Substances 0.000 claims abstract description 5
- 238000003384 imaging method Methods 0.000 claims abstract description 5
- 238000004891 communication Methods 0.000 claims abstract description 4
- 238000010586 diagram Methods 0.000 claims description 31
- 238000000034 method Methods 0.000 claims description 18
- 239000013598 vector Substances 0.000 claims description 12
- 238000011176 pooling Methods 0.000 claims description 11
- 230000004913 activation Effects 0.000 claims description 9
- 238000012800 visualization Methods 0.000 claims description 5
- 238000013507 mapping Methods 0.000 claims description 3
- 238000009877 rendering Methods 0.000 claims description 3
- 230000001537 neural effect Effects 0.000 claims description 2
- 210000005036 nerve Anatomy 0.000 abstract description 10
- 201000010099 disease Diseases 0.000 abstract description 4
- 208000037265 diseases, disorders, signs and symptoms Diseases 0.000 abstract description 4
- 210000003484 anatomy Anatomy 0.000 abstract description 3
- 238000003745 diagnosis Methods 0.000 abstract description 3
- 230000006870 function Effects 0.000 description 18
- 238000013473 artificial intelligence Methods 0.000 description 5
- 238000001356 surgical procedure Methods 0.000 description 4
- 238000013459 approach Methods 0.000 description 3
- 238000002594 fluoroscopy Methods 0.000 description 3
- 238000003709 image segmentation Methods 0.000 description 3
- 238000011161 development Methods 0.000 description 2
- 230000018109 developmental process Effects 0.000 description 2
- 238000005516 engineering process Methods 0.000 description 2
- 238000000605 extraction Methods 0.000 description 2
- 210000001519 tissue Anatomy 0.000 description 2
- 208000010392 Bone Fractures Diseases 0.000 description 1
- 206010010214 Compression fracture Diseases 0.000 description 1
- 206010017076 Fracture Diseases 0.000 description 1
- 206010029174 Nerve compression Diseases 0.000 description 1
- 206010033799 Paralysis Diseases 0.000 description 1
- 206010046543 Urinary incontinence Diseases 0.000 description 1
- 230000002159 abnormal effect Effects 0.000 description 1
- 230000009286 beneficial effect Effects 0.000 description 1
- 230000015556 catabolic process Effects 0.000 description 1
- 238000003759 clinical diagnosis Methods 0.000 description 1
- 238000013170 computed tomography imaging Methods 0.000 description 1
- 230000006837 decompression Effects 0.000 description 1
- 238000013135 deep learning Methods 0.000 description 1
- 238000006731 degradation reaction Methods 0.000 description 1
- 239000006185 dispersion Substances 0.000 description 1
- 230000000694 effects Effects 0.000 description 1
- 230000003902 lesion Effects 0.000 description 1
- 210000003141 lower extremity Anatomy 0.000 description 1
- 230000001717 pathogenic effect Effects 0.000 description 1
- 230000008447 perception Effects 0.000 description 1
- 230000005855 radiation Effects 0.000 description 1
- 238000011160 research Methods 0.000 description 1
- 210000004872 soft tissue Anatomy 0.000 description 1
- 208000005198 spinal stenosis Diseases 0.000 description 1
- 208000024891 symptom Diseases 0.000 description 1
- 238000012549 training Methods 0.000 description 1
- 230000000007 visual effect Effects 0.000 description 1
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T17/00—Three dimensional [3D] modelling, e.g. data description of 3D objects
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/04—Architecture, e.g. interconnection topology
- G06N3/045—Combinations of networks
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/04—Architecture, e.g. interconnection topology
- G06N3/048—Activation functions
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/08—Learning methods
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
- G06T7/0002—Inspection of images, e.g. flaw detection
- G06T7/0012—Biomedical image inspection
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
- G06T7/10—Segmentation; Edge detection
- G06T7/11—Region-based segmentation
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/10—Image acquisition modality
- G06T2207/10072—Tomographic images
- G06T2207/10088—Magnetic resonance imaging [MRI]
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/20—Special algorithmic details
- G06T2207/20081—Training; Learning
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/20—Special algorithmic details
- G06T2207/20084—Artificial neural networks [ANN]
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/30—Subject of image; Context of image processing
- G06T2207/30004—Biomedical image processing
Landscapes
- Engineering & Computer Science (AREA)
- Physics & Mathematics (AREA)
- Theoretical Computer Science (AREA)
- General Physics & Mathematics (AREA)
- General Health & Medical Sciences (AREA)
- Health & Medical Sciences (AREA)
- Software Systems (AREA)
- Computing Systems (AREA)
- Artificial Intelligence (AREA)
- Evolutionary Computation (AREA)
- Computational Linguistics (AREA)
- Molecular Biology (AREA)
- Biophysics (AREA)
- General Engineering & Computer Science (AREA)
- Biomedical Technology (AREA)
- Mathematical Physics (AREA)
- Data Mining & Analysis (AREA)
- Life Sciences & Earth Sciences (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Computer Graphics (AREA)
- Geometry (AREA)
- Medical Informatics (AREA)
- Nuclear Medicine, Radiotherapy & Molecular Imaging (AREA)
- Radiology & Medical Imaging (AREA)
- Quality & Reliability (AREA)
- Magnetic Resonance Imaging Apparatus (AREA)
Abstract
The invention discloses a lumbosacral plexus nerve segmentation and three-dimensional visualization method, which comprises the following steps: (1) processing data; acquiring MRI (magnetic resonance imaging) image data of a patient, storing the MRI image data in a DICOM (digital imaging and communications in medicine) format, converting the acquired data into a gray image, storing the gray image in a png format, and performing normalization processing on the image; (2) adopting a deep neural network to perform segmentation processing on the normalized image obtained in the step (1); (3) three-dimensional reconstruction; stacking the divided two-dimensional plane images, giving an original voxel space, performing three-dimensional reconstruction by using a VTK package and adopting a moving cube algorithm, and displaying the internal details of a reconstructed object to a user by using an isosurface structure. The invention can visually display the relative position of the three-dimensional anatomical structure of the nerve in the spine and can assist in clinical disease diagnosis.
Description
Technical Field
The invention relates to the technical field of three-dimensional visualization of medical images, in particular to a lumbosacral plexus nerve segmentation and three-dimensional visualization method.
Background
With the development of the field of artificial intelligence, the application of the artificial intelligence method in the medical field is gradually widened. Meanwhile, the computer algorithm and the medicine are crossed and penetrated into the whole medical field, and the number of the system developed by utilizing the artificial intelligence is not as large as possible. Medical safety is very important in the process of clinical diagnosis and surgery, and only in combination with the development of clinical application is a valuable artificial intelligence product. Most of the current spinal related procedures are performed in a recumbent position. The operation difficulty is increased while the convenience is provided for the operation. For patients with spinal stenosis, physicians often perform spinal decompression internal fixation procedures to reduce nerve compression and thereby relieve the patient's symptoms. For patients with compression fracture of the vertebral body, a doctor usually performs percutaneous vertebroplasty to enhance the hardness of the vertebral body and prevent secondary fracture, and for serious patients, the doctor improves the fractured vertebral body by applying internal fixing screws. The above operations have a common characteristic, the operation process is also performed through posterior approach, the operation mode can reach nerves in the vertebral canal to a great extent, the problems of lower limb paralysis, incontinence of urine and stool and the like of a patient can be caused under severe conditions, the fact that the patient is difficult to accept is brought, and medical disputes can also be caused. Therefore, how to utilize a visualization technology is important to enable a doctor to have a sufficiently clear understanding of the relative position of the nerve and other tissues before an operation, and especially for the advanced planning of some rare cases, the operation efficiency of the doctor can be greatly improved, and the risk of operation failure can be reduced. For most spine surgeons, the position of the lesion is determined by combining unilateral X-ray film, CT and MRI on an experienced basis for the patients to operate, and the relative position between the patients is difficult to completely show by the plane imaging mode. Even though the X-ray fluoroscopy is only performed by using a C-arm machine during the operation, the X-ray fluoroscopy cannot present related soft tissues such as nerves, and the radiation quantity of doctors and patients is greatly increased due to multiple X-ray fluoroscopy. In addition, the operation is inevitably performed by experience only, so that certain contingency exists, and the operation risk is greatly increased. The application of deep learning in the medical field becomes a key tool for medical image segmentation. In recent years, interest in spine segmentation using MR images has been renewed. Much research has focused on the location, identification, and segmentation of vertebral bodies. However, only a few studies have focused on the segmentation of the lumbosacral plexus. The segmentation of the lumbosacral plexus nerve is a key step for searching abnormal structures and possible pathogenic factors of the lumbosacral plexus nerve diseases. However, segmenting the lumbosacral plexus from MR images remains a challenging task. There is generally low contrast around the lumbosacral plexus, and secondly, the lumbosacral nerve plexus is extremely complex in structure, with different patient neuromorphies. Furthermore, manually marking the lumbosacral plexus from the MR image is not only time consuming, but is also prone to error even for the surgeon. Therefore, there is a need for a method capable of performing preoperative three-dimensional visualization to reduce the risk of surgery, thereby improving the success rate of surgery in clinical spinal surgery.
Disclosure of Invention
The invention aims to provide a lumbosacral plexus nerve segmentation and three-dimensional visualization method, which can intuitively display the relative position of a three-dimensional anatomical structure of a nerve in a spine and can assist in clinical disease diagnosis.
In order to solve the technical problem, the invention provides a lumbosacral plexus nerve segmentation and three-dimensional visualization method, which comprises the following steps:
(1) processing data; acquiring MRI (magnetic resonance imaging) image data of a patient, storing the MRI image data in a DICOM (digital imaging and communications in medicine) format, converting the acquired data into a gray image, storing the gray image in a png format, and performing normalization processing on the image;
(2) adopting a deep neural network to perform segmentation processing on the normalized image obtained in the step (1);
(3) three-dimensional reconstruction; stacking the divided two-dimensional plane images, giving an original voxel space, performing three-dimensional reconstruction by using a VTK package and adopting a moving cube algorithm, and displaying the internal details of a reconstructed object to a user by using an isosurface structure.
Preferably, in the step (2), the deep neural network comprises an encoder, a decoder, a spatial convolutional coding module, a residual jump linking module and a scale attention module; the preprocessed neural image firstly enters an encoder, the encoder comprises four spatial domain convolutional coding modules with different sizes, an input image obtains feature maps with different scales, then a last layer of feature map enters a decoder, the feature maps with different scales in the encoder are respectively fused with the feature maps with corresponding scales in the decoder, finally the decoder with different scales obtains the feature map with the same dimension as the input image through bilinear interpolation processing, and the feature maps after the cascade connection are processed through a scale attention module and a residual jump link module to obtain a final segmentation image.
Preferably, in step (2), each multi-aperture encoder module first obtains a feature map through a convolution layer with kernel size of 3 × 3, the feature map uses different convolution rates and convolution kernels to calculate the receptive field, which is defined as follows:
wherein r isnRepresenting the receptive field of the current layer, rn-1Upper layer, s, representing the receptive fieldiDenotes the step size of the i-th convolution, k denotes the size of the convolution kernel;
the spatial convolution function F (x) has 4 cascade branches with different convolution rates and convolution kernels, wherein the convolution rates are 1 to 1, 3 and 5, and the convolution kernels comprise 3 x 3 and 1 x 1;
[r3:r7:r9:r19]=F(x)
wherein [: represents cascade;
then, the receptive fields for each branch would be set to 3,7,9,19 and a cascade profile r is obtained3:r7:r9:r19]The cascade signature applies a 1 × 1 convolution and a linear activation function; finally, the input feature map is added with the previous feature map to obtain a new resolution level feature map, and the final output is obtained through a convolution layer with the kernel size of 3 × 3 and a pooling layer with the kernel size of 2 × 2. By introducing a plurality of convolution layers with different convolution rates into four different resolution levels, the receptive field of the feature map can be well expanded, the parameters of the feature map at each resolution level are greatly reduced, and more semantic information is learned in the lumbosacral plexus MRI image.
Preferably, in the step (2), in the residual jump linking module, the features with different dimensions are mapped to a convolution layer with 3 × 3 kernels, a batch normalization layer and a ReLU as an active layer; meanwhile, feature mapping of different dimensions also becomes a convolution layer with 3 cores and a layer; adding the characteristic diagram H (X) and the original characteristic diagram X in a short circuit mode to form a new characteristic diagram, and putting the new characteristic diagram into a ReLU activation function to obtain a new characteristic diagram beta:
β=ReLU(H(X)+E(X))
wherein ReLU () represents an activation function;
in addition, a 1 × 1 convolutional layer is added in the short circuit, which can provide some additional spatial features e (x).
Preferably, in the step (2), in the scale attention module, the bilinear interpolation method is used to resample the feature maps of different scales obtained by the decoder to the size of the predicted image, and the three feature maps of different scales are concatenated as inputAfter cascade connectionRepresenting a feature map with an input size of C H W, where C represents the input channel, H and W represent the height and width, respectively, of the feature map, and the inputRespectively passing through a global average pooling GAP layer and a global maximum pooling GMP layer, in order to simplify the characteristic diagram parameters and obtain the weight information of each channel, and outputting PGAP(X)∈R1×1×CAnd PGMP(X)∈R1 ×1×CMultilayer perceptual MLP is implemented with two fully connected layers to obtain a scale attention coefficient and X1And X2Are shared between them. The scale attention coefficient alpha is belonged to [0,1 ]]Comprises the following steps:
α=Sigmoid(X1+X2)
in the scale attention module, spatial attention is additionally usedObtaining space information by a block, wherein the space information comprises a 3 multiplied by 3 convolution layer and a 1 multiplied by 1 convolution layer, then applying a Dropout layer to enhance generalization capability, multiplying an output characteristic diagram by alpha to obtain a new characteristic diagram through a Sigmoid function, adding the characteristic diagram and an original characteristic diagram in a space attention module to obtain an output characteristic diagram gamma, and finally adding the characteristic diagram gamma and an original input characteristic diagram to obtain the output characteristic diagram of the multi-scale attention blockComprises the following steps:
finally, the feature map subjected to scale attention is subjected to a one-dimensional convolution with the kernel size of 1 × 1 and a Sigmoid function to obtain an output segmentation image which has the same shape as the input image.
Preferably, in step (3), stacking the segmented two-dimensional plane images, giving an original voxel space, performing three-dimensional reconstruction by using a VTK package and a marching cube algorithm, and displaying internal details of the reconstructed object to a user by using an isosurface structure specifically includes: firstly, reading all the iconography sequence data of the patient in a layered reading mode, scanning two layers of data, constructing voxels one by one, wherein 8 corner points in the voxels are taken from four pixel points which correspond to the upper part and the lower part of two slices, the two slices form a cube, and the isosurface is defined as follows:
Fi,j,k=Fi,j,k(xi,yj,zk)
{(x,y,z)|F(x,y,z)=N}
where F is the value of each pixel, xi,yj,zkIf the function value F of the vertex is more than or equal to N, the vertex is positioned in the equivalent surface and is marked as plus; if the function value F of the vertex<N, then the vertex is outside the iso-surface, marked-and then the corner function value of the voxel is compared with the junction after the iso-surfaceIf the state table of the voxel is constructed, the boundary voxel with the intersection point with the isosurface can be obtained from the obtained state table, and then the position coordinates (x, y, z) of the intersection point can be obtained by calculating the isosurface and the cube through a linear interpolation method:
to display the iso-surface image using graphics hardware, the normal components of the triangular patches of the iso-surface are generated, however for each point on the iso-surface, the gradient component of the tangent in the direction of the triangular surface is zero, then the gradient vector for that point represents the normal vector of the iso-surface at that point, assuming the voxel vertex is (i, j, k) and thus the gradient values are represented as:
wherein G isx、Gy、GzRespectively representing gradient values of voxel vertexes in x, y and z directions;
then the obtained G isx、Gy、GzCarrying out normalization treatment:
and finally, drawing an isosurface image according to each vertex coordinate and the normal vector on the triangular surface patch, and rendering by using a light renderer to obtain a final surface model so as to realize the three-dimensional drawing of the final image.
The beneficial effects of the invention are as follows: the segmentation algorithm can more quickly and effectively segment the lumbosacral plexus nerve from the spine MRI medical image, and the accuracy of the segmented tissue is higher, so that the accuracy of the three-dimensional structure is greatly increased; the lumbosacral plexus nerve segmentation and three-dimensional visualization algorithm can enable a doctor to have a sufficiently clear understanding of the relative position of the lumbosacral plexus before an operation, and especially can greatly improve the operation efficiency of the doctor and reduce the risk of operation failure for the advance planning of some rare cases; the invention not only can intuitively display the relative position of the three-dimensional anatomical structure of the nerve in the spine, but also can assist in clinical disease diagnosis.
Drawings
FIG. 1 is a general block diagram of a method implementation of the present invention.
FIG. 2 is a schematic structural diagram of a spatial convolutional coding module according to the present invention.
Fig. 3 is a schematic diagram of a residual jump connection structure according to the present invention.
Fig. 4 is a structural diagram of a scale attention module according to the present invention.
Fig. 5 is a schematic diagram of a three-dimensional reconstruction result according to the present invention.
Detailed Description
A lumbosacral plexus nerve segmentation and three-dimensional visualization method comprises the following steps:
(1) processing data; acquiring MRI (magnetic resonance imaging) image data of a patient, storing the MRI image data in a DICOM (digital imaging and communications in medicine) format, converting the acquired data into a gray image, storing the gray image in a png format, and performing normalization processing on the image;
(2) adopting a deep neural network to perform segmentation processing on the normalized image obtained in the step (1);
(3) three-dimensional reconstruction; stacking the divided two-dimensional plane images, giving an original voxel space, performing three-dimensional reconstruction by using a VTK package and adopting a moving cube algorithm, and displaying the internal details of a reconstructed object to a user by using an isosurface structure.
In the data processing part, firstly, MRI image data of a patient is collected and stored in a DICOM format, the obtained data is converted into a gray image and stored in a png format, and the image is normalized for use in the subsequent process. In the segmentation part, a deep neural network popular in the field of artificial intelligence is adopted in the segmentation method, and in the neural network, a U-net classical biomedical image segmentation network is used as a basic frame and is combined with a spatial domain convolution coding module, a residual jump link module and a scale attention module to complete lumbosacral plexus nerve segmentation. The U-shaped structure segmentation network is a coding-decoding network, can perform feature extraction on an original image and restore the original image to the size. The spatial convolution is essential to the wide-range semantic segmentation in the image processing field, and the effective improvement is achieved in the aspect of feature extraction. So in this algorithm we introduce spatial convolution into the encoder. However, the different feature levels may lose a large amount of segmentation information due to the continuous pooling between lumbosacral nerve plexus segmentations. Furthermore, for a lumbosacral plexus MR image, nerves account for only a small portion of the image. To overcome this limitation, we use hole convolution at different convolution rates in the feature encoder to extend the receptive field of the feature map at four different resolution levels, as shown in FIG. 1. Each multi-aperture encoder module first obtains a feature map X over a convolutional layer of kernel size 3 × 3. The feature maps use different convolution rates and convolution kernels to calculate the receptive field. In particular, the receptive fields are defined as follows:
wherein,rnrepresenting the receptive field of the current layer, rn-1Upper layer, s, representing the receptive fieldiDenotes the step size of the i-th convolution and k denotes the size of the convolution kernel.
The spatial convolution function f (x) has 4 cascaded branches of different convolution rates and convolution kernels, including convolution rates of 1 to 1, 3, 5, and convolution kernels including both 3 x 3 and 1 x 1.
[r3:r7:r9:r19]=F(x)
Wherein [: represents cascade.
Then, the receptive field for each branch will be set to 3,7,9,19, and a cascade profile r is obtained3:r7:r9:r19]The cascade signature applies a 1 × 1 convolution and a linear activation function. Finally, the input feature map is added to the previous feature map to obtain a new resolution level feature map, and the final output is obtained through a convolution layer with a kernel size of 3 × 3 and a pooling layer with a kernel size of 2 × 2. By introducing a plurality of convolution layers with different convolution rates into four different resolution levels, the receptive field of the feature map can be well expanded, the parameters of the feature map at each resolution level are greatly reduced, and more semantic information is learned in the lumbosacral plexus MRI image.
The residual error network can solve the problem of network degradation to a certain extent and relieve the problem of gradient dispersion to a certain extent, and jump layer connection used in the residual error network achieves the purposes of increasing feature diversity and accelerating training. Residual skipping concatenations can introduce feature information at the corresponding scale into the upsampling or deconvolution to improve the accuracy of the segmentation. Thus, in our approach, we introduce a residual skip concatenation operation between the encoder and decoder to integrate the spatial features lost to pooling in the encoder and the high-level feature maps in the decoder. The residual skips the connection, besides, the detailed information can be recovered from the original image, and multi-scale and multi-level information can be provided for image segmentation. The architecture of the residual jump connection is shown in fig. 2. Features of different dimensions are mapped to a convolutional layer with 3 x 3 kernels, a bulk normalization layer, and a ReLU as the active layer. Meanwhile, feature mapping of different dimensions also becomes a convolution layer with 3 cores and a layer. Adding the characteristic diagram H (X) and the original characteristic diagram X in a short circuit mode to form a new characteristic diagram, and putting the new characteristic diagram into a ReLU activation function to obtain a new characteristic diagram beta:
β=ReLU(H(X)+E(X))
in addition, they have added 1 × 1 convolutional layers in the shorts, and e (x) may provide some additional spatial features.
In order to achieve the best segmentation effect, a scale attention model is usually introduced in the computer vision model. In our approach, we introduce a scale attention module to better process the lumbosacral plexus to obtain segmentation results at different scales. The multi-scale attention block we propose is shown in fig. 3. We first resample the different scales of feature maps obtained by the decoder to the predicted image size using bilinear interpolation. We first concatenate three feature maps of different scales as inputAfter cascade connectionRepresents a feature map with an input size of C × H × W, where C represents the input channel and H and W represent the height and width of the feature map, respectively. Input deviceRespectively through a Global Average Pooling (GAP) layer and a Global Maximum Pooling (GMP) layer. The purpose is to simplify the characteristic diagram parameters and obtain the weight information of each channel, and the output is respectively expressed as PGAP(X)∈R1 ×1×*And PGMP(X)∈R1×1×C. Multi-layer perception (MLP) is implemented by two fully-connected layers to obtain the scale attention coefficient and at X1And X2Are shared between them. The scale attention coefficient alpha is belonged to [0,1 ]]Comprises the following steps:
α=Sigmoid(X1+X2)
in the scale attention Block, IThey additionally use the spatial attention block to obtain spatial information. It consists of one 3 x 3 convolutional layer and one 1 x 1 convolutional layer, and then a Dropout layer is applied to enhance the generalization capability. And multiplying the output feature map by alpha to obtain a new feature map through a Sigmoid function. The feature map and the original feature map in the spatial attention module are added to obtain an output feature map gamma, and finally the feature map gamma is added with the original input feature map to finally obtain the output of the multi-scale attention blockComprises the following steps:
finally, the feature map subjected to scale attention is subjected to a one-dimensional convolution with the kernel size of 1 × 1 and a Sigmoid function to obtain an output segmentation image which has the same shape as the input image.
The three-dimensional technology can give visual impact to people, people can feel objective objects more visually, segmented two-dimensional plane images are stacked in the system, an original voxel space is given, three-dimensional reconstruction is carried out by using a VTK package and a moving cube algorithm, the method utilizes an isosurface structure, internal details of the reconstructed object can be displayed like a user, and therefore doctors can observe the nerve condition of patients more visually. The basic idea of the algorithm is to process each voxel in a data field, classify the body elements intersected with the isosurface, calculate the intersection point and normal vector with the isosurface by using a linear difference method, and finally draw the isosurface by using related graphic software. Firstly, reading all the iconography sequence data of the patient in a layered reading mode, scanning two layers of data, and constructing voxels one by one, wherein 8 corner points in the voxels are taken from four pixel points which correspond to the upper part and the lower part of two slices. The two slices form a cube, and the iso-surface is defined as follows:
Fi,j,k=Fi,j,k(xi,yj,zk)
{(x,y,z)|F(x,y,z)=N}
where F is the value of each pixel, xi,yj,zkIf the function value F of the vertex is more than or equal to N, the vertex is positioned in the equivalent surface and is marked as plus; if the function value F of the vertex<N, then the vertex is outside the iso-surface, labeled-and then the result of comparing the corner function value of the voxel with the iso-surface constructs the state table for that voxel. The boundary voxel with the intersection point with the isosurface can be obtained from the obtained state table, and then the intersection point position coordinate (x, y, z) can be obtained by calculating the isosurface and the cube through a linear interpolation method:
in order to display our iso-surface image using graphics hardware, we must generate the normal components of the triangular patches of the iso-surface, however, for each point on the iso-surface, the gradient component of the tangent in the direction of the triangular surface is zero, then the gradient vector for that point represents the normal vector of the iso-surface at that point, assuming the voxel vertex is (i, j, k) and thus the gradient values are represented as:
wherein G isx、Gy、GzRespectively representing gradient values of voxel vertexes in x, y and z directions;
then the obtained Gx、Gy、GzCarrying out normalization treatment:
the normalized result is taken as the unit normal vector of the voxel vertex (i, j, k). And then normal vectors on all vertexes of the triangular patch can be calculated according to the linear difference function. And finally, drawing an isosurface image according to each vertex coordinate and normal vector on the triangular surface patch, and rendering by using a light renderer to obtain a final surface model to realize three-dimensional drawing of the final image, as shown in fig. 5. Thereby facilitating the preoperative judgment of doctors and leading patients to better know the focus parts of themselves.
Claims (6)
1. A lumbosacral plexus nerve segmentation and three-dimensional visualization method is characterized by comprising the following steps:
(1) processing data; acquiring MRI (magnetic resonance imaging) image data of a patient, storing the MRI image data in a DICOM (digital imaging and communications in medicine) format, converting the acquired data into a gray image, storing the gray image in a png format, and performing normalization processing on the image;
(2) adopting a deep neural network to perform segmentation processing on the normalized image obtained in the step (1);
(3) three-dimensional reconstruction; stacking the divided two-dimensional plane images, giving an original voxel space, performing three-dimensional reconstruction by using a VTK package and adopting a moving cube algorithm, and displaying the internal details of a reconstructed object to a user by using an isosurface structure.
2. The lumbosacral plexus nerve segmentation and three-dimensional visualization method of claim 1, wherein in the step (2), the deep neural network comprises an encoder, a decoder, a spatial convolutional coding module, a residual jump linking module, and a scale attention module; the preprocessed neural image firstly enters an encoder, the encoder comprises four spatial domain convolutional coding modules with different sizes, an input image obtains feature maps with different scales, then a last layer of feature map enters a decoder, the feature maps with different scales in the encoder are respectively fused with the feature maps with corresponding scales in the decoder, finally the decoder with different scales obtains the feature map with the same dimension as the input image through bilinear interpolation processing, and the feature maps after the cascade connection are processed through a scale attention module and a residual jump link module to obtain a final segmentation image.
3. The lumbosacral plexus nerve segmentation and three-dimensional visualization method as claimed in claim 2, wherein in step (2), each multi-aperture encoder module first obtains a feature map X through a convolution layer with kernel size of 3X 3, the feature map using different convolution rates and convolution kernels to calculate the receptive field, the receptive field being defined as follows:
wherein r isnShowing the receptive field of the current layer, rn-1Upper layer, s, representing the receptive fieldiDenotes the step size of the i-th convolution, k denotes the size of the convolution kernel;
the spatial convolution function F (x) has 4 cascade branches with different convolution rates and convolution kernels, wherein the convolution rates are 1 to 1, 3 and 5, and the convolution kernels comprise 3 x 3 and 1 x 1;
[r3:r7:r9:r19]=F(x)
wherein [: represents cascade;
then, the receptive fields for each branch would be set to 3,7,9,19 and a cascade profile r is obtained3:r7:r9:r19]The cascade signature applies a 1 × 1 convolution and a linear activation function; finally, the input feature map is added to the previous feature map,obtaining a new resolution level characteristic diagram, and obtaining the final output through a convolution layer with the kernel size of 3 multiplied by 3 and a pooling layer with the kernel size of 2 multiplied by 2.
4. The lumbosacral plexus nerve segmentation and three-dimensional visualization method as claimed in claim 2, wherein in step (2), in the residual jump linking module, features of different dimensions are mapped to a convolution layer having 3 x 3 kernels, a batch normalization layer and a ReLU as an activation layer; meanwhile, feature mapping of different dimensions also becomes a convolution layer with 3 cores and a layer; adding the characteristic diagram H (X) and the original characteristic diagram X in a short circuit mode to form a new characteristic diagram, and putting the new characteristic diagram into a ReLU activation function to obtain a new characteristic diagram beta:
β=ReLU(H(X)+E(X))
wherein ReLU () represents an activation function;
in addition, a 1 × 1 convolutional layer is added in the short circuit, which can provide some additional spatial features e (x).
5. The method for lumbosacral plexus segmentation and three-dimensional visualization as claimed in claim 2, wherein in the step (2), the scale attention module resamples the feature maps of different scales obtained from the decoder to the size of the predicted image by using bilinear interpolation, and concatenates the three feature maps of different scales as inputAfter cascade connectionRepresenting a feature map with an input size of C H W, where C represents the input channel, H and W represent the height and width, respectively, of the feature map, and the inputRespectively pass through a global average pooling GAP layer and a global maximum pooling GMP layer in order to simplify the feature map parameters and obtainObtaining the weight information of each channel, and outputting the weight information which is respectively expressed as PGAP(X)∈R1×1×CAnd PGMP(X)∈R1×1×CMultilayer perceptual MLP is implemented with two fully connected layers to obtain a scale attention coefficient and X1And X2Share between them, the scale attention coefficient alpha is in [0,1 ]]Comprises the following steps:
α=Sigmoid(X1+X2)
in the scale attention module, a space attention block is additionally used for acquiring space information, the space information consists of a 3 x 3 convolution layer and a 1 x 1 convolution layer, then a Dropout layer is applied to enhance the generalization capability, an output feature map is multiplied by alpha to obtain a new feature map through a Sigmoid function, the feature map is added with an original feature map in the space attention module to obtain an output feature map gamma, and finally the feature map gamma is added with an original input feature map to obtain the output feature map of the multi-scale attention blockComprises the following steps:
finally, the feature map subjected to scale attention is subjected to a one-dimensional convolution with the kernel size of 1 × 1 and a Sigmoid function to obtain an output segmentation image which has the same shape as the input image.
6. The lumbosacral plexus nerve segmentation and three-dimensional visualization method as claimed in claim 1, wherein in step (3), the segmented two-dimensional plane images are stacked, an original voxel space is given, three-dimensional reconstruction is performed by using a moving cube algorithm using a VTK package, and internal details of a reconstructed object are displayed to a user by using an isosurface structure, specifically: firstly, reading all the iconography sequence data of the patient in a layered reading mode, scanning two layers of data, constructing voxels one by one, wherein 8 corner points in the voxels are taken from four pixel points which correspond to the upper part and the lower part of two slices, the two slices form a cube, and the isosurface is defined as follows:
Fi,j,k=Fi,j,k(xi,yj,zk)
{(x,y,z)|F(x,y,z)=N}
where F is the value of each pixel, xi,yj,zkIf the function value F of the vertex is more than or equal to N, the vertex is positioned in the equivalent surface and is marked as plus; if the function value F of the vertex<N, the vertex is positioned outside the isosurface and is marked as-, then a state table of the voxel is constructed by comparing the corner point function value of the voxel with the isosurface, the boundary voxel with an intersection point with the isosurface can be obtained by the obtained state table, and then the position coordinate (x, y, z) of the intersection point is obtained by calculating the isosurface and the cube through a linear interpolation method:
to display the iso-surface image using graphics hardware, the normal components of the triangular patches of the iso-surface are generated, however for each point on the iso-surface, the gradient component of the tangent in the direction of the triangular surface is zero, then the gradient vector for that point represents the normal vector of the iso-surface at that point, assuming the voxel vertex is (i, j, k) and thus the gradient values are represented as:
wherein G isx、Gy、GzRespectively representing gradient values of voxel vertexes in x, y and z directions;
then the obtained Gx、Gy、GzCarrying out normalization treatment:
and finally, drawing an isosurface image according to each vertex coordinate and the normal vector on the triangular surface patch, and rendering by using a light renderer to obtain a final surface model so as to realize the three-dimensional drawing of the final image.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202210016806.4A CN114519770A (en) | 2022-01-07 | 2022-01-07 | Lumbosacral plexus nerve segmentation and three-dimensional visualization method |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202210016806.4A CN114519770A (en) | 2022-01-07 | 2022-01-07 | Lumbosacral plexus nerve segmentation and three-dimensional visualization method |
Publications (1)
Publication Number | Publication Date |
---|---|
CN114519770A true CN114519770A (en) | 2022-05-20 |
Family
ID=81595978
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN202210016806.4A Pending CN114519770A (en) | 2022-01-07 | 2022-01-07 | Lumbosacral plexus nerve segmentation and three-dimensional visualization method |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN114519770A (en) |
Cited By (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN117111013A (en) * | 2023-08-22 | 2023-11-24 | 南京慧尔视智能科技有限公司 | Radar target tracking track starting method, device, equipment and medium |
-
2022
- 2022-01-07 CN CN202210016806.4A patent/CN114519770A/en active Pending
Cited By (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN117111013A (en) * | 2023-08-22 | 2023-11-24 | 南京慧尔视智能科技有限公司 | Radar target tracking track starting method, device, equipment and medium |
CN117111013B (en) * | 2023-08-22 | 2024-04-30 | 南京慧尔视智能科技有限公司 | Radar target tracking track starting method, device, equipment and medium |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
Yi et al. | Generative adversarial network in medical imaging: A review | |
US10997725B2 (en) | Image processing method, image processing apparatus, and computer-program product | |
JP7114737B2 (en) | Image processing device, image processing method, and program | |
US11704796B2 (en) | Estimating bone mineral density from plain radiograph by assessing bone texture with deep learning | |
JP2022077991A (en) | Medical image processing apparatus, medical image processing method, medical image processing program, model training apparatus, and training method | |
US20230214664A1 (en) | Learning apparatus, method, and program, image generation apparatus, method, and program, trained model, virtual image, and recording medium | |
CN113361689A (en) | Training method of super-resolution reconstruction network model and scanning image processing method | |
CN116797519A (en) | Brain glioma segmentation and three-dimensional visualization model training method and system | |
CN114519770A (en) | Lumbosacral plexus nerve segmentation and three-dimensional visualization method | |
Wang et al. | TRCT-GAN: CT reconstruction from biplane X-rays using transformer and generative adversarial networks | |
Roberts et al. | Imaging evaluation of a proposed 3D generative model for MRI to CT translation in the lumbar spine | |
Zhao et al. | Residual-atrous attention network for lumbosacral plexus segmentation with MR image | |
Bukas et al. | Patient-specific virtual spine straightening and vertebra inpainting: an automatic framework for osteoplasty planning | |
CN117078658A (en) | Method for detecting vertebral center point in CT image based on deep learning | |
CN116758087A (en) | Lumbar vertebra CT bone window side recess gap detection method and device | |
Zhang et al. | A spine segmentation method under an arbitrary field of view based on 3d swin transformer | |
Antico et al. | Deep Learning-based automatic segmentation for reconstructing vertebral anatomy of Healthy Adolescents and Patients with Adolescent Idiopathic Scoliosis (AIS) Using MRI Data | |
CN110570416A (en) | Method for visualization and 3D printing of multi-modal cardiac images | |
CN113299369B (en) | Medical image window adjusting optimization method | |
CN113902912B (en) | CBCT image processing method, neural network system creation method and device | |
Xu et al. | An intelligent system for craniomaxillofacial defecting reconstruction | |
Ma et al. | A frequency domain constraint for synthetic and real x-ray image super resolution | |
Subsol et al. | 3d image processing for the study of the evolution of the shape of the human skull: presentation of the tools and preliminary results | |
Hoffmeister et al. | Three-dimensional surface reconstructions using a general purpose image processing system | |
Pandey et al. | A Framework for Mathematical Methods in Medical Image Processing |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination |