CN113469961A - Neural network-based carpal tunnel image segmentation method and system - Google Patents

Neural network-based carpal tunnel image segmentation method and system Download PDF

Info

Publication number
CN113469961A
CN113469961A CN202110704545.0A CN202110704545A CN113469961A CN 113469961 A CN113469961 A CN 113469961A CN 202110704545 A CN202110704545 A CN 202110704545A CN 113469961 A CN113469961 A CN 113469961A
Authority
CN
China
Prior art keywords
layer
carpal tunnel
convolution
block
segmentation
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN202110704545.0A
Other languages
Chinese (zh)
Inventor
卢荟
周海英
蒋帅
胡贤良
白琪
金倩君
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
First Affiliated Hospital of Zhejiang University School of Medicine
Original Assignee
First Affiliated Hospital of Zhejiang University School of Medicine
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by First Affiliated Hospital of Zhejiang University School of Medicine filed Critical First Affiliated Hospital of Zhejiang University School of Medicine
Priority to CN202110704545.0A priority Critical patent/CN113469961A/en
Publication of CN113469961A publication Critical patent/CN113469961A/en
Pending legal-status Critical Current

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/0002Inspection of images, e.g. flaw detection
    • G06T7/0012Biomedical image inspection
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/24Classification techniques
    • G06F18/241Classification techniques relating to the classification model, e.g. parametric or non-parametric approaches
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/045Combinations of networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/048Activation functions
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/08Learning methods
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/10Segmentation; Edge detection
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/10Image acquisition modality
    • G06T2207/10072Tomographic images
    • G06T2207/10088Magnetic resonance imaging [MRI]
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/20Special algorithmic details
    • G06T2207/20081Training; Learning
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/20Special algorithmic details
    • G06T2207/20084Artificial neural networks [ANN]
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/30Subject of image; Context of image processing
    • G06T2207/30004Biomedical image processing
    • G06T2207/30101Blood vessel; Artery; Vein; Vascular

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Data Mining & Analysis (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Artificial Intelligence (AREA)
  • General Health & Medical Sciences (AREA)
  • Health & Medical Sciences (AREA)
  • Evolutionary Computation (AREA)
  • General Engineering & Computer Science (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Molecular Biology (AREA)
  • Software Systems (AREA)
  • Biomedical Technology (AREA)
  • Biophysics (AREA)
  • Computational Linguistics (AREA)
  • Mathematical Physics (AREA)
  • Computing Systems (AREA)
  • Evolutionary Biology (AREA)
  • Bioinformatics & Cheminformatics (AREA)
  • Bioinformatics & Computational Biology (AREA)
  • Medical Informatics (AREA)
  • Nuclear Medicine, Radiotherapy & Molecular Imaging (AREA)
  • Radiology & Medical Imaging (AREA)
  • Quality & Reliability (AREA)
  • Image Analysis (AREA)

Abstract

The invention provides a neural network-based carpal tunnel image segmentation method, which comprises the following steps of: constructing and training a processing model by using a convolutional neural network, wherein the processing model at least comprises a classification module and a segmentation module; acquiring a carpal tunnel image to be segmented; preprocessing a carpal tunnel image to be segmented; using a classification module to perform morphological classification on the preprocessed image; and according to the form classification result, using an adaptive segmentation module to segment the carpal tunnel image to be segmented and outputting a segmentation result. The invention classifies and segments the images of the carpal tunnel to be segmented, thereby improving the accuracy of the segmentation of the carpal tunnel images by a processing model.

Description

Neural network-based carpal tunnel image segmentation method and system
Technical Field
The invention relates to a carpal tunnel image segmentation method and system, in particular to a carpal tunnel image segmentation method and system based on a neural network, and belongs to the technical field of medical image processing.
Background
Carpal tunnel syndrome, caused by the compression of the median nerve as it passes through the carpal tunnel, is manifested by pain, numbness, muscle weakness in the innervated area, is the most common clinically compressive focal mononeuropathy, and the incidence increases year by year. Early non-surgical treatment can bring certain curative effect, and more patients can obtain remarkable improvement of symptoms from surgical treatment. However, due to the non-specificity of the symptoms, misdiagnosis and missed diagnosis occur frequently, and the condition is continuously progressed due to the delay of diagnosis, and finally irreversible nerve injury and muscle nutritional loss atrophy are developed, so that defects on the appearance and the function of hands of patients are brought.
Clinical assessment is currently considered the gold standard for clinical and research, but it depends on the clinical experience and expertise of the physician, and therefore more objective imaging results are emerging. The Magnetic Resonance Imaging (MRI) technology can accurately measure the diameter of the carpal tunnel and the nerve, so as to distinguish whether the median nerve has swelling and thickening when entering the carpal tunnel and whether the median nerve enters the carpal tunnel and is pressed to be flat, and meanwhile, the ischemia and fibrosis of the nerve can show different degrees of signal changes in the MRI. However, it is time-consuming and labor-consuming to segment the carpal tunnel image manually, and different doctors may obtain different results when performing image segmentation, and even if the same doctor segments the same hand MRI image at different time periods, there are slight differences.
In recent years, a neural network has been developed rapidly in an image processing direction, and a large number of segmentation models have been proposed, such as network structures of FCN, UNet, SegNet, deep lab, and the like. Medical image processing is also increasingly applied to neural networks. Therefore, research and development of a neural network-based carpal tunnel image segmentation method and system are of great significance to the field.
Disclosure of Invention
The invention aims to provide a neural network-based carpal tunnel image segmentation method, which can obtain objective and stable image segmentation results, assist doctors in making better judgment on disease conditions and solve the problems in the background art.
The invention further aims to provide a carpal tunnel image segmentation system based on the neural network.
The technical scheme adopted by the invention for solving the technical problems is as follows:
a carpal tunnel image segmentation method based on a neural network comprises the following steps:
constructing and training a processing model by using a convolutional neural network, wherein the processing model at least comprises a classification module and a segmentation module;
acquiring a carpal tunnel image to be segmented;
preprocessing a carpal tunnel image to be segmented;
using a classification module to perform morphological classification on the preprocessed image;
and according to the form classification result, using an adaptive segmentation module to segment the carpal tunnel image to be segmented and outputting a segmentation result.
Preferably, the processing model comprises a plurality of segmentation modules, and is used for segmenting the carpal tunnel images to be segmented in different forms according to the form classification result and outputting the segmentation result.
Preferably, the classification module and the segmentation module each include at least a convolution block, and the convolution block includes a first convolution layer, a first batch normalization layer, a first active layer, a second convolution layer, a second batch normalization layer, and a second active layer, which are sequentially connected and encapsulated.
Preferably, the parameters of the volume block are set as:
a first winding layer: inputting a channel number in _ ch and an output channel number out _ ch, wherein the size of the adopted convolution kernel is 3, padding is 1 and stride is 1;
a first batch normalization layer and a second batch normalization layer: the number of channels is out _ ch;
first and second active layers: the activation function is a ReLU function, and the infill is True;
a second convolution layer: the number of input channels out _ ch and the number of output channels out _ ch are 3, padding is 1, and stride is 1.
Preferably, the classification module includes a first volume block, a first pooling layer, a second volume block, a second pooling layer, a first full-connection layer and a second full-connection layer, which are sequentially connected and encapsulated.
Preferably, the parameters of the classification module are set as follows:
a first rolling block: the number of input channels is 1, and the number of output channels is 16;
first and second pooling layers: adopting maximum pooling, wherein the size of a pooling nucleus is 8;
a first rolling block: the number of input channels is 16, and the number of output channels is 32;
first fully-connected layer: input feature number 960, output feature number 64;
second full connection layer: the feature number 64 is input, and the feature number 1 is output.
Preferably, the dividing module includes a first convolution block, a first pooling layer, a second convolution block, a second pooling layer, a third convolution block, a third pooling layer, a fourth convolution block, a fourth pooling layer, a fifth convolution block, a first transposing convolution layer, a sixth convolution block, a second transposing convolution layer, a seventh convolution block, a third transposing convolution layer, an eighth convolution block, a fourth transposing convolution layer, a ninth convolution block, and a first convolution layer, which are connected and encapsulated in sequence.
Preferably, the parameters of the segmentation module are set as follows:
a first rolling block: the number of input channels is 1, and the number of output channels is 32;
a first, second, third, and fourth pooling layer: adopting maximum pooling, wherein the size of a pooling nucleus is 2;
a second rolling block: input channel number 32, output channel number 64;
a third rolling block: input lane number 64, output lane number 128;
a fourth rolling block: input lane number 128, output lane number 256;
a fifth rolling block: input lane number 256, output lane number 512;
a first transfer buildup layer: the number of input channels is 512, the number of output channels is 256, the size of the adopted convolution kernel is 2, padding is 0, and stride is 2;
a sixth rolling block: splicing the output results of the fourth convolution block and the first transfer convolution layer as input, inputting the channel number 512 and outputting the channel number 256;
the second transpose convolution layer: 256 input channels and 128 output channels, wherein the size of the adopted convolution kernel is 2, padding is 0 and stride is 2;
a seventh rolling block: the output result of the third convolution block and the second transposition convolution layer is spliced as input, the number of input channels is 256, and the number of output channels is 128;
the third transpose convolution layer: the number of input channels is 128, the number of output channels is 64, the size of the adopted convolution kernel is 2, padding is 0, and stride is 2;
eighth volume block: the output result of the second convolution block and the third transposed convolution layer is spliced as input, the number of input channels is 128, and the number of output channels is 64;
the fourth transpose convolution layer: the number of input channels is 64, the number of output channels is 32, the size of the adopted convolution kernel is 2, padding is 0, and stride is 2;
ninth volume block: splicing the output results of the first convolution block and the fourth transposed convolution layer as input, and inputting 64 channels and 32 channels;
a first winding layer: the number of input channels 32, the number of output channels 1, the size of the adopted convolution kernel is 1, padding is 0, stride is 1, and the segmentation result is given at the pixel level.
Preferably, the training process model comprises the steps of:
acquiring a training set, wherein the training set comprises images and labels corresponding to the images;
inputting the image into a processing model to obtain a segmentation result, and calculating a loss value and a gradient corresponding to each module parameter of the processing model according to the label;
updating parameters of each module of the processing model according to the gradient;
and repeating the steps until the training iteration threshold of the processing model is reached.
Preferably, the training set of the classification module comprises a carpal tunnel image to be segmented and a corresponding first label, and the first label is a morphological category of the carpal tunnel image to be segmented; the training set of the segmentation module comprises various forms of carpal tunnel images to be segmented and corresponding second labels, and the second labels are carpal tunnel segmentation results marked by professional doctors; the loss value calculation of the classification module adopts a two-classification cross entropy loss function, and the loss value calculation of the segmentation module adopts a Dice loss function.
Preferably, the classification module and the segmentation module are trained by adopting an Adam optimizer, and the learning rate is set to be 0.001.
A neural network-based carpal tunnel image segmentation system, comprising:
the input module is used for acquiring a carpal tunnel image to be segmented;
the preprocessing module is used for preprocessing the carpal tunnel image to be segmented;
the classification module is used for carrying out morphological classification on the preprocessed image;
and the segmentation module is used for segmenting the carpal tunnel image to be segmented according to the form classification result and outputting the segmentation result.
The invention has the beneficial effects that:
(1) according to the method, the carpal tunnel images to be segmented are classified firstly and then segmented, so that the accuracy of the segmentation of the carpal tunnel images by the processing model is improved;
(2) the invention applies the image processing technology of deep learning to the medical image processing based on the processing model constructed and trained by the convolutional neural network, and realizes end-to-end carpal tunnel image segmentation by utilizing the high accuracy of the segmentation result, so that doctors without professional experience can finish the diagnosis of the carpal tunnel syndrome.
Drawings
In order to more clearly illustrate the embodiments of the present invention or the technical solutions in the prior art, the drawings used in the description of the embodiments or the prior art will be briefly described below, it is obvious that the drawings in the following description are only embodiments of the present invention, and for those skilled in the art, other drawings can be obtained according to the provided drawings without creative efforts.
FIG. 1 is a schematic flow diagram of the process of the present invention;
FIG. 2 is a schematic diagram of the system of the present invention;
FIG. 3 is a schematic view of a carpal tunnel image to be segmented in accordance with the present invention;
FIG. 4 is a schematic view of a pre-processed carpal tunnel image to be segmented in accordance with the present invention;
FIG. 5 is a schematic diagram of the operation of the classification module and segmentation module of the present invention;
FIG. 6 is a schematic diagram of the classification module of the present invention;
FIG. 7 is a schematic diagram of the partitioning module operation of the present invention.
In the figure: 1. the device comprises an input module, 2, a preprocessing module, 3, a classification module, 4 and a segmentation module.
Detailed Description
The technical solution of the present invention is further specifically described below by way of specific examples in conjunction with the accompanying drawings. It is to be understood that the practice of the invention is not limited to the following examples, and that any variations and/or modifications may be made thereto without departing from the scope of the invention.
In the present invention, all parts and percentages are by weight, unless otherwise specified, and the equipment and materials used are commercially available or commonly used in the art. The methods in the following examples are conventional in the art unless otherwise specified. The components or devices in the following examples are, unless otherwise specified, standard parts or parts known to those skilled in the art, the structure and principle of which are known to those skilled in the art through technical manuals or through routine experimentation.
Example 1:
fig. 1 shows a neural network-based carpal tunnel image segmentation method, which comprises the following steps:
constructing and training a processing model by using a convolutional neural network, wherein the processing model at least comprises a classification module and a segmentation module;
acquiring a carpal tunnel image to be segmented;
preprocessing a carpal tunnel image to be segmented;
using a classification module to perform morphological classification on the preprocessed image;
and according to the form classification result, using an adaptive segmentation module to segment the carpal tunnel image to be segmented and outputting a segmentation result.
With the method, a carpal tunnel image segmentation system based on a neural network is shown in fig. 2, and comprises:
the input module 1 is used for acquiring a carpal tunnel image to be segmented;
the preprocessing module 2 is used for preprocessing a carpal tunnel image to be segmented;
the classification module 3 is used for carrying out morphological classification on the preprocessed image;
and the segmentation module 4 is used for segmenting the carpal tunnel image to be segmented according to the form classification result and outputting the segmentation result.
By the technical scheme, the carpal tunnel image to be segmented is classified firstly and then segmented, so that the accuracy of segmenting the carpal tunnel image by the processing model is improved.
Example 2:
fig. 1 shows a neural network-based carpal tunnel image segmentation method, which comprises the following steps:
constructing and training a processing model by using a convolutional neural network, wherein the processing model at least comprises a classification module and a segmentation module;
acquiring a carpal tunnel image to be segmented;
preprocessing a carpal tunnel image to be segmented;
using a classification module to perform morphological classification on the preprocessed image;
and according to the form classification result, using an adaptive segmentation module to segment the carpal tunnel image to be segmented and outputting a segmentation result.
The processing model comprises a plurality of segmentation modules, and is used for segmenting the carpal tunnel images to be segmented in different forms according to form classification results and outputting segmentation results. The embodiment selects two segmentation modules: the first segmentation module and the second segmentation module are used for respectively processing and segmenting the carpal tunnel images to be segmented in two forms.
The obtained images of the carpal tunnel to be segmented are shown in fig. 3, and all the images are gray images.
The method for preprocessing the carpal tunnel image to be segmented comprises the following steps: changing the input image to a uniform size 512x 512; keeping the central point of the image unchanged, and cutting the image into a uniform size of 320x 384; adaptive histogram equalization is used on the image. In order to reduce the number of training rounds of the neural network, the example adopts a batch input method to input data into the neural network, so that the sizes of pictures need to be unified. In addition, in order to reduce the calculation amount and memory occupation of the neural network, the original image is subjected to center cropping in the present example, and only the portion of the center containing useful information is intercepted, so that the unified size is 320x 384. To enhance the contrast of the grey scale map, an adaptive histogram equalization technique is used in this example. The picture is modified to a uniform size of 512x512 using the pile library in a Python3 environment and center clipping is performed. And then performing adaptive histogram equalization by using a sketch library. The pre-processed image is shown in fig. 4.
The working of the classification module and the segmentation module is schematically shown in fig. 5. In fig. 5, MRI is an input carpal tunnel image to be segmented, classification is a classification module, Category1 is a first form, Category2 is a second form, Seg1 is a first segmentation module, Seg2 is a second segmentation module, Predict is a segmentation result, and Output is an Output segmentation result.
As shown in fig. 6, the sorting module includes a first stacking block, a first pooling layer, a second stacking block, a second pooling layer, a first full-connection layer, and a second full-connection layer, which are sequentially connected from front to back and encapsulated. In fig. 6, Conv Block is a volume Block, Pooling is a Pooling layer, Flatten is a Flatten layer (for flattening input and unifying multidimensional input), full connectivity is a full Connection layer, and Category is an output morphological classification result.
The parameters of the classification module are set as follows:
a first rolling block: the number of input channels is 1, and the number of output channels is 16;
first and second pooling layers: adopting maximum pooling, wherein the size of a pooling nucleus is 8;
a first rolling block: the number of input channels is 16, and the number of output channels is 32;
first fully-connected layer: input feature number 960, output feature number 64;
second full connection layer: the feature number 64 is input, and the feature number 1 is output.
After a preprocessed image is input into the classification module, the classification module processes the preprocessed image to obtain the type of the image form. The specific process is as follows:
inputting a grayscale image of 1x320x 384;
the output shape is 16x320x384 after the first volume block action;
the output shape is 16x40x48 after the first pooling layer;
the output shape is 32x40x48 after the action of the first volume block;
the output shape is 32x5x6 by the action of the first pooling layer;
after the data is converted into one dimension, the obtained shape is 960;
the output shape is 64 under the action of the first full connection layer;
and under the action of the second full-connection layer, the output shape is 2, namely the type of the input image.
As shown in fig. 7, each partitioning module includes a first convolution block, a first pooling layer, a second convolution block, a second pooling layer, a third convolution block, a third pooling layer, a fourth convolution block, a fourth pooling layer, a fifth convolution block, a first transposing convolution layer, a sixth convolution block, a second transposing convolution layer, a seventh convolution block, a third transposing convolution layer, an eighth convolution block, a fourth transposing convolution layer, a ninth convolution block, and a first convolution layer, which are sequentially connected from front to back and encapsulated. In FIG. 7, Conv Block is a convolutional Block, Max Pooling is a Pooling layer (Max Pooling), Transposed Conv is a Transposed convolutional layer, and Skip connection is Skip connection.
The parameters of the segmentation module are set as follows:
a first rolling block: the number of input channels is 1, and the number of output channels is 32;
a first, second, third, and fourth pooling layer: adopting maximum pooling, wherein the size of a pooling nucleus is 2;
a second rolling block: input channel number 32, output channel number 64;
a third rolling block: input lane number 64, output lane number 128;
a fourth rolling block: input lane number 128, output lane number 256;
a fifth rolling block: input lane number 256, output lane number 512;
a first transfer buildup layer: the number of input channels is 512, the number of output channels is 256, the size of the adopted convolution kernel is 2, padding is 0, and stride is 2;
a sixth rolling block: splicing the output results of the fourth convolution block and the first transfer convolution layer as input, inputting the channel number 512 and outputting the channel number 256;
the second transpose convolution layer: 256 input channels and 128 output channels, wherein the size of the adopted convolution kernel is 2, padding is 0 and stride is 2;
a seventh rolling block: the output result of the third convolution block and the second transposition convolution layer is spliced as input, the number of input channels is 256, and the number of output channels is 128;
the third transpose convolution layer: the number of input channels is 128, the number of output channels is 64, the size of the adopted convolution kernel is 2, padding is 0, and stride is 2;
eighth volume block: the output result of the second convolution block and the third transposed convolution layer is spliced as input, the number of input channels is 128, and the number of output channels is 64;
the fourth transpose convolution layer: the number of input channels is 64, the number of output channels is 32, the size of the adopted convolution kernel is 2, padding is 0, and stride is 2;
ninth volume block: splicing the output results of the first convolution block and the fourth transposed convolution layer as input, and inputting 64 channels and 32 channels;
a first winding layer: the number of input channels 32, the number of output channels 1, the size of the adopted convolution kernel is 1, padding is 0, stride is 1, and the segmentation result is given at the pixel level.
And calling the corresponding segmentation module to segment the carpal tunnel image to be segmented according to the form classification result of the classification module and outputting the segmentation result. The specific process is as follows:
inputting a grayscale image of 1x320x 384;
the output shape is 32x320x384 under the action of the first volume block, and the output shape is 32x160x192 under the action of the first pooling layer;
the output shape is 64x160x192 through the second convolution block, and the output shape is 64x80x96 through the second pooling layer;
the output shape is 128x80x96 through the third convolution block, and the output shape is 128x40x48 through the third pooling layer;
the output shape is 256x40x48 under the action of a fourth convolution block, and the output shape is 256x20x24 under the action of a fourth pooling layer;
the output shape is 512x20x24 through the function of a fifth convolution block, and the output shape is 256x40x48 through the function of a first transposition convolution layer;
splicing the output of the first transpose convolution layer and the output of the fourth convolution block, wherein the output shape is 256x40x48 through the action of a sixth convolution block, and the output shape is 128x80x96 through the action of the second transpose convolution layer;
splicing the output of the second transpose convolutional layer and the output of the third convolutional block, wherein the output shape is 128x80x96 through the action of a seventh convolutional block, and the output shape is 64x160x192 through the action of the third transpose convolutional layer;
the output of the third transposed convolutional layer and the output of the second convolutional block are spliced, the output shape is 64x160x192 through the action of an eighth convolutional block, and the output shape is 32x320x384 through the action of a fourth transposed convolutional layer;
and splicing the output of the fourth transposed convolution layer and the output of the first convolution block, wherein the output shape is 32x320x384 through the action of the ninth convolution block, and the output shape is 1x320x384 through the action of the first convolution layer, namely the hand carpal tunnel image segmentation result.
Wherein, the structure of all the above-mentioned volume blocks all includes: the first rolling layer, the first batch normalization layer, the first activation layer, the second rolling layer, the second batch normalization layer and the second activation layer are connected and packaged in sequence. The parameters of the volume block are set as:
a first winding layer: inputting a channel number in _ ch and an output channel number out _ ch, wherein the size of the adopted convolution kernel is 3, padding is 1 and stride is 1;
a first batch normalization layer and a second batch normalization layer: the number of channels is out _ ch;
first and second active layers: the activation function is a ReLU function, and the infill is True;
a second convolution layer: the number of input channels out _ ch and the number of output channels out _ ch are 3, padding is 1, and stride is 1.
Wherein, training the treatment model includes the following steps:
acquiring a training set, wherein the training set comprises images and labels corresponding to the images;
inputting the image into a processing model to obtain a segmentation result, and calculating a loss value and a gradient corresponding to each module parameter of the processing model according to the label;
updating parameters of each module of the processing model according to the gradient;
and repeating the steps until the training iteration threshold of the processing model is reached.
The training set of the classification module comprises a carpal tunnel image to be segmented and a corresponding first label, and the first label is the morphology category of the carpal tunnel image to be segmented. The loss value calculation of the classification module adopts a two-classification cross entropy loss function, and the expression is as follows:
Figure BDA0003131682540000101
where y is the true category of the image,
Figure BDA0003131682540000102
is the class predicted by the classification module.
The optimizer adopts an Adam optimizer, and the learning rate is set to be 0.001; the number of batch input data is 12, 30 epochs are trained, and after 20 epochs, the learning rate is reduced to one tenth of the original rate.
The training set of the first segmentation module and the training set of the second segmentation module comprise carpal tunnel images to be segmented in two forms and corresponding second labels, and the second labels are carpal tunnel segmentation results marked by professional doctors. The loss value calculation of the segmentation module adopts a Dice loss function, and the expression is as follows:
Figure BDA0003131682540000103
wherein X is a carpal tunnel segmentation label, Y is a neural network prediction result, | X | and | Y | represent the number of elements in X and Y, respectively, | X | N.Y | represents the number in the intersection of X and Y, and epsilon is a sufficiently small number for ensuring that the denominator is not 0.
The optimizer adopts an Adam optimizer, and the learning rate is set to be 0.001; the number of batch input data is 12, 120 epochs are trained, and after 90 epochs, the learning rate is reduced to one tenth of the original rate.
According to the technical scheme, the processing model constructed and trained on the basis of the convolutional neural network is used for applying the image processing technology of deep learning to medical image processing, and end-to-end carpal tunnel image segmentation is realized by utilizing the high accuracy of segmentation results, so that doctors without professional experience can finish diagnosis of carpal tunnel syndrome.
The above-described embodiments are only preferred embodiments of the present invention, and are not intended to limit the present invention in any way, and other variations and modifications may be made without departing from the spirit of the invention as set forth in the claims.

Claims (10)

1. A carpal tunnel image segmentation method based on a neural network is characterized in that: the method comprises the following steps:
constructing and training a processing model by using a convolutional neural network, wherein the processing model at least comprises a classification module and a segmentation module;
acquiring a carpal tunnel image to be segmented;
preprocessing a carpal tunnel image to be segmented;
using a classification module to perform morphological classification on the preprocessed image;
and according to the form classification result, using an adaptive segmentation module to segment the carpal tunnel image to be segmented and outputting a segmentation result.
2. The neural network-based carpal tunnel image segmentation method as set forth in claim 1, wherein: the processing model comprises a plurality of segmentation modules, and is used for segmenting the carpal tunnel images to be segmented in different forms according to form classification results and outputting segmentation results.
3. The neural network-based carpal tunnel image segmentation method as set forth in claim 1, wherein: the classification module and the segmentation module at least comprise a rolling block, and the rolling block comprises a first rolling layer, a first batch normalization layer, a first activation layer, a second rolling layer, a second batch normalization layer and a second activation layer which are sequentially connected and packaged.
4. The neural network-based carpal tunnel image segmentation method as set forth in claim 1, wherein: the classification module comprises a first volume block, a first pooling layer, a second volume block, a second pooling layer, a first full-connection layer and a second full-connection layer which are sequentially connected and packaged.
5. The neural network-based carpal tunnel image segmentation method as set forth in claim 4, wherein: the parameters of the classification module are set as follows:
a first rolling block: the number of input channels is 1, and the number of output channels is 16;
first and second pooling layers: adopting maximum pooling, wherein the size of a pooling nucleus is 8;
a first rolling block: the number of input channels is 16, and the number of output channels is 32;
first fully-connected layer: input feature number 960, output feature number 64;
second full connection layer: the feature number 64 is input, and the feature number 1 is output.
6. The neural network-based carpal tunnel image segmentation method as set forth in claim 1, wherein: the segmentation module comprises a first convolution block, a first pooling layer, a second convolution block, a second pooling layer, a third convolution block, a third pooling layer, a fourth convolution block, a fourth pooling layer, a fifth convolution block, a first transposition convolution layer, a sixth convolution block, a second transposition convolution layer, a seventh convolution block, a third transposition convolution layer, an eighth convolution block, a fourth transposition convolution layer, a ninth convolution block and a first convolution layer which are sequentially connected and packaged.
7. The neural network-based carpal tunnel image segmentation method as set forth in claim 6, wherein: the parameters of the segmentation module are set as follows:
a first rolling block: the number of input channels is 1, and the number of output channels is 32;
a first, second, third, and fourth pooling layer: adopting maximum pooling, wherein the size of a pooling nucleus is 2;
a second rolling block: input channel number 32, output channel number 64;
a third rolling block: input lane number 64, output lane number 128;
a fourth rolling block: input lane number 128, output lane number 256;
a fifth rolling block: input lane number 256, output lane number 512;
a first transfer buildup layer: the number of input channels is 512, the number of output channels is 256, the size of the adopted convolution kernel is 2, padding is 0, and stride is 2;
a sixth rolling block: splicing the output results of the fourth convolution block and the first transfer convolution layer as input, inputting the channel number 512 and outputting the channel number 256;
the second transpose convolution layer: 256 input channels and 128 output channels, wherein the size of the adopted convolution kernel is 2, padding is 0 and stride is 2;
a seventh rolling block: the output result of the third convolution block and the second transposition convolution layer is spliced as input, the number of input channels is 256, and the number of output channels is 128;
the third transpose convolution layer: the number of input channels is 128, the number of output channels is 64, the size of the adopted convolution kernel is 2, padding is 0, and stride is 2;
eighth volume block: the output result of the second convolution block and the third transposed convolution layer is spliced as input, the number of input channels is 128, and the number of output channels is 64;
the fourth transpose convolution layer: the number of input channels is 64, the number of output channels is 32, the size of the adopted convolution kernel is 2, padding is 0, and stride is 2;
ninth volume block: splicing the output results of the first convolution block and the fourth transposed convolution layer as input, and inputting 64 channels and 32 channels;
a first winding layer: the number of input channels 32, the number of output channels 1, the size of the adopted convolution kernel is 1, padding is 0, stride is 1, and the segmentation result is given at the pixel level.
8. The neural network-based carpal tunnel image segmentation method as set forth in claim 1, wherein: the training process model comprises the following steps:
acquiring a training set, wherein the training set comprises images and labels corresponding to the images;
inputting the image into a processing model to obtain a segmentation result, and calculating a loss value and a gradient corresponding to each module parameter of the processing model according to the label;
updating parameters of each module of the processing model according to the gradient;
and repeating the steps until the training iteration threshold of the processing model is reached.
9. The neural network-based carpal tunnel image segmentation method as set forth in claim 8, wherein: the training set of the classification module comprises a carpal tunnel image to be segmented and a corresponding first label, wherein the first label is the morphological category of the carpal tunnel image to be segmented; the training set of the segmentation module comprises various forms of carpal tunnel images to be segmented and corresponding second labels, and the second labels are carpal tunnel segmentation results marked by professional doctors; the loss value calculation of the classification module adopts a two-classification cross entropy loss function, and the loss value calculation of the segmentation module adopts a Dice loss function.
10. A carpal tunnel image segmentation system based on a neural network is characterized in that: the carpal tunnel image segmentation system based on the neural network comprises:
the input module is used for acquiring a carpal tunnel image to be segmented;
the preprocessing module is used for preprocessing the carpal tunnel image to be segmented;
the classification module is used for carrying out morphological classification on the preprocessed image;
and the segmentation module is used for segmenting the carpal tunnel image to be segmented according to the form classification result and outputting the segmentation result.
CN202110704545.0A 2021-06-24 2021-06-24 Neural network-based carpal tunnel image segmentation method and system Pending CN113469961A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202110704545.0A CN113469961A (en) 2021-06-24 2021-06-24 Neural network-based carpal tunnel image segmentation method and system

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202110704545.0A CN113469961A (en) 2021-06-24 2021-06-24 Neural network-based carpal tunnel image segmentation method and system

Publications (1)

Publication Number Publication Date
CN113469961A true CN113469961A (en) 2021-10-01

Family

ID=77872792

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202110704545.0A Pending CN113469961A (en) 2021-06-24 2021-06-24 Neural network-based carpal tunnel image segmentation method and system

Country Status (1)

Country Link
CN (1) CN113469961A (en)

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN114742917A (en) * 2022-04-25 2022-07-12 桂林电子科技大学 CT image segmentation method based on convolutional neural network
CN117952962A (en) * 2024-03-25 2024-04-30 南京科进实业有限公司 Bone mineral density detection image processing method and system

Cited By (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN114742917A (en) * 2022-04-25 2022-07-12 桂林电子科技大学 CT image segmentation method based on convolutional neural network
CN114742917B (en) * 2022-04-25 2024-04-26 桂林电子科技大学 CT image segmentation method based on convolutional neural network
CN117952962A (en) * 2024-03-25 2024-04-30 南京科进实业有限公司 Bone mineral density detection image processing method and system

Similar Documents

Publication Publication Date Title
CN111369563B (en) Semantic segmentation method based on pyramid void convolutional network
CN108898175B (en) Computer-aided model construction method based on deep learning gastric cancer pathological section
CN108268870B (en) Multi-scale feature fusion ultrasonic image semantic segmentation method based on counterstudy
WO2022077917A1 (en) Instance segmentation model sample screening method and apparatus, computer device and medium
CN111161273B (en) Medical ultrasonic image segmentation method based on deep learning
CN111242288B (en) Multi-scale parallel deep neural network model construction method for lesion image segmentation
CN109272048B (en) Pattern recognition method based on deep convolutional neural network
CN111369565B (en) Digital pathological image segmentation and classification method based on graph convolution network
CN110826596A (en) Semantic segmentation method based on multi-scale deformable convolution
CN110223304B (en) Image segmentation method and device based on multipath aggregation and computer-readable storage medium
CN112488234A (en) End-to-end histopathology image classification method based on attention pooling
CN114494195B (en) Small sample attention mechanism parallel twin method for fundus image classification
CN112150476A (en) Coronary artery sequence vessel segmentation method based on space-time discriminant feature learning
CN113469961A (en) Neural network-based carpal tunnel image segmentation method and system
CN108629772A (en) Image processing method and device, computer equipment and computer storage media
CN113344933B (en) Glandular cell segmentation method based on multi-level feature fusion network
CN111161271A (en) Ultrasonic image segmentation method
CN110930378A (en) Emphysema image processing method and system based on low data demand
CN108491863A (en) Color image processing method based on Non-negative Matrix Factorization and convolutional neural networks
CN114037699B (en) Pathological image classification method, equipment, system and storage medium
CN115100165A (en) Colorectal cancer T staging method and system based on tumor region CT image
CN114897782A (en) Gastric cancer pathological section image segmentation prediction method based on generating type countermeasure network
CN110910388A (en) Cancer cell image segmentation method based on U-Net and density estimation
CN114612709A (en) Multi-scale target detection method guided by image pyramid characteristics
CN115082840B (en) Action video classification method and device based on data combination and channel correlation

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination