CN111144474A - Multi-view, multi-scale and multi-task lung nodule classification method - Google Patents

Multi-view, multi-scale and multi-task lung nodule classification method Download PDF

Info

Publication number
CN111144474A
CN111144474A CN201911354214.8A CN201911354214A CN111144474A CN 111144474 A CN111144474 A CN 111144474A CN 201911354214 A CN201911354214 A CN 201911354214A CN 111144474 A CN111144474 A CN 111144474A
Authority
CN
China
Prior art keywords
nodule
lung
semantic features
scale
nodules
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN201911354214.8A
Other languages
Chinese (zh)
Other versions
CN111144474B (en
Inventor
黄青松
张帅威
刘利军
冯旭鹏
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Kunming University of Science and Technology
Original Assignee
Kunming University of Science and Technology
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Kunming University of Science and Technology filed Critical Kunming University of Science and Technology
Priority to CN201911354214.8A priority Critical patent/CN111144474B/en
Publication of CN111144474A publication Critical patent/CN111144474A/en
Application granted granted Critical
Publication of CN111144474B publication Critical patent/CN111144474B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/24Classification techniques
    • G06F18/241Classification techniques relating to the classification model, e.g. parametric or non-parametric approaches
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/045Combinations of networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/08Learning methods
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V2201/00Indexing scheme relating to image or video recognition or understanding
    • G06V2201/03Recognition of patterns in medical or anatomical images

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • Data Mining & Analysis (AREA)
  • Evolutionary Computation (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Artificial Intelligence (AREA)
  • General Physics & Mathematics (AREA)
  • General Engineering & Computer Science (AREA)
  • Computing Systems (AREA)
  • Software Systems (AREA)
  • Molecular Biology (AREA)
  • Computational Linguistics (AREA)
  • Biophysics (AREA)
  • Biomedical Technology (AREA)
  • Mathematical Physics (AREA)
  • General Health & Medical Sciences (AREA)
  • Health & Medical Sciences (AREA)
  • Bioinformatics & Cheminformatics (AREA)
  • Bioinformatics & Computational Biology (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Evolutionary Biology (AREA)
  • Apparatus For Radiation Diagnosis (AREA)
  • Image Analysis (AREA)

Abstract

The invention relates to a multi-view, multi-scale and multi-task lung nodule classification method, and belongs to the technical field of medical image processing. The invention comprises the following steps: firstly, extracting 2D nodule slices of 9 views from a 3-D view; extracting 10mm, 20mm and 40mm patches on the 2D nodule slices; constructing 3 convolutional neural network models, training the patches extracted from each plane view according to the scale, and then fusing the full-link layers of the three models to perform feature fusion of each image of the lung nodule; and performing combined training on the semantic features of the lung nodules on the full-connection layer to obtain a classification result of the semantic features. The method classifies a plurality of semantic features of the pulmonary nodules, and finally accurately marks out the semantic features of the nodules by combining multiple classifications.

Description

Multi-view, multi-scale and multi-task lung nodule classification method
Technical Field
The invention relates to a multi-view, multi-scale and multi-task lung nodule classification method, and belongs to the technical field of medical image processing.
Background
Lung cancer is one of the leading causes of cancer death worldwide, with new cases estimated for lung cancer accounting for 13% of new cases of all cancers in the united states (45 cancers total) and mortality rates as high as 63% in 2019. The american National Lung Screening Test (NLST) indicates that lung cancer mortality can be reduced using low-dose CT screening, and therefore, detection and diagnosis of early stage lung cancer is very important for later stage treatment. With the advancement of screening techniques, the detection rate of nodules is increasing, but the diagnosis of lung nodules by radiologists is highly subjective, based on their level and experience.
Providing meaningful diagnostic features to enhance the objectivity of diagnostic decisions is a significant challenge. Research shows that semantic features of lung nodules can be obtained through image feature learning, and correlation exists among the semantic features. These semantic features include fineness, internal structure, calcification, sphericity, edges, lobulation, spiculation, texture, and malignancy. However, in the study of the relationship between the degree of malignancy and other semantic features, each semantic feature is independent of the other semantic features, and the semantic features are not combined.
Disclosure of Invention
The invention provides a multi-view, multi-scale and multi-task lung nodule classification method, which is used for classifying a plurality of semantic features of lung nodules and solving the problem of unbalanced combined multi-task classification data.
The technical scheme of the invention is as follows: the method for classifying the pulmonary nodules based on multiple views, multiple scales and multiple tasks comprises the following specific steps:
step1, extracting 2D nodule slices of 9 views from the 3-D view;
step2, extracting 10mm, 20mm and 40mm patches on the 2D nodule slice;
step3, constructing 3 convolutional neural network models, training the patches extracted from each plane view according to the scale, and then fusing the full-link layers of the three models to perform feature fusion of each image of the lung nodule;
and Step4, performing combined training on the semantic features of the lung nodules at the full-connection layer to obtain a classification result of the semantic features.
Further, the Step1 includes the specific steps of:
step1.1, extracting a 40X40mm cube centered on a nodule for each nodule candidate, the cube size being chosen to contain all nodule information and to include sufficient background information;
step1.2, in order to extract the pulmonary nodule image, nine slices are extracted on the plane corresponding to the 9 symmetric cubic planes; the three symmetrical cubic planes are planes parallel to the cube and are commonly called sagittal plane, coronal plane and axial plane respectively; the other six planes are planes of symmetry that cut diagonally two opposite faces of the cube, such planes containing two opposite edges and four vertices of the cube.
Further, in Step2, to be able to extract quantitative features from the original nodule block, rather than segmenting the nodule, we consider a nodule slice centered on the nodule with a slice interval thickness of 0.702 mm. The image of the lung nodule is at the center of each slice. Since the largest nodule diameter in our subset is 30mm, selecting a length of 40mm would allow all nodules to be completely contained in the slice, and then we applied the data expansion method;
for a lung nodule we first extracted 9 slices and then 10mm, 20mm, 40mm slices centered on the nodule. Thus, the effect of segmenting the pulmonary nodules is achieved, meanwhile, the original edge information of the pulmonary nodules is kept, and then a data balance method is adopted;
for extraction of nodule features on three scales, we adopt a VGG-NET network, and we try to train the Multi-SVT model by adding training samples, and supplement the learning process given limited training samples. We enhance the nodule by an image rotation operation. The rotation method is that the 3-D vertical shaft, the coronal shaft and the horizontal shaft are respectively used as central shafts to rotate for a certain angle, the rotating angle range is [ 0-360 degrees ], and then the dices are respectively carried out. A series of 40mm lung nodule slices with nodules distributed in the center were obtained. Finally, cutting out lung nodule slices of 10mm and 20mm by taking the nodule as the center on the basis of 40mm slices;
further, the Step3 includes the specific steps of:
step3.1, constructing 3 convolutional neural network models and training by using patches extracted from each plane view according to scales;
step3.2, respectively inputting the nodule slices of each scale into a Model A, a Model B and a Model C, and then performing feature fusion of each image of the pulmonary nodule on a full-connection layer;
step3.3, the input slice is processed through four convolution layers and one full join layer, and finally a softmax layer.
Further, the Step4 includes the specific steps of:
the image characteristics of the pulmonary nodules are fused in Step4.1 and Step3, and further 64-dimensional information is output by the convolutional neural network at the last full-connected layer through the convolutional layer of the convolutional neural network;
hiding the semantic features of Step4.2 and 9 types in the feature vector, and finally performing softmax classification on each semantic feature respectively; the loss value is obtained through calculation according to the softmax classification result and the lung nodule true label, 9 semantic features are considered simultaneously in combination with multitask, and back propagation is carried out according to the total loss value of the 9 semantic features.
Further, the semantic features comprise fineness, internal structure, calcification, sphericity, edge, lobular sign, burr sign, texture and malignancy parameters, and the internal structure comprises soft tissue, liquid, fat and air parameters.
The invention provides a Multi-SVT method for classifying the malignancy degree of lung nodules by combining nine semantic features, and the classification of other semantic features enhances the objectivity of lung nodule diagnosis decision. The Multi-SVT method takes a plurality of CT images centered on a nodule as model input, and outputs nine semantic features as models. According to the method, the 3D characteristics of the lung nodule are represented by nine views of the lung nodule, and for each view, a lung nodule image with three scales (10mm, 20mm and 40mm) is captured by taking the lung nodule as a center to represent nodule edge information and background information, instead of segmenting the lung nodule, so that the integrity of the lung nodule edge information is ensured. And then, respectively filling the nodule images of each scale into different convolutional neural networks, so as to ensure that model parameters of the lung nodule images of the same scale are shared, then, fusing the output features of the convolutional neural networks of three scales in a full-link layer, finally, performing combined multi-task classification on the nine semantic features of the lung nodules after the fused full-link layer, and training the model by minimizing a global loss function.
The steps of Step2, Step3 and Step4 are further described as follows:
1) the data expansion method comprises the following steps:
studies have shown that the more data the deep learning algorithm accesses is more efficient, while the overfitting of the model can be reduced by data enhancement. Our study can learn not only lung nodule 3D features through multiple views, but also expand the dataset through different combinations of views. In the first step, we reconstruct the whole lung in three dimensions using the DICOM file of the patient in the LIDC dataset, with the help of the software 3D Slicer. In the second step, we obtain the center coordinates L (X, Y, Z) of the nodule from the position information of the lung nodule marked by the radiologist, then cut a cube with the size of 40X40X40mm by taking L as the center, and then slice the cube by weight, with the image of the nodule at the center of each slice. Finally, the vertical shaft, the coronal shaft and the horizontal shaft are respectively used as central shafts to rotate for a certain angle, and the range of the rotation angle is [ 0-360 degrees ], and then the second step of operation is carried out.
The data set can be expanded through the three steps of operations, and the third diagram shows the combination of the original lung nodule slices and the combination of the lung nodule slices which rotate by 45 degrees respectively by taking a vertical axis, a coronal axis and a horizontal axis as central axes. In addition, combining slices at different rotation angles can further expand the data set, with the results shown in fig. 3, and the specific method shown in algorithm 1 in table 1.
TABLE 1 data expansion
Figure BDA0002335447870000031
Figure BDA0002335447870000041
2) The data balance method comprises the following steps:
when the training data set is unbalanced, i.e., the available data is not evenly distributed among different classes, the accuracy of the image classification technique is significantly reduced. In the LIDC dataset, the data distribution of semantic features is unbalanced, for example, 70% of nodules of a lobular feature are labeled as lab le 0, 30% of nodules are labeled as lab le 1, and the same is true for other semantic features. Our model combines nine semantic features of lung nodules, and processing any one of them changes the distribution of the other semantic features. Therefore, a proportion ranking algorithm is proposed, and nine semantic features are combined to relieve the condition of training set imbalance.
The nine semantic features have the same weight for the dataset as a whole. In the first step, the score of each nodule is calculated according to equation (1), and the nodules are then sorted from large to small according to this score. And secondly, selecting M nodes with the last ranking, copying the nodes with the last M nodes and adding the nodes to the whole data set. Finally, the above steps are repeated until the data sets are relatively balanced in terms of nine semantic features. Formula (1) is shown below:
Figure BDA0002335447870000042
in the formula, N is the total classification 46 of the 9 attributes of the lung nodule, d indicates whether the nodule belongs to the class, if yes, 1, otherwise 0, m is the total number of nodules, and N is the number of nodules belonging to the class.
For the lung nodules newly added in the above process, the data is not oversampled using the composite few sample oversampling technique (SMOTE), but rather the data is oversampled by rotating a series of new lung nodule images by a slight angle with the vertical axis, the coronal axis, and the horizontal axis of the nodule as the central axis, and the algorithm is shown in table 2 below.
TABLE 2 data Balancing Algorithm
Figure BDA0002335447870000051
The invention has the beneficial effects that:
1. according to the multi-view, multi-scale and multi-task lung nodule classification method, a 2D image is adopted to represent a 3D image, and multi-scale and multi-view images are fused to better represent pathological information of lung nodules;
2. the lung nodule semantic feature classification has the problem of data distribution imbalance, and a method is provided for solving the problem of data imbalance of combined multitask classification;
3. classifying a plurality of semantic features of the pulmonary nodule according to the lung nodule features learned by the model, and finally accurately marking out the nodule semantic features by combining multiple classifications.
Drawings
FIG. 1 is a flow chart of the present invention;
FIG. 2 is a flow chart of a multi-view, multi-scale, multi-tasking pulmonary nodule classification method of the present invention;
FIG. 3 is a diagram illustrating the result of the data expansion method of the present invention;
FIG. 4 is a result diagram of the prediction of nine semantic features of lung nodules in the present invention, wherein (a) is a case where the model predicts that all nine semantic features of lung nodules are correct, and (b) is a case where part of the model predicts that a semantic feature is incorrect;
FIG. 5 shows the selection of lung nodules and background information on the 30mm and 40mm scale involved in the present invention;
Detailed Description
Example 1: as shown in fig. 1 to 5, the multi-view, multi-scale and multi-task based lung nodule classification method specifically includes the following steps:
step1, extracting 2D nodule slices of 9 views from the 3-D view;
further, the Step1 includes the specific steps of:
step1.1, extracting a 40X40mm cube centered on a nodule for each nodule candidate, the cube size being chosen to contain all nodule information and to include sufficient background information;
wherein, a large public data set, lung image database alliance (LIDC-IDRI), is adopted to train and verify the proposed CAD system. LIDC-IDRI contained 1018 heterogeneous cases from 7 institutions. The layer thickness of the CT images varied from 0.6mm to 5.0mm with a median of 2.0 mm. For the images in each example, two-stage diagnostic labeling was performed by 4 experienced chest radiologists. In the first stage, each physician independently diagnoses and labels the location of the patient, wherein three categories are labeled according to the size of the lung nodule diameter: 1) 3mm nodes, 2) <3mm nodes, 3) > <3mm non-nodes. In the second stage, each physician independently reviews the labels of the other three physicians and gives the final diagnosis result. The two-stage labeling can completely label all results as far as possible on the premise of avoiding forced consensus. Each case consists of a DICOM image of the associated CT scan and an associated extensible markup language (XML) file containing nodule annotations for up to 4 physicians. The labeling of each semantic feature of a lung nodule depends entirely on past experience of physicians, and different physicians may give different semantic features for the same lung nodule.
Step1.2, in order to extract a required pulmonary nodule image, a three-dimensional integral pulmonary nodule is firstly established by a three-dimensional reconstruction method, then a range of 40x40x40mm is cut by taking the center of the pulmonary nodule as a cube center, and nine slices are extracted on a plane corresponding to a symmetric cube plane. Three of these planes of symmetry are planes parallel to the cube, commonly referred to as sagittal, coronal, and axial planes. The other six planes are symmetry planes which cut out two opposite faces of the cube on diagonal lines. Such a plane contains two opposite edges and four vertices of a cube. Thus, we obtained 9 slice views of each nodule.
The method comprises the steps of performing three-dimensional reconstruction on the whole lung of a patient through a software 3D slicer, finding out a nodule center coordinate L (X, Y, Z) according to position information of a lung nodule marked by an expert after reconstruction, and cutting a cube with the size of 40X40mm by taking the L as the center;
step2, extracting 10mm, 20mm and 40mm patches on the 2D nodule slice;
further, in Step2, to be able to extract quantitative features from the original nodule block, rather than segmenting the nodule, we consider a nodule slice centered on the nodule with a slice interval thickness of 0.702 mm. The image of the lung nodule is at the center of each slice. Since the largest nodule diameter in our subset is 30mm, selecting a length of 40mm would allow all nodules to be completely contained in the slice, and then we applied the data expansion method;
for a lung nodule we first extracted 9 slices and then 10mm, 20mm, 40mm slices centered on the nodule. Thus, the effect of segmenting the pulmonary nodules is achieved, meanwhile, the original edge information of the pulmonary nodules is kept, and then a data balance method is adopted;
since the definition of nodule boundaries is ambiguous, variability between radiologists' subjective readers makes accurate nodule delineation a challenging task. So to be able to extract quantitative features from the original nodule block, rather than segmenting the nodule, we consider a nodule slice centered on the nodule. For a lung nodule we first extracted 9 slices and then 10mm, 20mm, 40mm slices centered on the nodule. Therefore, the effect of segmenting the lung nodule is achieved, and the original edge information of the lung nodule is kept.
For extraction of nodule features on three scales, we adopt a VGG-NET network, and we try to train the Multi-SVT model by adding training samples, and supplement the learning process given limited training samples. We enhance the nodule by an image rotation operation. The rotation method is that the 3-D vertical shaft, the coronal shaft and the horizontal shaft are respectively used as central shafts to rotate for a certain angle, the rotating angle range is [ 0-360 degrees ], and then the dices are respectively carried out. A series of 40mm lung nodule slices with nodules distributed in the center were obtained. Finally, cutting out lung nodule slices of 10mm and 20mm by taking the nodule as the center on the basis of 40mm slices;
step3, constructing 3 convolutional neural network models, training the patches extracted from each plane view according to the scale, and then fusing the full-link layers of the three models to perform feature fusion of each image of the lung nodule;
further, the Step3 includes the specific steps of:
step3.1, constructing 3 convolutional neural network models and training by using patches extracted from each plane view according to scales;
step3.2, respectively inputting the nodule slices of each scale into a Model A, a Model B and a Model C, and then performing feature fusion of each image of the pulmonary nodule on a full-connection layer;
step3.3, the input slice is processed through four convolution layers and one full join layer, and finally a softmax layer.
And Step4, performing combined training on the semantic features of the lung nodules at the full-connection layer to obtain a classification result of the semantic features.
Further, the Step4 includes the specific steps of:
the image characteristics of the pulmonary nodules are fused in Step4.1 and Step3, and further 64-dimensional information is output by the convolutional neural network at the last full-connected layer through the convolutional layer of the convolutional neural network;
hiding the semantic features of Step4.2 and 9 types in the feature vector, and finally performing softmax classification on each semantic feature respectively; the loss value is obtained through calculation according to the softmax classification result and the lung nodule true label, 9 semantic features are considered simultaneously in combination with multitask, and back propagation is carried out according to the total loss value of the 9 semantic features.
Further, the semantic features comprise fineness, internal structure, calcification, sphericity, edge, lobular sign, burr sign, texture and malignancy parameters, and the internal structure comprises soft tissue, liquid, fat and air parameters.
To illustrate the effects of the present invention, the following experiments were performed:
experiment one: by preprocessing the pulmonary nodule annotation information, the grades of nine semantic features of the pulmonary nodule by four experts are unified, and errors caused by non-unified semantic feature grades are well solved. Since previous studies showed a correlation between nine semantic features of lung nodules, the nine semantic features of lung nodules were trained simultaneously in one model by a multitasking method. The Multi-SVT method not only can train a plurality of semantic features of the lung nodules through one model, but also can show positive influence on the classification of malignancy. For validation, the Multi-SV method was compared (as shown in FIG. 2). The Multi-SV method uses the same model structure as the Multi-SVT method, but the output of the model is only a feature tag of Maligancy, and the experimental results are shown in Table 3.
Table 3 comparison of the two methods.
Figure BDA0002335447870000081
The Multi-SV method and Multi-SVT method in table 3 summarize the average AUC score, accuracy, sensitivity and specificity of the two models. The average AUC of the Multi-SVT model is 0.952, the average accuracy is 0.913, the average sensitivity is 0.843, and the average specificity is 0.929; while the average AUC of the Multi-SV model was 0.932, the average accuracy was 0.905, the average sensitivity was 0.816, and the average specificity was 0.926. The results show that the proposed Multi-SVT method achieves better performance in predicting malignancy than the Multi-SV method.
Our results were also compared to the currently best deep-learning model for lung nodule malignancy prediction. The best performance is achieved at present in the classification of the degree of malignancy, and the model is explained by the classification of the other five semantic features. The average AUC scores, accuracy, sensitivity and specificity of the nine semantic features of the two models are summarized in table 4. On the semantic feature of the malignancy, the average AUC of the Multi-SVT model is 0.952, the average accuracy is 0.913, the average sensitivity is 0.843, and the average specificity is 0.929; the average AUC of the model of the method A is 0.856, the average accuracy is 0.842, the average sensitivity is 0.705, and the average specificity is 0.889. Index evaluation shows that the proposed Multi-SVT method achieves better performance in predicting malignancy than the best-performing method.
Table 4 representation of the method on other semantic features.
Figure BDA0002335447870000091
In addition, for other semantic feature prediction results, in addition to the malignancy considering other five semantic features (Subtlety, location, locality, Margin, Texture), our method considers all other eight semantic features. The same threshold is taken above five common semantic features, while we define the threshold of three features for the three semantic features of Internal Structure, Lobulation, and Spicification. As shown by the results in the table, we obtained average accuracies of 0.728,0.985,0.956,0.644, 0.802,0.750,0.764,0.911 for eight semantic features (subtitle, Internal Structure, location, locality, segmentation, Texture); mean AUC scores of 0.803,0.908,0.935,0.595,0.787,0.721, 0.781, 0.848; average sensitivities of 0.676,0.50,0.999,0.824,0.892,0.395,0.605, 0.955; the average specificity was 0.866,0.990,0.771,0.276,0.479,0.855,0.805, 0.595. Through comparison of the Multi-SVT method results and the HSCNN method in the table, the method not only realizes classification of all semantic features, but also has better classification results than the best method at present on the basis of five semantic features considered together.
The figure shows the working of the Multi-SVT model, by visualizing the View1-9 of lung nodules on a 40mm scale, while presenting the model's predictions of nine semantic features and the expert annotation results. Fig. 4(a) shows that the Multi-SVT model classifies the pulmonary nodules as benign, the contrast of the nodule region with the surrounding regions is large, the interior is composed of soft tissue, calcified, spherical, sharp-edged, burr-free, paginated, solid, and the prediction of these nine semantic features is the same as the true label.
Fig. 4(b) shows that the Multi-SVT model wrongly classifies a lung nodule as benign (the reference label is malignant), and also wrongly classifies 5 semantic features (subtlety, specificity, distribution, differentiation, and malignance) of the nodule, which indicates that the classification of the Malignancy is wrong, and the Multi-SVT model can also reasonably explain.
Experiment two: multiple-view comparative experiment
In our Multi-SVT model, we decomposed each 3-D lung nodule into 9 fixed views and segmented the 9 views at three different scales, resulting in 27 2D images for one nodule. To prove that multi-view ensemble learning is effective, we tested 3 sets of different view combinations.
Table 5 discusses multiple views
Figure BDA0002335447870000101
The classification performance given in table 5 shows that the more views we use, the higher the classification accuracy and AUC we obtained. The results are not surprising, as more and more informative nodules can be used in our model with more and more perspectives.
Experiment three: comparative experiment of multiple scales
We used images of lung nodules at three dimensions of 10mm, 20mm, and 40 mm. To prove that multi-scale integrated learning is effective, we tested 3 sets of different scale combinations. Since the 40mm scale can contain all lung nodules, we use this scale as the reference scale.
Table 6 discusses multiscale
Figure BDA0002335447870000102
The classification performance given in table 6 shows that the more scales we use, the higher the classification accuracy and AUC we obtained. The results are not surprising, since we have taken a multi-scale original nodule patch, rather than segmenting the region, providing evidence that the information obtained from the original nodule patch is valuable for lung nodule diagnosis.
Experiment four: comparative experiments with different fusion modes
We discuss the combination of different model input streams:
1) Committee-Fusion: the Committee-Fusion method is to merge the outputs of the fully connected layers of each stream and finally classify them. Each ConvNet flow is trained separately using the patches of a particular view. In this configuration, the convolutional layer parameters of different streams are not shared.
2) Late-Fusion: the Late-Fusion method concatenates the outputs of the first fully connected layer and connects the concatenated output directly to the sorting layer. In this configuration, the convolutional layer parameters of different streams are shared.
3) Mixed-Fusion: Mixed-Fusion is a combination of the first two approaches. We split the 27 patches into three independent sets, each set containing 9 different patches of the same scale, and within each set the Late-Fusion method was used, and between the sets the commit-Fusion method was used. In this configuration, the convolutional layer parameters for different streams within a set are shared, and the convolutional layer parameters for different streams between sets are not shared.
Table 7 discusses different fusion modes
Figure BDA0002335447870000111
The results are shown in Table 7. The table shows that we use a mechanism of dividing input stream sharing parameters by scale, and the obtained classification accuracy and AUC are higher. The results are not surprising, as scaling the model parameters more amenable to model parameter training.
Experiment five: selection of dimensions
For the design of the Multi-SVT model, the Multi-scale selection is chosen to better characterize the lung nodule image and surrounding background information, and we next discuss the reason for choosing the 10, 20, 40mm combination. The table shows the distribution of the diameter of the lung nodules, wherein the ratio of the area of 3-10mm is 81%, the ratio of the area of 10-20mm is 14%, the ratio of the area of 20-30mm is 3%, the ratio of the area of 30-40mm is 0.3%, and the rest is nodules with the diameter less than 3 mm. Therefore, choosing a maximum dimension of 40mm may absolutely encompass the entire nodule and surrounding tissue around the nodule, and not too much noise. As shown in fig. 5, when the maximum dimension is selected to be 30mm, part of the lung nodule information and the lung nodule background information are lost.
Consider that selecting 30mm as the largest dimension covers most lung nodules, but some of the nodule information is lost. Choosing 50mm as the largest dimension would cover the entire lung nodule information, but would introduce more noise. Next we discuss experiments with three combinations of 10&20&30mm, 10&20&40mm and 10&20&50 mm. For the three combined experiments, we selected a fixed combination of 9 views, a fused mode of Mixed-Fusion, and a fixed model structure. The results are shown in the table.
Table 8 discusses the choice of dimensions.
Figure RE-GDA0002435985540000121
Analysis of the results in Table 8 shows that the best results are obtained when a combination of 10&20&40mm is chosen. As analyzed, selecting a 10&20&30mm combination may lose background information for some nodules, whereas selecting 10&20&50mm may introduce more noise.
While the present invention has been described in detail with reference to the embodiments shown in the drawings, the present invention is not limited to the embodiments, and various changes can be made without departing from the spirit of the present invention within the knowledge of those skilled in the art.

Claims (5)

1. The lung nodule classification method based on multiple views, multiple scales and multiple tasks is characterized by comprising the following steps of: the method for classifying the pulmonary nodules based on multiple views, multiple scales and multiple tasks comprises the following specific steps:
step1, extracting 2D nodule slices of 9 views from the 3-D view;
step2, extracting 10mm, 20mm and 40mm patches on the 2D nodule slice;
step3, constructing 3 convolutional neural network models, training the patches extracted from each plane view according to the scale, and then fusing the full-link layers of the three models to perform feature fusion of each image of the lung nodule;
and Step4, performing combined training on the semantic features of the lung nodules at the full-connection layer to obtain a classification result of the semantic features.
2. The multi-view, multi-scale, multi-tasking lung nodule classification method of claim 1, wherein: the specific steps of Step1 are as follows:
step1.1, extracting a 40X40mm cube centered on a nodule for each nodule candidate, the size of the cube being selected to contain all nodule information and to include sufficient background information;
step1.2, in order to extract the pulmonary nodule image, nine slices are extracted on the plane corresponding to the 9 symmetric cubic planes; the three symmetrical cubic planes are planes parallel to the cube and are commonly called sagittal plane, coronal plane and axial plane respectively; the other six planes are planes of symmetry that cut diagonally two opposite faces of the cube, such planes containing two opposite edges and four vertices of the cube.
3. The multi-view, multi-scale, multi-tasking lung nodule classification method of claim 1, wherein: the specific steps of Step3 are as follows:
step3.1, constructing 3 convolutional neural network models and training by using patches extracted from each plane view according to scales;
step3.2, respectively inputting the nodule slices of each scale into a Model A, a Model B and a Model C, and then performing feature fusion of each image of the pulmonary nodule on a full-connection layer;
step3.3, the input slice is processed through four convolution layers and one full join layer, and finally a softmax layer.
4. The multi-view, multi-scale, multi-tasking lung nodule classification method of claim 1, wherein: the specific steps of Step4 are as follows:
the image characteristics of the pulmonary nodules are fused in Step4.1 and Step3, and further 64-dimensional information is output by the convolutional neural network at the last full-connected layer through the convolutional layer of the convolutional neural network;
hiding the semantic features of Step4.2 and 9 types in the feature vector, and finally performing softmax classification on each semantic feature respectively; the loss value is obtained through calculation according to the softmax classification result and the lung nodule real label, 9 semantic features are considered simultaneously in combination with multitask, and back propagation is carried out according to the total loss value of the 9 semantic features.
5. The multi-view, multi-scale, multi-tasking lung nodule classification method of claim 1 or 4, wherein: the semantic features comprise fineness, internal structure, calcification, sphericity, edge, lobulation sign, burr sign, texture and malignancy parameter, and the internal structure comprises soft tissue, liquid, fat and air parameter.
CN201911354214.8A 2019-12-25 2019-12-25 Multi-view, multi-scale and multi-task lung nodule classification method Active CN111144474B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201911354214.8A CN111144474B (en) 2019-12-25 2019-12-25 Multi-view, multi-scale and multi-task lung nodule classification method

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201911354214.8A CN111144474B (en) 2019-12-25 2019-12-25 Multi-view, multi-scale and multi-task lung nodule classification method

Publications (2)

Publication Number Publication Date
CN111144474A true CN111144474A (en) 2020-05-12
CN111144474B CN111144474B (en) 2022-06-14

Family

ID=70519868

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201911354214.8A Active CN111144474B (en) 2019-12-25 2019-12-25 Multi-view, multi-scale and multi-task lung nodule classification method

Country Status (1)

Country Link
CN (1) CN111144474B (en)

Cited By (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN111723817A (en) * 2020-06-30 2020-09-29 重庆大学 Pulmonary nodule auxiliary detection method
CN111767952A (en) * 2020-06-30 2020-10-13 重庆大学 Interpretable classification method for benign and malignant pulmonary nodules
CN112116562A (en) * 2020-08-26 2020-12-22 重庆市中迪医疗信息科技股份有限公司 Method, device, equipment and medium for detecting focus based on lung image data
CN112633336A (en) * 2020-12-10 2021-04-09 重庆大学 Pulmonary nodule identification and classification method based on multi-scale feature fusion
CN112990344A (en) * 2021-04-09 2021-06-18 重庆大学 Multi-view classification method for pulmonary nodules
CN113191391A (en) * 2021-04-07 2021-07-30 浙江省交通运输科学研究院 Road disease classification method aiming at three-dimensional ground penetrating radar map
CN113436726A (en) * 2021-06-29 2021-09-24 南开大学 Automatic lung pathological sound analysis method based on multi-task classification

Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN106650830A (en) * 2017-01-06 2017-05-10 西北工业大学 Deep model and shallow model decision fusion-based pulmonary nodule CT image automatic classification method
CN106940816A (en) * 2017-03-22 2017-07-11 杭州健培科技有限公司 Connect the CT image Lung neoplasm detecting systems of convolutional neural networks entirely based on 3D
CN108389178A (en) * 2018-01-11 2018-08-10 上海交通大学 Lung CT preprocess method based on convolutional neural networks and system
CN110309860A (en) * 2019-06-06 2019-10-08 昆明理工大学 The method classified based on grade malignancy of the convolutional neural networks to Lung neoplasm

Patent Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN106650830A (en) * 2017-01-06 2017-05-10 西北工业大学 Deep model and shallow model decision fusion-based pulmonary nodule CT image automatic classification method
CN106940816A (en) * 2017-03-22 2017-07-11 杭州健培科技有限公司 Connect the CT image Lung neoplasm detecting systems of convolutional neural networks entirely based on 3D
CN108389178A (en) * 2018-01-11 2018-08-10 上海交通大学 Lung CT preprocess method based on convolutional neural networks and system
CN110309860A (en) * 2019-06-06 2019-10-08 昆明理工大学 The method classified based on grade malignancy of the convolutional neural networks to Lung neoplasm

Non-Patent Citations (2)

* Cited by examiner, † Cited by third party
Title
LIJUN LIU 等: "MobMivs: Implementing an Efficient Medical Image Visualization System for Mobile Telemedicine", 《2018 9TH INTERNATIONAL CONFERENCE ON INFORMATION TECHNOLOGY IN MEDICINE AND EDUCATION》 *
吴保荣: ""基于深度卷积神经网络的肺结节自动检测和分类方法研究 "", 《中国优秀硕士学位论文全文数据库 (医药卫生科技辑)》 *

Cited By (11)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN111723817A (en) * 2020-06-30 2020-09-29 重庆大学 Pulmonary nodule auxiliary detection method
CN111767952A (en) * 2020-06-30 2020-10-13 重庆大学 Interpretable classification method for benign and malignant pulmonary nodules
CN111723817B (en) * 2020-06-30 2023-09-29 重庆大学 Auxiliary detection method for pulmonary nodules
CN111767952B (en) * 2020-06-30 2024-03-29 重庆大学 Interpretable lung nodule benign and malignant classification method
CN112116562A (en) * 2020-08-26 2020-12-22 重庆市中迪医疗信息科技股份有限公司 Method, device, equipment and medium for detecting focus based on lung image data
CN112633336A (en) * 2020-12-10 2021-04-09 重庆大学 Pulmonary nodule identification and classification method based on multi-scale feature fusion
CN113191391A (en) * 2021-04-07 2021-07-30 浙江省交通运输科学研究院 Road disease classification method aiming at three-dimensional ground penetrating radar map
CN112990344A (en) * 2021-04-09 2021-06-18 重庆大学 Multi-view classification method for pulmonary nodules
CN112990344B (en) * 2021-04-09 2022-05-24 重庆大学 Multi-view classification method for pulmonary nodules
CN113436726A (en) * 2021-06-29 2021-09-24 南开大学 Automatic lung pathological sound analysis method based on multi-task classification
CN113436726B (en) * 2021-06-29 2022-03-04 南开大学 Automatic lung pathological sound analysis method based on multi-task classification

Also Published As

Publication number Publication date
CN111144474B (en) 2022-06-14

Similar Documents

Publication Publication Date Title
CN111144474B (en) Multi-view, multi-scale and multi-task lung nodule classification method
Guan et al. Deep convolutional neural network VGG-16 model for differential diagnosing of papillary thyroid carcinomas in cytological images: a pilot study
CN107016665B (en) CT pulmonary nodule detection method based on deep convolutional neural network
Wang et al. Using deep convolutional neural networks for multi-classification of thyroid tumor by histopathology: a large-scale pilot study
CN109583440A (en) It is identified in conjunction with image and reports the medical image aided diagnosis method edited and system
Niazi et al. Visually meaningful histopathological features for automatic grading of prostate cancer
Dequidt et al. Exploring radiologic criteria for glioma grade classification on the BraTS dataset
CN111767952A (en) Interpretable classification method for benign and malignant pulmonary nodules
CN101551854A (en) A processing system of unbalanced medical image and processing method thereof
Wang et al. 3D Inception U‐net with asymmetric loss for cancer detection in automated breast ultrasound
CN111798424B (en) Medical image-based nodule detection method and device and electronic equipment
Feng et al. Deep learning for chest radiology: a review
Yuan et al. An efficient multi-path 3D convolutional neural network for false-positive reduction of pulmonary nodule detection
Zhi et al. Deep neural network pulmonary nodule segmentation methods for CT images: Literature review and experimental comparisons
Tiwari et al. Optimized Ensemble of Hybrid RNN-GAN Models for Accurate and Automated Lung Tumour Detection from CT Images
Cao et al. 3D convolutional neural networks fusion model for lung nodule detection onclinical CT scans
Yu et al. Convolutional neural network design for breast cancer medical image classification
Ma et al. Automatic pulmonary ground‐glass opacity nodules detection and classification based on 3D neural network
Mir et al. Artificial intelligence-based techniques for analysis of body cavity fluids: a review
Parvatikar et al. Prototypical models for classifying high-risk atypical breast lesions
Joshua et al. Lung Cancer Classification with Improvised Three Parameter Logistic Type Distribution Model.
Devi et al. Brain Tumour Detection with Feature Extraction and Tumour Cell Classification Model using Machine Learning–A Survey
Prithvika et al. A Review on Identification and Classification of Lung Nodules
Khouadja et al. Lung Cancer Detection with Machine Learning and Deep Learning: A Narrative Review
Mohamed Aarif et al. Deep MammoNet: Early Diagnosis of Breast Cancer Using Multi-layer Hierarchical Features of Deep Transfer Learned Convolutional Neural Network

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant