CN113066576A - Lung cancer screening method based on three-dimensional mask-area convolutional neural network - Google Patents
Lung cancer screening method based on three-dimensional mask-area convolutional neural network Download PDFInfo
- Publication number
- CN113066576A CN113066576A CN202110515164.8A CN202110515164A CN113066576A CN 113066576 A CN113066576 A CN 113066576A CN 202110515164 A CN202110515164 A CN 202110515164A CN 113066576 A CN113066576 A CN 113066576A
- Authority
- CN
- China
- Prior art keywords
- lung
- data
- neural network
- convolution
- dimensional
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Pending
Links
- 238000000034 method Methods 0.000 title claims abstract description 38
- 206010058467 Lung neoplasm malignant Diseases 0.000 title claims abstract description 29
- 201000005202 lung cancer Diseases 0.000 title claims abstract description 29
- 208000020816 lung neoplasm Diseases 0.000 title claims abstract description 29
- 238000013527 convolutional neural network Methods 0.000 title claims abstract description 22
- 238000012216 screening Methods 0.000 title claims abstract description 21
- 206010056342 Pulmonary mass Diseases 0.000 claims abstract description 26
- 210000004072 lung Anatomy 0.000 claims description 30
- 238000001514 detection method Methods 0.000 claims description 27
- 238000005070 sampling Methods 0.000 claims description 24
- 238000012549 training Methods 0.000 claims description 18
- 238000011176 pooling Methods 0.000 claims description 14
- 208000037170 Delayed Emergence from Anesthesia Diseases 0.000 claims description 11
- 238000013145 classification model Methods 0.000 claims description 11
- 230000003211 malignant effect Effects 0.000 claims description 11
- 238000013528 artificial neural network Methods 0.000 claims description 8
- 238000000605 extraction Methods 0.000 claims description 7
- 238000001914 filtration Methods 0.000 claims description 6
- 238000010276 construction Methods 0.000 claims description 5
- 238000007781 pre-processing Methods 0.000 claims description 5
- 238000012545 processing Methods 0.000 claims description 4
- 238000001574 biopsy Methods 0.000 claims description 3
- 210000004204 blood vessel Anatomy 0.000 claims description 3
- 238000004364 calculation method Methods 0.000 claims description 3
- 238000010586 diagram Methods 0.000 claims description 3
- 230000000694 effects Effects 0.000 claims description 3
- 230000000877 morphologic effect Effects 0.000 claims description 3
- 230000002685 pulmonary effect Effects 0.000 claims description 3
- 230000004913 activation Effects 0.000 claims description 2
- 238000002156 mixing Methods 0.000 claims description 2
- 230000009466 transformation Effects 0.000 claims 1
- 201000011510 cancer Diseases 0.000 abstract description 11
- 206010028980 Neoplasm Diseases 0.000 abstract description 9
- 230000003902 lesion Effects 0.000 abstract description 4
- 230000006870 function Effects 0.000 description 4
- 238000003745 diagnosis Methods 0.000 description 3
- 238000013135 deep learning Methods 0.000 description 2
- 230000009286 beneficial effect Effects 0.000 description 1
- 230000015556 catabolic process Effects 0.000 description 1
- 238000003759 clinical diagnosis Methods 0.000 description 1
- 238000006731 degradation reaction Methods 0.000 description 1
- 238000013461 design Methods 0.000 description 1
- 238000011161 development Methods 0.000 description 1
- 238000011156 evaluation Methods 0.000 description 1
- 230000006698 induction Effects 0.000 description 1
- 230000036210 malignancy Effects 0.000 description 1
- 239000003550 marker Substances 0.000 description 1
- 238000005457 optimization Methods 0.000 description 1
- 230000004044 response Effects 0.000 description 1
- 230000000630 rising effect Effects 0.000 description 1
- 230000035945 sensitivity Effects 0.000 description 1
- 238000006467 substitution reaction Methods 0.000 description 1
- 238000013519 translation Methods 0.000 description 1
Images
Classifications
-
- G—PHYSICS
- G16—INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR SPECIFIC APPLICATION FIELDS
- G16H—HEALTHCARE INFORMATICS, i.e. INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR THE HANDLING OR PROCESSING OF MEDICAL OR HEALTHCARE DATA
- G16H50/00—ICT specially adapted for medical diagnosis, medical simulation or medical data mining; ICT specially adapted for detecting, monitoring or modelling epidemics or pandemics
- G16H50/20—ICT specially adapted for medical diagnosis, medical simulation or medical data mining; ICT specially adapted for detecting, monitoring or modelling epidemics or pandemics for computer-aided diagnosis, e.g. based on medical expert systems
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F18/00—Pattern recognition
- G06F18/20—Analysing
- G06F18/24—Classification techniques
- G06F18/241—Classification techniques relating to the classification model, e.g. parametric or non-parametric approaches
- G06F18/2415—Classification techniques relating to the classification model, e.g. parametric or non-parametric approaches based on parametric or probabilistic models, e.g. based on likelihood ratio or false acceptance rate versus a false rejection rate
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/04—Architecture, e.g. interconnection topology
- G06N3/045—Combinations of networks
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/08—Learning methods
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
- G06T7/0002—Inspection of images, e.g. flaw detection
- G06T7/0012—Biomedical image inspection
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V10/00—Arrangements for image or video recognition or understanding
- G06V10/20—Image preprocessing
- G06V10/25—Determination of region of interest [ROI] or a volume of interest [VOI]
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V10/00—Arrangements for image or video recognition or understanding
- G06V10/40—Extraction of image or video features
- G06V10/44—Local feature extraction by analysis of parts of the pattern, e.g. by detecting edges, contours, loops, corners, strokes or intersections; Connectivity analysis, e.g. of connected components
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/10—Image acquisition modality
- G06T2207/10072—Tomographic images
- G06T2207/10081—Computed x-ray tomography [CT]
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/20—Special algorithmic details
- G06T2207/20081—Training; Learning
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/20—Special algorithmic details
- G06T2207/20092—Interactive image processing based on input by user
- G06T2207/20104—Interactive definition of region of interest [ROI]
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/30—Subject of image; Context of image processing
- G06T2207/30004—Biomedical image processing
- G06T2207/30061—Lung
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/30—Subject of image; Context of image processing
- G06T2207/30004—Biomedical image processing
- G06T2207/30096—Tumor; Lesion
Abstract
A lung cancer screening method based on a three-dimensional mask-area convolutional neural network is used for screening lung cancer. Because clinical CT of lung cancer shows that the lung cancer is lung nodules, the method comprises two steps, wherein a regression frame of suspected lung nodules is found in the first step, and the suspected nodules are classified in the second step, so that the nodules suspected as cancer lesions are found. The method detects lung nodules based on a three-dimensional mask-region convolutional neural network and enables the method to be applicable to CT data. Since not all nodules are cancerous, it is desirable to take into account the image signs of the patient to infer their likelihood of developing cancer. In order to improve the classification accuracy, the three-dimensional data is used for obtaining the surrounding structure of the suspected nodule, and the probability that the suspected nodule belongs to the cancer focus is comprehensively judged.
Description
Technical Field
The invention belongs to the field of artificial intelligent identification and auxiliary diagnosis of medical images, and particularly relates to a lung cancer screening method based on a three-dimensional mask-region convolutional neural network.
Background
The lung cancer patients in China are many, the incidence rate is high, and the medical cost is high. According to the rank of the number of the patients, lung cancer accounts for 20.3 percent of all cancers and is the first to be attacked by malignant tumors. According to the national cancer report of 2019 issued by the national tumor quality control center, the death of malignant tumors reaches 23.91 percent of the total death causes of residents, the morbidity and the mortality of the malignant tumors are in a continuously rising state in recent decades, and the medical cost caused by the malignant tumors exceeds 2200 hundred million each year. In the clinical diagnosis process, doctors need to screen possible lung cancer focuses one by one in hundreds of CT images, which not only depends greatly on the clinical experience of the doctors, but also is extremely time-consuming and labor-consuming. Therefore, there is an urgent need to design a computer-aided lung cancer screening system to simplify the screening work of doctors and shorten the diagnosis time.
With the rapid development of deep learning, lung cancer screening methods based on convolutional neural networks are widely studied. The traditional lung cancer detection method detects candidate nodules based on some simple prior information, for example, the nodules are approximately circular on a 2D image and have a high CT value, the prior information is summarized into mathematical characteristics, and then classification is performed through a classifier to obtain a detection result. However, due to the height difference of the lung nodules in shape, size and texture, it is difficult to capture and distinguish features through low-level feature extraction through artificial induction, resulting in poor detection results. Meanwhile, the judgment of the quality and the malignancy of the nodule only from a single image is limited, and although some 2D or 2.5D deep neural networks have good expression in the aspect of reducing false positive, the detection sensitivity can be effectively improved by using 3D CT data and combining three-dimensional spatial information of the lesion, which can be achieved by a few methods at present.
Disclosure of Invention
The invention aims to provide a lung cancer screening method based on a three-dimensional mask-region convolutional neural network, which is used for effectively judging the benign and malignant lung nodules and accurately carrying out early screening on lung cancer.
According to an aspect of the present invention, a lung cancer screening method based on a three-dimensional mask-area convolutional neural network comprises the following steps:
the first step is as follows: data marking, namely, carrying out CT thin-layer scanning data on the lungs of a plurality of different patients, recording the three-dimensional position of a lung nodule in the CT data, and marking the benign and malignant degree of the corresponding lung nodule according to clinical experience and biopsy information;
the second step is that: data preprocessing, namely interpolating the CT data into equal resolution in three directions, and cutting the data to only reserve a lung region;
the third step: constructing and training a detection model, and using a three-dimensional mask-region convolution-based neural network as a lung nodule detection model;
the fourth step: according to the trained model, carrying out lung nodule detection on the CT data, and outputting a pixel cube of a suspected lung nodule, wherein the pixel cube comprises the coordinates and the size of the center point of the lung nodule;
the fifth step: constructing and training a classification model, and using a three-dimensional convolutional neural network as a classification model of benign and malignant pulmonary nodules;
and a sixth step: and outputting the coordinates and the size of the lung nodule center point with the highest malignant probability for further clinical report.
Preferably, in the second step, in the data preprocessing:
2.1) obtaining the spatial resolution of the data according to the CT data in the DICOM format; taking the minimum resolution as a reference, and enabling the resolution of the data to be the same in three directions by using linear interpolation;
2.2) the lung of the human body is divided into two parts, namely a left lung and a right lung; meanwhile, most of the lung is air, and the CT value of the air is 0, all cavities with different CT values of 0 and sizes in the interpolated data are found through a region growing algorithm, and the volume of each cavity is calculated according to the resolution;
2.3) the volume of the human lung is between 500 and 10000cm according to experience3Within this volumeThe two cavities of range and largest are considered to be the lungs;
2.4) filling the cavity caused by the blood vessels in the lung by using morphological filtering, wherein the filtering range is set to be 5 pixels;
2.5) cutting the data after interpolation, only reserving the data of the lung area, and reducing the subsequent calculation time.
The invention has the beneficial effects that: a method for detecting and classifying lung nodules in a CT image by using a deep learning method is provided, and the method is used for detecting the lung nodules based on a three-dimensional Mask-region convolutional neural network (3D Mask R-CNN) and is suitable for CT data. Since not all nodules are cancerous, it is desirable to take into account the image signs of the patient to infer their likelihood of developing cancer. In order to improve the classification accuracy, the three-dimensional data is used for obtaining the surrounding structure of the suspected nodule, and the probability that the suspected nodule belongs to the cancer focus is comprehensively judged.
Drawings
FIG. 1 is a network structure of a detection model according to the present invention;
FIG. 2 is a schematic diagram of a ResNet-FPN network;
FIG. 3 is a classification model network structure;
FIG. 4 detects the FROC curve obtained by the model;
FIG. 5 areas of suspected lung cancer detected in actual cases;
FIG. 6 shows that the numerical value of the classification result of the suspected lung cancer region in the actual case is closer to 1, which indicates that the probability of malignant lesion (lung cancer) is higher;
fig. 7 shows the result visualized in a medical image viewing system, in the form of a red box, the lesion area of suspected lung cancer.
Detailed Description
The invention is further described with reference to the following description and examples.
In fig. 1-7, a lung cancer screening method based on a three-dimensional mask-area convolutional neural network includes the following steps: the first step is as follows: and (3) data marking, namely preparing lung CT thin-layer scanning data of a plurality of different patients, recording the three-dimensional position of a lung nodule in the CT data, and marking the benign and malignant degree of the corresponding lung nodule according to clinical experience and biopsy information.
The second step is that: and (3) data preprocessing, wherein the CT data are interpolated into equal resolution in three directions, and only the lung region is reserved by cutting the data.
The third step: and (3) constructing and training a detection model, and using a three-dimensional mask-area convolution-based neural network as a lung nodule detection model.
The fourth step: and according to the trained model, carrying out lung nodule detection on the CT data, and outputting a pixel cube of a suspected lung nodule, wherein the pixel cube comprises the coordinates and the size of the center point of the lung nodule.
The fifth step: and (3) constructing and training a classification model, and using a three-dimensional convolutional neural network as the classification model of benign and malignant pulmonary nodules.
And a sixth step: and outputting the coordinates and the size of the lung nodule center point with the highest malignant probability for further clinical report.
In some embodiments, in the second step, data is pre-processed
1.1 obtaining the spatial resolution of the data according to the CT data in the DICOM format; based on the minimum resolution, linear interpolation is used to make the data have the same resolution in three directions.
1.2 the lung of the human body is divided into 2 parts, namely a left lung and a right lung; meanwhile, most of the lung is air, and the CT value of the air is 0, so that all cavities with different CT values of 0 and sizes in the interpolated data are found through a region growing algorithm, and the volume of each cavity is calculated according to the resolution.
1.3 empirically, the volume of the human lung is 500 to 10000cm3The largest two cavities belonging to this volume range are considered to be the lungs
1.4 fill up the cavity due to the blood vessels in the lung using morphological filtering, with a filtering range set to 5 pixels.
And 1.5, cutting the data after interpolation, only reserving the data of the lung region, and reducing the subsequent calculation time.
In some embodiments, in the third step, the construction and training of the detection model:
2.1 the detection model uses a convolution neural network based on three-dimensional mask-area, the network structure is shown in fig. 1 and comprises three parts: the first part is a hierarchical connection structure consisting of ResNet-FPN, wherein the hierarchical connection structure comprises an up-sampling structure, a down-sampling structure and a cascade structure connected among layers; the second part is that the regional suggestion network generates regional suggestions for the extracted feature maps and carries out Region of interest alignment (ROI alignment); the third part is to calculate the loss function.
The 2.2ResNet-FPN network includes down-sampling from bottom to top, up-sampling from top to bottom, and cascading sections horizontally. The down-sampling path from bottom to top, consisting of a series of convolutional layers and max-pooling layers, as shown in fig. 2, is used for feature map extraction. The feature extraction is divided into 5 stages according to the size of the output feature map, and the final layer of output convolution 2, convolution 3, convolution 4 and convolution 5 of stage 2, stage 3, stage 4 and stage 5 are respectively defined as C2, C3, C4 and C5. In the downsampling stage of the FPN, the convolution layer is first, the convolution kernel size is 7 × 7 × 7, the step size is 2, and then a maximum pooling layer of 3 × 3 × 3 size is used to perform rough feature extraction on the image. The layers are connected by different numbers of residual blocks (ResBlock), C2 is connected by 3 residual blocks, C3 is connected by 4 residual blocks, C4 is connected by 6 residual blocks, and C5 is connected by 3 residual blocks. The final output of the residual unit is summed by the plurality of convolutional layer outputs and by the ingress and egress X, and then activated by the ReLU. The training parameters and the training calculated amount of the network are not increased in the process, but the training speed of the model can be greatly increased, and the training effect is improved. Meanwhile, when the layer number of the model is deepened, the structure can well solve the problem of network degradation.
Up-sampling from top to bottom, starting from the highest layer, using nearest neighbor sampling. Compared with common deconvolution operation, nearest neighbor sampling is simpler, training parameters can be reduced, and training time is shortened. The upsampling process is also divided into 5 stages corresponding to the downsampling, with M2, M3, M4 and M5 as the results. The results of the downsampling process, i.e., C2, C3, C4, and C5, were convolved each layer with a 1 x 1, sized convolution kernel, with the output channel consistent set to 36, and then summed with the corresponding signature. This is to keep the upsampling to get the same feature dimension. And then the characteristics after the convolution kernel addition of 3 multiplied by 3 are adopted for processing, so as to eliminate the mixing effect of the up-sampling. By the method, targets with different sizes are detected on the feature maps with different resolutions, and the detection accuracy of the targets with smaller sizes can be obviously improved.
2.3 obtaining regression frames generated by 9 kinds of anchor points in the characteristic diagram obtained by the FPN network, and then judging which anchor points are target candidate frames needing to be detected by using the RPN network. The detection image is convolved by a 3 multiplied by 3 shared weight with 128 channels of 36, and then the last feature map is compressed by one-dimensional convolution of 6 128 channels. Each anchor point can be a foreground or a background, foreground anchor points are obtained through Softmax classification, and regression frames of the detected target candidate regions are extracted preliminarily. The real bounding box of the detection target and the (Ax, Ay, Az, Ar) anchor bounding box are denoted by (Gx, Gy, Gz, Gr), where the first three elements represent the coordinate regression box of the center point and the last element represents the nodule size. A cross-over ratio (IoU) is used to determine the label of each anchor point. Anchors with targets greater than 0.5 and IoU less than 0.05 are considered positive and negative examples, respectively, and are ignored during retraining of other examples. The amount of translation (dx, dy, dz) between the front anchor and the true marker and the scale factor (dr) are as follows:
dx=(Gx–Ax)/Ar,
dy=(Gy–Ay)/Ar,
dz=(Gz–Az)/Ar,
dr=log(Gr/Ar)
2.4 feature location positioning can be more accurate using ROI alignment (Align) instead of ROI Pooling due to the feature map dimensions that are compressed by the Pooling (Pooling) operation, resulting in feature location inaccuracies. Since the region suggestions given by the RPN network are generally non-integer numbers, it is necessary to make the region integers, that is, divide the boundary region into n units on average, and perform rounding processing on each unit. However, after the integer number is processed, the obtained candidate frame is already deviated from the candidate frame proposed by the initial RPN network, thereby affecting the accuracy of the final regression frame. Compared with pooling operation, ROI alignment obtains coordinates by using a bilinear interpolation method, the values of pixel points are obtained, and then sampling is carried out. For example, the number of sampling points obtained in the feature map is 4, the cell is divided into four small squares, and the center of each small square is the sampling point. And performing maximum pooling operation on the sampling points in each cell to obtain a final ROI alignment result.
2.5 in the iterative optimization of the training model parameters, the loss function consists of three parts: regression loss, classification loss and mask loss. The overall loss function is defined as follows:
the prediction probability and label of an anchor point are denoted by p i and p, respectively, p ∈ {0,1} (negative samples are 0, positive samples are 1). The following table i in the penalty function is the table below for the anchor points, pi represents the probability that the region proposal is foreground, and pi is the predicted probability of the true regression box compared to the anchor points.
In the fifth step, the construction and training of the classification model:
3.1 the output of the detection network is used as the input of the classification network, and in order to better utilize three-dimensional data, the classification model uses a three-dimensional convolution neural network.
3.2 classification network As shown in FIG. 3, it consists of 6 three-dimensional convolutional layers, 3 three-dimensional max pooling layers, 2 full link layers and the final activation layer. 2 missing layers (Dropout Layer) were added simultaneously to avoid overfitting.
In consideration of clinical requirements, the performance evaluation of the screening method mainly comprises two indexes, namely recall rate and the number of false positives, and the recall rate is as high as possible within the acceptable range of the false positives, which indicates that missed diagnosis can be reduced as much as possible through a model in clinic. Considering that a patient may have multiple lung nodules that need to be detected, the present invention uses a free response receive operating curve (FROC) to evaluate model performance, with the results shown in fig. 4.
The following table demonstrates a comparison of the results of the present method with other methods, which achieved an average FROC score of 0.804 over other methods. In addition, the inventive method achieved optimal results at 2 FPs. In clinical practice, 2FPs is of greater concern because false positives in this case do not place too much of a diagnostic burden on the physician. The method has the recall rate of 89.9 percent at 2 FPs.
The foregoing is a more detailed description of the invention in connection with specific preferred embodiments and it is not intended that the invention be limited to these specific details. For those skilled in the art to which the invention pertains, several simple deductions or substitutions can be made without departing from the spirit of the invention, and all shall be considered as belonging to the protection scope of the invention.
Claims (8)
1. A lung cancer screening method based on a three-dimensional mask-area convolutional neural network is characterized by comprising the following steps:
the first step is as follows: data marking, namely, carrying out CT thin-layer scanning data on the lungs of a plurality of different patients, recording the three-dimensional position of a lung nodule in the CT data, and marking the benign and malignant degree of the corresponding lung nodule according to clinical experience and biopsy information;
the second step is that: data preprocessing, namely interpolating the CT data into equal resolution in three directions, and cutting the data to only reserve a lung region;
the third step: constructing and training a detection model, and using a three-dimensional mask-region convolution-based neural network as a lung nodule detection model;
the fourth step: according to the trained model, carrying out lung nodule detection on the CT data, and outputting a pixel cube of a suspected lung nodule, wherein the pixel cube comprises the coordinates and the size of the center point of the lung nodule;
the fifth step: constructing and training a classification model, and using a three-dimensional convolutional neural network as a classification model of benign and malignant pulmonary nodules;
and a sixth step: and outputting the coordinates and the size of the lung nodule center point with the highest malignant probability for further clinical report.
2. The lung cancer screening method based on the three-dimensional mask-area convolutional neural network as claimed in claim 1, wherein in the second step, in the data preprocessing:
2.1) obtaining the spatial resolution of the data according to the CT data in the DICOM format; taking the minimum resolution as a reference, and enabling the resolution of the data to be the same in three directions by using linear interpolation;
2.2) the lung of the human body is divided into two parts, namely a left lung and a right lung; meanwhile, most of the lung is air, and the CT value of the air is 0, all cavities with different CT values of 0 and sizes in the interpolated data are found through a region growing algorithm, and the volume of each cavity is calculated according to the resolution;
2.3) the volume of the human lung is 500 to 10000cm3The two largest cavities belonging to this volume range are considered to be the lungs;
2.4) filling the cavity caused by the blood vessels in the lung by using morphological filtering, wherein the filtering range is set to be five pixels;
2.5) cutting the data after interpolation, only reserving the data of the lung area, and reducing the subsequent calculation time.
3. The lung cancer screening method based on the three-dimensional mask-area convolutional neural network as claimed in claim 1, wherein in the third step, the construction and training of the detection model:
the detection model uses a three-dimensional mask-based, area convolution neural network, and comprises three parts: the first part is a hierarchical connection structure formed by a ResNet-FPN network, wherein the hierarchical connection structure comprises an up-sampling structure, a down-sampling structure and a cascade structure connected between layers; the second part is that the regional suggestion network generates regional suggestions for the extracted feature maps and carries out region-of-interest alignment; the third part is to calculate the loss function.
4. The lung cancer screening method based on the three-dimensional mask-area convolutional neural network as claimed in claim 3, wherein in the third step, the construction and training of the detection model:
the ResNet-FPN network comprises a down-sampling part from bottom to top, an up-sampling part from top to bottom and a transverse cascading part; and the down-sampling path from bottom to top consists of a series of convolution layers and maximum pooling layers and is used for feature map extraction.
5. The method of claim 4 for screening lung cancer based on a three-dimensional mask-area convolutional neural network, wherein: according to the size of the output feature map, feature extraction is divided into five stages, and the last output convolution two, convolution three, convolution four and convolution five in the stage two, stage three, stage four and stage five are respectively defined as C2, C3, C4 and C5; in the downsampling stage of FPN, firstly, a convolution layer is formed, the size of the convolution kernel is 7 multiplied by 7, the step length is 2, then, a maximum pooling layer with the size of 3 multiplied by 3 is formed, and rough feature extraction is carried out on an image; the layers are connected by different numbers of residual blocks, C2 is connected by three residual blocks, C3 is connected by four residual blocks, C4 is connected by six residual blocks, and C5 is connected by three residual blocks; the final output of the residual error unit is added by the output of a plurality of convolution layers and the input and output X, and then activated by the ReLU;
sampling from top to bottom from the highest layer, and using nearest neighbor sampling; corresponding to the down sampling, the up sampling process is also divided into no stages, and M2, M3, M4 and M5 are used as the results; convolving the results of the downsampling process, i.e., C2, C3, C4, and C5, each layer with a 1 × 1 convolution kernel, with the output channels consistently set to 36, and then adding the corresponding feature maps; and then the characteristics after the addition of the convolution kernels of 3 multiplied by 3 are adopted for processing so as to eliminate the mixing effect of the up-sampling.
6. The method of claim 5 for screening lung cancer based on a three-dimensional mask-area convolutional neural network, wherein: obtaining a regression frame generated by nine anchor points in a characteristic diagram obtained by the FPN network, and then judging which anchor points are target candidate frames needing to be detected by using the RPN network; the detection image is convolved by a 3 multiplied by 3 shared weight with 128 channels of which the number is 36, and then the last feature map is compressed by one-dimensional convolution of 6 128 channels; each anchor point can be a foreground or a background, foreground anchor points are obtained through Softmax classification, and a regression frame of a detection target candidate region is preliminarily extracted; representing a real boundary box of a detection target and an anchor point boundary box (Ax, Ay, Az, Ar) by (Gx, Gy, Gz, Gr), wherein the first three elements represent a coordinate regression box of a central point and the last element represents the size of a nodule; using the cross-over ratio to determine the label of each anchor point; anchors with targets greater than 0.5 and IoU less than 0.05 are considered positive and negative examples, respectively, and are ignored during retraining of other examples.
7. The method of claim 4 for screening lung cancer based on a three-dimensional mask-area convolutional neural network, wherein: pooling operation can compress feature map dimensions, cause inaccuracy of feature positions, and the feature positions can be more accurately positioned by replacing ROI pooling through ROI alignment; the region suggestions given by the RPN are generally non-integer numbers, and the region suggestions need to be integer numbers, namely, the boundary region is averagely divided into n units, and rounding processing is carried out on each unit; however, after the integer transformation, the obtained candidate frame has a deviation from the candidate frame proposed by the initial RPN network, thereby affecting the accuracy of the final regression frame; compared with pooling operation, ROI alignment obtains coordinates by using a bilinear interpolation method, the values of pixel points are obtained, and then sampling is carried out.
8. The method of claim 1 for lung cancer screening based on a three-dimensional mask-area convolutional neural network, wherein: in the fifth step, the construction and training of a classification model,
5.1) taking the output result of the detection network as the input of a classification network, wherein in order to better utilize three-dimensional data, a three-dimensional convolution neural network is used as a classification model;
5.1) the classification network comprises six three-dimensional convolution layers, three-dimensional maximum pooling layers, two full-connection layers and a final activation layer; two missing layers are added simultaneously to avoid overfitting.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202110515164.8A CN113066576A (en) | 2021-05-12 | 2021-05-12 | Lung cancer screening method based on three-dimensional mask-area convolutional neural network |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202110515164.8A CN113066576A (en) | 2021-05-12 | 2021-05-12 | Lung cancer screening method based on three-dimensional mask-area convolutional neural network |
Publications (1)
Publication Number | Publication Date |
---|---|
CN113066576A true CN113066576A (en) | 2021-07-02 |
Family
ID=76568739
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN202110515164.8A Pending CN113066576A (en) | 2021-05-12 | 2021-05-12 | Lung cancer screening method based on three-dimensional mask-area convolutional neural network |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN113066576A (en) |
Cited By (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN113990494A (en) * | 2021-12-24 | 2022-01-28 | 浙江大学 | Tic disorder auxiliary screening system based on video data |
CN115409815A (en) * | 2022-09-02 | 2022-11-29 | 山东财经大学 | Pulmonary nodule detection method based on three-dimensional deformable transformer |
Citations (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN108805209A (en) * | 2018-06-14 | 2018-11-13 | 清华大学深圳研究生院 | A kind of Lung neoplasm screening method based on deep learning |
CN112116603A (en) * | 2020-09-14 | 2020-12-22 | 中国科学院大学宁波华美医院 | Pulmonary nodule false positive screening method based on multitask learning |
-
2021
- 2021-05-12 CN CN202110515164.8A patent/CN113066576A/en active Pending
Patent Citations (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN108805209A (en) * | 2018-06-14 | 2018-11-13 | 清华大学深圳研究生院 | A kind of Lung neoplasm screening method based on deep learning |
CN112116603A (en) * | 2020-09-14 | 2020-12-22 | 中国科学院大学宁波华美医院 | Pulmonary nodule false positive screening method based on multitask learning |
Non-Patent Citations (2)
Title |
---|
林华明 等: "基于特征金字塔网络的肺结节检测", 《中国计量大学学报》, vol. 31, no. 3, 30 September 2020 (2020-09-30), pages 363 - 369 * |
高智勇 等: "基于特征金字塔网络的肺结节检测", 《计算机应用》, vol. 40, no. 09, 30 September 2020 (2020-09-30), pages 2571 - 2576 * |
Cited By (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN113990494A (en) * | 2021-12-24 | 2022-01-28 | 浙江大学 | Tic disorder auxiliary screening system based on video data |
CN113990494B (en) * | 2021-12-24 | 2022-03-25 | 浙江大学 | Tic disorder auxiliary screening system based on video data |
CN115409815A (en) * | 2022-09-02 | 2022-11-29 | 山东财经大学 | Pulmonary nodule detection method based on three-dimensional deformable transformer |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN109063710B (en) | 3D CNN nasopharyngeal carcinoma segmentation method based on multi-scale feature pyramid | |
CN111242174B (en) | Liver cancer image feature extraction and pathological classification method based on imaging omics | |
Hepsağ et al. | Using deep learning for mammography classification | |
CN110321920A (en) | Image classification method, device, computer readable storage medium and computer equipment | |
CN107492071A (en) | Medical image processing method and equipment | |
CN109363698A (en) | A kind of method and device of breast image sign identification | |
Fan et al. | Lung nodule detection based on 3D convolutional neural networks | |
US20030161522A1 (en) | Method, and corresponding apparatus, for automatic detection of regions of interest in digital images of biological tissue | |
CN113223005B (en) | Thyroid nodule automatic segmentation and grading intelligent system | |
CN113066576A (en) | Lung cancer screening method based on three-dimensional mask-area convolutional neural network | |
CN113592794B (en) | Spine graph segmentation method of 2D convolutional neural network based on mixed attention mechanism | |
CN110838114B (en) | Pulmonary nodule detection method, device and computer storage medium | |
CN111415728A (en) | CT image data automatic classification method and device based on CNN and GAN | |
CN112819747A (en) | Method for automatically diagnosing benign and malignant nodules based on lung tomography image | |
US20230005140A1 (en) | Automated detection of tumors based on image processing | |
Tan et al. | GLCM-CNN: gray level co-occurrence matrix based CNN model for polyp diagnosis | |
CN111260639A (en) | Multi-view information-collaborative breast benign and malignant tumor classification method | |
CN116030325A (en) | Lung nodule CT image recognition method based on deep hybrid learning framework | |
CN112270667A (en) | TI-RADS-based integrated deep learning multi-tag identification method | |
Weimin et al. | Enhancing Liver Segmentation: A Deep Learning Approach with EAS Feature Extraction and Multi-Scale Fusion | |
CN112785603A (en) | Brain tissue segmentation method based on Unet and superpixel | |
Haiying et al. | False-positive reduction of pulmonary nodule detection based on deformable convolutional neural networks | |
CN111784638A (en) | Pulmonary nodule false positive screening method and system based on convolutional neural network | |
CN112634308A (en) | Nasopharyngeal carcinoma target area and endangered organ delineation method based on different receptive fields | |
CN111724345A (en) | Pneumonia picture verification device and method capable of adaptively adjusting size of receptive field |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
WD01 | Invention patent application deemed withdrawn after publication | ||
WD01 | Invention patent application deemed withdrawn after publication |
Application publication date: 20210702 |