CN115471838A - Cervical squamous lesion cell detection method based on depth self-adaptive feature extraction - Google Patents

Cervical squamous lesion cell detection method based on depth self-adaptive feature extraction Download PDF

Info

Publication number
CN115471838A
CN115471838A CN202211161135.7A CN202211161135A CN115471838A CN 115471838 A CN115471838 A CN 115471838A CN 202211161135 A CN202211161135 A CN 202211161135A CN 115471838 A CN115471838 A CN 115471838A
Authority
CN
China
Prior art keywords
cells
lesion
detection
cervical squamous
feature extraction
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN202211161135.7A
Other languages
Chinese (zh)
Inventor
李佐勇
彭中华
陈春强
胡蓉
林滨
樊好义
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Minjiang University
Original Assignee
Minjiang University
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Minjiang University filed Critical Minjiang University
Priority to CN202211161135.7A priority Critical patent/CN115471838A/en
Publication of CN115471838A publication Critical patent/CN115471838A/en
Pending legal-status Critical Current

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V20/00Scenes; Scene-specific elements
    • G06V20/60Type of objects
    • G06V20/69Microscopic objects, e.g. biological cells or cellular parts
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/08Learning methods
    • G06N3/082Learning methods modifying the architecture, e.g. adding, deleting or silencing nodes or connections
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/40Extraction of image or video features
    • G06V10/42Global feature extraction by analysis of the whole pattern, e.g. using frequency domain transformations or autocorrelation
    • G06V10/422Global feature extraction by analysis of the whole pattern, e.g. using frequency domain transformations or autocorrelation for representing the structure of the pattern or shape of an object therefor
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/70Arrangements for image or video recognition or understanding using pattern recognition or machine learning
    • G06V10/82Arrangements for image or video recognition or understanding using pattern recognition or machine learning using neural networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V20/00Scenes; Scene-specific elements
    • G06V20/60Type of objects
    • G06V20/69Microscopic objects, e.g. biological cells or cellular parts
    • G06V20/695Preprocessing, e.g. image segmentation
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V2201/00Indexing scheme relating to image or video recognition or understanding
    • G06V2201/07Target detection

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Health & Medical Sciences (AREA)
  • Multimedia (AREA)
  • General Health & Medical Sciences (AREA)
  • Molecular Biology (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Biomedical Technology (AREA)
  • Evolutionary Computation (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Computing Systems (AREA)
  • Software Systems (AREA)
  • Artificial Intelligence (AREA)
  • General Engineering & Computer Science (AREA)
  • Data Mining & Analysis (AREA)
  • Mathematical Physics (AREA)
  • Computational Linguistics (AREA)
  • Biophysics (AREA)
  • Databases & Information Systems (AREA)
  • Medical Informatics (AREA)
  • Image Analysis (AREA)
  • Investigating Or Analysing Biological Materials (AREA)

Abstract

The invention relates to a cervical squamous lesion cell detection method based on depth self-adaptive feature extraction. On the basis of YOLOv5, a Transformer module is used in a high-level feature layer of a backbone network, so that the detection accuracy of 4 types (ASCUS, ASCH, HSIL and LISL) of cervical squamous lesion cells can be effectively distinguished; in addition, because the smear background is complex when detecting cells, the smear background is full of free cytoplasm, nuclei and other impurities; aiming at the problem of the situation, the invention adopts a bidirectional feature pyramid (BiFPN) to adaptively extract features of different layers in the picture so as to reduce the influence of a complex background on the detection of the squamous lesion cells. Furthermore, because cells are stacked and agglomerated, the sizes of detection frames for independent cells and agglomerated cells are different, and therefore, in the invention, attention volume blocks (CBAM) are used to adaptively learn the size of the final cell detection frame. The invention can improve the detection precision of the cervical squamous lesion cells.

Description

Cervical squamous lesion cell detection method based on depth adaptive feature extraction
Technical Field
The invention belongs to the technical field of image processing, and particularly relates to a cervical squamous lesion cell detection method based on depth adaptive feature extraction.
Background
Currently, cancer is a leading cause of death in people less than 70 years of age worldwide. According to the data published by GLOBOCAN in 2020, the number of cancer attacks is nearly 1930 ten thousand and the number of deaths is nearly 1000 ten thousand globally. The number of onset and death of cervical cancer is 60.4 ten thousand and 46.6 ten thousand, respectively, the number of onset accounts for 6.5% of new cases of cancer worldwide, and 7 th is the average of the number of onset and death. With a higher ranking of female cancer morbidity and mortality being 4 th. Studies according to WTO have shown that if women in developing countries are able to do a precautionary check for the pre-cervical cancer every 5 years, the overall prevalence can drop by 60% or more. At present, according to the most effective cervical cancer screening method proven by clinical medicine, doctors mainly analyze morphological structures of cell images by observing cervical smears through an optical microscope, and finally, diagnosis is performed according to the cytopathology principle. But this method undoubtedly requires a large number of experienced medical workers and the task of reviewing smears is very tedious, laborious and laborious.
The current assisted screening system for cervical cell smears is mainly assisted from three aspects: cervical cell classification, detection and segmentation. The smear classification comprises cell classification and smear classification, and is different from the cell classification, and the purpose of the smear classification is mainly to predict the category of the whole smear; the division includes the strategy of firstly dividing and then adjusting, and firstly detecting and then finely dividing; the cervical cell detection is mainly divided into two types, one type is one-stage target detection, the model is mainly based on variant models of YOLO, retinaNet and EfficientNet, the characteristics of the models are obvious, the overall reasoning speed is particularly high because the part is not like ROI recommendation and classification of two stages, but the defect is obvious, and the precision is relatively lower than that of the detection model of two stages. The other is a two-stage detection model, which is mainly based on Fast R-CNN and Fast R-CNN variant models, and the models are characterized by higher detection accuracy than the one-stage model as mentioned above. However, due to the continuous iterative updating of the model, the precision of the current partial one-stage detection model can be comparable to that of the current two-stage model, the current examination volume is extremely large, and the smear is time-efficient, so that a fast and accurate detection model is urgently needed at present. Early classification and segmentation tasks were based on a few single-cell datasets that had been trimmed. The classification task at that time is usually to analyze the distribution state of the cell nucleus and cytoplasm, and the analysis method is usually to train the segmented cell nucleus and cytoplasm information to a specific classifier to obtain a classification model, while the segmentation methods used are usually algorithms such as threshold segmentation, watershed segmentation and C-means. For example, thanatip uses a fuzzy C-means clustering method to segment cellular images, and finally trains its own classifier. Willian, by training out the Weka segmentation classifier and then combining the fuzzy C-means algorithm to classify. The above conventional algorithms are based on extracting the cell tissue features by using a segmentation algorithm in advance, in other words, a segmentation algorithm capable of well segmenting the cell nucleus and the cell cytoplasm is required as the guarantee of the overall classification accuracy. Since CNN exhibits excellent ability in object feature extraction and the use of CNN to extract cell features does not require the use of segmentation algorithms to acquire cell tissue information in advance, many CNN-based classifications have emerged in recent research. These models have achieved superior results to earlier studies in the classification dataset of Herlev and HEMLBC (H & E stabilized manual LBC), but are difficult to apply in realistic clinical medicine to assist medical staff in analytical classification because the datasets they apply are relatively simple. Most of cervical cell smears are overlapped, most of cells are still required to be cut into single cells or cell groups for classification in the classification process, the overall reasoning time is in direct proportion to the number of the cells, and the cutting of the smear is a very tedious matter, so that the auxiliary effect of the classification task on the actual clinical medicine is extremely limited.
Of course, in the former work, there are many target detection models based on one stage, which are mostly modified and supplemented based on YOLO, retinaNet and EfficientNet, and they have used their own data sets to sufficiently train and optimize the models, but in the final result, the effect on the detection of squamous cells is not particularly ideal. Since the features between squamous cells are so similar that they are indistinguishable, as shown in figure 1, there is a strong desire in both current model research as well as clinical medicine to address this problem.
The challenges of image detection of squamous cervical lesion cells are: (1) In clinical medicine, squamous lesion cells are very similar in their characteristics, and there are often cases that are difficult to distinguish, which are not so friendly for subsequent pathological analysis; (2) Because cervical smears often present poor quality samples, there are cases where multiple cells are stacked together, leading to detection difficulties; (3) The criteria for determining the lesion cells in different age groups are different, and therefore it is often necessary to consider other cells around one cell to determine whether it is a lesion squamous cell. (4) The smear background is complex when detecting cells, and is full of free cytoplasm, nucleus and other impurities, which can affect the detection result of cervical squamous lesion cells to different degrees.
Disclosure of Invention
The invention aims to improve the detection precision of cervical squamous lesion cells and provides a cervical squamous lesion cell detection method based on depth adaptive feature extraction.
In order to realize the purpose, the technical scheme of the invention is as follows: a cervical squamous lesion cell detection method based on depth adaptive feature extraction uses a Transformer module in a high-level feature layer of a trunk network on the basis of YOLOv 5; the neck network adopts a bidirectional feature pyramid Bi-FPN to adaptively extract features of different levels in the picture; the prediction head module uses the attention convolution block CBAM to adaptively learn the size of the final cell detection box.
Compared with the prior art, the invention has the following beneficial effects:
in order to enhance the acquisition capability of the model for global information and better extract the cell characteristics in a self-adaptive manner, the invention uses a Transformer module in a high-level characteristic layer of a backbone network on the basis of YOLOv5, so that the detection accuracy of the cervical squamous lesion cells of 4 types (ASCUS, ASCH, HSIL and LISL) can be effectively distinguished. In addition, because the background of the smear is complex when detecting cells, it is filled with some free cytoplasm, nuclei and other impurities. Aiming at the problem of the situation, the invention adopts a bidirectional feature pyramid (BiFPN) to adaptively extract features of different levels in the picture so as to reduce the influence of a complex background on the detection of squamous lesion cells. Furthermore, because cells are stacked and agglomerated, the sizes of detection frames for independent cells and agglomerated cells are different, and therefore, in the invention, attention volume blocks (CBAM) are used to adaptively learn the size of the final cell detection frame. The algorithm of the invention is proved to have different degrees of improvement on 4 types of lesion cells in a large number of squamous cell detection experiments.
Drawings
FIG. 1 is an example of a white blood cell image prepared by rapid and standard staining.
Fig. 2 is a block diagram of the algorithm of the present invention.
Fig. 3 is a backbone network: (left) original backbone network structure; (right) backbone network structure after optimization using Transformer Block.
Fig. 4 is a neck network: (a) a feature pyramid; (b) a pixel aggregation network; (c) a feature pyramid network of neural architecture search; (d) a bidirectional feature pyramid network.
Fig. 5 shows a prediction head module: attention convolution module.
FIG. 6 is a random affine example graph: (a) a translation transformation; (b) a scaling transform; (c) a shear transformation; and (d) rotating and transforming.
FIG. 7 is an exemplary illustration of mixing enhancement: (a) original 1; (b) Effect after performing Mosaic enhancement together with other sample pictures; (c) original figure 2; (d) Effect after MixUp enhancement with other sample pictures.
Fig. 8 is sample 1 and the corresponding label.
FIG. 9 shows the results of the detection of the present invention on sample 1.
Fig. 10 is sample 2 and the corresponding label.
FIG. 11 shows the results of the test of the present invention on sample 2.
Fig. 12 shows sample 3 and the corresponding label.
FIG. 13 shows the results of the detection of the present invention on sample 3.
Detailed Description
The technical scheme of the invention is specifically explained below with reference to the accompanying drawings.
The invention relates to a cervical squamous lesion cell detection method based on depth adaptive feature extraction, which uses a Transformer module in a high-level feature layer of a trunk network on the basis of YOLOv 5; the neck network adopts a bidirectional feature pyramid Bi-FPN to adaptively extract features of different levels in the picture; the prediction head module uses the attention convolution block CBAM to adaptively learn the size of the final cell detection box.
The following is a specific implementation process of the present invention.
As shown in FIG. 1, in order to enhance the acquisition capability of the model for global information and better extract the cell features in an adaptive manner, the invention uses a Transformer module in a high-level feature layer of a backbone network on the basis of YOLOv5, so that the detection accuracy of 4 types (ASCUS, ASCH, HSIL, LISL) of cervical squamous lesion cells can be effectively distinguished. In addition, because the background of the smear is complex when detecting cells, it is filled with some free cytoplasm, nuclei and other impurities. Aiming at the problem of the situation, the invention adopts a bidirectional feature pyramid (BiFPN) to adaptively extract features of different layers in the picture so as to reduce the influence of a complex background on the detection of the squamous lesion cells. Furthermore, because cells are stacked and clustered, the sizes of detection frames for independent cells and clustered cells are different, and therefore, in the invention, the attention volume block (CBAM) is used for adaptively learning the size of the final cell detection frame.
1. Network architecture
1.1, backbone network
The invention is a model optimized after modularization modification on the basis of YOLOv5, a main network uses CSP-Darknet53, but the 9 th layer and the 11 th layer are modified, and a Transformer Block capable of better acquiring global information is used for replacing an original C3Block, and the model is specifically shown in FIG. 3. The reason why the replacement is carried out at the end of the backbone is that ZFNET proves that the deep neural network extracts high-dimensional features which can represent overall information more than low-dimensional features, and the calculation amount after the replacement is not increased too much. Furthermore, since ViT uses Transformer in the image classification task of computer vision and embodies the strong feature extraction capability of Transformer, we also want to use Transformer to help the model distinguish the features between different types of squamous lesion cells.
1.2 neck network
According to the description of EfficientDet, although FPN and PAN are widely applied to multi-scale feature fusion of targets in various target detection models. But for the importance of features of different scales, neither FPN nor PAN do a detailed analysis of this, and almost all treat this information indiscriminately. In order to solve the problem, the invention designs Bi-FPN by a weighting method, and the structure chart is shown in figure 4 (d), wherein the Bi-FPN introduces an extra weight to each input of the network, deletes the node with only one input edge and adds a jump connection between the input node and the output node in the same scale. In general, bi-FPN is equivalent to endowing different weights to each layer for fusion, so that the network focuses more on important layers, and node connection of some unnecessary layers is reduced. The authors demonstrated the efficient and accurate performance of this model in a large number of experiments. Therefore, when a model for detecting cervical squamous lesion cells is designed, the original PAN is replaced by the Bi-FPN, so that the overall model can better process features of different scales. And the final experiment shows that the use of Bi-FPN can increase the identification accuracy of the model for different classes while increasing the calculation amount as little as possible.
1.3 prediction of head modules
An attention mechanism module representing a convolution module. Is an attention mechanism module combining space (spatial) and channel (channel). Compared with the attention mechanism that SeNet only focuses on channels, the attention mechanism can achieve better effect. The module is simple and efficient, and can carry out attention extraction on the intermediate characteristic diagram from different angles, so that the characteristics of the input characteristic diagram can be refined in a self-adaptive mode. Moreover, because the CBAM is an end-to-end model structure, the CBAM can be flexibly applied to different models, and the CBAM has high-efficiency performance in extracting target features by using the CBAM in a large number of classical models and carrying out experiments in different data sets. The details of the model are in figure 5. Since most of the images we detect have complex backgrounds, such as free nuclei, cytoplasm and mucus silk or other impurities, and there are cases of large cell stacking, we use CBAM module to assist YOLO capture on the target object to be detected.
1.4 data enhancement
1.4.1 random affine enhancement
We have considered at the beginning of the training of the model that the network misidentifies cells because the deformation of the samples may cause the cells to change from positive to negative samples. We have not initially been enhanced with random biomimetic transformation data that would cause cell deformation, such as Translation (Translation), scale (Scale), flip (Flip), rotation (Rotation), and Shear (Shear), as shown in fig. 6. But the overall effect is quite poor, in one chance we erroneously use data enhancement that may lead to deformation, and the final effect is particularly excellent. Then we begin to explore the data enhancement mode most suitable for this task, and finally add the parameters of the method in table 1, so that the overall mAP is 2.5mAP higher than the case of using default data enhancement.
Table 1: random affine data enhancing corresponding parameters
Figure BDA0003858275870000051
1.4.2, mixed data enhancement
In order to enable the model of the invention to have stronger robustness under some complex environments, more complex data enhancement is used for combining different samples, a) Mosaic fuses a plurality of pictures into one picture, thereby increasing the diversity of the samples; b) The MixUp is to fuse two pictures together according to a certain transparency, so that the training difficulty is increased, and the model has better semantic analysis capability. The specific effect is shown in fig. 7.
1.5 loss function
The loss function of the present invention consists of three parts
1.5.1 Object Score loss
When detecting the object, the model uses two-class cross entropy loss which can judge whether the object is in the frame of the target frame, wherein o is a label vector, t is a prediction vector:
Figure BDA0003858275870000061
1.5.2 Class Probasic Score loss
When detecting classes, the model uses two-class cross entropy loss that can determine whether the target frame has correct class in the frame, as with Object Score, where o is the label vector, t is the predictor vector:
Figure BDA0003858275870000062
1.5.3 Bounding Box loss
When detecting whether the target object can be compactly framed in the prediction frame, the model uses the CIOU Loss with more excellent performance:
Figure BDA0003858275870000063
wherein α is:
Figure BDA0003858275870000064
v is:
Figure BDA0003858275870000065
wherein D is 2 To predict the distance between the frame and the center point of the target frame, D c Is the diagonal distance of the minimum circumscribed rectangle; w and h are the height and width of the prediction box, w gt And h gt Is the height and width of the real box.
To evaluate the performance of the leukocyte segmentation algorithm, in this experiment, we used a data set of CDetector [1] containing a total of 7410 cervical microscopic image slices of 11 classes, all of which were cropped from the slide microscopic image, wherein a total of 6666 cervical microscopic image slices were obtained from the training set and 744 cervical microscopic image slices were obtained from the test set. The slide micrographs in the data set were taken using a panoramic MIDI II digital slide scanner, and the corresponding specimens were prepared using the Thinprep method and stained using a Papanicolaou stain. The expert labels all the sections of the cervical microscopy images according to the TBS category and provides the corresponding bounding box and category names, for a total of 11 category names. Respectively, being an antigenic square cells Canot exact HSIL (ASCH), an antigenic square cells of undetermined design (ASCUS), a high-grade square intraepithelial version (HSIL), a low-grade square intraepithelial version (LSIL), a flo, a Trichlonas (TRICH), an Antigenic Great Cells (AGC), an Actinomyces (ACTIN), a Candida (CAND), a hers, a squamous-cell carbon (SCC). We selected only 4 types of squamous lesion cells (ASCUS, ASCH, HSIL, LSIL) in CDetector for training and analysis. The 4 types of pictures are 3502 in the training set and 505 in the test set, and the training set is divided into a training set and a verification set according to the proportion of 9. The algorithm of the invention is compared with the existing various detection methods, the result is shown in Table 2, and three Loss values of different methods are shown in FIG. 8; then, various algorithms were quantitatively compared on the CDetector dataset using the measure mAP of paslc VOCs.
The mAP is the average value of AP detected by various types of cells of a model. Where AP is the area under the curve for Precision and Recall Recall:
Figure BDA0003858275870000071
Figure BDA0003858275870000072
wherein r is 1 ,r 2 …r n Is the corresponding Recall value at the first interpolation of the Precison interpolation segment arranged in ascending order, and Precision and Recall are respectively:
Figure BDA0003858275870000073
Figure BDA0003858275870000074
wherein TP represents True Positive, represents the Positive sample predicted as Positive by the model, FP represents False Positive, represents the negative sample predicted as Positive by the model, FN represents False negative, represents the Positive sample predicted as negative by the model.
Higher values of mAP represent better detection. The experimental environment of the invention is Ubuntu 20.04.3 LTS operating system, NVIDIA GEFORCE RTX 3080Ti display card is adopted for operation, the display memory is 12GB and 128GB RAM, the CPU processor is Intel (R) Xeon (R) Silver4210 CPU @2.20GHz, and all models are based on python 3.8 operating environment, pytrch 1.11.0 deep learning frame, CUDA 11.6 and CUDNN 8.3 acceleration library.
2. Effect display
Fig. 8 shows sample 1 and the corresponding labels. FIG. 9 shows the results of the detection of the present invention on sample 1. Fig. 10 is sample 2 and the corresponding label. FIG. 11 shows the results of the test of the present invention on sample 2. Fig. 12 shows sample 3 and the corresponding label. FIG. 13 shows the results of the detection of the present invention on sample 3.
3. Qualitative comparison
In order to qualitatively compare the detection effect of our algorithm with that of other excellent baseline algorithms on 4 kinds of squamous cervical lesion cells, we performed comparative experiments, and respectively show the effect images of our model on the detection of ASCUS, ASCH, HSIL and LSIL samples, and the detection results are shown in fig. 8-13, which shows that our invention can effectively and accurately cut out the lesion cells of these kinds.
As can be seen from fig. 8, 10, 12, the characteristics of these 4 types of squamous lesion cells in the true microscopic smear are very similar, and it is almost difficult for a person without specialized training to distinguish the differences among them. However, the model can distinguish the lesion cells in the samples with excellent accuracy, and although individual misjudgment also appears, the model is also applicable as a cervical lesion cell detection system for providing pathology-assisted analysis for clinical medical detection at present.
As shown in Table 2, compared with the fast R-CNN and RetinaNet models with a large number of parameters, the models of other series of YOLO, and the models which are also YOLOv5 but have larger scale and more parameter quantity, the model parameters are smaller, the detection accuracy of the 4 types of squamous epithelial cells is higher, and the reasoning speed is Faster.
TABLE 2 quantitative comparison of the results of the present invention with other excellent models for detecting diseased cells
Figure BDA0003858275870000081
4. Quantitative comparison
To quantitatively compare the accuracy of detection of cervical lesions with five methods (i.e., fast R-CNN, retinaNet, YOLOv3, YOLOv5, and the algorithm of the present invention) and corresponding models of different parameter scales, we performed experiments on a data set consisting of 3150 cervical cell images stained with Papanicolaou stain and quantitatively evaluated the segmentation results with two measures, mAP.5 and mAP.5: 0.95. Table 2 shows the results of quantitative evaluation on the current data set for the squamous cell test results in 4, with the best measure in each column of data shown in bold. It can be easily seen that the algorithm of our invention has the best detection effect on cervical lesion cells with almost the same parameters and calculation amount. Compared with other models with larger-scale parameter quantity and calculated quantity, the method still has more excellent performance. Therefore, the algorithm of the present invention has the best detection effect on the four cervical lesion cells in combination.
The above are preferred embodiments of the present invention, and all changes made according to the technical scheme of the present invention that produce functional effects do not exceed the scope of the technical scheme of the present invention belong to the protection scope of the present invention.

Claims (6)

1. A cervical squamous lesion cell detection method based on depth adaptive feature extraction is characterized in that a Transformer module is used in a high-level feature layer of a trunk network on the basis of YOLOv 5; the neck network adopts a bidirectional feature pyramid Bi-FPN to adaptively extract features of different levels in the picture; the prediction head module uses the attention convolution block CBAM to adaptively learn the size of the final cell detection box.
2. The method for detecting the cervical squamous lesion cells based on the depth adaptive feature extraction as claimed in claim 1, wherein the main network uses CSP-Darknet53, and Transformer Block is used for replacing C3Block at the 9 th layer and the 11 th layer.
3. The cervical squamous lesion cell detection method based on depth adaptive feature extraction as claimed in claim 1, wherein the neck network replaces PAN with Bi-FPN designed by a weighted method; the Bi-FPN introduces an extra weight to each input of the network, deleting nodes with only one input edge and adding a jump connection between the input node and the output node of the same scale.
4. The cervical squamous lesion cell detection method based on depth adaptive feature extraction as claimed in claim 1, wherein the attention convolution block CBAM assists YOLOv5 to capture on the target object to be detected.
5. The cervical squamous lesion cell detection method based on depth adaptive feature extraction as claimed in claim 1, characterized in that the method adopts a mixed data enhancement mode to combine different samples, specifically: a) A plurality of pictures are fused into one picture by adopting Mosaic, so that the diversity of the sample is increased; b) And the MixUp is adopted to fuse the two pictures together according to the preset transparency, so that the training difficulty is increased, and the model has better semantic analysis capability.
6. The cervical squamous lesion cell detection method based on depth adaptive feature extraction as claimed in claim 1, wherein the loss function adopted by the method is composed of three parts, specifically:
1) Object Score loss
When detecting the object, two-class cross entropy loss which can judge whether the object in the target frame is in the frame is used, wherein o is a label vector, t is a prediction vector:
Figure FDA0003858275860000011
2) Class Probasic Score loss
In detecting the class, as well as the Object Score loss, two classes of cross entropy loss are used to determine whether the target frame is the correct class in the frame, where o is the label vector, t is the prediction vector:
Figure FDA0003858275860000012
3) Bounding Box loss
When detecting whether the target object can be compactly framed in the prediction frame, the CIOU Loss is used:
Figure FDA0003858275860000021
wherein α is:
Figure FDA0003858275860000022
v is:
Figure FDA0003858275860000023
wherein D is 2 To predict the distance between the frame and the center point of the target frame,D c is the diagonal distance of the minimum circumscribed rectangle; w and h are the height and width of the prediction box, w gt And h gt Is the height and width of the real box.
CN202211161135.7A 2022-09-22 2022-09-22 Cervical squamous lesion cell detection method based on depth self-adaptive feature extraction Pending CN115471838A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202211161135.7A CN115471838A (en) 2022-09-22 2022-09-22 Cervical squamous lesion cell detection method based on depth self-adaptive feature extraction

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202211161135.7A CN115471838A (en) 2022-09-22 2022-09-22 Cervical squamous lesion cell detection method based on depth self-adaptive feature extraction

Publications (1)

Publication Number Publication Date
CN115471838A true CN115471838A (en) 2022-12-13

Family

ID=84334303

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202211161135.7A Pending CN115471838A (en) 2022-09-22 2022-09-22 Cervical squamous lesion cell detection method based on depth self-adaptive feature extraction

Country Status (1)

Country Link
CN (1) CN115471838A (en)

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN116704317A (en) * 2023-08-09 2023-09-05 深圳华付技术股份有限公司 Target detection method, storage medium and computer device
CN117152746A (en) * 2023-10-27 2023-12-01 南方医科大学 Method for acquiring cervical cell classification parameters based on YOLOV5 network

Cited By (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN116704317A (en) * 2023-08-09 2023-09-05 深圳华付技术股份有限公司 Target detection method, storage medium and computer device
CN116704317B (en) * 2023-08-09 2024-04-19 深圳华付技术股份有限公司 Target detection method, storage medium and computer device
CN117152746A (en) * 2023-10-27 2023-12-01 南方医科大学 Method for acquiring cervical cell classification parameters based on YOLOV5 network
CN117152746B (en) * 2023-10-27 2024-03-26 南方医科大学 Method for acquiring cervical cell classification parameters based on YOLOV5 network

Similar Documents

Publication Publication Date Title
US5987158A (en) Apparatus for automated identification of thick cell groupings on a biological specimen
Ramesh et al. Isolation and two-step classification of normal white blood cells in peripheral blood smears
CN110765855B (en) Pathological image processing method and system
CN115471838A (en) Cervical squamous lesion cell detection method based on depth self-adaptive feature extraction
CN111062296B (en) Automatic white blood cell identification and classification method based on computer
EP4290451A1 (en) Deep neural network-based method for detecting living cell morphology, and related product
CN114945941A (en) Non-tumor segmentation for supporting tumor detection and analysis
CN112365471B (en) Cervical cancer cell intelligent detection method based on deep learning
Mao et al. A deep convolutional neural network trained on representative samples for circulating tumor cell detection
CN116580394A (en) White blood cell detection method based on multi-scale fusion and deformable self-attention
CN112784767A (en) Cell example segmentation algorithm based on leukocyte microscopic image
CN111062346A (en) Automatic leukocyte positioning detection and classification recognition system and method
Rachmad et al. Mycobacterium tuberculosis images classification based on combining of convolutional neural network and support vector machine
Du et al. Automatic classification of cells in microscopic fecal images using convolutional neural networks
Zhang et al. Research on application of classification model based on stack generalization in staging of cervical tissue pathological images
CN114387596A (en) Automatic interpretation system for cytopathology smear
Mustafa et al. Malaria parasite diagnosis using computational techniques: a comprehensive review
CN115775226B (en) Medical image classification method based on transducer
Kim et al. Nucleus segmentation and recognition of uterine cervical pap-smears
Angulo et al. Microscopic image analysis using mathematical morphology: Application to haematological cytology
CN114898866A (en) Thyroid cell auxiliary diagnosis method, equipment and storage medium
CN114037868A (en) Image recognition model generation method and device
Song et al. Automatic vaginal bacteria segmentation and classification based on superpixel and deep learning
Boonsiri et al. 3D gray level co-occurrence matrix based classification of favor benign and borderline types in follicular neoplasm images
Ma et al. CHS-NET: A cascaded neural network with semi-focal loss for mitosis detection

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination