CN116386034A - Cervical cell classification method based on multiscale attention feature enhancement - Google Patents

Cervical cell classification method based on multiscale attention feature enhancement Download PDF

Info

Publication number
CN116386034A
CN116386034A CN202310131751.6A CN202310131751A CN116386034A CN 116386034 A CN116386034 A CN 116386034A CN 202310131751 A CN202310131751 A CN 202310131751A CN 116386034 A CN116386034 A CN 116386034A
Authority
CN
China
Prior art keywords
attention
pyramid
scale
features
feature
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN202310131751.6A
Other languages
Chinese (zh)
Inventor
刘娟
金钰
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Wuhan University WHU
Original Assignee
Wuhan University WHU
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Wuhan University WHU filed Critical Wuhan University WHU
Priority to CN202310131751.6A priority Critical patent/CN116386034A/en
Publication of CN116386034A publication Critical patent/CN116386034A/en
Pending legal-status Critical Current

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V20/00Scenes; Scene-specific elements
    • G06V20/60Type of objects
    • G06V20/69Microscopic objects, e.g. biological cells or cellular parts
    • G06V20/698Matching; Classification
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/08Learning methods
    • G06N3/084Backpropagation, e.g. using gradient descent
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/20Image preprocessing
    • G06V10/26Segmentation of patterns in the image field; Cutting or merging of image elements to establish the pattern region, e.g. clustering-based techniques; Detection of occlusion
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/40Extraction of image or video features
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/70Arrangements for image or video recognition or understanding using pattern recognition or machine learning
    • G06V10/764Arrangements for image or video recognition or understanding using pattern recognition or machine learning using classification, e.g. of video objects
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/70Arrangements for image or video recognition or understanding using pattern recognition or machine learning
    • G06V10/77Processing image or video features in feature spaces; using data integration or data reduction, e.g. principal component analysis [PCA] or independent component analysis [ICA] or self-organising maps [SOM]; Blind source separation
    • G06V10/80Fusion, i.e. combining data from various sources at the sensor level, preprocessing level, feature extraction level or classification level
    • G06V10/806Fusion, i.e. combining data from various sources at the sensor level, preprocessing level, feature extraction level or classification level of extracted features
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/70Arrangements for image or video recognition or understanding using pattern recognition or machine learning
    • G06V10/82Arrangements for image or video recognition or understanding using pattern recognition or machine learning using neural networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V2201/00Indexing scheme relating to image or video recognition or understanding
    • G06V2201/03Recognition of patterns in medical or anatomical images

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Multimedia (AREA)
  • Health & Medical Sciences (AREA)
  • General Health & Medical Sciences (AREA)
  • Evolutionary Computation (AREA)
  • Computing Systems (AREA)
  • Software Systems (AREA)
  • Artificial Intelligence (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Medical Informatics (AREA)
  • Databases & Information Systems (AREA)
  • Molecular Biology (AREA)
  • Biomedical Technology (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Biophysics (AREA)
  • Computational Linguistics (AREA)
  • Data Mining & Analysis (AREA)
  • General Engineering & Computer Science (AREA)
  • Mathematical Physics (AREA)
  • Image Analysis (AREA)

Abstract

The invention discloses a cervical cell classification method based on multi-scale attention feature enhancement, which comprises the following steps: s1: extracting multi-scale features by using a deep convolutional neural network; s2: constructing a multi-scale feature pyramid according to a feature fusion method in the FPN; s3: respectively calculating spatial attention and channel attention for each layer of features in pyramid features to generate a spatial attention pyramid and a channel attention vector; s4: the masks obtained by segmentation of the spatial attention pyramid threshold are used for enhancing the spatial attention; s5: the attention weighting is carried out on the multi-scale pyramid features in the S2 by using the enhanced spatial attention pyramid and the attention vectors of the channels of each layer, so as to obtain a multi-scale attention feature pyramid; s6: respectively constructing a classifier for each layer of features in the multi-scale attention feature pyramid; s7: and (3) performing gradient descent optimization training on the whole network and performing classified prediction on cervical cells. The invention provides a cervical cell classification model with higher classification accuracy.

Description

Cervical cell classification method based on multiscale attention feature enhancement
Technical Field
The invention relates to the field of deep learning and medical intersection, in particular to a cervical cell classification method based on multi-scale attention feature enhancement, and belongs to the application of deep learning in the medical field.
Background
Cervical cancer is one of the most common gynaecological malignancies, with extremely high morbidity and mortality. The advanced age of carcinoma in situ is 30-35 years, the invasive carcinoma is 45-55 years, and the onset of carcinoma in situ has a tendency to be younger in recent years. Clinical practice shows that the effective cervical cell screening means can lead cervical cancer and precancerous lesions to be found and treated earlier, and obviously reduces the incidence and death rate of cervical cancer.
Cervical cells can be classified into the following 5 classes according to the different stages of cervical cancer precancerous lesions: NORMAL, ASC-US (atypical squamous cells of undetermined significance, atypical squamous cell of undefined diagnostic significance), ASC-H (atypical squamous cells but cannot rule out high-grade squamous intra-epithelial lesion, atypical squamous cell of high squamous intraepithelial lesions cannot be excluded), LSIL (low-grade squamous intra-epithelial lesion, low squamous intraepithelial lesions), HSIL (high-grade squamous intra-epithelial lesion, high squamous intraepithelial lesions). These 5 cell types are generally similar in appearance, but as the course of the disease progresses, they differ primarily in that their pathological characteristics change to varying degrees, for example, in abnormal cells there is often uncontrolled division leading to the appearance of a heterogeneous nucleus. Thus, in clinical diagnosis, cervical cell screening is often performed by a cytologist observing the pathological features of the nuclei in cervical cell smears under a microscope to determine the extent of cytopathy. However, manual cervical cytological screening is highly specialized and experienced by cytologists and is time-consuming and labor-consuming. With the rapid development of machine learning methods, particularly deep learning, more and more researchers are beginning to use computer-aided diagnosis (CAD) technology to make up for the shortfall of manual cytology screening methods.
Many automatic cervical cell classification tasks have been presented, however, these methods have not directly solved two major difficulties faced by cervical cell classification. First, the differentiation between the different classes of cervical cells is mainly manifested on the nucleus, so the classification network needs to be more focused on the nucleus area. Second, cervical cell classification is a fine-grained image classification task, and the similarity between different classes of cells is large, and due to inconsistent staining conditions or differences in personal preferences of cervical smear producers, there may be large differences in cells of the same class, which makes the cervical cell classification task more challenging.
Disclosure of Invention
Aiming at the technical problems, the invention discloses a cervical cell classification method based on multi-scale attention feature enhancement, which is characterized in that a multi-scale feature pyramid with bidirectional feature transmission is built for extracting a multi-scale feature pyramid simultaneously comprising low-level detail features and high-level semantic features, a plurality of joint classifiers are built by using the richer features, the discrimination features among various cervical cells with fine granularity can be more effectively excavated, and then the attention of multiple scales is enhanced by combining with the attention masks generated by threshold segmentation, so that more accurate cervical cell classification is finally realized.
The technical scheme provided by the invention is as follows:
a cervical cell classification method based on multi-scale attention feature enhancement comprises the following steps:
s1: extracting multi-scale features from the cervical cell image using a deep convolutional neural network comprising a plurality of blocks;
s2: constructing a multi-scale feature pyramid according to a feature fusion method in the FPN;
s3: calculating spatial attention and channel attention for each layer of features in the multi-scale pyramid features by using APN, and generating a spatial attention pyramid and channel attention vectors of each layer;
s4: the masks obtained by segmentation of the spatial attention pyramid threshold are used for enhancing the spatial attention;
s5: the enhanced spatial attention pyramid and the attention vectors of the channels of each layer are used for carrying out attention weighting on the multi-scale pyramid features in the S2, so that a multi-scale attention feature pyramid for classification is obtained;
s6: respectively constructing a classifier for each layer of features in the multi-scale attention feature pyramid;
s7: and performing end-to-end gradient descent optimization training on the whole network by using three loss functions, and performing classification prediction on cervical cells by using a trained classification network.
Further, the deep convolutional neural network in step S1 is a backhaul of any general classification network, which is divided into n blocks, and the output feature map of each block has a different scale. Preferred classification networks include ResNet, denseNet, VGG, etc.
Further, in step S2, after sampling the feature map output by the deep block to the same size as the feature map output by the shallow block, adding the feature map to the feature map output by the shallow block to realize fusion of features of different scales from top to bottom, and constructing a multi-scale feature pyramid by using the features fused by each block output feature and the deep feature.
Further, step S3 comprises the sub-steps of:
s3.1: for each layer of features P in a multi-scale pyramid i Performing 3×3 deconvolution and sigmoid normalization to generate a spatial attention map corresponding to each layer of features
Figure SMS_1
S3.2: for each layer of features F in a multi-scale pyramid i Generating initial channel attention after global average pooling and two full connection layers in turn
Figure SMS_2
S3.3: initial channel attention using current layer
Figure SMS_3
Channel attention to the shallow layer>
Figure SMS_4
Adding and fusing to obtain final channel attention of the current layer>
Figure SMS_5
And the attention characteristic fusion of different scales from bottom to top is realized.
Further, step S4 comprises the sub-steps of:
s4.1: computing per-layer spatial attention patterns
Figure SMS_6
Mean value theta of (2) i
S4.2: using theta i Spatial attention as a threshold
Figure SMS_7
Dividing to obtain->
Figure SMS_8
Corresponding segmentation mask M i
S4.3: by M i Space attention force diagram
Figure SMS_9
And weighting is carried out, and attention scores at points with smaller attention weights are removed, so that the enhanced multi-scale space attention characteristics are obtained.
Further, in step S4.2, θ is used i Spatial attention as a threshold
Figure SMS_10
The method for dividing comprises the following steps: />
Figure SMS_11
Is greater than or equal to theta i The value of (2) is assigned 1, (-)>
Figure SMS_12
Values less than are assigned as0。
Further, in step S5, each value in the channel attention vector corresponds to one channel in the current layer feature map, and each attention value is weighted to all feature values on its corresponding channel by multiplication.
Further, the classifier in step S6 is composed of one global average pooling layer and several fully connected layers.
Further, in step S7, the three-term loss function is obtained by the joint cross entropy loss L of all classifiers CE Smoothing loss L of masks SM Loss of channel feature diversity L CDis Is a weighted sum composition of (i.e., l=l) CE +λL CDis +μL SM Where λ and μ are constants used to balance the losses.
Further, in step S7, the smoothing loss L SM Boundary smoothing for constraint-type saturation mask, which is the scale-type saturation mask M i The first-order gradient sum has the formula:
Figure SMS_13
wherein n is the total number of the attention masks, W i And H i The length and width of the ith saturation mask, m and n represent the coordinate index of a point in the saturation mask;
Figure SMS_14
representing the gradient at the (m, n) point coordinate point in the ith saturation mask, a gradient operator is used in the present invention to determine.
Further, the channel feature diversity loss in step S7 is lost L CDis The feature channel used for forcing the last convolution layer output in the classifier has the discrimination to the specific class, thereby increasing the diversity of the features of each channel, and the specific calculation steps comprise:
(1) Dividing each layer of classification features into C groups according to channels, wherein each group comprises n=c/C channel features, C is the channel number of the classification features of the current layer, and C is the cervical cell class number;
(2) Aggregating n channel features in each group into single channel features by using cross-channel averaging pooling, calculating a class response value for each group of features by using global averaging pooling, and normalizing all c response values by using a softmax function;
(3) And calculating the channel diversity loss of the current layer characteristic of the cross entropy loss value of the normalized vector formed by c response values and the current image category single thermal coding, and then calculating the average value of the channel diversity loss of each layer as the channel diversity loss of the current network.
Further, in step S7, the result of the classification prediction is the average value of all the classifier prediction results.
Further, in step S7, cervical cells are classified into 5 classes, including NORMAL, ASC-US, ASC-H, LSIL, and HSIL.
The beneficial effects of the invention are as follows:
(1) According to the invention, the multi-scale feature pyramid with the bidirectional feature transfer is constructed to extract the multi-scale feature pyramid simultaneously containing the low-level detail features and the high-level semantic features, and the identification features among various cervical cells with fine granularity can be more effectively excavated by constructing a plurality of joint classifiers by using the richer features;
(2) The attention weight is corrected based on the characteristic enhancement method of the attention mask, so that attention to a key area is further promoted, and more attention is added to a nucleus area. In the implementation flow, firstly, multiscale features and multiscale attention of cervical cell images are extracted through FPN (Feature Pyramid Network) and APN (Attention Pyramid Network), then, the multiscale attention is enhanced by using attention masks generated by threshold segmentation, and finally, a classifier is constructed by using features generated by weighting the attention of the multiscale features by the enhanced attention, so that the accuracy of cervical cell classification is improved from the aspect of improving the interpretability.
Drawings
In order to more clearly illustrate the embodiments of the present invention or the technical solutions in the prior art, the drawings used in the embodiments or the description of the prior art will be briefly described, and it is obvious that the drawings in the following description are some embodiments of the present invention, and other drawings may be obtained according to these drawings without inventive effort for a person skilled in the art.
FIG. 1 is an exemplary view of various cervical cell images used in accordance with an embodiment of the invention;
FIG. 2 is a general flow chart of an embodiment of the present invention;
FIG. 3 is a detailed view of the implementation of the modules in an embodiment of the invention.
Detailed Description
For the purpose of making the objects, technical solutions and advantages of the embodiments of the present invention more apparent, the technical solutions of the embodiments of the present invention will be clearly and completely described below with reference to the accompanying drawings in the embodiments of the present invention, and it is apparent that the described embodiments are some embodiments of the present invention, but not all embodiments of the present invention. All other embodiments, which can be made by those skilled in the art based on the embodiments of the invention without making any inventive effort, are intended to be within the scope of the invention.
Examples
The embodiment of the invention provides a cervical cell classification method (based on ResNet 50) based on multi-scale attention feature enhancement, and referring to a general flow structure of FIG. 2, the method comprises the following steps:
s1: backbone using ResNet50 as the multi-scale feature extractor, using the output feature map { B > of 4 blocks therein 1 ,B 2 ,B 3 ,B 4 -forming a multi-scale feature;
s2: multi-scale feature { B } is processed according to feature fusion method in FPN 1 ,B 2 ,B 3 ,B 4 Performing feature fusion from top to bottom to obtain a multi-scale feature pyramid { P } 2 ,P 3 ,P 4 Specific implementation details are shown in FF of FIG. 3 i Module, note P 4 Directly from B 4 The 1×1 convolution results in that P1 is not taken into consideration because the large feature map size requires significant computational resources and thus can be selectively implementedConstructing a feature pyramid by using the features of the partial scale;
s3: calculating spatial attention and channel attention for each layer of features in the multi-scale pyramid feature by using APN, generating spatial attention pyramid and channel attention vector of each layer, and specific implementation details are shown in CA in FIG. 3 i &SA i A module comprising the steps of:
s3.1: for each layer of features P in a multi-scale pyramid i Performing 3×3 deconvolution and sigmoid normalization to generate a spatial attention map corresponding to each layer of features
Figure SMS_15
S3.2: for each layer of features F in a multi-scale pyramid i Generating initial channel attention after global average pooling and two full connection layers in turn
Figure SMS_16
S3.3: initial channel attention using current layer
Figure SMS_17
Channel attention to the shallow layer>
Figure SMS_18
Adding and fusing to obtain final channel attention of the current layer>
Figure SMS_19
Realizing the fusion of attention features with different scales from bottom to top, the final channel attention of the shallowest layer features is the initial channel attention, namely +.>
Figure SMS_20
S4: using masks obtained by threshold segmentation of the spatial attention pyramid to enhance spatial attention, see Mask in FIG. 3 for specific implementation details i A module comprising the steps of:
s4.1: computing per-layer spatial attention patterns
Figure SMS_21
Mean value theta of (2) i
S4.2: using theta i Spatial attention as a threshold
Figure SMS_22
Dividing, i.e.)>
Figure SMS_23
Is greater than or equal to theta i The value of (2) is assigned 1, i.e. +.>
Figure SMS_24
A value less than or equal to 0, giving +.>
Figure SMS_25
Corresponding segmentation mask M i In the implementation process, M can be conveniently updated in order to facilitate back propagation i M can be made by a sigmoid function i The specific calculation formula is as follows:
Figure SMS_26
wherein ω is set to a very large constant to ensure M i The elements in (2) are close to 1 or 0; normally ω=10 can be set 12
S4.3: spatial attention seeking Using Mi
Figure SMS_27
Weighting, i.e. by->
Figure SMS_28
Eliminating attention scores at points with smaller attention weights to obtain enhanced multiscale spatial attention features +.>
Figure SMS_29
S5: using the enhanced spatial attention pyramid and the attention vectors of each layer of channels to perform multi-scale pyramid characteristics in S2Attention weighting, resulting in a multi-scale attention feature pyramid { P 'for classification' 2 ,P′ 3 ,P′ 4 Specific implementation details are shown in FIG. 3 FW i A module, expressed in terms of a formula, may be:
Figure SMS_30
wherein->
Figure SMS_31
Is a broadcast addition, ->
Figure SMS_32
Is bit-wise multiplication;
s6: respectively constructing classifiers for each layer of features in the multi-scale attention feature pyramid, wherein each classifier consists of a global average pooling layer and two full-connection layers;
s7: dividing the cervical cell image dataset with five types of cervical cells shown in fig. 1 into a training set, a verification set and a test set according to the ratio of 7:1:2, performing gradient descent optimization training on the whole network by using three loss functions, and performing classification prediction on the cervical cells on the test set by using the network with the highest classification accuracy on the verification set after training for 200 rounds, wherein the final prediction result is the average value of all classifier prediction results. Wherein the three-term loss function is obtained by the joint cross entropy loss L of all classifiers CE Smoothing loss L of masks SM Loss of channel feature diversity L CDis Is a weighted sum composition of (i.e., l=l) CE +λL CDis +μL SM Where λ and μ are constants used to balance the losses.
In the specific implementation process, the smoothing loss L SM Boundary smoothing for constraint-type saturation mask, which is the scale-type saturation mask M i The first-order gradient sum has the formula:
Figure SMS_33
wherein n is the total number of the attention masks, W i And H i Respectively the ith saturation masThe length and width of k, m and n represent the coordinate index of a point in the attribute mask;
Figure SMS_34
representing the gradient at the (m, n) point coordinate point in the ith saturation mask, a gradient operator is used in the present invention to determine. Channel feature diversity loss L CDis The feature channel used to force the last convolutional layer output in the classifier has discrimination to specific class, thus increasing the diversity of each channel feature, the calculation process includes:
firstly, classifying each layer of classification features into 5 groups (cervical cell class number) according to channels, wherein each group comprises n=C/5 channel features, and C is the channel number of the classification features of the current layer;
then, aggregating n channel features in each group into single channel features by using cross-channel average pooling, calculating a class response value for each group of features by using global average pooling, and normalizing all 5 response values by using a softmax function;
and finally, calculating the channel diversity loss of the current layer characteristics of the vector formed by the normalized 5 response values and the cross entropy loss value of the current image category single thermal coding, and then calculating the average value of the channel diversity losses of each layer as the channel diversity loss of the current network.
Today, where cervical cancer patients tend to be younger, efficient and automated cervical cell classification techniques are critical to reducing the incidence and mortality of cervical cancer. Aiming at two difficulties faced by the automatic classification task of cervical cells at present, the invention provides a cervical cell classification method based on multi-scale attention feature enhancement, which uses the attention feature enhanced by the attention mask for classifying cervical cells, realizes the purpose of simultaneously utilizing the high-level semantic features and the more detailed low-level features of the general classification network, and uses the attention mask to prompt the network to pay more attention to the nucleus region which is critical to classifying cervical cells, thereby further improving the accuracy of classifying cervical cells from the aspect of improving the interpretability.
The above embodiments are only for illustrating the technical solution of the present invention, and are not limiting; although the invention has been described in detail with reference to the foregoing embodiments, it will be understood by those of ordinary skill in the art that: the technical scheme described in the foregoing embodiments can be modified or some technical features thereof can be replaced by equivalents; such modifications and substitutions do not depart from the spirit and scope of the technical solutions of the embodiments of the present invention.

Claims (10)

1. A cervical cell classification method based on multi-scale attention feature enhancement is characterized by comprising the following steps:
s1: extracting multi-scale features from the cervical cell image using a deep convolutional neural network comprising a plurality of blocks;
s2: constructing a multi-scale feature pyramid according to a feature fusion method in the FPN;
s3: calculating spatial attention and channel attention for each layer of features in the multi-scale pyramid features by using APN, and generating a spatial attention pyramid and channel attention vectors of each layer;
s4: the masks obtained by segmentation of the spatial attention pyramid threshold are used for enhancing the spatial attention;
s5: the enhanced spatial attention pyramid and the attention vectors of the channels of each layer are used for carrying out attention weighting on the multi-scale pyramid features in the S2, so that a multi-scale attention feature pyramid for classification is obtained;
s6: respectively constructing a classifier for each layer of features in the multi-scale attention feature pyramid;
s7: and performing end-to-end gradient descent optimization training on the whole network by using three loss functions, and performing classification prediction on cervical cells by using a trained classification network.
2. The method according to claim 1, characterized in that: the deep convolutional neural network in step S1 is a backbone of any general classification network, which is divided into n blocks, the output feature map of each block having a different scale.
3. The method according to claim 1, characterized in that: in step S2, the feature map output by the deep block is sampled to be the same size as the feature map output by the shallow block, and then added with the feature map output by the shallow block to realize the fusion of features with different scales from top to bottom, and the features fused by each block output feature and the deep feature are used for constructing a multi-scale feature pyramid.
4. The method according to claim 1, characterized in that: step S3 comprises the following sub-steps:
s3.1: for each layer of features P in a multi-scale pyramid i Performing 3×3 deconvolution and sigmoid normalization to generate a spatial attention map corresponding to each layer of features
Figure FDA0004084285740000011
S3.2: for each layer of features F in a multi-scale pyramid i Generating initial channel attention after global average pooling and two full connection layers in turn
Figure FDA0004084285740000012
S3.3: initial channel attention using current layer
Figure FDA0004084285740000013
Channel attention to the shallow layer>
Figure FDA0004084285740000014
Adding and fusing to obtain final channel attention of the current layer>
Figure FDA0004084285740000015
And the attention characteristic fusion of different scales from bottom to top is realized.
5. The method according to claim 1, characterized in that: step S4 comprises the following sub-steps:
s4.1: computing per-layer spatial attention patterns
Figure FDA0004084285740000016
Mean value theta of (2) i
S4.2: using theta i Spatial attention as a threshold
Figure FDA0004084285740000017
Dividing to obtain->
Figure FDA0004084285740000018
Corresponding split mask M i
S4.3: by M i Space attention force diagram
Figure FDA0004084285740000019
And weighting is carried out, and attention scores at points with smaller attention weights are removed, so that the enhanced multi-scale space attention characteristics are obtained.
6. The method according to claim 5, wherein: in step S4.2, θ is used i Spatial attention as a threshold
Figure FDA0004084285740000021
The method for dividing comprises the following steps: />
Figure FDA0004084285740000022
Is greater than or equal to theta i The value of (2) is assigned 1, (-)>
Figure FDA0004084285740000023
Values less than are assigned 0.
7. The method according to claim 1, characterized in that: in step S5, each value in the channel attention vector corresponds to a channel in the current layer feature map, and each attention value is weighted by multiplication to all feature values on its corresponding channel.
8. The method according to claim 1, characterized in that: the classifier in step S6 is composed of a global average pooling layer and a number of fully connected layers.
9. The method according to claim 1, characterized in that: in step S7, the three-term loss function is obtained by the joint cross entropy loss L of all classifiers CE Smoothing loss L of masks SM Loss of channel feature diversity L CDis Is a weighted sum composition of (i.e., l=l) CE +λL CDis +μL SM Where λ and μ are constants used to balance the losses.
10. The method according to claim 1, characterized in that: in step S7, the result of the classification prediction is the average value of all the classifier prediction results.
CN202310131751.6A 2023-02-16 2023-02-16 Cervical cell classification method based on multiscale attention feature enhancement Pending CN116386034A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202310131751.6A CN116386034A (en) 2023-02-16 2023-02-16 Cervical cell classification method based on multiscale attention feature enhancement

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202310131751.6A CN116386034A (en) 2023-02-16 2023-02-16 Cervical cell classification method based on multiscale attention feature enhancement

Publications (1)

Publication Number Publication Date
CN116386034A true CN116386034A (en) 2023-07-04

Family

ID=86975812

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202310131751.6A Pending CN116386034A (en) 2023-02-16 2023-02-16 Cervical cell classification method based on multiscale attention feature enhancement

Country Status (1)

Country Link
CN (1) CN116386034A (en)

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN117671357A (en) * 2023-12-01 2024-03-08 广东技术师范大学 Pyramid algorithm-based prostate cancer ultrasonic video classification method and system

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN117671357A (en) * 2023-12-01 2024-03-08 广东技术师范大学 Pyramid algorithm-based prostate cancer ultrasonic video classification method and system
CN117671357B (en) * 2023-12-01 2024-07-05 广东技术师范大学 Pyramid algorithm-based prostate cancer ultrasonic video classification method and system

Similar Documents

Publication Publication Date Title
Li et al. Graph-feature-enhanced selective assignment network for hyperspectral and multispectral data classification
CN112288706B (en) Automatic chromosome karyotype analysis and abnormality detection method
Wang et al. Adaptive pruning of transfer learned deep convolutional neural network for classification of cervical pap smear images
Kainz et al. Semantic segmentation of colon glands with deep convolutional neural networks and total variation segmentation
CN111723675A (en) Remote sensing image scene classification method based on multiple similarity measurement deep learning
Pal et al. Histopathological image classification using enhanced bag-of-feature with spiral biogeography-based optimization
Liang et al. Comparison detector for cervical cell/clumps detection in the limited data scenario
Wang et al. Weakly supervised learning for whole slide lung cancer image classification
CN110188763B (en) Image significance detection method based on improved graph model
CN110633708A (en) Deep network significance detection method based on global model and local optimization
CN109102498B (en) Method for segmenting cluster type cell nucleus in cervical smear image
CN112163599B (en) Image classification method based on multi-scale and multi-level fusion
CN113032613B (en) Three-dimensional model retrieval method based on interactive attention convolution neural network
CN111524140B (en) Medical image semantic segmentation method based on CNN and random forest method
CN108765374B (en) Method for screening abnormal nuclear area in cervical smear image
CN114510594A (en) Traditional pattern subgraph retrieval method based on self-attention mechanism
CN111767952A (en) Interpretable classification method for benign and malignant pulmonary nodules
CN116386034A (en) Cervical cell classification method based on multiscale attention feature enhancement
Huang et al. Segmentation of cervical cell images based on generative adversarial networks
Zhang et al. Research on application of classification model based on stack generalization in staging of cervical tissue pathological images
CN116229112A (en) Twin network target tracking method based on multiple attentives
CN116525075A (en) Thyroid nodule computer-aided diagnosis method and system based on few sample learning
CN111815582A (en) Two-dimensional code area detection method for improving background prior and foreground prior
Wu et al. A new multiple‐distribution GAN model to solve complexity in end‐to‐end chromosome karyotyping
CN116311387B (en) Cross-modal pedestrian re-identification method based on feature intersection

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination