CN113469119B - Cervical cell image classification method based on visual converter and image convolution network - Google Patents

Cervical cell image classification method based on visual converter and image convolution network Download PDF

Info

Publication number
CN113469119B
CN113469119B CN202110820463.2A CN202110820463A CN113469119B CN 113469119 B CN113469119 B CN 113469119B CN 202110820463 A CN202110820463 A CN 202110820463A CN 113469119 B CN113469119 B CN 113469119B
Authority
CN
China
Prior art keywords
image
formula
cell image
layer
cervical
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN202110820463.2A
Other languages
Chinese (zh)
Other versions
CN113469119A (en
Inventor
史骏
唐昆铭
祝新宇
孙宇
李俊
郑钰山
姜志国
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Hefei University of Technology
Original Assignee
Hefei University of Technology
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Hefei University of Technology filed Critical Hefei University of Technology
Priority to CN202110820463.2A priority Critical patent/CN113469119B/en
Publication of CN113469119A publication Critical patent/CN113469119A/en
Application granted granted Critical
Publication of CN113469119B publication Critical patent/CN113469119B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/21Design or setup of recognition systems or techniques; Extraction of features in feature space; Blind source separation
    • G06F18/214Generating training patterns; Bootstrap methods, e.g. bagging or boosting
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/24Classification techniques
    • G06F18/245Classification techniques relating to the decision surface
    • G06F18/2451Classification techniques relating to the decision surface linear, e.g. hyperplane
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/045Combinations of networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/08Learning methods

Abstract

The invention discloses a cervical cell image classification method based on a visual converter and a graph convolution network, which comprises the following steps: 1. acquiring a cervical liquid-based cell image dataset and preprocessing the cervical liquid-based cell image dataset; 2. for the visual converter branch, the input image is partitioned, then the position coding and the visual converter are used for feature coding, and the long-distance dependency relationship in the image is mined; 3. for the graph convolution network branches, performing superpixel segmentation on an input image, performing pooling processing on convolution neural network characteristics corresponding to each superpixel and using the pooling processing as nodes in a graph structure, modeling the graph structure inside cells by using the graph convolution network, and sensing the spatial topological relation inside the cells; 4. the features resulting from fusing the two branches form the final feature representation of the cervical cells and are used for classification. According to the invention, the characterization extraction and classification of cervical cells are completed by extracting the long-distance dependency relationship and the spatial topological structure information in the cervical fluid-based cell structure.

Description

Cervical cell image classification method based on visual converter and graph convolution network
Technical Field
The invention relates to the technical field of computer vision, in particular to a cervical cell image classification method based on a vision converter and a graph convolution neural network.
Background
Classification of cervical cells has been widely used in clinical work as a primary screening means for gynecological cervical lesions. In actual diagnosis, a pathologist needs to visually examine tens of thousands of cells under a microscope. Each pathologist needs to process a large number of patient specimens every day, so that the fatigue of film reading is often caused, and misdiagnosis sometimes happens. Therefore, there is a need for an efficient and quantitative cervical cell diagnosis method that reduces the burden of pathologists on reading and improves the accuracy of cervical cell identification. At present, algorithms for cervical cell image classification mainly include two types, namely a traditional feature classification algorithm and a supervised learning classification algorithm based on deep learning.
The traditional classification method generally utilizes manually designed features to model image features, and different classes of features are distinguished by training a classifier. Common manual visual features include low-level features such as colors, textures, edges and the like, and potential high-level semantic information of the image is difficult to describe. However, the characteristics of complex pathological image structure and variable types lead to that the manual feature extraction cannot perform better modeling on the image features, so that the classification effect is unsatisfactory.
In recent years, deep learning models have achieved significant effects in various fields of computer vision, and some researchers have applied convolutional neural networks to cervical cell classification tasks, such as residual error network (ResNet) and dense convolutional network (densnet), but they only use neural networks to extract global features of a single picture, and ignore local features of cervical cells, so that it is difficult to better solve the problems of large intra-cell intra-class differences and high inter-class similarities.
Disclosure of Invention
The invention provides a cervical cell image classification method based on a vision converter and a graph convolution network, aiming at solving the problem of difficult cervical cell image classification caused by complex and variable pathological image structure and rich characteristic information.
In order to achieve the purpose, the invention adopts the following technical scheme:
the invention relates to a cervical cell image classification method based on a visual converter and a graph convolution network, which is characterized by comprising the following steps:
step 1, marking the cell nucleus position and the cell nucleus type in the cervical digital slice image, and cutting according to the marked cell nucleus position to obtain each cell image and the corresponding type thereof, which are marked as
Figure BDA0003171814880000011
Wherein the content of the first and second substances,
Figure BDA0003171814880000021
representing the ith cell image, C, H and W represent the number of channels, height and width of the image, respectively, y i Represents the ith cell image X i A corresponding category; i =1,2, …, N; n represents the number of cell images;
step 2, constructing a visual converter ViT composed of L encoders as a first branch, wherein each encoder comprises: two normalization layers, a multi-head attention mechanism layer and a multi-layer sensor;
step 2.1, image X of ith cell i Performing block processing to obtain a sequence containing m image blocks
Figure BDA0003171814880000022
Wherein the content of the first and second substances,
Figure BDA0003171814880000023
represents the ith cell image X i The jth image block of (1);
Figure BDA0003171814880000024
p × p denotes the dimension of each image block, and m = H × W/p 2
Step 2.2, setting a learnable classification mark x class And m image blocks and a classification mark x are obtained by using the formula (1) class D-dimensional embedded representation z 0 And as input to the 1 st encoder;
Figure BDA0003171814880000025
in the formula (1), E pos Representing m image blocks and a class mark x class In the ith cell image X i (iii) a spatial position of; e denotes the set embedding matrix;
step 2.3, obtaining m image blocks and classification marks x by using the formula (2) class Output z of the multi-head attention mechanism layer at the I-th encoder l ′:
z l ′=MSA(LN(z l-1 ))+z l-1 ,l=1,…,L (2)
In the formula (2), LN () represents processing of the normalization layer; z is a radical of l-1 Represents the output of the l-1 st encoder;
step 2.4, obtaining the output z of the multi-layer perceptron of the first encoder by using the formula (3) l To obtain the output z of the last encoder L
z l =MLP(LN(z′ l ))+z′ l ,l=1,…,L (3)
Step 2.5, outputting z containing m + 1D dimensional characteristics L Obtaining the ith cell image X by averaging i ViT feature representation
Figure BDA0003171814880000026
Step 3, constructing a graph convolution network as a second branch;
step 3.1, image X of ith cell by SLIC algorithm i Performing over-segmentation to obtain the ith cell image X i A set of superpixels;
step 3.2, extracting the ith cell image X by using a pre-trained residual error network ResNet i CNN profile of (a);
3.3, processing the CNN characteristic in the CNN characteristic graph where each super pixel is located by utilizing a maximum pooling mode according to the position of the super pixel set on the CNN characteristic graph, thereby obtaining the CNN characteristic vector representation corresponding to each super pixel;
step 3.4, the CNN characteristic vector representation corresponding to each super pixel is respectively used as the nodes of the graph structure, thereby forming a node set
Figure BDA0003171814880000031
Wherein S represents the number of superpixels, and t is the dimension of the CNN feature corresponding to each superpixel;
step 3.5, constructing the adjacency matrix by using the formula (4)
Figure BDA0003171814880000032
Figure BDA0003171814880000033
In the formula (4), B α Representing the alpha-th super pixel, B β Denotes the beta-th super pixel, A αβ Representing the alpha-th super-pixel B in the adjacency matrix A α And the beta-th super pixel B β Spatial neighbor relationships between;
and 3.6, updating the relation perception representation of the nodes by using an equation (5) according to the hierarchical propagation rule of the graph convolution network:
Figure BDA0003171814880000034
in the formula (5), the reaction mixture is,
Figure BDA0003171814880000035
is a q-th level trainable weight, σ (-) is an activation function;
Figure BDA0003171814880000036
is a characteristic of a node of the q-th layer, d q Is the dimension of the characteristics of the nodes at the q-th layer;
Figure BDA0003171814880000037
represents the characteristics of the node of the q +1 th layer, and
Figure BDA0003171814880000038
Figure BDA0003171814880000039
the characteristic of the s-th node output in the q + 1-th layer is represented by letting H be when q =0 (0) = B; and comprises the following components:
Figure BDA00031718148800000310
in the formula (6), D represents a diagonal matrix, and
Figure BDA00031718148800000311
D ss representing the elements of the s-th row and s-th column of the diagonal matrix D, A us Elements representing the uth row and the s column in the adjacency matrix A; i represents an identity matrix;
step 3.7, obtaining the ith cell image X obtained by aggregating the node output characteristics of the last Q layer in the graph convolution network by using the formula (7) i Is a graph convolution feature of (a) represents z g
Figure BDA00031718148800000312
In equation (7), concat (. Cndot.) represents the feature vector splicing operation, S represents the total number of nodes,
Figure BDA00031718148800000313
features representing an s-th node in the Q-th layer;
step 4, obtaining the ith cell image X by using the formula (8) i Final characterization of (a):
Figure BDA0003171814880000041
and 5, performing linear transformation on the final characterization z by using the formula (9) to obtain an output result p of the linear classifier pred
p pred =Linear(z) (9)
In the formula (9), linear (·) represents a Linear classification function;
Figure BDA0003171814880000042
c represents the number of classes of cells;
step 6, constructing a cross entropy loss function L by using the formula (10), and training a network formed by two branches and a linear classifier by using a gradient descent algorithm to ensure that the cross entropy loss function L reaches convergence, thereby obtaining a trained cervical cell image classification model:
Figure BDA0003171814880000043
in formula (10), y label The cell image sample is a corresponding real category, and N is the total number of the cell image samples.
Compared with the prior art, the invention has the following advantages:
1. according to the method, the vision converter is used for learning the characteristics of the cervical cell image, and the vision converter can learn the long-distance dependency relationship in the image, so that the dependency relationship among all parts in the cervical cell structure is established, and the classification accuracy is improved;
2. the invention utilizes the graph convolution network to model the spatial topological relation of the cell structure, and the graph convolution neural network can well learn and fit the spatial topological structure inside the image, thereby modeling the whole cell image to obtain the representation and improving the characteristic representation capability of the cell image;
3. the invention integrates two groups of characteristics with the remote dependence relationship inside the cell and the spatial topological structure information to obtain a group of more robust cell image representation which is far superior to the global characteristics extracted by the traditional CNN network, thereby enhancing the identification capability of the characteristics;
4. the invention adopts a supervision training mode to train, and the semantic information is embedded into the characteristics of the image, thereby improving the classification accuracy.
Drawings
FIG. 1 is a block diagram of a network in accordance with the present invention;
FIG. 2 is a general flow diagram of the present invention;
fig. 3 is a diagram of a cervical cell image training sample according to the present invention.
Detailed Description
In this embodiment, a cervical cell image classification method based on a visual converter and a graph convolution network comprehensively considers the long-distance dependency relationship and the spatial topology structure inside a cell structure, and is to input an input image into a visual converter branch and a graph convolution network branch respectively to obtain ViT features and GCN features, fuse the two sets of features to obtain a final cervical cell image feature representation, and send the feature representation into a classifier for classification, thereby completing classification of a cervical cell image, as shown in fig. 2, the specific steps are as follows:
step 1, marking the nucleus position and the nucleus type in the cervical digital slice image, and cutting according to the marked nucleus position, thereby obtaining each cell image and the corresponding type thereof, and marking as
Figure BDA0003171814880000051
Wherein, the first and the second end of the pipe are connected with each other,
Figure BDA0003171814880000052
representing the ith cell image, C, H and W represent the number of channels, height and width of the image, respectively, y i Represents the ith cell image X i A corresponding category; i =1,2, …, N; n represents the number of cell images; the data used in this example contained 7 categories of normal superficial cells, normal intermediate basal cells, granulocytes, glandular cells, atypical squamous cells, hollowed-out cells and high-cytoplasmic ratio cells, as shown in fig. 3, 500 images per category, each cell image size being 224 × 224, so C =3, H =224 and W =224. 80% of each class in the dataset was used for training and the remaining 20% was used for testing;
and 2, establishing a deep learning network model based on the visual converter and the graph convolution network as shown in the figure 1, wherein the deep learning network comprises two branches of the visual converter ViT and the graph convolution network. First, a vision converter ViT made up of L encoders is constructed as a first branch, each encoder comprising: two normalization layers, a multi-head attention mechanism layer and a multi-layer sensor;
step 2.1, image X of ith cell i Performing block processing to obtain a sequence containing m image blocks
Figure BDA0003171814880000053
Wherein the content of the first and second substances,
Figure BDA0003171814880000054
represents the ith cell image X i The jth image block of (1);
Figure BDA0003171814880000055
p × p denotes the dimension of each image block, and m = H × W/p 2 (ii) a In the present embodiment, each image block size is 16 × 16, so p =16, and m =196.
Step 2.2, setting a learnable classification mark x class And m image blocks and a classification mark x are obtained by using the formula (1) class D-dimensional embedded representation z 0 And as input to the 1 st encoder;
Figure BDA0003171814880000056
in the formula (1), E pos Representing m image blocks and a class mark x class In the ith cell image X i (iii) a spatial position of; e denotes the set embedding matrix; in the present example, D =768,x class Is a 768-dimensional vector consisting of 768 random numbers, E is a matrix of 768 × 768 random numbers with 768 rows and 768 columns, and E is a function of the number of rows and columns of the matrix pos The random number matrix is a matrix formed by 197 x 768 random numbers, the number of rows of the matrix is 197, and the number of columns of the matrix is 768.
Step 2.3, obtaining m image blocks and classification marks x by using the formula (2) class Output z 'of multi-head attention device layer at l-th encoder' l
z′ l =MSA(LN(z l-1 ))+z l-1 ,l=1,…,L (2)
In the formula (2), LN () represents processing of the normalization layer; z is a radical of l-1 The output of the L-1 st encoder is shown, L =12 in the present embodiment.
Step 2.4, obtaining the output z of the multi-layer perceptron of the first encoder by using the formula (3) l To obtain the output z of the last encoder L
z l =MLP(LN(z′ l ))+z′ l ,l=1,…,L (3)
In this embodiment, the multi-layer perceptron comprises two layers of networks and one GELU non-linear active layer.
Step 2.5, outputting z containing m + 1D dimensional characteristics L Obtaining the ith cell image X by averaging i ViT feature representation of
Figure BDA0003171814880000063
Step 3, constructing a graph convolution network as a second branch;
step 3.1, image X of ith cell by SLIC algorithm i Performing over-segmentation to obtain the ith cell image X i A set of superpixels;
step 3.2, extracting the ith cell image X by using a pre-trained residual error network ResNet i CNN feature map of (1), in this exampleExtracting a feature map of the cell image by using ResNet 18;
3.3, processing the CNN characteristic in the CNN characteristic graph where each super pixel is located by utilizing a maximum pooling mode according to the position of the super pixel set on the CNN characteristic graph, thereby obtaining the CNN characteristic vector representation corresponding to each super pixel;
step 3.4, the CNN characteristic vector representation corresponding to each super pixel is respectively used as the nodes of the graph structure, thereby forming a node set
Figure BDA0003171814880000061
Wherein S represents the number of superpixels, and t is the dimension of the CNN feature corresponding to each superpixel, in this embodiment, the number S of superpixels is 16, and the dimension t of the CNN feature corresponding to each superpixel is 512;
step 3.5, constructing the adjacency matrix by using the formula (4)
Figure BDA0003171814880000062
Figure BDA0003171814880000071
In the formula (4), B α Representing the alpha-th super pixel, B β Denotes the beta-th super pixel, A αβ Representing the alpha-th super-pixel B in the adjacency matrix A α And the beta-th super pixel B β The spatial neighbor relation between them, in this embodiment, the spatial 4 neighborhood represents 4 spatial relations of up, down, left and right;
and 3.6, updating the relation perception representation of the nodes by using a formula (5) according to the hierarchical propagation rule of the graph convolution network:
Figure BDA0003171814880000072
in the formula (5), the reaction mixture is,
Figure BDA0003171814880000073
is a q-th level trainable weight, σ(. Is an activation function;
Figure BDA0003171814880000074
is a characteristic of a node of the q-th layer, d q Is the dimension of the characteristics of the nodes at the q-th layer;
Figure BDA0003171814880000075
represents the characteristics of the node of the q +1 th layer, and
Figure BDA0003171814880000076
Figure BDA0003171814880000077
the characteristic of the s-th node output in the q + 1-th layer is shown, when q =0, let H (0) = B; and comprises the following components:
Figure BDA0003171814880000078
in the formula (6), D represents a diagonal matrix, and
Figure BDA0003171814880000079
D ss representing the elements of the s-th row and s-th column of the diagonal matrix D, A us The element of the w-th row and s-th column in the adjacent matrix A is represented, and I represents an identity matrix; in the present embodiment, sigmoid function is used as the activation function σ ();
step 3.7, obtaining the ith cell image X obtained by aggregating the node output characteristics of the last Q layer in the graph convolution network by using the formula (7) i Is a graph convolution feature of (a) g
Figure BDA00031718148800000710
In equation (7), concat (. Cndot.) represents the feature vector splicing operation, S represents the total number of nodes,
Figure BDA00031718148800000711
features representing the s-th node in the Q-th layer, bookIn the embodiment, the number Q of layers of the graph convolution network is 2;
step 4, obtaining the ith cell image X by using the formula (8) i Final characterization of (a):
Figure BDA00031718148800000712
and 5, performing linear transformation on the final characterization z by using the formula (9) to obtain an output result p of the linear classifier pred
p pred =Linear(z) (9)
In formula (9), linear (·) represents a Linear classification function;
Figure BDA0003171814880000081
c represents the number of classes of cells;
step 6, constructing a cross entropy loss function L by using the formula (10), and training a network formed by two branches and a linear classifier by using a gradient descent algorithm to ensure that the cross entropy loss function L is converged, thereby obtaining a trained cervical cell image classification model:
Figure BDA0003171814880000082
in the formula (10), y label And N is the actual category corresponding to the cell image sample, and is the total number of the cell image samples.

Claims (1)

1. A cervical cell image classification method based on a visual converter and a graph convolution network is characterized by comprising the following steps:
step 1, marking the nucleus position and the nucleus type in the cervical digital slice image, and cutting according to the marked nucleus position, thereby obtaining each cell image and the corresponding type thereof, and marking as
Figure FDA0003171814870000011
Wherein, the first and the second end of the pipe are connected with each other,
Figure FDA0003171814870000012
representing the ith cell image, C, H and W represent the number of channels, height and width of the image, respectively, y i Represents the ith cell image X i A corresponding category; i =1,2, …, N; n represents the number of cell images;
step 2, constructing a visual converter ViT composed of L encoders as a first branch, wherein each encoder comprises: two normalization layers, a multi-head attention mechanism layer and a multi-layer sensor;
step 2.1, image X of ith cell i Performing block division to obtain a sequence containing m image blocks
Figure FDA0003171814870000013
Wherein the content of the first and second substances,
Figure FDA0003171814870000014
represents the ith cell image X i The jth image block of (1);
Figure FDA0003171814870000015
p × p denotes the dimension of each image block, and m = H × W/p 2
Step 2.2, setting a learnable classification mark x class And obtaining m image blocks and a classification mark x by using the formula (1) class D-dimensional embedded representation z 0 And as input to the 1 st encoder;
Figure FDA0003171814870000016
in the formula (1), E pos Representing m image blocks and a class mark x class In the ith cell image X i (ii) a spatial position of; e represents the set embedding matrix;
step 2.3, obtaining m image blocks and classification marks x by using the formula (2) class Multiple attention mechanism layer at the first encoderOutput z' l
z′ l =MSA(LN(z l-1 ))+z l-1 ,l=1,…,L (2)
In the formula (2), LN () represents processing of the normalization layer; z is a radical of l-1 Represents the output of the l-1 encoder;
step 2.4, obtaining the output z of the multilayer perceptron of the first encoder by using the formula (3) l To obtain the output z of the last encoder L
z l =MLP(LN(z′ l ))+z′ l ,l=1,…,L (3)
Step 2.5, outputting z containing m + 1D dimensional characteristics L Obtaining the ith cell image X by averaging i ViT feature representation
Figure FDA0003171814870000017
Step 3, constructing a graph convolution network as a second branch;
step 3.1, image X of ith cell by SLIC algorithm i Performing over-segmentation to obtain the ith cell image X i A set of superpixels;
step 3.2, extracting the ith cell image X by using a pre-trained residual error network ResNet i CNN profile of (a);
3.3, processing the CNN characteristic in the CNN characteristic graph where each super pixel is located by utilizing a maximum pooling mode according to the position of the super pixel set on the CNN characteristic graph, thereby obtaining the CNN characteristic vector representation corresponding to each super pixel;
step 3.4, the CNN characteristic vector representation corresponding to each super pixel is respectively used as the nodes of the graph structure, thereby forming a node set
Figure FDA0003171814870000021
Wherein S represents the number of superpixels, and t is the dimension of the CNN characteristic corresponding to each superpixel;
step 3.5, constructing the adjacency matrix by using the formula (4)
Figure FDA0003171814870000022
Figure FDA0003171814870000023
In the formula (4), B α Denotes the alpha th super pixel, B β Denotes the beta-th super pixel, A αβ Representing the alpha-th super-pixel B in the adjacency matrix A α And the beta-th super pixel B β Spatial proximity relation between them;
and 3.6, updating the relation perception representation of the nodes by using an equation (5) according to the hierarchical propagation rule of the graph convolution network:
Figure FDA0003171814870000024
in the formula (5), the reaction mixture is,
Figure FDA0003171814870000025
is a q-th level trainable weight, σ (-) is an activation function;
Figure FDA0003171814870000026
is a characteristic of a node of the q-th layer, d q Is the dimension of the characteristics of the nodes at the q-th layer;
Figure FDA0003171814870000027
represents the characteristics of the node of the q +1 th layer, and
Figure FDA0003171814870000028
Figure FDA0003171814870000029
the characteristic of the s-th node output in the q + 1-th layer is shown, when q =0, let H (0) = B; and comprises the following components:
Figure FDA00031718148700000210
in the formula (6), D represents a diagonal matrix, and
Figure FDA00031718148700000211
D ss representing the elements of the s-th row and s-th column of the diagonal matrix D, A us Elements representing the w-th row and s-th column in the adjacency matrix A; i represents an identity matrix;
step 3.7, obtaining the ith cell image X obtained by aggregating the node output characteristics of the last Q-th layer in the graph convolution network by using the formula (7) i Is a graph convolution feature of (a) g
Figure FDA0003171814870000031
In equation (7), concat (. Cndot.) represents the feature vector splicing operation, S represents the total number of nodes,
Figure FDA0003171814870000032
features representing an s-th node in the Q-th layer;
step 4, obtaining the ith cell image X by using the formula (8) i Final characterization of (a):
Figure FDA0003171814870000033
and 5, performing linear transformation on the final characterization z by using the formula (9) to obtain an output result p of the linear classifier pred
p pred =Linear(z) (9)
In the formula (9), linear (·) represents a Linear classification function;
Figure FDA0003171814870000034
c represents the number of classes of cells;
step 6, constructing a cross entropy loss function L by using the formula (10), and training a network formed by two branches and a linear classifier by using a gradient descent algorithm to ensure that the cross entropy loss function L is converged, thereby obtaining a trained cervical cell image classification model:
Figure FDA0003171814870000035
in the formula (10), y label And N is the actual category corresponding to the cell image sample, and is the total number of the cell image samples.
CN202110820463.2A 2021-07-20 2021-07-20 Cervical cell image classification method based on visual converter and image convolution network Active CN113469119B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202110820463.2A CN113469119B (en) 2021-07-20 2021-07-20 Cervical cell image classification method based on visual converter and image convolution network

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202110820463.2A CN113469119B (en) 2021-07-20 2021-07-20 Cervical cell image classification method based on visual converter and image convolution network

Publications (2)

Publication Number Publication Date
CN113469119A CN113469119A (en) 2021-10-01
CN113469119B true CN113469119B (en) 2022-10-04

Family

ID=77881307

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202110820463.2A Active CN113469119B (en) 2021-07-20 2021-07-20 Cervical cell image classification method based on visual converter and image convolution network

Country Status (1)

Country Link
CN (1) CN113469119B (en)

Families Citing this family (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN114119977B (en) * 2021-12-01 2022-12-30 昆明理工大学 Graph convolution-based Transformer gastric cancer canceration region image segmentation method
CN115661688B (en) * 2022-10-09 2024-04-26 武汉大学 Unmanned aerial vehicle target re-identification method, system and equipment with rotation invariance
CN117036811A (en) * 2023-08-14 2023-11-10 桂林电子科技大学 Intelligent pathological image classification system and method based on double-branch fusion network

Citations (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2017215284A1 (en) * 2016-06-14 2017-12-21 山东大学 Gastrointestinal tumor microscopic hyper-spectral image processing method based on convolutional neural network
CN110363188A (en) * 2019-04-18 2019-10-22 麦克奥迪(厦门)医疗诊断系统有限公司 Cervical cell image classification method based on convolutional neural networks
CN111274903A (en) * 2020-01-15 2020-06-12 合肥工业大学 Cervical cell image classification method based on graph convolution neural network
CN111461258A (en) * 2020-04-26 2020-07-28 武汉大学 Remote sensing image scene classification method of coupling convolution neural network and graph convolution network
CN112750115A (en) * 2021-01-15 2021-05-04 杭州电子科技大学 Multi-modal cervical carcinoma pre-lesion image recognition method based on graph neural network
CN113052257A (en) * 2021-04-13 2021-06-29 中国电子科技集团公司信息科学研究院 Deep reinforcement learning method and device based on visual converter
JPWO2021132633A1 (en) * 2019-12-26 2021-07-01

Patent Citations (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2017215284A1 (en) * 2016-06-14 2017-12-21 山东大学 Gastrointestinal tumor microscopic hyper-spectral image processing method based on convolutional neural network
CN110363188A (en) * 2019-04-18 2019-10-22 麦克奥迪(厦门)医疗诊断系统有限公司 Cervical cell image classification method based on convolutional neural networks
JPWO2021132633A1 (en) * 2019-12-26 2021-07-01
CN111274903A (en) * 2020-01-15 2020-06-12 合肥工业大学 Cervical cell image classification method based on graph convolution neural network
CN111461258A (en) * 2020-04-26 2020-07-28 武汉大学 Remote sensing image scene classification method of coupling convolution neural network and graph convolution network
CN112750115A (en) * 2021-01-15 2021-05-04 杭州电子科技大学 Multi-modal cervical carcinoma pre-lesion image recognition method based on graph neural network
CN113052257A (en) * 2021-04-13 2021-06-29 中国电子科技集团公司信息科学研究院 Deep reinforcement learning method and device based on visual converter

Non-Patent Citations (2)

* Cited by examiner, † Cited by third party
Title
Cervical cell classification with graph convolutional network;Jun Shi et al;《Computer Methods and Programs in Biomedicine》;20210131;全文 *
基于深度图卷积胶囊网络的图分类模型;刘海潮等;《计算机科学》;20200930;全文 *

Also Published As

Publication number Publication date
CN113469119A (en) 2021-10-01

Similar Documents

Publication Publication Date Title
CN113469119B (en) Cervical cell image classification method based on visual converter and image convolution network
CN110555446B (en) Remote sensing image scene classification method based on multi-scale depth feature fusion and migration learning
CN108985238B (en) Impervious surface extraction method and system combining deep learning and semantic probability
CN111368896B (en) Hyperspectral remote sensing image classification method based on dense residual three-dimensional convolutional neural network
CN107506761B (en) Brain image segmentation method and system based on significance learning convolutional neural network
Panchal et al. Biogeography based satellite image classification
CN104992223B (en) Intensive Population size estimation method based on deep learning
CN110135459B (en) Zero sample classification method based on double-triple depth measurement learning network
CN108090472B (en) Pedestrian re-identification method and system based on multi-channel consistency characteristics
CN112347970B (en) Remote sensing image ground object identification method based on graph convolution neural network
CN111914907A (en) Hyperspectral image classification method based on deep learning space-spectrum combined network
CN108052966A (en) Remote sensing images scene based on convolutional neural networks automatically extracts and sorting technique
CN113378792B (en) Weak supervision cervical cell image analysis method fusing global and local information
CN111090764A (en) Image classification method and device based on multitask learning and graph convolution neural network
CN111222519B (en) Construction method, method and device of hierarchical colored drawing manuscript line extraction model
CN111161244B (en) Industrial product surface defect detection method based on FCN + FC-WXGboost
CN113988147B (en) Multi-label classification method and device for remote sensing image scene based on graph network, and multi-label retrieval method and device
CN113780292A (en) Semantic segmentation network model uncertainty quantification method based on evidence reasoning
CN110619059A (en) Building marking method based on transfer learning
CN111652273A (en) Deep learning-based RGB-D image classification method
CN113298129A (en) Polarized SAR image classification method based on superpixel and graph convolution network
CN114299324A (en) Pathological image classification method and system based on multi-scale domain confrontation network
CN114863348A (en) Video target segmentation method based on self-supervision
CN114692732A (en) Method, system, device and storage medium for updating online label
Bhimavarapu et al. Analysis and characterization of plant diseases using transfer learning

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant