CN115810191A - Pathological cell classification method based on multi-attention fusion and high-precision segmentation network - Google Patents

Pathological cell classification method based on multi-attention fusion and high-precision segmentation network Download PDF

Info

Publication number
CN115810191A
CN115810191A CN202211710618.8A CN202211710618A CN115810191A CN 115810191 A CN115810191 A CN 115810191A CN 202211710618 A CN202211710618 A CN 202211710618A CN 115810191 A CN115810191 A CN 115810191A
Authority
CN
China
Prior art keywords
pathological
pathological cell
medical image
cells
classification
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN202211710618.8A
Other languages
Chinese (zh)
Inventor
胡鹤轩
丁秋阳
黄倩
杨天金
胡强
巫义锐
张晔
狄峰
胡震云
周晓军
沈勤
吕京澴
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Jiuyisanluling Medical Technology Nanjing Co ltd
Hohai University HHU
Original Assignee
Jiuyisanluling Medical Technology Nanjing Co ltd
Hohai University HHU
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Jiuyisanluling Medical Technology Nanjing Co ltd, Hohai University HHU filed Critical Jiuyisanluling Medical Technology Nanjing Co ltd
Priority to CN202211710618.8A priority Critical patent/CN115810191A/en
Publication of CN115810191A publication Critical patent/CN115810191A/en
Pending legal-status Critical Current

Links

Images

Landscapes

  • Image Analysis (AREA)

Abstract

The invention discloses a pathological cell classification method based on a multi-attention fusion mechanism and a high-precision segmentation network, which comprises the following steps of: collecting and preprocessing pathological cell images; constructing a multi-attention fusion module, which comprises a construction compression excitation module and a space attention module; constructing a high-precision segmentation network, and generating an interested region on the pathological cell image through the high-precision segmentation network; building a depth network model based on multi-attention fusion and high-precision segmentation network, and training the model by utilizing the preprocessed pathological cell image to obtain a network model with the highest pathological cell classification accuracy; and (4) sending the pathological cell image to be detected into the optimized network model to obtain a classification result of the pathological cell sample. The invention enables the model to learn the channel weight and the space weight in the cell image, and can extract the characteristics of pathological cell areas in the image for classification by using a high-precision segmentation network, thereby effectively reducing the calculated amount and the noise influence.

Description

Pathological cell classification method based on multi-attention fusion and high-precision segmentation network
Technical Field
The invention belongs to the technical field of digital image processing and medical science intersection, and particularly relates to a pathological cell classification method based on multi-attention fusion and high-precision segmentation network.
Background
With the development of science and technology, cancer is no longer an incurable disease, the early cancer cure rate is more than 90%, and the late cancer cure rate is only about 10%. Therefore, early screening for cancer is very important for the prevention and treatment of cancer. Pap smear testing is a common and effective way to screen for cancer. The professional physician observes the cell morphology in the pap smear under a microscope, classifies each cell, and determines whether the cell sample is cancerous. The cost for classifying pathological cells by manpower is high, and the classification result often has subjective deviation and influences the screening accuracy, so that the research and development of the technology for automatically classifying the medical images of the pathological cells have great significance.
Some pathological cell medical image classification methods include a pathological cell medical image classification method based on context modeling, a pathological cell medical image classification method based on a atlas neural network, and the like. Chinese patent application (CN 112200253A) "cervical cell image classification method based on SENet", which uses SENet to realize classification of cervical pathological cell medical images. The method mainly has the following defects: (1) SENET uses a channel attention mechanism, only channel weight is calculated, and space weight is not calculated, so that the attention of a network model in the method to key areas of pathological cell medical images is not enough, cell features cannot be fully extracted by the network model, and the accuracy of cell classification is poor; (2) According to the method, the global features extracted from the whole cell image by the network model are sent to the classifier for classification, so that the calculated amount is large, and the global features often contain some noises which do not belong to cells, so that the accuracy of model classification is influenced.
Disclosure of Invention
The invention aims to overcome the defects of the prior art and provides a pathological cell classification method based on multi-attention fusion and a high-precision segmentation network.
In order to solve the technical problems, the invention adopts the following technical scheme.
A pathological cell classification method based on a multi-attention fusion mechanism and a high-precision segmentation network comprises the following steps:
step 1, collecting and preprocessing a pathological cell medical image;
step 2, constructing a multi-attention fusion module, which comprises constructing a compression excitation module so that a network model can learn the relationship among different channels; constructing a spatial attention module to help the network model to identify key feature regions in the cell image;
step 3, constructing a high-precision segmentation network, and generating an interested area on the pathological cell medical image through the high-precision segmentation network;
step 4, building a depth network model based on a multi-attention fusion and high-precision segmentation network, wherein the depth network model comprises a multi-attention fusion module and a high-precision segmentation network module, and training the depth network model by utilizing a preprocessed pathological cell medical image to obtain a depth network model with the highest pathological cell classification accuracy;
and 5, sending the pathological cell medical image to be detected into the depth network model optimized in the step 4 to obtain a classification result of the pathological cell sample.
Specifically, the step 1 process comprises:
step 1.1. Collect the images of pap smears made with pathological cells under a microscope, classify the pathological cells in each pathological cell medical image by a specialist as follows: cells with high karyoplasmic ratio morphology meeting HSIL standard, abnormal keratinocytes in HSIL or SCC, abnormal naked nuclei in HSIL or SCC, hollowed cells, non-hollowed cells with morphology meeting LSIL standard, abnormal glandular cells, abnormal metaplastic cells, normal naked nuclei, normal mesodermal cells, pathogenic microbial cells, normal parabasal layer cells, classical squamous metaplasia, classical keratinization;
step 1.2, performing image enhancement on the collected pathological cell medical image, including random angle rotation, turning, cutting, displacement, scaling and the like, so that the obtained pathological cell medical image data set is richer;
and 1.3, labeling each pathological cell medical image by using a labelme labeling tool, wherein the labeling comprises a classification label and a cell outline coordinate of pathological cells in the pathological cell medical image.
Specifically, the step 2 includes:
step 2.1. Construct the compression excitation module, including:
compression: performing global average pooling on the pathological cell medical images, compressing the pathological cell medical images from H multiplied by W multiplied by C to 1 multiplied by C after compression operation, and aggregating cross-space characteristics; wherein H represents the height of the pathological cell medical image, W represents the width of the pathological cell medical image, and C represents the number of channels of the pathological cell medical image; the calculation formula for the compression operation is:
Figure BDA0004026122280000021
wherein u is c (i, j) pixel values representing different channels of the pathological cell medical image at the (i, j) position;
excitation: learning the weight of each channel of the pathological cell medical image, wherein the weight comprises two fully-connected layers and finally outputs a 1 × 1 × C vector; the calculation formula of the excitation operation is:
S=F ex (z,W)=σ(g(z,w))=σ(W 2 δ(W 1 z)) (2)
wherein sigma represents a sigmoid activation function, and delta represents a ReLu function; w 1 Refers to the weight parameter, W, of the first fully-connected layer 2 Refers to the weight parameter of the second fully-connected layer;
scale: multiplying the weight of each channel learned in the excitation operation by the original features, the Scale operation is calculated by the formula:
u c (i,j)=S×u c (i,j) (3)
wherein u is c Is a medical image of pathological cells input into the Scale operation formula, u c Representing Scale operation disclosureOutputting the pathological cell medical image;
step 2.2, constructing a space attention module: firstly, aggregating the channel characteristics of the pathological cell medical image by using a global average pooling method, wherein the global average pooling along the channel direction is represented by the following formula:
Figure BDA0004026122280000031
wherein x (i, j) represents a feature of a point (i, j) on the pathological cell medical image;
the weight G of the attention map is then learned with a 1 × 1 convolution, which computes the formula:
u=F Conv =Mx(i,j) (5)
where M is a weight matrix learned by 1 × 1 convolution;
and finally, carrying out Scale operation, wherein the calculation formula is as follows:
u c (i,j)=G×u c (i,j) (6)。
specifically, the step 3 includes:
generating a plurality of rectangular anchor frames on each pixel point on the pathological cell medical image, wherein the anchor frames select 3 sizes and 3 different length-width ratios, and the total number of the anchor frames is 9; classifying each anchor frame and performing an offset regression task; wherein the anchor frame classification task uses a binary cross entropy loss function:
L cls (p,y)=-y log(p)-(1-y)log(1-p) (7)
where y is the true value and p is the predicted value;
the loss function used for the anchor frame offset regression is:
Figure BDA0004026122280000032
wherein y is Is a predicted value, delta is a constant, and the value can be determined according to the situation;
taking the weighted sum of the penalty function used by the anchor box classification task and the penalty function used by the anchor box offset as the overall penalty function:
Figure BDA0004026122280000033
i is the index of each anchor frame; alpha is a constant determined according to the situation and used for controlling the proportion of the loss of the anchor frame classification task in the total loss; p is a radical of * 、t * Respectively representing true values, and p and t respectively representing predicted values; n is a radical of cls Representing the number of anchor frames;
calculating the position and the size of the proposed frame by using the anchor frame offset obtained by the anchor frame regression, and further calculating the intersection ratio of the proposed frame and the real frame, wherein the formula is as follows:
Figure BDA0004026122280000034
wherein, F IoU Representing the intersection ratio of the suggestion frame and the real frame, A representing the area of the real frame, B representing the area of the suggestion frame; weighting and summing the classification scores obtained by the anchor frame classification task with the cross-over ratio, wherein the calculation formula is as follows:
Ts=F cls +λF IoU (11)
wherein Ts represents a fusion score, F cls Representing classification scores, wherein lambda is a coefficient and can be obtained by model learning, and the weight of intersection ratio in the fusion scores is measured;
and processing the fusion score by adopting a non-maximum inhibition algorithm, and screening to obtain the most appropriate suggestion box.
Specifically, the step 4 includes:
step 4.1, adding a multi-attention fusion module in the ResNet18 network; the network model is based on a ResNet18 network, the ResNet18 network has 5 stages, the first stage is convolution of 7 multiplied by 7, and four stages consisting of residual blocks; the compression excitation module and the spatial attention module are serially embedded into the second to fourth stages of the ResNet18 network, and the ResNet18 network finally outputs a characteristic diagram obtained by extracting a pathological cell medical image;
step 4.2, sending the characteristic diagram into a high-precision segmentation network to obtain an interested area, namely an anchor frame containing pathological cells; obtaining the characteristics contained in the region of interest according to the characteristic diagram obtained in the step 4.1 and the anchor frame obtained in the previous step, and sending the characteristics to a classifier for classification tasks after pooling;
and 4.3, performing pretreatment on the medical image of the pathological cells according to the ratio of 8:2, dividing a training set and a test set; the training set is sent into the network model set up in the front for training, the loss function used in the training process adopts a focus loss function, and the formula is as follows:
Figure BDA0004026122280000041
where FL represents the focus loss function value,
Figure BDA0004026122280000042
representing a predicted value, p representing a true value, gamma being a weight, and confirming an optimal value through experiments;
and after the training is finished, testing the network model by using the test set, and selecting the model with the highest average accuracy as the optimal pathological cell medical image classification model.
Specifically, the step 5 includes:
step 5.1, the medical image of the pathological cell to be detected is sent to the network model optimized in the step 4, and the classification result of the network model to the inspection sample is obtained;
and 5.2, judging a classification result: judging the sample as canceration positive if the classification result belongs to one of cells with high nuclear-to-cytoplasmic ratio and morphology meeting HSIL standard, abnormal keratinocytes in HSIL or SCC, abnormal naked nuclei, hollowed cells, non-hollowed cells with morphology meeting LSIL standard, abnormal glandular cells and abnormal metaplasia cells in HSIL or SCC; and if the classification result belongs to one of normal naked nucleus, normal mesodermal cell, pathogenic microorganism cell, normal accessory basal layer cell, normal cervical canal cell and typical parakeratosis, judging the result as canceration negative.
Compared with the prior art, the invention has the following advantages and beneficial effects:
1. according to the method, a multi-attention mechanism is introduced, a channel attention mechanism is adopted to enable a network model to learn weights among different channels, a space attention mechanism is adopted to enable the network model to learn weights of different regions of cell images, and the network model can better extract key features in the cell images through a method of fusing the two attention mechanisms, so that a classification task can be favorably carried out, and the classification accuracy is improved.
2. The invention uses a high-precision segmentation network, adds a module for combining classification scores and cross-over ratios on the basis of the traditional regional suggestion network, and processes the fusion scores by using a non-maximum suppression algorithm to obtain the most appropriate suggestion frame. The high-precision segmentation network in the invention processes the classification score, which is a single variable, by adopting a non-maximum suppression algorithm so as to screen the most suitable suggestion frame, so that the algorithm not only needs to pay attention to whether the classification of the suggestion frame is correct, but also needs to meet the requirement of higher cross-over ratio as far as possible, thereby overcoming the problem that the cross-over ratio of the suggestion frames obtained by the traditional region suggestion network is poor in effect, enabling the model to more accurately position the region of interest, being capable of accurately extracting the features of the region of interest for classification, effectively reducing the calculated amount, eliminating the noise interference outside the region of interest and improving the classification accuracy.
3. The invention introduces a focus loss function, and the focus loss function adds a prepositive term in front of the cross entropy function, so that the weight of a small number of samples in a data set in the overall loss function can be increased. Compared with a cross entropy function, the focus loss function can effectively solve the problem of unbalanced training sample types in the model training process by improving the weight of a small number of samples in the loss function.
Drawings
FIG. 1 is a flow chart of a method according to an embodiment of the present invention.
Fig. 2 is a block diagram of a compression excitation module according to the prior art.
Fig. 3 is a block diagram of a spatial attention module according to the prior art.
Fig. 4 is a diagram of a high-precision segmentation network architecture according to an embodiment of the present invention.
Detailed Description
The invention discloses a pathological cell classification method based on multi-attention fusion and high-precision segmentation network, which comprises the following steps: collecting a pap smear medical image dataset made with pathological cells; constructing a multi-attention fusion module for learning the weight and the spatial region weight among different channels of the medical image; constructing a high-precision segmentation network module to extract the characteristics of a pathological cell area in a medical image; constructing a depth network model based on a multi-attention fusion module and a high-precision segmentation network; and classifying pathological cells by using the trained deep network model with the optimal parameters, and further judging whether the cell sample in the medical image is cancerated or not. The invention not only needs less calculation amount, but also greatly improves the accuracy of pathological cell classification.
The present invention will be described in further detail with reference to the accompanying drawings.
FIG. 1 is a flow chart of a pathological cell classification method based on a multi-attention fusion mechanism and a high-precision segmentation network provided by the invention. As shown in fig. 1, the method of this embodiment includes the following steps:
step 1, collecting and preprocessing a pathological cell medical image. The process comprises the following steps:
step 1.1, collecting images of a pap smear made of pathological cells under a microscope field, classifying the pathological cells in each image by a professional physician into 13 categories of cells with high nuclear-to-cytoplasmic ratio morphology meeting HSIL standards, abnormal keratinocytes in HSIL or SCC, abnormal naked nuclei in HSIL or SCC, hollowed cells, non-hollowed cells with morphology meeting LSIL standards, abnormal glandular cells, abnormal metaplastic cells, normal naked nuclei, normal mesodermal cells, pathogenic microbial cells, normal parabasal layer cells, typical squamous metaplasia and typical parakeratosis;
step 1.2, performing image enhancement on the collected pathological cell medical image, including random angle rotation, turning, cutting, displacement, scaling and the like, so that the obtained pathological cell medical image data set is richer;
and 1.3, labeling each pathological cell medical image by using a labelme labeling tool, wherein the labeling comprises a classification label and a cell outline coordinate of pathological cells in the pathological cell medical image.
Step 2, constructing a multi-attention fusion module, which comprises constructing a compression excitation module, so that a network model can learn the relationship between different channels of the pathological cell medical image; and constructing a spatial attention module to help the network model to identify key characteristic regions in the pathological cell medical image. The method specifically comprises the following steps:
and 2.1, constructing a compression excitation module. The compression excitation module can learn the relation among different channels of the pathological cell medical image, and the function of the compression excitation module is realized through three steps of compression, excitation and Scale.
The compression operation is to perform global average pooling on the pathological cell medical images, and after the compression operation, the size of the pathological cell medical images is compressed from H multiplied by W multiplied by C to 1 multiplied by C, so that the features across the space are aggregated. Where H denotes the height of the pathological cell medical image, W denotes the width of the pathological cell medical image, and C denotes the number of channels of the pathological cell medical image. The calculation formula for the compression operation is:
Figure BDA0004026122280000061
wherein u is c (i, j) represents pixel values of different channels of the pathological cell medical image at the (i, j) position.
The excitation operation learns the weights of each channel of the pathological cell medical image, which contains two fully-connected layers, ultimately outputting a 1 × 1 × C vector. The calculation formula of the excitation operation is:
S=F ex (z,W)=σ(g(z,w))=σ(W 2 δ(W 1 z)) (2)
wherein sigmaRepresents sigmoid activation function, and δ represents ReLu function. W 1 Weight parameter referring to the first fully connected layer, c 2 Refers to the weight parameter of the second fully connected layer.
The Scale operation is to multiply the weight of each channel learned in the excitation operation by the original feature. The formula for Scale operation is:
u c (i,j)=S×u c (i,j) (3)
wherein u c Is a medical image of pathological cells input into the Scale operation formula, u c Representing a pathological cell medical image output by a Scale operation formula;
and 2.2, constructing a space attention module. Firstly, aggregating the channel characteristics of the pathological cell medical image by using a global average pooling method, wherein the global average pooling formula along the channel direction is as follows:
Figure BDA0004026122280000071
where x (i, j) represents the characteristic of point (i, j) on the pathological cell medical image.
The weight G of the attention map is then learned by a 1 × 1 convolution, which is calculated by the formula:
G=F Conv =Mx(i,j) (5)
where M is a weight matrix learned by 1 × 1 convolution.
And finally, carrying out Scale operation, wherein the calculation formula is as follows:
u′ c (i,j)=u×u c (i,j) (6)
and 3, constructing a high-precision segmentation network, and generating an interested region on the pathological cell medical image through the high-precision segmentation network, so that the next classification task can be conveniently carried out. The method specifically comprises the following steps:
a plurality of rectangular anchor frames are generated on each pixel point on the pathological cell medical image, and the anchor frames are selected from 3 sizes and 3 different length-width ratios, and 9 anchor frames are calculated. And then, classifying and performing an offset regression task on each anchor frame. Wherein the anchor frame classification task uses a binary cross entropy loss function:
L cls (p,y)=-y log(p)-(1-y)log(1-p) (7)
where y is the true value and p is the predicted value.
The loss function used for the anchor frame offset regression is:
Figure BDA0004026122280000072
wherein y is δ is a constant value for the predicted value, and its value may be determined as appropriate.
Taking the weighted sum of the penalty function used by the anchor box classification task and the penalty function used by the anchor box offset as the overall penalty function:
Figure BDA0004026122280000073
i is the index of each anchor frame; a is a constant determined according to the situation and used for controlling the proportion of the loss of the classification task in the total loss; p is a radical of * 、t * Respectively representing true values, and p and t respectively representing predicted values; n is a radical of cls The anchor frame number is indicated.
The position and the size of the suggested frame can be calculated by using the anchor frame offset obtained by the anchor frame regression, and then the intersection ratio of the suggested frame and the real frame can be calculated, wherein the calculation formula is as follows:
Figure BDA0004026122280000074
wherein, F IoU Represents the intersection ratio of the suggestion frame and the real frame, A represents the area of the real frame, and B represents the area of the suggestion frame.
Weighting and summing the classification scores obtained by the classification tasks with the cross-over ratio, wherein the calculation formula is as follows:
Ts=F cls +λF IoU (11)
whereinTs represents the fusion score, F cls Represents the classification score, and λ is a coefficient that can be learned by the model and measures the weight of the cross-over ratio in the fusion score.
And processing the fusion score by adopting a non-maximum inhibition algorithm, and screening to obtain the most appropriate suggestion box.
The high-precision segmentation network used by the invention is added with a module for combining the classification scores and the cross-over ratio on the basis of the traditional regional suggestion network, and the fusion scores are processed by using a non-maximum suppression algorithm to obtain the most appropriate suggestion frame. The high-precision segmentation network in the invention processes the classification score, which is a single variable, by adopting a non-maximum suppression algorithm so as to screen the most suitable suggestion frame, so that the algorithm not only needs to pay attention to whether the classification of the suggestion frame is correct, but also needs to meet the requirement of higher cross-over ratio as far as possible, thereby overcoming the problem that the cross-over ratio of the suggestion frames obtained by the traditional region suggestion network is poor in effect, enabling the model to more accurately position the region of interest, being capable of accurately extracting the features of the region of interest for classification, effectively reducing the calculated amount, eliminating the noise interference outside the region of interest and improving the classification accuracy.
And 4, building a depth network model based on multi-attention fusion and a high-precision segmentation network, and training the depth network model by utilizing the preprocessed pathological cell medical image to obtain the depth network model with the highest pathological cell classification accuracy. The method specifically comprises the following steps:
and 4.1, adding a multi-attention fusion module in the ResNet18 network. The network model is based on a ResNet18 network, the ResNet18 network has 5 stages, the first stage is convolution of 7 multiplied by 7, and the last 4 stages are composed of residual blocks. And the compressed excitation module and the spatial attention module are serially embedded into the second to fourth stages of the ResNet18 network, and the feature extraction part of the network model finally outputs a feature map obtained by extracting a pathological cell medical image.
And 4.2, sending the characteristic diagram into a high-precision segmentation network to obtain an interested area, namely an anchor frame containing pathological cells. And (4) acquiring the characteristics contained in the region of interest according to the characteristic diagram obtained in the step (4.1) and the anchor frame obtained in the previous step, and sending the characteristics to a classifier for classification after pooling.
And 4.3, performing the data set obtained in the step 1 according to the ratio of 8: and 2, dividing a training set and a test set. The training set is sent to the network model set up in the front for training, the loss function used in the training process adopts a focus loss function, and the calculation formula is as follows:
Figure BDA0004026122280000081
where FL represents the focus loss function value,
Figure BDA0004026122280000082
the predicted value is represented, p represents the true value, gamma is a weight, and the optimal value needs to be confirmed through experiments.
The focus loss function is preceded by a prefix term in the conventional cross-entropy function, which can increase the weight of a small number of samples in the data set in the overall loss function. Compared with a cross entropy function, the focus loss function can effectively solve the problem of unbalanced training sample types in the model training process by improving the weight of a small number of samples in the loss function.
And after the training is finished, testing the network model by using the test set, and selecting the model with the highest average accuracy as the optimal network model.
And 5, sending the pathological cell medical image to be detected into the network model optimized in the step 4 to obtain a classification result of the pathological cell sample. The method specifically comprises the following steps:
and 5.1, sending the pathological cell medical image to be detected into the network model optimized in the step 4 to obtain a classification result of the network model on the pathological cell medical image.
And 5.2, judging the classification result. If the classification result belongs to one of cells with high nuclear-to-mass ratio and morphology meeting HSIL standard, abnormal keratinocytes in HSIL or SCC, abnormal naked nuclei, hollow cells in HSIL or SCC, non-hollow cells with morphology meeting LSIL standard, abnormal gland cells and abnormal metaplasia cells, judging the sample as cancerization positive; and if the classification result belongs to one of normal naked nucleus, normal mesodermal cell, pathogenic microorganism cell, normal accessory basal layer cell, normal cervical canal cell and typical parakeratosis, judging the result as canceration negative.

Claims (6)

1. A pathological cell classification method based on a multi-attention fusion mechanism and a high-precision segmentation network is characterized by comprising the following steps:
step 1, collecting and preprocessing a pathological cell medical image;
step 2, constructing a multi-attention fusion module, which comprises constructing a compression excitation module so that a network model can learn the relationship among different channels; constructing a spatial attention module to help the network model identify key feature regions in the cell image;
step 3, constructing a high-precision segmentation network, and generating an interested area on the pathological cell medical image through the high-precision segmentation network;
step 4, building a depth network model based on a multi-attention fusion and high-precision segmentation network, wherein the depth network model comprises a multi-attention fusion module and a high-precision segmentation network module, and training the depth network model by utilizing a preprocessed pathological cell medical image to obtain a depth network model with the highest pathological cell classification accuracy;
and 5, sending the pathological cell medical image to be detected into the depth network model optimized in the step 4 to obtain a classification result of the pathological cell sample.
2. The method for classifying pathological cells based on the multi-attention fusion mechanism and the high-precision segmentation network as claimed in claim 1, wherein the step 1 process comprises:
step 1.1. Collect the images of pap smears made with pathological cells under a microscope, classify the pathological cells in each pathological cell medical image by a specialist as follows: cells with high karyoplasmic ratio morphology meeting HSIL standard, abnormal keratinocytes in HSIL or SCC, abnormal naked nuclei in HSIL or SCC, hollowed cells, non-hollowed cells with morphology meeting LSIL standard, abnormal glandular cells, abnormal metaplastic cells, normal naked nuclei, normal mesodermal cells, pathogenic microbial cells, normal parabasal layer cells, classical squamous metaplasia, classical keratinization;
step 1.2, performing image enhancement on the collected pathological cell medical image, including random angle rotation, turning, cutting, displacement, scaling and the like, so that the obtained pathological cell medical image data set is richer;
and 1.3, labeling each pathological cell medical image by using a labelme labeling tool, wherein the labeling comprises a classification label and a cell outline coordinate of pathological cells in the pathological cell medical image.
3. The method for classifying pathological cells based on the multi-attention fusion mechanism and the high-precision segmentation network as claimed in claim 1, wherein the step 2 is specifically as follows:
step 2.1. Construct the compression excitation module, including:
compression: performing global average pooling on the pathological cell medical images, compressing the pathological cell medical images from H multiplied by W multiplied by C to 1 multiplied by C after compression operation, and aggregating cross-space characteristics; wherein H represents the height of the pathological cell medical image, W represents the width of the pathological cell medical image, and C represents the number of channels of the pathological cell medical image; the calculation formula for the compression operation is:
Figure FDA0004026122270000021
wherein u is c (i, j) pixel values representing different channels of the pathological cell medical image at the (i, j) position;
excitation: learning the weight of each channel of the pathological cell medical image, wherein the weight comprises two fully-connected layers and finally outputs a 1 × 1 × C vector; the calculation formula of the excitation operation is:
S=F ex (z,W)=σ(g(z,w))=σ(W 2 δ(W 1 z)) (2)
wherein sigma represents a sigmoid activation function, and delta represents a ReLu function; w1 refers to the weight parameter of the first fully-connected layer, W 2 Refers to the weight parameter of the second fully-connected layer;
scale: multiplying the weight of each channel learned in the excitation operation by the original features, and the calculation formula of Scale operation is:
u′ c (i,j)=S×u c (i,j) (3)
wherein u is c Is a pathological cell medical image u 'input to the Scale operation formula' c Representing a pathological cell medical image output by a Scale operation formula;
step 2.2, constructing a space attention module: firstly, aggregating the channel characteristics of the pathological cell medical image by using a global average pooling method, wherein the global average pooling along the channel direction is represented by the following formula:
Figure FDA0004026122270000022
wherein x (i, j) represents a feature of a point (i, j) on the pathological cell medical image;
the weight G of the attention map is then learned with a 1 × 1 convolution, which computes the formula:
G=F Conv =Mx(i,j) (5)
where M is a weight matrix learned by 1 × 1 convolution;
and finally, carrying out Scale operation, wherein the calculation formula is as follows:
u′ c (i,j)=G×u c (i,j) (6)。
4. the method for classifying pathological cells based on the multi-attention fusion mechanism and the high-precision segmentation network as claimed in claim 1, wherein the step 3 is specifically as follows:
generating a plurality of rectangular anchor frames on each pixel point on the pathological cell medical image, wherein the anchor frames select 3 sizes and 3 different length-width ratios, and the total number of the anchor frames is 9; classifying each anchor frame and performing an offset regression task; wherein the anchor frame classification task uses a binary cross entropy loss function:
L cls (p,y)=-ylog(p)-(1-y)log(1-p) (7)
where y is the true value and p is the predicted value;
the loss function used for the anchor frame offset regression is:
Figure FDA0004026122270000031
wherein y' is a predicted value and δ is a constant, the value of which can be determined as appropriate;
taking the weighted sum of the penalty function used by the anchor box classification task and the penalty function used by the anchor box offset as the overall penalty function:
Figure FDA0004026122270000032
i is the index of each anchor frame; alpha is a constant determined according to the situation and used for controlling the proportion of the loss of the anchor frame classification task in the total loss; p is a radical of * 、t * Respectively representing true values, and p and t respectively representing predicted values; n is a radical of cls Representing the number of anchor frames;
calculating the position and the size of the proposed frame by using the anchor frame offset obtained by the anchor frame regression, and further calculating the intersection ratio of the proposed frame and the real frame, wherein the formula is as follows:
Figure FDA0004026122270000033
wherein, F IoU Representing the intersection ratio of the suggestion frame and the real frame, A representing the area of the real frame, B representing the area of the suggestion frame;
weighting and summing the classification scores obtained by the anchor frame classification task with the cross-over ratio, wherein the calculation formula is as follows:
Ts=F cls +λF IoU (11)
wherein Ts represents a fusion score, F cls Representing classification scores, wherein lambda is a coefficient and can be obtained by model learning, and the weight of intersection ratio in the fusion scores is measured;
and processing the fusion score by adopting a non-maximum inhibition algorithm, and screening to obtain the most appropriate suggestion box.
5. The method for classifying pathological cells according to claim 1, wherein the step 4 comprises:
step 4.1, adding a multi-attention fusion module in the ResNet18 network; the network model is based on a ResNet18 network, the ResNet18 network has 5 stages, the first stage is convolution of 7 multiplied by 7, and four stages consisting of residual blocks; the compression excitation module and the spatial attention module are serially embedded into the second to fourth stages of the ResNet18 network, and the ResNet18 network finally outputs a characteristic diagram obtained by extracting a pathological cell medical image;
step 4.2, sending the characteristic diagram into a high-precision segmentation network to obtain an interested area, namely an anchor frame containing pathological cells; obtaining the characteristics contained in the region of interest according to the characteristic diagram obtained in the step 4.1 and the anchor frame obtained in the previous step, and sending the characteristics to a classifier for classification tasks after pooling;
and 4.3, performing pretreatment on the medical image of the pathological cells according to the ratio of 8:2, dividing a training set and a test set; the training set is sent into the network model set up in the front for training, the loss function used in the training process adopts a focus loss function, and the formula is as follows:
Figure FDA0004026122270000041
where FL represents the focus loss function value,
Figure FDA0004026122270000042
representing a predicted value, p representing a true value, gamma being a weight, and confirming an optimal value through experiments;
and after the training is finished, testing the network model by using the test set, and selecting the model with the highest average accuracy as the optimal pathological cell medical image classification model.
6. The method for classifying pathological cells based on a multi-attention fusion mechanism and a high-precision segmentation network as claimed in claim 1, wherein the step 5 specifically comprises:
step 5.1, the medical image of the pathological cell to be detected is sent to the network model optimized in the step 4, and the classification result of the network model to the inspection sample is obtained;
and 5.2, judging a classification result: judging the sample as canceration positive if the classification result belongs to one of cells with high nuclear-to-cytoplasmic ratio and morphology meeting HSIL standard, abnormal keratinocytes in HSIL or SCC, abnormal naked nuclei, hollowed cells, non-hollowed cells with morphology meeting LSIL standard, abnormal glandular cells and abnormal metaplasia cells in HSIL or SCC; and if the classification result belongs to one of normal naked nucleus, normal mesodermal cell, pathogenic microorganism cell, normal accessory basal layer cell, normal cervical canal cell and typical parakeratosis, judging the result as canceration negative.
CN202211710618.8A 2022-12-29 2022-12-29 Pathological cell classification method based on multi-attention fusion and high-precision segmentation network Pending CN115810191A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202211710618.8A CN115810191A (en) 2022-12-29 2022-12-29 Pathological cell classification method based on multi-attention fusion and high-precision segmentation network

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202211710618.8A CN115810191A (en) 2022-12-29 2022-12-29 Pathological cell classification method based on multi-attention fusion and high-precision segmentation network

Publications (1)

Publication Number Publication Date
CN115810191A true CN115810191A (en) 2023-03-17

Family

ID=85487091

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202211710618.8A Pending CN115810191A (en) 2022-12-29 2022-12-29 Pathological cell classification method based on multi-attention fusion and high-precision segmentation network

Country Status (1)

Country Link
CN (1) CN115810191A (en)

Cited By (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN115984296A (en) * 2023-03-21 2023-04-18 译企科技(成都)有限公司 Medical image segmentation method and system applying multi-attention mechanism
CN116563848A (en) * 2023-07-12 2023-08-08 北京大学 Abnormal cell identification method, device, equipment and storage medium
CN116844161A (en) * 2023-09-04 2023-10-03 深圳市大数据研究院 Cell detection classification method and system based on grouping prompt learning

Cited By (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN115984296A (en) * 2023-03-21 2023-04-18 译企科技(成都)有限公司 Medical image segmentation method and system applying multi-attention mechanism
CN115984296B (en) * 2023-03-21 2023-06-13 译企科技(成都)有限公司 Medical image segmentation method and system applying multi-attention mechanism
CN116563848A (en) * 2023-07-12 2023-08-08 北京大学 Abnormal cell identification method, device, equipment and storage medium
CN116563848B (en) * 2023-07-12 2023-11-10 北京大学 Abnormal cell identification method, device, equipment and storage medium
CN116844161A (en) * 2023-09-04 2023-10-03 深圳市大数据研究院 Cell detection classification method and system based on grouping prompt learning
CN116844161B (en) * 2023-09-04 2024-03-05 深圳市大数据研究院 Cell detection classification method and system based on grouping prompt learning

Similar Documents

Publication Publication Date Title
CN112308158B (en) Multi-source field self-adaptive model and method based on partial feature alignment
CN111652038B (en) Remote sensing sea ice image classification method based on convolutional neural network
CN115810191A (en) Pathological cell classification method based on multi-attention fusion and high-precision segmentation network
CN111191660B (en) Colon cancer pathology image classification method based on multi-channel collaborative capsule network
CN108986058B (en) Image fusion method for brightness consistency learning
CN111986125B (en) Method for multi-target task instance segmentation
CN112395442B (en) Automatic identification and content filtering method for popular pictures on mobile internet
CN111401426A (en) Small sample hyperspectral image classification method based on pseudo label learning
CN112507800A (en) Pedestrian multi-attribute cooperative identification method based on channel attention mechanism and light convolutional neural network
CN109472733A (en) Image latent writing analysis method based on convolutional neural networks
CN113743353A (en) Cervical cell classification method based on spatial, channel and scale attention fusion learning
CN113159159A (en) Small sample image classification method based on improved CNN
CN110136113B (en) Vagina pathology image classification method based on convolutional neural network
CN116452862A (en) Image classification method based on domain generalization learning
CN113989256A (en) Detection model optimization method, detection method and detection device for remote sensing image building
CN110163855A (en) A kind of color image quality evaluation method based on multichannel depth convolutional neural networks
CN113486857A (en) Ascending safety detection method and system based on YOLOv4
CN112418115A (en) Mask wearing detection method based on improved SSD model
CN114565593B (en) Full-field digital image classification and detection method based on semi-supervision and attention
CN116543414A (en) Tongue color classification and tongue redness and purple quantification method based on multi-model fusion
CN112508958B (en) Lightweight multi-scale biomedical image segmentation method
CN115100509A (en) Image identification method and system based on multi-branch block-level attention enhancement network
CN114549536A (en) Microbial colony segmentation method based on attention mechanism
CN114821174A (en) Power transmission line aerial image data cleaning method based on content perception
Tang et al. Ethnic costume grayscale image coloring method with improved Pix2Pix

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination