CN111985536B - Based on weak supervised learning gastroscopic pathology image Classification method - Google Patents

Based on weak supervised learning gastroscopic pathology image Classification method Download PDF

Info

Publication number
CN111985536B
CN111985536B CN202010690425.5A CN202010690425A CN111985536B CN 111985536 B CN111985536 B CN 111985536B CN 202010690425 A CN202010690425 A CN 202010690425A CN 111985536 B CN111985536 B CN 111985536B
Authority
CN
China
Prior art keywords
image
feature extraction
gastroscope
pathological
network model
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN202010690425.5A
Other languages
Chinese (zh)
Other versions
CN111985536A (en
Inventor
丁偕
张敬谊
刘全祥
王瑜
韦金江
刘鸣
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Shanghai Fugao Computer Technology Co ltd
WONDERS INFORMATION CO Ltd
Original Assignee
Shanghai Fugao Computer Technology Co ltd
WONDERS INFORMATION CO Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Shanghai Fugao Computer Technology Co ltd, WONDERS INFORMATION CO Ltd filed Critical Shanghai Fugao Computer Technology Co ltd
Priority to CN202010690425.5A priority Critical patent/CN111985536B/en
Publication of CN111985536A publication Critical patent/CN111985536A/en
Application granted granted Critical
Publication of CN111985536B publication Critical patent/CN111985536B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/24Classification techniques
    • G06F18/243Classification techniques relating to the number of classes
    • G06F18/24323Tree-organised classifiers
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/045Combinations of networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/08Learning methods
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/20Image preprocessing
    • G06V10/26Segmentation of patterns in the image field; Cutting or merging of image elements to establish the pattern region, e.g. clustering-based techniques; Detection of occlusion
    • G06V10/267Segmentation of patterns in the image field; Cutting or merging of image elements to establish the pattern region, e.g. clustering-based techniques; Detection of occlusion by performing operations on regions, e.g. growing, shrinking or watersheds
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/40Extraction of image or video features
    • G06V10/44Local feature extraction by analysis of parts of the pattern, e.g. by detecting edges, contours, loops, corners, strokes or intersections; Connectivity analysis, e.g. of connected components
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/40Extraction of image or video features
    • G06V10/56Extraction of image or video features relating to colour
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F21/00Security arrangements for protecting computers, components thereof, programs or data against unauthorised activity
    • G06F21/60Protecting data
    • G06F21/62Protecting access to data via a platform, e.g. using keys or access control rules
    • G06F21/6218Protecting access to data via a platform, e.g. using keys or access control rules to a system of files or objects, e.g. local or distributed file system or database
    • G06F21/6245Protecting personal data, e.g. for financial or medical purposes
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T5/00Image enhancement or restoration
    • G06T5/20Image enhancement or restoration by the use of local operators
    • G06T5/30Erosion or dilatation, e.g. thinning
    • G06T5/73

Abstract

The invention provides a based on weak supervision learned gastroscopic pathology an image classification method. Digital pathology for gastroscope the cost of image annotation is high, the problem of difficulty in obtaining large labeled datasets, the invention utilizes easily obtained coarse-grained labels to construct large-scale labels and training the gastroscope pathological image data set to obtain a weak supervision network model. And extracting the characteristics of the gastroscopic pathology image through the obtained weak supervision network model, further fusing the global and local qualitative characteristics of the gastroscopic pathology image, and finally realizing the negative and positive classification of the whole gastroscopic pathology image through a random forest classifier. By using the gastroscope pathological image classification method based on weak supervision, provided by the invention, pathological data screening information can be provided for a pathologist, diagnosis work of the pathologist is assisted, and working efficiency of the pathologist is improved.

Description

Based on weak supervised learning gastroscopic pathology image Classification method
Technical Field
The invention relates to a gastroscope pathological image classification method based on weak supervision learning, and belongs to the technical field of medical image computer-aided diagnosis.
Background
The full-view digital pathological section (Whole Slide Image, WSI) is a full-view digital image generated by fully scanning the traditional glass section with high precision and without seamless splicing by using a full-automatic microscope scanning platform. The original digital pathology image is quite large in size, and a 40-fold-magnified electronic pathology is usually composed of more than billions of pixels. This makes it necessary for the doctor to spend a great deal of time examining the digital pathology image, identifying tiny cancer cells in pictures of billions of pixels, and the effort is very great.
Deep Learning (DL) is a very popular machine Learning method in recent years, particularly convolutional neural networks (Convolutional Neural Network, CNN) have been widely applied in the field of medical imaging in recent years, and have made great progress in the aspects of classification and identification of targets of medical images, positioning and detection, segmentation of tissues and organs and lesions, and the like, so as to provide scientific methods and advanced techniques for screening, diagnosis, treatment planning, efficacy evaluation and follow-up of various serious diseases in clinical medicine. Deep learning in the current medical field mainly uses a supervised learning mode to perform model training and later test application, and has achieved great success, but the deep learning is seriously dependent on data resources, and a large number of marked data samples are required to play a role. However, for digital pathology images, the acquisition of large labeling datasets is very difficult, requiring finding and labeling the region of interest over 40-fold enlarged digital pathology images, while in order to meet the effectiveness of model learning, these datasets need to have a broad specificity of cases, which would make time and labor costs very high.
The weakly supervised learning (Weakly Supervised Learning) is a machine learning method for solving the problem of insufficient labeling sample, and can be generally divided into three categories:
first category: incomplete supervision (incomplete supervision): only a portion of the training data has labels.
The second category: uncertain supervision (inexact supervision): the training data only has coarse-grained labels.
Third category: inaccuracy supervision (inaccurate supervision): the labels of the training data are not always true.
The deep learning model can learn effective features from a large enough data set only by taking coarse granularity labels at the image level as labels of training data without definite supervision learning, so that the extensive specificity of the model is ensured. The gastric cancer pathology image has obvious specific characteristics, obvious glandular or cell aberration expression, and characteristic information is obviously different from negative pathology images.
Disclosure of Invention
The purpose of the invention is that: by utilizing the characteristic that weak supervision learning is sensitive to specific information and through reasonable algorithm model design, the problem of insufficient labeling samples is solved, and automatic classification and discrimination of gastroscope digital pathological images are realized.
In order to achieve the above purpose, the technical scheme of the invention is to provide a gastroscope pathological image classification method based on weak supervision learning, which is characterized by comprising the following steps:
step 1, obtaining pathological images and constructing a database
Collecting and acquiring a gastroscope biopsy digital pathology image, cleaning the collected data, and carrying out coarse granularity labeling on the gastroscope biopsy digital pathology image by clinical specialists, wherein the coarse granularity labeling only relates to benign and malignant classification of the gastroscope biopsy digital pathology image, so as to form a gastroscope pathology image database;
step 2, obtaining a plurality of gastroscope biopsy digital pathological images used for training the feature extraction network model from a gastroscope pathological image database, and entering step 3 after corresponding labeling results;
step 3, image preprocessing:
extracting tissue parts of each gastroscope biopsy digital pathological image, removing invalid areas, performing small block image cutting processing on the extracted images, and performing color standardization and data enhancement processing on cut image small blocks;
step 4, after the gastroscope biopsy digital pathology image obtained in the step 2 is subjected to the image preprocessing process in the step 3, a training data set consisting of a group of multi-example package bag with classification labels is obtained, wherein each gastroscope biopsy digital pathology image is regarded as a multi-example package bag, each multi-example package bag contains a plurality of example instances without classification labels, and each example instance is an image patch obtained in the step 3; if the multi-instance package bag contains at least one positive instance, the multi-instance package bag is marked as a positive type multi-instance package, and if all the instances of the multi-instance package bag are negative instances, the multi-instance package bag is marked as a negative type multi-instance package;
step 5, obtaining a feature extraction network model based on weak supervision learning:
constructing a feature extraction network model architecture, designing corresponding feature output, training a feature extraction network model and obtaining a trained feature extraction network model, wherein:
the feature extraction network model adopts a multi-example learning algorithm, the mapping relation between example instance in a multi-example packet bag and labels of the multi-example packet bag in a training data set is learned through the feature extraction network model, specific image data is subjected to feature extraction through an encoder and a plurality of feature extraction convolution layers, finally, result output is achieved, and feature vectors and probability values of image small blocks are output simultaneously through the feature extraction network model:
feature vector information: after the image small blocks pass through a feature extraction module in the feature extraction network model, the dimension reduction is realized, the features of the small block images are automatically extracted, and finally, a feature vector with one dimension is output;
probability value: the characteristic vector of the image small block is obtained through the steps, then the characteristic vector is compressed into a vector with the length of two through full connection, the first position element value of the vector with the length of two represents the weight of the image small block which is a negative image small block, the second position element value represents the weight of the image small block which is a positive image small block, finally the vector element value with the length of two is mapped to a (0, 1) interval through a normalized exponential function, and the probability value that the image small block is negative and positive is obtained and output;
step 6, after obtaining the real-time input digital pathology image of the gastroscope biopsy, preprocessing the digital pathology image of the gastroscope biopsy through the step 3, constructing and training a feature extraction network model by the step 5 through a plurality of example package bag obtained after preprocessing, and outputting corresponding feature vector information and probability values by the feature extraction network model;
step 7, feature extraction and feature fusion, comprising the following steps:
step 701, first feature extraction:
acquiring a thermodynamic diagram of a full-field gastroscope biopsy digital pathological image, extracting suspicious tissue regions according to the thermodynamic diagram, and then extracting characteristic information of the suspicious tissue regions of the pathological image as first characteristics, wherein the method specifically comprises the following steps of:
step 7011 generates a thermodynamic diagram: utilizing the feature extraction network model to output a set of probability values, and splicing the probability values by the reverse process of overlapping the image small blocks to obtain a thermodynamic diagram hetmap of the whole gastroscope biopsy digital pathological image;
step 7012, determining a mask of a suspicious lesion area: resetting a pixel value which is larger than or equal to a preset threshold value I in the thermodynamic diagram to be 1, and resetting a pixel value which is smaller than the threshold value I to be 0 to obtain a mask of the suspicious region;
step 7013, suspicious tissue region extraction: selecting an image with the lowest resolution from the digital pathology image of the gastroscope biopsy, converting the image into a gray image, and combining a mask of a suspicious region to extract the pathology image of the suspicious region, wherein the mask of the suspicious region is subjected to a nearest neighbor interpolation algorithm, and resampling the mask size to the size of the pathology image with the lowest resolution;
step 7014, extracting suspicious tissue region characteristics;
step 702, second feature extraction:
screening N image small blocks with highest probability values by using probability values of the image small blocks in each multi-example packet bag, and obtaining feature vector information corresponding to the current image small blocks output by the feature extraction network model according to indexes to serve as second features;
step 703, third feature extraction:
screening N small block images with highest probability values by using probability values of the small blocks in each multi-example packet bag, and extracting cell characteristics of the N small blocks as third characteristics, wherein the method comprises the following steps:
step 7031, calculating cell number: converting the selected image small blocks into gray images, converting the gray images into binary images according to corresponding threshold values, setting the pixel value of a target area to be 1, setting the pixel value of a background area to be 0, taking the binary images as an effective area mask, performing image processing operation based on mathematical morphology on the obtained binary images, performing morphological opening operation, namely firstly corroding and then expanding to remove fine impurity areas, separating cells with sticky conditions by using a watershed algorithm to obtain a discrete cell mask image, and counting cells in the discrete cell mask;
step 7032, calculating cell perimeter area: taking the outline of the cell as a standard ellipse approximately, counting the number of pixel points of each cell in the discrete cell mask obtained in the steps as the cell area, counting the longest axis and the shortest axis of the cell, and obtaining the perimeter of the outline of the cell by using an ellipse perimeter calculation formula;
step 7033, cell texture feature statistics: converting the selected image small blocks into gray level images, positioning cell areas on the gray level images by using a discrete cell mask, and extracting cell texture features by using a gray level symbiotic matrix in the current cell areas, wherein the common texture features defined by the gray level symbiotic matrix are as follows:
texture second order matrix ASM: corresponding to the uniformity of the texture, the smaller the ASM value, the more uniform the staining presented by the cell nucleus;
entropy ENT: the texture information is one of image information, and the image texture richness and the entropy value are in positive correlation;
inverse differential moment IDM: describing the amount of local change of the cell image texture, wherein the larger the value is, the smaller the change of different areas of the cell texture is, namely the local is very uniform;
contrast CON: reflecting the definition of the cell image and the depth of the texture groove, the greater the CON value is, the more pixels with large contrast are;
step 7034, feature fusion:
splicing the second and third features extracted from each selected image small block with the first features extracted from the whole pathological image according to rows to obtain feature vectors of the current small block image, and splicing the feature vectors of the N image small blocks according to columns to obtain a feature matrix of the current gastroscope biopsy digital pathological image; carrying out normalization processing on the feature matrix, and mapping the features of different attributes to the same distribution space so that the features of different attributes have the same initial weight; averaging the normalized feature matrix according to columns, and compressing the feature matrix into a one-dimensional feature vector;
step 8, distinguishing and classifying gastroscope pathological images:
inputting the one-dimensional feature vector obtained in the step 7034 into a pre-trained random forest classifier to obtain benign and malignant classification of the current gastroscope biopsy digital pathological image.
Preferably, the step 1 includes the steps of:
step 101, data acquisition and screening:
determining a unified dyeing mode, collecting a digital pathology image of the gastroscope biopsy, screening the collected digital pathology image of the gastroscope biopsy, and eliminating pathology data with wrong image or inaccurate information;
step 102, data desensitization:
desensitizing each acquired digital pathology image of gastroscope biopsy;
step 103, data marking:
the professional pathologist carries out image-level coarse granularity labeling on the gastroscope biopsy digital pathological images screened in the step 101 and desensitized in the step 102, wherein the coarse granularity labeling only relates to benign and malignant classification of the gastroscope biopsy digital pathological images, and does not need pixel-level image segmentation labeling in a pathological region; the marking result of the gastroscope biopsy digital pathological image which is judged to be positive is '1', and the marking result of the gastroscope biopsy digital pathological image which is judged to be negative is '0'; the original gastroscope biopsy digital pathology image and the corresponding labeling result together form the gastroscope digital pathology image database.
Preferably, the step 3 includes the steps of:
step 301, background and invalid region removal:
extracting tissue areas of the digital pathological images of the gastroscope biopsy by using an Ojin method and a quality control method, and filtering invalid tissue areas;
step 302, graph cutting processing:
for the gastroscope biopsy digital pathology image processed in the step 301, adopting a mode of overlapping and cutting images under the same size, and transferring the pathology image into an image small block with a fixed size, wherein the size of the image small block is matched with the size of an input image of a feature extraction network model based on weak supervision learning;
step 303, color standardization processing:
mapping the image small blocks corresponding to the effective tissues to the same color gamut space by using a color standardization algorithm, ensuring that the pixel value distribution of the image small blocks with the same tissue structure is subjected to the forward distribution, and eliminating the potential influence caused by inconsistent imaging of the gastroscopic pathological images due to uncontrollable differences;
step 304, image enhancement:
the random image enhancement technology is adopted, and on the basis of guaranteeing the invariance of space translation, the operations of random rotation, random translation, random mirroring and random distortion are carried out on the image small blocks subjected to the color standardization processing, so that gastroscopic pathological images under different fields of view are simulated, and the characteristics of the image small blocks are guaranteed to be fully learned and extracted by a model.
Preferably, in the step 5, the training data set is divided into three sets of sub-data sets of a training set, a verification set and a test set according to a certain proportion, so as to perform model training, and a final trained feature extraction network model is obtained, wherein the training set is used for training the feature extraction network model, the verification data set determines super parameters and training stop positions of the feature extraction network model, and the test data set verifies the final effect of the feature extraction network model.
Aiming at the problems that the cost for marking the gastroscope digital pathological image is high and a large-scale marked data set is difficult to obtain, the invention trains and obtains a weak supervision network model by utilizing the large-scale gastroscope pathological image data set constructed by the easily-obtained coarse-granularity labels. And extracting the characteristics of the gastroscopic pathology image through the obtained weak supervision network model, further fusing the global and local qualitative characteristics of the gastroscopic pathology image, and finally realizing the negative and positive classification of the whole gastroscopic pathology image through a random forest classifier. By using the gastroscope pathological image classification method based on weak supervision, provided by the invention, pathological data screening information can be provided for a pathologist, diagnosis work of the pathologist is assisted, and working efficiency of the pathologist is improved.
Drawings
FIG. 1 is a flow chart of an overall framework of the present invention;
FIG. 2 is a database construction flow;
FIG. 3 is an image preprocessing flow;
FIG. 4 is a CNN-based feature extraction flow;
fig. 5 is a feature extraction and feature fusion flow.
Detailed Description
The invention will be further illustrated with reference to specific examples. It is to be understood that these examples are illustrative of the present invention and are not intended to limit the scope of the present invention. Further, it is understood that various changes and modifications may be made by those skilled in the art after reading the teachings of the present invention, and such equivalents are intended to fall within the scope of the claims appended hereto.
The gastroscope pathological image classification method based on weak supervision learning provided by the invention utilizes the advantage that coarse granularity labeling of pathological images is easy to obtain, collects gastroscope pathological images to form a large gastroscope pathological database, and then combines the weak supervision method and a random forest classifier to classify the gastroscope pathological images. Firstly, the whole pathology image is diced through image preprocessing, then the characteristics of small blocks of the image are automatically extracted through weak supervision learning, the global and local qualitative characteristics of the gastroscopic pathology image are further fused, and finally, the negative and positive classification of the whole gastroscopic pathology image is realized through a random forest classifier.
Specifically, as shown in fig. 1, the present invention includes the steps of:
s1, obtaining pathological images and constructing a database
Collecting and acquiring a gastroscope biopsy digital pathology image, cleaning the collected data, and marking the gastroscope biopsy digital pathology image by a clinical expert to form a gastroscope pathology image database.
As shown in fig. 2, the step S1 specifically includes the following steps:
step 101, data acquisition and screening:
and determining a unified staining mode, and collecting digital pathological images of the gastroscope biopsy. And screening the acquired digital pathology images of the gastroscope biopsy, and removing pathology data with wrong images or inaccurate information.
Step 102, data desensitization:
pathology data and pathology reports typically involve private information of the patient, and the use of pathology data that is not desensitized would result in disclosure of the patient's personal privacy. Therefore, the sensitive information is desensitized by anonymizing, information conversion and other modes on each case of collected pathological data;
step 103, data marking:
the digital pathology image of gastroscope biopsy which is screened in step 101 and desensitized in step 102 is marked with coarse granularity in the image level by a professional pathologist. Coarse-granularity labeling only relates to the classification of benign and malignant images of digital pathology images of gastroscopy, and does not need labeling of pixel-level image segmentation in a lesion area. The labeling result of the gastroscope biopsy digital pathological image judged to be positive is '1', and the labeling result of the gastroscope biopsy digital pathological image judged to be negative is '0'. The original gastroscope biopsy digital pathology image and the corresponding labeling result together form a gastroscope digital pathology image database.
Step S2, image preprocessing:
obtaining a plurality of gastroscope biopsy digital pathological images from a gastroscope pathological image database, extracting a tissue part of each gastroscope biopsy digital pathological image, removing an invalid area, and performing small block cutting image processing on the extracted images. And performing color standardization and data enhancement processing on the cut image small blocks, wherein the image small blocks corresponding to all the processed gastroscope biopsy digital pathology images are used for forming a training data set of a training feature extraction network model.
As shown in fig. 3, the step S2 specifically includes the following steps:
step 201, background and invalid region removal:
after a plurality of gastroscopic biopsy digital pathological images are obtained from a gastroscopic pathological image database, tissue areas of each gastroscopic biopsy digital pathological image are extracted by an Otsu method (OTSU) and a quality control method, and invalid tissue areas with handwriting, overlapping, smudge and focusing blurring are filtered. The method aims to remove invalid information areas irrelevant to the pathological image tissues and pathological change texture forms of the gastroscope and reduce errors generated in the process of positioning the pathological image distortion areas of the model.
Step 202, cutting a graph:
and (3) for the gastroscope biopsy digital pathology image processed in the step (201), adopting a mode of overlapping and cutting images under the same size, and transferring the pathology image with the hundred million-level pixel level into an image small block with a fixed size, wherein the size of the image small block is matched with the size of an input image of the feature extraction network model based on weak supervision learning. The overlapping ground cut map endows the continuity of the space features between the adjacent image small blocks, and improves the utilization rate of the weak supervision model on the edge features of the image small blocks.
Step 203, color normalization processing:
and mapping the image small blocks corresponding to the effective tissues into the same color gamut space by using a color standardization algorithm, ensuring that the pixel value distribution of the image small blocks with the same tissue structure is subjected to the positive too distribution, and eliminating the potential influence caused by inconsistent imaging of the gastroscope pathological images due to uncontrollable differences of different film picking instruments, dyeing depths and the like.
Step 204, image enhancement: the random image enhancement technology is adopted, and on the basis of guaranteeing the invariance of space translation, the operations of random rotation, random translation, random mirroring and random distortion are carried out on the image small blocks subjected to the color standardization processing, so that gastroscopic pathological images under different fields of view are simulated, and the characteristics of the image small blocks are guaranteed to be fully learned and extracted by a model.
Step S3, obtaining a feature extraction network model based on weak supervision learning:
constructing a feature extraction network model architecture, designing corresponding feature output, training a feature extraction network model and obtaining a trained network model.
The multi-instance learning (Multiple Instance Learning, MILs) algorithm is a very effective and inexact supervision algorithm that can use simple coarse-grained labeling data, in combination with a deep learning network, to enable statistical analysis of image features of several gigapixels. In multi-instance learning, the training set consists of a set of multi-instance packages bag with classification labels, each of which contains a number of instance instances without classification labels. If the multi-instance package bag contains at least one positive instance, the multi-instance package bag is marked as a positive-class multi-instance package (positive package). If all examples of the multi-example package bag are negative examples, the multi-example package bag is marked as a negative-type multi-example package (negative package).
The whole model is processed as shown in fig. 4, and comprises the following steps:
step 301, constructing a training data set:
each gastroscopic biopsy digital pathology image in the gastroscopic pathology image database processed in the step S2 can be regarded as a multi-example bag, wherein the multi-example bag contains all image small blocks of the current gastroscopic biopsy digital pathology image, and each image small block serves as an example instance. This results in a training dataset in units of multiple example package bag whose labels are of the original gastroscopic biopsy digital pathology image.
Step 302, designing a network model:
and building a feature extraction network model adopting a multi-example learning algorithm, and learning a mapping relation between the example instance in the multi-example package bag and the labels of the multi-example package bag in the training data set through the feature extraction network model. The specific image data is subjected to feature extraction through an Encoder (Encoder) and a plurality of feature extraction convolution layers, and finally the result output is realized.
Step 303, outputting a result:
in order to ensure the analysis requirement of the subsequent feature extraction, the feature extraction network model outputs the feature vector and the probability value of the image small block at the same time.
Feature vector information: after the image small blocks pass through a feature extraction module in the feature extraction network model, the dimension reduction is realized, the features such as textures, tissues, characters and the like of the small block images are automatically extracted, and finally a feature vector with one dimension is output;
probability value: the characteristic vector of the small image block is obtained through the steps, and then the small image block is compressed into a vector with the length of two through full connection. The vector first position element value represents the weight of the image patch as a negative image patch, and the second position element value represents the weight of the image patch as a positive image patch. Finally, mapping the vector element values with the lengths of two to a (0, 1) interval through a normalized exponential function (Soft Max Function), obtaining probability values of negative and positive of the image small blocks, and outputting the probability values.
Step 304, model training: and dividing the training data set into three groups of sub data sets of a training set, a verification set and a test set according to a certain proportion for model training, and obtaining a final trained feature extraction network model. The training set is used for training the feature extraction network model, the verification data set is used for determining the super parameters and the training stopping positions of the feature extraction network model, and the test data set is used for checking the final effect of the feature extraction network model.
Step S4, feature extraction and feature fusion
In order to ensure the reliability of the classification model, qualitative analysis feature extraction is added on the basis of weak supervision learning automatic feature extraction, wherein the feature extraction comprises global features of a whole gastroscopic pathological image and local features of a suspected pathological region small image. A flow chart of feature extraction is shown in fig. 5.
After obtaining a real-time input gastroscope biopsy digital pathology image, preprocessing the gastroscope biopsy digital pathology image through the step S2, constructing and training a feature extraction network model through the step S3 by using a multi-example packet bag obtained after preprocessing, outputting corresponding feature vector information and probability values by using the feature extraction network model, and then, comprising the following steps of:
step 401, first feature extraction:
obtaining a thermodynamic diagram of the full-field gastroscope biopsy digital pathological image, extracting suspicious tissue regions according to the thermodynamic diagram, and then extracting characteristic information of the suspicious tissue regions of the pathological image as first characteristics. The method specifically comprises the following steps:
a. generating a thermodynamic diagram: and (3) utilizing the set of probability values output by the feature extraction network model, and splicing by the reverse process of overlapping the image small blocks to obtain the thermodynamic diagram hetmap of the whole gastroscope biopsy digital pathological image.
b. Determining a lesion suspicious region mask: setting a threshold value, resetting a pixel value which is larger than or equal to the threshold value in the thermodynamic diagram to be 1, and resetting a pixel value which is smaller than the threshold value to be 0, thereby obtaining a mask of the suspicious region.
c. Suspicious tissue region extraction: and selecting an image with the lowest resolution from the digital pathology image of the gastroscope biopsy, converting the image into a gray image, and extracting the pathology image of the suspicious region by combining the mask of the suspicious region. The mask of the suspicious region needs to be resampled to the size of the pathological image with the lowest resolution through the nearest neighbor interpolation algorithm.
d. Suspicious tissue region feature extraction: including but not limited to, the maximum pixel value, the average pixel value, the variance of the pixel values, the skewness and kurtosis of the distribution of the pixel values in the suspicious region, the maximum pixel value, the average pixel value, the variance of the pixel values, the skewness and kurtosis of the distribution of the pixel values in the boundary of the suspicious region, the area ratio of the suspicious region to the whole tissue region, the longest diameter of the maximum connected domain of the suspicious region, the number of pixels of the suspicious region, and the like. Wherein the suspicious region boundary is defined as: and a pixel point set formed by four adjacent pixel points from top to bottom, left to right around the center of the pixel point on the boundary line of the suspicious region.
Step 402, second feature extraction:
and screening 5 image small blocks with highest probability values by using probability values of the image small blocks in each multi-example packet bag, and obtaining feature vector information corresponding to the current image small block output by the feature extraction network model according to indexes to serve as second features. Wherein, the small image block screening quantity can be correspondingly increased or decreased according to the classification performance requirement.
Step 403, third feature extraction:
and screening 5 small block images with highest probability values by using probability values of the small image blocks in each multi-example packet bag, and extracting cell characteristics of the 5 small image blocks as third characteristics.
Quantitative cell characteristic information plays an important role in processing and analyzing pathological images. In clinic, the rapid, accurate, reliable and objective parameter analysis can avoid the influence caused by subjective factors. Based on the method, the cell characteristic extraction is carried out, and more commonly extracted characteristic parameters include cell area, perimeter, centroid, length of long and short axes, color, texture and the like.
Step 403 specifically includes the steps of:
a. cell number calculation: and converting the selected small image block into a gray image, converting into a binary image according to a corresponding threshold value, setting the pixel value of the target area to be 1, and setting the pixel value of the background area to be 0, wherein the pixel value of the background area is used as an effective area mask. Performing image processing operation based on mathematical morphology on the obtained binary image, performing morphological open operation, namely firstly corroding and then expanding to remove fine impurity regions, then separating cells with sticky condition by using a watershed algorithm to obtain a discrete cell mask image, and counting cells in the discrete cell mask.
b. Cell perimeter area calculation: taking the cell outline as a standard ellipse approximately, counting the number of pixel points of each cell in the discrete cell mask obtained by the steps as the cell area, counting the longest axis and the shortest axis of the cell, and obtaining the perimeter of the cell outline by using an ellipse perimeter calculation formula. The calculation formula of the approximate perimeter of the ellipse is as follows:
where a represents the minor axis radius of the ellipse, b represents the major axis radius of the ellipse, and C is the approximate perimeter of the ellipse.
c. Cell texture feature statistics: and converting the selected small image blocks into gray level images, and positioning cell areas on the gray level images by using a discrete cell mask. Extracting cell texture features in the current cell area by using a gray level co-occurrence matrix, wherein the common texture features defined by the gray level co-occurrence matrix are as follows:
texture second order matrix (ASM): corresponding to texture uniformity, the smaller ASM value indicates the more uniform staining exhibited by the nuclei.
Entropy (ENT): the texture information is one kind of image information, and the image texture richness and entropy value are in positive correlation.
Inverse Differential Moment (IDM): describing how much the texture of the cell image changes locally, the larger the value, the smaller the change in different areas of the cell texture, i.e. locally very uniform.
Contrast (CON): the greater the CON value, the more pixels with large contrast are, reflecting the sharpness of the cell image and the depth of the texture grooves.
Step 404, feature fusion:
and splicing the second and third features extracted from each selected image small block with the first features extracted from the whole pathological image according to rows to obtain feature vectors of the current small block image, and splicing the feature vectors of the 5 image small blocks according to columns to obtain a feature matrix of the current gastroscope biopsy digital pathological image. And carrying out normalization processing on the feature matrix, and mapping the features of different attributes to the same distribution space so that the features of different attributes have the same initial weight. The normalized feature matrix is averaged according to columns, the feature matrix is compressed into a one-dimensional feature vector, and the one-dimensional feature vector is fused with the global and local features of the current digital pathology image of the gastroscope biopsy, so that the whole pathology image can be well described.
Step S5, distinguishing and classifying gastroscope pathological images:
and distinguishing and classifying the current gastroscope biopsy digital pathological images by using a trained random forest classifier. The classifier selects Random Forest (Random Forest) because the Random Forest proves convergence by using strong number law, and the Random Forest cannot be over-fitted due to the increase of trees, is insensitive to noise data, and keeps the characteristic of small generalization error. Because the decision tree node dividing features can be randomly selected, the model can be efficiently trained when the feature dimension of the sample is very high, and a stable model can be obtained.
Specifically, the one-dimensional feature vector obtained in step 404 is input into a pre-trained random forest classifier to obtain the benign and malignant classification of the current gastroscope biopsy digital pathology image.

Claims (4)

1. The gastroscope pathological image classification method based on weak supervision learning is characterized by comprising the following steps of:
step 1, obtaining pathological images and constructing a database
Collecting and acquiring a gastroscope biopsy digital pathology image, cleaning the collected data, and carrying out coarse granularity labeling on the gastroscope biopsy digital pathology image by clinical specialists, wherein the coarse granularity labeling only relates to benign and malignant classification of the gastroscope biopsy digital pathology image, so as to form a gastroscope pathology image database;
step 2, obtaining a plurality of gastroscope biopsy digital pathological images used for training the feature extraction network model from a gastroscope pathological image database, and entering step 3 after corresponding labeling results;
step 3, image preprocessing:
extracting tissue parts of each gastroscope biopsy digital pathological image, removing invalid areas, performing small block image cutting processing on the extracted images, and performing color standardization and data enhancement processing on cut image small blocks;
step 4, after the gastroscope biopsy digital pathology image obtained in the step 2 is subjected to the image preprocessing process in the step 3, a training data set consisting of a group of multi-example package bag with classification labels is obtained, wherein each gastroscope biopsy digital pathology image is regarded as a multi-example package bag, each multi-example package bag contains a plurality of example instances without classification labels, and each example instance is an image patch obtained in the step 3; if the multi-instance package bag contains at least one positive instance, the multi-instance package bag is marked as a positive type multi-instance package, and if all the instances of the multi-instance package bag are negative instances, the multi-instance package bag is marked as a negative type multi-instance package;
step 5, obtaining a feature extraction network model based on weak supervision learning:
constructing a feature extraction network model architecture, designing corresponding feature output, training a feature extraction network model and obtaining a trained feature extraction network model, wherein:
the feature extraction network model adopts a multi-example learning algorithm, the mapping relation between example instance in a multi-example packet bag and labels of the multi-example packet bag in a training data set is learned through the feature extraction network model, specific image data is subjected to feature extraction through an encoder and a plurality of feature extraction convolution layers, finally, result output is achieved, and feature vectors and probability values of image small blocks are output simultaneously through the feature extraction network model:
feature vector information: after the image small blocks pass through a feature extraction module in the feature extraction network model, the dimension reduction is realized, the features of the small block images are automatically extracted, and finally, a feature vector with one dimension is output;
probability value: the characteristic vector of the image small block is obtained through the steps, then the characteristic vector is compressed into a vector with the length of two through full connection, the first position element value of the vector with the length of two represents the weight of the image small block which is a negative image small block, the second position element value represents the weight of the image small block which is a positive image small block, finally the vector element value with the length of two is mapped to a (0, 1) interval through a normalized exponential function, and the probability value that the image small block is negative and positive is obtained and output;
step 6, after obtaining the real-time input digital pathology image of the gastroscope biopsy, preprocessing the digital pathology image of the gastroscope biopsy through the step 3, constructing and training a feature extraction network model by the step 5 through a plurality of example package bag obtained after preprocessing, and outputting corresponding feature vector information and probability values by the feature extraction network model;
step 7, feature extraction and feature fusion, comprising the following steps:
step 701, first feature extraction:
acquiring a thermodynamic diagram of a full-field gastroscope biopsy digital pathological image, extracting suspicious tissue regions according to the thermodynamic diagram, and then extracting characteristic information of the suspicious tissue regions of the pathological image as first characteristics, wherein the method specifically comprises the following steps of:
step 7011 generates a thermodynamic diagram: utilizing the feature extraction network model to output a set of probability values, and splicing the probability values by the reverse process of overlapping the image small blocks to obtain a thermodynamic diagram hetmap of the whole gastroscope biopsy digital pathological image;
step 7012, determining a mask of a suspicious lesion area: resetting a pixel value which is larger than or equal to a preset threshold value in the thermodynamic diagram to be 1, and resetting a pixel value which is smaller than the threshold value to be 0 to obtain a mask of the suspicious region;
step 7013, suspicious tissue region extraction: selecting an image with the lowest resolution from the digital pathology image of the gastroscope biopsy, converting the image into a gray image, and combining a mask of a suspicious region to extract the pathology image of the suspicious region, wherein the mask of the suspicious region is subjected to a nearest neighbor interpolation algorithm, and resampling the mask size to the size of the pathology image with the lowest resolution;
step 7014, extracting suspicious tissue region characteristics;
step 702, second feature extraction:
screening N image small blocks with highest probability values by using probability values of the image small blocks in each multi-example packet bag, and obtaining feature vector information corresponding to the current image small blocks output by the feature extraction network model according to indexes to serve as second features;
step 703, third feature extraction:
screening N small block images with highest probability values by using probability values of the small blocks in each multi-example packet bag, and extracting cell characteristics of the N small blocks as third characteristics, wherein the method comprises the following steps:
step 7031, calculating cell number: converting the selected image small blocks into gray images, converting the gray images into binary images according to corresponding threshold values, setting the pixel value of a target area to be 1, setting the pixel value of a background area to be 0, taking the binary images as an effective area mask, performing image processing operation based on mathematical morphology on the obtained binary images, performing morphological opening operation, namely firstly corroding and then expanding to remove fine impurity areas, separating cells with sticky conditions by using a watershed algorithm to obtain a discrete cell mask image, and counting cells in the discrete cell mask;
step 7032, calculating cell perimeter area: taking the outline of the cell as a standard ellipse approximately, counting the number of pixel points of each cell in the discrete cell mask obtained in the steps as the cell area, counting the longest axis and the shortest axis of the cell, and obtaining the perimeter of the outline of the cell by using an ellipse perimeter calculation formula;
step 7033, cell texture feature statistics: converting the selected image small blocks into gray level images, positioning cell areas on the gray level images by using a discrete cell mask, and extracting cell texture features by using a gray level symbiotic matrix in the current cell areas, wherein the common texture features defined by the gray level symbiotic matrix are as follows:
texture second order matrix ASM: corresponding to the uniformity of the texture, the smaller the ASM value, the more uniform the staining presented by the cell nucleus;
entropy ENT: the texture information is one of image information, and the image texture richness and the entropy value are in positive correlation;
inverse differential moment IDM: describing the amount of local change of the cell image texture, wherein the larger the value is, the smaller the change of different areas of the cell texture is, namely the local is very uniform;
contrast CON: reflecting the definition of the cell image and the depth of the texture groove, the greater the CON value is, the more pixels with large contrast are;
step 7034, feature fusion:
splicing the second and third features extracted from each selected image small block with the first features extracted from the whole pathological image according to rows to obtain feature vectors of the current small block image, and splicing the feature vectors of the N image small blocks according to columns to obtain a feature matrix of the current gastroscope biopsy digital pathological image; carrying out normalization processing on the feature matrix, and mapping the features of different attributes to the same distribution space so that the features of different attributes have the same initial weight; averaging the normalized feature matrix according to columns, and compressing the feature matrix into a one-dimensional feature vector;
step 8, distinguishing and classifying gastroscope pathological images:
inputting the one-dimensional feature vector obtained in the step 7034 into a pre-trained random forest classifier to obtain benign and malignant classification of the current gastroscope biopsy digital pathological image.
2. The gastroscopic pathology image classification method based on weak supervised learning as set forth in claim 1, wherein the step 1 comprises the steps of:
step 101, data acquisition and screening:
determining a unified dyeing mode, collecting a digital pathology image of the gastroscope biopsy, screening the collected digital pathology image of the gastroscope biopsy, and eliminating pathology data with wrong image or inaccurate information;
step 102, data desensitization:
desensitizing each acquired digital pathology image of gastroscope biopsy;
step 103, data marking:
the professional pathologist carries out image-level coarse granularity labeling on the gastroscope biopsy digital pathological images screened in the step 101 and desensitized in the step 102, wherein the coarse granularity labeling only relates to benign and malignant classification of the gastroscope biopsy digital pathological images, and does not need pixel-level image segmentation labeling in a pathological region; the marking result of the gastroscope biopsy digital pathological image which is judged to be positive is '1', and the marking result of the gastroscope biopsy digital pathological image which is judged to be negative is '0'; the original gastroscopic biopsy digital pathology image and the corresponding labeling result jointly form the gastroscopic pathology image database.
3. The gastroscopic pathology image classification method based on weak supervised learning as set forth in claim 1, wherein the step 3 comprises the steps of:
step 301, background and invalid region removal:
extracting tissue areas of the digital pathological images of the gastroscope biopsy by using an Ojin method and a quality control method, and filtering invalid tissue areas;
step 302, graph cutting processing:
for the gastroscope biopsy digital pathology image processed in the step 301, adopting a mode of overlapping and cutting images under the same size, and transferring the pathology image into an image small block with a fixed size, wherein the size of the image small block is matched with the size of an input image of a feature extraction network model based on weak supervision learning;
step 303, color standardization processing:
mapping the image small blocks corresponding to the effective tissues to the same color gamut space by using a color standardization algorithm, ensuring that the pixel value distribution of the image small blocks with the same tissue structure is subjected to the forward distribution, and eliminating the potential influence caused by inconsistent imaging of the gastroscopic pathological images due to uncontrollable differences;
step 304, image enhancement:
the random image enhancement technology is adopted, and on the basis of guaranteeing the invariance of space translation, the operations of random rotation, random translation, random mirroring and random distortion are carried out on the image small blocks subjected to the color standardization processing, so that gastroscopic pathological images under different fields of view are simulated, and the characteristics of the image small blocks are guaranteed to be fully learned and extracted by a model.
4. The gastroscope pathological image classification method based on weak supervision learning according to claim 1, wherein in the step 5, a training dataset is divided into three sub-datasets of a training dataset, a verification dataset and a test dataset according to a certain proportion for model training, and a final trained feature extraction network model is obtained, wherein the training dataset is used for training the feature extraction network model, the verification dataset determines super parameters and training stop positions of the feature extraction network model, and the test dataset checks the final effect of the feature extraction network model.
CN202010690425.5A 2020-07-17 2020-07-17 Based on weak supervised learning gastroscopic pathology image Classification method Active CN111985536B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202010690425.5A CN111985536B (en) 2020-07-17 2020-07-17 Based on weak supervised learning gastroscopic pathology image Classification method

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202010690425.5A CN111985536B (en) 2020-07-17 2020-07-17 Based on weak supervised learning gastroscopic pathology image Classification method

Publications (2)

Publication Number Publication Date
CN111985536A CN111985536A (en) 2020-11-24
CN111985536B true CN111985536B (en) 2024-02-09

Family

ID=73438801

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202010690425.5A Active CN111985536B (en) 2020-07-17 2020-07-17 Based on weak supervised learning gastroscopic pathology image Classification method

Country Status (1)

Country Link
CN (1) CN111985536B (en)

Families Citing this family (13)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN112560957B (en) * 2020-12-17 2023-04-07 北京邮电大学 Neural network training and detecting method, device and equipment
CN113096079B (en) * 2021-03-30 2023-12-29 四川大学华西第二医院 Image analysis system and construction method thereof
CN113221978A (en) * 2021-04-27 2021-08-06 浙江师范大学 Colorectal cancer digital pathological image discrimination method and system based on weak supervised learning
CN113256572B (en) * 2021-05-12 2023-04-07 中国科学院自动化研究所 Gastroscope image analysis system, method and equipment based on restoration and selective enhancement
CN113674288B (en) * 2021-07-05 2024-02-02 华南理工大学 Automatic segmentation method for digital pathological image tissue of non-small cell lung cancer
CN113378792B (en) * 2021-07-09 2022-08-02 合肥工业大学 Weak supervision cervical cell image analysis method fusing global and local information
CN113436191B (en) * 2021-08-26 2021-11-30 深圳科亚医疗科技有限公司 Pathological image classification method, pathological image classification system and readable medium
CN114511523A (en) * 2022-01-21 2022-05-17 中山大学 Gastric cancer molecular subtype classification method and device based on self-supervision learning
CN115063592B (en) * 2022-08-16 2022-12-06 之江实验室 Multi-scale-based full-scanning pathological feature fusion extraction method and system
CN115861212B (en) * 2022-11-25 2023-07-14 中国医学科学院肿瘤医院 System and apparatus for distinguishing related disorders based on gastric mucosa biopsy samples
CN116580397B (en) * 2023-07-12 2023-11-10 北京大学 Pathological image recognition method, device, equipment and storage medium
CN117038023A (en) * 2023-10-08 2023-11-10 中国医学科学院北京协和医院 dMMR germ line mutation subtype classification method and system based on colorectal cancer pathological image
CN117392468B (en) * 2023-12-11 2024-02-13 山东大学 Cancer pathology image classification system, medium and equipment based on multi-example learning

Citations (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN105550651A (en) * 2015-12-14 2016-05-04 中国科学院深圳先进技术研究院 Method and system for automatically analyzing panoramic image of digital pathological section
CN109492711A (en) * 2018-12-07 2019-03-19 杭州电子科技大学 Malignant mela noma and non-malignant melanotic nevus classification method based on deep learning
CN109670510A (en) * 2018-12-21 2019-04-23 万达信息股份有限公司 A kind of gastroscopic biopsy pathological data screening system and method based on deep learning
CN109754879A (en) * 2019-01-04 2019-05-14 湖南兰茜生物科技有限公司 A kind of lung cancer computer aided detection method and system based on deep learning
CN110472676A (en) * 2019-08-05 2019-11-19 首都医科大学附属北京朝阳医院 Stomach morning cancerous tissue image classification system based on deep neural network
CN110751172A (en) * 2019-09-10 2020-02-04 中南大学 Weakly supervised learning pathology full-image category inference method and system
CN111265234A (en) * 2020-03-05 2020-06-12 上海市肺科医院(上海市职业病防治院) Method and system for judging properties of lung mediastinal lymph nodes

Family Cites Families (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US10672115B2 (en) * 2016-12-06 2020-06-02 Siemens Corporation Weakly supervised anomaly detection and segmentation in images
US10445879B1 (en) * 2018-03-23 2019-10-15 Memorial Sloan Kettering Cancer Center Systems and methods for multiple instance learning for classification and localization in biomedical imaging
US11730387B2 (en) * 2018-11-02 2023-08-22 University Of Central Florida Research Foundation, Inc. Method for detection and diagnosis of lung and pancreatic cancers from imaging scans

Patent Citations (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN105550651A (en) * 2015-12-14 2016-05-04 中国科学院深圳先进技术研究院 Method and system for automatically analyzing panoramic image of digital pathological section
CN109492711A (en) * 2018-12-07 2019-03-19 杭州电子科技大学 Malignant mela noma and non-malignant melanotic nevus classification method based on deep learning
CN109670510A (en) * 2018-12-21 2019-04-23 万达信息股份有限公司 A kind of gastroscopic biopsy pathological data screening system and method based on deep learning
CN109754879A (en) * 2019-01-04 2019-05-14 湖南兰茜生物科技有限公司 A kind of lung cancer computer aided detection method and system based on deep learning
CN110472676A (en) * 2019-08-05 2019-11-19 首都医科大学附属北京朝阳医院 Stomach morning cancerous tissue image classification system based on deep neural network
CN110751172A (en) * 2019-09-10 2020-02-04 中南大学 Weakly supervised learning pathology full-image category inference method and system
CN111265234A (en) * 2020-03-05 2020-06-12 上海市肺科医院(上海市职业病防治院) Method and system for judging properties of lung mediastinal lymph nodes

Non-Patent Citations (6)

* Cited by examiner, † Cited by third party
Title
Constrained Deep Weak Supervision for Histopathology Image Segmentation;Zhipeng Jia等;IEEE Transactions on Medical Imaging;第36卷(第11期);第2376-2388页 *
Deep Generative Models for Weakly-Supervised Multi-Label Classification;Hong-Min Chu等;ECCV 2018: Computer Vision;第11206卷;第409-425页 *
一种基于树搜索的层次多标签乳腺疾病分类诊断方法;金程笑等;智能计算机与应用;第10卷(第2期);第34-39页 *
一种病理图像自动标注的机器学习方法;张钢等;《计算机研究与发展》;第52卷(第9期);第2135-2144页 *
基于弱监督学习的癌症病理图像分割;胡孔涛;中国优秀硕士学位论文全文数据库 医药卫生科技辑(第6期);第E072-31页 *
多核多示例学习的糖尿病视网膜病变诊断;任福龙等;中国图象图形学报(第4期);第90-101页 *

Also Published As

Publication number Publication date
CN111985536A (en) 2020-11-24

Similar Documents

Publication Publication Date Title
CN111985536B (en) Based on weak supervised learning gastroscopic pathology image Classification method
Li et al. A comprehensive review of computer-aided whole-slide image analysis: from datasets to feature extraction, segmentation, classification and detection approaches
CN110599448B (en) Migratory learning lung lesion tissue detection system based on MaskScoring R-CNN network
CN110120040B (en) Slice image processing method, slice image processing device, computer equipment and storage medium
CN112101451B (en) Breast cancer tissue pathological type classification method based on generation of antagonism network screening image block
Kowal et al. Computer-aided diagnosis of breast cancer based on fine needle biopsy microscopic images
WO2021139258A1 (en) Image recognition based cell recognition and counting method and apparatus, and computer device
Pan et al. Mitosis detection techniques in H&E stained breast cancer pathological images: A comprehensive review
CN111798425B (en) Intelligent detection method for mitotic image in gastrointestinal stromal tumor based on deep learning
Tang et al. Segnet-based gland segmentation from colon cancer histology images
CN111415352B (en) Cancer metastasis panoramic pathological section analysis method based on deep cascade network
CN110766670A (en) Mammary gland molybdenum target image tumor localization algorithm based on deep convolutional neural network
CN113034462B (en) Method and system for processing gastric cancer pathological section image based on graph convolution
CN112102332A (en) Cancer WSI segmentation method based on local classification neural network
CN112348059A (en) Deep learning-based method and system for classifying multiple dyeing pathological images
US20230005140A1 (en) Automated detection of tumors based on image processing
CN110021019B (en) AI-assisted hair thickness distribution analysis method for AGA clinical image
Ning et al. Multiscale context-cascaded ensemble framework (MsC 2 EF): application to breast histopathological image
CN112927215A (en) Automatic analysis method for digestive tract biopsy pathological section
CN115880266B (en) Intestinal polyp detection system and method based on deep learning
CN114782948A (en) Global interpretation method and system for cervical liquid-based cytology smear
CN116524315A (en) Mask R-CNN-based lung cancer pathological tissue section identification and segmentation method
CN109948706B (en) Micro-calcification cluster detection method combining deep learning and feature multi-scale fusion
CN111783571A (en) Cervical cell automatic classification model establishment and cervical cell automatic classification method
Lakshmanan Classification of Cancerous and Non Cancerous Cells in H & E Breast Cancer Images Using Structure Descriptors

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant