CN108564114B - Human body fecal leucocyte automatic identification method based on machine learning - Google Patents
Human body fecal leucocyte automatic identification method based on machine learning Download PDFInfo
- Publication number
- CN108564114B CN108564114B CN201810262889.9A CN201810262889A CN108564114B CN 108564114 B CN108564114 B CN 108564114B CN 201810262889 A CN201810262889 A CN 201810262889A CN 108564114 B CN108564114 B CN 108564114B
- Authority
- CN
- China
- Prior art keywords
- image
- area
- pixels
- value range
- binary image
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Expired - Fee Related
Links
- 238000000034 method Methods 0.000 title claims abstract description 49
- 238000010801 machine learning Methods 0.000 title claims abstract description 11
- 230000002550 fecal effect Effects 0.000 title claims description 9
- 210000000265 leukocyte Anatomy 0.000 claims abstract description 39
- 238000012545 processing Methods 0.000 claims abstract description 9
- 238000012549 training Methods 0.000 claims description 26
- 239000013598 vector Substances 0.000 claims description 21
- 238000012706 support-vector machine Methods 0.000 claims description 15
- 238000013145 classification model Methods 0.000 claims description 13
- 238000007781 pre-processing Methods 0.000 claims description 12
- 230000008569 process Effects 0.000 claims description 10
- 238000001514 detection method Methods 0.000 claims description 9
- 239000012535 impurity Substances 0.000 claims description 8
- 210000004027 cell Anatomy 0.000 claims description 6
- 230000006870 function Effects 0.000 claims description 6
- 230000000877 morphologic effect Effects 0.000 claims description 6
- 238000000513 principal component analysis Methods 0.000 claims description 6
- 238000004364 calculation method Methods 0.000 claims description 4
- 238000002790 cross-validation Methods 0.000 claims description 4
- 230000007797 corrosion Effects 0.000 claims description 3
- 238000005260 corrosion Methods 0.000 claims description 3
- 238000001000 micrograph Methods 0.000 claims description 3
- 238000010187 selection method Methods 0.000 claims description 3
- 238000003756 stirring Methods 0.000 claims description 3
- 238000012360 testing method Methods 0.000 claims description 3
- 230000010339 dilation Effects 0.000 claims description 2
- 238000007865 diluting Methods 0.000 claims description 2
- 230000001376 precipitating effect Effects 0.000 claims description 2
- 238000005070 sampling Methods 0.000 claims description 2
- 238000005516 engineering process Methods 0.000 abstract description 4
- 210000003608 fece Anatomy 0.000 description 4
- 210000000170 cell membrane Anatomy 0.000 description 2
- 230000007850 degeneration Effects 0.000 description 2
- 230000000717 retained effect Effects 0.000 description 2
- 210000004369 blood Anatomy 0.000 description 1
- 239000008280 blood Substances 0.000 description 1
- 230000001413 cellular effect Effects 0.000 description 1
- 238000013461 design Methods 0.000 description 1
- 238000010586 diagram Methods 0.000 description 1
- 238000010790 dilution Methods 0.000 description 1
- 239000012895 dilution Substances 0.000 description 1
- 239000012530 fluid Substances 0.000 description 1
- 239000010794 food waste Substances 0.000 description 1
- 210000003714 granulocyte Anatomy 0.000 description 1
- 208000035861 hematochezia Diseases 0.000 description 1
- 208000015181 infectious disease Diseases 0.000 description 1
- 230000004660 morphological change Effects 0.000 description 1
- 210000003097 mucus Anatomy 0.000 description 1
- 230000007935 neutral effect Effects 0.000 description 1
- 238000001556 precipitation Methods 0.000 description 1
- 238000003672 processing method Methods 0.000 description 1
- 238000012216 screening Methods 0.000 description 1
- 210000002700 urine Anatomy 0.000 description 1
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F18/00—Pattern recognition
- G06F18/20—Analysing
- G06F18/24—Classification techniques
- G06F18/241—Classification techniques relating to the classification model, e.g. parametric or non-parametric approaches
- G06F18/2411—Classification techniques relating to the classification model, e.g. parametric or non-parametric approaches based on the proximity to a decision surface, e.g. support vector machines
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F18/00—Pattern recognition
- G06F18/20—Analysing
- G06F18/21—Design or setup of recognition systems or techniques; Extraction of features in feature space; Blind source separation
- G06F18/214—Generating training patterns; Bootstrap methods, e.g. bagging or boosting
Landscapes
- Engineering & Computer Science (AREA)
- Data Mining & Analysis (AREA)
- Theoretical Computer Science (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Bioinformatics & Cheminformatics (AREA)
- Bioinformatics & Computational Biology (AREA)
- Artificial Intelligence (AREA)
- Evolutionary Biology (AREA)
- Evolutionary Computation (AREA)
- Physics & Mathematics (AREA)
- General Engineering & Computer Science (AREA)
- General Physics & Mathematics (AREA)
- Life Sciences & Earth Sciences (AREA)
- Investigating Or Analysing Biological Materials (AREA)
- Image Analysis (AREA)
Abstract
The invention discloses a human body excrement leukocyte automatic identification method based on machine learning, and relates to a method for automatically identifying and detecting leukocytes in human body excrement based on a digital image processing technology, in particular to a machine learning technology. The method has the advantages of accurate recognition, high speed, high efficiency and small calculated amount.
Description
Technical Field
The invention relates to a method for automatically identifying and detecting white blood cells in human excrement based on a digital image processing technology, in particular to a machine learning technology.
Background
Leukocytes in human feces are usually detected in mucus and purulent blood stool, and are mostly neutral paging nuclear granulocytes. Cells become more swollen and unclear due to their degeneration. If the amount is large, the cells are piled up and the cell membrane is incomplete or broken, which is also called as pus cell, indicating serious infection. Different from blood and urine of a human body, the excrement of the human body contains a large amount of impurities such as food residues and the like, so that the image background is complicated; part of the leukocyte cell membrane is not intact and adheres to part of the impurities. The above reasons all result in the difficult division of leukocytes. The morphology of leukocytes is variable due to cellular degeneration and is not easily recognized using morphology-based methods.
The applicant has filed a patent in 14 years for an automatic detection method (patent number 2014103696359) for white blood cells in a bronchoalveolar lavage smear, and the invention discloses that the white blood cells are finally identified by carrying out graying and binarization processing on a microscopic image of a bronchoalveolar lavage fluid smear collected by a microscope, simultaneously screening by utilizing the appearance characteristics and the internal characteristics of the white blood cells, and the white blood cells in a sample are detected by adopting a pure image processing method, so that the calculation amount is large in practical application.
Disclosure of Invention
The invention aims to design an automatic identification method for white blood cells in human excrement aiming at complex background of an excrement microscopic image and the morphological change diversity of the white blood cells.
The technical scheme provided by the invention is a human body fecal leucocyte automatic identification method based on machine learning, which comprises the following steps:
step 1: diluting, stirring and precipitating human excrement, and acquiring a microscopic image of a human excrement sample by using a biological microscope;
step 2: carrying out graying processing on the image in the step 1, and converting the image into a gray image;
and step 3: performing local reverse binarization processing on the gray level image in the step 2 to obtain a binarized image;
and 4, step 4: performing hole filling operation on the binary image in the step 3, and filling an area with an area smaller than S1, wherein the value range of S1 is 180-230 pixels;
and 5: performing morphological dilation operation on the binary image in the step 4, and adopting a circular template with the radius of R1; filling holes in the expanded binary image; performing morphological corrosion operation on the binary image after the hole is filled, and adopting a circular template with the radius of R2; the value range of R1 is 2-5 pixels, the value range of R2 is 1-3 pixels, and R1 is larger than R2;
step 6: marking the binary image connected region in the step 5, and calculating the area characteristic and the circumscribed rectangle of the connected region; the reserved area is larger than S2, and the area proportion of the reserved area to the circumscribed rectangle is larger than D1; the width and height of the circumscribed rectangle are larger than H1 and smaller than H2; the value range of S2 is 1400-1550 pixels, the value range of D1 is 60% -80%, the value range of H1 is 30-50 pixels, and the value range of H2 is 100-120 pixels;
and 7: cutting a binary image according to the circumscribed rectangle for the area reserved in the step 6; for the cut binary image, firstly, calculating the center and the radius of the maximum inscribed circle of the binary image, and only reserving the area of which the radius of the inscribed circle is larger than R3; further operating on the binary image, and only keeping points which are less than L1 away from the circle center; the value range of R3 is 15-20 pixels, and the value range of L1 is 20-30 pixels;
and 8: calculating the minimum circumcircle of the binary image in the step 7, and reserving the binary image of which the area of the binary region accounts for more than the area D2 of the circumcircle; the value range of D2 is 72-90%
And step 9: recalculating the binary image in the step 8 and correcting the coordinates of the circumscribed rectangle; reserving a region of which the width and the height of the circumscribed rectangle are larger than H3 and the difference of the absolute values of the width and the height is smaller than H4; the value range of H3 is 50-57 pixels, and the value range of H4 is 12-18 pixels;
step 10: cutting the corresponding gray level image and color image according to the newly calculated external rectangular coordinates of the area reserved in the step 9; calculating the variance, the gray mean value and the definition of the gray image; reserving an area with variance larger than F1, gray average larger than G1 and definition larger than Q1; the value range of F1 is 295-308, the value range of G1 is 55-70, and the value range of Q1 is 11-20;
the method comprises two processes of training a classifier and actually detecting;
step 11 to step 16 are classifier training processes; step 17-step 20 are actual detection processes;
step 11: comparing the area coordinates reserved in the step 10 with corresponding manual frame selection leukocyte area coordinates, removing real leukocytes, and intercepting the color images of the remaining areas as impurities of the negative sample set; the training positive sample set adopts white blood cells selected by manual frame; all the parameters adopted in the steps 4 to 10 can ensure that the white blood cells in all the samples enter the reserved area after the step 10;
step 12: normalizing all sample images in the sample sets, and then respectively extracting Gabor characteristics and LBP characteristics of the three color channels;
step 13: preprocessing the features extracted from the sample set in the step 12;
step 13-1: firstly, carrying out Principal Component Analysis (PCA) (principal Component analysis) on the feature vector, and only retaining 99% of principal components;
step 13-2: normalizing the characteristics obtained in the step 13-1 by 'L2';
step 13-3: standardizing the characteristics obtained in the step 13-2;
step 14: equally dividing the sample set into N parts by using a k-fold cross validation method, wherein N is more than 3, using one or two parts as a training set each time, and using the rest parts as a test set;
step 15: training the training set in the step 14 by using an SVM (support Vector machine) support Vector machine to obtain a classification model 1;
step 16: training a classification model 2 according to the classification model 1 obtained in the step 15;
and step 17: cutting the color image corresponding to the reserved area in the step 10, extracting and combining Gabor and LBP characteristics, and performing data preprocessing according to the method in the step 13 to obtain a characteristic vector to be detected;
step 18: inputting the feature vector to be detected in the step 17 into a classifier 1 to obtain a classification result;
step 19: extracting the HOG (histogram of organized gradient) features of the color image with the result of the leucocyte in the step 18, performing data preprocessing according to the method in the step 13, and sending the processed feature vector to the classifier 2 to obtain a final classification result;
step 20: all detected leukocytes in step 19 were marked on the color microscope image of step 1.
Further, the local area binarization method in the step 3 is that the local area is processed to be square, the side length is 11 pixels, the threshold value selection method is a gray average value, and the threshold value scaling coefficient is 0.98.
Further, the sharpness calculation in step 10 adopts a secondary blurring method, and specifically includes the following steps:
step 10-1: and calculating the gradient value of each pixel point of the gray image to obtain an original gradient map.
Step 10-2: the gray scale was gaussian blurred, the template size was 9 × 9, and the variance was taken to be 1.5.
Step 10-3: and (3) calculating the gradient value of each pixel point of the gray level image after the step (10-2) of blurring to obtain a blurring gradient image.
Step 10-4: and (4) calculating the average absolute difference value of the original gradient map obtained in the step 10-1 and the fuzzy gradient map obtained in the step 10-3 to obtain a definition value.
Further, the specific method of step 12 is as follows:
scaling the image to 80 × 80, and then extracting Gabor features and LBP features of three color channels respectively; wherein the Gabor characteristic scale is selected 15; the direction is selected from {0 degrees, 45 degrees, 90 degrees, 135 degrees, 180 degrees, 225 degrees, 270 degrees, 315 degrees }8 directions; the wavelength is pi/4; the spatial aspect ratio is 1.0; taking the standard deviation as 1.5; the phase offset is 0; simultaneously, 8 times of down sampling is adopted to reduce the characteristic dimension; wherein the LBP (local Binary Pattern) feature adopts an LBP equivalent mode (Uniform Pattern), and combines the Gabor feature and the LBP feature.
Further, in the step 15:
when an SVM support vector machine model is used for training, an RBF kernel function is selected, gamma is automatically selected, a penalty coefficient C is taken as 2, and the weight ratio of positive samples to negative samples is 2: 1;
the specific steps in the step 16 are as follows:
step 16-1: and taking all false negative samples of the model 1 in the step 15 as a negative sample set of the model 2. The positive sample set is the white blood cells of the manual frame selection after the scaling.
Step 16-2: HOG (histogram of Oriented gradient) histogram features are extracted from the gray scale map of the sample set. The Cell size is 8 × 8, and the gradient is in 9 directions; the size of Block is 2 x 2; the step size is taken to be 8. A 3600-dimensional feature vector is obtained.
Step 16-3: and (4) adopting a data preprocessing mode in the step 13 for the sample set characteristics to obtain 2320-dimensional characteristic vectors.
Step 16-4: and (3) obtaining a classification model 2 by adopting the same training method as the model 1 and using an SVM model, thereby achieving the purpose of further removing false detection impurities on the basis of keeping white blood cells. Finally, selecting an RBF kernel function, wherein gamma is selected automatically; taking 1 as a penalty coefficient C; the positive and negative sample weight ratio is 6:1, and the others are default parameters.
The method has the advantages of accurate recognition, high speed, high efficiency and small calculated amount.
Drawings
FIG. 1 is a flow chart of the detection of the automatic identification method for leukocytes in human feces.
Fig. 2 is a flow chart of a corresponding training classification model.
Fig. 3 is a diagram of the final automatic recognition result.
Detailed Description
A method for detecting human fecal leukocytes, the method comprising the steps of:
step 1: the human body feces are subjected to pretreatment operations such as dilution, stirring, precipitation and the like, and a biological microscope is used for collecting microscopic images of human body feces samples.
Step 2: and (3) carrying out gray processing on the image in the step (1) to convert the image into a gray image.
And step 3: and (3) carrying out local reverse binarization processing on the gray level image in the step (2) to obtain a binarized image. The area of the treatment part is square, and the radius is 11; the threshold value selection method is a gray average value, and the threshold value scaling coefficient is 0.98.
And 4, step 4: and (4) carrying out hole filling operation on the binary image in the step (3), and then removing the area smaller than 200.
And 5: and (4) performing morphological expansion operation on the binary image in the step (4), and adopting a circular template with the radius of 4. And filling holes in the expanded binary image. And (4) performing morphological corrosion operation on the binary image after the hole is filled, and adopting a circular template with the radius of 2.
Step 6: and (5) marking the binary image connected region in the step (5), and calculating the area characteristics and the circumscribed rectangle of the connected region. The reserved area is more than 1500, and the area proportion of the reserved area in the circumscribed rectangle is more than 70%; the width and height of the circumscribed rectangle are larger than 35 and smaller than 110.
And 7: and 6, cutting the area reserved in the step 6 into a binary image according to the circumscribed rectangle. For the cut binary image, firstly, the center and the radius of the maximum inscribed circle of the binary image are calculated, and only the area of which the radius of the inscribed circle is larger than 18 is reserved. And further operating on the binary image, and only keeping points which are less than 24 away from the circle center.
And 8: and (4) calculating the minimum circumcircle of the binary image in the step (7), and reserving the binary image of which the area of the binary region accounts for more than 84% of the area of the circumcircle.
And step 9: and (4) recalculating the binary image in the step (8) and correcting the coordinates of the circumscribed rectangle. And reserving an area of which the width and height of the circumscribed rectangle are more than 53 and the difference of the absolute values of the width and the height is less than 15.
Step 10: and (4) cutting the corresponding gray-scale image and color image according to the newly calculated external rectangular coordinates of the area reserved in the step (9). And calculating the variance, the gray mean and the definition of the gray map. Regions with variance greater than 300, mean of gray greater than 60, and sharpness greater than 15 are retained.
The following processes are classified into a training classifier and an actual detection. Step 11 to step 16 are classifier training processes. Step 17 to step 20 are actual detection processes.
Step 11: and (3) comparing the area coordinates reserved in the step (10) with the corresponding manual frame selection leukocyte area coordinates, removing real leukocytes, and intercepting the color images of the remaining areas as impurities of the negative sample set. The positive sample set used for training was white blood cells framed manually. All the parameters adopted in the steps 4 to 10 can ensure that the white blood cells in all the samples enter the reserved area after the step 10.
Step 12: for all sample sets, the images were scaled to 80 × 80 size, and then Gabor and LBP features were extracted for the three color channels, respectively. Wherein the Gabor characteristic dimension is selected 15; the direction is selected from {0 degrees, 45 degrees, 90 degrees, 135 degrees, 180 degrees, 225 degrees, 270 degrees, 315 degrees }8 directions; the wavelength is pi/4; the spatial aspect ratio is 1.0; taking the standard deviation as 1.5; the phase offset is 0. And meanwhile, 8 times of downsampling is adopted to reduce the characteristic dimension. A total of 2400-dimensional features are obtained. Wherein, the LBP (local Binary Pattern) feature adopts an LBP equivalent mode (Uniform Pattern) to extract 78-dimensional features. The Gabor features and LBP features are combined to obtain a 2478-dimensional feature vector.
Step 13: preprocessing the features extracted from the sample set in step 12 to obtain 1342-dimensional feature vectors.
Step 14: using the k-fold cross validation method, the sample set was divided equally into 5 parts, 1 of which was used as the training set and the remaining 4 as the test set.
Step 15: and (3) training the training set in the step 14 by using an SVM (support Vector machine) support Vector machine model to obtain a classification model 1. Finally selecting an RBF kernel function according to the result of the cross validation, wherein gamma is selected automatically; the penalty coefficient C takes the value of 2; the positive and negative sample weight ratio is 2:1, and the others are default parameters.
Step 16: based on the classification model 1 obtained in step 15, a classification model 2 is trained.
And step 17: and (3) cutting the color image corresponding to the reserved area in the step 10, extracting and combining Gabor and LBP characteristics, and performing data preprocessing according to the method in the step 13 to obtain a characteristic vector to be detected.
Step 18: and inputting the feature vector to be detected in the step 17 into a classifier 1 to obtain a classification result.
Step 19: extracting the HOG (histogram of organized gradient) features of the color image with the result of the white blood cells in the step 18, performing data preprocessing according to the method in the step 13, and sending the processed feature vectors into the classifier 2 to obtain a final classification result.
Step 20: all detected leukocytes in step 19 were marked on the color microscope image of step 1.
The definition calculation in the step 10 adopts a secondary fuzzy method, and the specific steps are as follows:
step 10-1: and calculating the gradient value of each pixel point of the gray image to obtain an original gradient image.
Step 10-2: the gray scale was gaussian blurred, the template size was 9 × 9, and the variance was taken to be 1.5.
Step 10-3: and (3) calculating the gradient value of each pixel point of the gray level image after the step (10-2) of blurring to obtain a blurring gradient image.
Step 10-4: and (4) calculating the average absolute difference value of the original gradient map obtained in the step 10-1 and the fuzzy gradient map obtained in the step 10-3 to obtain a definition value.
The specific steps in step 13 are as follows:
step 13-1: first, pca (principal Component analysis) principal Component analysis is performed on the 2478-dimensional feature vector, and only 99% of the principal components are retained to obtain a 1342-dimensional feature vector.
Step 13-2: the characteristics obtained in step 13-1 were normalized by 'L2'.
Step 13-3: the features obtained in step 13-2 are normalized.
The specific steps in the step 16 are as follows:
step 16-1: and taking all false negative samples of the model 1 in the step 15 as a negative sample set of the model 2. The positive sample set was a scaled manual frame for white blood cells.
Step 16-2: HOG (histogram of Oriented gradient) histogram features are extracted from the gray scale map of the sample set. The Cell size is 8 × 8, and the gradient is in 9 directions; the size of Block is 2 x 2; the step size is taken to be 8. A 3600-dimensional feature vector is obtained.
Step 16-3: and (3) preprocessing the sample set features in step 13 to obtain 2320-dimensional feature vectors.
Step 16-4: and (3) obtaining a classification model 2 by adopting the same training method as the model 1 and using an SVM model, thereby achieving the purpose of further removing false detection impurities on the basis of keeping white blood cells. Finally, selecting an RBF kernel function, wherein gamma is selected automatically; taking a penalty coefficient C as 1; the positive and negative sample weight ratio is 6:1, and the others are default parameters.
Claims (5)
1. A human body fecal leucocyte automatic identification method based on machine learning comprises the following steps:
step 1: diluting, stirring and precipitating human excrement, and acquiring a microscopic image of a human excrement sample by using a biological microscope;
step 2: carrying out graying processing on the image in the step 1 to convert the image into a grayscale image;
and step 3: performing local reverse binarization processing on the gray level image in the step 2 to obtain a binarized image;
and 4, step 4: performing hole filling operation on the binary image in the step 3, and filling an area with an area smaller than S1, wherein the value range of S1 is 180-230 pixels;
and 5: performing morphological dilation operation on the binary image in the step 4, and adopting a circular template with the radius of R1; filling holes in the expanded binary image; performing morphological corrosion operation on the binary image after the hole is filled, and adopting a circular template with the radius of R2; the value range of R1 is 2-5 pixels, the value range of R2 is 1-3 pixels, and R1 is larger than R2;
step 6: marking the binary image connected region in the step 5, and calculating the area characteristic and the circumscribed rectangle of the connected region; the reserved area is larger than S2, and the area ratio of the reserved area to the circumscribed rectangle is larger than D1; the width and height of the circumscribed rectangle are larger than H1 and smaller than H2; the value range of S2 is 1400-1550 pixels, the value range of D1 is 60% -80%, the value range of H1 is 30-50 pixels, and the value range of H2 is 100-120 pixels;
and 7: cutting a binary image according to the circumscribed rectangle for the area reserved in the step 6; for the cut binary image, firstly, calculating the center and the radius of the maximum inscribed circle of the binary image, and only reserving the area of which the radius of the inscribed circle is larger than R3; further operating on the binary image, and only keeping points which are less than L1 away from the circle center; the value range of R3 is 15-20 pixels, and the value range of L1 is 20-30 pixels;
and 8: calculating the minimum circumcircle of the binary image in the step 7, and reserving the binary image of which the area of the binary region accounts for more than the area D2 of the circumcircle; the value range of D2 is 72% -90%;
and step 9: recalculating the binary image in the step 8 and correcting the coordinates of the circumscribed rectangle; reserving a region of which the width and the height of the circumscribed rectangle are larger than H3 and the difference of the absolute values of the width and the height is smaller than H4; h3 ranges from 50 pixels to 57 pixels, and H4 ranges from 12 pixels to 18 pixels;
step 10: cutting the corresponding gray level image and color image according to the newly calculated external rectangular coordinates of the area reserved in the step 9; calculating the variance, the gray mean value and the definition of the gray image; reserving an area with variance larger than F1, gray average larger than G1 and definition larger than Q1; the value range of F1 is 295-308, the value range of G1 is 55-70, and the value range of Q1 is 11-20;
the method comprises two processes of training a classifier and actually detecting;
step 11 to step 16 are classifier training processes; step 17-step 20 are actual detection processes;
step 11: comparing the area coordinates reserved in the step 10 with corresponding manual frame selection leukocyte area coordinates, removing real leukocytes, and intercepting the color images of the remaining areas as impurities of the negative sample set; the training positive sample set adopts white blood cells selected by manual frame; all the parameters adopted in the steps 4 to 10 can ensure that the white blood cells in all the samples enter the reserved area after the step 10;
step 12: normalizing all sample images in the sample sets, and then respectively extracting Gabor characteristics and LBP characteristics of the three color channels;
step 13: preprocessing the features extracted from the sample set in the step 12;
step 13-1: firstly, PCA principal component analysis is carried out on the feature vector, and only 99% of principal components are reserved;
step 13-2: normalizing the characteristics obtained in the step 13-1 by 'L2';
step 13-3: standardizing the characteristics obtained in the step 13-2;
step 14: equally dividing the sample set into N parts by using a k-fold cross validation method, wherein N is more than 3, using one or two parts as a training set each time, and using the rest parts as a test set;
step 15: training the training set in the step 14 by using an SVM (support vector machine) model to obtain a classification model 1;
step 16: training a classification model 2 according to the classification model 1 obtained in the step 15;
step 16-1: taking all false-detected negative samples of the model 1 in the step 15 as a negative sample set of the model 2, and taking the positive sample set as the scaled artificial frame selection white blood cells;
step 16-2: extracting HOG direction gradient histogram features from a gray level image of a sample set; the size of the Cell is 8 × 8, the gradient is 9 directions, the size of the Block is 2 × 2, the step length is 8, and 3600-dimensional feature vectors are obtained;
step 16-3: adopting a data preprocessing mode in the step 13 for the sample set characteristics to obtain 2320-dimensional characteristic vectors;
step 16-4: a classification model 2 is obtained by adopting the same training method as the model 1 and using an SVM model, so that the aim of further removing false detection impurities on the basis of keeping white blood cells is fulfilled; finally, selecting an RBF kernel function, wherein gamma is selected automatically; taking 1 as a penalty coefficient C; the weight ratio of the positive sample to the negative sample is 6:1, and the other samples are default parameters;
and step 17: cutting the color image corresponding to the reserved area in the step 10, extracting and combining Gabor and LBP characteristics, and performing data preprocessing according to the method in the step 13 to obtain a characteristic vector to be detected;
step 18: inputting the feature vector to be detected in the step 17 into a classifier 1 to obtain a classification result;
step 19: extracting the HOG features of the color image with the result of the white blood cells in the step 18, preprocessing the data according to the method in the step 13, and sending the processed feature vectors into the classifier 2 to obtain a final classification result;
step 20: all detected leukocytes in step 19 were marked on the color microscope image of step 1.
2. The method for automatically identifying human fecal leucocytes based on machine learning as claimed in claim 1, wherein the method for binarizing the local region in step 3 is to process the local region as a square with a side length of 11 pixels, the threshold value selection method is a gray average value, and the threshold scaling factor is 0.98.
3. The method for automatically identifying human fecal leucocytes based on machine learning according to claim 1, wherein the sharpness calculation in the step 10 adopts a secondary fuzzy method, comprising the following steps:
step 10-1: calculating the gradient value of each pixel point of the gray level image to obtain an original gradient map;
step 10-2: performing Gaussian blur on the gray level image, wherein the size of the template is 9 x 9, and the variance is 1.5;
step 10-3: calculating the gradient value of each pixel point of the gray level image after the step 10-2 to obtain a fuzzy gradient map;
step 10-4: and (4) calculating the average absolute difference value of the original gradient map obtained in the step (10-1) and the fuzzy gradient map obtained in the step (10-3) to obtain a definition value.
4. The method for automatically identifying human fecal leucocytes based on machine learning as claimed in claim 1, wherein the specific method of the step 12 is as follows:
scaling the image to 80 × 80, and then extracting Gabor features and LBP features of three color channels respectively; wherein the Gabor characteristic scale is selected 15; the directions are selected from {0 °,45 °,90 °,135 °,180 °,225 °,270 °,315 ° }8 directions; the wavelength is pi/4; the spatial aspect ratio is 1.0; taking the standard deviation as 1.5; the phase offset is 0; simultaneously, 8 times of down sampling is adopted to reduce the characteristic dimension; wherein the LBP features employ an LBP equivalence mode, combining the Gabor features and the LBP features.
5. The method for automatically identifying human fecal leukocytes based on machine learning according to claim 1, wherein in the step 15:
when an SVM support vector machine model is used for training, an RBF kernel function is selected, gamma is automatically selected, a punishment coefficient C is taken as 2, and the weight ratio of positive samples to negative samples is 2: 1.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201810262889.9A CN108564114B (en) | 2018-03-28 | 2018-03-28 | Human body fecal leucocyte automatic identification method based on machine learning |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201810262889.9A CN108564114B (en) | 2018-03-28 | 2018-03-28 | Human body fecal leucocyte automatic identification method based on machine learning |
Publications (2)
Publication Number | Publication Date |
---|---|
CN108564114A CN108564114A (en) | 2018-09-21 |
CN108564114B true CN108564114B (en) | 2022-05-27 |
Family
ID=63533547
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN201810262889.9A Expired - Fee Related CN108564114B (en) | 2018-03-28 | 2018-03-28 | Human body fecal leucocyte automatic identification method based on machine learning |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN108564114B (en) |
Families Citing this family (8)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN109447119A (en) * | 2018-09-26 | 2019-03-08 | 电子科技大学 | Cast recognition methods in the arena with SVM is cut in a kind of combining form credit |
CN109919054B (en) * | 2019-02-25 | 2023-04-07 | 电子科技大学 | Machine vision-based reagent card automatic classification detection method |
CN109975196B (en) * | 2019-03-01 | 2021-10-08 | 深圳大学 | Reticulocyte detection method and system |
CN110070024B (en) * | 2019-04-16 | 2020-05-05 | 温州医科大学 | Method and system for identifying skin pressure injury thermal imaging image and mobile phone |
CN110070138B (en) * | 2019-04-26 | 2021-09-21 | 河南萱闱堂医疗信息科技有限公司 | Method for automatically scoring excrement picture before endoscope detection of colon |
CN110415212A (en) * | 2019-06-18 | 2019-11-05 | 平安科技(深圳)有限公司 | Abnormal cell detection method, device and computer readable storage medium |
CN112540039A (en) * | 2020-12-31 | 2021-03-23 | 北京博奥体质宝健康科技有限公司 | Method for directly calculating number of adherent living cells |
CN113989628B (en) * | 2021-10-27 | 2022-08-26 | 哈尔滨工程大学 | Underwater signal lamp positioning method based on weak direction gradient |
Citations (13)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN103154732A (en) * | 2010-08-05 | 2013-06-12 | 艾博特健康公司 | Method and apparatus for automated whole blood sample analyses from microscopy images |
CN103679184A (en) * | 2013-12-06 | 2014-03-26 | 河海大学 | Method for leukocyte automatic identification based on relevant vector machine |
CN104156951A (en) * | 2014-07-30 | 2014-11-19 | 电子科技大学 | Leukocyte detecting method aiming at bronchoalveolar lavage smear |
CN104198355A (en) * | 2014-07-16 | 2014-12-10 | 电子科技大学 | Automatic detection method for red cells in feces |
CN105404887A (en) * | 2015-07-05 | 2016-03-16 | 中国计量学院 | White blood count five-classification method based on random forest |
WO2016086023A1 (en) * | 2014-11-24 | 2016-06-02 | Massachusetts Institute Of Technology | Systems, apparatus, and methods for analyzing blood cell dynamics |
CN105654107A (en) * | 2015-09-21 | 2016-06-08 | 长春迪瑞医疗科技股份有限公司 | Visible component classification method based on SVM |
CN106295588A (en) * | 2016-08-17 | 2017-01-04 | 电子科技大学 | The automatic identifying method of leukocyte in a kind of leucorrhea micro-image |
CN106682633A (en) * | 2016-12-30 | 2017-05-17 | 四川沃文特生物技术有限公司 | Method for classifying and identifying visible components of microscopic excrement examination images based on machine vision |
CN106897682A (en) * | 2017-02-15 | 2017-06-27 | 电子科技大学 | Leucocyte automatic identifying method in a kind of leukorrhea based on convolutional neural networks |
CN106980839A (en) * | 2017-03-31 | 2017-07-25 | 宁波摩视光电科技有限公司 | A kind of method of automatic detection bacillus in leukorrhea based on HOG features |
WO2017145172A1 (en) * | 2016-02-23 | 2017-08-31 | Sigtuple Technologies Private Limited | System and method for extraction and analysis of samples under a microscope |
CN107408197A (en) * | 2015-03-11 | 2017-11-28 | 西门子公司 | The system and method for the classification of cell image and video based on deconvolution network |
-
2018
- 2018-03-28 CN CN201810262889.9A patent/CN108564114B/en not_active Expired - Fee Related
Patent Citations (13)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN103154732A (en) * | 2010-08-05 | 2013-06-12 | 艾博特健康公司 | Method and apparatus for automated whole blood sample analyses from microscopy images |
CN103679184A (en) * | 2013-12-06 | 2014-03-26 | 河海大学 | Method for leukocyte automatic identification based on relevant vector machine |
CN104198355A (en) * | 2014-07-16 | 2014-12-10 | 电子科技大学 | Automatic detection method for red cells in feces |
CN104156951A (en) * | 2014-07-30 | 2014-11-19 | 电子科技大学 | Leukocyte detecting method aiming at bronchoalveolar lavage smear |
WO2016086023A1 (en) * | 2014-11-24 | 2016-06-02 | Massachusetts Institute Of Technology | Systems, apparatus, and methods for analyzing blood cell dynamics |
CN107408197A (en) * | 2015-03-11 | 2017-11-28 | 西门子公司 | The system and method for the classification of cell image and video based on deconvolution network |
CN105404887A (en) * | 2015-07-05 | 2016-03-16 | 中国计量学院 | White blood count five-classification method based on random forest |
CN105654107A (en) * | 2015-09-21 | 2016-06-08 | 长春迪瑞医疗科技股份有限公司 | Visible component classification method based on SVM |
WO2017145172A1 (en) * | 2016-02-23 | 2017-08-31 | Sigtuple Technologies Private Limited | System and method for extraction and analysis of samples under a microscope |
CN106295588A (en) * | 2016-08-17 | 2017-01-04 | 电子科技大学 | The automatic identifying method of leukocyte in a kind of leucorrhea micro-image |
CN106682633A (en) * | 2016-12-30 | 2017-05-17 | 四川沃文特生物技术有限公司 | Method for classifying and identifying visible components of microscopic excrement examination images based on machine vision |
CN106897682A (en) * | 2017-02-15 | 2017-06-27 | 电子科技大学 | Leucocyte automatic identifying method in a kind of leukorrhea based on convolutional neural networks |
CN106980839A (en) * | 2017-03-31 | 2017-07-25 | 宁波摩视光电科技有限公司 | A kind of method of automatic detection bacillus in leukorrhea based on HOG features |
Non-Patent Citations (4)
Title |
---|
Acute leukemia classification by using SVM and K-Means clustering;Jakkrich Laosai等;《2014 International Electrical Engineering Congress (iEECON)》;20141016;第1-4页 * |
Analyzing Microscopic Images of Peripheral Blood Smear Using Deep Learning;Dheeraj Mundhra等;《DLMIA 2017, ML-CDS 2017: Deep Learning in Medical Image Analysis and Multimodal Learning for Clinical Decision Support》;20170909;第178-185页 * |
基于多空间混合属性融合的白细胞图像识别方法研究;郝连旺;《中国优秀博硕士学位论文全文数据库(博士)医药卫生科技辑》;20150115(第(2015)01期);第E059-26页 * |
白细胞图像的特征提取与分类算法研究;张敏淑;《中国优秀博硕士学位论文全文数据库(硕士)医药卫生科技辑》;20170415(第(2017)04期);第E060-37页 * |
Also Published As
Publication number | Publication date |
---|---|
CN108564114A (en) | 2018-09-21 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN108564114B (en) | Human body fecal leucocyte automatic identification method based on machine learning | |
CN107256558B (en) | Unsupervised automatic cervical cell image segmentation method and system | |
CN108647648A (en) | A kind of Ship Recognition system and method under visible light conditions based on convolutional neural networks | |
CN108090906B (en) | Cervical image processing method and device based on region nomination | |
CN111462076B (en) | Full-slice digital pathological image fuzzy region detection method and system | |
CN108765402B (en) | Non-woven fabric defect detection and classification method | |
CN111046880A (en) | Infrared target image segmentation method and system, electronic device and storage medium | |
CN108537099A (en) | A kind of licence plate recognition method of complex background | |
CN111598856A (en) | Chip surface defect automatic detection method and system based on defect-oriented multi-point positioning neural network | |
CN112215800A (en) | Machine learning-based overlapped chromosome recognition and segmentation method | |
CN111754538B (en) | Threshold segmentation method for USB surface defect detection | |
CN113256624A (en) | Continuous casting round billet defect detection method and device, electronic equipment and readable storage medium | |
CN110349125A (en) | A kind of LED chip open defect detection method and system based on machine vision | |
CN109003275A (en) | The dividing method of weld defect image | |
CN115294377A (en) | System and method for identifying road cracks | |
CN114627116B (en) | Fabric defect identification method and system based on artificial intelligence | |
CN113435460A (en) | Method for identifying brilliant particle limestone image | |
Theerapattanakul et al. | An efficient method for segmentation step of automated white blood cell classifications | |
CN102521582A (en) | Human upper body detection and splitting method applied to low-contrast video | |
CN103177244B (en) | Method for quickly detecting target organisms in underwater microscopic images | |
CN105354547A (en) | Pedestrian detection method in combination of texture and color features | |
CN112396580B (en) | Method for detecting defects of round part | |
CN116188786B (en) | Image segmentation system for hepatic duct and biliary tract calculus | |
Zhang et al. | Design of tire damage image recognition system based on deep learning | |
Danilescu et al. | Road anomalies detection using basic morphological algorithms |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
GR01 | Patent grant | ||
GR01 | Patent grant | ||
CF01 | Termination of patent right due to non-payment of annual fee |
Granted publication date: 20220527 |
|
CF01 | Termination of patent right due to non-payment of annual fee |