CN113312978A - Method and system for accurately identifying and segmenting target under microscopic image - Google Patents

Method and system for accurately identifying and segmenting target under microscopic image Download PDF

Info

Publication number
CN113312978A
CN113312978A CN202110482525.3A CN202110482525A CN113312978A CN 113312978 A CN113312978 A CN 113312978A CN 202110482525 A CN202110482525 A CN 202110482525A CN 113312978 A CN113312978 A CN 113312978A
Authority
CN
China
Prior art keywords
target
microscopic image
semantic segmentation
module
loss function
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN202110482525.3A
Other languages
Chinese (zh)
Inventor
肖立
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Institute of Computing Technology of CAS
Original Assignee
Institute of Computing Technology of CAS
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Institute of Computing Technology of CAS filed Critical Institute of Computing Technology of CAS
Priority to CN202110482525.3A priority Critical patent/CN113312978A/en
Publication of CN113312978A publication Critical patent/CN113312978A/en
Pending legal-status Critical Current

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V20/00Scenes; Scene-specific elements
    • G06V20/40Scenes; Scene-specific elements in video content
    • G06V20/41Higher-level, semantic clustering, classification or understanding of video scenes, e.g. detection, labelling or Markovian modelling of sport events or news items
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/24Classification techniques
    • G06F18/241Classification techniques relating to the classification model, e.g. parametric or non-parametric approaches
    • G06F18/2415Classification techniques relating to the classification model, e.g. parametric or non-parametric approaches based on parametric or probabilistic models, e.g. based on likelihood ratio or false acceptance rate versus a false rejection rate
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/10Segmentation; Edge detection
    • G06T7/11Region-based segmentation
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/10Image acquisition modality
    • G06T2207/10056Microscopic image
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/30Subject of image; Context of image processing
    • G06T2207/30004Biomedical image processing
    • G06T2207/30096Tumor; Lesion

Landscapes

  • Engineering & Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • Theoretical Computer Science (AREA)
  • General Physics & Mathematics (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Data Mining & Analysis (AREA)
  • General Engineering & Computer Science (AREA)
  • Bioinformatics & Computational Biology (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Evolutionary Biology (AREA)
  • Evolutionary Computation (AREA)
  • Artificial Intelligence (AREA)
  • Bioinformatics & Cheminformatics (AREA)
  • Probability & Statistics with Applications (AREA)
  • Computational Linguistics (AREA)
  • Software Systems (AREA)
  • Multimedia (AREA)
  • Image Analysis (AREA)

Abstract

The invention discloses a method and a system for accurately identifying and segmenting targets under microscopic images, wherein the method comprises the following steps: identifying a target in a microscopic image based on a Focal local target detection algorithm to obtain a target substance; performing semantic segmentation on the target substance based on a semantic segmentation algorithm, and acquiring an image area only containing the target substance from the microscopic image; and step three, visualizing the image area. The method carries out single target identification and high-quality reproduction in the microscopic image based on deep learning target detection and semantic segmentation, and solves the difficult problems of time and labor waste in manual microscopic image target identification and fine judgment.

Description

Method and system for accurately identifying and segmenting target under microscopic image
Technical Field
The invention relates to the technical fields of computer vision image processing, microscopic image analysis and the like, in particular to a method and a system for accurately identifying and segmenting a target under a microscopic image.
Background
Object recognition in microscopic images has important applications in biology and medicine. The existing microscopic image analysis mainly depends on manual judgment, and is time-consuming and labor-consuming. For example, karyotyping requires the identification and arrangement of all 46 chromosomes, while blood cell examination requires the identification and statistical averaging of all classes of bone marrow cells in multiple fields of view.
The manual analysis method is the most direct one, however, since a single field of view under a microscopic image usually has tens to hundreds of targets, each analysis usually needs to count a plurality of fields of view to achieve statistical accuracy, and is time-consuming and labor-consuming. Some subtle lesion features are difficult to find effectively because the targets tend to come together.
The prior art has the problems and needs to be solved urgently.
Disclosure of Invention
The invention aims to provide a method and a system for accurately identifying and segmenting a target under a microscopic image, which are used for identifying and reproducing a single target in the microscopic image with high quality based on deep learning target detection and semantic segmentation and solve the problem that manual microscopic image target identification and fine judgment are time-consuming and labor-consuming.
In order to achieve the above object, the present invention provides a method for accurately identifying and segmenting a target under a microscopic image, comprising:
identifying a target in a microscopic image based on a Focal local target detection algorithm to obtain a target substance;
performing semantic segmentation on the target substance based on a semantic segmentation algorithm, and acquiring an image area only containing the target substance from the microscopic image; and
and step three, visualizing the image area.
The method described above, wherein before the step one is executed, further comprising: and performing unified pretreatment on the microscopic image.
The method, wherein the uniform preprocessing is performed by randomly flipping the microscopic image at a probability of 50%, randomly rotating between-5 ° and 5 °, and randomly varying the brightness and contrast.
In the first step, the step of identifying the target in a microscopic image based on the Focal mass target detection algorithm includes:
constructing a network model for target area detection;
identifying the microscopic image according to the network model to obtain a candidate rectangular frame;
and obtaining a rectangular frame corresponding to the target substance according to the candidate rectangular frame.
In the second step, the semantic segmentation of the target substance based on the semantic segmentation algorithm includes:
constructing a semantic segmentation network by adding deep persistence into a standard Unet;
combining the cross entropy loss function with the bias-dice loss function to obtain a loss function;
and performing semantic segmentation on the target substance according to the semantic segmentation network and the loss function.
In order to achieve the above object, the present invention further provides a system for accurately identifying and segmenting a target under a microscopic image, comprising:
the target identification module is used for identifying a target in a microscopic image based on a Focal local target detection algorithm to obtain a target substance;
the semantic segmentation module is used for performing semantic segmentation on the target substance based on a semantic segmentation algorithm and acquiring an image area only containing the target substance from the microscopic image; and
and the visualization module is used for visualizing the image area.
The system, wherein, still include: and the preprocessing module is used for uniformly preprocessing the microscopic image.
The system of (1), wherein the pre-processing module processes the microscopic image using random flipping, random rotation between-5 ° and random shading and contrast variation at a probability of 50%.
The system of (a), wherein the object recognition module further comprises:
the model construction module is used for constructing a network model for target area detection;
the area prediction module is used for identifying the microscopic image according to the network model to obtain a candidate rectangular frame;
and the target area module is used for obtaining a rectangular frame corresponding to the target substance according to the candidate rectangular frame.
The system of (a), wherein the semantic segmentation module further comprises:
the segmentation network module is used for building a semantic segmentation network in a manner of adding deep persistence into the standard Unet;
the loss function module is used for combining the cross entropy loss function with the bias-dice loss function to obtain a loss function;
and the segmentation module is used for performing semantic segmentation on the target substance according to the semantic segmentation network and the loss function.
Compared with the prior art, the invention has the technical effects that:
1. the invention utilizes a target detection algorithm based on Focal local to position and judge the type of the target in the microscopic image, thereby solving the problem of influence of sample unbalance on the detection precision.
2. According to the invention, the image area corresponding to each rectangular frame is extracted, and the image area only containing the target substance part is segmented for visualization by utilizing a semantic segmentation algorithm, so that the operation of manual image matting by a doctor is omitted, and the time and the energy of the doctor are saved. And the single target presentation is more beneficial for a doctor to observe the target fine lesion, and the diagnosis precision is improved.
Drawings
FIG. 1 is a flow chart of a method for accurately identifying and segmenting a target under a microscopic image according to the present invention;
FIG. 2 is a diagram of the precise identification and segmentation of a target under a microscopic image according to the present invention;
FIG. 3 is a schematic diagram of a standard Unet incorporating the deep super vision architecture of the present invention;
FIG. 4 is a block diagram of a system for accurately identifying and segmenting objects under a microscopic image in accordance with the present invention.
Detailed Description
The invention is described in detail below with reference to the drawings and specific examples, but the invention is not limited thereto.
Fig. 1 is a flow chart of the method for accurately identifying and segmenting the target under the microscopic image according to the present invention.
Identification and fine segmentation in microscopic images is a common requirement in the study of biomedical images. If the research on the cell biological phenomenon needs to finely identify organelles and judge abnormal conditions, a single chromosome needs to be counted and finely divided to be presented in a report when the karyotype analysis is carried out, and the cell category under a single visual field needs to be counted when the bone marrow cell analysis is carried out in the blood examination.
These processes currently rely primarily on manual analysis and are inefficient. Through research, the defect can be realized by a deep learning target detection and analysis method. When a deep learning algorithm is investigated, the Focal local and a current main deep learning target detection model are well combined, and the influence of sample imbalance on detection precision is solved. The visual effect can be improved by adjusting the punishment degree of the target accuracy and the recall degree in the microscopic image in the design of the loss function of the segmentation model.
The method selects Focal local to be applied to a deep learning target detection framework, and ensures that the model can accurately identify the sample under the condition of unbalance. And applying Bce-bias-dice loss to make the model add punishment to the pixel points with missing prediction, thereby ensuring that the original information of the sample is kept on the visual map as much as possible.
Therefore, the microscopic image diagnosis method based on the target detection and semantic segmentation algorithm provided by the invention mainly aims to utilize the deep learning technology to help doctors, complete the complex work of detecting and segmenting the microscopic image target substance, and save the time and energy of the doctors. In the flow chart shown in fig. 1, the following steps are included:
step 10, identifying a target in a microscopic image based on a Focal local target detection algorithm to obtain a target substance;
in one embodiment, the step requires that the input microscopic image is uniformly preprocessed, and then the target in the microscopic image is accurately identified by using a target detection algorithm based on Focal loss. The target detection algorithm classifies the position and size of the target substance and determines the category to which each target substance belongs.
Step 20, performing semantic segmentation on the target substance based on a semantic segmentation algorithm, and acquiring an image area only containing the target substance from the microscopic image;
in one embodiment, after the target detection process is completed, the rectangular box corresponding to each target substance is detected. According to the invention, the image region corresponding to each rectangular frame is extracted, and the image region only containing the target substance part is segmented by utilizing a semantic segmentation algorithm for subsequent visualization, so that the operation of manual image matting by a doctor is omitted, and the time and the energy of the doctor are saved. And the single target presentation is more beneficial for a doctor to observe the target fine lesion, and the diagnosis precision is improved.
Step 30, visualizing the image area.
The method firstly completes the training of a target detection algorithm and a semantic segmentation algorithm, and then applies a trained model to predict the category of a target substance in a microscopic image and segment the target substance, as shown in FIG. 2.
The method for accurately identifying and segmenting the target under the microscopic image of the invention is further described by combining an embodiment with fig. 1 and 2, and mainly comprises the following four steps:
step 100, training data is preprocessed.
For all pictures, gaussian filtering is first used to filter out some of the noise in the picture, thereby reducing the error introduced by the image noise. For training pictures, the invention processes the pictures in a way of random inversion with a probability of 50%, random rotation between-5 degrees and random brightness and contrast change, thereby achieving the effect of data augmentation.
And 200, constructing a target detection model.
(201) VarifocalNet network
A VarifocalNet network, called VFNet for short, is a network structure for target region detection based on an FPN model. The model uses a VarifocalNet network to predict candidate rectangular boxes (bounding boxes) of the target region.
When the approximation degree of the candidate rectangular frame and the gt frame (ground-route), namely the prediction accuracy, is calculated, an Intersection over Union (IoU) is adopted, and the specific calculation method is as follows:
Figure BDA0003049821340000051
wherein IoU (A, B) indicates the overlapping degree of A and B, which reflects the accuracy of model prediction, A, B corresponds to the predicted candidate rectangular box region and gt box region, respectively.
The classification scores of the traditional target detection methods are often difficult to accurately measure the prediction quality. The confidence S of the candidate box that the prediction is accurate is low, so that it is not selected when Non-Maximum Suppression (Non-Maximum Suppression) is present.
The VarifocalNet network is adopted to combine the quality prediction of the target position with the classification score, and IoU-related classification score (IACS) is used as the detection score.
The Loss function adopted by the network training is variable local, and the Loss function is evolved according to the local. Focal local processes positive and negative samples are the same. The varifocalcanet network modifies the Focal local, and the expression form is as follows:
Figure BDA0003049821340000061
where VFL (p, q) is the Loss of the VarifocalNet network, p is the predicted IACS score, α is the weighted weight of the negative samples, and q is the IoU score of the target, i.e., q is IoU of gt box and predicted candidate rectangular box in positive samples, and q is 0 in negative samples, r is the attenuation coefficient of the Focal Loss, typically greater than 1. In the Varifocal local, p is used only for negative samplesrAnd attenuation is carried out, so that the supervision signals of the positive samples can be fully utilized, and the problem that the number of the positive samples in all the samples is small is effectively solved.
In the model, the candidate rectangular frame specifically adopts a representation form of a star-shaped bounding box (star-shaped bounding box) for predicting the target region. 9 fixed sampling points were used. Specifically, given a sampling point (x, y), x, y are coordinates of the sampling point, a 3 × 3 convolution regression is used to obtain a candidate rectangular box, a star bounding box is encoded as a four-dimensional vector (l ', t', r ', b') respectively representing distances from the sampling point to four edges, and then 9 (x, y), (x, y + b '), (x, y-t'), (x + r ', y), (x + r', y + b '), (x + r', y-t '), (x-l', y), (x-l ', y + b'), (x-l ', y-t') are heuristically selected. Because the points are obtained by manual setting and direct calculation, the prediction load is not required to be increased, and the calculation efficiency is high.
For the update of the star bounding box, it is a problem of residual learning. And learning scaling factors (delta l, delta t, delta r and delta b) of 4 distances in the training process, wherein the scaling factors (delta l, delta t, delta r and delta b) are used for changing the area of the initial star bounding box, and the specific updating is changed into a four-dimensional vector (l, t, r and b) ═ delta l × l ', delta t × t', delta r × r ', delta b × b'), which respectively represents the distances from the sampling point to the four edges, so that the updated star bounding box is closer to the gt box.
(202) Fast R-CNN network
The invention has compatibility to various target detection networks. The target detection network may also employ the Faster R-CNN network. The Faster R-CNN can be viewed as a combination of RPN and Fast R-CNN networks.
The Faster R-CNN Network uses an RPN (Region pro positive Network) to replace a Selective Search method to generate a rectangular frame, and the calculation speed is improved by 10 times. Each Feature point on a Feature map (Feature map) obtained after passing through a series of convolutional layers and pooling layers is mapped back to the central point of the receptive field in the original image as a reference point. A plurality of candidate boxes of different area and aspect ratios are selected based on each fiducial point. Wherein the area is selected from 3 kinds, respectively {128 }2,2562,5122}; the length-width ratios are 3, namely {1:1,1:2, 2: 1}. The two are combined with each other, and 9 candidate frames with different scales are preliminarily selected for each reference point.
The sampling of training data in the Faster R-CNN network was based on IoU scores. Dividing the samples into positive samples and negative samples, wherein the positive samples are selected as the samples with IoU values of the candidate rectangular box and the gt box being greater than 0.7, and if the samples with the highest score are not greater than 0.7, selecting the samples with the highest score as the positive samples; negative examples are candidate rectangular boxes with IoU values less than 0.3 for all gt boxes. Other samples are discarded.
Training the RPN network requires training classification tasks and regression tasks. The total loss is a linear weighted sum of two part loss functions.
The Loss function of the classification task takes the form of Focal local:
Figure BDA0003049821340000071
wherein L isclsRefers to the loss function of the classification task, i represents each template box (anchor), N refers to the total number of template boxes, alpha refers to the weighting weight, p refers to the total number of template boxesiRefers to the prediction probability, q, for the ith stencil maskiThe i-th template frame is a real label, and gamma is an attenuation coefficient.
The loss function of the regression task takes the form:
Figure BDA0003049821340000072
for any x, define
Figure BDA0003049821340000073
Wherein L isregRefers to the loss function of the regression task,
Figure BDA0003049821340000074
i-th coordinate value, v, of a predicted rectangular frame of the u-th class objectiThe ith coordinate value corresponding to the real target rectangular frame is indicated.
Loss of regression tasks using those commonly used in object detection models
Figure BDA0003049821340000075
Calculating a function, wherein x, y, w and h are respectively in the template frameCoordinates of the center point and the length and width, x, of the stencil frame*,y*,w*,h*The coordinate of the central point of the real target rectangular frame and the length width of the real target rectangular frame are respectively. In order to ensure the translation invariance and the length and width consistency of the coordinates, the coordinates are parameterized to generate a 4-dimensional vector t ═ tx,ty,tw,th}. The specific calculation method is as follows:
Figure BDA0003049821340000076
Figure BDA0003049821340000081
Figure BDA0003049821340000082
Figure BDA0003049821340000083
wherein, subscript a represents a stencil box (anchor), superscript denotes a real target rectangle box, tx,ty,tw,tyCoordinates representing the template frame subjected to the parameterization processing,
Figure BDA0003049821340000084
and coordinates representing the real target rectangular frame subjected to parameterization processing.
The complete loss function of the RPN network is a weighted sum of the two task loss functions, which is specifically expressed as follows:
Figure BDA0003049821340000085
wherein, L ({ p)i},{ti}) is a weighted sum of the sum,
Figure BDA0003049821340000086
in order to classify the loss function of the task,
Figure BDA0003049821340000087
as a loss function of the regression task, tiParameterize the processed coordinate vector for the ith stencil frame,
Figure BDA0003049821340000088
parameterizing a real target rectangular frame corresponding to the ith template frame to obtain a processed coordinate vector piThe predicted probability of the ith stencil box,
Figure BDA0003049821340000089
for the real label corresponding to the ith template frame (if the template frame is a positive example value, 1, otherwise 0), NclsFor the number of all stencil boxes, NregFor the number of all cases, { p }iRepresents the set of prediction probabilities for all stencil boxes, { t }iRepresents the set of predicted coordinate vectors, λ, for all stencil boxes1The weighting coefficient is defaulted to 1 in the present invention.
In a Fast R-CNN network part, a candidate rectangular frame obtained by the Fast R-CNN network through RPN network prediction is sampled to a feature with the same size, specifically a 7 x 7 feature map, through RoI Pooling Region of Interest Pooling (Region of Interest), and the feature map is flattened and then passes through a series of full-connection layers to obtain a prediction result. And performing classification prediction and regression prediction on the obtained feature vectors. The two parts are similar to the calculation method of the classification regression task of the RPN network. The specific loss function is expressed as follows:
L(p,u,tu,v)=Lcls(p,u)+λ2[u≥1]Lreg(tu,v)
wherein, L (p, u, t)uV) is a loss function, Lcls(p, u) is a loss function of the classification task, Lreg(tuV) is the loss function of the regression task, p is the softmax probability distribution predicted by the classifier, u is the true classification label of the corresponding target, tuCoordinate vector of corresponding category u predicted by regressor corresponding to prediction rectangular box
Figure BDA0003049821340000091
v is the coordinate vector corresponding to the real target rectangular box (vx,vy,vw,vh),λ2Is a weighting coefficient, and defaults to 1 in the present invention.
As the RPN network and the Fast R-CNN network both need to use the CNN network to extract image features, in the Fast R-CNN network, the two parts share one feature extraction network. Sharing CNN network part weights.
As described above with respect to step 200, the process of identifying the target in a microscopic image based on the Focal target detection algorithm is performed, and in the actual process of detecting the target, the variacalnet network mainly reduces the weight of the negative sample through the Focal distance, and is mainly used for detecting the single-type target in the present invention; for multi-class target detection, the invention adopts a fast R-CNN network based on Focal local, and the effect is better.
And step 300, constructing a semantic segmentation network.
The semantic segmentation network part of the invention is mainly improved on UNet and a new loss function is designed.
(301) UNet with deep Suvision
In the standard UNet, the model is divided into two parts, an encoder and a decoder. The encoder receives the original image and performs the convolution and downsampling operation layer by layer, thereby encoding the original image into a plurality of feature maps. And the decoder receives the result of the encoder, fuses and upsamples layer by layer, and finally obtains a segmentation result. With the codec model, the network can process multi-scale features and produce edge-smoothed segmentation results.
According to the method, a deep hierarchy structure is added in a standard UNet, the output of each layer of decoder is subjected to upsampling and convolution to form a predicted segmentation result, then the predicted segmentation result is subjected to loss calculation by a label, and loss values of all layers are added to obtain a final loss value. The structure is shown in figure 3.
In order to quantify the segmentation result, the invention uses a mean dice index to evaluate the result, and the calculation formula is as follows:
Figure BDA0003049821340000092
here A, B correspond to the predicted pixel region and the true labeled pixel region, respectively.
In the implementation process, the segmentation network receives pictures with the size of 512 × 512 as input, and performs convolution pooling on each downsampling layer to reduce the length and width of the feature map by half and double the number of channels, so as to gradually increase the receptive field. At each upsampling layer, the feature map transmitted by the encoder is combined and the convolution and bilinear interpolation are utilized, so that the length and the width of the feature map are doubled and gradually restored to the size of the original image. The dimensions of each layer feature map are as follows:
Name Shape
Input
512*512*3
Downsample1 256*256*128
Downsample2 128*128*256
Downsample3 64*64*512
Downsample4 32*32*1024
Upsample1 64*64*512
Upsample2 128*128*256
Upsample3 256*256*128
Output 512*512*N
wherein, the sizes of the feature layers corresponding to the dowsamples 1 to 4 after four times of down sampling, and the sizes of the feature layers corresponding to the upsamples 1 to 3 after four times of up sampling are shown in fig. 3.
Through the analysis of a plurality of different scale feature layers, the model can more effectively extract features of different scales, and the segmentation accuracy is improved.
(302) Bce-bias-dice loss function
The Bias-dice loss function is as follows:
Figure BDA0003049821340000101
in the Bias-dice loss function, piRepresenting the probability of prediction as class i, giGT is represented as class i, N is the total number of classes, beta is a number greater than 1 to increase the model's false negative penalty, and epsilon is a very small number to prevent the denominator from being zero. To avoid bias terms with scores of 0 above and below. Then p isigiIndicates True Positive (TP, True Positive), pi(1-gi) Indicates False Positive (FP, False Positive, g)i(1-pi) Representing False negatives (FN, False Negative), in the formula, by modulation of betaThe size can control the influence degree of FN on the whole Loss. As β increases, the FN contribution to overall Loss increases, thereby achieving the effect of reducing FN, increasing recall, and thus reducing the probability of missing pixels in the prediction.
Although the Bias-dice loss function can enable the model to retain more foreground information, the visualization effect is enhanced. However, the Bias-dice loss function deliberately reduces the penalty of false positive, and on the other hand, most of the microscopic image analysis has the background occupying the main body, which causes that the Bias-dice loss function alone will not predict the penalty of the background, which brings the disadvantages of unstable training and large fluctuation of the loss value. Aiming at the problem, the Bce-Bias-dice loss function is formed by combining the cross entropy loss function and the Bias-dice loss function, so that the problem that the Bias-dice loss function is not stable enough is solved, and the final loss function is as follows.
Figure BDA0003049821340000111
The Bce-bias-dice loss proposed by the present invention will be a function of the loss of UNet based on the deep super vision structure.
The technical effects of the method are shown in the following aspects:
1. the method utilizes a Focal local-based target detection algorithm to position and judge the type of the target in the microscopic image, and performs experiments on a Kaggle 2018data science bowl nucleus segmentation public data set which comprises 670 pictures. According to the following steps of 6: and 4, dividing the image into a training set and a testing set to obtain 402 images as training data and 268 images as testing data. The mAP reached 66.68 on the test set.
2. According to the invention, a deep learning semantic segmentation algorithm is utilized to perform semantic segmentation on a detected target, and an original image is cut to obtain 41363 subgraphs. According to the following steps of 6: 4, dividing the images into a training set and a testing set to obtain 29372 images as training data and 11991 images as testing data. Mean dice on test set reached 93.08.
Fig. 4 is a schematic diagram of a system for accurately identifying and segmenting an object under a microscopic image according to the present invention.
In fig. 4, a system 400 includes:
the target identification module 41 is used for identifying a target in a microscopic image based on a Focal local target detection algorithm to obtain a target substance;
the semantic segmentation module 42 is configured to perform semantic segmentation on the target substance based on a semantic segmentation algorithm, and acquire an image region only including the target substance from the microscopic image; and
a visualization module 43 for visualizing the image region.
Further, the system 400 further comprises: the preprocessing module 40 is used for uniformly preprocessing the microscopic image, namely processing the microscopic image in a random turning mode with a probability of 50%, a random rotation mode between-5 degrees and 5 degrees, and a random brightness and contrast change mode.
Further, the object recognition module 41 includes:
a model building module 411, configured to build a network model for target area detection;
the area prediction module 412 is configured to identify the microscopic image according to the network model to obtain a candidate rectangular frame;
and a target area module 413, configured to obtain a rectangular frame corresponding to the target substance according to the candidate rectangular frame.
The above-described object recognition in the embodiment of the precise recognition and segmentation of the object under the microscopic image is also applicable to the object recognition module 41.
Further, the semantic segmentation module 42 includes:
the segmentation network module 421 is configured to build a semantic segmentation network by adding deep persistence to the standard uet;
a loss function module 422, configured to combine the cross entropy loss function with the bias-dice loss function to obtain a loss function;
a segmentation module 423, configured to perform semantic segmentation on the target substance according to the semantic segmentation network and the loss function.
The semantic segmentation method in the above-described embodiment of accurately identifying and segmenting the target in the microscopic image is also applicable to the semantic segmentation module 42.
Compared with the prior art, the invention provides a method for accurately identifying and segmenting targets under a microscopic image, which is a microscopic image universal analysis method based on deep learning, can identify single targets in the microscopic image and reproduce high quality based on deep learning target detection and semantic segmentation, and solves the problem that time and labor are wasted when the microscopic image target identification and fine judgment are carried out manually.
The present invention may be embodied in other specific forms without departing from the spirit or essential attributes thereof, and it should be understood that various changes and modifications can be effected therein by one skilled in the art without departing from the spirit and scope of the invention as defined in the appended claims.

Claims (10)

1. A method for accurately identifying and segmenting a target under a microscopic image is characterized by comprising the following steps:
identifying a target in a microscopic image based on a Focal local target detection algorithm to obtain a target substance;
performing semantic segmentation on the target substance based on a semantic segmentation algorithm, and acquiring an image area only containing the target substance from the microscopic image; and
and step three, visualizing the image area.
2. The method of claim 1, further comprising, before said step one is performed: and performing unified pretreatment on the microscopic image.
3. The method of claim 2, wherein the uniform preprocessing is performed by using random flipping, random rotation between-5 ° and random shading and contrast variation on the microscopic image with a probability of 50%.
4. The method according to claim 1, wherein in the first step, the step of identifying the target in a microscopic image based on a Focal mass target detection algorithm comprises:
constructing a network model for target area detection;
identifying the microscopic image according to the network model to obtain a candidate rectangular frame;
and obtaining a rectangular frame corresponding to the target substance according to the candidate rectangular frame.
5. The method according to claim 1, wherein in the second step, the step of semantically segmenting the target substance based on the semantic segmentation algorithm comprises:
constructing a semantic segmentation network by adding deep persistence into a standard Unet;
combining the cross entropy loss function with the bias-dice loss function to obtain a loss function;
and performing semantic segmentation on the target substance according to the semantic segmentation network and the loss function.
6. A system for accurately identifying and segmenting a target under a microscopic image is characterized by comprising:
the target identification module is used for identifying a target in a microscopic image based on a Focal local target detection algorithm to obtain a target substance;
the semantic segmentation module is used for performing semantic segmentation on the target substance based on a semantic segmentation algorithm and acquiring an image area only containing the target substance from the microscopic image; and
and the visualization module is used for visualizing the image area.
7. The system of claim 6, further comprising:
and the preprocessing module is used for uniformly preprocessing the microscopic image.
8. The system of claim 7, wherein the pre-processing module processes the microscopic images using random flipping, random rotation between-5 ° and 5 °, and random shading and contrast variation with a probability of 50%.
9. The system of claim 6, wherein the object recognition module further comprises:
the model construction module is used for constructing a network model for target area detection;
the area prediction module is used for identifying the microscopic image according to the network model to obtain a candidate rectangular frame;
and the target area module is used for obtaining a rectangular frame corresponding to the target substance according to the candidate rectangular frame.
10. The system of claim 6, wherein the semantic segmentation module further comprises:
the segmentation network module is used for building a semantic segmentation network in a manner of adding deep persistence into the standard Unet;
the loss function module is used for combining the cross entropy loss function with the bias-dice loss function to obtain a loss function;
and the segmentation module is used for performing semantic segmentation on the target substance according to the semantic segmentation network and the loss function.
CN202110482525.3A 2021-04-30 2021-04-30 Method and system for accurately identifying and segmenting target under microscopic image Pending CN113312978A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202110482525.3A CN113312978A (en) 2021-04-30 2021-04-30 Method and system for accurately identifying and segmenting target under microscopic image

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202110482525.3A CN113312978A (en) 2021-04-30 2021-04-30 Method and system for accurately identifying and segmenting target under microscopic image

Publications (1)

Publication Number Publication Date
CN113312978A true CN113312978A (en) 2021-08-27

Family

ID=77371338

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202110482525.3A Pending CN113312978A (en) 2021-04-30 2021-04-30 Method and system for accurately identifying and segmenting target under microscopic image

Country Status (1)

Country Link
CN (1) CN113312978A (en)

Citations (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN110443808A (en) * 2019-07-04 2019-11-12 杭州深睿博联科技有限公司 Medical image processing method and device, equipment, storage medium for the detection of brain middle line
US20190385021A1 (en) * 2018-06-18 2019-12-19 Drvision Technologies Llc Optimal and efficient machine learning method for deep semantic segmentation
CN110874860A (en) * 2019-11-21 2020-03-10 哈尔滨工业大学 Target extraction method of symmetric supervision model based on mixed loss function
CN111402254A (en) * 2020-04-03 2020-07-10 杭州华卓信息科技有限公司 CT image pulmonary nodule high-performance automatic detection method and device
CN112116599A (en) * 2020-08-12 2020-12-22 南京理工大学 Sputum smear tubercle bacillus semantic segmentation method and system based on weak supervised learning
CN112200091A (en) * 2020-10-13 2021-01-08 深圳市悦动天下科技有限公司 Tongue region identification method and device and computer storage medium
CN112258488A (en) * 2020-10-29 2021-01-22 山西三友和智慧信息技术股份有限公司 Medical image focus segmentation method
CN112634261A (en) * 2020-12-30 2021-04-09 上海交通大学医学院附属瑞金医院 Stomach cancer focus detection method and device based on convolutional neural network

Patent Citations (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20190385021A1 (en) * 2018-06-18 2019-12-19 Drvision Technologies Llc Optimal and efficient machine learning method for deep semantic segmentation
CN110443808A (en) * 2019-07-04 2019-11-12 杭州深睿博联科技有限公司 Medical image processing method and device, equipment, storage medium for the detection of brain middle line
CN110874860A (en) * 2019-11-21 2020-03-10 哈尔滨工业大学 Target extraction method of symmetric supervision model based on mixed loss function
CN111402254A (en) * 2020-04-03 2020-07-10 杭州华卓信息科技有限公司 CT image pulmonary nodule high-performance automatic detection method and device
CN112116599A (en) * 2020-08-12 2020-12-22 南京理工大学 Sputum smear tubercle bacillus semantic segmentation method and system based on weak supervised learning
CN112200091A (en) * 2020-10-13 2021-01-08 深圳市悦动天下科技有限公司 Tongue region identification method and device and computer storage medium
CN112258488A (en) * 2020-10-29 2021-01-22 山西三友和智慧信息技术股份有限公司 Medical image focus segmentation method
CN112634261A (en) * 2020-12-30 2021-04-09 上海交通大学医学院附属瑞金医院 Stomach cancer focus detection method and device based on convolutional neural network

Non-Patent Citations (4)

* Cited by examiner, † Cited by third party
Title
HAOYANG ZHANG 等: "VarifocalNet: An IoU-aware Dense Object Detector", ARXIV *
YIFEI XU 等: "FFU-Net: Feature Fusion U-Net for Lesion Segmentation of Diabetic Retinopathy", 《BIOMED RESEARCH INTERNATIONAL》 *
胡满满 等: "基于动态采样和迁移学习的疾病预测模型", 《计算机学报》 *
高智勇等: "基于特征金字塔网络的肺结节检测", 《计算机应用》 *

Similar Documents

Publication Publication Date Title
CN110210463B (en) Precise ROI-fast R-CNN-based radar target image detection method
CN110428432B (en) Deep neural network algorithm for automatically segmenting colon gland image
CN108364288B (en) Segmentation method and device for breast cancer pathological image
CN111524138B (en) Microscopic image cell identification method and device based on multitask learning
CN109389129B (en) Image processing method, electronic device and storage medium
CN111524137B (en) Cell identification counting method and device based on image identification and computer equipment
CN111462076B (en) Full-slice digital pathological image fuzzy region detection method and system
Ismael et al. Medical image classification using different machine learning algorithms
CN112819821B (en) Cell nucleus image detection method
CN110796661B (en) Fungal microscopic image segmentation detection method and system based on convolutional neural network
CN110135271A (en) A kind of cell sorting method and device
CN113096096B (en) Microscopic image bone marrow cell counting method and system fusing morphological characteristics
WO2022167005A1 (en) Deep neural network-based method for detecting living cell morphology, and related product
CN111951288A (en) Skin cancer lesion segmentation method based on deep learning
CN112365497A (en) High-speed target detection method and system based on Trident Net and Cascade-RCNN structures
CN112348059A (en) Deep learning-based method and system for classifying multiple dyeing pathological images
CN113902945A (en) Multi-modal breast magnetic resonance image classification method and system
CN114445356A (en) Multi-resolution-based full-field pathological section image tumor rapid positioning method
CN114783604A (en) Method, system and storage medium for predicting sentinel lymph node metastasis of breast cancer
CN115909006A (en) Mammary tissue image classification method and system based on convolution Transformer
CN114332473A (en) Object detection method, object detection device, computer equipment, storage medium and program product
WO2023220913A1 (en) Cell image processing method, electronic device and storage medium
CN113312978A (en) Method and system for accurately identifying and segmenting target under microscopic image
CN115131628A (en) Mammary gland image classification method and equipment based on typing auxiliary information
CN114897823A (en) Cytology sample image quality control method, system and storage medium

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination