CN110837836B - Semi-supervised semantic segmentation method based on maximized confidence - Google Patents

Semi-supervised semantic segmentation method based on maximized confidence Download PDF

Info

Publication number
CN110837836B
CN110837836B CN201911071629.4A CN201911071629A CN110837836B CN 110837836 B CN110837836 B CN 110837836B CN 201911071629 A CN201911071629 A CN 201911071629A CN 110837836 B CN110837836 B CN 110837836B
Authority
CN
China
Prior art keywords
loss
image
class probability
network
segmentation
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN201911071629.4A
Other languages
Chinese (zh)
Other versions
CN110837836A (en
Inventor
金�一
陈怀安
陈林
陈宇旋
竺长安
陈恩红
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Jin Yi
Zhu Changan
University of Science and Technology of China USTC
Original Assignee
University of Science and Technology of China USTC
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by University of Science and Technology of China USTC filed Critical University of Science and Technology of China USTC
Priority to CN201911071629.4A priority Critical patent/CN110837836B/en
Publication of CN110837836A publication Critical patent/CN110837836A/en
Application granted granted Critical
Publication of CN110837836B publication Critical patent/CN110837836B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/20Image preprocessing
    • G06V10/26Segmentation of patterns in the image field; Cutting or merging of image elements to establish the pattern region, e.g. clustering-based techniques; Detection of occlusion
    • G06V10/267Segmentation of patterns in the image field; Cutting or merging of image elements to establish the pattern region, e.g. clustering-based techniques; Detection of occlusion by performing operations on regions, e.g. growing, shrinking or watersheds
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/045Combinations of networks

Abstract

The invention discloses a semi-supervised semantic segmentation method based on maximized confidence, which comprises the following steps: selecting a part of images from the existing training data set as marked images, and using the rest images as unmarked images; constructing a network model, and predicting a prediction class probability map of a marked image and an unmarked image through a segmentation network in the network model; adopting a mode of supervised learning and generation countermeasure to maximize the confidence of the labeled image prediction class probability map; predicting a segmentation error region in the unmarked image prediction class probability map by adopting an unsupervised learning mode; training the network model by combining the loss of supervised learning and the loss of unsupervised learning; and in the testing stage, inputting the unmarked image to be segmented into the trained network model to obtain the segmented semantic image. According to the scheme of the embodiment of the invention, the unmarked image can be accurately subjected to semantic segmentation.

Description

Semi-supervised semantic segmentation method based on maximized confidence
Technical Field
The invention relates to the field of image semantic segmentation, in particular to a semi-supervised semantic segmentation method based on maximized confidence.
Background
The image segmentation means that an image is divided into a plurality of mutually disjoint areas according to characteristics such as gray scale, color, spatial texture, geometric shape and the like, so that the characteristics show consistency or similarity in the same area and obviously differ among different areas. In a simple way, in an image, different objects are separated from the background, and it is clear from the segmentation result what object is segmented. Overall, semantic segmentation is a difficult task aimed at scene understanding. Scene understanding is widely applied in the information society of today as a core problem of computer vision. These applications include: autopilot, human-computer interaction, computer photography, image search engines, and augmented reality. These problems have been attempted to be solved using a variety of computer vision and machine learning methods.
Recently, methods using convolutional neural networks have achieved the most advanced performance in image semantic segmentation. These methods extract features of the neural network from a model trained on large-scale pixel-level annotated datasets. For example, PSPNet (pyramid scene analysis network), FCN (full Convolutional neural network), and the like. However, annotating accurate pixel-level tags on large-scale data is very time consuming, labor intensive, and inefficient. To reduce the need to construct an accurate pixel-level annotation data set, the unsupervised learning approach seems to be a more suitable approach. However, to date, unsupervised learning approaches have not been successful due to the lack of detailed information about the semantic segmentation task. Therefore, weakly supervised and semi-supervised learning methods have also been proposed for semantic segmentation. These methods typically use unmarked or weakly marked data, and sometimes they also use additional fully annotated data to improve performance. Weakly labeled images may be partially annotated, but all may be annotated with some limited area, such as image-level annotations, box annotations, graffiti annotations, and so forth. However, this approach also has its non-negligible drawbacks, such as:
1) due to the lack of detailed boundary location information, the weakly supervised approach performs far less well than the fully supervised approach.
2) Some semi-supervised learning methods are inefficient in using unlabeled data because they ignore the large amount of available misclassification information.
Disclosure of Invention
The invention aims to provide a semi-supervised semantic segmentation method based on maximized confidence coefficient, which can accurately perform semantic segmentation on an unmarked image.
The purpose of the invention is realized by the following technical scheme:
a semi-supervised semantic segmentation method based on maximized confidence includes:
constructing a training data set by using the marked images and the unmarked images in a specified proportion;
constructing a network model, and predicting a prediction class probability map of a marked image and an unmarked image through a segmentation network in the network model; adopting a supervised learning mode to maximize the confidence of the labeled image prediction class probability map; predicting a segmentation error region in the unmarked image prediction class probability map by adopting an unsupervised learning mode;
training the network model by combining the loss of supervised learning and the loss of unsupervised learning to obtain a trained segmented network;
and in the testing stage, inputting the unmarked image to be segmented into the trained segmentation network model, obtaining a predicted class probability map, and searching the index of the maximum value in the channel dimension in the predicted class probability map to obtain a segmented semantic image.
According to the technical scheme provided by the invention, the method improves the accuracy of semantic segmentation from the perspective of enhancing the confidence degree of the class probability map and paying attention to the wrongly classified area, and researches the data distribution of unmarked data through a segmentation network so as to generate a more reliable prediction result for the unmarked image.
Drawings
In order to more clearly illustrate the technical solutions of the embodiments of the present invention, the drawings required to be used in the description of the embodiments are briefly introduced below, and it is obvious that the drawings in the description below are only some embodiments of the present invention, and it is obvious for those skilled in the art that other drawings can be obtained according to the drawings without creative efforts.
FIG. 1 is a schematic diagram illustrating comparison of information entropy between a validation set and a training set according to an embodiment of the present invention;
FIG. 2 is a flowchart of a semi-supervised semantic segmentation method based on maximized confidence level according to an embodiment of the present invention;
FIG. 3 is a schematic diagram of a network model according to an embodiment of the present invention;
FIG. 4 is a schematic diagram of a tag error map and a predicted segmentation error map according to an embodiment of the present invention;
fig. 5 is a schematic representation of the performance of the protocols participating in the comparative experiment on the PASCAL VOC 2012 validation set provided by the example of the present invention;
fig. 6 is a schematic diagram of a performance result of the scheme participating in the comparative experiment on the PASCAL-CONTEXT validation set according to the embodiment of the present invention.
Detailed Description
The technical solutions in the embodiments of the present invention are clearly and completely described below with reference to the drawings in the embodiments of the present invention, and it is obvious that the described embodiments are only a part of the embodiments of the present invention, and not all embodiments. All other embodiments, which can be derived by a person skilled in the art from the embodiments of the present invention without making any creative effort, shall fall within the protection scope of the present invention.
The embodiment of the invention provides a semi-supervised semantic segmentation method based on maximized confidence coefficient, which provides a semi-supervised learning framework, combines supervised learning and unsupervised learning, and solves the problem from the perspective of enhancing the confidence coefficient of a class probability map. At the same time, the regions of misclassification, in particular in the border regions, are of greater interest. Also, the data distribution of the unlabeled data is studied by the segmentation network to produce a more reliable prediction of the unlabeled image.
In the embodiment of the invention, a generation countermeasure framework is constructed for the marked image. The segmentation network is considered as a generator that takes the image as input and outputs a prediction class probability map. The recognizer is constructed in a full convolution mode and is used for distinguishing whether the input is from a prediction class probability map of the prediction marked image or a class probability map consisting of 0 and 1 generated by a label map; the generator and recognizer compete with each other with the goal of maximizing the confidence of the prediction class probability map (i.e., the confidence of the segmented network). For unlabeled data, segmentation networks trained using labeled images achieve high confidence for correctly classified pixels with the help of resistant learning. Therefore, classified pixels with high uncertainty are considered segmentation error pixels. Next, an entropy of information representing the segmentation probability map is calculated to infer a segmentation error map. When the information entropy of a pixel is maximized, its prediction class probability approximates a uniform probability distribution, indicating that the feature under study cannot classify the pixel, and the weights of the model should be optimized to obtain a more representative feature.
Part (a) in fig. 1 shows the entropy of information on the verification set and part (b) in fig. 1 shows the entropy of information on the training set, and it is apparent that the entropy of information on the verification set is greater than the entropy of information on the training set, which indicates that the segmented network is less reliable when the image is predicted without prior training, especially in the boundary region. In this work, the mean entropy of information in the misclassified region of unlabeled data is calculated and used as an additional supervised learning signal to optimize the segmented network. Therefore, the present invention focuses more on misclassified regions, especially in the border regions. Segmentation networks study the data distribution of unlabeled data to produce more reliable predictions for the unlabeled images.
As shown in fig. 2, it is a flowchart of a semi-supervised semantic segmentation method based on maximized confidence provided in the embodiment of the present invention; it mainly comprises:
step 1, constructing a training data set by using marked images and unmarked images in a specified proportion.
Typically, the training data set is constructed using a small number of labeled images and a large number of unlabeled images, and the images used may be from an existing training data set. The specific proportion of the marked image to the unmarked image can be set by the user according to the actual situation.
Illustratively, a more challenging data set may be chosen: PASCAL VOC 2012 and pascalontext. The PASCAL VOC 2012 data set includes 20 foreground object classes and a background class that contains 1464, 1456 and 1449 pixel-level annotation images for training, testing and validation, respectively, and additionally, 10582 training images are obtained using additional annotation images from the segmentation boundary data Set (SBD) for the enhancement data set. The PASCAL-CONTEXT dataset provides detailed pixel-level annotations on two objects (e.g., cars) and filler (e.g., sky), and the present invention evaluates the 59 most frequent classes and one background class in the dataset, resulting in 4998 training images. Finally, 10%, 30% and 50% of the images were randomly drawn from the training data set as labeled images, and the remaining data was used for unlabeled images. It should be noted that the data sets and the ratio of marked images to unmarked images mentioned herein are examples and are not intended to be limiting.
Step 2, constructing a network model, and predicting a prediction class probability map of a marked image and an unmarked image through a segmentation network in the network model; adopting a supervised learning mode to maximize the confidence of the labeled image prediction class probability map; and predicting the segmentation error region in the unmarked image prediction class probability map by adopting an unsupervised learning mode.
As shown in fig. 3, a schematic diagram of a network model structure is constructed. The network model is divided into a main network and a sub-network. The master network is a segmented network, i.e. a generator in a generating confrontation network, whose inputs are labeled images and unlabeled images and whose output is a prediction class probability map. The sub-network is a fully-convoluted recognizer, i.e. a recognizer in the generation countermeasure network, which takes as input a prediction class probability map output by the segmentation network or a class probability map consisting of 0 and 1 generated by the label map, and outputs a two-channel class probability map to distinguish whether the input is from the prediction class probability map or the class probability map consisting of 0 and 1 generated by the label map.
The following description is directed to a split network and a full convolution neural network.
1) Network partitioning: in the present example, the Deeplab-v2 model pre-trained on the MSCOCO and ImageNet datasets served as the baseline network. However, to simplify the experiment and reduce memory consumption, the Conditional Random Field (CRF) and the multi-scale input to maximize fusion are not used, and only the ASPP output layer (the aperture Spatial Pyramid bulking space Pyramid Pooling layer) is preserved. To match the size of the input image, an upsampling layer and a Softmax function are applied to predict the final class probability map.
2) Full convolution neural network: in the embodiment of the invention, the input of the full convolution neural network is two, one is a class probability map generated after the label map is subjected to down-sampling and Onehot coding, and the other is a prediction class probability map generated after the marked image is subjected to a segmentation network and Softmax. The full convolution neural network as the identifier is composed of 5 expansion convolution layers with step size 1 containing 3 x 3 convolution kernel and {64, 128, 256, 512, 2} channels, and the expansion ratio is set to {1,1,2,4,1} in each layer respectively. In addition, each of the swelling convolutional layers, except the last layer, is followed by a ReLU activation function.
It should be noted that the structural forms and the values of the relevant parameters mentioned in the above descriptions for the segmented network and the full convolutional neural network are only examples and are not limiting.
The following describes a supervised learning mode, an unsupervised learning mode and a related loss function.
1) There are ways of supervised learning and associated loss functions.
Supervised learning has two main goals: the first is the basic task of assigning semantic labels to each pixel, and the second is to maximize the confidence of the prediction class probability map using a way of generating countermeasures. To this end, a generative confrontation framework is constructed, wherein the generator is a segmentation network and the recognizer is a full convolution neural network. In the generation of the countermeasure network, the segmentation network is used as a generator to predict the prediction class probability map of the marked image; the full convolution neural network is used as a recognizer, the input of the full convolution neural network is a labeled image prediction class probability graph, and a class probability graph consisting of 0 and 1 generated after downsampling and onehot coding (one-hot coding) of a label graph, and the type of the input is recognized through the recognizer; the generator and recognizer compete against each other with the goal of maximizing the confidence of the prediction class probability map.
In the generator network, spatially multi-class cross entropy penalties are used to force the segmentation network to independently predict the correct semantic label class at each pixel location, expressed as:
Figure GDA0003698490930000051
wherein x is n For the marked image input to the segmented network, y n Encoding the label map for onehot of the corresponding labeled image, (h) 1 ,w 1 ,c 1 ) The predicted class probability map is of size H for the position coordinates of the pixels in the map 1 ×W 1 ×C 1 ,H 1 、W 1 Respectively representing the height, width, C of the image 1 Represents the number of categories (channels); s (x) n ) Tagged image x predicted for segmented networks n The prediction class probability map of (1).
In the recognizer, a spatial binary class entropy penalty is used to distinguish whether the input is a predicted labeled image prediction class probability map or a class probability map generated from a label map, the spatial binary class entropy penalty being expressed as:
Figure GDA0003698490930000052
Y n =one_hot(ones(H 2 ,W 2 )×SG)
wherein p is n A labeled image prediction class probability map representing a prediction or a class probability map generated from a label map, D (-) representing a recognizer, Y · n Is a comment for distinguishing the input source, C 2 2 because the recognizer is a binary classification network; one _ hot (·) is the onehot encoding function, ons (H) 2 ,W 2 ) For generating a size of H 2 ×W 2 Matrix of (H) 2 、W 2 Respectively representing the row number and the column number of the matrix, wherein the values of all elements are 1; SG is 0, and represents a prediction class probability map of the marked image input as prediction by the recognizer; SG is 1, and represents that the recognizer inputs are class probability maps composed of 0 and 1 generated by the label map; the above-mentioned entropy loss of the spatial binary class is mainly used to train the recognizer.
Adding an antagonistic loss to the split network facilitates it to increase the predicted class probability to near 1. The resistance loss can be written as follows:
Figure GDA0003698490930000061
in embodiments of the invention, loss is calculated when the input is from a split network adv . In addition, to confuse the recognition network, SG is set to 1.
2) Unsupervised learning mode and associated loss function.
The information entropy of the unlabeled image class probability map characterizes the uncertainty of the segmentation result of the image, which is closely related to the segmentation error map of the image. Therefore, the invention uses the information entropy of the predicted class probability map to infer the segmentation error map, and FIG. 4 shows the labeled error map and the predicted segmentation error map. Wrongly classified pixels are mainly located around the boundary, which means that the segmentation error map contains rich classification information, especially in the boundary region. Fig. 4 shows an original image in part (a), an error map label in part (b), and a predicted segmentation error map in part (c); the artwork in fig. 4 is from the data set PASCAL VOC 2012. After obtaining the segmentation error map, the average entropy of information in the error classification region is calculated as unsupervised loss.
Given size H 1 ×W 1 X 3 unmarked image x n ', segmentation of network predicted unmarked image x n ' the prediction class probability map is S (x) n ') and an information entropy diagram H (x) is calculated in the following manner n '):
Figure GDA0003698490930000062
Wherein E [. C]Represents to all C 1 (ii) a desire for a category;
the information entropy indicates the uncertainty of the prediction of the segmentation network, given an uncertainty threshold T, a binary map is obtained representing the segmentation error map EM (x) n ') expressed as:
Figure GDA0003698490930000071
wherein the content of the first and second substances,
Figure GDA0003698490930000072
representing (h) in an information entropy diagram 1 ,w 1 ) The pixel point at the position indicates EM (x) n ') is obtained by determining whether each pixel value of the information entropy map is greater than a threshold value.
By means of an information entropy diagram H (x) n ') and a segmentation error map EM (x) n ') get unsupervised losses and feed back to the split network, the unsupervised losses are expressed as:
Figure GDA0003698490930000073
in the embodiment of the invention, a mixing loss function is used, and the mixing loss function combines spatial multi-class entropy loss, antagonism loss and unsupervised loss. The amount of mixing loss was calculated as follows:
loss seg =loss mceadv loss advinf loss inf
therein, loss mce ,loss adv And loss inf Respectively representing spatial multi-class entropy loss, antagonism loss and unsupervised loss of a maximized unmarked image prediction class probability graph; lambda [ alpha ] adv And λ inf Are two weights that balance the corresponding losses. loss mce And loss adv For guiding supervised learning, but loss inf Is used as an unsupervised learning signal to study the data distribution of the unlabeled images.
And 3, training the network model by combining the loss of supervised learning and the loss of unsupervised learning to obtain a trained segmented network.
The marked images and the unmarked images are combined together according to the batch size as input, various hyper-parameters of a network model are set, a weight initialization mode is set, a supervised learning loss and an unsupervised learning loss are combined, a segmentation network is trained by using a Stochastic Gradient Descent (SGD) method and a polygon learning rate strategy, a recognizer is trained by using an Adam optimizer and an exponential decay learning rate strategy, and the trained model weight is stored.
By way of example, some specific settings for network model training are given below:
the proposed network is implemented by a sensor-Flow framework running on a GPU (Tesla V100). The training images obtained in the step 1 are randomly scaled and cut to 321 × 321 pixel Size, and the training models are combined together as input according to the Batch Size (Batch Size) of 10 for 20K iterations. With respect to the hyper-parameter of the proposed method, λ adv Is set to 0.02 and lambda inf Set to 0.1. Further, the threshold T for obtaining the segmentation error map is set to 0.2.
In training the segmented network, Stochastic Gradient Descent (SGD) optimization is applied, with a momentum of 0.9 and a weight decay of 5E-4. And saving the trained model weight.
When the recognition network is trained, an Adam optimizer is adopted: the initial learning rate is set to 1E-4. And saving the trained model weight.
And 4, in the testing stage, inputting the unmarked image to be segmented into the trained segmentation network model to obtain a predicted class probability map, and searching the index of the maximum value in the channel dimension in the predicted class probability map to obtain the segmented semantic image.
The scheme of the embodiment of the invention has the following beneficial effects:
1) the invention develops a generative confrontation framework, treats the segmented network as a generator, and uses the full convolutional network as a recognizer. With the help of this generation of the countermeasure framework, the segmentation network can generate class probability maps with higher confidence.
2) The invention provides an unsupervised learning method for researching data distribution of unmarked images. In order to focus the unsupervised learning signal on misclassified regions, particularly at boundary regions, segmentation error regions of the unlabeled image are predicted instead of predicting potentially reliable regions, and then the uncertainty of the unlabeled image prediction is minimized.
3) We propose a semi-supervised learning framework that combines supervised learning and unsupervised learning. The experimental results on the PASCAL VOC 2012 and the PASCAL-CONTEXT data sets show that the semi-supervised learning method provided by the invention is competitive.
In order to demonstrate the performance of the above-described embodiment of the present invention, a comparative experiment was performed as follows.
In the experiment, images are selected in a manner similar to the step 1 to form a verification set. For example, 1449 images of the standard validation set were obtained in the PASCAL VOC 2012 data set to evaluate the trained network model, and 5105 images of the standard validation set were used in the PASCAL-CONTEXT data set to evaluate the trained network model.
Protocols participating in comparative experiments include: (1) a baseline network; (2) baseline network + loss adv (ii) a (3) Baseline network + loss adv +loss inf
And (3) analysis of experimental results:
1) results on PASCAL VOC 2012. The quantitative results of the method on the PASCAL VOC 2012 validation set are shown in table 1. The qualitative results of some sample images are shown in fig. 5. In fig. 5, (a) shows an original image, and (b) shows a label of a semantic division image. (c) Sections (a) to (e) correspond to schemes (1) to (3) in this order.
Figure GDA0003698490930000081
Figure GDA0003698490930000091
Table 1 results on PASCAL VOC 2012 validation set
As shown in Table 1, loss of antagonism adv Resulting in an improvement of mIOU (average cross-over ratio) of 1.1% to 1.4%. This indicates a loss of antagonism loss adv Segmentation performance can be improved by increasing the confidence of the prediction on the labeled image. Incorporating unsupervised loss inf Minimizing prediction on unmarked picturesThe proposed method achieves an improvement of 1.9% to 2.7% over the baseline network. The qualitative results presented in fig. 5 show that using models for loss tolerance and unsupervised loss achieves some improvement in the misclassified regions of the baseline network, particularly in some of the border regions.
2) The result of PASCAL-CONTEXT. The results of the quantitative assessment on the PASCAL CONTEXT dataset are shown in table 2. Furthermore, the qualitative results of some sample images are visualized in fig. 6. In fig. 6, (a) shows an original image, and (b) shows a semantic division image tag. (c) Sections (a) to (e) correspond to schemes (1) to (3) in this order.
Data volume 10% 30% 50% 100%
Baseline network 34.6 38.0 40.1 42.3
Baseline network + loss adv 35.1 38.7 40.8 42.9
Baseline network + loss adv +loss inf 35.9 39.6 41.3
TABLE 2 results on PASCAL-CONTEXT validation set
It can be found that the proposed method is still effective, and in a complex scenario, the proposed method improves the average cross-binding by 1.2% to 1.6%, with the antagonism loss accounting for about 0.5% to 0.7% of the performance improvement. The performance assessment of the PASCAL-CONTEXT dataset was worse than that on the PASCAL VOC 2012 dataset. This is because the PASCAL-CONTEXT data set containing objects and fill annotations is more complex, resulting in the proposed method not being able to accurately infer tag mismappings.
3) Compared to the most advanced methods. The method proposed by the present invention is first compared to several of the most advanced methods of weak supervision. All of these weakly supervised methods used a DeepLab-v2 based on ResNet-101 as the baseline network. The weak surveillance method was trained on the PASCAL VOC 2012 dataset using image level annotation, while the proposed method was trained on the same dataset using 440 pixel level annotation images and 10142 unlabeled images. As shown in Table 3, the proposed method has a mIOU (mean cross-over ratio) of 68.9%, which is at least 4.0% better than all weakly supervised methods. These large improvements can be attributed to the proposed method of obtaining more detailed information of the border area. Weakly supervised learning methods directly use image-level annotations, which lead to difficulties in locating the boundary regions. In contrast, the proposed method first learns how to locate the bounding regions by adversarial learning with limited pixel-level annotations; it then predicts segmentation error regions of the unlabeled image, making the unsupervised learning signal more focused on misclassified regions, especially in the border regions. Thus, the proposed method achieves a more competitive performance compared to the weakly supervised method.
Figure GDA0003698490930000101
Table 3 results of the inventive method compared to the advanced weak surveillance method on the PASCAL VOC 2012 validation set
Through the above description of the embodiments, it is clear to those skilled in the art that the above embodiments can be implemented by software, and can also be implemented by software plus a necessary general hardware platform. With this understanding, the technical solutions of the embodiments can be embodied in the form of a software product, which can be stored in a non-volatile storage medium (which can be a CD-ROM, a usb disk, a removable hard disk, etc.), and includes several instructions for enabling a computer device (which can be a personal computer, a server, or a network device, etc.) to execute the methods according to the embodiments of the present invention.
The above description is only for the preferred embodiment of the present invention, but the scope of the present invention is not limited thereto, and any changes or substitutions that can be easily conceived by those skilled in the art within the technical scope of the present invention are included in the scope of the present invention. Therefore, the protection scope of the present invention shall be subject to the protection scope of the claims.

Claims (5)

1. A semi-supervised semantic segmentation method based on maximized confidence is characterized by comprising the following steps:
constructing a training data set by using the marked images and the unmarked images in a specified proportion;
constructing a network model, and predicting a prediction class probability map of a marked image and an unmarked image through a segmentation network in the network model; adopting a supervised learning mode to maximize the confidence of the labeled image prediction class probability map; predicting a segmentation error region in the unmarked image prediction class probability map by adopting an unsupervised learning mode;
training the network model by combining the loss of supervised learning and the loss of unsupervised learning to obtain a trained segmented network;
in the testing stage, an unlabeled image to be segmented is input into a trained segmentation network model, a predicted class probability graph is obtained, and then an index of the maximum value in the channel dimension in the predicted class probability graph is searched to obtain a segmented semantic image;
the method for predicting the segmentation error region in the class probability map by adopting an unsupervised learning mode for the unmarked image comprises the following steps:
the information entropy of the unmarked image class probability graph represents the uncertainty of the segmentation result of the corresponding image, which is related to the segmentation error graph;
deducing a segmentation error map by using the information entropy of the predicted class probability map, and calculating the average information entropy in the error classification area as unsupervised loss after obtaining the segmentation error map;
given size H 1 ×W 1 ×C 1 Unmarked image x of n ', segmentation of network predicted unmarked image x n ' the prediction class probability map is S (x) n ') and an information entropy diagram H (x) is calculated in the following manner n '):
Figure FDA0003698490920000011
Wherein E [. C]Represents to all C 1 (ii) a desire for a category;
the information entropy indicates the uncertainty of the prediction of the segmentation network, given an uncertainty threshold T, a binary map is obtained representing the segmentation error map EM (x) n ') expressed as:
Figure FDA0003698490920000012
where h 1 ∈H 1 ,w 1 ∈W 1
wherein (h) 1 ,w 1 ) Representing the position coordinates of the pixel points;
by information entropy diagram H (x) n ') and a segmentation error map EM (x) n ') get unsupervised losses and feed back to the split network, the unsupervised losses are expressed as:
Figure FDA0003698490920000021
2. the semi-supervised semantic segmentation method based on maximizing confidence of claim 1, wherein the maximizing confidence of the prediction class probability map by adopting supervised learning for the labeled image comprises:
for the marked images, a generating countermeasure network is adopted, and the confidence of the prediction class probability map is maximized through a generating countermeasure mode in supervised learning;
the generation countermeasure network is composed of a segmentation network and a full convolution neural network in a network model;
in the generation of the countermeasure network, the segmentation network is used as a generator to predict a class probability graph of the marked image; the full convolution neural network is used as a recognizer, the input of the full convolution neural network is a marked image prediction class probability graph and a class probability graph generated after downsampling and onehot coding of a label graph, and the type of the input is recognized through the recognizer;
the generator and recognizer compete against each other with the goal of maximizing the confidence of the prediction class probability map.
3. The semi-supervised semantic segmentation method based on maximized confidence as claimed in claim 2, wherein the loss of supervised learning comprises: multi-class cross entropy loss and antagonism loss;
the multi-class cross entropy loss is used to facilitate the segmentation network to independently predict the correct semantic label class at each pixel position, and is expressed as:
Figure FDA0003698490920000022
wherein x is n For the marked image input to the segmented network, y n Encoding the label map for onehot of the corresponding labeled image, (h) 1 ,w 1 ,c 1 ) The predicted class probability map is of size H for the position coordinates of the pixels in the map 1 ×W 1 ×C 1 ,H 1 、W 1 Respectively representing the height, width, C of the image 1 Representing the number of categories, namely the number of channels; s (x) n ) Tagged image x predicted for segmented networks n A prediction class probability map of (1);
in the recognizer, a spatial binary class entropy penalty is used to distinguish whether the input is a predicted labeled image prediction class probability map or a class probability map generated from a label map, the spatial binary class entropy penalty being expressed as:
Figure FDA0003698490920000023
Y n =one_hot(ones(H 2 ,W 2 )×SG)
wherein p is n A labeled image prediction class probability map representing a prediction or a class probability map generated from a label map, D (-) representing a recognizer, Y · n Is a comment for distinguishing the source of the input, C 2 2, one _ hot (·) is an onehot coding function, ons (H) 2 ,W 2 ) For generating a size of H 2 ×W 2 Matrix of (H) 2 、W 2 Respectively representing the row number and the column number of the matrix, wherein the values of all elements are 1; SG equals 0, indicating the labeled image prediction class probability map for which the recognizer input is prediction; SG is 1, representing the recognizer input as a class probability map generated from the tag map; the spatial binary entropy loss is used for training a recognizer;
the resistance loss is expressed as:
Figure FDA0003698490920000031
4. the semi-supervised semantic segmentation method based on maximized confidence as claimed in claim 1, wherein the supervised learning loss and the unsupervised learning loss constitute a total loss of the network model, and are expressed as:
loss seg =loss mceadv loss advinf loss inf
wherein the supervised learning penalty comprises a spatial multi-class entropy penalty loss that facilitates a segmentation network to independently predict correct semantic label classes at each pixel location mce And a loss of antagonism loss (loss) to maximize confidence in the labeled image prediction class probability map adv (ii) a Unsupervised loss to maximize unmarked image prediction class probability map inf ;λ adv And λ inf Are two weights that balance the corresponding losses.
5. The semi-supervised semantic segmentation method based on maximized confidence as recited in claim 2,
the marked images and the unmarked images are combined together according to batch size as input, each hyper-parameter of a network model is set, a weight initialization mode is set, a random gradient descent method and a polygon learning rate strategy are used for training a segmentation network in combination with the loss of supervised learning and the loss of unsupervised learning, and an Adam optimizer and a Poly learning rate strategy are used for training a recognizer.
CN201911071629.4A 2019-11-05 2019-11-05 Semi-supervised semantic segmentation method based on maximized confidence Active CN110837836B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201911071629.4A CN110837836B (en) 2019-11-05 2019-11-05 Semi-supervised semantic segmentation method based on maximized confidence

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201911071629.4A CN110837836B (en) 2019-11-05 2019-11-05 Semi-supervised semantic segmentation method based on maximized confidence

Publications (2)

Publication Number Publication Date
CN110837836A CN110837836A (en) 2020-02-25
CN110837836B true CN110837836B (en) 2022-09-02

Family

ID=69576198

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201911071629.4A Active CN110837836B (en) 2019-11-05 2019-11-05 Semi-supervised semantic segmentation method based on maximized confidence

Country Status (1)

Country Link
CN (1) CN110837836B (en)

Families Citing this family (17)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN111651998B (en) * 2020-05-07 2022-09-06 中国科学技术大学 Weak supervision deep learning semantic analysis method under virtual reality and augmented reality scenes
CN111612010A (en) * 2020-05-21 2020-09-01 京东方科技集团股份有限公司 Image processing method, device, equipment and computer readable storage medium
CN111666953B (en) * 2020-06-04 2022-07-01 电子科技大学 Tidal zone surveying and mapping method and device based on semantic segmentation
CN111798471B (en) * 2020-07-27 2024-04-02 中科智脑(北京)技术有限公司 Training method of image semantic segmentation network
CN111870279B (en) * 2020-07-31 2022-01-28 西安电子科技大学 Method, system and application for segmenting left ventricular myocardium of ultrasonic image
CN112132149B (en) * 2020-09-10 2023-09-05 武汉汉达瑞科技有限公司 Semantic segmentation method and device for remote sensing image
CN112419327B (en) * 2020-12-10 2023-08-04 复旦大学附属肿瘤医院 Image segmentation method, system and device based on generation countermeasure network
CN112801107A (en) * 2021-02-01 2021-05-14 联想(北京)有限公司 Image segmentation method and electronic equipment
CN113269197B (en) * 2021-04-25 2024-03-08 南京三百云信息科技有限公司 Certificate image vertex coordinate regression system and identification method based on semantic segmentation
CN113516130B (en) * 2021-07-19 2024-01-05 闽江学院 Semi-supervised image semantic segmentation method based on entropy minimization
CN113537365B (en) * 2021-07-20 2024-02-06 北京航空航天大学 Information entropy dynamic weighting-based multi-task learning self-adaptive balancing method
CN113610807B (en) * 2021-08-09 2024-02-09 西安电子科技大学 New coronaries pneumonia segmentation method based on weak supervision multitask learning
CN114004817B (en) * 2021-11-03 2024-04-02 深圳大学 Semi-supervised training method, system, equipment and storage medium for segmentation network
CN114118167B (en) * 2021-12-04 2024-02-27 河南大学 Action sequence segmentation method aiming at behavior recognition and based on self-supervision less sample learning
CN114565755B (en) * 2022-01-17 2023-04-18 北京新氧科技有限公司 Image segmentation method, device, equipment and storage medium
CN115100491B (en) 2022-08-25 2022-11-18 山东省凯麟环保设备股份有限公司 Abnormal robust segmentation method and system for complex automatic driving scene
CN116403074A (en) * 2023-04-03 2023-07-07 上海锡鼎智能科技有限公司 Semi-automatic image labeling method and device based on active labeling

Family Cites Families (13)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN104091333A (en) * 2014-07-01 2014-10-08 黄河科技学院 Multi-class unsupervised color texture image segmentation method based on credible regional integration
CN104537676B (en) * 2015-01-12 2017-03-22 南京大学 Gradual image segmentation method based on online learning
US10402690B2 (en) * 2016-11-07 2019-09-03 Nec Corporation System and method for learning random-walk label propagation for weakly-supervised semantic segmentation
US10600185B2 (en) * 2017-03-08 2020-03-24 Siemens Healthcare Gmbh Automatic liver segmentation using adversarial image-to-image network
US10699412B2 (en) * 2017-03-23 2020-06-30 Petuum Inc. Structure correcting adversarial network for chest X-rays organ segmentation
CN108062753B (en) * 2017-12-29 2020-04-17 重庆理工大学 Unsupervised domain self-adaptive brain tumor semantic segmentation method based on deep counterstudy
CN108549895A (en) * 2018-04-17 2018-09-18 深圳市唯特视科技有限公司 A kind of semi-supervised semantic segmentation method based on confrontation network
CN109409240B (en) * 2018-09-28 2022-02-11 北京航空航天大学 SegNet remote sensing image semantic segmentation method combined with random walk
CN109614921B (en) * 2018-12-07 2022-09-30 安徽大学 Cell segmentation method based on semi-supervised learning of confrontation generation network
CN109741332B (en) * 2018-12-28 2021-06-04 天津大学 Man-machine cooperative image segmentation and annotation method
US10430946B1 (en) * 2019-03-14 2019-10-01 Inception Institute of Artificial Intelligence, Ltd. Medical image segmentation and severity grading using neural network architectures with semi-supervised learning techniques
CN109993770B (en) * 2019-04-09 2022-07-15 西南交通大学 Target tracking method for adaptive space-time learning and state recognition
CN110363772B (en) * 2019-08-22 2021-08-13 西南大学 Heart MRI segmentation method and system based on antagonistic network

Also Published As

Publication number Publication date
CN110837836A (en) 2020-02-25

Similar Documents

Publication Publication Date Title
CN110837836B (en) Semi-supervised semantic segmentation method based on maximized confidence
CN109934293B (en) Image recognition method, device, medium and confusion perception convolutional neural network
US10410353B2 (en) Multi-label semantic boundary detection system
CN110619369B (en) Fine-grained image classification method based on feature pyramid and global average pooling
US11670071B2 (en) Fine-grained image recognition
Kao et al. Hierarchical aesthetic quality assessment using deep convolutional neural networks
Xiao et al. A weakly supervised semantic segmentation network by aggregating seed cues: the multi-object proposal generation perspective
Hu et al. Qs-attn: Query-selected attention for contrastive learning in i2i translation
CN111126488B (en) Dual-attention-based image recognition method
CN112100346B (en) Visual question-answering method based on fusion of fine-grained image features and external knowledge
US7672915B2 (en) Method and system for labelling unlabeled data records in nodes of a self-organizing map for use in training a classifier for data classification in customer relationship management systems
CN109063719B (en) Image classification method combining structure similarity and class information
US11640714B2 (en) Video panoptic segmentation
Guan et al. A unified probabilistic model for global and local unsupervised feature selection
US9569698B2 (en) Method of classifying a multimodal object
CN110826609B (en) Double-current feature fusion image identification method based on reinforcement learning
EP1903479A1 (en) Method and system for data classification using a self-organizing map
CN116844179A (en) Emotion analysis method based on multi-mode cross attention mechanism image-text fusion
CN112257758A (en) Fine-grained image recognition method, convolutional neural network and training method thereof
JP2014106736A (en) Information processor and control method thereof
Li et al. Co-saliency detection based on hierarchical consistency
CN114332544A (en) Image block scoring-based fine-grained image classification method and device
Sun et al. Perceptual multi-channel visual feature fusion for scene categorization
Zhou et al. Semantic image segmentation using low-level features and contextual cues
CN116109649A (en) 3D point cloud instance segmentation method based on semantic error correction

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant
TR01 Transfer of patent right

Effective date of registration: 20231227

Address after: 230026 Jinzhai Road, Baohe District, Hefei, Anhui Province, No. 96

Patentee after: University of Science and Technology of China

Patentee after: Zhu Changan

Patentee after: Jin Yi

Address before: 230026 Jinzhai Road, Baohe District, Hefei, Anhui Province, No. 96

Patentee before: University of Science and Technology of China

TR01 Transfer of patent right