CN116028878A - Diversified query active learning method and device for image classification - Google Patents

Diversified query active learning method and device for image classification Download PDF

Info

Publication number
CN116028878A
CN116028878A CN202211589310.2A CN202211589310A CN116028878A CN 116028878 A CN116028878 A CN 116028878A CN 202211589310 A CN202211589310 A CN 202211589310A CN 116028878 A CN116028878 A CN 116028878A
Authority
CN
China
Prior art keywords
image
sample
classification
image set
unlabeled
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN202211589310.2A
Other languages
Chinese (zh)
Inventor
陈子晞
杜博
王增茂
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Wuhan University WHU
Original Assignee
Wuhan University WHU
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Wuhan University WHU filed Critical Wuhan University WHU
Priority to CN202211589310.2A priority Critical patent/CN116028878A/en
Publication of CN116028878A publication Critical patent/CN116028878A/en
Pending legal-status Critical Current

Links

Images

Classifications

    • YGENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
    • Y02TECHNOLOGIES OR APPLICATIONS FOR MITIGATION OR ADAPTATION AGAINST CLIMATE CHANGE
    • Y02DCLIMATE CHANGE MITIGATION TECHNOLOGIES IN INFORMATION AND COMMUNICATION TECHNOLOGIES [ICT], I.E. INFORMATION AND COMMUNICATION TECHNOLOGIES AIMING AT THE REDUCTION OF THEIR OWN ENERGY USE
    • Y02D10/00Energy efficient computing, e.g. low power processors, power management or thermal management

Landscapes

  • Information Retrieval, Db Structures And Fs Structures Therefor (AREA)

Abstract

The invention provides an image classification-oriented diversified query active learning method and device, wherein the method comprises the following steps: generating a pseudo tag with high confidence by using the auxiliary classification network as a main classification network, and removing redundant samples which are similar to the tag data from the non-tag data; in each active learning cycle, the main classification network and the auxiliary classification network exchange respective high-confidence pseudo-mark samples in a cooperative training mode, meanwhile, unlabeled samples predicted by the main classification network and the auxiliary classification network as different labels are added into a candidate pool for active learning, the auxiliary learning and multi-level diversity screening strategy is utilized for carrying out artificial expert marking on more diversified samples, and a small number of labeled samples are utilized for carrying out active learning cycle training on the classification model, so that the training effect close to using a large number of labeled samples can be achieved.

Description

Diversified query active learning method and device for image classification
Technical Field
The invention relates to the technical field of image processing, in particular to an image classification-oriented diversified query active learning method and device.
Background
The depth model learns the characteristics with strong robustness and expressive property from a large number of marked image samples, so that higher classification precision can be obtained on an image classification task. However, the labor and time costs required to annotate a large number of images are very high. In many practical applications, the number of available tagged images is very limited. Therefore, how to improve classification performance in the case of limited labeling data has become a challenge in the field of machine learning. Active learning is a very popular technique at present, which can train a depth model with high generalization capability by iteratively screening sample images and delivering limited annotation data obtained by expert annotation. Sample query strategies are the core of active learning methods. If the query function can select the sample with the largest information amount in each active learning cycle, the generalization capability of the classification model can be obviously improved through a small number of marked samples, so that the manual labeling cost is reduced. The query strategy in active learning is designed based mainly on two criteria, uncertainty and representativeness. Uncertainty criteria mainly indicate whether the predictions of the samples have a high confidence, and many active learning methods are designed based on this criteria, which select the samples with the largest information content and the lowest classification confidence, such as marginal sampling and entropy sampling, in each cycle. While the uncertainty criteria are very popular for use in active learning, it does not effectively improve the performance of classification. For example, if a batch of samples is selected in each cycle, there may be redundancy between samples with similar uncertainties; however, if only the least deterministic sample is selected in each cycle, the number of active learning cycles will increase significantly, resulting in an unacceptable time cost. To more fully explore the structural information in the unlabeled data, a representative method of selecting independent co-distributed samples in an exponential search space is proposed. But the computational cost of selecting a batch of independent co-distributed samples increases significantly over a large data set. Therefore, how to design an active learning strategy which is efficient and can improve the generalization capability of the depth model and reduce the cost of manual labeling is a problem with important practical significance.
Disclosure of Invention
In order to solve at least one technical problem in the prior art, the invention provides a diversified query active learning method and device for image classification.
In a first aspect, the present invention provides an image-classification-oriented diversified query active learning method, including:
step 1, acquiring an unlabeled image set U consisting of a first number of unlabeled images and a labeled image set L consisting of a second number of labeled images, wherein the first number is greater than the second number;
step 2, using the labeled image set L as a training set to respectively classify the main classification model f 1 And an auxiliary classification model f 2 Training to obtain f 1' and f2 ' and utilize f 1' and f2 ' extracting and predicting the characteristics of all the unlabeled images in the unlabeled image set U;
step 3, from f 1' and f2 ' predicting all the unlabeled images in the unlabeled image set to obtain the target image of the target image 1' and f2 ' image set S consisting of unlabeled images predicted to be different labels and composed of a quilt f 1' and f2 ' image set H of unlabeled images predicted to be the same label;
step 4, clustering, sorting and screening the unlabeled images in the image set S by using a multi-level diversity criterion to form an unlabeled sample candidate set to be labeled;
Step 5, labeling the label-free sample candidate set based on labeling operation, and removing the labeled sample from the label-free image set U and adding the labeled sample into the label-free image set L;
step 6, the images in the image set H are respectively processed according to f 1' and f2 Sequencing the prediction confidence degrees, and respectively screening out a sample set H with highest confidence degree 1 and H2 And respectively giving the samples of the sample with the values of f 1' and f2 'pseudo tag output' to obtain sample set H 1' and H2 ';
Step 7Using sample sets H 2 ' sum tagged image set L vs f 1 ' training to obtain f 1 ", use sample set H 1 ' sum tagged image set L vs f 2 ' training to obtain f 2 ”;
Step 8, f 1” and f2 "as the main classification model f 1 And an auxiliary classification model f 2 Returning to the step 2 until the main classification model f 1 Stopping iteration when the expected classification performance or the marked sample exceeds the set upper limit, and entering a step 9;
step 9, utilizing the main classification model f 1 And performing image classification prediction tasks.
In a second aspect, the present invention further provides an image-classification-oriented diversified query active learning apparatus, where the image-classification-oriented diversified query active learning apparatus includes:
the acquisition module is used for acquiring an unlabeled image set U consisting of a first number of unlabeled images and a labeled image set L consisting of a second number of labeled images, wherein the first number is larger than the second number;
A first training module for using the labeled image set L as training set to respectively classify the main classification model f 1 And an auxiliary classification model f 2 Training to obtain f 1' and f2 ' and utilize f 1′ and f2 ' extracting and predicting the characteristics of all the unlabeled images in the unlabeled image set U;
a first building block for building up the first building block from f 1' and f2 ' predicting all the unlabeled images in the unlabeled image set to obtain the target image of the target image 1′ and f2 ' image set S consisting of unlabeled images predicted to be different labels and composed of a quilt f 1' and f2 ' image set H of unlabeled images predicted to be the same label;
the second construction module is used for clustering, sorting and screening the unlabeled images in the image set S by using a multi-level diversity criterion to form an unlabeled sample candidate set to be labeled;
the labeling module is used for labeling the label-free sample candidate set based on labeling operation, removing the labeled sample from the label-free image set U and adding the labeled sample into the label-free image set L;
a third construction module for respectively setting the images in the image set H according to f 1' and f2 Sequencing the prediction confidence degrees, and respectively screening out a sample set H with highest confidence degree 1 and H2 And respectively giving the samples of the sample with the values of f 1' and f2 'pseudo tag output' to obtain sample set H 1' and H2 ′;
A second training module for utilizing the sample set H 2 ' sum tagged image set L vs f 1 ' training to obtain f 1 ", use sample set H 1 ' sum tagged image set L vs f 2 ' training to obtain f 2 ”;
A circulation module for at f 1” and f2 "as the main classification model f 1 And an auxiliary classification model f 2 Returning to the step executed by the first training module until the main classification model f 1 Stopping iteration when the expected classification performance or the labeling sample exceeds a set upper limit;
a prediction module for utilizing the main classification model f 1 And performing image classification prediction tasks.
In the present invention, the following steps are performed:
step 1, acquiring an unlabeled image set U consisting of a first number of unlabeled images and a labeled image set L consisting of a second number of labeled images, wherein the first number is greater than the second number;
step 2, using the labeled image set L as a training set to respectively classify the main classification model f 1 And an auxiliary classification model f 2 Training to obtain f 1' and f2 ' and utilize f 1' and f2 ' extracting and predicting the characteristics of all the unlabeled images in the unlabeled image set U;
step 3, from f 1' and f2 ' predicting all the unlabeled images in the unlabeled image set to obtain the target image of the target image 1' and f2 ' image set S consisting of unlabeled images predicted to be different labels and composed of a quilt f 1′ and f2 ' image set H of unlabeled images predicted to be the same label;
step 4, clustering, sorting and screening the unlabeled images in the image set S by using a multi-level diversity criterion to form an unlabeled sample candidate set to be labeled;
step 5, labeling the label-free sample candidate set based on labeling operation, and removing the labeled sample from the label-free image set U and adding the labeled sample into the label-free image set L;
step 6, the images in the image set H are respectively processed according to f 1' and f2 Sequencing the prediction confidence degrees, and respectively screening out a sample set H with highest confidence degree 1 and H2 And respectively giving the samples of the sample with the values of f 1′ and f2 'pseudo tag output' to obtain sample set H 1′ and H2 ′;
Step 7, utilizing sample set H 2 ' sum tagged image set L vs f 1 ' training to obtain f 1 ", use sample set H 1 ' sum tagged image set L vs f 2 ' training to obtain f 2 ”;
Step 8, f 1” and f2 "as the main classification model f 1 And an auxiliary classification model f 2 Returning to the step 2 until the main classification model f 1 Stopping iteration when the expected classification performance or the marked sample exceeds the set upper limit, and entering a step 9;
Step 9, utilizing the main classification model f 1 And performing image classification prediction tasks.
According to the invention, a large amount of manual labeling cost is reduced, and the training effect close to that of using a large amount of labeled samples can be achieved by using a small amount of labeled samples to perform active learning and circulating training on the classification model.
Drawings
FIG. 1 is a flow chart of an embodiment of an image classification oriented diversified query active learning method of the present invention;
fig. 2 is a schematic diagram of functional modules of an embodiment of the image classification-oriented diversified query active learning device according to the present invention.
The achievement of the objects, functional features and advantages of the present invention will be further described with reference to the accompanying drawings, in conjunction with the embodiments.
Detailed Description
It should be understood that the specific embodiments described herein are for purposes of illustration only and are not intended to limit the scope of the invention.
In a first aspect, an embodiment of the present invention provides an image classification-oriented active learning method for diversified queries.
In an embodiment, referring to fig. 1, fig. 1 is a flowchart illustrating an embodiment of a method for learning a diversified query for image classification according to the present invention. As shown in fig. 1, the method for actively learning the diversified queries for image classification includes:
Step 1, acquiring an unlabeled image set U consisting of a first number of unlabeled images and a labeled image set L consisting of a second number of labeled images, wherein the first number is greater than the second number;
in this embodiment, the first number is much larger than the second number, i.e. the set of unlabeled images U comprises a large number of unlabeled images and the set of labeled images L comprises a small number of labeled images.
Step 2, using the labeled image set L as a training set to respectively classify the main classification model f 1 And an auxiliary classification model f 2 Training to obtain f 1′ and f2 ' and utilize f 1′ and f2 ' extracting and predicting the characteristics of all the unlabeled images in the unlabeled image set U;
in this embodiment, the labeled image set L is used as a training set, and the main classification model f is respectively identified 1 And an auxiliary classification model f 2 Training to obtain f 1′ and f2 ′。
Further, in an embodiment, the labeled image set L is used as a training set, and the main classification model f is respectively calculated 1 And an auxiliary classification model f 2 Proceeding withTraining to obtain f 1′ and f2 The' step comprises:
for the ith labeled image x in the labeled image set L i Obtaining x i Are respectively f 1 and f2 Classification probability for class j:
Figure BDA0003991525340000051
Figure BDA0003991525340000052
wherein ,W1 and W2 Respectively f 1 and f2 Parameters of (2);
Figure BDA0003991525340000053
is based on f 1 The x obtained i Classification probability belonging to class j +.>
Figure BDA0003991525340000054
Is based on f 2 The x obtained i Classification probability belonging to class j;
by minimizing the primary classification model f 1 And an auxiliary classification model f 2 Cross entropy loss on labeled image set L
Figure BDA0003991525340000055
and />
Figure BDA0003991525340000056
For the main classification model f 1 And an auxiliary classification model f 2 Training to obtain f 1′ and f2 ' wherein for tagged images in the tagged image set L, the model f is classified in the main classification 1 And an auxiliary classification model f 2 The cross entropy losses above are respectively:
Figure BDA0003991525340000057
Figure BDA0003991525340000058
wherein ,
Figure BDA0003991525340000059
and />
Figure BDA00039915253400000510
Respectively, training a main classification model f 1 And an auxiliary classification model f 2 Is a target function of (2); c is the total number of categories contained in the tagged image set L; l is the number of tagged images contained by the tagged image set L; y is i Is the ith labeled image x i Is a real tag of (1); j is x i Is a predictive tag of (1); />
Figure BDA00039915253400000511
Is based on the x obtained by the main classification model i The prediction probabilities belonging to class j,
Figure BDA0003991525340000061
is obtained based on auxiliary classification model i Prediction probability belonging to class j; i { y } i =j } is an indication function if y i Equal to j, 1, otherwise 0.
After obtaining f 1′ and f2 ' after using f 1′ and f2 Feature extraction and prediction are performed on all the unlabeled images in the unlabeled image set U. I.e. all the unlabeled images in the unlabeled image set U are sequentially input into f 1′ and f2 ' obtain f 1′ and f2 ' prediction result for each unlabeled image, and f 1 ' feature extraction results for each unlabeled image.
Further, in one embodiment, f is utilized 1′ and f2 The step of' feature extraction and prediction of all the unlabeled images in the unlabeled image set U comprises:
for the ith unlabeled image z in the unlabeled image set U i Obtaining z i Are respectively f 1′ and f2 The classification probability of' class j:
Figure BDA0003991525340000062
Figure BDA0003991525340000063
wherein ,W1′ and W2 ' f is respectively 1′ and f2 ' parameters;
Figure BDA0003991525340000064
is based on f 1 ' z obtained i Classification probability belonging to class j +.>
Figure BDA0003991525340000065
Is based on f 2 ' z obtained i Classification probability belonging to class j;
the method comprises the following steps:
Figure BDA0003991525340000066
Figure BDA0003991525340000067
obtaining f 1′ and f2 ' respectively to z i Is a predictive pseudo tag of (1)
Figure BDA0003991525340000068
and />
Figure BDA0003991525340000069
According to f 1 The output of the last pooling layer yields the pair z i Extracted feature orientationAmount of the components.
Step 3, from f 1′ and f2 ' predicting all the unlabeled images in the unlabeled image set to obtain the target image of the target image 1′ and f2 ' image set S consisting of unlabeled images predicted to be different labels and composed of a quilt f 1′ and f2 ' image set H of unlabeled images predicted to be the same label;
step 4, clustering, sorting and screening the unlabeled images in the image set S by using a multi-level diversity criterion to form an unlabeled sample candidate set to be labeled;
In this embodiment, according to the feature vector corresponding to each unlabeled image in the image set S, the image set S is divided into C clusters by using a k-means clustering algorithm, which is expressed as u=u 1 ∪U 2 ∪...∪U C For the above C clusters, according to the sum of f in each cluster 1 The predicted category label number is further divided into a plurality of sub-clusters, the iterative clustering operation is carried out until only one category label is in each minimum sub-cluster, the clustering operation is stopped, m minimum sub-clusters are obtained and expressed as
Figure BDA00039915253400000610
And selecting one with highest uncertainty from each minimum sub-cluster according to a BvSB criterion to obtain a query data set F with m samples, wherein the BvSB criterion is expressed as:
Figure BDA0003991525340000071
Figure BDA0003991525340000072
Figure BDA0003991525340000073
wherein, for each smallest sub-cluster,
Figure BDA0003991525340000074
refers to samples within a cluster being f 1 ' classification probability of class j; r is (r) max1 Refer to f 1 ' class subscript with maximum probability of classifying samples, r max2 Refer to f 1 ' category subscript with second greatest probability of classifying samples->
Figure BDA0003991525340000075
Representing the maximum predictive probability of a sample +.>
Figure BDA0003991525340000076
And the second highest prediction probability->
Figure BDA0003991525340000077
The difference between them, p is chosen for each smallest sub-cluster diff The smallest sample forms the query dataset F, p diff The smallest sample is the sample with the highest uncertainty;
If t samples have the same pseudo tag k in F, then the threshold for selecting different samples with the same pseudo tag k is expressed as:
Figure BDA0003991525340000078
where Var (·) is the variance operator, β is the hyper-parameter used to adjust the threshold,
Figure BDA0003991525340000079
represents f 1 ' Classification probability for sample outputs in F with the same pseudo tag k, +.>
Figure BDA00039915253400000710
Represents f 1 ' Classification probability for sample outputs in F with the same pseudo tag k, +.>
Figure BDA00039915253400000711
Ordering F by uncertainty of the samples and defining as
Figure BDA00039915253400000712
Wherein the uncertainty of the sample is from->
Figure BDA00039915253400000713
Decrementing to +.>
Figure BDA00039915253400000714
From F u Sequentially selecting samples and adding the samples into a label-free sample candidate set Q, wherein +_>
Figure BDA00039915253400000715
Is F u The j sample in (1) is predicted as the k-th class, if t with a pseudo tag k exists in Q k Sample number, calculate->
Figure BDA00039915253400000716
And t k The similarity of the individual samples is:
Figure BDA00039915253400000717
if t k 0, then
Figure BDA00039915253400000718
Added to Q; otherwise, u k When greater than v k When in use, will->
Figure BDA00039915253400000719
Added to Q; repeating the screening process until h samples exist in the Q, wherein h is a hyper-parameter controlling the size of the active learning screening batch, and +.>
Figure BDA00039915253400000720
Represents f 1 ' classification probability for sample output with false tag k in Q.
Step 5, labeling the label-free sample candidate set based on labeling operation, and removing the labeled sample from the label-free image set U and adding the labeled sample into the label-free image set L;
In this embodiment, a manual expert specifically marks samples in the unlabeled sample candidate set, and a label can be added to each sample in the unlabeled sample candidate set based on the marking operation of the manual expert. Then, the marked sample is removed from the unlabeled image set U and added to the labeled image set L.
Step 6, the images in the image set H are respectively processed according to f 1′ and f2 Sequencing the prediction confidence degrees, and respectively screening out a sample set H with highest confidence degree 1 and H2 And respectively giving the samples of the sample with the values of f 1′ and f2 'pseudo tag output' to obtain sample set H 1′ and H2 ′;
In this embodiment, the images in the image set H are respectively expressed as f 1′ and f2 Sequencing the 'prediction confidence coefficient' to obtain two sequences, and screening a plurality of samples with highest confidence coefficient from the sequence 1 to form a sample set H 1 The method comprises the steps of carrying out a first treatment on the surface of the Likewise, a plurality of samples with highest confidence are screened from the sequence 2 to form a sample set H 2
For H 1 Each sample in (2) is assigned the value of f 1 'pseudo tag output' to obtain sample set H 1 'A'; similarly, for H 2 Each sample in (2) is assigned the value of f 2 'pseudo tag output' to obtain sample set H 2 ′。
Further, in an embodiment, the images in the image set H are respectively according to f 1′ and f2 Sequencing the prediction confidence degrees, and respectively screening out a sample set H with highest confidence degree 1 and H2 The method comprises the following steps:
the images in the image set H are processed according to f 1 Sequencing the 'prediction confidence degrees', screening out the top l multiplied by alpha images with the highest confidence degrees, and forming a sample set H 1
The images in the image set H are processed according to f 2 The 'confidence in prediction' is ranked,screening out the first l multiplied by alpha images with highest confidence coefficient to form a sample set H 2
wherein ,f1 The' confidence in prediction for the ith image in image set H is:
Figure BDA0003991525340000081
f 2 the' confidence in prediction for the ith image in image set H is:
Figure BDA0003991525340000082
Figure BDA0003991525340000083
is f for the ith image in the image set H 1 ' Classification probability of class j +.>
Figure BDA0003991525340000084
Is f for the ith image in the image set H 2 The' classification probability of j classes, i is the number of tagged images contained in the tagged image set L, and alpha is the super parameter.
Step 7, utilizing sample set H 2 ' sum tagged image set L vs f 1 ' training to obtain f 1 "utilize sample set H 1 ' sum tagged image set L vs f 2 ' training to obtain f 2 ″;
In this embodiment, the training process refers to the training process in step 2, and will not be described herein.
Step 8, f 1” and f2 "as the main classification model f 1 And an auxiliary classification model f 2 Returning to the step 2 until the main classification model f 1 Stopping iteration when the expected classification performance or the marked sample exceeds the set upper limit, and entering a step 9;
in the present embodiment, f is 1” and f2 "as the main classification model f 1 And an auxiliary classification model f 2 Returning to the step 2 until the stopping condition is met, stopping iteration, and obtaining f finally 1 As an image classification model, for performing image classification prediction tasks.
Step 9, utilizing the main classification model f 1 And performing image classification prediction tasks.
In this example, the final f is used 1 And performing image classification prediction tasks.
In this embodiment, the artificial expert labeling is performed on the more diversified samples by using the auxiliary learning and multi-level diversity screening strategy, so as to reduce the cost of artificial labeling. Specifically, in the auxiliary learning, the embodiment uses the auxiliary classification network as the main classification network to generate the pseudo tag with high confidence, and removes the redundant sample with similar representation with the tag data from the non-tag data; in each active learning cycle, the primary and secondary classification networks exchange respective highly-trusted pseudo-labeled samples in a co-training manner, while unlabeled samples predicted to be different labels by the primary and secondary classification networks are added to the candidate pool for active learning. Based on the predictions of the primary classification network, the present embodiment employs uncertainty criteria to select the most uncertain samples among them, followed by a multi-level diversity screening strategy to remove redundant samples from the uncertain samples. The multi-level diversity criterion comprises feature level diversity and class level diversity, wherein the feature level diversity can effectively mine the distribution structure of unlabeled data, unlabeled samples are clustered according to pseudo labels of a main classification model until the pseudo labels in each sub-cluster are the same, and the most uncertain sample is selected from each sub-cluster, so that samples with uncertainty and representativeness can be screened out; based on feature level diversity, the embodiment designs class level diversity criteria by considering redundancy inside query samples with the same pseudo tags, and effectively removes redundancy in uncertain samples with the same prediction tags, thereby ensuring that each query sample is beneficial to improving generalization capability of a classification network. Therefore, the multi-level diversity screening criteria adopted in the embodiment can ensure the representativeness and diversity of the selected samples, thereby reducing a large amount of manual labeling cost, and the training effect of using a large amount of labeled samples can be achieved by using a small amount of labeled samples to perform active learning and circulating training on the classification model.
In a second aspect, the embodiment of the invention further provides a diversified query active learning device oriented to image classification.
In an embodiment, referring to fig. 2, fig. 2 is a schematic functional block diagram of an embodiment of the image classification-oriented diversified query active learning device according to the present invention. As shown in fig. 2, the image classification-oriented diversified query active learning apparatus includes:
an acquisition module 10, configured to acquire an unlabeled image set U composed of a first number of unlabeled images and a labeled image set L composed of a second number of labeled images, where the first number is greater than the second number;
a first training module 20 for using the labeled image set L as training set for the main classification model f 1 And an auxiliary classification model f 2 Training to obtain f 1′ and f2 ' and utilize f 1′ and f2 ' extracting and predicting the characteristics of all the unlabeled images in the unlabeled image set U;
a first building block 30 for the first building block defined by f 1′ and f2 ' predicting all the unlabeled images in the unlabeled image set to obtain the target image of the target image 1′ and f2 ' image set S consisting of unlabeled images predicted to be different labels and composed of a quilt f 1′ and f2 ' image set H of unlabeled images predicted to be the same label;
The second construction module 40 is configured to cluster and sort the unlabeled images in the image set S by using a multi-level diversity criterion, so as to form an unlabeled sample candidate set to be labeled;
the labeling module 50 is configured to label the label-free sample candidate set based on a labeling operation, and remove the labeled sample from the label-free image set U and add the labeled sample to the labeled image set L;
third constructionA module 60 for respectively converting the images in the image set H according to f 1′ and f2 Sequencing the prediction confidence degrees, and respectively screening out a sample set H with highest confidence degree 1 and H2 And respectively giving the samples of the sample with the values of f 1′ and f2 'pseudo tag output' to obtain sample set H 1′ and H2 ′;
A second training module 70 for utilizing the sample set H 2 ' sum tagged image set L vs f 1 ' training to obtain f 1 ", use sample set H 1 ' sum tagged image set L vs f 2 ' training to obtain f 2 ”;
A circulation module 80 for at f 1” and f2 "as the main classification model f 1 And an auxiliary classification model f 2 Returning to the step executed by the first training module until the main classification model f 1 Stopping iteration when the expected classification performance or the labeling sample exceeds a set upper limit;
A prediction module 90 for utilizing a primary classification model f 1 And performing image classification prediction tasks.
Further, in an embodiment, the first training module 20 is configured to:
for the ith labeled image x in the labeled image set L i Obtaining x i Are respectively f 1 and f2 Classification probability for class j:
Figure BDA0003991525340000111
Figure BDA0003991525340000112
wherein ,W1 and W2 Respectively f 1 and f2 Parameters of (2);
Figure BDA0003991525340000113
is based on f 1 The x obtained i Classification probability belonging to class j +.>
Figure BDA0003991525340000114
Is based on f 2 The x obtained i Classification probability belonging to class j;
by minimizing the primary classification model f 1 And an auxiliary classification model f 2 Cross entropy loss on labeled image set L
Figure BDA0003991525340000115
and />
Figure BDA0003991525340000116
For the main classification model f1 and the auxiliary classification model f 2 Training to obtain f 1′ and f2 ' wherein for tagged images in the tagged image set L, the model f is classified in the main classification 1 And an auxiliary classification model f 2 The cross entropy losses above are respectively:
Figure BDA0003991525340000117
Figure BDA0003991525340000118
wherein ,
Figure BDA0003991525340000119
and />
Figure BDA00039915253400001110
Respectively, training a main classification model f 1 And an auxiliary classification model f 2 Is a target function of (2); c is the total number of categories contained in the tagged image set L; l is the number of tagged images contained by the tagged image set L; y is i Is the ith labeled image x i Is a real tag of (1); j is x i Is a predictive tag of (1); />
Figure BDA00039915253400001111
Is based on the x obtained by the main classification model i The prediction probabilities belonging to class j,
Figure BDA00039915253400001112
is obtained based on auxiliary classification model i Prediction probability belonging to class j; i { y } i =j } is an indication function if y i Equal to j, 1, otherwise 0.
Further, in an embodiment, the first training module 20 is configured to:
for the ith unlabeled image z in the unlabeled image set U i Obtaining z i Are respectively f 1′ and f2 The classification probability of' class j:
Figure BDA00039915253400001113
Figure BDA00039915253400001114
wherein ,W1′ and W2 ' f is respectively 1′ and f2 ' parameters;
Figure BDA00039915253400001115
is based on f 1 ' z obtained i Classification probability belonging to class j +.>
Figure BDA00039915253400001116
Is based on f 2 ' z obtained i Classification probability belonging to class j;
the method comprises the following steps:
Figure BDA00039915253400001117
Figure BDA0003991525340000121
obtaining f 1′ and f2 ' respectively to z i Is a predictive pseudo tag of (1)
Figure BDA0003991525340000122
and />
Figure BDA0003991525340000123
According to f 1 The output of the last pooling layer yields the pair z i The extracted feature vector.
Further, in an embodiment, the second construction module 40 is configured to:
according to the feature vector corresponding to each unlabeled image in the image set S, the image set S is divided into C clusters by using a k-means clustering algorithm, and the C clusters are expressed as U=U 1 ∪U 2 ∪...∪U C For the above C clusters, according to the sum of f in each cluster 1 The predicted category label number is further divided into a plurality of sub-clusters, the iterative clustering operation is carried out until only one category label is in each minimum sub-cluster, the clustering operation is stopped, m minimum sub-clusters are obtained and expressed as
Figure BDA00039915253400001217
And selecting one with highest uncertainty from each minimum sub-cluster according to a BvSB criterion to obtain a query data set F with m samples, wherein the BvSB criterion is expressed as:
Figure BDA0003991525340000124
Figure BDA0003991525340000125
Figure BDA0003991525340000126
wherein, for each smallest sub-cluster,
Figure BDA0003991525340000127
refers to samples within a cluster being f 1 ' classification probability of class j; r is (r) max1 Refer to f 1 ' class subscript with maximum probability of classifying samples, r max2 Refer to f 1 ' category subscript with second greatest probability of classifying samples->
Figure BDA0003991525340000128
Representing the maximum predictive probability of a sample +.>
Figure BDA0003991525340000129
And the second highest prediction probability->
Figure BDA00039915253400001210
The difference between them, p is chosen for each smallest sub-cluster diff The smallest sample forms the query dataset F, p diff The smallest sample is the sample with the highest uncertainty;
if t samples have the same pseudo tag k in F, then the threshold for selecting different samples with the same pseudo tag k is expressed as:
Figure BDA00039915253400001212
where Var (·) is the variance operator, β is the hyper-parameter used to adjust the threshold,
Figure BDA00039915253400001213
represents f 1 ' Classification probability for sample outputs in F with the same pseudo tag k, +.>
Figure BDA00039915253400001214
Represents f 1 ' Classification probability for sample outputs in F with the same pseudo tag k, +.>
Figure BDA00039915253400001215
Ordering F by uncertainty of the samples and defining as
Figure BDA00039915253400001216
Wherein the uncertainty of the sample is from- >
Figure BDA0003991525340000131
Decrementing to +.>
Figure BDA0003991525340000132
From F u Sequentially selecting samples and adding the samples into a label-free sample candidate set Q, wherein +_>
Figure BDA0003991525340000133
Is F u The j sample in (1) is predicted as the k-th class, if t with a pseudo tag k exists in Q k Sample number, calculate->
Figure BDA0003991525340000134
And t k The similarity of the individual samples is:
Figure BDA0003991525340000135
if t k 0, then
Figure BDA0003991525340000136
Added to Q; otherwise, u k When greater than v k When in use, will->
Figure BDA0003991525340000137
Added to Q; repeating the screening process until h samples exist in the Q, wherein h is a hyper-parameter controlling the size of the active learning screening batch, and +.>
Figure BDA0003991525340000138
Represents f 1 ' classification probability for sample output with false tag k in Q.
Further, in an embodiment, the third construction module 60 is configured to:
the images in the image set H are processed according to f 1 Sequencing the 'prediction confidence degrees', screening out the top l multiplied by alpha images with the highest confidence degrees, and forming a sample set H 1
The images in the image set H are processed according to f 2 Sequencing the 'prediction confidence degrees', screening out the top l multiplied by alpha images with the highest confidence degrees, and forming a sample set H 2
wherein ,f1 The' confidence in prediction for the ith image in image set H is:
Figure BDA0003991525340000139
f 2 the' confidence in prediction for the ith image in image set H is:
Figure BDA00039915253400001310
Figure BDA00039915253400001311
is f for the ith image in the image set H 1 ' Classification probability of class j +. >
Figure BDA00039915253400001312
Is f for the ith image in the image set H 2 The' classification probability of j classes, i is the number of tagged images contained in the tagged image set L, and alpha is the super parameter.
The function implementation of each module in the image-classification-oriented diversified query active learning device corresponds to each step in the embodiment of the image-classification-oriented diversified query active learning method, and the function and implementation process of the method are not described in detail herein.
It should be noted that, in this document, the terms "comprises," "comprising," or any other variation thereof, are intended to cover a non-exclusive inclusion, such that a process, method, article, or system that comprises a list of elements does not include only those elements but may include other elements not expressly listed or inherent to such process, method, article, or system. Without further limitation, an element defined by the phrase "comprising one … …" does not exclude the presence of other like elements in a process, method, article, or system that comprises the element.
The foregoing embodiment numbers of the present invention are merely for the purpose of description, and do not represent the advantages or disadvantages of the embodiments.
From the above description of the embodiments, it will be clear to those skilled in the art that the above-described embodiment method may be implemented by means of software plus a necessary general hardware platform, but of course may also be implemented by means of hardware, but in many cases the former is a preferred embodiment. Based on such understanding, the technical solution of the present invention may be embodied essentially or in a part contributing to the prior art in the form of a software product stored in a storage medium (e.g. ROM/RAM, magnetic disk, optical disk) as described above, comprising several instructions for causing a terminal device to perform the method according to the embodiments of the present invention.
The foregoing description is only of the preferred embodiments of the present invention, and is not intended to limit the scope of the invention, but rather is intended to cover any equivalents of the structures or equivalent processes disclosed herein or in the alternative, which may be employed directly or indirectly in other related arts.

Claims (10)

1. The image classification-oriented diversified query active learning method is characterized by comprising the following steps of:
Step 1, acquiring an unlabeled image set U consisting of a first number of unlabeled images and a labeled image set L consisting of a second number of labeled images, wherein the first number is greater than the second number;
step 2, using the labeled image set L as a training set to respectively classify the main classification model f 1 And an auxiliary classification model f 2 Training to obtain f 1′ and f2 ' and utilize f 1′ and f2 ' extracting and predicting the characteristics of all the unlabeled images in the unlabeled image set U;
step 3, from f 1′ and f2 ' predicting all the unlabeled images in the unlabeled image set to obtain the target image of the target image 1′ and f2 ' image set S consisting of unlabeled images predicted to be different labels and composed of a quilt f 1′ and f2 ' image set H of unlabeled images predicted to be the same label;
step 4, clustering, sorting and screening the unlabeled images in the image set S by using a multi-level diversity criterion to form an unlabeled sample candidate set to be labeled;
step 5, labeling the label-free sample candidate set based on labeling operation, and removing the labeled sample from the label-free image set U and adding the labeled sample into the label-free image set L;
step 6, the images in the image set H are respectively processed according to f 1′ and f2 Sequencing the prediction confidence degrees, and respectively screening out a sample set H with highest confidence degree 1 and H2 And respectively giving the samples of the sample with the values of f 1′ and f2 'pseudo tag output' to obtain sample set H 1′ and H2 ′;
Step 7, utilizing sample set H 2 ' sum tagged image set L vs f 1 ' training to obtain f 1 ", use sample set H 1 ' sum tagged image set L vs f 2 ' training to obtain f 2 ”;
Step 8, f 1” and f2 "as the main classification model f 1 And an auxiliary classification model f 2 Returning to the step 2 until the main classification model f 1 Stopping iteration when the expected classification performance or the marked sample exceeds the set upper limit, and entering a step 9;
step 9, utilizing the main classificationModel f 1 And performing image classification prediction tasks.
2. The method for actively learning image-classification-oriented diversified queries according to claim 1, wherein the labeled image set L is used as a training set for the main classification model f 1 And an auxiliary classification model f 2 Training to obtain f 1' and f2 The' step comprises:
for the ith labeled image x in the labeled image set L i Obtaining x i Are respectively f 1 and f2 Classification probability for class j:
Figure FDA0003991525330000021
Figure FDA0003991525330000022
wherein ,
Figure FDA0003991525330000023
and />
Figure FDA0003991525330000024
Respectively f 1 and f2 Parameters of (2); / >
Figure FDA0003991525330000025
Is based on f 1 The x obtained i Classification probability belonging to class j +.>
Figure FDA0003991525330000026
Is based on f 2 The x obtained i Classification probability belonging to class j;
by minimizing the primary classification model f 1 And an auxiliary classification model f 2 Cross entropy loss on labeled image set L
Figure FDA0003991525330000027
and />
Figure FDA0003991525330000028
For the main classification model f 1 And an auxiliary classification model f 2 Training to obtain f 1' and f2 ' wherein for tagged images in the tagged image set L, the model f is classified in the main classification 1 And an auxiliary classification model f 2 The cross entropy losses above are respectively:
Figure FDA0003991525330000029
/>
Figure FDA00039915253300000210
wherein ,
Figure FDA00039915253300000211
and />
Figure FDA00039915253300000212
Respectively, training a main classification model f 1 And an auxiliary classification model f 2 Is a target function of (2); c is the total number of categories contained in the tagged image set L; l is the number of tagged images contained by the tagged image set L; y is i Is the ith labeled image x i Is a real tag of (1); j is x i Is a predictive tag of (1); />
Figure FDA00039915253300000213
Is based on the x obtained by the main classification model i Prediction probability belonging to class j, +.>
Figure FDA00039915253300000214
Is obtained based on auxiliary classification model i Prediction probability belonging to class j;I{y i =j } is an indication function if y i Equal to j, 1, otherwise 0.
3. The method for actively learning image-classification-oriented diversified queries of claim 2 utilizing f 1′ and f2 The step of' feature extraction and prediction of all the unlabeled images in the unlabeled image set U comprises:
For the ith unlabeled image z in the unlabeled image set U i Obtaining z i Are respectively f 1' and f2 The classification probability of' class j:
Figure FDA00039915253300000215
Figure FDA00039915253300000216
wherein ,
Figure FDA00039915253300000217
and />
Figure FDA00039915253300000218
Respectively f 1′ and f2 ' parameters; />
Figure FDA00039915253300000219
Is based on f 1 ' z obtained i The probability of classification belonging to class j,
Figure FDA00039915253300000220
is based on f 2 ' z obtained i Classification probability belonging to class j;
the method comprises the following steps:
Figure FDA0003991525330000031
Figure FDA0003991525330000032
obtaining f 1′ and f2 ' respectively to z i Is a predictive pseudo tag of (1)
Figure FDA0003991525330000033
and />
Figure FDA0003991525330000034
According to f 1 The output of the last pooling layer yields the pair z i The extracted feature vector.
4. The method for actively learning image-classification-oriented diversified query of claim 3 wherein said step of clustering and sorting unlabeled images in image set S using a multi-level diversity criterion to form a set of unlabeled sample candidates to be labeled comprises:
according to the feature vector corresponding to each unlabeled image in the image set S, the image set S is divided into C clusters by using a k-means clustering algorithm, and the C clusters are expressed as U=U 1 ∪U 2 ∪…∪U C For the above C clusters, according to the sum of f in each cluster 1 The predicted category label number is further divided into a plurality of sub-clusters, the iterative clustering operation is carried out until only one category label is in each minimum sub-cluster, the clustering operation is stopped, m minimum sub-clusters are obtained and expressed as
Figure FDA0003991525330000035
And selecting one with highest uncertainty from each minimum sub-cluster according to a BvSB criterion to obtain a query data set F with m samples, wherein the BvSB criterion is expressed as:
Figure FDA0003991525330000036
Figure FDA0003991525330000037
/>
Figure FDA0003991525330000038
wherein, for each smallest sub-cluster,
Figure FDA0003991525330000039
refers to samples within a cluster being f 1 ' classification probability of class j; r is (r) max1 Refer to f 1 ' class subscript with maximum probability of classifying samples, r max2 Refer to f 1 ' category subscript with second greatest probability of classifying samples->
Figure FDA00039915253300000310
Representing the maximum predictive probability of a sample +.>
Figure FDA00039915253300000311
And the second highest prediction probability->
Figure FDA00039915253300000312
The difference between them, p is chosen for each smallest sub-cluster diff The smallest sample forms the query dataset F, p diff The smallest sample is the sample with the highest uncertainty;
if t samples have the same pseudo tag k in F, then the threshold for selecting different samples with the same pseudo tag k is expressed as:
Figure FDA00039915253300000313
where Var (·) is the variance operator, β is the hyper-parameter used to adjust the threshold,
Figure FDA0003991525330000041
represents f 1 ' Classification probability for sample outputs in F with the same pseudo tag k, +.>
Figure FDA0003991525330000042
Figure FDA0003991525330000043
Represents f 1 ' Classification probability for sample outputs in F with the same pseudo tag k, +.>
Figure FDA0003991525330000044
Ordering F by uncertainty of the samples and defining as
Figure FDA0003991525330000045
Wherein the uncertainty of the sample is from- >
Figure FDA0003991525330000046
Decrementing to +.>
Figure FDA0003991525330000047
From F u Sequentially selecting samples and adding the samples into a label-free sample candidate set Q, wherein +_>
Figure FDA0003991525330000048
Is F u The j sample in (1) is predicted as the k-th class, if t with a pseudo tag k exists in Q k Sample number, calculate->
Figure FDA0003991525330000049
And t k The similarity of the individual samples is:
Figure FDA00039915253300000410
if t k 0, then
Figure FDA00039915253300000411
Added to Q; otherwise, u k When greater than v k When in use, will->
Figure FDA00039915253300000412
Added to Q; repeating the screening process until h samples exist in the Q, wherein h is a hyper-parameter controlling the size of the active learning screening batch, and +.>
Figure FDA00039915253300000413
Represents f 1 ' classification probability for sample output with false tag k in Q.
5. The method for actively learning image-classification-oriented diversified queries of claim 4 wherein images in said image set H are each identified as f 1′ and f2 Sequencing the prediction confidence degrees, and respectively screening out a sample set H with highest confidence degree 1 and H2 The method comprises the following steps:
the images in the image set H are processed according to f 1 Sequencing the 'prediction confidence degrees', screening out the top l multiplied by alpha images with the highest confidence degrees, and forming a sample set H 1
The images in the image set H are processed according to f 2 Sequencing the 'prediction confidence degrees', screening out the top l multiplied by alpha images with the highest confidence degrees, and forming a sample set H 2
wherein ,f1 The' confidence in prediction for the ith image in image set H is:
Figure FDA00039915253300000414
f 2 the' confidence in prediction for the ith image in image set H is:
Figure FDA00039915253300000415
Figure FDA00039915253300000416
is f for the ith image in the image set H 1 ' Classification probability of class j +.>
Figure FDA00039915253300000417
Is f for the ith image in the image set H 2 The' classification probability of j classes, i is the number of tagged images contained in the tagged image set L, and alpha is the super parameter.
6. The utility model provides a diversified inquiry initiative learning device towards image classification which characterized in that, diversified inquiry initiative learning device towards image classification includes:
the acquisition module is used for acquiring an unlabeled image set U consisting of a first number of unlabeled images and a labeled image set L consisting of a second number of labeled images, wherein the first number is larger than the second number;
a first training module for using the labeled image set L as training set to respectively classify the main classification model f 1 And an auxiliary classification model f 2 Training to obtain f 1′ and f2 ' and utilize f 1′ and f2 ' extracting and predicting the characteristics of all the unlabeled images in the unlabeled image set U;
a first building block for building up the first building block from f 1′ and f2 ' predicting all the unlabeled images in the unlabeled image set to obtain the target image of the target image 1' and f2 ' image set S consisting of unlabeled images predicted to be different labels and composed of a quilt f 1′ and f2 ' predicted as phaseAn image set H consisting of unlabeled images of the same label;
the second construction module is used for clustering, sorting and screening the unlabeled images in the image set S by using a multi-level diversity criterion to form an unlabeled sample candidate set to be labeled;
the labeling module is used for labeling the label-free sample candidate set based on labeling operation, removing the labeled sample from the label-free image set U and adding the labeled sample into the label-free image set L;
a third construction module for respectively setting the images in the image set H according to f 1′ and f2 Sequencing the prediction confidence degrees, and respectively screening out a sample set H with highest confidence degree 1 and H2 And respectively giving the samples of the sample with the values of f 1′ and f2 'pseudo tag output' to obtain sample set H 1′ and H2 ′;
A second training module for utilizing the sample set H 2 ' sum tagged image set L vs f 1 ' training to obtain f 1 ", use sample set H 1 ' sum tagged image set L vs f 2 ' training to obtain f 2 ”;
A circulation module for at f 1” and f2 "as the main classification model f 1 And an auxiliary classification model f 2 Returning to the step executed by the first training module until the main classification model f 1 Stopping iteration when the expected classification performance or the labeling sample exceeds a set upper limit;
a prediction module for utilizing the main classification model f 1 And performing image classification prediction tasks.
7. The image classification-oriented diversified query active learning apparatus of claim 6 wherein the first training module is configured to:
for the ith labeled image x in the labeled image set L i Obtaining x i Are respectively f 1 and f2 Classification probability for class j:
Figure FDA0003991525330000061
Figure FDA0003991525330000062
wherein ,
Figure FDA0003991525330000063
and />
Figure FDA0003991525330000064
Respectively f 1 and f2 Parameters of (2); />
Figure FDA0003991525330000065
Is based on f 1 The x obtained i Classification probability belonging to class j +.>
Figure FDA0003991525330000066
Is based on f 2 The x obtained i Classification probability belonging to class j;
by minimizing the primary classification model f 1 And an auxiliary classification model f 2 Cross entropy loss on labeled image set L
Figure FDA0003991525330000067
and />
Figure FDA0003991525330000068
For the main classification model f 1 And an auxiliary classification model f 2 Training to obtain f 1′ and f2 ' wherein for tagged images in the tagged image set L, the model f is classified in the main classification 1 And an auxiliary classification model f 2 The cross entropy losses above are respectively:
Figure FDA0003991525330000069
Figure FDA00039915253300000610
wherein ,
Figure FDA00039915253300000611
and />
Figure FDA00039915253300000612
Respectively, training a main classification model f 1 And an auxiliary classification model f 2 Is a target function of (2); c is the total number of categories contained in the tagged image set L; l is the number of tagged images contained by the tagged image set L; y is i Is the ith labeled image x i Is a real tag of (1); j is x i Is a predictive tag of (1); />
Figure FDA00039915253300000613
Is based on the x obtained by the main classification model i Prediction probability belonging to class j, +.>
Figure FDA00039915253300000614
Is obtained based on auxiliary classification model i Prediction probability belonging to class j; i { y } i =j } is an indication function if y i Equal to j, 1, otherwise 0.
8. The image classification-oriented diversified query active learning apparatus of claim 7 wherein the first training module is configured to:
for the ith unlabeled image z in the unlabeled image set U i Obtaining z i Are respectively f 1' and f2 The classification probability of' class j:
Figure FDA00039915253300000615
Figure FDA00039915253300000616
wherein ,
Figure FDA00039915253300000617
and />
Figure FDA00039915253300000618
Respectively f 1' and f2 ' parameters; />
Figure FDA00039915253300000619
Is based on f 1 ' z obtained i The probability of classification belonging to class j,
Figure FDA00039915253300000620
is based on f 2 ' z obtained i Classification probability belonging to class j;
the method comprises the following steps:
Figure FDA0003991525330000071
Figure FDA0003991525330000072
obtaining f 1′ and f2 ' respectively to z i Is a predictive pseudo tag of (1)
Figure FDA0003991525330000073
and />
Figure FDA0003991525330000074
According to f 1 Output of the last pooling layerObtain the pair z i The extracted feature vector.
9. The image-classification-oriented diversified query active learning device of claim 8 wherein the second building module is configured to:
according to the feature vector corresponding to each unlabeled image in the image set S, the image set S is divided into C clusters by using a k-means clustering algorithm, and the C clusters are expressed as U=U 1 ∪U 2 ∪…∪U C For the above C clusters, according to the sum of f in each cluster 1 The predicted category label number is further divided into a plurality of sub-clusters, the iterative clustering operation is carried out until only one category label is in each minimum sub-cluster, the clustering operation is stopped, m minimum sub-clusters are obtained and expressed as
Figure FDA0003991525330000075
And selecting one with highest uncertainty from each minimum sub-cluster according to a BvSB criterion to obtain a query data set F with m samples, wherein the BvSB criterion is expressed as: />
Figure FDA0003991525330000076
Figure FDA0003991525330000077
Figure FDA0003991525330000078
Wherein, for each smallest sub-cluster,
Figure FDA0003991525330000079
refers to samples within a cluster being f 1 ' classification probability of class j; r is (r) max1 Refer to f 1 ' Classification of samplesCategory subscript with highest probability, r max2 Refer to f 1 ' category subscript with second greatest probability of classifying samples->
Figure FDA00039915253300000710
Representing the maximum predictive probability of a sample +.>
Figure FDA00039915253300000711
And the second highest prediction probability->
Figure FDA00039915253300000712
The difference between them, p is chosen for each smallest sub-cluster diff The smallest sample forms the query dataset F, p diff The smallest sample is the sample with the highest uncertainty;
if t samples have the same pseudo tag k in F, then the threshold for selecting different samples with the same pseudo tag k is expressed as:
Figure FDA00039915253300000713
where Var (·) is the variance operator, β is the hyper-parameter used to adjust the threshold,
Figure FDA00039915253300000714
Represents f 1 ' Classification probability for sample outputs in F with the same pseudo tag k, +.>
Figure FDA0003991525330000081
Figure FDA0003991525330000082
Represents f 1 ' Classification probability for sample outputs in F with the same pseudo tag k, +.>
Figure FDA0003991525330000083
Ordering F by uncertainty of the samples and defining as
Figure FDA0003991525330000084
Wherein the uncertainty of the sample is from->
Figure FDA0003991525330000085
Decrementing to +.>
Figure FDA0003991525330000086
From F u Sequentially selecting samples and adding the samples into a label-free sample candidate set Q, wherein +_>
Figure FDA0003991525330000087
Is F u The j sample in (1) is predicted as the k-th class, if t with a pseudo tag k exists in Q k Sample number, calculate->
Figure FDA0003991525330000088
And t k The similarity of the individual samples is:
Figure FDA0003991525330000089
if t k 0, then
Figure FDA00039915253300000810
Added to Q; otherwise, u k When greater than v k When in use, will->
Figure FDA00039915253300000811
Added to Q; repeating the screening process until h samples exist in the Q, wherein h is a hyper-parameter controlling the size of the active learning screening batch, and +.>
Figure FDA00039915253300000812
Represents f 1 ' classification probability for sample output with false tag k in Q.
10. The image classification-oriented diversified query active learning device of claim 9 wherein the third building module is configured to:
the images in the image set H are processed according to f 1 Sequencing the 'prediction confidence degrees', screening out the top l multiplied by alpha images with the highest confidence degrees, and forming a sample set H 1
The images in the image set H are processed according to f 2 Sequencing the 'prediction confidence degrees', screening out the top l multiplied by alpha images with the highest confidence degrees, and forming a sample set H 2
wherein ,f1 The' confidence in prediction for the ith image in image set H is:
Figure FDA00039915253300000813
f 2 the' confidence in prediction for the ith image in image set H is:
Figure FDA00039915253300000814
Figure FDA00039915253300000815
is f for the ith image in the image set H 1 ' Classification probability of class j +.>
Figure FDA00039915253300000816
Is f for the ith image in the image set H 2 The' classification probability of j classes, i is the number of tagged images contained in the tagged image set L, and alpha is the super parameter. />
CN202211589310.2A 2022-12-09 2022-12-09 Diversified query active learning method and device for image classification Pending CN116028878A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202211589310.2A CN116028878A (en) 2022-12-09 2022-12-09 Diversified query active learning method and device for image classification

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202211589310.2A CN116028878A (en) 2022-12-09 2022-12-09 Diversified query active learning method and device for image classification

Publications (1)

Publication Number Publication Date
CN116028878A true CN116028878A (en) 2023-04-28

Family

ID=86076747

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202211589310.2A Pending CN116028878A (en) 2022-12-09 2022-12-09 Diversified query active learning method and device for image classification

Country Status (1)

Country Link
CN (1) CN116028878A (en)

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN117708694A (en) * 2023-11-15 2024-03-15 佛山科学技术学院 Diagnosis method and equipment for semi-supervised learning of self-adaptive prototype network and pseudo tag

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN117708694A (en) * 2023-11-15 2024-03-15 佛山科学技术学院 Diagnosis method and equipment for semi-supervised learning of self-adaptive prototype network and pseudo tag

Similar Documents

Publication Publication Date Title
CN114241282B (en) Knowledge distillation-based edge equipment scene recognition method and device
CN111783831B (en) Complex image accurate classification method based on multi-source multi-label shared subspace learning
JP5234469B2 (en) Correspondence relationship learning device and method, correspondence relationship learning program, annotation device and method, annotation program, retrieval device and method, and retrieval program
CN110647907B (en) Multi-label image classification algorithm using multi-layer classification and dictionary learning
Bochinski et al. Deep active learning for in situ plankton classification
CN113657267B (en) Semi-supervised pedestrian re-identification method and device
CN113806582B (en) Image retrieval method, image retrieval device, electronic equipment and storage medium
CN115292532B (en) Remote sensing image domain adaptive retrieval method based on pseudo tag consistency learning
CN112734037A (en) Memory-guidance-based weakly supervised learning method, computer device and storage medium
CN116028878A (en) Diversified query active learning method and device for image classification
Kushnir et al. Diffusion-based deep active learning
CN114093445B (en) Patient screening marking method based on partial multi-marking learning
CN111191033A (en) Open set classification method based on classification utility
CN114329031A (en) Fine-grained bird image retrieval method based on graph neural network and deep hash
CN113705215A (en) Meta-learning-based large-scale multi-label text classification method
CN113658109A (en) Glass defect detection method based on field loss prediction active learning
CN114492581A (en) Method for classifying small sample pictures based on transfer learning and attention mechanism element learning application
CN113095229A (en) Unsupervised domain self-adaptive pedestrian re-identification system and method
CN117079017A (en) Credible small sample image identification and classification method
CN115795035A (en) Science and technology service resource classification method and system based on evolutionary neural network and computer readable storage medium thereof
CN115690566A (en) Deep sea animal new species identification method based on deep migration clustering learning
Xue et al. Fast and unsupervised neural architecture evolution for visual representation learning
CN115063831A (en) High-performance pedestrian retrieval and re-identification method and device
CN114418111A (en) Label prediction model training and sample screening method, device and storage medium
Li et al. A BYY scale-incremental EM algorithm for Gaussian mixture learning

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination