CN113269226B - Picture selection labeling method based on local and global information - Google Patents
Picture selection labeling method based on local and global information Download PDFInfo
- Publication number
- CN113269226B CN113269226B CN202110399472.9A CN202110399472A CN113269226B CN 113269226 B CN113269226 B CN 113269226B CN 202110399472 A CN202110399472 A CN 202110399472A CN 113269226 B CN113269226 B CN 113269226B
- Authority
- CN
- China
- Prior art keywords
- picture
- model
- objects
- information
- budget
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Active
Links
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F18/00—Pattern recognition
- G06F18/20—Analysing
- G06F18/21—Design or setup of recognition systems or techniques; Extraction of features in feature space; Blind source separation
- G06F18/214—Generating training patterns; Bootstrap methods, e.g. bagging or boosting
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F18/00—Pattern recognition
- G06F18/20—Analysing
- G06F18/23—Clustering techniques
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/04—Architecture, e.g. interconnection topology
- G06N3/045—Combinations of networks
Abstract
The invention discloses a picture selection and marking method based on local and global information, which can learn a model as good as possible by using pictures with marks as less as possible by enabling a learning model to automatically select partial pictures for marking. In order to reduce the requirement of image marking, the method utilizes the feature extraction capability of the depth model to construct the feature representation space of the image sample, and the effect of the sample on model updating is measured based on the local information of the image sample in the feature representation space. Meanwhile, the picture data space is divided into different areas based on the global information of the feature representation space, and the labeling budget is dynamically allocated according to the performance of the model on the different areas, so that the picture marking information is efficiently utilized, and the demand of picture marking is reduced.
Description
Technical Field
The invention relates to a picture selection labeling method based on local and global information, which can efficiently select objects to be labeled in a picture database by utilizing the local and global information of a feature representation space, train a better picture classification model with less labeling cost, and belongs to the technical field of computer artificial intelligent data analysis.
Background
With the continuous development of the internet, a large amount of picture data needs to be processed, such as face pictures in face recognition, road pictures in automatic driving, commodity pictures on e-commerce platforms, and the like. The picture data structure is complex, so the picture classification task is often completed by using a depth model. But training the depth model requires a large number of labeled pictures. In general, it is expensive to label these pictures with a lot of manpower and material resources. In order to reduce the labeling cost and improve the utilization efficiency of labeled pictures, one solution is to let the model automatically select important pictures to be labeled, and collect the labels of the pictures for updating the model, which is the basic idea of selecting labels. The current selection labeling method mainly considers the uncertainty and the representativeness of data when measuring the importance degree of the data. Wherein the lower the confidence of the model in the prediction of the data, the higher the uncertainty of the data. In addition, the modulo length of the data gradient can also be used to estimate the uncertainty of the data. Since the uncertainty-based approach only considers the uncertainty level of a single data, the model easily picks out a batch of data with high uncertainty but redundancy. This problem can be alleviated to some extent by taking into account the representativeness of the data. Typically, a representative-based approach groups the features of the data into clusters, and selects the center point of each cluster as a representative of the cluster. Therefore, the distribution condition of the whole data can be described by only using a small amount of data. However, in this method, since there is no information about the model as a guide, the selected data does not necessarily facilitate the update of the model.
Disclosure of Invention
The invention aims to: aiming at the problems and the defects in the prior art, the invention provides a picture selection and annotation method based on local and global information. The method can utilize the picture characteristics to represent local information in space, combines the prediction result of the model, measures the information quantity of the picture, and can avoid similar or redundant pictures to a certain extent. Meanwhile, the global information of the feature representation space is combined, the picture data are divided into a plurality of clusters, and the labeling budget is dynamically allocated according to the performance of the model on different clusters, so that the utilization efficiency of the picture label is further improved, and the labeling cost is reduced. When the same number of marked pictures are utilized, the model trained by the method has better performance compared with a general selection marking method.
The technical scheme is as follows: a picture selection labeling method based on local and global information comprises the following contents:
first, a user is required to create a library of picture objects. And then randomly selecting a part of picture objects from the picture object library, acquiring marks of the picture objects, and forming an initial training set. And setting the structure of the depth model, the number of the selected picture objects in each round and the total number of iteration rounds by the user.
Next, the deep learning model is trained based on the training set. And converting the picture objects in the picture object library into feature representations by using the depth model, namely extracting the features of the pictures in the picture object library. Where the output of the penultimate layer of the depth model is often represented as a feature of the corresponding picture object. The space composed of these feature representations is called a feature representation space.
Then, in the feature expression space, the information amount of each object is estimated according to a local information calculation method, and a labeling budget is allocated according to a global information budget allocation method. Based on the budget, a batch of picture objects with high information content are selected, and marks of the picture objects are collected. And updating the marked picture object set and the unmarked picture object set. Meanwhile, the depth model is retrained by using the marked picture object set, and the feature representation of the picture object is re-extracted by using the new model. These steps are iterated in turn to specify the number of rounds. And the model of the last round is the final depth model.
And finally, in the prediction stage, the user inputs the picture object to be tested into the depth model obtained by training, and the depth model returns the prediction result to the user.
Has the advantages that: compared with the prior art, the method and the device have the advantages that the local information and the global information in the characteristic representation space are combined, the local information of the picture object is considered to avoid selecting a redundant picture, the budget is allocated according to needs through the global information of the characteristic representation space, the utilization efficiency of the picture data marks is improved, and the marking cost is reduced.
Drawings
FIG. 1 is a flow chart of the present invention;
FIG. 2 is a flow chart of a local information computation method in the present invention;
FIG. 3 is a flowchart of a global information budget allocation method according to the present invention.
Detailed Description
The present invention is further illustrated by the following examples, which are intended to be purely exemplary and are not intended to limit the scope of the invention, as various equivalent modifications of the invention will occur to those skilled in the art upon reading the present disclosure and fall within the scope of the appended claims.
As shown in fig. 1, the method for selecting and labeling a picture based on local and global information includes the following steps:
wherein g (x; Θ) ═ g (x; Θ) 1 ,…,g(x;Θ) C );
105, aggregating the unmarked data into C clusters in the feature expression space according to the global information budget allocation method, and allocating budgets in different clusters (B) 1 ,…,B C ) In which B is j Budgets for the markers assigned to the jth cluster. As shown in fig. 3, the specific steps are as follows:
Wherein ∑ j α j τ is a temperature parameter used to adjust the degree of smoothing of the Gibbs distribution;
and step 108, using the model obtained by the training of the T-th round as a final model. And for the object to be measured, outputting the mark predicted by the model.
Claims (1)
1. A picture selection labeling method based on local and global information is characterized by comprising the following contents:
firstly, establishing a picture object library; then randomly selecting a part of picture objects from a picture object library, acquiring marks of the picture objects, and forming an initial training set; setting the structure of a depth model, the number of picture objects selected in each round and the total number of iteration rounds;
secondly, training a deep learning model based on a training set; converting the picture objects in the picture object library into feature representation by using the depth model, namely extracting the features of the pictures in the picture object library; the space composed by the feature representation is called a feature representation space;
then, in the feature representation space, estimating the information quantity of each object according to a local information calculation method, and allocating a labeling budget according to a global information budget allocation method; based on the budget, selecting a batch of picture objects with high information quantity, and collecting marks of the picture objects; updating a marked picture object set and an unmarked picture object set; meanwhile, retraining the depth model by using the marked picture object set, and extracting the feature representation of the picture object by using the new model; iteration is carried out to designate the number of rounds; the model of the last round is the final depth model;
finally, in a prediction stage, the user inputs the picture object to be tested into a depth model obtained by training, and the depth model returns a prediction result to the user;
recording the category number of data in the picture object library as C;representing a set of tagged picture objects,a set of representative unmarked picture objects; the selected depth model is denoted as f (·; Θ), whereIn order to be the parameters of the model,selecting the number B of samples selected in each round and the total number T of iteration rounds by a user as a full-connection layer parameter of the model and theta as other parameters in the model; using tagged picture objectsTraining a depth model, wherein the current round number t is 1; inputting unmarked picture objects into a depth model, and extracting characteristic representation r of the picture objects according to the depth model θ (x) Outputs f (x; theta) with the softmax layer;
the method for calculating the object information quantity by using the probability smoothing and the local information comprises the following specific steps:
step 1041, selecting a range E of the local neighbor area;
step 1042, for the unmarked picture object x, the softmax layer output is f (x; Θ) ═ p 1 ,…,p C ),Labelling of model f (x; Θ) predictionsProbability smoothing is performed as follows:
wherein g (x; Θ) ═ g (x; Θ) 1 ,…,g(x;Θ) C );
Step 1043, for unmarked picture objectAndcomputing information volume based on smoothed probabilities
Step 1044, recording the neighboring area of the picture object x asWherein r is θ (x) For the feature representation of the picture object x, the information content of the picture object x is
grouping the unmarked data into C clusters in the feature representation space according to the global information budget allocation method, and allocating budgets in different clusters (B) 1 ,…,B C ) In which B is j The specific steps for the marking budget allocated to the jth cluster are as follows:
step 1051, selecting a temperature parameter tau of Gibbs distribution by a user;
step 1052, using the kmeans + + method to group the feature representations of the unmarked picture objects into C clusters, the picture objects in the jth cluster are grouped into a set
Step 1053, estimating the performance of the model on different clusters, and recording the performance as gamma of the model on the jth cluster j
Step 1054, according to γ j Constructing a budget of Gibbs distribution α ═ (α) 1 ,…,α C )
Tau is a temperature parameter used for adjusting the smoothness degree of Gibbs distribution;
step 1055, sampling for B times according to the Gibbs distribution alpha to obtain the budget (B) allocated in each cluster 1 ,…,B C ) And output, wherein ∑ j B j B is the total marking budget.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202110399472.9A CN113269226B (en) | 2021-04-14 | 2021-04-14 | Picture selection labeling method based on local and global information |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202110399472.9A CN113269226B (en) | 2021-04-14 | 2021-04-14 | Picture selection labeling method based on local and global information |
Publications (2)
Publication Number | Publication Date |
---|---|
CN113269226A CN113269226A (en) | 2021-08-17 |
CN113269226B true CN113269226B (en) | 2022-09-23 |
Family
ID=77229077
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN202110399472.9A Active CN113269226B (en) | 2021-04-14 | 2021-04-14 | Picture selection labeling method based on local and global information |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN113269226B (en) |
Family Cites Families (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US8452087B2 (en) * | 2009-09-30 | 2013-05-28 | Microsoft Corporation | Image selection techniques |
CN106934055B (en) * | 2017-03-20 | 2020-05-19 | 南京大学 | Semi-supervised webpage automatic classification method based on insufficient modal information |
US11003892B2 (en) * | 2018-11-09 | 2021-05-11 | Sap Se | Landmark-free face attribute prediction |
CN111177384B (en) * | 2019-12-25 | 2023-01-20 | 南京理工大学 | Multi-mark Chinese emotion marking method based on global and local mark correlation |
CN112434736A (en) * | 2020-11-24 | 2021-03-02 | 成都潜在人工智能科技有限公司 | Deep active learning text classification method based on pre-training model |
-
2021
- 2021-04-14 CN CN202110399472.9A patent/CN113269226B/en active Active
Also Published As
Publication number | Publication date |
---|---|
CN113269226A (en) | 2021-08-17 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN111191732B (en) | Target detection method based on full-automatic learning | |
CN108985334B (en) | General object detection system and method for improving active learning based on self-supervision process | |
CN110245709B (en) | 3D point cloud data semantic segmentation method based on deep learning and self-attention | |
CN114067160B (en) | Small sample remote sensing image scene classification method based on embedded smooth graph neural network | |
CN112347970B (en) | Remote sensing image ground object identification method based on graph convolution neural network | |
CN109902761B (en) | Fishing situation prediction method based on marine environment factor fusion and deep learning | |
CN113223042B (en) | Intelligent acquisition method and equipment for remote sensing image deep learning sample | |
CN110738132B (en) | Target detection quality blind evaluation method with discriminant perception capability | |
CN113111716B (en) | Remote sensing image semiautomatic labeling method and device based on deep learning | |
CN116403058B (en) | Remote sensing cross-scene multispectral laser radar point cloud classification method | |
CN115292532B (en) | Remote sensing image domain adaptive retrieval method based on pseudo tag consistency learning | |
CN115471739A (en) | Cross-domain remote sensing scene classification and retrieval method based on self-supervision contrast learning | |
CN114863091A (en) | Target detection training method based on pseudo label | |
CN111239137B (en) | Grain quality detection method based on transfer learning and adaptive deep convolution neural network | |
CN110245723A (en) | A kind of safe and reliable image classification semi-supervised learning method and device | |
JP2009259109A (en) | Device, program and method for labeling, and recording medium recording labeling program | |
CN114579794A (en) | Multi-scale fusion landmark image retrieval method and system based on feature consistency suggestion | |
CN117572457A (en) | Cross-scene multispectral point cloud classification method based on pseudo tag learning | |
CN112528058B (en) | Fine-grained image classification method based on image attribute active learning | |
CN113869418A (en) | Small sample ship target identification method based on global attention relationship network | |
CN113034511A (en) | Rural building identification algorithm based on high-resolution remote sensing image and deep learning | |
CN113269226B (en) | Picture selection labeling method based on local and global information | |
Sun et al. | Automatic building age prediction from street view images | |
CN111783788B (en) | Multi-label classification method facing label noise | |
CN116012840B (en) | Three-dimensional point cloud semantic segmentation labeling method based on active learning and semi-supervision |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
GR01 | Patent grant | ||
GR01 | Patent grant |