CN110046666B - Mass picture labeling method - Google Patents
Mass picture labeling method Download PDFInfo
- Publication number
- CN110046666B CN110046666B CN201910312598.0A CN201910312598A CN110046666B CN 110046666 B CN110046666 B CN 110046666B CN 201910312598 A CN201910312598 A CN 201910312598A CN 110046666 B CN110046666 B CN 110046666B
- Authority
- CN
- China
- Prior art keywords
- labeling
- model
- prediction
- pictures
- initial model
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Active
Links
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F18/00—Pattern recognition
- G06F18/20—Analysing
- G06F18/24—Classification techniques
- G06F18/241—Classification techniques relating to the classification model, e.g. parametric or non-parametric approaches
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F18/00—Pattern recognition
- G06F18/20—Analysing
- G06F18/24—Classification techniques
- G06F18/241—Classification techniques relating to the classification model, e.g. parametric or non-parametric approaches
- G06F18/2411—Classification techniques relating to the classification model, e.g. parametric or non-parametric approaches based on the proximity to a decision surface, e.g. support vector machines
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F18/00—Pattern recognition
- G06F18/20—Analysing
- G06F18/24—Classification techniques
- G06F18/241—Classification techniques relating to the classification model, e.g. parametric or non-parametric approaches
- G06F18/2413—Classification techniques relating to the classification model, e.g. parametric or non-parametric approaches based on distances to training or reference patterns
- G06F18/24147—Distances to closest patterns, e.g. nearest neighbour classification
Landscapes
- Engineering & Computer Science (AREA)
- Data Mining & Analysis (AREA)
- Theoretical Computer Science (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Bioinformatics & Cheminformatics (AREA)
- Bioinformatics & Computational Biology (AREA)
- Artificial Intelligence (AREA)
- Evolutionary Biology (AREA)
- Evolutionary Computation (AREA)
- Physics & Mathematics (AREA)
- General Engineering & Computer Science (AREA)
- General Physics & Mathematics (AREA)
- Life Sciences & Earth Sciences (AREA)
- Image Analysis (AREA)
Abstract
The invention provides a method for labeling a mass of pictures, which comprises the following steps: step 1, establishing an initial model according to labeling information of a plurality of pictures; step 2, predicting the unmarked picture by using the initial model; step 3, carrying out confidence coefficient analysis on the prediction result in the step 3, thereby selecting reliable prediction; step 4, replacing the prediction result of reliable prediction with labeling information; and 5, updating the initial model by using the marking information obtained in the step 4, and returning to the step 1 to repeat iteration. The method can quickly and effectively label massive pictures so as to solve the problem of the real bottleneck of data labeling in deep learning.
Description
Technical Field
The invention relates to the technical field of picture marking, in particular to a massive picture marking method.
Background
Deep learning is currently finding increasingly widespread use in both academic and industrial areas. However, the deep learning model has three resource limitation problems: the data size of model training is large, and if the model training depends on network transmission, the network bandwidth becomes a bottleneck; a large amount of computing resources, many deep learning model parameters and complex operation process result in huge computation amount; a large amount of annotation data is required to train complex models.
Annotating data is a tedious process and is typically performed point-by-point with a brush aimed at a target, as shown in fig. 1, where (a) is the original picture and (b) is the annotated picture. Therefore, the existing mode wastes time and labor, and in the labeling process, the labeling effect is often not ideal due to artificial factors and differences.
The mass pictures are marked by consuming a large amount of manpower and material resources, but a method for quickly and effectively marking the mass pictures does not exist at present.
Disclosure of Invention
The invention provides a method for labeling a mass of pictures, which aims to solve at least one technical problem.
In order to solve the above problem, as an aspect of the present invention, a method for labeling a large number of pictures is provided, including: step 1, establishing an initial model according to labeling information of a plurality of pictures; step 2, predicting the unmarked picture by using the initial model; step 3, carrying out confidence coefficient analysis on the prediction result in the step 3, thereby selecting reliable prediction; step 4, replacing the prediction result of reliable prediction with labeling information; and 5, updating the initial model by using the marking information obtained in the step 4, and returning to the step 1 to repeat iteration.
Preferably, the labeling information in step 1 is obtained by manual labeling.
Preferably, step 3 comprises: and setting a confidence coefficient function for evaluating the reliability of the prediction, wherein the function utilizes the original image information and the output vector of the softmax layer of the model to evaluate the prediction result.
Preferably, step 3 comprises: and judging the matching degree of each point in the prediction result by utilizing the pixels in the 4x4 field around the point to calculate the information such as the variance and the gradient of the contrast and the output vector of the softmax layer of the position model, and if the matching degree is matched according to the preset confidence threshold, the prediction is reliable.
Preferably, the anastomotic match is characterized by a neural network or a support vector machine.
Preferably, the updating the initial model by using the label information obtained in step 4 includes: let the initial model be θ t The last used model is θ' t-1 Then the new model is θ = α θ' t-1 +(1-α)θ t Where α is an exponential smoothing coefficient.
By adopting the technical scheme, the method can quickly and effectively label massive pictures to solve the real bottleneck problem of labeling data in deep learning, firstly labels a small number of pictures, trains a deep learning model, then predicts the remaining unmarked pictures by using the model, evaluates the prediction result, converts accurate prediction into labeling, and then optimizes the model, so that the accuracy of the model is higher and the labeling accuracy is higher and higher in circulating iteration.
Drawings
FIG. 1 is a schematic diagram illustrating a prior art labeling of a picture;
FIG. 2 schematically illustrates a flow chart of the present invention;
fig. 3 schematically shows a schematic diagram of selecting reliable prediction results according to confidence.
Detailed Description
Embodiments of the invention will be described in detail below with reference to the drawings, but the invention can be implemented in many different ways as defined and covered by the claims.
An initial model is assumed to exist, and if not, a small number of pictures can be artificially marked to train the initial model. For convenience of the following explanation, we take defect detection as an example, as shown in fig. 1.
Step 1: the unlabeled data is predicted with the initial model.
Step 2: and carrying out confidence degree analysis on the predicted result. Since the model performance of step 1 is not ideal at this time, the predicted result is good or bad, and therefore, the purpose of performing confidence analysis on the predicted result is to select reliable prediction.
Confidence functionTo evaluate the reliability of model predictions. This function evaluates the curve part in (b) of fig. 1 using the original image information I and the output vector of the softmax layer of the model. Specifically, for each point on the curve, we calculate the variance, gradient, etc. of the contrast using the pixels in the 4 × 4 domain around the point, and at this position, the softmax layer output of the model represents the predicted defect strength. These two pieces of information, if matched with a high degree of agreement, indicate that the prediction is reliable. The mapping relationship can be characterized by a simple neural network or a support vector machine (support vector machine).
And step 3: and selecting a key sample. Step 2, confidence analysis is carried out on all predictions, and accordingly, reliable predictions can be selected. We can set a confidence threshold T, when the threshold is greater than T, indicating that the prediction is reliable, and leave, otherwise leave, as shown in fig. 3, the prediction confidence in the upper box in (b) is low, and leave in the lower box.
And 4, step 4: and converting the prediction result into the labeling information. The prediction result with high confidence and the artificial label are consistent on a large probability and can be converted into label information.
And 5: and updating the model. By using more labeled information, we can train out a model theta t The last used model is θ' t-1 Then the new model is θ = α θ' t-1 +(1-α)θ t Where α is an exponential smoothing coefficient. The model has better prediction performance than the previous model.
With the updated model, we return to step 1 to predict the unmarked part and update the existing prediction. The performance of the model is better and better by repeating the iteration, and the quality of the label is gradually improved.
The above description is only a preferred embodiment of the present invention and is not intended to limit the present invention, and various modifications and changes may be made by those skilled in the art. Any modification, equivalent replacement, or improvement made within the spirit and principle of the present invention should be included in the protection scope of the present invention.
Claims (2)
1. A method for labeling massive pictures is characterized by comprising the following steps:
step 1, establishing an initial model according to labeling information of a plurality of pictures, wherein the labeling information is obtained in a manual labeling mode;
step 2, directly utilizing the initial model to predict the unmarked picture;
step 3, performing confidence analysis on the prediction result in the step 2, thereby selecting reliable predictions, which comprises the following steps: setting a confidence function for evaluating the reliability of prediction, wherein the function utilizes original image information and an output vector of a softmax layer of a model to evaluate a prediction result, judging the matching degree of each point in the prediction result between the variance and the gradient of the contrast ratio of each point and the output vector of the softmax layer of the model by utilizing pixels in the 4x4 field around each point, and if the point is matched according to a preset confidence threshold value, the prediction is reliable;
step 4, replacing the prediction result of reliable prediction with labeling information;
and 5, updating the initial model by using the marking information obtained in the step 4, wherein the method comprises the following steps: setting an initial model asThe last used model isThen the new model isWhereinIs an exponential smoothing coefficient; then, the iteration is repeated by returning to the step 1.
2. The method for labeling the massive pictures according to claim 1, wherein the matching is characterized by being characterized by a neural network or a support vector machine.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201910312598.0A CN110046666B (en) | 2019-04-18 | 2019-04-18 | Mass picture labeling method |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201910312598.0A CN110046666B (en) | 2019-04-18 | 2019-04-18 | Mass picture labeling method |
Publications (2)
Publication Number | Publication Date |
---|---|
CN110046666A CN110046666A (en) | 2019-07-23 |
CN110046666B true CN110046666B (en) | 2022-12-02 |
Family
ID=67277763
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN201910312598.0A Active CN110046666B (en) | 2019-04-18 | 2019-04-18 | Mass picture labeling method |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN110046666B (en) |
Families Citing this family (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN112699908B (en) * | 2019-10-23 | 2022-08-05 | 武汉斗鱼鱼乐网络科技有限公司 | Method for labeling picture, electronic terminal, computer readable storage medium and equipment |
CN112101156B (en) * | 2020-09-02 | 2024-08-27 | 杭州海康威视数字技术股份有限公司 | Target identification method and device and electronic equipment |
Family Cites Families (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP4480156B2 (en) * | 2005-02-02 | 2010-06-16 | キヤノン株式会社 | Image processing apparatus and method |
CN106096627A (en) * | 2016-05-31 | 2016-11-09 | 河海大学 | The Polarimetric SAR Image semisupervised classification method that considering feature optimizes |
CN107977667B (en) * | 2016-10-21 | 2019-02-19 | 西安电子科技大学 | SAR target discrimination method based on semi-supervised coorinated training |
CN106897738B (en) * | 2017-01-22 | 2019-07-16 | 华南理工大学 | A kind of pedestrian detection method based on semi-supervised learning |
US10210403B2 (en) * | 2017-04-24 | 2019-02-19 | Here Global B.V. | Method and apparatus for pixel based lane prediction |
CN108764281A (en) * | 2018-04-18 | 2018-11-06 | 华南理工大学 | A kind of image classification method learning across task depth network based on semi-supervised step certainly |
-
2019
- 2019-04-18 CN CN201910312598.0A patent/CN110046666B/en active Active
Non-Patent Citations (1)
Title |
---|
Restoration algorithm of damaged mark image based on texture synthesis;Min Qi et al.;《2015 International Conference on Computer and Computational Sciences (ICCCS)》;20151221;第1-4页 * |
Also Published As
Publication number | Publication date |
---|---|
CN110046666A (en) | 2019-07-23 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN114241282B (en) | Knowledge distillation-based edge equipment scene recognition method and device | |
CN109741332B (en) | Man-machine cooperative image segmentation and annotation method | |
US12061991B2 (en) | Transfer learning with machine learning systems | |
CN110046666B (en) | Mass picture labeling method | |
CN111581116B (en) | Cross-project software defect prediction method based on hierarchical data screening | |
CN112861959B (en) | Automatic labeling method for target detection image | |
CN112052818A (en) | Unsupervised domain adaptive pedestrian detection method, unsupervised domain adaptive pedestrian detection system and storage medium | |
CN109919302B (en) | Training method and device for neural network of image | |
CN111160432A (en) | Automatic classification method and system for panel production defects | |
CN114511077A (en) | Training point cloud processing neural networks using pseudo-element based data augmentation | |
CN115277587B (en) | Network traffic identification method, device, equipment and medium | |
CN115205727A (en) | Experiment intelligent scoring method and system based on unsupervised learning | |
CN114782752B (en) | Small sample image integrated classification method and device based on self-training | |
CN117876383B (en) | Yolov5 l-based highway surface strip-shaped crack detection method | |
CN111125389A (en) | Data classification cleaning system and cleaning method based on dynamic progressive sampling | |
CN113223011B (en) | Small sample image segmentation method based on guide network and full-connection conditional random field | |
CN114186090A (en) | Intelligent quality inspection method and system for image annotation data | |
CN113554068A (en) | Semi-automatic labeling method and device for instance segmentation data set and readable medium | |
CN117703023A (en) | Seamless paving method for rock plates/tiles | |
CN116823193B (en) | Intelligent manufacturing flow management system based on big data | |
CN110929792B (en) | Image labeling method, device, electronic equipment and storage medium | |
CN112164040A (en) | Steel surface defect identification method based on semi-supervised deep learning algorithm | |
Sirhan et al. | Multilabel CNN model for asphalt distress classification | |
CN113111729A (en) | Training method, recognition method, system, device and medium of personnel recognition model | |
KR102474170B1 (en) | Apparatus and method for generating synthetic data for model training |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
GR01 | Patent grant | ||
GR01 | Patent grant |