CN113128472B - Multi-label labeling method based on intelligent collaborative learning - Google Patents

Multi-label labeling method based on intelligent collaborative learning Download PDF

Info

Publication number
CN113128472B
CN113128472B CN202110534862.2A CN202110534862A CN113128472B CN 113128472 B CN113128472 B CN 113128472B CN 202110534862 A CN202110534862 A CN 202110534862A CN 113128472 B CN113128472 B CN 113128472B
Authority
CN
China
Prior art keywords
label
image
training set
vectors
training
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN202110534862.2A
Other languages
Chinese (zh)
Other versions
CN113128472A (en
Inventor
赵海英
贺延钊
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Beijing University of Posts and Telecommunications
Original Assignee
Beijing University of Posts and Telecommunications
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Beijing University of Posts and Telecommunications filed Critical Beijing University of Posts and Telecommunications
Priority to CN202110534862.2A priority Critical patent/CN113128472B/en
Publication of CN113128472A publication Critical patent/CN113128472A/en
Application granted granted Critical
Publication of CN113128472B publication Critical patent/CN113128472B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V20/00Scenes; Scene-specific elements
    • G06V20/20Scenes; Scene-specific elements in augmented reality scenes
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/21Design or setup of recognition systems or techniques; Extraction of features in feature space; Blind source separation
    • G06F18/213Feature extraction, e.g. by transforming the feature space; Summarisation; Mappings, e.g. subspace methods
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/21Design or setup of recognition systems or techniques; Extraction of features in feature space; Blind source separation
    • G06F18/214Generating training patterns; Bootstrap methods, e.g. bagging or boosting
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/24Classification techniques
    • G06F18/243Classification techniques relating to the number of classes
    • G06F18/24323Tree-organised classifiers
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N20/00Machine learning
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/40Extraction of image or video features

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Data Mining & Analysis (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Artificial Intelligence (AREA)
  • Evolutionary Computation (AREA)
  • General Engineering & Computer Science (AREA)
  • Evolutionary Biology (AREA)
  • Bioinformatics & Computational Biology (AREA)
  • Bioinformatics & Cheminformatics (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Multimedia (AREA)
  • Software Systems (AREA)
  • Medical Informatics (AREA)
  • Computing Systems (AREA)
  • Mathematical Physics (AREA)
  • Image Analysis (AREA)

Abstract

The invention relates to a multi-label labeling algorithm taking an improved label embedding method as a core, which comprises the following steps: constructing a national costume pattern label system and a data set, forming a set of label system related to the national costume pattern by analyzing the characteristics and the content of the national costume pattern, and constructing a set of national costume pattern data set on the basis of the set of label system; performing feature extraction on the national costume pattern image by using a deep learning method; expressing the relation among the labels by constructing a label network, and optimizing the weight value representation of the label network; compressing the label network by a Laplace characteristic dimension reduction method and reserving the structure of the original label network; and automatically labeling the national costume patterns by combining the ensemble learning and the multi-label classification learner.

Description

Multi-label labeling method based on intelligent collaborative learning
Technical Field
The invention relates to the technical field of machine learning, in particular to a multi-label labeling method based on intelligent collaborative learning.
Background
The national dress pattern field, a certain image relates to a plurality of aspects of the cultural field, including fields such as name, meaning, configuration, color and the like. For such image data with ambiguity, all contents of the image data cannot be completely explained by using a single label, while multi-label labeling can accurately and completely describe cultural contents of national dress patterns, and a user can acquire rich cultural contents of the national dress patterns in an automatic labeling mode.
The multi-label labeling is a popular research direction in the field of current machine learning, and the multi-label labeling method based on label embedding provided by researchers is suitable for numerous and complicated national dress pattern data, so that the problems of overlarge label space and unbalanced sample label categories are solved well.
However, the label embedding method has some problems in terms of label space dimension reduction, and particularly when the number of labels is too large, problems of inaccurate dimension reduction extraction, too low speed and the like are generated, so that the pre-operation of label embedding and optimization of the pre-operation need to be performed.
Disclosure of Invention
The invention aims to overcome the defects in the prior art and provide a multi-label labeling method based on intelligent collaborative learning. The national costume pattern culture content can be automatically marked.
The invention provides a multi-label labeling method based on intelligent collaborative learning, which is characterized by comprising the following steps: the method comprises the following steps:
step 1, acquiring a national costume pattern image, and performing manual marking to form a set of complete national costume pattern data and corresponding label contents, wherein each label content comprises a plurality of independent labels;
step 2, dividing the ethnic clothing pattern data and the corresponding label content into a training set and a verification set;
step 3, training the algorithm model in the training set, and specifically comprising the following steps:
3.1, extracting characteristic vectors of the ethnic clothing pattern data in the training set through a residual error network, wherein all the characteristic vectors form an image characteristic space;
3.2, taking each independent label of the training set as a node in the network structure, forming an edge between two labels when the two labels simultaneously appear in a sample example, and forming the weight of the edge through the co-occurrence frequency of the labels to complete the construction of the label network, wherein the weight calculation formula of each edge is as follows:
Figure BDA0003069217210000021
Figure BDA0003069217210000022
w (i, j) represents the weight of the edge between label i and label j, wherein the numerator represents the ethnicity of label i and label j in the training setThe number of times of simultaneous occurrence in the label content of the clothing pattern, the denominator represents the total number of the occurrence of the label i and the label j in the label content of the national clothing pattern of the training set, y s,i 1 indicates the presence of the label i, y in the sample s s,j 1 indicates the presence of label j in sample s; w is a i ,w j Respectively representing weight vectors formed by the weights of edges taking the label i and the label j as end points in the label network,
Figure BDA0003069217210000023
c is a constant value, and c is more than 0 and less than 1;
3.3, reducing the dimension of the label network through Laplace feature mapping to obtain a label embedding space, and determining a label embedding vector corresponding to each image feature vector through a polymerization method;
3.4, training the decision tree integration regression model by using a training set, wherein input data are image feature vectors of each image in the training set and corresponding label embedded vectors;
step 4, using the verification set to label multiple labels
4.1, extracting and verifying feature vectors of the centralized ethnic clothing pattern data through a residual error network, wherein all the feature vectors form an image feature space;
4.2, performing regression on the image feature space by using the trained set-pair decision tree integrated regression model, and outputting a label embedded vector corresponding to the image feature vector;
and 4.3, performing multi-label classification on the label embedded vector obtained after regression by using a classifier to obtain a final prediction result of the verification set, namely an image labeling result.
The innovation of the invention is that:
1. and constructing a label network to express the relation between the labels, and optimizing the weight value of the label network through a kernel function.
2. The deep learning model, the embedding method, the ensemble learning method and the multi-label classification method are combined to form a new multi-label labeling algorithm, and a better labeling effect is achieved.
Drawings
In order to more clearly illustrate the embodiments of the present invention or the technical solutions in the prior art, the drawings used in the description of the embodiments or the prior art will be briefly described below, it is obvious that the drawings in the following description are only some embodiments of the present invention, and for those skilled in the art, other drawings can be obtained according to the drawings without creative efforts.
Fig. 1 is a flowchart of a multi-label labeling method based on intelligent collaborative learning according to an embodiment of the present invention.
Detailed Description
The technical solutions in the embodiments of the present invention will be clearly and completely described below with reference to the accompanying drawings in the embodiments of the present invention, and it is obvious that the described embodiments are only a part of the embodiments of the present invention, and not all of the embodiments. All other embodiments, which can be derived by a person skilled in the art from the embodiments given herein without making any creative effort, shall fall within the protection scope of the present invention.
Specifically, the multi-label labeling method based on intelligent collaborative learning provided by the embodiment includes the following steps:
s110, acquiring national clothing pattern images by means of digital scanning of books, and manually marking the images through marking software to form a set of complete national clothing pattern data and corresponding label contents, wherein each label content comprises a plurality of independent labels.
S120, processing the images and the label contents into a computer 0/1 adjacency matrix by using a data set construction program, and forming a training set and a verification set which are suitable for an algorithm model. In this example, the data set is randomly divided into two parts, a training set and a validation set, in a 9:1 ratio.
S130, training the algorithm model in a training set, and specifically comprising the following steps:
a1, extracting the characteristic vectors of the national costume pattern data in the training set through a residual error network, wherein all the characteristic vectors form an image characteristic space.
In the step, an objective function BCEWithLoitsLoss which is suitable for multi-label classification is utilized to extract image characteristics of the Resnet model which is pre-trained by ImageNet through a ethnic clothing pattern data set.
The specific process is as follows:
the input picture 224 x 224 sized area was randomly chosen and the input picture samples were increased with horizontal flipping and standard color enhancement was used. And removing the original full connection layer in Resnet, reconstructing the full connection layer, and extracting the image characteristics of the ethnic clothing pattern data set by using the Resnet model pre-trained by ImageNet by using a target function BCEWithLoctissoss suitable for multi-label classification. The 2048 dimensional image features are exported from the Resnet fully connected layer onwards. BCELoss may calculate a binary loss between the predicted output and the true output as a criterion for single label classification. BCEwithLogitsLoss is a combination of a Sigmoid function and a BCELoss function. The calculation models used in this step are all existing models, see the paper: he K, Zhang X, Ren S, et al. deep reactive Learning for Image registration [ C ]. IEEE Conference on Computer Vision and Pattern Registration (CVPR). IEEE Computer Society,2016 (see).
a2, taking each label in the training set as a node in the network structure, forming an edge between two labels when the two labels appear in a sample example at the same time, and forming the weight of the edge through the co-occurrence frequency of the labels to complete the construction of the label network, wherein the weight calculation formula of each edge is as follows:
Figure BDA0003069217210000041
Figure BDA0003069217210000051
w (i, j) represents the weight of the edge between the label i and the label j, wherein the numerator represents the number of times that the label i and the label j simultaneously appear in the label content of the ethnic clothing pattern of the training set, and the denominator represents the number of times that the label i and the label j simultaneously appear in the training setTotal number of occurrences in tag content of national dress pattern, y s,i 1 indicates the presence of the label i, y in the sample s s,j 1 indicates the presence of label j in sample s; w is a i ,w j Each weight vector represents a weight of an edge having a label i and a label j as end points in the label network, represents an approximation relationship between the label and another label, and considers the degree of association of the label from the viewpoint of a sample example. The incidence relation of the two labels can be further considered from the perspective of the labels by utilizing the kernel function, if the two labels coexist with more same labels, the two labels can be similar, otherwise, the two labels are more independent.
Figure BDA0003069217210000052
The weight value constructed by the kernel function has the characteristics that,
Figure BDA0003069217210000053
the larger the tag i is, the more similar the tag j is, and vice versa, the more independent. c is constant, and c is more than 0 and less than 1.
a3, reducing dimensions of the label network through Laplacian feature mapping to obtain a label embedding space, and determining a label embedding vector corresponding to each image feature vector through a polymerization method. In this step, the laplace feature dimension reduction method is a known method, and is described in the literature: belkin M, Niyogi P.Laplacian eigenmaps and spectral techniques for embedding and curing [ C]585-591, also known, see: for vector v i ,ξ(v i )=v i1 +v i2 +v i3 +...v in Mathematically, a vector aggregation operation (vector summation).
a4, selecting a more direct regression method to map the feature space to the label embedding space.
In this embodiment, the regression method adopts an integrated regression model for the decision tree, and the input data is the image feature vector and the corresponding label embedded vector of each image in the training set. The decision tree regression integration algorithm adopts the square average error as the judgment standard of the purity of the decision tree nodes, and the number N of the decision trees is 10.
Considering that the feature space dimension is high, the formed decision tree may have a deep depth and a poor generalization capability, the embodiment performs post-pruning on a single decision tree by using the CART pruning method. The decision tree integration regression model is an existing regression learning method, and is described in the literature: liaw A, Wiener M, classification and regression by random forest [ J ]. R news,2002,2(3):18-22.
Step 4, using the verification set to label multiple labels
b1, extracting the characteristic vectors of the ethnic clothing pattern data in the verification set through a residual error network, wherein all the characteristic vectors form an image characteristic space. This step can be referred to as step a1 and will not be described in detail herein.
b2, performing regression on the image feature space by using the trained set-pair decision tree integrated regression model, and outputting a label embedded vector corresponding to the image feature vector.
b3, performing multi-label classification on the label embedding vector obtained after regression by using a classifier to obtain a final prediction result of the verification set, namely an image labeling result.
In this example, the ML-KNN multi-label classification learner is used to perform multi-label classification. The ML-KNN multi-label classification learner firstly needs to use training set data for training, input data are label embedding vectors and original label vectors corresponding to image features in a training set, and the original label vectors are formed by independent labels contained in the content of each label. The original label vector and the regression result (label embedded vector) are used as the input of the ML-KNN multi-label classifier to train, so that the original characteristic input can be reserved to a certain extent, and the decoding deviation caused by regression error is reduced. The ML-KNN classifier belongs to the prior art category, see literature: zhang M L, Zhou Z H.ML-KNN A lazy learning approach to multi-label learning [ J ]. Pattern recognition,2007,40(7): 2038-.
Experiments were performed on a national costume pattern dataset. The smaller the value of One-error, Coverage, Ranking-loss and Hamming-Los indexes adopted by the evaluation indexes is, the better the performance of the algorithm on the indexes is, and the higher the value of Average-precision is, the better the performance of the algorithm on the indexes is.
The results of the experiment are as follows:
Figure BDA0003069217210000061
in addition to the above embodiments, the present invention may have other embodiments. All technical solutions formed by adopting equivalent substitutions or equivalent transformations fall within the protection scope of the claims of the present invention.

Claims (7)

1. A multi-label labeling method based on intelligent collaborative learning is characterized in that: the method comprises the following steps:
step 1, acquiring a national costume pattern image, and performing manual marking to form a set of complete national costume pattern data and corresponding label contents, wherein each label content comprises a plurality of independent labels;
step 2, dividing the ethnic clothing pattern data and the corresponding label content into a training set and a verification set;
step 3, training the algorithm model in the training set, and specifically comprising the following steps:
3.1, extracting characteristic vectors of the ethnic clothing pattern data in the training set through a residual error network, wherein all the characteristic vectors form an image characteristic space;
3.2, taking each independent label of the training set as a node in a network structure, forming an edge between two labels when the two labels simultaneously appear in a sample example, and forming the weight of the edge through the co-occurrence frequency of the labels to complete the construction of the label network, wherein the weight calculation formula of each edge is as follows:
Figure FDA0003069217200000011
Figure FDA0003069217200000012
w (i, j) represents the weight of the edge between the label i and the label j, wherein the numerator represents the number of times that the label i and the label j simultaneously appear in the label content of the national costume pattern of the training set, the denominator represents the total number of the label i and the label j appearing in the label content of the national costume pattern of the training set, y s,i 1 indicates the presence of the label i, y in the sample s s,j 1 indicates the presence of label j in sample s; w is a i ,w j Respectively representing weight vectors formed by the weights of edges taking the label i and the label j as end points in the label network,
Figure FDA0003069217200000013
c is a constant value, and c is more than 0 and less than 1;
3.3, reducing the dimension of the label network through Laplace feature mapping to obtain a label embedding space, and determining a label embedding vector corresponding to each image feature vector through a polymerization method;
3.4, training the decision tree integration regression model by using a training set, wherein input data are image feature vectors of each image in the training set and corresponding label embedded vectors;
step 4, using the verification set to label multiple labels
4.1, extracting and verifying feature vectors of the centralized ethnic clothing pattern data through a residual error network, wherein all the feature vectors form an image feature space;
4.2, performing regression on the image feature space by using the trained set-pair decision tree integrated regression model, and outputting a label embedded vector corresponding to the image feature vector;
and 4.3, performing multi-label classification on the label embedded vector obtained after regression by using a classifier to obtain a final prediction result of the verification set, namely an image labeling result.
2. The multi-label labeling method based on intelligent collaborative learning according to claim 1, characterized in that: in the step 1, after the book is scanned by using the scanner, the national clothing pattern is identified by manual visual observation through the image marking tool, and the corresponding marking file is obtained by manual marking.
3. The multi-label labeling method based on intelligent collaborative learning according to claim 1, characterized in that: in the step 1, the data set is randomly divided into a training set and a verification set according to a ratio of 9: 1.
4. The multi-label labeling method based on intelligent collaborative learning according to claim 1, characterized in that: in the step 3.1, the image feature extraction is carried out on the ethnic clothing pattern data set by the Resnet model after ImageNet pre-training by using the target function BCEWithLoitsLoss suitable for multi-label classification.
5. The multi-label labeling method based on intelligent collaborative learning according to claim 1, characterized in that: in step 3.4, the decision tree regression integration algorithm adopts the square mean error as the judgment standard of the node impure degree of the decision tree, and the number N of the decision trees is 10.
6. The multi-label labeling method based on intelligent collaborative learning according to claim 1, characterized in that: and 4.3, performing multi-label classification by using an ML-KNN multi-label classification learning device.
7. The multi-label labeling method based on intelligent collaborative learning according to claim 6, characterized in that: and 4.3, training the ML-KNN multi-label classification learning device by using the training set, wherein input data are label embedded vectors and original label vectors corresponding to the image features in the training set, and the original label vectors are formed by independent labels contained in the content of each label.
CN202110534862.2A 2021-05-17 2021-05-17 Multi-label labeling method based on intelligent collaborative learning Active CN113128472B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202110534862.2A CN113128472B (en) 2021-05-17 2021-05-17 Multi-label labeling method based on intelligent collaborative learning

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202110534862.2A CN113128472B (en) 2021-05-17 2021-05-17 Multi-label labeling method based on intelligent collaborative learning

Publications (2)

Publication Number Publication Date
CN113128472A CN113128472A (en) 2021-07-16
CN113128472B true CN113128472B (en) 2022-09-20

Family

ID=76782086

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202110534862.2A Active CN113128472B (en) 2021-05-17 2021-05-17 Multi-label labeling method based on intelligent collaborative learning

Country Status (1)

Country Link
CN (1) CN113128472B (en)

Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN104216993A (en) * 2014-09-10 2014-12-17 武汉科技大学 Tag-co-occurred tag clustering method
CN104881689A (en) * 2015-06-17 2015-09-02 苏州大学张家港工业技术研究院 Method and system for multi-label active learning classification
CN110163234A (en) * 2018-10-10 2019-08-23 腾讯科技(深圳)有限公司 A kind of model training method, device and storage medium

Patent Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN104216993A (en) * 2014-09-10 2014-12-17 武汉科技大学 Tag-co-occurred tag clustering method
CN104881689A (en) * 2015-06-17 2015-09-02 苏州大学张家港工业技术研究院 Method and system for multi-label active learning classification
CN110163234A (en) * 2018-10-10 2019-08-23 腾讯科技(深圳)有限公司 A kind of model training method, device and storage medium

Non-Patent Citations (2)

* Cited by examiner, † Cited by third party
Title
Double Attention for Multi-Label Image Classification;HAIYING ZHAO 等;《IEEE ACCESS》;20201214;第8卷;全文 *
基于多标签学习的卷积神经网络的图像标注方法;高耀东 等;《计算机应用》;20171231;全文 *

Also Published As

Publication number Publication date
CN113128472A (en) 2021-07-16

Similar Documents

Publication Publication Date Title
CN106599854B (en) Automatic facial expression recognition method based on multi-feature fusion
CN112131978B (en) Video classification method and device, electronic equipment and storage medium
CN107480620B (en) Remote sensing image automatic target identification method based on heterogeneous feature fusion
CN112734775A (en) Image annotation, image semantic segmentation and model training method and device
CN102054178A (en) Chinese painting image identifying method based on local semantic concept
CN109002755B (en) Age estimation model construction method and estimation method based on face image
CN110188708A (en) A kind of facial expression recognizing method based on convolutional neural networks
CN104850822B (en) Leaf identification method under simple background based on multi-feature fusion
CN105760858A (en) Pedestrian detection method and apparatus based on Haar-like intermediate layer filtering features
CN107622280B (en) Modularized processing mode image saliency detection method based on scene classification
CN109886161A (en) A kind of road traffic index identification method based on possibility cluster and convolutional neural networks
JP4553300B2 (en) Content identification device
CN110413825B (en) Street-clapping recommendation system oriented to fashion electronic commerce
CN105718932A (en) Colorful image classification method based on fruit fly optimization algorithm and smooth twinborn support vector machine and system thereof
CN110852327A (en) Image processing method, image processing device, electronic equipment and storage medium
CN110046544A (en) Digital gesture identification method based on convolutional neural networks
CN110599463A (en) Tongue image detection and positioning algorithm based on lightweight cascade neural network
CN113705579A (en) Automatic image annotation method driven by visual saliency
CN108647703B (en) Saliency-based classification image library type judgment method
CN114187520A (en) Building extraction model and application method thereof
CN106650798A (en) Indoor scene recognition method combining deep learning and sparse representation
CN112001438A (en) Multi-mode data clustering method for automatically selecting clustering number
CN114782979A (en) Training method and device for pedestrian re-recognition model, storage medium and terminal
CN109658523A (en) The method for realizing each function operation instruction of vehicle using the application of AR augmented reality
CN113128472B (en) Multi-label labeling method based on intelligent collaborative learning

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant